/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 564 * @attr_mask: the mask of attributes to modify 570 int attr_mask, struct ib_udata *udata) qib_modify_qp() 584 cur_state = attr_mask & IB_QP_CUR_STATE ? qib_modify_qp() 586 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; qib_modify_qp() 589 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) qib_modify_qp() 592 if (attr_mask & IB_QP_AV) { qib_modify_qp() 599 if (attr_mask & IB_QP_ALT_PATH) { qib_modify_qp() 608 if (attr_mask & IB_QP_PKEY_INDEX) qib_modify_qp() 612 if (attr_mask & IB_QP_MIN_RNR_TIMER) qib_modify_qp() 616 if (attr_mask & IB_QP_PORT) qib_modify_qp() 623 if (attr_mask & IB_QP_DEST_QPN) qib_modify_qp() 627 if (attr_mask & IB_QP_RETRY_CNT) qib_modify_qp() 631 if (attr_mask & IB_QP_RNR_RETRY) qib_modify_qp() 643 if (attr_mask & IB_QP_PATH_MTU) { qib_modify_qp() 674 if (attr_mask & IB_QP_PATH_MIG_STATE) { qib_modify_qp() 691 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) qib_modify_qp() 749 if (attr_mask & IB_QP_PKEY_INDEX) qib_modify_qp() 752 if (attr_mask & IB_QP_PORT) qib_modify_qp() 755 if (attr_mask & IB_QP_DEST_QPN) qib_modify_qp() 758 if (attr_mask & IB_QP_SQ_PSN) { qib_modify_qp() 766 if (attr_mask & IB_QP_RQ_PSN) qib_modify_qp() 769 if (attr_mask & IB_QP_ACCESS_FLAGS) qib_modify_qp() 772 if (attr_mask & IB_QP_AV) { qib_modify_qp() 777 if (attr_mask & IB_QP_ALT_PATH) { qib_modify_qp() 782 if (attr_mask & IB_QP_PATH_MIG_STATE) { qib_modify_qp() 791 if (attr_mask & IB_QP_PATH_MTU) { qib_modify_qp() 796 if (attr_mask & IB_QP_RETRY_CNT) { qib_modify_qp() 801 if (attr_mask & IB_QP_RNR_RETRY) { qib_modify_qp() 806 if (attr_mask & IB_QP_MIN_RNR_TIMER) qib_modify_qp() 809 if (attr_mask & IB_QP_TIMEOUT) { qib_modify_qp() 816 if (attr_mask & IB_QP_QKEY) qib_modify_qp() 819 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) qib_modify_qp() 822 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) qib_modify_qp() 856 int attr_mask, struct ib_qp_init_attr *init_attr) qib_query_qp() 569 qib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) qib_modify_qp() argument 855 qib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) qib_query_qp() argument
|
H A D | qib_srq.c | 208 * @attr_mask: indicates which attributes to modify 212 enum ib_srq_attr_mask attr_mask, qib_modify_srq() 219 if (attr_mask & IB_SRQ_MAX_WR) { qib_modify_srq() 226 ((attr_mask & IB_SRQ_LIMIT) ? qib_modify_srq() 299 if (attr_mask & IB_SRQ_LIMIT) qib_modify_srq() 333 } else if (attr_mask & IB_SRQ_LIMIT) { qib_modify_srq() 211 qib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) qib_modify_srq() argument
|
H A D | qib_verbs.h | 919 int attr_mask, struct ib_udata *udata); 922 int attr_mask, struct ib_qp_init_attr *init_attr); 997 enum ib_srq_attr_mask attr_mask,
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_verbs.c | 82 int attr_mask; ipoib_init_qp() local 91 attr_mask = ipoib_init_qp() 96 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp() 104 attr_mask &= ~IB_QP_PORT; ipoib_init_qp() 105 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp() 113 attr_mask |= IB_QP_SQ_PSN; ipoib_init_qp() 114 attr_mask &= ~IB_QP_PKEY_INDEX; ipoib_init_qp() 115 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 445 * @attr_mask: the mask of attributes to modify 451 int attr_mask, struct ib_udata *udata) ipath_modify_qp() 461 cur_state = attr_mask & IB_QP_CUR_STATE ? ipath_modify_qp() 463 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; ipath_modify_qp() 466 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) ipath_modify_qp() 469 if (attr_mask & IB_QP_AV) { ipath_modify_qp() 479 if (attr_mask & IB_QP_PKEY_INDEX) ipath_modify_qp() 483 if (attr_mask & IB_QP_MIN_RNR_TIMER) ipath_modify_qp() 487 if (attr_mask & IB_QP_PORT) ipath_modify_qp() 496 if ((attr_mask & IB_QP_PATH_MTU) && ipath_modify_qp() 501 if (attr_mask & IB_QP_PATH_MIG_STATE) ipath_modify_qp() 506 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) ipath_modify_qp() 550 if (attr_mask & IB_QP_PKEY_INDEX) ipath_modify_qp() 553 if (attr_mask & IB_QP_DEST_QPN) ipath_modify_qp() 556 if (attr_mask & IB_QP_SQ_PSN) { ipath_modify_qp() 561 if (attr_mask & IB_QP_RQ_PSN) ipath_modify_qp() 564 if (attr_mask & IB_QP_ACCESS_FLAGS) ipath_modify_qp() 567 if (attr_mask & IB_QP_AV) { ipath_modify_qp() 572 if (attr_mask & IB_QP_PATH_MTU) ipath_modify_qp() 575 if (attr_mask & IB_QP_RETRY_CNT) ipath_modify_qp() 578 if (attr_mask & IB_QP_RNR_RETRY) { ipath_modify_qp() 585 if (attr_mask & IB_QP_MIN_RNR_TIMER) ipath_modify_qp() 588 if (attr_mask & IB_QP_TIMEOUT) ipath_modify_qp() 591 if (attr_mask & IB_QP_QKEY) ipath_modify_qp() 594 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) ipath_modify_qp() 597 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) ipath_modify_qp() 622 int attr_mask, struct ib_qp_init_attr *init_attr) ipath_query_qp() 450 ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) ipath_modify_qp() argument 621 ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) ipath_query_qp() argument
|
H A D | ipath_srq.c | 212 * @attr_mask: indicates which attributes to modify 216 enum ib_srq_attr_mask attr_mask, ipath_modify_srq() 223 if (attr_mask & IB_SRQ_MAX_WR) { ipath_modify_srq() 230 ((attr_mask & IB_SRQ_LIMIT) ? ipath_modify_srq() 303 if (attr_mask & IB_SRQ_LIMIT) ipath_modify_srq() 333 } else if (attr_mask & IB_SRQ_LIMIT) { ipath_modify_srq() 215 ipath_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ipath_modify_srq() argument
|
H A D | ipath_verbs.h | 745 int attr_mask, struct ib_udata *udata); 748 int attr_mask, struct ib_qp_init_attr *init_attr); 799 enum ib_srq_attr_mask attr_mask,
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 299 int attr_mask) store_attrs() 301 if (attr_mask & IB_QP_PKEY_INDEX) store_attrs() 303 if (attr_mask & IB_QP_QKEY) store_attrs() 305 if (attr_mask & IB_QP_SQ_PSN) store_attrs() 328 int attr_mask) get_hw_access_flags() 334 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) get_hw_access_flags() 339 if (attr_mask & IB_QP_ACCESS_FLAGS) get_hw_access_flags() 544 const struct ib_qp_attr *attr, int attr_mask, __mthca_modify_qp() 567 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) __mthca_modify_qp() 588 else if (attr_mask & IB_QP_PATH_MTU) { __mthca_modify_qp() 615 if (attr_mask & IB_QP_DEST_QPN) { __mthca_modify_qp() 623 if (attr_mask & IB_QP_PORT) { __mthca_modify_qp() 630 if (attr_mask & IB_QP_PKEY_INDEX) { __mthca_modify_qp() 636 if (attr_mask & IB_QP_RNR_RETRY) { __mthca_modify_qp() 643 if (attr_mask & IB_QP_AV) { __mthca_modify_qp() 645 attr_mask & IB_QP_PORT ? attr->port_num : qp->port)) __mthca_modify_qp() 664 if (attr_mask & IB_QP_TIMEOUT) { __mthca_modify_qp() 669 if (attr_mask & IB_QP_ALT_PATH) { __mthca_modify_qp() 701 if (attr_mask & IB_QP_RETRY_CNT) { __mthca_modify_qp() 706 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mthca_modify_qp() 717 if (attr_mask & IB_QP_SQ_PSN) __mthca_modify_qp() 726 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mthca_modify_qp() 734 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) { __mthca_modify_qp() 735 qp_context->params2 |= get_hw_access_flags(qp, attr, attr_mask); __mthca_modify_qp() 746 if (attr_mask & IB_QP_MIN_RNR_TIMER) { __mthca_modify_qp() 750 if (attr_mask & IB_QP_RQ_PSN) __mthca_modify_qp() 763 if (attr_mask & IB_QP_QKEY) { __mthca_modify_qp() 773 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && __mthca_modify_qp() 786 if (attr_mask & IB_QP_ACCESS_FLAGS) __mthca_modify_qp() 788 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mthca_modify_qp() 790 if (attr_mask & IB_QP_PORT) __mthca_modify_qp() 792 if (attr_mask & IB_QP_ALT_PATH) __mthca_modify_qp() 796 store_attrs(to_msqp(qp), attr, attr_mask); __mthca_modify_qp() local 842 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, mthca_modify_qp() argument 851 if (attr_mask & IB_QP_CUR_STATE) { mthca_modify_qp() 861 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mthca_modify_qp() 863 if (!ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask, mthca_modify_qp() 868 attr_mask); mthca_modify_qp() 872 if ((attr_mask & IB_QP_PKEY_INDEX) && mthca_modify_qp() 879 if ((attr_mask & IB_QP_PORT) && mthca_modify_qp() 885 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mthca_modify_qp() 892 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mthca_modify_qp() 904 err = __mthca_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mthca_modify_qp() 298 store_attrs(struct mthca_sqp *sqp, const struct ib_qp_attr *attr, int attr_mask) store_attrs() argument 327 get_hw_access_flags(struct mthca_qp *qp, const struct ib_qp_attr *attr, int attr_mask) get_hw_access_flags() argument 543 __mthca_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mthca_modify_qp() argument
|
H A D | mthca_srq.c | 365 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mthca_modify_srq() 372 if (attr_mask & IB_SRQ_MAX_WR) mthca_modify_srq() 375 if (attr_mask & IB_SRQ_LIMIT) { mthca_modify_srq() 364 mthca_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mthca_modify_srq() argument
|
H A D | mthca_dev.h | 516 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata); 531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 1239 int attr_mask, int smi_reset2init) internal_modify_qp() 1293 (attr_mask & IB_QP_STATE) && internal_modify_qp() 1315 if ((attr_mask & IB_QP_CUR_STATE) && internal_modify_qp() 1327 my_qp, ibqp->qp_num, qp_cur_state, attr->qp_state, attr_mask); internal_modify_qp() 1329 qp_new_state = attr_mask & IB_QP_STATE ? attr->qp_state : qp_cur_state; internal_modify_qp() 1332 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) { internal_modify_qp() 1336 "ehca_qp=%p qp_num=%x attr_mask=%x", qp_new_state, internal_modify_qp() 1337 qp_cur_state, my_qp, ibqp->qp_num, attr_mask); internal_modify_qp() 1428 !(attr_mask & IB_QP_MAX_DEST_RD_ATOMIC)) { internal_modify_qp() 1434 if (attr_mask & IB_QP_PKEY_INDEX) { internal_modify_qp() 1445 if (attr_mask & IB_QP_PORT) { internal_modify_qp() 1482 if (attr_mask & IB_QP_QKEY) { internal_modify_qp() 1486 if (attr_mask & IB_QP_AV) { internal_modify_qp() 1534 if (attr_mask & IB_QP_PATH_MTU) { internal_modify_qp() 1540 if (attr_mask & IB_QP_TIMEOUT) { internal_modify_qp() 1544 if (attr_mask & IB_QP_RETRY_CNT) { internal_modify_qp() 1548 if (attr_mask & IB_QP_RNR_RETRY) { internal_modify_qp() 1552 if (attr_mask & IB_QP_RQ_PSN) { internal_modify_qp() 1556 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { internal_modify_qp() 1562 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { internal_modify_qp() 1569 if (attr_mask & IB_QP_ALT_PATH) { internal_modify_qp() 1648 if (attr_mask & IB_QP_MIN_RNR_TIMER) { internal_modify_qp() 1654 if (attr_mask & IB_QP_SQ_PSN) { internal_modify_qp() 1659 if (attr_mask & IB_QP_DEST_QPN) { internal_modify_qp() 1664 if (attr_mask & IB_QP_PATH_MIG_STATE) { internal_modify_qp() 1679 if (attr_mask & IB_QP_CAP) { internal_modify_qp() 1763 if (attr_mask & IB_QP_QKEY) internal_modify_qp() 1778 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, ehca_modify_qp() argument 1821 p->mask = attr_mask; ehca_modify_qp() 1834 ret = internal_modify_qp(ibqp, attr, attr_mask, 0); ehca_modify_qp() 1837 if ((ret == 0) && (attr_mask & IB_QP_STATE)) ehca_modify_qp() 2033 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ehca_modify_srq() 2052 if (attr_mask & IB_SRQ_LIMIT) { ehca_modify_srq() 2053 attr_mask &= ~IB_SRQ_LIMIT; ehca_modify_srq() 2062 /* by now, all bits in attr_mask should have been cleared */ ehca_modify_srq() 2063 if (attr_mask) { ehca_modify_srq() 2065 "attr_mask=%x", attr_mask); ehca_modify_srq() 1237 internal_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, int smi_reset2init) internal_modify_qp() argument 2032 ehca_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ehca_modify_srq() argument
|
H A D | ehca_iverbs.h | 149 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, 170 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 1228 int attr_mask) to_mlx4_access_flags() 1234 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) to_mlx4_access_flags() 1239 if (attr_mask & IB_QP_ACCESS_FLAGS) to_mlx4_access_flags() 1258 int attr_mask) store_sqp_attrs() 1260 if (attr_mask & IB_QP_PKEY_INDEX) store_sqp_attrs() 1262 if (attr_mask & IB_QP_QKEY) store_sqp_attrs() 1264 if (attr_mask & IB_QP_SQ_PSN) store_sqp_attrs() 1447 const struct ib_qp_attr *attr, int attr_mask, __mlx4_ib_modify_qp() 1461 if (attr_mask & IB_QP_ALT_PATH && __mlx4_ib_modify_qp() 1473 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) __mlx4_ib_modify_qp() 1500 } else if (attr_mask & IB_QP_PATH_MTU) { __mlx4_ib_modify_qp() 1530 if (attr_mask & IB_QP_DEST_QPN) __mlx4_ib_modify_qp() 1533 if (attr_mask & IB_QP_PORT) { __mlx4_ib_modify_qp() 1535 !(attr_mask & IB_QP_AV)) { __mlx4_ib_modify_qp() 1555 if (attr_mask & IB_QP_PKEY_INDEX) { __mlx4_ib_modify_qp() 1562 if (attr_mask & IB_QP_AV) { __mlx4_ib_modify_qp() 1563 if (mlx4_set_path(dev, attr, attr_mask, qp, &context->pri_path, __mlx4_ib_modify_qp() 1564 attr_mask & IB_QP_PORT ? __mlx4_ib_modify_qp() 1572 if (attr_mask & IB_QP_TIMEOUT) { __mlx4_ib_modify_qp() 1577 if (attr_mask & IB_QP_ALT_PATH) { __mlx4_ib_modify_qp() 1586 if (mlx4_set_alt_path(dev, attr, attr_mask, qp, __mlx4_ib_modify_qp() 1607 if (attr_mask & IB_QP_RNR_RETRY) { __mlx4_ib_modify_qp() 1612 if (attr_mask & IB_QP_RETRY_CNT) { __mlx4_ib_modify_qp() 1617 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mlx4_ib_modify_qp() 1624 if (attr_mask & IB_QP_SQ_PSN) __mlx4_ib_modify_qp() 1627 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mlx4_ib_modify_qp() 1634 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) { __mlx4_ib_modify_qp() 1635 context->params2 |= to_mlx4_access_flags(qp, attr, attr_mask); __mlx4_ib_modify_qp() 1642 if (attr_mask & IB_QP_MIN_RNR_TIMER) { __mlx4_ib_modify_qp() 1646 if (attr_mask & IB_QP_RQ_PSN) __mlx4_ib_modify_qp() 1650 if (attr_mask & IB_QP_QKEY) { __mlx4_ib_modify_qp() 1735 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && attr->en_sqd_async_notify) __mlx4_ib_modify_qp() 1771 if (attr_mask & IB_QP_ACCESS_FLAGS) __mlx4_ib_modify_qp() 1773 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mlx4_ib_modify_qp() 1775 if (attr_mask & IB_QP_PORT) { __mlx4_ib_modify_qp() 1779 if (attr_mask & IB_QP_ALT_PATH) __mlx4_ib_modify_qp() 1783 store_sqp_attrs(to_msqp(qp), attr, attr_mask); __mlx4_ib_modify_qp() local 1917 int attr_mask, struct ib_udata *udata) mlx4_ib_modify_qp() 1926 cur_state = attr_mask & IB_QP_CUR_STATE ? attr->cur_qp_state : qp->state; mlx4_ib_modify_qp() 1927 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mlx4_ib_modify_qp() 1932 int port = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx4_ib_modify_qp() 1937 attr_mask, ll)) { mlx4_ib_modify_qp() 1940 " attr_mask 0x%x\n", mlx4_ib_modify_qp() 1942 ibqp->qp_type, attr_mask); mlx4_ib_modify_qp() 1946 if (mlx4_is_bonded(dev->dev) && (attr_mask & IB_QP_PORT)) { mlx4_ib_modify_qp() 1958 attr_mask &= ~IB_QP_PORT; mlx4_ib_modify_qp() 1962 if ((attr_mask & IB_QP_PORT) && mlx4_ib_modify_qp() 1971 if ((attr_mask & IB_QP_PORT) && (ibqp->qp_type == IB_QPT_RAW_PACKET) && mlx4_ib_modify_qp() 1976 if (attr_mask & IB_QP_PKEY_INDEX) { mlx4_ib_modify_qp() 1977 int p = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx4_ib_modify_qp() 1987 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mlx4_ib_modify_qp() 1996 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mlx4_ib_modify_qp() 2010 err = __mlx4_ib_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mlx4_ib_modify_qp() 2012 if (mlx4_is_bonded(dev->dev) && (attr_mask & IB_QP_PORT)) mlx4_ib_modify_qp() 1227 to_mlx4_access_flags(struct mlx4_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) to_mlx4_access_flags() argument 1257 store_sqp_attrs(struct mlx4_ib_sqp *sqp, const struct ib_qp_attr *attr, int attr_mask) store_sqp_attrs() argument 1446 __mlx4_ib_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mlx4_ib_modify_qp() argument 1916 mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) mlx4_ib_modify_qp() argument
|
H A D | srq.c | 229 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx4_ib_modify_srq() 236 if (attr_mask & IB_SRQ_MAX_WR) mlx4_ib_modify_srq() 239 if (attr_mask & IB_SRQ_LIMIT) { mlx4_ib_modify_srq() 228 mlx4_ib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx4_ib_modify_srq() argument
|
H A D | mlx4_ib.h | 688 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata); 700 int attr_mask, struct ib_udata *udata);
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.h | 68 int attr_mask); 70 int attr_mask, struct ib_udata *udata);
|
H A D | ocrdma_hw.c | 2431 int attr_mask) ocrdma_set_av_params() 2474 if (attr_mask & IB_QP_VID) { ocrdma_set_av_params() 2497 struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_qp_params() 2502 if (attr_mask & IB_QP_PKEY_INDEX) { ocrdma_set_qp_params() 2507 if (attr_mask & IB_QP_QKEY) { ocrdma_set_qp_params() 2512 if (attr_mask & IB_QP_AV) { ocrdma_set_qp_params() 2513 status = ocrdma_set_av_params(qp, cmd, attrs, attr_mask); ocrdma_set_qp_params() 2525 if ((attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY) && ocrdma_set_qp_params() 2531 if (attr_mask & IB_QP_DEST_QPN) { ocrdma_set_qp_params() 2536 if (attr_mask & IB_QP_PATH_MTU) { ocrdma_set_qp_params() 2550 if (attr_mask & IB_QP_TIMEOUT) { ocrdma_set_qp_params() 2555 if (attr_mask & IB_QP_RETRY_CNT) { ocrdma_set_qp_params() 2561 if (attr_mask & IB_QP_MIN_RNR_TIMER) { ocrdma_set_qp_params() 2567 if (attr_mask & IB_QP_RNR_RETRY) { ocrdma_set_qp_params() 2573 if (attr_mask & IB_QP_SQ_PSN) { ocrdma_set_qp_params() 2577 if (attr_mask & IB_QP_RQ_PSN) { ocrdma_set_qp_params() 2581 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { ocrdma_set_qp_params() 2589 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { ocrdma_set_qp_params() 2605 struct ib_qp_attr *attrs, int attr_mask) ocrdma_mbx_modify_qp() 2616 if (attr_mask & IB_QP_STATE) { ocrdma_mbx_modify_qp() 2628 status = ocrdma_set_qp_params(qp, cmd, attrs, attr_mask); ocrdma_mbx_modify_qp() 2428 ocrdma_set_av_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_av_params() argument 2495 ocrdma_set_qp_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_qp_params() argument 2604 ocrdma_mbx_modify_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ib_qp_attr *attrs, int attr_mask) ocrdma_mbx_modify_qp() argument
|
H A D | ocrdma_hw.h | 115 struct ib_qp_attr *attrs, int attr_mask);
|
H A D | ocrdma_verbs.c | 1411 int attr_mask) _ocrdma_modify_qp() 1420 if (attr_mask & IB_QP_STATE) _ocrdma_modify_qp() 1427 status = ocrdma_mbx_modify_qp(dev, qp, attr, attr_mask); _ocrdma_modify_qp() 1433 int attr_mask, struct ib_udata *udata) ocrdma_modify_qp() 1449 if (attr_mask & IB_QP_STATE) ocrdma_modify_qp() 1455 if (!ib_modify_qp_is_ok(old_qps, new_qps, ibqp->qp_type, attr_mask, ocrdma_modify_qp() 1459 __func__, dev->id, attr_mask, qp->id, ibqp->qp_type, ocrdma_modify_qp() 1464 status = _ocrdma_modify_qp(ibqp, attr, attr_mask); ocrdma_modify_qp() 1503 int attr_mask, struct ib_qp_init_attr *qp_init_attr) ocrdma_query_qp() 1723 int attr_mask; ocrdma_destroy_qp() local 1734 attr_mask = IB_QP_STATE; ocrdma_destroy_qp() 1735 _ocrdma_modify_qp(ibqp, &attrs, attr_mask); ocrdma_destroy_qp() 1410 _ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask) _ocrdma_modify_qp() argument 1432 ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) ocrdma_modify_qp() argument 1501 ocrdma_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int attr_mask, struct ib_qp_init_attr *qp_init_attr) ocrdma_query_qp() argument
|
H A D | ocrdma_main.c | 569 int attr_mask = IB_QP_STATE; ocrdma_close() local 579 _ocrdma_modify_qp(&qp->ibqp, &attrs, attr_mask); ocrdma_close()
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
H A D | qp.c | 1322 int attr_mask) to_mlx5_access_flags() 1328 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) to_mlx5_access_flags() 1333 if (attr_mask & IB_QP_ACCESS_FLAGS) to_mlx5_access_flags() 1377 struct mlx5_qp_path *path, u8 port, int attr_mask, mlx5_set_path() 1387 if (attr_mask & IB_QP_PKEY_INDEX) mlx5_set_path() 1414 if (attr_mask & IB_QP_TIMEOUT) mlx5_set_path() 1566 const struct ib_qp_attr *attr, int attr_mask, __mlx5_ib_modify_qp() 1594 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) { __mlx5_ib_modify_qp() 1615 } else if (attr_mask & IB_QP_PATH_MTU) { __mlx5_ib_modify_qp() 1625 if (attr_mask & IB_QP_DEST_QPN) __mlx5_ib_modify_qp() 1628 if (attr_mask & IB_QP_PKEY_INDEX) __mlx5_ib_modify_qp() 1636 if (attr_mask & IB_QP_PORT) __mlx5_ib_modify_qp() 1639 if (attr_mask & IB_QP_AV) { __mlx5_ib_modify_qp() 1641 attr_mask & IB_QP_PORT ? attr->port_num : qp->port, __mlx5_ib_modify_qp() 1642 attr_mask, 0, attr); __mlx5_ib_modify_qp() 1647 if (attr_mask & IB_QP_TIMEOUT) __mlx5_ib_modify_qp() 1650 if (attr_mask & IB_QP_ALT_PATH) { __mlx5_ib_modify_qp() 1652 attr->alt_port_num, attr_mask, 0, attr); __mlx5_ib_modify_qp() 1665 if (attr_mask & IB_QP_RNR_RETRY) __mlx5_ib_modify_qp() 1668 if (attr_mask & IB_QP_RETRY_CNT) __mlx5_ib_modify_qp() 1671 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mlx5_ib_modify_qp() 1677 if (attr_mask & IB_QP_SQ_PSN) __mlx5_ib_modify_qp() 1680 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mlx5_ib_modify_qp() 1686 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) __mlx5_ib_modify_qp() 1687 context->params2 |= to_mlx5_access_flags(qp, attr, attr_mask); __mlx5_ib_modify_qp() 1689 if (attr_mask & IB_QP_MIN_RNR_TIMER) __mlx5_ib_modify_qp() 1692 if (attr_mask & IB_QP_RQ_PSN) __mlx5_ib_modify_qp() 1695 if (attr_mask & IB_QP_QKEY) __mlx5_ib_modify_qp() 1702 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && attr->en_sqd_async_notify) __mlx5_ib_modify_qp() 1726 optpar = ib_mask_to_mlx5_opt(attr_mask); __mlx5_ib_modify_qp() 1740 if (attr_mask & IB_QP_ACCESS_FLAGS) __mlx5_ib_modify_qp() 1742 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mlx5_ib_modify_qp() 1744 if (attr_mask & IB_QP_PORT) __mlx5_ib_modify_qp() 1746 if (attr_mask & IB_QP_ALT_PATH) __mlx5_ib_modify_qp() 1775 int attr_mask, struct ib_udata *udata) mlx5_ib_modify_qp() 1787 cur_state = attr_mask & IB_QP_CUR_STATE ? attr->cur_qp_state : qp->state; mlx5_ib_modify_qp() 1788 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mlx5_ib_modify_qp() 1791 !ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask, mlx5_ib_modify_qp() 1795 if ((attr_mask & IB_QP_PORT) && mlx5_ib_modify_qp() 1799 if (attr_mask & IB_QP_PKEY_INDEX) { mlx5_ib_modify_qp() 1800 port = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx5_ib_modify_qp() 1805 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mlx5_ib_modify_qp() 1809 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mlx5_ib_modify_qp() 1818 err = __mlx5_ib_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mlx5_ib_modify_qp() 1321 to_mlx5_access_flags(struct mlx5_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) to_mlx5_access_flags() argument 1376 mlx5_set_path(struct mlx5_ib_dev *dev, const struct ib_ah_attr *ah, struct mlx5_qp_path *path, u8 port, int attr_mask, u32 path_flags, const struct ib_qp_attr *attr) mlx5_set_path() argument 1565 __mlx5_ib_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mlx5_ib_modify_qp() argument 1774 mlx5_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) mlx5_ib_modify_qp() argument
|
H A D | srq.c | 345 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx5_ib_modify_srq() 352 if (attr_mask & IB_SRQ_MAX_WR) mlx5_ib_modify_srq() 355 if (attr_mask & IB_SRQ_LIMIT) { mlx5_ib_modify_srq() 344 mlx5_ib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx5_ib_modify_srq() argument
|
H A D | mlx5_ib.h | 539 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata); 548 int attr_mask, struct ib_udata *udata);
|
/linux-4.1.27/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_verbs.h | 46 int attr_mask, struct ib_udata *udata);
|
H A D | usnic_ib_verbs.c | 547 int attr_mask, struct ib_udata *udata) usnic_ib_modify_qp() 557 if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_INIT) { usnic_ib_modify_qp() 559 } else if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_RTR) { usnic_ib_modify_qp() 561 } else if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_RTS) { usnic_ib_modify_qp() 565 attr_mask & IB_QP_STATE, attr->qp_state); usnic_ib_modify_qp() 546 usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) usnic_ib_modify_qp() argument
|
/linux-4.1.27/include/uapi/rdma/ |
H A D | ib_user_verbs.h | 544 __u32 attr_mask; member in struct:ib_uverbs_query_qp 586 __u32 attr_mask; member in struct:ib_uverbs_modify_qp 873 __u32 attr_mask; member in struct:ib_uverbs_modify_srq
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 1814 int attr_mask, struct ib_udata *udata) c4iw_ib_modify_qp() 1824 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) c4iw_ib_modify_qp() 1825 attr_mask &= ~IB_QP_STATE; c4iw_ib_modify_qp() 1828 if (!attr_mask) c4iw_ib_modify_qp() 1843 mask |= (attr_mask & IB_QP_STATE) ? C4IW_QP_ATTR_NEXT_STATE : 0; c4iw_ib_modify_qp() 1844 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? c4iw_ib_modify_qp() 1856 mask |= (attr_mask & IB_QP_SQ_PSN) ? C4IW_QP_ATTR_SQ_DB : 0; c4iw_ib_modify_qp() 1857 mask |= (attr_mask & IB_QP_RQ_PSN) ? C4IW_QP_ATTR_RQ_DB : 0; c4iw_ib_modify_qp() 1872 int attr_mask, struct ib_qp_init_attr *init_attr) c4iw_ib_query_qp() 1813 c4iw_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) c4iw_ib_modify_qp() argument 1871 c4iw_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) c4iw_ib_query_qp() argument
|
H A D | iw_cxgb4.h | 1006 int attr_mask, struct ib_udata *udata); 1008 int attr_mask, struct ib_qp_init_attr *init_attr);
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 1047 int attr_mask, struct ib_udata *udata) iwch_ib_modify_qp() 1057 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) iwch_ib_modify_qp() 1058 attr_mask &= ~IB_QP_STATE; iwch_ib_modify_qp() 1061 if (!attr_mask) iwch_ib_modify_qp() 1076 mask |= (attr_mask & IB_QP_STATE) ? IWCH_QP_ATTR_NEXT_STATE : 0; iwch_ib_modify_qp() 1077 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? iwch_ib_modify_qp() 1046 iwch_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) iwch_ib_modify_qp() argument
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
H A D | c2_provider.c | 559 int attr_mask, struct ib_udata *udata) c2_modify_qp() 565 attr_mask); c2_modify_qp() 558 c2_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) c2_modify_qp() argument
|
H A D | c2_qp.c | 135 struct ib_qp_attr *attr, int attr_mask) c2_qp_modify() 163 if (attr_mask & IB_QP_STATE) { c2_qp_modify() 185 } else if (attr_mask & IB_QP_CUR_STATE) { c2_qp_modify() 134 c2_qp_modify(struct c2_dev *c2dev, struct c2_qp *qp, struct ib_qp_attr *attr, int attr_mask) c2_qp_modify() argument
|
H A D | c2.h | 493 struct ib_qp_attr *attr, int attr_mask);
|
/linux-4.1.27/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 2147 static uint32_t srp_compute_rq_tmo(struct ib_qp_attr *qp_attr, int attr_mask) srp_compute_rq_tmo() argument 2157 WARN_ON_ONCE((attr_mask & (IB_QP_TIMEOUT | IB_QP_RETRY_CNT)) != srp_compute_rq_tmo() 2180 int attr_mask = 0; srp_cm_rep_handler() local 2217 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); srp_cm_rep_handler() 2221 ret = ib_modify_qp(ch->qp, qp_attr, attr_mask); srp_cm_rep_handler() 2234 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); srp_cm_rep_handler() 2238 target->rq_tmo_jiffies = srp_compute_rq_tmo(qp_attr, attr_mask); srp_cm_rep_handler() 2240 ret = ib_modify_qp(ch->qp, qp_attr, attr_mask); srp_cm_rep_handler()
|
/linux-4.1.27/drivers/infiniband/core/ |
H A D | uverbs_cmd.c | 1938 ret = ib_query_qp(qp, attr, cmd.attr_mask, init_attr); ib_uverbs_query_qp() 2098 ret = ib_resolve_eth_l2_attrs(qp, attr, &cmd.attr_mask); ib_uverbs_modify_qp() 2102 modify_qp_mask(qp->qp_type, cmd.attr_mask), &udata); ib_uverbs_modify_qp() 2104 ret = ib_modify_qp(qp, attr, modify_qp_mask(qp->qp_type, cmd.attr_mask)); ib_uverbs_modify_qp() 3202 ret = srq->device->modify_srq(srq, &attr, cmd.attr_mask, &udata); ib_uverbs_modify_srq()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.c | 2820 int attr_mask, struct ib_qp_init_attr *init_attr) nes_query_qp() 2927 int attr_mask, struct ib_udata *udata) nes_modify_qp() 2950 " QP Access Flags=0x%X, attr_mask = 0x%0x\n", nes_modify_qp() 2952 nesqp->hw_tcp_state, attr->qp_access_flags, attr_mask); nes_modify_qp() 2954 if (attr_mask & IB_QP_STATE) { nes_modify_qp() 3090 if (attr_mask & IB_QP_ACCESS_FLAGS) { nes_modify_qp() 2819 nes_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) nes_query_qp() argument 2926 nes_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) nes_modify_qp() argument
|
/linux-4.1.27/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 988 int attr_mask; srpt_ch_qp_rtr() local 992 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); srpt_ch_qp_rtr() 998 ret = ib_modify_qp(qp, &qp_attr, attr_mask); srpt_ch_qp_rtr() 1018 int attr_mask; srpt_ch_qp_rts() local 1022 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); srpt_ch_qp_rts() 1028 ret = ib_modify_qp(qp, &qp_attr, attr_mask); srpt_ch_qp_rts()
|