Searched refs:attr_mask (Results 1 - 36 of 36) sorted by relevance

/linux-4.1.27/drivers/infiniband/hw/qib/
H A Dqib_qp.c564 * @attr_mask: the mask of attributes to modify
570 int attr_mask, struct ib_udata *udata) qib_modify_qp()
584 cur_state = attr_mask & IB_QP_CUR_STATE ? qib_modify_qp()
586 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; qib_modify_qp()
589 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) qib_modify_qp()
592 if (attr_mask & IB_QP_AV) { qib_modify_qp()
599 if (attr_mask & IB_QP_ALT_PATH) { qib_modify_qp()
608 if (attr_mask & IB_QP_PKEY_INDEX) qib_modify_qp()
612 if (attr_mask & IB_QP_MIN_RNR_TIMER) qib_modify_qp()
616 if (attr_mask & IB_QP_PORT) qib_modify_qp()
623 if (attr_mask & IB_QP_DEST_QPN) qib_modify_qp()
627 if (attr_mask & IB_QP_RETRY_CNT) qib_modify_qp()
631 if (attr_mask & IB_QP_RNR_RETRY) qib_modify_qp()
643 if (attr_mask & IB_QP_PATH_MTU) { qib_modify_qp()
674 if (attr_mask & IB_QP_PATH_MIG_STATE) { qib_modify_qp()
691 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) qib_modify_qp()
749 if (attr_mask & IB_QP_PKEY_INDEX) qib_modify_qp()
752 if (attr_mask & IB_QP_PORT) qib_modify_qp()
755 if (attr_mask & IB_QP_DEST_QPN) qib_modify_qp()
758 if (attr_mask & IB_QP_SQ_PSN) { qib_modify_qp()
766 if (attr_mask & IB_QP_RQ_PSN) qib_modify_qp()
769 if (attr_mask & IB_QP_ACCESS_FLAGS) qib_modify_qp()
772 if (attr_mask & IB_QP_AV) { qib_modify_qp()
777 if (attr_mask & IB_QP_ALT_PATH) { qib_modify_qp()
782 if (attr_mask & IB_QP_PATH_MIG_STATE) { qib_modify_qp()
791 if (attr_mask & IB_QP_PATH_MTU) { qib_modify_qp()
796 if (attr_mask & IB_QP_RETRY_CNT) { qib_modify_qp()
801 if (attr_mask & IB_QP_RNR_RETRY) { qib_modify_qp()
806 if (attr_mask & IB_QP_MIN_RNR_TIMER) qib_modify_qp()
809 if (attr_mask & IB_QP_TIMEOUT) { qib_modify_qp()
816 if (attr_mask & IB_QP_QKEY) qib_modify_qp()
819 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) qib_modify_qp()
822 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) qib_modify_qp()
856 int attr_mask, struct ib_qp_init_attr *init_attr) qib_query_qp()
569 qib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) qib_modify_qp() argument
855 qib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) qib_query_qp() argument
H A Dqib_srq.c208 * @attr_mask: indicates which attributes to modify
212 enum ib_srq_attr_mask attr_mask, qib_modify_srq()
219 if (attr_mask & IB_SRQ_MAX_WR) { qib_modify_srq()
226 ((attr_mask & IB_SRQ_LIMIT) ? qib_modify_srq()
299 if (attr_mask & IB_SRQ_LIMIT) qib_modify_srq()
333 } else if (attr_mask & IB_SRQ_LIMIT) { qib_modify_srq()
211 qib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) qib_modify_srq() argument
H A Dqib_verbs.h919 int attr_mask, struct ib_udata *udata);
922 int attr_mask, struct ib_qp_init_attr *init_attr);
997 enum ib_srq_attr_mask attr_mask,
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c82 int attr_mask; ipoib_init_qp() local
91 attr_mask = ipoib_init_qp()
96 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp()
104 attr_mask &= ~IB_QP_PORT; ipoib_init_qp()
105 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp()
113 attr_mask |= IB_QP_SQ_PSN; ipoib_init_qp()
114 attr_mask &= ~IB_QP_PKEY_INDEX; ipoib_init_qp()
115 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); ipoib_init_qp()
/linux-4.1.27/drivers/infiniband/hw/ipath/
H A Dipath_qp.c445 * @attr_mask: the mask of attributes to modify
451 int attr_mask, struct ib_udata *udata) ipath_modify_qp()
461 cur_state = attr_mask & IB_QP_CUR_STATE ? ipath_modify_qp()
463 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; ipath_modify_qp()
466 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) ipath_modify_qp()
469 if (attr_mask & IB_QP_AV) { ipath_modify_qp()
479 if (attr_mask & IB_QP_PKEY_INDEX) ipath_modify_qp()
483 if (attr_mask & IB_QP_MIN_RNR_TIMER) ipath_modify_qp()
487 if (attr_mask & IB_QP_PORT) ipath_modify_qp()
496 if ((attr_mask & IB_QP_PATH_MTU) && ipath_modify_qp()
501 if (attr_mask & IB_QP_PATH_MIG_STATE) ipath_modify_qp()
506 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) ipath_modify_qp()
550 if (attr_mask & IB_QP_PKEY_INDEX) ipath_modify_qp()
553 if (attr_mask & IB_QP_DEST_QPN) ipath_modify_qp()
556 if (attr_mask & IB_QP_SQ_PSN) { ipath_modify_qp()
561 if (attr_mask & IB_QP_RQ_PSN) ipath_modify_qp()
564 if (attr_mask & IB_QP_ACCESS_FLAGS) ipath_modify_qp()
567 if (attr_mask & IB_QP_AV) { ipath_modify_qp()
572 if (attr_mask & IB_QP_PATH_MTU) ipath_modify_qp()
575 if (attr_mask & IB_QP_RETRY_CNT) ipath_modify_qp()
578 if (attr_mask & IB_QP_RNR_RETRY) { ipath_modify_qp()
585 if (attr_mask & IB_QP_MIN_RNR_TIMER) ipath_modify_qp()
588 if (attr_mask & IB_QP_TIMEOUT) ipath_modify_qp()
591 if (attr_mask & IB_QP_QKEY) ipath_modify_qp()
594 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) ipath_modify_qp()
597 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) ipath_modify_qp()
622 int attr_mask, struct ib_qp_init_attr *init_attr) ipath_query_qp()
450 ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) ipath_modify_qp() argument
621 ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) ipath_query_qp() argument
H A Dipath_srq.c212 * @attr_mask: indicates which attributes to modify
216 enum ib_srq_attr_mask attr_mask, ipath_modify_srq()
223 if (attr_mask & IB_SRQ_MAX_WR) { ipath_modify_srq()
230 ((attr_mask & IB_SRQ_LIMIT) ? ipath_modify_srq()
303 if (attr_mask & IB_SRQ_LIMIT) ipath_modify_srq()
333 } else if (attr_mask & IB_SRQ_LIMIT) { ipath_modify_srq()
215 ipath_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ipath_modify_srq() argument
H A Dipath_verbs.h745 int attr_mask, struct ib_udata *udata);
748 int attr_mask, struct ib_qp_init_attr *init_attr);
799 enum ib_srq_attr_mask attr_mask,
/linux-4.1.27/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c299 int attr_mask) store_attrs()
301 if (attr_mask & IB_QP_PKEY_INDEX) store_attrs()
303 if (attr_mask & IB_QP_QKEY) store_attrs()
305 if (attr_mask & IB_QP_SQ_PSN) store_attrs()
328 int attr_mask) get_hw_access_flags()
334 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) get_hw_access_flags()
339 if (attr_mask & IB_QP_ACCESS_FLAGS) get_hw_access_flags()
544 const struct ib_qp_attr *attr, int attr_mask, __mthca_modify_qp()
567 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) __mthca_modify_qp()
588 else if (attr_mask & IB_QP_PATH_MTU) { __mthca_modify_qp()
615 if (attr_mask & IB_QP_DEST_QPN) { __mthca_modify_qp()
623 if (attr_mask & IB_QP_PORT) { __mthca_modify_qp()
630 if (attr_mask & IB_QP_PKEY_INDEX) { __mthca_modify_qp()
636 if (attr_mask & IB_QP_RNR_RETRY) { __mthca_modify_qp()
643 if (attr_mask & IB_QP_AV) { __mthca_modify_qp()
645 attr_mask & IB_QP_PORT ? attr->port_num : qp->port)) __mthca_modify_qp()
664 if (attr_mask & IB_QP_TIMEOUT) { __mthca_modify_qp()
669 if (attr_mask & IB_QP_ALT_PATH) { __mthca_modify_qp()
701 if (attr_mask & IB_QP_RETRY_CNT) { __mthca_modify_qp()
706 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mthca_modify_qp()
717 if (attr_mask & IB_QP_SQ_PSN) __mthca_modify_qp()
726 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mthca_modify_qp()
734 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) { __mthca_modify_qp()
735 qp_context->params2 |= get_hw_access_flags(qp, attr, attr_mask); __mthca_modify_qp()
746 if (attr_mask & IB_QP_MIN_RNR_TIMER) { __mthca_modify_qp()
750 if (attr_mask & IB_QP_RQ_PSN) __mthca_modify_qp()
763 if (attr_mask & IB_QP_QKEY) { __mthca_modify_qp()
773 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && __mthca_modify_qp()
786 if (attr_mask & IB_QP_ACCESS_FLAGS) __mthca_modify_qp()
788 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mthca_modify_qp()
790 if (attr_mask & IB_QP_PORT) __mthca_modify_qp()
792 if (attr_mask & IB_QP_ALT_PATH) __mthca_modify_qp()
796 store_attrs(to_msqp(qp), attr, attr_mask); __mthca_modify_qp() local
842 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, mthca_modify_qp() argument
851 if (attr_mask & IB_QP_CUR_STATE) { mthca_modify_qp()
861 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mthca_modify_qp()
863 if (!ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask, mthca_modify_qp()
868 attr_mask); mthca_modify_qp()
872 if ((attr_mask & IB_QP_PKEY_INDEX) && mthca_modify_qp()
879 if ((attr_mask & IB_QP_PORT) && mthca_modify_qp()
885 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mthca_modify_qp()
892 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mthca_modify_qp()
904 err = __mthca_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mthca_modify_qp()
298 store_attrs(struct mthca_sqp *sqp, const struct ib_qp_attr *attr, int attr_mask) store_attrs() argument
327 get_hw_access_flags(struct mthca_qp *qp, const struct ib_qp_attr *attr, int attr_mask) get_hw_access_flags() argument
543 __mthca_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mthca_modify_qp() argument
H A Dmthca_srq.c365 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mthca_modify_srq()
372 if (attr_mask & IB_SRQ_MAX_WR) mthca_modify_srq()
375 if (attr_mask & IB_SRQ_LIMIT) { mthca_modify_srq()
364 mthca_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mthca_modify_srq() argument
H A Dmthca_dev.h516 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
/linux-4.1.27/drivers/infiniband/hw/ehca/
H A Dehca_qp.c1239 int attr_mask, int smi_reset2init) internal_modify_qp()
1293 (attr_mask & IB_QP_STATE) && internal_modify_qp()
1315 if ((attr_mask & IB_QP_CUR_STATE) && internal_modify_qp()
1327 my_qp, ibqp->qp_num, qp_cur_state, attr->qp_state, attr_mask); internal_modify_qp()
1329 qp_new_state = attr_mask & IB_QP_STATE ? attr->qp_state : qp_cur_state; internal_modify_qp()
1332 attr_mask, IB_LINK_LAYER_UNSPECIFIED)) { internal_modify_qp()
1336 "ehca_qp=%p qp_num=%x attr_mask=%x", qp_new_state, internal_modify_qp()
1337 qp_cur_state, my_qp, ibqp->qp_num, attr_mask); internal_modify_qp()
1428 !(attr_mask & IB_QP_MAX_DEST_RD_ATOMIC)) { internal_modify_qp()
1434 if (attr_mask & IB_QP_PKEY_INDEX) { internal_modify_qp()
1445 if (attr_mask & IB_QP_PORT) { internal_modify_qp()
1482 if (attr_mask & IB_QP_QKEY) { internal_modify_qp()
1486 if (attr_mask & IB_QP_AV) { internal_modify_qp()
1534 if (attr_mask & IB_QP_PATH_MTU) { internal_modify_qp()
1540 if (attr_mask & IB_QP_TIMEOUT) { internal_modify_qp()
1544 if (attr_mask & IB_QP_RETRY_CNT) { internal_modify_qp()
1548 if (attr_mask & IB_QP_RNR_RETRY) { internal_modify_qp()
1552 if (attr_mask & IB_QP_RQ_PSN) { internal_modify_qp()
1556 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { internal_modify_qp()
1562 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { internal_modify_qp()
1569 if (attr_mask & IB_QP_ALT_PATH) { internal_modify_qp()
1648 if (attr_mask & IB_QP_MIN_RNR_TIMER) { internal_modify_qp()
1654 if (attr_mask & IB_QP_SQ_PSN) { internal_modify_qp()
1659 if (attr_mask & IB_QP_DEST_QPN) { internal_modify_qp()
1664 if (attr_mask & IB_QP_PATH_MIG_STATE) { internal_modify_qp()
1679 if (attr_mask & IB_QP_CAP) { internal_modify_qp()
1763 if (attr_mask & IB_QP_QKEY) internal_modify_qp()
1778 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, ehca_modify_qp() argument
1821 p->mask = attr_mask; ehca_modify_qp()
1834 ret = internal_modify_qp(ibqp, attr, attr_mask, 0); ehca_modify_qp()
1837 if ((ret == 0) && (attr_mask & IB_QP_STATE)) ehca_modify_qp()
2033 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ehca_modify_srq()
2052 if (attr_mask & IB_SRQ_LIMIT) { ehca_modify_srq()
2053 attr_mask &= ~IB_SRQ_LIMIT; ehca_modify_srq()
2062 /* by now, all bits in attr_mask should have been cleared */ ehca_modify_srq()
2063 if (attr_mask) { ehca_modify_srq()
2065 "attr_mask=%x", attr_mask); ehca_modify_srq()
1237 internal_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, int smi_reset2init) internal_modify_qp() argument
2032 ehca_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) ehca_modify_srq() argument
H A Dehca_iverbs.h149 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
170 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
/linux-4.1.27/drivers/infiniband/hw/mlx4/
H A Dqp.c1228 int attr_mask) to_mlx4_access_flags()
1234 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) to_mlx4_access_flags()
1239 if (attr_mask & IB_QP_ACCESS_FLAGS) to_mlx4_access_flags()
1258 int attr_mask) store_sqp_attrs()
1260 if (attr_mask & IB_QP_PKEY_INDEX) store_sqp_attrs()
1262 if (attr_mask & IB_QP_QKEY) store_sqp_attrs()
1264 if (attr_mask & IB_QP_SQ_PSN) store_sqp_attrs()
1447 const struct ib_qp_attr *attr, int attr_mask, __mlx4_ib_modify_qp()
1461 if (attr_mask & IB_QP_ALT_PATH && __mlx4_ib_modify_qp()
1473 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) __mlx4_ib_modify_qp()
1500 } else if (attr_mask & IB_QP_PATH_MTU) { __mlx4_ib_modify_qp()
1530 if (attr_mask & IB_QP_DEST_QPN) __mlx4_ib_modify_qp()
1533 if (attr_mask & IB_QP_PORT) { __mlx4_ib_modify_qp()
1535 !(attr_mask & IB_QP_AV)) { __mlx4_ib_modify_qp()
1555 if (attr_mask & IB_QP_PKEY_INDEX) { __mlx4_ib_modify_qp()
1562 if (attr_mask & IB_QP_AV) { __mlx4_ib_modify_qp()
1563 if (mlx4_set_path(dev, attr, attr_mask, qp, &context->pri_path, __mlx4_ib_modify_qp()
1564 attr_mask & IB_QP_PORT ? __mlx4_ib_modify_qp()
1572 if (attr_mask & IB_QP_TIMEOUT) { __mlx4_ib_modify_qp()
1577 if (attr_mask & IB_QP_ALT_PATH) { __mlx4_ib_modify_qp()
1586 if (mlx4_set_alt_path(dev, attr, attr_mask, qp, __mlx4_ib_modify_qp()
1607 if (attr_mask & IB_QP_RNR_RETRY) { __mlx4_ib_modify_qp()
1612 if (attr_mask & IB_QP_RETRY_CNT) { __mlx4_ib_modify_qp()
1617 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mlx4_ib_modify_qp()
1624 if (attr_mask & IB_QP_SQ_PSN) __mlx4_ib_modify_qp()
1627 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mlx4_ib_modify_qp()
1634 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) { __mlx4_ib_modify_qp()
1635 context->params2 |= to_mlx4_access_flags(qp, attr, attr_mask); __mlx4_ib_modify_qp()
1642 if (attr_mask & IB_QP_MIN_RNR_TIMER) { __mlx4_ib_modify_qp()
1646 if (attr_mask & IB_QP_RQ_PSN) __mlx4_ib_modify_qp()
1650 if (attr_mask & IB_QP_QKEY) { __mlx4_ib_modify_qp()
1735 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && attr->en_sqd_async_notify) __mlx4_ib_modify_qp()
1771 if (attr_mask & IB_QP_ACCESS_FLAGS) __mlx4_ib_modify_qp()
1773 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mlx4_ib_modify_qp()
1775 if (attr_mask & IB_QP_PORT) { __mlx4_ib_modify_qp()
1779 if (attr_mask & IB_QP_ALT_PATH) __mlx4_ib_modify_qp()
1783 store_sqp_attrs(to_msqp(qp), attr, attr_mask); __mlx4_ib_modify_qp() local
1917 int attr_mask, struct ib_udata *udata) mlx4_ib_modify_qp()
1926 cur_state = attr_mask & IB_QP_CUR_STATE ? attr->cur_qp_state : qp->state; mlx4_ib_modify_qp()
1927 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mlx4_ib_modify_qp()
1932 int port = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx4_ib_modify_qp()
1937 attr_mask, ll)) { mlx4_ib_modify_qp()
1940 " attr_mask 0x%x\n", mlx4_ib_modify_qp()
1942 ibqp->qp_type, attr_mask); mlx4_ib_modify_qp()
1946 if (mlx4_is_bonded(dev->dev) && (attr_mask & IB_QP_PORT)) { mlx4_ib_modify_qp()
1958 attr_mask &= ~IB_QP_PORT; mlx4_ib_modify_qp()
1962 if ((attr_mask & IB_QP_PORT) && mlx4_ib_modify_qp()
1971 if ((attr_mask & IB_QP_PORT) && (ibqp->qp_type == IB_QPT_RAW_PACKET) && mlx4_ib_modify_qp()
1976 if (attr_mask & IB_QP_PKEY_INDEX) { mlx4_ib_modify_qp()
1977 int p = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx4_ib_modify_qp()
1987 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mlx4_ib_modify_qp()
1996 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mlx4_ib_modify_qp()
2010 err = __mlx4_ib_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mlx4_ib_modify_qp()
2012 if (mlx4_is_bonded(dev->dev) && (attr_mask & IB_QP_PORT)) mlx4_ib_modify_qp()
1227 to_mlx4_access_flags(struct mlx4_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) to_mlx4_access_flags() argument
1257 store_sqp_attrs(struct mlx4_ib_sqp *sqp, const struct ib_qp_attr *attr, int attr_mask) store_sqp_attrs() argument
1446 __mlx4_ib_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mlx4_ib_modify_qp() argument
1916 mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) mlx4_ib_modify_qp() argument
H A Dsrq.c229 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx4_ib_modify_srq()
236 if (attr_mask & IB_SRQ_MAX_WR) mlx4_ib_modify_srq()
239 if (attr_mask & IB_SRQ_LIMIT) { mlx4_ib_modify_srq()
228 mlx4_ib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx4_ib_modify_srq() argument
H A Dmlx4_ib.h688 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
700 int attr_mask, struct ib_udata *udata);
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.h68 int attr_mask);
70 int attr_mask, struct ib_udata *udata);
H A Docrdma_hw.c2431 int attr_mask) ocrdma_set_av_params()
2474 if (attr_mask & IB_QP_VID) { ocrdma_set_av_params()
2497 struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_qp_params()
2502 if (attr_mask & IB_QP_PKEY_INDEX) { ocrdma_set_qp_params()
2507 if (attr_mask & IB_QP_QKEY) { ocrdma_set_qp_params()
2512 if (attr_mask & IB_QP_AV) { ocrdma_set_qp_params()
2513 status = ocrdma_set_av_params(qp, cmd, attrs, attr_mask); ocrdma_set_qp_params()
2525 if ((attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY) && ocrdma_set_qp_params()
2531 if (attr_mask & IB_QP_DEST_QPN) { ocrdma_set_qp_params()
2536 if (attr_mask & IB_QP_PATH_MTU) { ocrdma_set_qp_params()
2550 if (attr_mask & IB_QP_TIMEOUT) { ocrdma_set_qp_params()
2555 if (attr_mask & IB_QP_RETRY_CNT) { ocrdma_set_qp_params()
2561 if (attr_mask & IB_QP_MIN_RNR_TIMER) { ocrdma_set_qp_params()
2567 if (attr_mask & IB_QP_RNR_RETRY) { ocrdma_set_qp_params()
2573 if (attr_mask & IB_QP_SQ_PSN) { ocrdma_set_qp_params()
2577 if (attr_mask & IB_QP_RQ_PSN) { ocrdma_set_qp_params()
2581 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { ocrdma_set_qp_params()
2589 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { ocrdma_set_qp_params()
2605 struct ib_qp_attr *attrs, int attr_mask) ocrdma_mbx_modify_qp()
2616 if (attr_mask & IB_QP_STATE) { ocrdma_mbx_modify_qp()
2628 status = ocrdma_set_qp_params(qp, cmd, attrs, attr_mask); ocrdma_mbx_modify_qp()
2428 ocrdma_set_av_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_av_params() argument
2495 ocrdma_set_qp_params(struct ocrdma_qp *qp, struct ocrdma_modify_qp *cmd, struct ib_qp_attr *attrs, int attr_mask) ocrdma_set_qp_params() argument
2604 ocrdma_mbx_modify_qp(struct ocrdma_dev *dev, struct ocrdma_qp *qp, struct ib_qp_attr *attrs, int attr_mask) ocrdma_mbx_modify_qp() argument
H A Docrdma_hw.h115 struct ib_qp_attr *attrs, int attr_mask);
H A Docrdma_verbs.c1411 int attr_mask) _ocrdma_modify_qp()
1420 if (attr_mask & IB_QP_STATE) _ocrdma_modify_qp()
1427 status = ocrdma_mbx_modify_qp(dev, qp, attr, attr_mask); _ocrdma_modify_qp()
1433 int attr_mask, struct ib_udata *udata) ocrdma_modify_qp()
1449 if (attr_mask & IB_QP_STATE) ocrdma_modify_qp()
1455 if (!ib_modify_qp_is_ok(old_qps, new_qps, ibqp->qp_type, attr_mask, ocrdma_modify_qp()
1459 __func__, dev->id, attr_mask, qp->id, ibqp->qp_type, ocrdma_modify_qp()
1464 status = _ocrdma_modify_qp(ibqp, attr, attr_mask); ocrdma_modify_qp()
1503 int attr_mask, struct ib_qp_init_attr *qp_init_attr) ocrdma_query_qp()
1723 int attr_mask; ocrdma_destroy_qp() local
1734 attr_mask = IB_QP_STATE; ocrdma_destroy_qp()
1735 _ocrdma_modify_qp(ibqp, &attrs, attr_mask); ocrdma_destroy_qp()
1410 _ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask) _ocrdma_modify_qp() argument
1432 ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) ocrdma_modify_qp() argument
1501 ocrdma_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int attr_mask, struct ib_qp_init_attr *qp_init_attr) ocrdma_query_qp() argument
H A Docrdma_main.c569 int attr_mask = IB_QP_STATE; ocrdma_close() local
579 _ocrdma_modify_qp(&qp->ibqp, &attrs, attr_mask); ocrdma_close()
/linux-4.1.27/drivers/infiniband/hw/mlx5/
H A Dqp.c1322 int attr_mask) to_mlx5_access_flags()
1328 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) to_mlx5_access_flags()
1333 if (attr_mask & IB_QP_ACCESS_FLAGS) to_mlx5_access_flags()
1377 struct mlx5_qp_path *path, u8 port, int attr_mask, mlx5_set_path()
1387 if (attr_mask & IB_QP_PKEY_INDEX) mlx5_set_path()
1414 if (attr_mask & IB_QP_TIMEOUT) mlx5_set_path()
1566 const struct ib_qp_attr *attr, int attr_mask, __mlx5_ib_modify_qp()
1594 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) { __mlx5_ib_modify_qp()
1615 } else if (attr_mask & IB_QP_PATH_MTU) { __mlx5_ib_modify_qp()
1625 if (attr_mask & IB_QP_DEST_QPN) __mlx5_ib_modify_qp()
1628 if (attr_mask & IB_QP_PKEY_INDEX) __mlx5_ib_modify_qp()
1636 if (attr_mask & IB_QP_PORT) __mlx5_ib_modify_qp()
1639 if (attr_mask & IB_QP_AV) { __mlx5_ib_modify_qp()
1641 attr_mask & IB_QP_PORT ? attr->port_num : qp->port, __mlx5_ib_modify_qp()
1642 attr_mask, 0, attr); __mlx5_ib_modify_qp()
1647 if (attr_mask & IB_QP_TIMEOUT) __mlx5_ib_modify_qp()
1650 if (attr_mask & IB_QP_ALT_PATH) { __mlx5_ib_modify_qp()
1652 attr->alt_port_num, attr_mask, 0, attr); __mlx5_ib_modify_qp()
1665 if (attr_mask & IB_QP_RNR_RETRY) __mlx5_ib_modify_qp()
1668 if (attr_mask & IB_QP_RETRY_CNT) __mlx5_ib_modify_qp()
1671 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC) { __mlx5_ib_modify_qp()
1677 if (attr_mask & IB_QP_SQ_PSN) __mlx5_ib_modify_qp()
1680 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) { __mlx5_ib_modify_qp()
1686 if (attr_mask & (IB_QP_ACCESS_FLAGS | IB_QP_MAX_DEST_RD_ATOMIC)) __mlx5_ib_modify_qp()
1687 context->params2 |= to_mlx5_access_flags(qp, attr, attr_mask); __mlx5_ib_modify_qp()
1689 if (attr_mask & IB_QP_MIN_RNR_TIMER) __mlx5_ib_modify_qp()
1692 if (attr_mask & IB_QP_RQ_PSN) __mlx5_ib_modify_qp()
1695 if (attr_mask & IB_QP_QKEY) __mlx5_ib_modify_qp()
1702 attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY && attr->en_sqd_async_notify) __mlx5_ib_modify_qp()
1726 optpar = ib_mask_to_mlx5_opt(attr_mask); __mlx5_ib_modify_qp()
1740 if (attr_mask & IB_QP_ACCESS_FLAGS) __mlx5_ib_modify_qp()
1742 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) __mlx5_ib_modify_qp()
1744 if (attr_mask & IB_QP_PORT) __mlx5_ib_modify_qp()
1746 if (attr_mask & IB_QP_ALT_PATH) __mlx5_ib_modify_qp()
1775 int attr_mask, struct ib_udata *udata) mlx5_ib_modify_qp()
1787 cur_state = attr_mask & IB_QP_CUR_STATE ? attr->cur_qp_state : qp->state; mlx5_ib_modify_qp()
1788 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; mlx5_ib_modify_qp()
1791 !ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask, mlx5_ib_modify_qp()
1795 if ((attr_mask & IB_QP_PORT) && mlx5_ib_modify_qp()
1799 if (attr_mask & IB_QP_PKEY_INDEX) { mlx5_ib_modify_qp()
1800 port = attr_mask & IB_QP_PORT ? attr->port_num : qp->port; mlx5_ib_modify_qp()
1805 if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC && mlx5_ib_modify_qp()
1809 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC && mlx5_ib_modify_qp()
1818 err = __mlx5_ib_modify_qp(ibqp, attr, attr_mask, cur_state, new_state); mlx5_ib_modify_qp()
1321 to_mlx5_access_flags(struct mlx5_ib_qp *qp, const struct ib_qp_attr *attr, int attr_mask) to_mlx5_access_flags() argument
1376 mlx5_set_path(struct mlx5_ib_dev *dev, const struct ib_ah_attr *ah, struct mlx5_qp_path *path, u8 port, int attr_mask, u32 path_flags, const struct ib_qp_attr *attr) mlx5_set_path() argument
1565 __mlx5_ib_modify_qp(struct ib_qp *ibqp, const struct ib_qp_attr *attr, int attr_mask, enum ib_qp_state cur_state, enum ib_qp_state new_state) __mlx5_ib_modify_qp() argument
1774 mlx5_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) mlx5_ib_modify_qp() argument
H A Dsrq.c345 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx5_ib_modify_srq()
352 if (attr_mask & IB_SRQ_MAX_WR) mlx5_ib_modify_srq()
355 if (attr_mask & IB_SRQ_LIMIT) { mlx5_ib_modify_srq()
344 mlx5_ib_modify_srq(struct ib_srq *ibsrq, struct ib_srq_attr *attr, enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) mlx5_ib_modify_srq() argument
H A Dmlx5_ib.h539 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
548 int attr_mask, struct ib_udata *udata);
/linux-4.1.27/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.h46 int attr_mask, struct ib_udata *udata);
H A Dusnic_ib_verbs.c547 int attr_mask, struct ib_udata *udata) usnic_ib_modify_qp()
557 if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_INIT) { usnic_ib_modify_qp()
559 } else if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_RTR) { usnic_ib_modify_qp()
561 } else if ((attr_mask & IB_QP_STATE) && attr->qp_state == IB_QPS_RTS) { usnic_ib_modify_qp()
565 attr_mask & IB_QP_STATE, attr->qp_state); usnic_ib_modify_qp()
546 usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) usnic_ib_modify_qp() argument
/linux-4.1.27/include/uapi/rdma/
H A Dib_user_verbs.h544 __u32 attr_mask; member in struct:ib_uverbs_query_qp
586 __u32 attr_mask; member in struct:ib_uverbs_modify_qp
873 __u32 attr_mask; member in struct:ib_uverbs_modify_srq
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
H A Dqp.c1814 int attr_mask, struct ib_udata *udata) c4iw_ib_modify_qp()
1824 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) c4iw_ib_modify_qp()
1825 attr_mask &= ~IB_QP_STATE; c4iw_ib_modify_qp()
1828 if (!attr_mask) c4iw_ib_modify_qp()
1843 mask |= (attr_mask & IB_QP_STATE) ? C4IW_QP_ATTR_NEXT_STATE : 0; c4iw_ib_modify_qp()
1844 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? c4iw_ib_modify_qp()
1856 mask |= (attr_mask & IB_QP_SQ_PSN) ? C4IW_QP_ATTR_SQ_DB : 0; c4iw_ib_modify_qp()
1857 mask |= (attr_mask & IB_QP_RQ_PSN) ? C4IW_QP_ATTR_RQ_DB : 0; c4iw_ib_modify_qp()
1872 int attr_mask, struct ib_qp_init_attr *init_attr) c4iw_ib_query_qp()
1813 c4iw_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) c4iw_ib_modify_qp() argument
1871 c4iw_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) c4iw_ib_query_qp() argument
H A Diw_cxgb4.h1006 int attr_mask, struct ib_udata *udata);
1008 int attr_mask, struct ib_qp_init_attr *init_attr);
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
H A Diwch_provider.c1047 int attr_mask, struct ib_udata *udata) iwch_ib_modify_qp()
1057 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) iwch_ib_modify_qp()
1058 attr_mask &= ~IB_QP_STATE; iwch_ib_modify_qp()
1061 if (!attr_mask) iwch_ib_modify_qp()
1076 mask |= (attr_mask & IB_QP_STATE) ? IWCH_QP_ATTR_NEXT_STATE : 0; iwch_ib_modify_qp()
1077 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? iwch_ib_modify_qp()
1046 iwch_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) iwch_ib_modify_qp() argument
/linux-4.1.27/drivers/infiniband/hw/amso1100/
H A Dc2_provider.c559 int attr_mask, struct ib_udata *udata) c2_modify_qp()
565 attr_mask); c2_modify_qp()
558 c2_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) c2_modify_qp() argument
H A Dc2_qp.c135 struct ib_qp_attr *attr, int attr_mask) c2_qp_modify()
163 if (attr_mask & IB_QP_STATE) { c2_qp_modify()
185 } else if (attr_mask & IB_QP_CUR_STATE) { c2_qp_modify()
134 c2_qp_modify(struct c2_dev *c2dev, struct c2_qp *qp, struct ib_qp_attr *attr, int attr_mask) c2_qp_modify() argument
H A Dc2.h493 struct ib_qp_attr *attr, int attr_mask);
/linux-4.1.27/drivers/infiniband/ulp/srp/
H A Dib_srp.c2147 static uint32_t srp_compute_rq_tmo(struct ib_qp_attr *qp_attr, int attr_mask) srp_compute_rq_tmo() argument
2157 WARN_ON_ONCE((attr_mask & (IB_QP_TIMEOUT | IB_QP_RETRY_CNT)) != srp_compute_rq_tmo()
2180 int attr_mask = 0; srp_cm_rep_handler() local
2217 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); srp_cm_rep_handler()
2221 ret = ib_modify_qp(ch->qp, qp_attr, attr_mask); srp_cm_rep_handler()
2234 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); srp_cm_rep_handler()
2238 target->rq_tmo_jiffies = srp_compute_rq_tmo(qp_attr, attr_mask); srp_cm_rep_handler()
2240 ret = ib_modify_qp(ch->qp, qp_attr, attr_mask); srp_cm_rep_handler()
/linux-4.1.27/drivers/infiniband/core/
H A Duverbs_cmd.c1938 ret = ib_query_qp(qp, attr, cmd.attr_mask, init_attr); ib_uverbs_query_qp()
2098 ret = ib_resolve_eth_l2_attrs(qp, attr, &cmd.attr_mask); ib_uverbs_modify_qp()
2102 modify_qp_mask(qp->qp_type, cmd.attr_mask), &udata); ib_uverbs_modify_qp()
2104 ret = ib_modify_qp(qp, attr, modify_qp_mask(qp->qp_type, cmd.attr_mask)); ib_uverbs_modify_qp()
3202 ret = srq->device->modify_srq(srq, &attr, cmd.attr_mask, &udata); ib_uverbs_modify_srq()
/linux-4.1.27/drivers/infiniband/hw/nes/
H A Dnes_verbs.c2820 int attr_mask, struct ib_qp_init_attr *init_attr) nes_query_qp()
2927 int attr_mask, struct ib_udata *udata) nes_modify_qp()
2950 " QP Access Flags=0x%X, attr_mask = 0x%0x\n", nes_modify_qp()
2952 nesqp->hw_tcp_state, attr->qp_access_flags, attr_mask); nes_modify_qp()
2954 if (attr_mask & IB_QP_STATE) { nes_modify_qp()
3090 if (attr_mask & IB_QP_ACCESS_FLAGS) { nes_modify_qp()
2819 nes_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_qp_init_attr *init_attr) nes_query_qp() argument
2926 nes_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata) nes_modify_qp() argument
/linux-4.1.27/drivers/infiniband/ulp/srpt/
H A Dib_srpt.c988 int attr_mask; srpt_ch_qp_rtr() local
992 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); srpt_ch_qp_rtr()
998 ret = ib_modify_qp(qp, &qp_attr, attr_mask); srpt_ch_qp_rtr()
1018 int attr_mask; srpt_ch_qp_rts() local
1022 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); srpt_ch_qp_rts()
1028 ret = ib_modify_qp(qp, &qp_attr, attr_mask); srpt_ch_qp_rts()

Completed in 1798 milliseconds