Home
last modified time | relevance | path

Searched refs:ibqp (Results 1 – 85 of 85) sorted by relevance

/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_mcast.c59 int ehca_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ehca_attach_mcast() argument
61 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); in ehca_attach_mcast()
62 struct ehca_shca *shca = container_of(ibqp->device, struct ehca_shca, in ehca_attach_mcast()
67 if (ibqp->qp_type != IB_QPT_UD) { in ehca_attach_mcast()
68 ehca_err(ibqp->device, "invalid qp_type=%x", ibqp->qp_type); in ehca_attach_mcast()
73 ehca_err(ibqp->device, "invalid mulitcast gid"); in ehca_attach_mcast()
76 ehca_err(ibqp->device, "invalid mulitcast lid=%x", lid); in ehca_attach_mcast()
89 ehca_err(ibqp->device, in ehca_attach_mcast()
91 "h_ret=%lli", my_qp, ibqp->qp_num, h_ret); in ehca_attach_mcast()
96 int ehca_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ehca_detach_mcast() argument
[all …]
Dehca_qp.c1237 static int internal_modify_qp(struct ib_qp *ibqp, in internal_modify_qp() argument
1245 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); in internal_modify_qp()
1247 container_of(ibqp->pd->device, struct ehca_shca, ib_device); in internal_modify_qp()
1258 ehca_err(ibqp->device, "Could not get zeroed page for mqpcb " in internal_modify_qp()
1259 "ehca_qp=%p qp_num=%x ", my_qp, ibqp->qp_num); in internal_modify_qp()
1268 ehca_err(ibqp->device, "hipz_h_query_qp() failed " in internal_modify_qp()
1270 my_qp, ibqp->qp_num, h_ret); in internal_modify_qp()
1274 if (ibqp->uobject) in internal_modify_qp()
1281 ehca_err(ibqp->device, "Invalid current ehca_qp_state=%x " in internal_modify_qp()
1283 mqpcb->qp_state, my_qp, ibqp->qp_num); in internal_modify_qp()
[all …]
Dehca_iverbs.h154 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
181 u64 ehca_define_sqp(struct ehca_shca *shca, struct ehca_qp *ibqp,
Dehca_main.c555 struct ib_qp *ibqp; in ehca_create_aqp1() local
594 ibqp = ib_create_qp(&shca->pd->ib_pd, &qp_init_attr); in ehca_create_aqp1()
595 if (IS_ERR(ibqp)) { in ehca_create_aqp1()
597 ret = PTR_ERR(ibqp); in ehca_create_aqp1()
600 sport->ibqp_sqp[IB_QPT_GSI] = ibqp; in ehca_create_aqp1()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_ud.c51 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); in qib_ud_loopback()
68 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? in qib_ud_loopback()
69 IB_QPT_UD : sqp->ibqp.qp_type; in qib_ud_loopback()
70 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? in qib_ud_loopback()
71 IB_QPT_UD : qp->ibqp.qp_type; in qib_ud_loopback()
82 if (qp->ibqp.qp_num > 1) { in qib_ud_loopback()
94 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback()
106 if (qp->ibqp.qp_num) { in qib_ud_loopback()
118 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback()
154 if (qp->ibqp.qp_num == 0) in qib_ud_loopback()
[all …]
Dqib_ruc.c90 rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_init_sge()
91 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in qib_init_sge()
121 wc.qp = &qp->ibqp; in qib_init_sge()
123 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in qib_init_sge()
150 if (qp->ibqp.srq) { in qib_get_rwqe()
151 srq = to_isrq(qp->ibqp.srq); in qib_get_rwqe()
213 ev.device = qp->ibqp.device; in qib_get_rwqe()
214 ev.element.srq = qp->ibqp.srq; in qib_get_rwqe()
239 ev.device = qp->ibqp.device; in qib_migrate_qp()
240 ev.element.qp = &qp->ibqp; in qib_migrate_qp()
[all …]
Dqib_qp.c225 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in insert_qp()
227 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); in insert_qp()
232 if (qp->ibqp.qp_num == 0) in insert_qp()
234 else if (qp->ibqp.qp_num == 1) in insert_qp()
250 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in remove_qp()
251 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); in remove_qp()
357 if (qp->ibqp.qp_num == qpn) { in qib_lookup_qpn()
437 if (qp->ibqp.qp_type == IB_QPT_UD || in clear_mr_refs()
438 qp->ibqp.qp_type == IB_QPT_SMI || in clear_mr_refs()
439 qp->ibqp.qp_type == IB_QPT_GSI) in clear_mr_refs()
[all …]
Dqib_verbs_mcast.c225 int qib_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in qib_multicast_attach() argument
227 struct qib_qp *qp = to_iqp(ibqp); in qib_multicast_attach()
228 struct qib_ibdev *dev = to_idev(ibqp->device); in qib_multicast_attach()
234 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in qib_multicast_attach()
254 ibp = to_iport(ibqp->device, qp->port_num); in qib_multicast_attach()
283 int qib_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in qib_multicast_detach() argument
285 struct qib_qp *qp = to_iqp(ibqp); in qib_multicast_detach()
286 struct qib_ibdev *dev = to_idev(ibqp->device); in qib_multicast_detach()
287 struct qib_ibport *ibp = to_iport(ibqp->device, qp->port_num); in qib_multicast_detach()
294 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) in qib_multicast_detach()
Dqib_keys.c255 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_rkey_ok()
266 struct qib_pd *pd = to_ipd(qp->ibqp.pd); in qib_rkey_ok()
289 if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) in qib_rkey_ok()
343 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_reg_mr()
344 struct qib_pd *pd = to_ipd(qp->ibqp.pd); in qib_reg_mr()
361 if (unlikely(mrg == NULL || qp->ibqp.pd != mrg->pd)) in qib_reg_mr()
Dqib_verbs.c368 } else if (qp->ibqp.qp_type == IB_QPT_UC) { in qib_post_one_send()
371 } else if (qp->ibqp.qp_type != IB_QPT_RC) { in qib_post_one_send()
377 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in qib_post_one_send()
397 rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_post_one_send()
398 pd = to_ipd(qp->ibqp.pd); in qib_post_one_send()
401 if (qp->ibqp.qp_type != IB_QPT_UC && in qib_post_one_send()
402 qp->ibqp.qp_type != IB_QPT_RC) in qib_post_one_send()
437 if (qp->ibqp.qp_type == IB_QPT_UC || in qib_post_one_send()
438 qp->ibqp.qp_type == IB_QPT_RC) { in qib_post_one_send()
441 } else if (wqe->length > (dd_from_ibdev(qp->ibqp.device)->pport + in qib_post_one_send()
[all …]
Dqib_uc.c333 if (qp->ibqp.event_handler) { in qib_uc_rcv()
336 ev.device = qp->ibqp.device; in qib_uc_rcv()
337 ev.element.qp = &qp->ibqp; in qib_uc_rcv()
339 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_uc_rcv()
407 wc.qp = &qp->ibqp; in qib_uc_rcv()
417 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in qib_uc_rcv()
Dqib_rc.c233 struct qib_ibdev *dev = to_idev(qp->ibqp.device); in qib_make_rc_req()
652 struct qib_devdata *dd = dd_from_ibdev(qp->ibqp.device); in qib_send_rc_ack()
653 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_send_rc_ack()
888 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_restart_rc()
914 ibp = to_iport(qp->ibqp.device, qp->port_num); in rc_timeout()
1027 wc.qp = &qp->ibqp; in qib_rc_send_complete()
1028 qib_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in qib_rc_send_complete()
1083 wc.qp = &qp->ibqp; in do_rc_completion()
1084 qib_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_completion()
1155 ibp = to_iport(qp->ibqp.device, qp->port_num); in do_rc_ack()
[all …]
Dqib_verbs.h429 struct ib_qp ibqp; member
841 static inline struct qib_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument
843 return container_of(ibqp, struct qib_qp, ibqp); in to_iqp()
910 int qib_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
912 int qib_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
924 int qib_destroy_qp(struct ib_qp *ibqp);
928 int qib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
931 int qib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
Dqib_driver.c370 switch (qp->ibqp.qp_type) { in qib_rcv_hdrerr()
Dqib_sdma.c669 if (qp->ibqp.qp_type == IB_QPT_RC) { in qib_sdma_verbs_send()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dud.c70 struct hfi1_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); in ud_loopback()
90 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? in ud_loopback()
91 IB_QPT_UD : sqp->ibqp.qp_type; in ud_loopback()
92 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? in ud_loopback()
93 IB_QPT_UD : qp->ibqp.qp_type; in ud_loopback()
104 if (qp->ibqp.qp_num > 1) { in ud_loopback()
116 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback()
128 if (qp->ibqp.qp_num) { in ud_loopback()
140 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback()
176 if (qp->ibqp.qp_num == 0) in ud_loopback()
[all …]
Dqp.c234 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in insert_qp()
240 if (qp->ibqp.qp_num <= 1) { in insert_qp()
241 rcu_assign_pointer(ibp->qp[qp->ibqp.qp_num], qp); in insert_qp()
243 u32 n = qpn_hash(dev->qp_dev, qp->ibqp.qp_num); in insert_qp()
259 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in remove_qp()
260 u32 n = qpn_hash(dev->qp_dev, qp->ibqp.qp_num); in remove_qp()
422 if (qp->ibqp.qp_type == IB_QPT_UD || in clear_mr_refs()
423 qp->ibqp.qp_type == IB_QPT_SMI || in clear_mr_refs()
424 qp->ibqp.qp_type == IB_QPT_GSI) in clear_mr_refs()
435 if (qp->ibqp.qp_type != IB_QPT_RC) in clear_mr_refs()
[all …]
Druc.c108 rkt = &to_idev(qp->ibqp.device)->lk_table; in init_sge()
109 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in init_sge()
139 wc.qp = &qp->ibqp; in init_sge()
141 hfi1_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in init_sge()
168 if (qp->ibqp.srq) { in hfi1_get_rwqe()
169 srq = to_isrq(qp->ibqp.srq); in hfi1_get_rwqe()
231 ev.device = qp->ibqp.device; in hfi1_get_rwqe()
232 ev.element.srq = qp->ibqp.srq; in hfi1_get_rwqe()
258 ev.device = qp->ibqp.device; in hfi1_migrate_qp()
259 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp()
[all …]
Dverbs_mcast.c242 int hfi1_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in hfi1_multicast_attach() argument
244 struct hfi1_qp *qp = to_iqp(ibqp); in hfi1_multicast_attach()
245 struct hfi1_ibdev *dev = to_idev(ibqp->device); in hfi1_multicast_attach()
251 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in hfi1_multicast_attach()
271 ibp = to_iport(ibqp->device, qp->port_num); in hfi1_multicast_attach()
300 int hfi1_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in hfi1_multicast_detach() argument
302 struct hfi1_qp *qp = to_iqp(ibqp); in hfi1_multicast_detach()
303 struct hfi1_ibdev *dev = to_idev(ibqp->device); in hfi1_multicast_detach()
304 struct hfi1_ibport *ibp = to_iport(ibqp->device, qp->port_num); in hfi1_multicast_detach()
311 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in hfi1_multicast_detach()
Dqp.h114 if (qp->ibqp.qp_num == qpn) in hfi1_lookup_qpn()
155 int hfi1_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
158 int hfi1_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
191 int hfi1_destroy_qp(struct ib_qp *ibqp);
Dverbs.c370 struct hfi1_devdata *dd = dd_from_ibdev(qp->ibqp.device); in post_one_send()
386 if (qp->ibqp.qp_type == IB_QPT_UC) { in post_one_send()
389 } else if (qp->ibqp.qp_type != IB_QPT_RC) { in post_one_send()
395 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in post_one_send()
413 rkt = &to_idev(qp->ibqp.device)->lk_table; in post_one_send()
414 pd = to_ipd(qp->ibqp.pd); in post_one_send()
418 if (qp->ibqp.qp_type != IB_QPT_UC && in post_one_send()
419 qp->ibqp.qp_type != IB_QPT_RC) in post_one_send()
451 if (qp->ibqp.qp_type == IB_QPT_UC || in post_one_send()
452 qp->ibqp.qp_type == IB_QPT_RC) { in post_one_send()
[all …]
Drc.c262 struct hfi1_ibdev *dev = to_idev(qp->ibqp.device); in hfi1_make_rc_req()
688 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_send_rc_ack()
766 trace_output_ibhdr(dd_from_ibdev(qp->ibqp.device), &hdr); in hfi1_send_rc_ack()
900 ibp = to_iport(qp->ibqp.device, qp->port_num); in restart_rc()
926 ibp = to_iport(qp->ibqp.device, qp->port_num); in rc_timeout()
1041 wc.qp = &qp->ibqp; in hfi1_rc_send_complete()
1042 hfi1_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in hfi1_rc_send_complete()
1098 wc.qp = &qp->ibqp; in do_rc_completion()
1099 hfi1_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_completion()
1186 ibp = to_iport(qp->ibqp.device, qp->port_num); in do_rc_ack()
[all …]
Dkeys.c277 struct hfi1_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in hfi1_rkey_ok()
288 struct hfi1_pd *pd = to_ipd(qp->ibqp.pd); in hfi1_rkey_ok()
310 if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) in hfi1_rkey_ok()
Dverbs.h436 struct ib_qp ibqp; member
825 static inline struct hfi1_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument
827 return container_of(ibqp, struct hfi1_qp, ibqp); in to_iqp()
921 int hfi1_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
923 int hfi1_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
Dtrace.h300 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
306 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device))
308 __entry->qpn = qp->ibqp.qp_num;
334 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
339 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device))
340 __entry->qpn = qp->ibqp.qp_num;
1259 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
1270 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device))
1271 __entry->qpn = qp->ibqp.qp_num;
Ddriver.c326 switch (qp->ibqp.qp_type) { in rcv_hdrerr()
391 switch (qp->ibqp.qp_type) { in rcv_hdrerr()
446 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in process_ecn()
451 switch (qp->ibqp.qp_type) { in process_ecn()
Duc.c447 wc.qp = &qp->ibqp; in hfi1_uc_rcv()
468 hfi1_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in hfi1_uc_rcv()
Ddiag.c1640 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in snoop_send_pio_handler()
1773 } else if (qp->ibqp.qp_type == IB_QPT_RC) { in snoop_send_pio_handler()
/linux-4.4.14/drivers/staging/rdma/ipath/
Dipath_ruc.c88 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_insert_rnr_queue()
148 wc.qp = &qp->ibqp; in ipath_init_sge()
150 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_init_sge()
176 if (qp->ibqp.srq) { in ipath_get_rwqe()
177 srq = to_isrq(qp->ibqp.srq); in ipath_get_rwqe()
235 ev.device = qp->ibqp.device; in ipath_get_rwqe()
236 ev.element.srq = qp->ibqp.srq; in ipath_get_rwqe()
261 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); in ipath_ruc_loopback()
317 if (sqp->ibqp.qp_type == IB_QPT_RC) in ipath_ruc_loopback()
438 wc.qp = &qp->ibqp; in ipath_ruc_loopback()
[all …]
Dipath_qp.c217 qp->ibqp.qp_num = ret; in ipath_alloc_qpn()
250 qpp = &qpt->table[qp->ibqp.qp_num % qpt->max]; in ipath_free_qp()
308 if (qp->ibqp.qp_num == qpn) { in ipath_lookup_qpn()
378 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_error_qp()
399 wc.qp = &qp->ibqp; in ipath_error_qp()
405 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp()
428 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp()
433 } else if (qp->ibqp.event_handler) in ipath_error_qp()
449 int ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in ipath_modify_qp() argument
452 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_modify_qp()
[all …]
Dipath_ud.c51 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); in ipath_ud_loopback()
78 if (unlikely(qp->ibqp.qp_num && in ipath_ud_loopback()
106 if (qp->ibqp.srq) { in ipath_ud_loopback()
107 srq = to_isrq(qp->ibqp.srq); in ipath_ud_loopback()
168 ev.device = qp->ibqp.device; in ipath_ud_loopback()
169 ev.element.srq = qp->ibqp.srq; in ipath_ud_loopback()
214 wc.qp = &qp->ibqp; in ipath_ud_loopback()
215 wc.src_qp = sqp->ibqp.qp_num; in ipath_ud_loopback()
226 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in ipath_ud_loopback()
242 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_make_ud_req()
[all …]
Dipath_keys.c124 struct ipath_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in ipath_lkey_ok()
136 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); in ipath_lkey_ok()
151 qp->ibqp.pd != mr->pd)) { in ipath_lkey_ok()
202 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_rkey_ok()
216 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); in ipath_rkey_ok()
234 qp->ibqp.pd != mr->pd)) { in ipath_rkey_ok()
Dipath_verbs_mcast.c235 int ipath_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ipath_multicast_attach() argument
237 struct ipath_qp *qp = to_iqp(ibqp); in ipath_multicast_attach()
238 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_multicast_attach()
284 int ipath_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ipath_multicast_detach() argument
286 struct ipath_qp *qp = to_iqp(ibqp); in ipath_multicast_detach()
287 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_multicast_detach()
Dipath_verbs.h364 struct ib_qp ibqp; member
686 static inline struct ipath_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument
688 return container_of(ibqp, struct ipath_qp, ibqp); in to_iqp()
734 int ipath_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
736 int ipath_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
748 int ipath_destroy_qp(struct ib_qp *ibqp);
752 int ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
755 int ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
Dipath_verbs.c345 struct ipath_devdata *dd = to_idev(qp->ibqp.device)->dd; in ipath_post_one_send()
349 if (qp->ibqp.qp_type != IB_QPT_SMI && in ipath_post_one_send()
368 if (qp->ibqp.qp_type == IB_QPT_UC) { in ipath_post_one_send()
371 } else if (qp->ibqp.qp_type == IB_QPT_UD) { in ipath_post_one_send()
377 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in ipath_post_one_send()
399 if (qp->ibqp.qp_type != IB_QPT_UC && in ipath_post_one_send()
400 qp->ibqp.qp_type != IB_QPT_RC) in ipath_post_one_send()
431 if (qp->ibqp.qp_type == IB_QPT_UC || in ipath_post_one_send()
432 qp->ibqp.qp_type == IB_QPT_RC) { in ipath_post_one_send()
435 } else if (wqe->length > to_idev(qp->ibqp.device)->dd->ipath_ibmtu) in ipath_post_one_send()
[all …]
Dipath_rc.c68 dev = to_idev(qp->ibqp.device); in ipath_init_restart()
215 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_make_rc_req()
614 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in send_rc_ack()
821 dev = to_idev(qp->ibqp.device); in ipath_restart_rc()
860 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in do_rc_ack()
956 wc.qp = &qp->ibqp; in do_rc_ack()
960 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_ack()
1537 ev.device = qp->ibqp.device; in ipath_rc_error()
1538 ev.element.qp = &qp->ibqp; in ipath_rc_error()
1540 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in ipath_rc_error()
[all …]
Dipath_uc.c213 ipath_make_ruc_header(to_idev(qp->ibqp.device), in ipath_make_uc_req()
415 wc.qp = &qp->ibqp; in ipath_uc_rcv()
420 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in ipath_uc_rcv()
/linux-4.4.14/drivers/infiniband/hw/mthca/
Dmthca_qp.c260 event.element.qp = &qp->ibqp; in mthca_qp_event()
261 if (qp->ibqp.event_handler) in mthca_qp_event()
262 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event()
424 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, in mthca_query_qp() argument
427 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_query_qp()
428 struct mthca_qp *qp = to_mqp(ibqp); in mthca_query_qp()
543 static int __mthca_modify_qp(struct ib_qp *ibqp, in __mthca_modify_qp() argument
547 struct mthca_dev *dev = to_mdev(ibqp->device); in __mthca_modify_qp()
548 struct mthca_qp *qp = to_mqp(ibqp); in __mthca_modify_qp()
609 if (qp->ibqp.uobject) in __mthca_modify_qp()
[all …]
Dmthca_mcg.c120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_attach() argument
122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach()
165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { in mthca_multicast_attach()
167 ibqp->qp_num); in mthca_multicast_attach()
171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); in mthca_multicast_attach()
214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_detach() argument
216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach()
242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) in mthca_multicast_detach()
249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num); in mthca_multicast_detach()
Dmthca_provider.h262 struct ib_qp ibqp; member
334 static inline struct mthca_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument
336 return container_of(ibqp, struct mthca_qp, ibqp); in to_mqp()
Dmthca_dev.h529 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
533 int mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
535 int mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
537 int mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
539 int mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
573 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
574 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
Dmthca_cq.c536 entry->qp = &(*cur_qp)->ibqp; in mthca_poll_one()
544 } else if ((*cur_qp)->ibqp.srq) { in mthca_poll_one()
545 struct mthca_srq *srq = to_msrq((*cur_qp)->ibqp.srq); in mthca_poll_one()
Dmthca_provider.c588 qp->ibqp.qp_num = qp->qpn; in mthca_create_qp()
602 qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1; in mthca_create_qp()
608 qp->ibqp.qp_num, init_attr->port_num, in mthca_create_qp()
628 return &qp->ibqp; in mthca_create_qp()
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Dqp.c248 if (qp->ibqp.qp_type == IB_QPT_UD) { in post_nop_wqe()
252 av->port_pd = cpu_to_be32((qp->port << 24) | to_mpd(qp->ibqp.pd)->pdn); in post_nop_wqe()
289 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx4_ib_qp_event() local
294 if (ibqp->event_handler) { in mlx4_ib_qp_event()
295 event.device = ibqp->device; in mlx4_ib_qp_event()
296 event.element.qp = ibqp; in mlx4_ib_qp_event()
328 ibqp->event_handler(&event, ibqp->qp_context); in mlx4_ib_qp_event()
975 if (qp->ibqp.qp_type == IB_QPT_XRC_TGT) in get_pd()
976 return to_mpd(to_mxrcd(qp->ibqp.xrcd)->pd); in get_pd()
978 return to_mpd(qp->ibqp.pd); in get_pd()
[all …]
Dmlx4_ib.h287 struct ib_qp ibqp; member
660 static inline struct mlx4_ib_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument
662 return container_of(ibqp, struct mlx4_ib_qp, ibqp); in to_mqp()
747 int mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
749 int mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
751 int mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
753 int mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
Dmain.c1229 static int add_gid_entry(struct ib_qp *ibqp, union ib_gid *gid) in add_gid_entry() argument
1231 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in add_gid_entry()
1232 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in add_gid_entry()
1705 static int mlx4_ib_mcg_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx4_ib_mcg_attach() argument
1708 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_attach()
1710 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in mlx4_ib_mcg_attach()
1742 err = add_gid_entry(ibqp, gid); in mlx4_ib_mcg_attach()
1783 static int mlx4_ib_mcg_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx4_ib_mcg_detach() argument
1786 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_detach()
1788 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in mlx4_ib_mcg_detach()
[all …]
Dcq.c584 ib_dma_sync_single_for_cpu(qp->ibqp.device, in use_tunnel_data()
626 wc->qp = &qp->ibqp; in mlx4_ib_qp_sw_comp()
732 wc->qp = &(*cur_qp)->ibqp; in mlx4_ib_poll_one()
756 } else if ((*cur_qp)->ibqp.srq) { in mlx4_ib_poll_one()
757 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx4_ib_poll_one()
/linux-4.4.14/drivers/infiniband/hw/usnic/
Dusnic_ib_qp_grp.h49 struct ib_qp ibqp; member
128 struct usnic_ib_qp_grp *to_uqp_grp(struct ib_qp *ibqp) in to_uqp_grp() argument
130 return container_of(ibqp, struct usnic_ib_qp_grp, ibqp); in to_uqp_grp()
Dusnic_ib_verbs.h63 int usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
82 int usnic_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
84 int usnic_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
Dusnic_ib_qp_grp.c85 qp_grp->ibqp.qp_num, in usnic_ib_qp_grp_dump_rows()
493 ib_event.element.qp = &qp_grp->ibqp; in usnic_ib_qp_grp_modify()
498 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify()
499 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
503 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify()
504 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
510 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify()
511 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify()
733 qp_grp->ibqp.qp_num = qp_grp->grp_id; in usnic_ib_qp_grp_create()
Dusnic_ib_verbs.c390 switch (qp_grp->ibqp.qp_type) { in usnic_ib_query_qp()
395 usnic_err("Unexpected qp_type %d\n", qp_grp->ibqp.qp_type); in usnic_ib_query_qp()
534 return &qp_grp->ibqp; in usnic_ib_create_qp()
565 int usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in usnic_ib_modify_qp() argument
572 qp_grp = to_uqp_grp(ibqp); in usnic_ib_modify_qp()
755 int usnic_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in usnic_ib_post_send() argument
762 int usnic_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr, in usnic_ib_post_recv() argument
Dusnic_ib_sysfs.c258 qp_grp->ibqp.qp_num, in summary_show()
/linux-4.4.14/drivers/staging/rdma/amso1100/
Dc2_provider.h108 struct ib_qp ibqp; member
158 static inline struct c2_qp *to_c2qp(struct ib_qp *ibqp) in to_c2qp() argument
160 return container_of(ibqp, struct c2_qp, ibqp); in to_c2qp()
Dc2_cm.c44 struct ib_qp *ibqp; in c2_llp_connect() local
54 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); in c2_llp_connect()
55 if (!ibqp) in c2_llp_connect()
57 qp = to_c2qp(ibqp); in c2_llp_connect()
292 struct ib_qp *ibqp; in c2_llp_accept() local
298 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); in c2_llp_accept()
299 if (!ibqp) in c2_llp_accept()
301 qp = to_c2qp(ibqp); in c2_llp_accept()
Dc2_provider.c205 static void c2_add_ref(struct ib_qp *ibqp) in c2_add_ref() argument
208 BUG_ON(!ibqp); in c2_add_ref()
209 qp = to_c2qp(ibqp); in c2_add_ref()
213 static void c2_rem_ref(struct ib_qp *ibqp) in c2_rem_ref() argument
216 BUG_ON(!ibqp); in c2_rem_ref()
217 qp = to_c2qp(ibqp); in c2_rem_ref()
232 return (qp?&qp->ibqp:NULL); in c2_get_qp()
278 return &qp->ibqp; in c2_create_qp()
566 static int c2_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in c2_modify_qp() argument
572 c2_qp_modify(to_c2dev(ibqp->device), to_c2qp(ibqp), attr, in c2_modify_qp()
[all …]
Dc2_ae.c235 ib_event.element.qp = &qp->ibqp; in c2_ae_event()
238 if (qp->ibqp.event_handler) in c2_ae_event()
239 qp->ibqp.event_handler(&ib_event, in c2_ae_event()
240 qp->ibqp. in c2_ae_event()
Dc2_qp.c432 qp->ibqp.qp_num = qp->qpn; in c2_alloc_qp()
433 qp->ibqp.qp_type = IB_QPT_RC; in c2_alloc_qp()
605 send_cq = to_c2cq(qp->ibqp.send_cq); in c2_free_qp()
606 recv_cq = to_c2cq(qp->ibqp.recv_cq); in c2_free_qp()
790 int c2_post_send(struct ib_qp *ibqp, struct ib_send_wr *ib_wr, in c2_post_send() argument
793 struct c2_dev *c2dev = to_c2dev(ibqp->device); in c2_post_send()
794 struct c2_qp *qp = to_c2qp(ibqp); in c2_post_send()
944 int c2_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *ib_wr, in c2_post_receive() argument
947 struct c2_dev *c2dev = to_c2dev(ibqp->device); in c2_post_receive()
948 struct c2_qp *qp = to_c2qp(ibqp); in c2_post_receive()
Dc2.h496 extern int c2_post_send(struct ib_qp *ibqp, struct ib_send_wr *ib_wr,
498 extern int c2_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *ib_wr,
Dc2_cq.c158 entry->qp = &qp->ibqp; in c2_poll_one()
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
Diwch_ev.c93 event.element.qp = &qhp->ibqp; in post_qp_event()
95 if (qhp->ibqp.event_handler) in post_qp_event()
96 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
129 iwch_qp_add_ref(&qhp->ibqp); in iwch_ev_dispatch()
229 iwch_qp_rem_ref(&qhp->ibqp); in iwch_ev_dispatch()
Diwch_provider.h162 struct ib_qp ibqp; member
179 static inline struct iwch_qp *to_iwch_qp(struct ib_qp *ibqp) in to_iwch_qp() argument
181 return container_of(ibqp, struct iwch_qp, ibqp); in to_iwch_qp()
329 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
331 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
Diwch_provider.c75 static int iwch_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in iwch_multicast_attach() argument
80 static int iwch_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in iwch_multicast_detach() argument
1029 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp()
1036 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp()
1053 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp()
1065 qhp->ibqp.qp_num = qhp->wq.qpid; in iwch_create_qp()
1072 return &qhp->ibqp; in iwch_create_qp()
1075 static int iwch_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in iwch_ib_modify_qp() argument
1083 PDBG("%s ib_qp %p\n", __func__, ibqp); in iwch_ib_modify_qp()
1094 qhp = to_iwch_qp(ibqp); in iwch_ib_modify_qp()
Diwch_qp.c351 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in iwch_post_send() argument
366 qhp = to_iwch_qp(ibqp); in iwch_post_send()
466 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, in iwch_post_receive() argument
476 qhp = to_iwch_qp(ibqp); in iwch_post_receive()
861 if (qhp->ibqp.uobject) { in flush_qp()
918 if (!qhp->ibqp.uobject) in rdma_init()
1057 if (qhp->ibqp.uobject) in iwch_modify_qp()
Diwch_cq.c82 wc->qp = &qhp->ibqp; in iwch_poll_cq_one()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Dodp.c157 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_page_fault_resume()
183 struct mlx5_ib_dev *mib_dev = to_mdev(qp->ibqp.pd->device); in pagefault_single_data_segment()
212 if (mr->ibmr.pd != qp->ibqp.pd) { in pagefault_single_data_segment()
323 if (receive_queue && qp->ibqp.srq) in pagefault_data_segments()
387 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_mr_initiator_pfault_handler()
434 switch (qp->ibqp.qp_type) { in mlx5_ib_mr_initiator_pfault_handler()
477 qp->ibqp.qp_type, opcode); in mlx5_ib_mr_initiator_pfault_handler()
492 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_mr_responder_pfault_handler()
496 if (qp->ibqp.srq) { in mlx5_ib_mr_responder_pfault_handler()
511 switch (qp->ibqp.qp_type) { in mlx5_ib_mr_responder_pfault_handler()
[all …]
Dqp.c119 struct ib_device *ibdev = qp->ibqp.device; in mlx5_ib_read_user_wqe()
131 qp->ibqp.qp_type); in mlx5_ib_read_user_wqe()
172 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx5_ib_qp_event() local
178 if (ibqp->event_handler) { in mlx5_ib_qp_event()
179 event.device = ibqp->device; in mlx5_ib_qp_event()
180 event.element.qp = ibqp; in mlx5_ib_qp_event()
211 ibqp->event_handler(&event, ibqp->qp_context); in mlx5_ib_qp_event()
1108 return to_mpd(qp->ibqp.pd); in get_pd()
1114 switch (qp->ibqp.qp_type) { in get_cqs()
1121 *send_cq = to_mcq(qp->ibqp.send_cq); in get_cqs()
[all …]
Dmlx5_ib.h175 struct ib_qp ibqp; member
476 static inline struct mlx5_ib_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument
478 return container_of(ibqp, struct mlx5_ib_qp, ibqp); in to_mqp()
527 int mlx5_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
529 int mlx5_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
532 int mlx5_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
534 int mlx5_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
Dcq.c174 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in handle_responder()
180 if (qp->ibqp.srq || qp->ibqp.xrcd) { in handle_responder()
183 if (qp->ibqp.xrcd) { in handle_responder()
188 srq = to_msrq(qp->ibqp.srq); in handle_responder()
231 if (unlikely(is_qp1(qp->ibqp.qp_type))) { in handle_responder()
456 if (!*cur_qp || (qpn != (*cur_qp)->ibqp.qp_num)) { in mlx5_poll_one()
471 wc->qp = &(*cur_qp)->ibqp; in mlx5_poll_one()
510 if ((*cur_qp)->ibqp.srq) { in mlx5_poll_one()
511 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx5_poll_one()
Dmain.c838 static int mlx5_ib_mcg_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx5_ib_mcg_attach() argument
840 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_mcg_attach()
843 err = mlx5_core_attach_mcg(dev->mdev, gid, ibqp->qp_num); in mlx5_ib_mcg_attach()
846 ibqp->qp_num, gid->raw); in mlx5_ib_mcg_attach()
851 static int mlx5_ib_mcg_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx5_ib_mcg_detach() argument
853 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_mcg_detach()
856 err = mlx5_core_detach_mcg(dev->mdev, gid, ibqp->qp_num); in mlx5_ib_mcg_detach()
859 ibqp->qp_num, gid->raw); in mlx5_ib_mcg_detach()
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
Dev.c109 event.element.qp = &qhp->ibqp; in post_qp_event()
110 if (qhp->ibqp.event_handler) in post_qp_event()
111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
153 c4iw_qp_add_ref(&qhp->ibqp); in c4iw_ev_dispatch()
218 c4iw_qp_rem_ref(&qhp->ibqp); in c4iw_ev_dispatch()
Diw_cxgb4.h467 struct ib_qp ibqp; member
481 static inline struct c4iw_qp *to_c4iw_qp(struct ib_qp *ibqp) in to_c4iw_qp() argument
483 return container_of(ibqp, struct c4iw_qp, ibqp); in to_c4iw_qp()
946 int c4iw_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
948 int c4iw_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
994 int c4iw_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
996 int c4iw_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
Dqp.c739 int c4iw_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in c4iw_post_send() argument
753 qhp = to_c4iw_qp(ibqp); in c4iw_post_send()
867 int c4iw_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, in c4iw_post_receive() argument
878 qhp = to_c4iw_qp(ibqp); in c4iw_post_receive()
1181 rchp = to_c4iw_cq(qhp->ibqp.recv_cq); in flush_qp()
1182 schp = to_c4iw_cq(qhp->ibqp.send_cq); in flush_qp()
1185 if (qhp->ibqp.uobject) { in flush_qp()
1312 if (!qhp->ibqp.uobject) in rdma_init()
1797 qhp->ibqp.qp_num = qhp->wq.sq.qid; in c4iw_create_qp()
1805 return &qhp->ibqp; in c4iw_create_qp()
[all …]
Dprovider.c72 static int c4iw_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in c4iw_multicast_attach() argument
77 static int c4iw_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in c4iw_multicast_detach() argument
Dcq.c240 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); in c4iw_flush_sq()
709 wc->qp = &qhp->ibqp; in c4iw_poll_cq_one()
Ddevice.c1344 c4iw_qp_add_ref(&qp->ibqp); in add_and_ref_qp()
1361 c4iw_qp_rem_ref(&qp_list->qps[idx]->ibqp); in deref_qps()
Dcm.c154 c4iw_qp_rem_ref(&ep->com.qp->ibqp); in deref_qp()
161 c4iw_qp_add_ref(&ep->com.qp->ibqp); in ref_qp()
/linux-4.4.14/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1463 qp->ibqp.qp_num = qp->id; in ocrdma_create_qp()
1465 return &qp->ibqp; in ocrdma_create_qp()
1481 int _ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in _ocrdma_modify_qp() argument
1489 qp = get_ocrdma_qp(ibqp); in _ocrdma_modify_qp()
1490 dev = get_ocrdma_dev(ibqp->device); in _ocrdma_modify_qp()
1503 int ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in ocrdma_modify_qp() argument
1512 qp = get_ocrdma_qp(ibqp); in ocrdma_modify_qp()
1513 dev = get_ocrdma_dev(ibqp->device); in ocrdma_modify_qp()
1526 if (!ib_modify_qp_is_ok(old_qps, new_qps, ibqp->qp_type, attr_mask, in ocrdma_modify_qp()
1530 __func__, dev->id, attr_mask, qp->id, ibqp->qp_type, in ocrdma_modify_qp()
[all …]
Docrdma.h392 struct ib_qp ibqp; member
479 static inline struct ocrdma_qp *get_ocrdma_qp(struct ib_qp *ibqp) in get_ocrdma_qp() argument
481 return container_of(ibqp, struct ocrdma_qp, ibqp); in get_ocrdma_qp()
Docrdma_hw.c730 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent()
735 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent()
739 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent()
743 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent()
765 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent()
781 if (qp->ibqp.event_handler) in ocrdma_dispatch_ibevent()
782 qp->ibqp.event_handler(&ib_evt, qp->ibqp.qp_context); in ocrdma_dispatch_ibevent()
2115 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_flush_qp()
2505 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_set_av_params()
2570 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_set_qp_params()
/linux-4.4.14/drivers/infiniband/hw/nes/
Dnes_verbs.c212 static int nes_bind_mw(struct ib_qp *ibqp, struct ib_mw *ibmw, in nes_bind_mw() argument
216 struct nes_vnic *nesvnic = to_nesvnic(ibqp->device); in nes_bind_mw()
219 struct nes_qp *nesqp = to_nesqp(ibqp); in nes_bind_mw()
1245 nesqp->ibqp.qp_num = nesqp->hwqp.qp_id; in nes_create_qp()
1385 nes_add_ref(&nesqp->ibqp); in nes_create_qp()
1402 return &nesqp->ibqp; in nes_create_qp()
1442 static int nes_destroy_qp(struct ib_qp *ibqp) in nes_destroy_qp() argument
1444 struct nes_qp *nesqp = to_nesqp(ibqp); in nes_destroy_qp()
1458 nes_modify_qp(&nesqp->ibqp, &attr, IB_QP_STATE, NULL); in nes_destroy_qp()
1482 if ((ibqp->uobject)&&(ibqp->uobject->context)) { in nes_destroy_qp()
[all …]
Dnes.c266 void nes_add_ref(struct ib_qp *ibqp) in nes_add_ref() argument
270 nesqp = to_nesqp(ibqp); in nes_add_ref()
272 ibqp->qp_num, atomic_read(&nesqp->refcount)); in nes_add_ref()
309 void nes_rem_ref(struct ib_qp *ibqp) in nes_rem_ref() argument
313 struct nes_vnic *nesvnic = to_nesvnic(ibqp->device); in nes_rem_ref()
319 nesqp = to_nesqp(ibqp); in nes_rem_ref()
323 __func__, ibqp->qp_num, nesqp->last_aeq); in nes_rem_ref()
371 return &nesadapter->qp_table[qpn - NES_FIRST_QPN]->ibqp; in nes_get_qp()
Dnes_cm.c1738 nes_rem_ref(&nesqp->ibqp); in rem_ref_cm_node()
3007 nes_add_ref(&nesqp->ibqp); in nes_cm_disconn()
3027 nes_rem_ref(&nesqp->ibqp); in nes_disconnect_worker()
3066 nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_cm_disconn_true()
3111 ibevent.device = nesqp->ibqp.device; in nes_cm_disconn_true()
3113 ibevent.element.qp = &nesqp->ibqp; in nes_cm_disconn_true()
3114 if (nesqp->ibqp.event_handler) in nes_cm_disconn_true()
3115 nesqp->ibqp.event_handler(&ibevent, nesqp->ibqp.qp_context); in nes_cm_disconn_true()
3178 nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_disconnect()
3220 struct ib_qp *ibqp; in nes_accept() local
[all …]
Dnes.h510 static inline struct nes_qp *to_nesqp(struct ib_qp *ibqp) in to_nesqp() argument
512 return container_of(ibqp, struct nes_qp, ibqp); in to_nesqp()
Dnes_verbs.h139 struct ib_qp ibqp; member
Dnes_hw.c3449 struct nes_vnic *nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_terminate_done()