| /linux-4.4.14/drivers/staging/rdma/ehca/ |
| D | ehca_mcast.c | 59 int ehca_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ehca_attach_mcast() argument 61 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); in ehca_attach_mcast() 62 struct ehca_shca *shca = container_of(ibqp->device, struct ehca_shca, in ehca_attach_mcast() 67 if (ibqp->qp_type != IB_QPT_UD) { in ehca_attach_mcast() 68 ehca_err(ibqp->device, "invalid qp_type=%x", ibqp->qp_type); in ehca_attach_mcast() 73 ehca_err(ibqp->device, "invalid mulitcast gid"); in ehca_attach_mcast() 76 ehca_err(ibqp->device, "invalid mulitcast lid=%x", lid); in ehca_attach_mcast() 89 ehca_err(ibqp->device, in ehca_attach_mcast() 91 "h_ret=%lli", my_qp, ibqp->qp_num, h_ret); in ehca_attach_mcast() 96 int ehca_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ehca_detach_mcast() argument [all …]
|
| D | ehca_qp.c | 1237 static int internal_modify_qp(struct ib_qp *ibqp, in internal_modify_qp() argument 1245 struct ehca_qp *my_qp = container_of(ibqp, struct ehca_qp, ib_qp); in internal_modify_qp() 1247 container_of(ibqp->pd->device, struct ehca_shca, ib_device); in internal_modify_qp() 1258 ehca_err(ibqp->device, "Could not get zeroed page for mqpcb " in internal_modify_qp() 1259 "ehca_qp=%p qp_num=%x ", my_qp, ibqp->qp_num); in internal_modify_qp() 1268 ehca_err(ibqp->device, "hipz_h_query_qp() failed " in internal_modify_qp() 1270 my_qp, ibqp->qp_num, h_ret); in internal_modify_qp() 1274 if (ibqp->uobject) in internal_modify_qp() 1281 ehca_err(ibqp->device, "Invalid current ehca_qp_state=%x " in internal_modify_qp() 1283 mqpcb->qp_state, my_qp, ibqp->qp_num); in internal_modify_qp() [all …]
|
| D | ehca_iverbs.h | 154 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, 181 u64 ehca_define_sqp(struct ehca_shca *shca, struct ehca_qp *ibqp,
|
| D | ehca_main.c | 555 struct ib_qp *ibqp; in ehca_create_aqp1() local 594 ibqp = ib_create_qp(&shca->pd->ib_pd, &qp_init_attr); in ehca_create_aqp1() 595 if (IS_ERR(ibqp)) { in ehca_create_aqp1() 597 ret = PTR_ERR(ibqp); in ehca_create_aqp1() 600 sport->ibqp_sqp[IB_QPT_GSI] = ibqp; in ehca_create_aqp1()
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_ud.c | 51 struct qib_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); in qib_ud_loopback() 68 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? in qib_ud_loopback() 69 IB_QPT_UD : sqp->ibqp.qp_type; in qib_ud_loopback() 70 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? in qib_ud_loopback() 71 IB_QPT_UD : qp->ibqp.qp_type; in qib_ud_loopback() 82 if (qp->ibqp.qp_num > 1) { in qib_ud_loopback() 94 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback() 106 if (qp->ibqp.qp_num) { in qib_ud_loopback() 118 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback() 154 if (qp->ibqp.qp_num == 0) in qib_ud_loopback() [all …]
|
| D | qib_ruc.c | 90 rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_init_sge() 91 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in qib_init_sge() 121 wc.qp = &qp->ibqp; in qib_init_sge() 123 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in qib_init_sge() 150 if (qp->ibqp.srq) { in qib_get_rwqe() 151 srq = to_isrq(qp->ibqp.srq); in qib_get_rwqe() 213 ev.device = qp->ibqp.device; in qib_get_rwqe() 214 ev.element.srq = qp->ibqp.srq; in qib_get_rwqe() 239 ev.device = qp->ibqp.device; in qib_migrate_qp() 240 ev.element.qp = &qp->ibqp; in qib_migrate_qp() [all …]
|
| D | qib_qp.c | 225 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in insert_qp() 227 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); in insert_qp() 232 if (qp->ibqp.qp_num == 0) in insert_qp() 234 else if (qp->ibqp.qp_num == 1) in insert_qp() 250 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in remove_qp() 251 unsigned n = qpn_hash(dev, qp->ibqp.qp_num); in remove_qp() 357 if (qp->ibqp.qp_num == qpn) { in qib_lookup_qpn() 437 if (qp->ibqp.qp_type == IB_QPT_UD || in clear_mr_refs() 438 qp->ibqp.qp_type == IB_QPT_SMI || in clear_mr_refs() 439 qp->ibqp.qp_type == IB_QPT_GSI) in clear_mr_refs() [all …]
|
| D | qib_verbs_mcast.c | 225 int qib_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in qib_multicast_attach() argument 227 struct qib_qp *qp = to_iqp(ibqp); in qib_multicast_attach() 228 struct qib_ibdev *dev = to_idev(ibqp->device); in qib_multicast_attach() 234 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in qib_multicast_attach() 254 ibp = to_iport(ibqp->device, qp->port_num); in qib_multicast_attach() 283 int qib_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in qib_multicast_detach() argument 285 struct qib_qp *qp = to_iqp(ibqp); in qib_multicast_detach() 286 struct qib_ibdev *dev = to_idev(ibqp->device); in qib_multicast_detach() 287 struct qib_ibport *ibp = to_iport(ibqp->device, qp->port_num); in qib_multicast_detach() 294 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) in qib_multicast_detach()
|
| D | qib_keys.c | 255 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_rkey_ok() 266 struct qib_pd *pd = to_ipd(qp->ibqp.pd); in qib_rkey_ok() 289 if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) in qib_rkey_ok() 343 struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_reg_mr() 344 struct qib_pd *pd = to_ipd(qp->ibqp.pd); in qib_reg_mr() 361 if (unlikely(mrg == NULL || qp->ibqp.pd != mrg->pd)) in qib_reg_mr()
|
| D | qib_verbs.c | 368 } else if (qp->ibqp.qp_type == IB_QPT_UC) { in qib_post_one_send() 371 } else if (qp->ibqp.qp_type != IB_QPT_RC) { in qib_post_one_send() 377 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in qib_post_one_send() 397 rkt = &to_idev(qp->ibqp.device)->lk_table; in qib_post_one_send() 398 pd = to_ipd(qp->ibqp.pd); in qib_post_one_send() 401 if (qp->ibqp.qp_type != IB_QPT_UC && in qib_post_one_send() 402 qp->ibqp.qp_type != IB_QPT_RC) in qib_post_one_send() 437 if (qp->ibqp.qp_type == IB_QPT_UC || in qib_post_one_send() 438 qp->ibqp.qp_type == IB_QPT_RC) { in qib_post_one_send() 441 } else if (wqe->length > (dd_from_ibdev(qp->ibqp.device)->pport + in qib_post_one_send() [all …]
|
| D | qib_uc.c | 333 if (qp->ibqp.event_handler) { in qib_uc_rcv() 336 ev.device = qp->ibqp.device; in qib_uc_rcv() 337 ev.element.qp = &qp->ibqp; in qib_uc_rcv() 339 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_uc_rcv() 407 wc.qp = &qp->ibqp; in qib_uc_rcv() 417 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in qib_uc_rcv()
|
| D | qib_rc.c | 233 struct qib_ibdev *dev = to_idev(qp->ibqp.device); in qib_make_rc_req() 652 struct qib_devdata *dd = dd_from_ibdev(qp->ibqp.device); in qib_send_rc_ack() 653 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_send_rc_ack() 888 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_restart_rc() 914 ibp = to_iport(qp->ibqp.device, qp->port_num); in rc_timeout() 1027 wc.qp = &qp->ibqp; in qib_rc_send_complete() 1028 qib_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in qib_rc_send_complete() 1083 wc.qp = &qp->ibqp; in do_rc_completion() 1084 qib_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_completion() 1155 ibp = to_iport(qp->ibqp.device, qp->port_num); in do_rc_ack() [all …]
|
| D | qib_verbs.h | 429 struct ib_qp ibqp; member 841 static inline struct qib_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument 843 return container_of(ibqp, struct qib_qp, ibqp); in to_iqp() 910 int qib_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 912 int qib_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 924 int qib_destroy_qp(struct ib_qp *ibqp); 928 int qib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 931 int qib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
|
| D | qib_driver.c | 370 switch (qp->ibqp.qp_type) { in qib_rcv_hdrerr()
|
| D | qib_sdma.c | 669 if (qp->ibqp.qp_type == IB_QPT_RC) { in qib_sdma_verbs_send()
|
| /linux-4.4.14/drivers/staging/rdma/hfi1/ |
| D | ud.c | 70 struct hfi1_ibport *ibp = to_iport(sqp->ibqp.device, sqp->port_num); in ud_loopback() 90 sqptype = sqp->ibqp.qp_type == IB_QPT_GSI ? in ud_loopback() 91 IB_QPT_UD : sqp->ibqp.qp_type; in ud_loopback() 92 dqptype = qp->ibqp.qp_type == IB_QPT_GSI ? in ud_loopback() 93 IB_QPT_UD : qp->ibqp.qp_type; in ud_loopback() 104 if (qp->ibqp.qp_num > 1) { in ud_loopback() 116 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback() 128 if (qp->ibqp.qp_num) { in ud_loopback() 140 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback() 176 if (qp->ibqp.qp_num == 0) in ud_loopback() [all …]
|
| D | qp.c | 234 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in insert_qp() 240 if (qp->ibqp.qp_num <= 1) { in insert_qp() 241 rcu_assign_pointer(ibp->qp[qp->ibqp.qp_num], qp); in insert_qp() 243 u32 n = qpn_hash(dev->qp_dev, qp->ibqp.qp_num); in insert_qp() 259 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in remove_qp() 260 u32 n = qpn_hash(dev->qp_dev, qp->ibqp.qp_num); in remove_qp() 422 if (qp->ibqp.qp_type == IB_QPT_UD || in clear_mr_refs() 423 qp->ibqp.qp_type == IB_QPT_SMI || in clear_mr_refs() 424 qp->ibqp.qp_type == IB_QPT_GSI) in clear_mr_refs() 435 if (qp->ibqp.qp_type != IB_QPT_RC) in clear_mr_refs() [all …]
|
| D | ruc.c | 108 rkt = &to_idev(qp->ibqp.device)->lk_table; in init_sge() 109 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in init_sge() 139 wc.qp = &qp->ibqp; in init_sge() 141 hfi1_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in init_sge() 168 if (qp->ibqp.srq) { in hfi1_get_rwqe() 169 srq = to_isrq(qp->ibqp.srq); in hfi1_get_rwqe() 231 ev.device = qp->ibqp.device; in hfi1_get_rwqe() 232 ev.element.srq = qp->ibqp.srq; in hfi1_get_rwqe() 258 ev.device = qp->ibqp.device; in hfi1_migrate_qp() 259 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp() [all …]
|
| D | verbs_mcast.c | 242 int hfi1_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in hfi1_multicast_attach() argument 244 struct hfi1_qp *qp = to_iqp(ibqp); in hfi1_multicast_attach() 245 struct hfi1_ibdev *dev = to_idev(ibqp->device); in hfi1_multicast_attach() 251 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in hfi1_multicast_attach() 271 ibp = to_iport(ibqp->device, qp->port_num); in hfi1_multicast_attach() 300 int hfi1_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in hfi1_multicast_detach() argument 302 struct hfi1_qp *qp = to_iqp(ibqp); in hfi1_multicast_detach() 303 struct hfi1_ibdev *dev = to_idev(ibqp->device); in hfi1_multicast_detach() 304 struct hfi1_ibport *ibp = to_iport(ibqp->device, qp->port_num); in hfi1_multicast_detach() 311 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) { in hfi1_multicast_detach()
|
| D | qp.h | 114 if (qp->ibqp.qp_num == qpn) in hfi1_lookup_qpn() 155 int hfi1_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 158 int hfi1_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 191 int hfi1_destroy_qp(struct ib_qp *ibqp);
|
| D | verbs.c | 370 struct hfi1_devdata *dd = dd_from_ibdev(qp->ibqp.device); in post_one_send() 386 if (qp->ibqp.qp_type == IB_QPT_UC) { in post_one_send() 389 } else if (qp->ibqp.qp_type != IB_QPT_RC) { in post_one_send() 395 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in post_one_send() 413 rkt = &to_idev(qp->ibqp.device)->lk_table; in post_one_send() 414 pd = to_ipd(qp->ibqp.pd); in post_one_send() 418 if (qp->ibqp.qp_type != IB_QPT_UC && in post_one_send() 419 qp->ibqp.qp_type != IB_QPT_RC) in post_one_send() 451 if (qp->ibqp.qp_type == IB_QPT_UC || in post_one_send() 452 qp->ibqp.qp_type == IB_QPT_RC) { in post_one_send() [all …]
|
| D | rc.c | 262 struct hfi1_ibdev *dev = to_idev(qp->ibqp.device); in hfi1_make_rc_req() 688 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_send_rc_ack() 766 trace_output_ibhdr(dd_from_ibdev(qp->ibqp.device), &hdr); in hfi1_send_rc_ack() 900 ibp = to_iport(qp->ibqp.device, qp->port_num); in restart_rc() 926 ibp = to_iport(qp->ibqp.device, qp->port_num); in rc_timeout() 1041 wc.qp = &qp->ibqp; in hfi1_rc_send_complete() 1042 hfi1_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in hfi1_rc_send_complete() 1098 wc.qp = &qp->ibqp; in do_rc_completion() 1099 hfi1_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_completion() 1186 ibp = to_iport(qp->ibqp.device, qp->port_num); in do_rc_ack() [all …]
|
| D | keys.c | 277 struct hfi1_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in hfi1_rkey_ok() 288 struct hfi1_pd *pd = to_ipd(qp->ibqp.pd); in hfi1_rkey_ok() 310 if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) in hfi1_rkey_ok()
|
| D | verbs.h | 436 struct ib_qp ibqp; member 825 static inline struct hfi1_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument 827 return container_of(ibqp, struct hfi1_qp, ibqp); in to_iqp() 921 int hfi1_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 923 int hfi1_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
|
| D | trace.h | 300 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 306 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)) 308 __entry->qpn = qp->ibqp.qp_num; 334 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 339 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)) 340 __entry->qpn = qp->ibqp.qp_num; 1259 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 1270 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)) 1271 __entry->qpn = qp->ibqp.qp_num;
|
| D | driver.c | 326 switch (qp->ibqp.qp_type) { in rcv_hdrerr() 391 switch (qp->ibqp.qp_type) { in rcv_hdrerr() 446 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in process_ecn() 451 switch (qp->ibqp.qp_type) { in process_ecn()
|
| D | uc.c | 447 wc.qp = &qp->ibqp; in hfi1_uc_rcv() 468 hfi1_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in hfi1_uc_rcv()
|
| D | diag.c | 1640 struct hfi1_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in snoop_send_pio_handler() 1773 } else if (qp->ibqp.qp_type == IB_QPT_RC) { in snoop_send_pio_handler()
|
| /linux-4.4.14/drivers/staging/rdma/ipath/ |
| D | ipath_ruc.c | 88 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_insert_rnr_queue() 148 wc.qp = &qp->ibqp; in ipath_init_sge() 150 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_init_sge() 176 if (qp->ibqp.srq) { in ipath_get_rwqe() 177 srq = to_isrq(qp->ibqp.srq); in ipath_get_rwqe() 235 ev.device = qp->ibqp.device; in ipath_get_rwqe() 236 ev.element.srq = qp->ibqp.srq; in ipath_get_rwqe() 261 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); in ipath_ruc_loopback() 317 if (sqp->ibqp.qp_type == IB_QPT_RC) in ipath_ruc_loopback() 438 wc.qp = &qp->ibqp; in ipath_ruc_loopback() [all …]
|
| D | ipath_qp.c | 217 qp->ibqp.qp_num = ret; in ipath_alloc_qpn() 250 qpp = &qpt->table[qp->ibqp.qp_num % qpt->max]; in ipath_free_qp() 308 if (qp->ibqp.qp_num == qpn) { in ipath_lookup_qpn() 378 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_error_qp() 399 wc.qp = &qp->ibqp; in ipath_error_qp() 405 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp() 428 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp() 433 } else if (qp->ibqp.event_handler) in ipath_error_qp() 449 int ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in ipath_modify_qp() argument 452 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_modify_qp() [all …]
|
| D | ipath_ud.c | 51 struct ipath_ibdev *dev = to_idev(sqp->ibqp.device); in ipath_ud_loopback() 78 if (unlikely(qp->ibqp.qp_num && in ipath_ud_loopback() 106 if (qp->ibqp.srq) { in ipath_ud_loopback() 107 srq = to_isrq(qp->ibqp.srq); in ipath_ud_loopback() 168 ev.device = qp->ibqp.device; in ipath_ud_loopback() 169 ev.element.srq = qp->ibqp.srq; in ipath_ud_loopback() 214 wc.qp = &qp->ibqp; in ipath_ud_loopback() 215 wc.src_qp = sqp->ibqp.qp_num; in ipath_ud_loopback() 226 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in ipath_ud_loopback() 242 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_make_ud_req() [all …]
|
| D | ipath_keys.c | 124 struct ipath_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table; in ipath_lkey_ok() 136 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); in ipath_lkey_ok() 151 qp->ibqp.pd != mr->pd)) { in ipath_lkey_ok() 202 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_rkey_ok() 216 struct ipath_pd *pd = to_ipd(qp->ibqp.pd); in ipath_rkey_ok() 234 qp->ibqp.pd != mr->pd)) { in ipath_rkey_ok()
|
| D | ipath_verbs_mcast.c | 235 int ipath_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ipath_multicast_attach() argument 237 struct ipath_qp *qp = to_iqp(ibqp); in ipath_multicast_attach() 238 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_multicast_attach() 284 int ipath_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in ipath_multicast_detach() argument 286 struct ipath_qp *qp = to_iqp(ibqp); in ipath_multicast_detach() 287 struct ipath_ibdev *dev = to_idev(ibqp->device); in ipath_multicast_detach()
|
| D | ipath_verbs.h | 364 struct ib_qp ibqp; member 686 static inline struct ipath_qp *to_iqp(struct ib_qp *ibqp) in to_iqp() argument 688 return container_of(ibqp, struct ipath_qp, ibqp); in to_iqp() 734 int ipath_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 736 int ipath_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 748 int ipath_destroy_qp(struct ib_qp *ibqp); 752 int ipath_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 755 int ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
|
| D | ipath_verbs.c | 345 struct ipath_devdata *dd = to_idev(qp->ibqp.device)->dd; in ipath_post_one_send() 349 if (qp->ibqp.qp_type != IB_QPT_SMI && in ipath_post_one_send() 368 if (qp->ibqp.qp_type == IB_QPT_UC) { in ipath_post_one_send() 371 } else if (qp->ibqp.qp_type == IB_QPT_UD) { in ipath_post_one_send() 377 if (qp->ibqp.pd != ud_wr(wr)->ah->pd) in ipath_post_one_send() 399 if (qp->ibqp.qp_type != IB_QPT_UC && in ipath_post_one_send() 400 qp->ibqp.qp_type != IB_QPT_RC) in ipath_post_one_send() 431 if (qp->ibqp.qp_type == IB_QPT_UC || in ipath_post_one_send() 432 qp->ibqp.qp_type == IB_QPT_RC) { in ipath_post_one_send() 435 } else if (wqe->length > to_idev(qp->ibqp.device)->dd->ipath_ibmtu) in ipath_post_one_send() [all …]
|
| D | ipath_rc.c | 68 dev = to_idev(qp->ibqp.device); in ipath_init_restart() 215 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in ipath_make_rc_req() 614 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in send_rc_ack() 821 dev = to_idev(qp->ibqp.device); in ipath_restart_rc() 860 struct ipath_ibdev *dev = to_idev(qp->ibqp.device); in do_rc_ack() 956 wc.qp = &qp->ibqp; in do_rc_ack() 960 ipath_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in do_rc_ack() 1537 ev.device = qp->ibqp.device; in ipath_rc_error() 1538 ev.element.qp = &qp->ibqp; in ipath_rc_error() 1540 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in ipath_rc_error() [all …]
|
| D | ipath_uc.c | 213 ipath_make_ruc_header(to_idev(qp->ibqp.device), in ipath_make_uc_req() 415 wc.qp = &qp->ibqp; in ipath_uc_rcv() 420 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, in ipath_uc_rcv()
|
| /linux-4.4.14/drivers/infiniband/hw/mthca/ |
| D | mthca_qp.c | 260 event.element.qp = &qp->ibqp; in mthca_qp_event() 261 if (qp->ibqp.event_handler) in mthca_qp_event() 262 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 424 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, in mthca_query_qp() argument 427 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_query_qp() 428 struct mthca_qp *qp = to_mqp(ibqp); in mthca_query_qp() 543 static int __mthca_modify_qp(struct ib_qp *ibqp, in __mthca_modify_qp() argument 547 struct mthca_dev *dev = to_mdev(ibqp->device); in __mthca_modify_qp() 548 struct mthca_qp *qp = to_mqp(ibqp); in __mthca_modify_qp() 609 if (qp->ibqp.uobject) in __mthca_modify_qp() [all …]
|
| D | mthca_mcg.c | 120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_attach() argument 122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach() 165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { in mthca_multicast_attach() 167 ibqp->qp_num); in mthca_multicast_attach() 171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); in mthca_multicast_attach() 214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_detach() argument 216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach() 242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) in mthca_multicast_detach() 249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num); in mthca_multicast_detach()
|
| D | mthca_provider.h | 262 struct ib_qp ibqp; member 334 static inline struct mthca_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument 336 return container_of(ibqp, struct mthca_qp, ibqp); in to_mqp()
|
| D | mthca_dev.h | 529 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, 533 int mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 535 int mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, 537 int mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 539 int mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, 573 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid); 574 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid);
|
| D | mthca_cq.c | 536 entry->qp = &(*cur_qp)->ibqp; in mthca_poll_one() 544 } else if ((*cur_qp)->ibqp.srq) { in mthca_poll_one() 545 struct mthca_srq *srq = to_msrq((*cur_qp)->ibqp.srq); in mthca_poll_one()
|
| D | mthca_provider.c | 588 qp->ibqp.qp_num = qp->qpn; in mthca_create_qp() 602 qp->ibqp.qp_num = init_attr->qp_type == IB_QPT_SMI ? 0 : 1; in mthca_create_qp() 608 qp->ibqp.qp_num, init_attr->port_num, in mthca_create_qp() 628 return &qp->ibqp; in mthca_create_qp()
|
| /linux-4.4.14/drivers/infiniband/hw/mlx4/ |
| D | qp.c | 248 if (qp->ibqp.qp_type == IB_QPT_UD) { in post_nop_wqe() 252 av->port_pd = cpu_to_be32((qp->port << 24) | to_mpd(qp->ibqp.pd)->pdn); in post_nop_wqe() 289 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx4_ib_qp_event() local 294 if (ibqp->event_handler) { in mlx4_ib_qp_event() 295 event.device = ibqp->device; in mlx4_ib_qp_event() 296 event.element.qp = ibqp; in mlx4_ib_qp_event() 328 ibqp->event_handler(&event, ibqp->qp_context); in mlx4_ib_qp_event() 975 if (qp->ibqp.qp_type == IB_QPT_XRC_TGT) in get_pd() 976 return to_mpd(to_mxrcd(qp->ibqp.xrcd)->pd); in get_pd() 978 return to_mpd(qp->ibqp.pd); in get_pd() [all …]
|
| D | mlx4_ib.h | 287 struct ib_qp ibqp; member 660 static inline struct mlx4_ib_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument 662 return container_of(ibqp, struct mlx4_ib_qp, ibqp); in to_mqp() 747 int mlx4_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 749 int mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 751 int mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 753 int mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
| D | main.c | 1229 static int add_gid_entry(struct ib_qp *ibqp, union ib_gid *gid) in add_gid_entry() argument 1231 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in add_gid_entry() 1232 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in add_gid_entry() 1705 static int mlx4_ib_mcg_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx4_ib_mcg_attach() argument 1708 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_attach() 1710 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in mlx4_ib_mcg_attach() 1742 err = add_gid_entry(ibqp, gid); in mlx4_ib_mcg_attach() 1783 static int mlx4_ib_mcg_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx4_ib_mcg_detach() argument 1786 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_detach() 1788 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in mlx4_ib_mcg_detach() [all …]
|
| D | cq.c | 584 ib_dma_sync_single_for_cpu(qp->ibqp.device, in use_tunnel_data() 626 wc->qp = &qp->ibqp; in mlx4_ib_qp_sw_comp() 732 wc->qp = &(*cur_qp)->ibqp; in mlx4_ib_poll_one() 756 } else if ((*cur_qp)->ibqp.srq) { in mlx4_ib_poll_one() 757 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx4_ib_poll_one()
|
| /linux-4.4.14/drivers/infiniband/hw/usnic/ |
| D | usnic_ib_qp_grp.h | 49 struct ib_qp ibqp; member 128 struct usnic_ib_qp_grp *to_uqp_grp(struct ib_qp *ibqp) in to_uqp_grp() argument 130 return container_of(ibqp, struct usnic_ib_qp_grp, ibqp); in to_uqp_grp()
|
| D | usnic_ib_verbs.h | 63 int usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 82 int usnic_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 84 int usnic_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
| D | usnic_ib_qp_grp.c | 85 qp_grp->ibqp.qp_num, in usnic_ib_qp_grp_dump_rows() 493 ib_event.element.qp = &qp_grp->ibqp; in usnic_ib_qp_grp_modify() 498 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify() 499 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 503 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify() 504 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 510 qp_grp->ibqp.event_handler(&ib_event, in usnic_ib_qp_grp_modify() 511 qp_grp->ibqp.qp_context); in usnic_ib_qp_grp_modify() 733 qp_grp->ibqp.qp_num = qp_grp->grp_id; in usnic_ib_qp_grp_create()
|
| D | usnic_ib_verbs.c | 390 switch (qp_grp->ibqp.qp_type) { in usnic_ib_query_qp() 395 usnic_err("Unexpected qp_type %d\n", qp_grp->ibqp.qp_type); in usnic_ib_query_qp() 534 return &qp_grp->ibqp; in usnic_ib_create_qp() 565 int usnic_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in usnic_ib_modify_qp() argument 572 qp_grp = to_uqp_grp(ibqp); in usnic_ib_modify_qp() 755 int usnic_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in usnic_ib_post_send() argument 762 int usnic_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr, in usnic_ib_post_recv() argument
|
| D | usnic_ib_sysfs.c | 258 qp_grp->ibqp.qp_num, in summary_show()
|
| /linux-4.4.14/drivers/staging/rdma/amso1100/ |
| D | c2_provider.h | 108 struct ib_qp ibqp; member 158 static inline struct c2_qp *to_c2qp(struct ib_qp *ibqp) in to_c2qp() argument 160 return container_of(ibqp, struct c2_qp, ibqp); in to_c2qp()
|
| D | c2_cm.c | 44 struct ib_qp *ibqp; in c2_llp_connect() local 54 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); in c2_llp_connect() 55 if (!ibqp) in c2_llp_connect() 57 qp = to_c2qp(ibqp); in c2_llp_connect() 292 struct ib_qp *ibqp; in c2_llp_accept() local 298 ibqp = c2_get_qp(cm_id->device, iw_param->qpn); in c2_llp_accept() 299 if (!ibqp) in c2_llp_accept() 301 qp = to_c2qp(ibqp); in c2_llp_accept()
|
| D | c2_provider.c | 205 static void c2_add_ref(struct ib_qp *ibqp) in c2_add_ref() argument 208 BUG_ON(!ibqp); in c2_add_ref() 209 qp = to_c2qp(ibqp); in c2_add_ref() 213 static void c2_rem_ref(struct ib_qp *ibqp) in c2_rem_ref() argument 216 BUG_ON(!ibqp); in c2_rem_ref() 217 qp = to_c2qp(ibqp); in c2_rem_ref() 232 return (qp?&qp->ibqp:NULL); in c2_get_qp() 278 return &qp->ibqp; in c2_create_qp() 566 static int c2_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in c2_modify_qp() argument 572 c2_qp_modify(to_c2dev(ibqp->device), to_c2qp(ibqp), attr, in c2_modify_qp() [all …]
|
| D | c2_ae.c | 235 ib_event.element.qp = &qp->ibqp; in c2_ae_event() 238 if (qp->ibqp.event_handler) in c2_ae_event() 239 qp->ibqp.event_handler(&ib_event, in c2_ae_event() 240 qp->ibqp. in c2_ae_event()
|
| D | c2_qp.c | 432 qp->ibqp.qp_num = qp->qpn; in c2_alloc_qp() 433 qp->ibqp.qp_type = IB_QPT_RC; in c2_alloc_qp() 605 send_cq = to_c2cq(qp->ibqp.send_cq); in c2_free_qp() 606 recv_cq = to_c2cq(qp->ibqp.recv_cq); in c2_free_qp() 790 int c2_post_send(struct ib_qp *ibqp, struct ib_send_wr *ib_wr, in c2_post_send() argument 793 struct c2_dev *c2dev = to_c2dev(ibqp->device); in c2_post_send() 794 struct c2_qp *qp = to_c2qp(ibqp); in c2_post_send() 944 int c2_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *ib_wr, in c2_post_receive() argument 947 struct c2_dev *c2dev = to_c2dev(ibqp->device); in c2_post_receive() 948 struct c2_qp *qp = to_c2qp(ibqp); in c2_post_receive()
|
| D | c2.h | 496 extern int c2_post_send(struct ib_qp *ibqp, struct ib_send_wr *ib_wr, 498 extern int c2_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *ib_wr,
|
| D | c2_cq.c | 158 entry->qp = &qp->ibqp; in c2_poll_one()
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
| D | iwch_ev.c | 93 event.element.qp = &qhp->ibqp; in post_qp_event() 95 if (qhp->ibqp.event_handler) in post_qp_event() 96 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event() 129 iwch_qp_add_ref(&qhp->ibqp); in iwch_ev_dispatch() 229 iwch_qp_rem_ref(&qhp->ibqp); in iwch_ev_dispatch()
|
| D | iwch_provider.h | 162 struct ib_qp ibqp; member 179 static inline struct iwch_qp *to_iwch_qp(struct ib_qp *ibqp) in to_iwch_qp() argument 181 return container_of(ibqp, struct iwch_qp, ibqp); in to_iwch_qp() 329 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 331 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
| D | iwch_provider.c | 75 static int iwch_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in iwch_multicast_attach() argument 80 static int iwch_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in iwch_multicast_detach() argument 1029 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp() 1036 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp() 1053 iwch_destroy_qp(&qhp->ibqp); in iwch_create_qp() 1065 qhp->ibqp.qp_num = qhp->wq.qpid; in iwch_create_qp() 1072 return &qhp->ibqp; in iwch_create_qp() 1075 static int iwch_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in iwch_ib_modify_qp() argument 1083 PDBG("%s ib_qp %p\n", __func__, ibqp); in iwch_ib_modify_qp() 1094 qhp = to_iwch_qp(ibqp); in iwch_ib_modify_qp()
|
| D | iwch_qp.c | 351 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in iwch_post_send() argument 366 qhp = to_iwch_qp(ibqp); in iwch_post_send() 466 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, in iwch_post_receive() argument 476 qhp = to_iwch_qp(ibqp); in iwch_post_receive() 861 if (qhp->ibqp.uobject) { in flush_qp() 918 if (!qhp->ibqp.uobject) in rdma_init() 1057 if (qhp->ibqp.uobject) in iwch_modify_qp()
|
| D | iwch_cq.c | 82 wc->qp = &qhp->ibqp; in iwch_poll_cq_one()
|
| /linux-4.4.14/drivers/infiniband/hw/mlx5/ |
| D | odp.c | 157 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_page_fault_resume() 183 struct mlx5_ib_dev *mib_dev = to_mdev(qp->ibqp.pd->device); in pagefault_single_data_segment() 212 if (mr->ibmr.pd != qp->ibqp.pd) { in pagefault_single_data_segment() 323 if (receive_queue && qp->ibqp.srq) in pagefault_data_segments() 387 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_mr_initiator_pfault_handler() 434 switch (qp->ibqp.qp_type) { in mlx5_ib_mr_initiator_pfault_handler() 477 qp->ibqp.qp_type, opcode); in mlx5_ib_mr_initiator_pfault_handler() 492 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.pd->device); in mlx5_ib_mr_responder_pfault_handler() 496 if (qp->ibqp.srq) { in mlx5_ib_mr_responder_pfault_handler() 511 switch (qp->ibqp.qp_type) { in mlx5_ib_mr_responder_pfault_handler() [all …]
|
| D | qp.c | 119 struct ib_device *ibdev = qp->ibqp.device; in mlx5_ib_read_user_wqe() 131 qp->ibqp.qp_type); in mlx5_ib_read_user_wqe() 172 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx5_ib_qp_event() local 178 if (ibqp->event_handler) { in mlx5_ib_qp_event() 179 event.device = ibqp->device; in mlx5_ib_qp_event() 180 event.element.qp = ibqp; in mlx5_ib_qp_event() 211 ibqp->event_handler(&event, ibqp->qp_context); in mlx5_ib_qp_event() 1108 return to_mpd(qp->ibqp.pd); in get_pd() 1114 switch (qp->ibqp.qp_type) { in get_cqs() 1121 *send_cq = to_mcq(qp->ibqp.send_cq); in get_cqs() [all …]
|
| D | mlx5_ib.h | 175 struct ib_qp ibqp; member 476 static inline struct mlx5_ib_qp *to_mqp(struct ib_qp *ibqp) in to_mqp() argument 478 return container_of(ibqp, struct mlx5_ib_qp, ibqp); in to_mqp() 527 int mlx5_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 529 int mlx5_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, 532 int mlx5_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 534 int mlx5_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
| D | cq.c | 174 struct mlx5_ib_dev *dev = to_mdev(qp->ibqp.device); in handle_responder() 180 if (qp->ibqp.srq || qp->ibqp.xrcd) { in handle_responder() 183 if (qp->ibqp.xrcd) { in handle_responder() 188 srq = to_msrq(qp->ibqp.srq); in handle_responder() 231 if (unlikely(is_qp1(qp->ibqp.qp_type))) { in handle_responder() 456 if (!*cur_qp || (qpn != (*cur_qp)->ibqp.qp_num)) { in mlx5_poll_one() 471 wc->qp = &(*cur_qp)->ibqp; in mlx5_poll_one() 510 if ((*cur_qp)->ibqp.srq) { in mlx5_poll_one() 511 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx5_poll_one()
|
| D | main.c | 838 static int mlx5_ib_mcg_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx5_ib_mcg_attach() argument 840 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_mcg_attach() 843 err = mlx5_core_attach_mcg(dev->mdev, gid, ibqp->qp_num); in mlx5_ib_mcg_attach() 846 ibqp->qp_num, gid->raw); in mlx5_ib_mcg_attach() 851 static int mlx5_ib_mcg_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mlx5_ib_mcg_detach() argument 853 struct mlx5_ib_dev *dev = to_mdev(ibqp->device); in mlx5_ib_mcg_detach() 856 err = mlx5_core_detach_mcg(dev->mdev, gid, ibqp->qp_num); in mlx5_ib_mcg_detach() 859 ibqp->qp_num, gid->raw); in mlx5_ib_mcg_detach()
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
| D | ev.c | 109 event.element.qp = &qhp->ibqp; in post_qp_event() 110 if (qhp->ibqp.event_handler) in post_qp_event() 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event() 153 c4iw_qp_add_ref(&qhp->ibqp); in c4iw_ev_dispatch() 218 c4iw_qp_rem_ref(&qhp->ibqp); in c4iw_ev_dispatch()
|
| D | iw_cxgb4.h | 467 struct ib_qp ibqp; member 481 static inline struct c4iw_qp *to_c4iw_qp(struct ib_qp *ibqp) in to_c4iw_qp() argument 483 return container_of(ibqp, struct c4iw_qp, ibqp); in to_c4iw_qp() 946 int c4iw_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, 948 int c4iw_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, 994 int c4iw_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 996 int c4iw_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
|
| D | qp.c | 739 int c4iw_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, in c4iw_post_send() argument 753 qhp = to_c4iw_qp(ibqp); in c4iw_post_send() 867 int c4iw_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, in c4iw_post_receive() argument 878 qhp = to_c4iw_qp(ibqp); in c4iw_post_receive() 1181 rchp = to_c4iw_cq(qhp->ibqp.recv_cq); in flush_qp() 1182 schp = to_c4iw_cq(qhp->ibqp.send_cq); in flush_qp() 1185 if (qhp->ibqp.uobject) { in flush_qp() 1312 if (!qhp->ibqp.uobject) in rdma_init() 1797 qhp->ibqp.qp_num = qhp->wq.sq.qid; in c4iw_create_qp() 1805 return &qhp->ibqp; in c4iw_create_qp() [all …]
|
| D | provider.c | 72 static int c4iw_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in c4iw_multicast_attach() argument 77 static int c4iw_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in c4iw_multicast_detach() argument
|
| D | cq.c | 240 struct c4iw_cq *chp = to_c4iw_cq(qhp->ibqp.send_cq); in c4iw_flush_sq() 709 wc->qp = &qhp->ibqp; in c4iw_poll_cq_one()
|
| D | device.c | 1344 c4iw_qp_add_ref(&qp->ibqp); in add_and_ref_qp() 1361 c4iw_qp_rem_ref(&qp_list->qps[idx]->ibqp); in deref_qps()
|
| D | cm.c | 154 c4iw_qp_rem_ref(&ep->com.qp->ibqp); in deref_qp() 161 c4iw_qp_add_ref(&ep->com.qp->ibqp); in ref_qp()
|
| /linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma_verbs.c | 1463 qp->ibqp.qp_num = qp->id; in ocrdma_create_qp() 1465 return &qp->ibqp; in ocrdma_create_qp() 1481 int _ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in _ocrdma_modify_qp() argument 1489 qp = get_ocrdma_qp(ibqp); in _ocrdma_modify_qp() 1490 dev = get_ocrdma_dev(ibqp->device); in _ocrdma_modify_qp() 1503 int ocrdma_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, in ocrdma_modify_qp() argument 1512 qp = get_ocrdma_qp(ibqp); in ocrdma_modify_qp() 1513 dev = get_ocrdma_dev(ibqp->device); in ocrdma_modify_qp() 1526 if (!ib_modify_qp_is_ok(old_qps, new_qps, ibqp->qp_type, attr_mask, in ocrdma_modify_qp() 1530 __func__, dev->id, attr_mask, qp->id, ibqp->qp_type, in ocrdma_modify_qp() [all …]
|
| D | ocrdma.h | 392 struct ib_qp ibqp; member 479 static inline struct ocrdma_qp *get_ocrdma_qp(struct ib_qp *ibqp) in get_ocrdma_qp() argument 481 return container_of(ibqp, struct ocrdma_qp, ibqp); in get_ocrdma_qp()
|
| D | ocrdma_hw.c | 730 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent() 735 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent() 739 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent() 743 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent() 765 ib_evt.element.qp = &qp->ibqp; in ocrdma_dispatch_ibevent() 781 if (qp->ibqp.event_handler) in ocrdma_dispatch_ibevent() 782 qp->ibqp.event_handler(&ib_evt, qp->ibqp.qp_context); in ocrdma_dispatch_ibevent() 2115 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_flush_qp() 2505 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_set_av_params() 2570 struct ocrdma_dev *dev = get_ocrdma_dev(qp->ibqp.device); in ocrdma_set_qp_params()
|
| /linux-4.4.14/drivers/infiniband/hw/nes/ |
| D | nes_verbs.c | 212 static int nes_bind_mw(struct ib_qp *ibqp, struct ib_mw *ibmw, in nes_bind_mw() argument 216 struct nes_vnic *nesvnic = to_nesvnic(ibqp->device); in nes_bind_mw() 219 struct nes_qp *nesqp = to_nesqp(ibqp); in nes_bind_mw() 1245 nesqp->ibqp.qp_num = nesqp->hwqp.qp_id; in nes_create_qp() 1385 nes_add_ref(&nesqp->ibqp); in nes_create_qp() 1402 return &nesqp->ibqp; in nes_create_qp() 1442 static int nes_destroy_qp(struct ib_qp *ibqp) in nes_destroy_qp() argument 1444 struct nes_qp *nesqp = to_nesqp(ibqp); in nes_destroy_qp() 1458 nes_modify_qp(&nesqp->ibqp, &attr, IB_QP_STATE, NULL); in nes_destroy_qp() 1482 if ((ibqp->uobject)&&(ibqp->uobject->context)) { in nes_destroy_qp() [all …]
|
| D | nes.c | 266 void nes_add_ref(struct ib_qp *ibqp) in nes_add_ref() argument 270 nesqp = to_nesqp(ibqp); in nes_add_ref() 272 ibqp->qp_num, atomic_read(&nesqp->refcount)); in nes_add_ref() 309 void nes_rem_ref(struct ib_qp *ibqp) in nes_rem_ref() argument 313 struct nes_vnic *nesvnic = to_nesvnic(ibqp->device); in nes_rem_ref() 319 nesqp = to_nesqp(ibqp); in nes_rem_ref() 323 __func__, ibqp->qp_num, nesqp->last_aeq); in nes_rem_ref() 371 return &nesadapter->qp_table[qpn - NES_FIRST_QPN]->ibqp; in nes_get_qp()
|
| D | nes_cm.c | 1738 nes_rem_ref(&nesqp->ibqp); in rem_ref_cm_node() 3007 nes_add_ref(&nesqp->ibqp); in nes_cm_disconn() 3027 nes_rem_ref(&nesqp->ibqp); in nes_disconnect_worker() 3066 nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_cm_disconn_true() 3111 ibevent.device = nesqp->ibqp.device; in nes_cm_disconn_true() 3113 ibevent.element.qp = &nesqp->ibqp; in nes_cm_disconn_true() 3114 if (nesqp->ibqp.event_handler) in nes_cm_disconn_true() 3115 nesqp->ibqp.event_handler(&ibevent, nesqp->ibqp.qp_context); in nes_cm_disconn_true() 3178 nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_disconnect() 3220 struct ib_qp *ibqp; in nes_accept() local [all …]
|
| D | nes.h | 510 static inline struct nes_qp *to_nesqp(struct ib_qp *ibqp) in to_nesqp() argument 512 return container_of(ibqp, struct nes_qp, ibqp); in to_nesqp()
|
| D | nes_verbs.h | 139 struct ib_qp ibqp; member
|
| D | nes_hw.c | 3449 struct nes_vnic *nesvnic = to_nesvnic(nesqp->ibqp.device); in nes_terminate_done()
|