| /linux-4.4.14/drivers/infiniband/hw/mthca/ |
| D | mthca_srq.c | 72 static void *get_wqe(struct mthca_srq *srq, int n) in get_wqe() argument 74 if (srq->is_direct) in get_wqe() 75 return srq->queue.direct.buf + (n << srq->wqe_shift); in get_wqe() 77 return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf + in get_wqe() 78 ((n << srq->wqe_shift) & (PAGE_SIZE - 1)); in get_wqe() 97 struct mthca_srq *srq, in mthca_tavor_init_srq_context() argument 102 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); in mthca_tavor_init_srq_context() 104 context->lkey = cpu_to_be32(srq->mr.ibmr.lkey); in mthca_tavor_init_srq_context() 115 struct mthca_srq *srq, in mthca_arbel_init_srq_context() argument 126 max = srq->max; in mthca_arbel_init_srq_context() [all …]
|
| D | mthca_provider.c | 444 struct mthca_srq *srq; in mthca_create_srq() local 450 srq = kmalloc(sizeof *srq, GFP_KERNEL); in mthca_create_srq() 451 if (!srq) in mthca_create_srq() 469 srq->mr.ibmr.lkey = ucmd.lkey; in mthca_create_srq() 470 srq->db_index = ucmd.db_index; in mthca_create_srq() 474 &init_attr->attr, srq); in mthca_create_srq() 483 if (context && ib_copy_to_udata(udata, &srq->srqn, sizeof (__u32))) { in mthca_create_srq() 484 mthca_free_srq(to_mdev(pd->device), srq); in mthca_create_srq() 489 return &srq->ibsrq; in mthca_create_srq() 492 kfree(srq); in mthca_create_srq() [all …]
|
| D | mthca_dev.h | 249 struct mthca_array srq; member 507 struct mthca_srq *srq); 513 struct ib_srq_attr *attr, struct mthca_srq *srq); 514 void mthca_free_srq(struct mthca_dev *dev, struct mthca_srq *srq); 517 int mthca_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr); 521 void mthca_free_srq_wqe(struct mthca_srq *srq, u32 wqe_addr); 522 int mthca_tavor_post_srq_recv(struct ib_srq *srq, struct ib_recv_wr *wr, 524 int mthca_arbel_post_srq_recv(struct ib_srq *srq, struct ib_recv_wr *wr,
|
| D | mthca_cq.c | 279 struct mthca_srq *srq) in mthca_cq_clean() argument 311 if (srq && is_recv_cqe(cqe)) in mthca_cq_clean() 312 mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe)); in mthca_cq_clean() 544 } else if ((*cur_qp)->ibqp.srq) { in mthca_poll_one() 545 struct mthca_srq *srq = to_msrq((*cur_qp)->ibqp.srq); in mthca_poll_one() local 548 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one() 549 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one() 550 mthca_free_srq_wqe(srq, wqe); in mthca_poll_one()
|
| D | mthca_qp.c | 743 if (ibqp->srq) in __mthca_modify_qp() 768 if (ibqp->srq) in __mthca_modify_qp() 770 to_msrq(ibqp->srq)->srqn); in __mthca_modify_qp() 820 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in __mthca_modify_qp() 1457 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in mthca_free_qp() 2236 if (qp->ibqp.srq && !is_send) { in mthca_free_err_wqe()
|
| D | mthca_eq.c | 147 } __attribute__((packed)) srq; member 302 mthca_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) & 0xffffff, in mthca_eq_int()
|
| /linux-4.4.14/drivers/infiniband/hw/mlx4/ |
| D | srq.c | 42 static void *get_wqe(struct mlx4_ib_srq *srq, int n) in get_wqe() argument 44 return mlx4_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 47 static void mlx4_ib_srq_event(struct mlx4_srq *srq, enum mlx4_event type) in mlx4_ib_srq_event() argument 50 struct ib_srq *ibsrq = &to_mibsrq(srq)->ibsrq; in mlx4_ib_srq_event() 54 event.element.srq = ibsrq; in mlx4_ib_srq_event() 64 "on SRQ %06x\n", type, srq->srqn); in mlx4_ib_srq_event() 77 struct mlx4_ib_srq *srq; in mlx4_ib_create_srq() local 92 srq = kmalloc(sizeof *srq, GFP_KERNEL); in mlx4_ib_create_srq() 93 if (!srq) in mlx4_ib_create_srq() 96 mutex_init(&srq->mutex); in mlx4_ib_create_srq() [all …]
|
| D | cq.c | 663 struct mlx4_ib_srq *srq; in mlx4_ib_poll_one() local 756 } else if ((*cur_qp)->ibqp.srq) { in mlx4_ib_poll_one() 757 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx4_ib_poll_one() 759 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one() 760 mlx4_ib_free_srq_wqe(srq, wqe_ctr); in mlx4_ib_poll_one() 762 srq = to_mibsrq(msrq); in mlx4_ib_poll_one() 764 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one() 765 mlx4_ib_free_srq_wqe(srq, wqe_ctr); in mlx4_ib_poll_one() 934 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) in __mlx4_ib_cq_clean() argument 962 if (srq && !(cqe->owner_sr_opcode & MLX4_CQE_IS_SEND_MASK)) in __mlx4_ib_cq_clean() [all …]
|
| D | Makefile | 3 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
|
| D | mlx4_ib.h | 725 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 726 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 737 int mlx4_ib_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr); 738 int mlx4_ib_destroy_srq(struct ib_srq *srq); 739 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index);
|
| D | qp.c | 608 return !attr->srq; in qp_has_rq() 1045 qp->ibqp.srq ? to_msrq(qp->ibqp.srq): NULL); in destroy_qp_common() 1746 if (ibqp->srq) in __mlx4_ib_modify_qp() 1777 if (ibqp->srq) in __mlx4_ib_modify_qp() 1778 context->srqn = cpu_to_be32(1 << 24 | to_msrq(ibqp->srq)->msrq.srqn); in __mlx4_ib_modify_qp() 1914 ibqp->srq ? to_msrq(ibqp->srq) : NULL); in __mlx4_ib_modify_qp()
|
| D | main.c | 467 props->max_srq = dev->dev->quotas.srq; in mlx4_ib_query_device() 2689 if (!mqp->ibqp.srq) { in mlx4_ib_handle_catas_error()
|
| /linux-4.4.14/drivers/infiniband/hw/mlx5/ |
| D | srq.c | 46 static void *get_wqe(struct mlx5_ib_srq *srq, int n) in get_wqe() argument 48 return mlx5_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 51 static void mlx5_ib_srq_event(struct mlx5_core_srq *srq, enum mlx5_event type) in mlx5_ib_srq_event() argument 54 struct ib_srq *ibsrq = &to_mibsrq(srq)->ibsrq; in mlx5_ib_srq_event() 58 event.element.srq = ibsrq; in mlx5_ib_srq_event() 68 type, srq->srqn); in mlx5_ib_srq_event() 76 static int create_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq, in create_srq_user() argument 103 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); in create_srq_user() 105 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, buf_size, in create_srq_user() 107 if (IS_ERR(srq->umem)) { in create_srq_user() [all …]
|
| D | cq.c | 175 struct mlx5_ib_srq *srq; in handle_responder() local 180 if (qp->ibqp.srq || qp->ibqp.xrcd) { in handle_responder() 186 srq = to_mibsrq(msrq); in handle_responder() 188 srq = to_msrq(qp->ibqp.srq); in handle_responder() 190 if (srq) { in handle_responder() 192 wc->wr_id = srq->wrid[wqe_ctr]; in handle_responder() 193 mlx5_ib_free_srq_wqe(srq, wqe_ctr); in handle_responder() 508 struct mlx5_ib_srq *srq; in mlx5_poll_one() local 510 if ((*cur_qp)->ibqp.srq) { in mlx5_poll_one() 511 srq = to_msrq((*cur_qp)->ibqp.srq); in mlx5_poll_one() [all …]
|
| D | Makefile | 3 mlx5_ib-y := main.o cq.o doorbell.o qp.o mem.o srq.o mr.o ah.o mad.o
|
| D | mlx5_ib.h | 504 void __mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq); 505 void mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq); 506 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index); 521 int mlx5_ib_destroy_srq(struct ib_srq *srq);
|
| D | odp.c | 323 if (receive_queue && qp->ibqp.srq) in pagefault_data_segments() 496 if (qp->ibqp.srq) { in mlx5_ib_mr_responder_pfault_handler()
|
| D | qp.c | 403 attr->qp_type == IB_QPT_XRC_TGT || attr->srq || in qp_has_rq() 839 if (attr->srq || (attr->qp_type == IB_QPT_XRC_TGT) || in get_rx_type() 1003 if (init_attr->srq) { in create_qp_common() 1005 in->ctx.rq_type_srqn |= cpu_to_be32(to_msrq(init_attr->srq)->msrq.srqn); in create_qp_common() 1168 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in destroy_qp_common() 1741 ibqp->srq ? to_msrq(ibqp->srq) : NULL); in __mlx5_ib_modify_qp()
|
| /linux-4.4.14/drivers/staging/rdma/hfi1/ |
| D | srq.c | 68 struct hfi1_srq *srq = to_isrq(ibsrq); in hfi1_post_srq_receive() local 78 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in hfi1_post_srq_receive() 84 spin_lock_irqsave(&srq->rq.lock, flags); in hfi1_post_srq_receive() 85 wq = srq->rq.wq; in hfi1_post_srq_receive() 87 if (next >= srq->rq.size) in hfi1_post_srq_receive() 90 spin_unlock_irqrestore(&srq->rq.lock, flags); in hfi1_post_srq_receive() 96 wqe = get_rwqe_ptr(&srq->rq, wq->head); in hfi1_post_srq_receive() 104 spin_unlock_irqrestore(&srq->rq.lock, flags); in hfi1_post_srq_receive() 123 struct hfi1_srq *srq; in hfi1_create_srq() local 140 srq = kmalloc(sizeof(*srq), GFP_KERNEL); in hfi1_create_srq() [all …]
|
| D | ruc.c | 109 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in init_sge() 162 struct hfi1_srq *srq; in hfi1_get_rwqe() local 168 if (qp->ibqp.srq) { in hfi1_get_rwqe() 169 srq = to_isrq(qp->ibqp.srq); in hfi1_get_rwqe() 170 handler = srq->ibsrq.event_handler; in hfi1_get_rwqe() 171 rq = &srq->rq; in hfi1_get_rwqe() 173 srq = NULL; in hfi1_get_rwqe() 226 if (n < srq->limit) { in hfi1_get_rwqe() 229 srq->limit = 0; in hfi1_get_rwqe() 232 ev.element.srq = qp->ibqp.srq; in hfi1_get_rwqe() [all …]
|
| D | qp.c | 906 attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1; in hfi1_query_qp() 930 init_attr->srq = qp->ibqp.srq; in hfi1_query_qp() 951 if (qp->ibqp.srq) { in hfi1_compute_aeth() 1033 if (!init_attr->srq) { in hfi1_create_qp() 1069 if (init_attr->srq) { in hfi1_create_qp() 1070 struct hfi1_srq *srq = to_isrq(init_attr->srq); in hfi1_create_qp() local 1072 if (srq->rq.max_sge > 1) in hfi1_create_qp() 1074 (srq->rq.max_sge - 1); in hfi1_create_qp() 1092 if (init_attr->srq) in hfi1_create_qp()
|
| D | Makefile | 12 qp.o qsfp.o rc.o ruc.o sdma.o srq.o sysfs.o trace.o twsi.o \
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_srq.c | 51 struct qib_srq *srq = to_isrq(ibsrq); in qib_post_srq_receive() local 61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in qib_post_srq_receive() 67 spin_lock_irqsave(&srq->rq.lock, flags); in qib_post_srq_receive() 68 wq = srq->rq.wq; in qib_post_srq_receive() 70 if (next >= srq->rq.size) in qib_post_srq_receive() 73 spin_unlock_irqrestore(&srq->rq.lock, flags); in qib_post_srq_receive() 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); in qib_post_srq_receive() 87 spin_unlock_irqrestore(&srq->rq.lock, flags); in qib_post_srq_receive() 106 struct qib_srq *srq; in qib_create_srq() local 123 srq = kmalloc(sizeof(*srq), GFP_KERNEL); in qib_create_srq() [all …]
|
| D | qib_ruc.c | 91 pd = to_ipd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in qib_init_sge() 144 struct qib_srq *srq; in qib_get_rwqe() local 150 if (qp->ibqp.srq) { in qib_get_rwqe() 151 srq = to_isrq(qp->ibqp.srq); in qib_get_rwqe() 152 handler = srq->ibsrq.event_handler; in qib_get_rwqe() 153 rq = &srq->rq; in qib_get_rwqe() 155 srq = NULL; in qib_get_rwqe() 208 if (n < srq->limit) { in qib_get_rwqe() 211 srq->limit = 0; in qib_get_rwqe() 214 ev.element.srq = qp->ibqp.srq; in qib_get_rwqe() [all …]
|
| D | qib_qp.c | 870 attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1; in qib_query_qp() 894 init_attr->srq = qp->ibqp.srq; in qib_query_qp() 915 if (qp->ibqp.srq) { in qib_compute_aeth() 1004 if (!init_attr->srq) { in qib_create_qp() 1041 if (init_attr->srq) { in qib_create_qp() 1042 struct qib_srq *srq = to_isrq(init_attr->srq); in qib_create_qp() local 1044 if (srq->rq.max_sge > 1) in qib_create_qp() 1046 (srq->rq.max_sge - 1); in qib_create_qp() 1064 if (init_attr->srq) in qib_create_qp()
|
| /linux-4.4.14/drivers/staging/rdma/ipath/ |
| D | ipath_srq.c | 51 struct ipath_srq *srq = to_isrq(ibsrq); in ipath_post_srq_receive() local 61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in ipath_post_srq_receive() 67 spin_lock_irqsave(&srq->rq.lock, flags); in ipath_post_srq_receive() 68 wq = srq->rq.wq; in ipath_post_srq_receive() 70 if (next >= srq->rq.size) in ipath_post_srq_receive() 73 spin_unlock_irqrestore(&srq->rq.lock, flags); in ipath_post_srq_receive() 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); in ipath_post_srq_receive() 87 spin_unlock_irqrestore(&srq->rq.lock, flags); in ipath_post_srq_receive() 106 struct ipath_srq *srq; in ipath_create_srq() local 126 srq = kmalloc(sizeof(*srq), GFP_KERNEL); in ipath_create_srq() [all …]
|
| D | ipath_ud.c | 56 struct ipath_srq *srq; in ipath_ud_loopback() local 106 if (qp->ibqp.srq) { in ipath_ud_loopback() 107 srq = to_isrq(qp->ibqp.srq); in ipath_ud_loopback() 108 handler = srq->ibsrq.event_handler; in ipath_ud_loopback() 109 rq = &srq->rq; in ipath_ud_loopback() 111 srq = NULL; in ipath_ud_loopback() 163 if (n < srq->limit) { in ipath_ud_loopback() 166 srq->limit = 0; in ipath_ud_loopback() 169 ev.element.srq = qp->ibqp.srq; in ipath_ud_loopback() 171 handler(&ev, srq->ibsrq.srq_context); in ipath_ud_loopback()
|
| D | ipath_ruc.c | 170 struct ipath_srq *srq; in ipath_get_rwqe() local 176 if (qp->ibqp.srq) { in ipath_get_rwqe() 177 srq = to_isrq(qp->ibqp.srq); in ipath_get_rwqe() 178 handler = srq->ibsrq.event_handler; in ipath_get_rwqe() 179 rq = &srq->rq; in ipath_get_rwqe() 181 srq = NULL; in ipath_get_rwqe() 230 if (n < srq->limit) { in ipath_get_rwqe() 233 srq->limit = 0; in ipath_get_rwqe() 236 ev.element.srq = qp->ibqp.srq; in ipath_get_rwqe() 238 handler(&ev, srq->ibsrq.srq_context); in ipath_get_rwqe()
|
| D | ipath_qp.c | 635 attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1; in ipath_query_qp() 659 init_attr->srq = qp->ibqp.srq; in ipath_query_qp() 680 if (qp->ibqp.srq) { in ipath_compute_aeth() 764 if (!init_attr->srq) { in ipath_create_qp() 795 if (init_attr->srq) { in ipath_create_qp() 796 struct ipath_srq *srq = to_isrq(init_attr->srq); in ipath_create_qp() local 798 if (srq->rq.max_sge > 1) in ipath_create_qp() 800 (srq->rq.max_sge - 1); in ipath_create_qp() 819 if (init_attr->srq) { in ipath_create_qp()
|
| /linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/ |
| D | srq.c | 45 struct mlx5_core_srq *srq; in mlx5_srq_event() local 49 srq = radix_tree_lookup(&table->tree, srqn); in mlx5_srq_event() 50 if (srq) in mlx5_srq_event() 51 atomic_inc(&srq->refcount); in mlx5_srq_event() 55 if (!srq) { in mlx5_srq_event() 60 srq->event(srq, event_type); in mlx5_srq_event() 62 if (atomic_dec_and_test(&srq->refcount)) in mlx5_srq_event() 63 complete(&srq->free); in mlx5_srq_event() 137 struct mlx5_core_srq *srq; in mlx5_core_get_srq() local 141 srq = radix_tree_lookup(&table->tree, srqn); in mlx5_core_get_srq() [all …]
|
| D | Makefile | 4 health.o mcg.o cq.o srq.o alloc.o qp.o port.o mr.o pd.o \
|
| /linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
| D | srq.c | 46 struct mlx4_srq *srq; in mlx4_srq_event() local 50 srq = radix_tree_lookup(&srq_table->tree, srqn & (dev->caps.num_srqs - 1)); in mlx4_srq_event() 51 if (srq) in mlx4_srq_event() 52 atomic_inc(&srq->refcount); in mlx4_srq_event() 56 if (!srq) { in mlx4_srq_event() 61 srq->event(srq, event_type); in mlx4_srq_event() 63 if (atomic_dec_and_test(&srq->refcount)) in mlx4_srq_event() 64 complete(&srq->free); in mlx4_srq_event() 166 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq) in mlx4_srq_alloc() argument 174 err = mlx4_srq_alloc_icm(dev, &srq->srqn); in mlx4_srq_alloc() [all …]
|
| D | resource_tracker.c | 112 struct res_srq *srq; member 447 dev->quotas.srq = dev->caps.num_srqs - dev->caps.reserved_srqs; in mlx4_init_quotas() 458 dev->quotas.srq = in mlx4_init_quotas() 1634 enum res_srq_states state, struct res_srq **srq) in srq_res_start_move_to() argument 1660 if (srq) in srq_res_start_move_to() 1661 *srq = r; in srq_res_start_move_to() 2646 int srq = (be32_to_cpu(qpc->srqn) >> 24) & 1; in qp_get_mtt_size() local 2657 rq_size = (srq|rss|xrc) ? 0 : (1 << (log_rq_size + log_rq_stride + 4)); in qp_get_mtt_size() 2893 struct res_srq *srq; in mlx4_RST2INIT_QP_wrapper() local 2933 err = get_res(dev, slave, srqn, RES_SRQ, &srq); in mlx4_RST2INIT_QP_wrapper() [all …]
|
| D | Makefile | 5 srq.o resource_tracker.o
|
| D | eq.c | 552 be32_to_cpu(eqe->event.srq.srqn) in mlx4_eq_int() 563 be32_to_cpu(eqe->event.srq.srqn), in mlx4_eq_int() 574 mlx4_srq_event(dev, be32_to_cpu(eqe->event.srq.srqn) & in mlx4_eq_int()
|
| D | main.c | 845 dev->quotas.srq = func_cap.srq_quota; in mlx4_slave_cap()
|
| /linux-4.4.14/drivers/infiniband/core/ |
| D | verbs.c | 455 struct ib_srq *srq; in ib_create_srq() local 460 srq = pd->device->create_srq(pd, srq_init_attr, NULL); in ib_create_srq() 462 if (!IS_ERR(srq)) { in ib_create_srq() 463 srq->device = pd->device; in ib_create_srq() 464 srq->pd = pd; in ib_create_srq() 465 srq->uobject = NULL; in ib_create_srq() 466 srq->event_handler = srq_init_attr->event_handler; in ib_create_srq() 467 srq->srq_context = srq_init_attr->srq_context; in ib_create_srq() 468 srq->srq_type = srq_init_attr->srq_type; in ib_create_srq() 469 if (srq->srq_type == IB_SRQT_XRC) { in ib_create_srq() [all …]
|
| D | uverbs_cmd.c | 269 static void put_srq_read(struct ib_srq *srq) in put_srq_read() argument 271 put_uobj_read(srq->uobject); in put_srq_read() 1762 struct ib_srq *srq = NULL; in create_qp() local 1794 srq = idr_read_srq(cmd->srq_handle, in create_qp() 1796 if (!srq || srq->srq_type != IB_SRQT_BASIC) { in create_qp() 1827 attr.srq = srq; in create_qp() 1877 qp->srq = attr.srq; in create_qp() 1886 if (attr.srq) in create_qp() 1887 atomic_inc(&attr.srq->usecnt); in create_qp() 1925 if (srq) in create_qp() [all …]
|
| D | ucm.c | 219 ureq->srq = kreq->srq; in ib_ucm_event_req_get() 241 urep->srq = krep->srq; in ib_ucm_event_rep_get() 773 param.srq = cmd.srq; in ib_ucm_send_req() 815 param.srq = cmd.srq; in ib_ucm_send_rep()
|
| D | cm_msgs.h | 284 static inline void cm_req_set_srq(struct cm_req_msg *req_msg, u8 srq) in cm_req_set_srq() argument 287 ((srq & 0x1) << 3)); in cm_req_set_srq() 610 static inline void cm_rep_set_srq(struct cm_rep_msg *rep_msg, u8 srq) in cm_rep_set_srq() argument 613 ((srq & 0x1) << 4)); in cm_rep_set_srq()
|
| D | cma.c | 227 u8 srq; member 697 id_priv->srq = (qp->srq != NULL); in rdma_create_qp() 1520 event->param.conn.srq = rep_data->srq; in cma_set_rep_event_data() 1714 event->param.conn.srq = req_data->srq; in cma_set_req_event_data() 3226 req.srq = id_priv->srq ? 1 : 0; in cma_connect_ib() 3292 id_priv->srq = conn_param->srq; in rdma_connect() 3338 rep.srq = id_priv->srq ? 1 : 0; in cma_accept_ib() 3403 id_priv->srq = conn_param->srq; in rdma_accept()
|
| D | uverbs_main.c | 258 struct ib_srq *srq = uobj->object; in ib_uverbs_cleanup_ucontext() local 263 ib_destroy_srq(srq); in ib_uverbs_cleanup_ucontext() 564 uobj = container_of(event->element.srq->uobject, in ib_uverbs_srq_event_handler()
|
| D | ucma.c | 246 dst->srq = src->srq; in ucma_copy_conn_event() 1008 dst->srq = src->srq; in ucma_copy_conn_param()
|
| D | cm.c | 1091 cm_req_set_srq(req_msg, param->srq); in cm_format_req() 1390 param->srq = cm_req_get_srq(req_msg); in cm_format_req_event() 1706 cm_rep_set_srq(rep_msg, param->srq); in cm_format_rep() 1848 param->srq = cm_rep_get_srq(rep_msg); in cm_format_rep_event()
|
| /linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma_verbs.c | 1232 if (!attrs->srq && (attrs->cap.max_recv_wr > dev->attr.max_rqe)) { in ocrdma_check_qp_params() 1287 int dpp_credit_lmt, int srq) in ocrdma_copy_qp_uresp() argument 1304 if (!srq) { in ocrdma_copy_qp_uresp() 1331 if (!srq) { in ocrdma_copy_qp_uresp() 1458 (attrs->srq != NULL)); in ocrdma_create_qp() 1659 static void ocrdma_srq_toggle_bit(struct ocrdma_srq *srq, unsigned int idx) in ocrdma_srq_toggle_bit() argument 1664 srq->idx_bit_fields[i] ^= mask; in ocrdma_srq_toggle_bit() 1726 if (is_hw_sq_empty(qp) && (!qp->srq && is_hw_rq_empty(qp))) in ocrdma_discard_cqes() 1743 if (qp->srq) { in ocrdma_discard_cqes() 1746 qp->srq->rq.max_wqe_idx; in ocrdma_discard_cqes() [all …]
|
| D | ocrdma_hw.c | 753 ib_evt.element.srq = &qp->srq->ibsrq; in ocrdma_dispatch_ibevent() 759 ib_evt.element.srq = &qp->srq->ibsrq; in ocrdma_dispatch_ibevent() 787 if (qp->srq->ibsrq.event_handler) in ocrdma_dispatch_ibevent() 788 qp->srq->ibsrq.event_handler(&ib_evt, in ocrdma_dispatch_ibevent() 789 qp->srq->ibsrq. in ocrdma_dispatch_ibevent() 913 if (qp->srq) in _ocrdma_qp_buddy_cq_handler() 2121 if (!qp->srq) { in ocrdma_flush_qp() 2368 if (!attrs->srq) { in ocrdma_get_create_qp_rsp() 2411 if (attrs->srq) { in ocrdma_mbx_create_qp() 2412 struct ocrdma_srq *srq = get_ocrdma_srq(attrs->srq); in ocrdma_mbx_create_qp() local [all …]
|
| D | ocrdma.h | 415 struct ocrdma_srq *srq; member
|
| /linux-4.4.14/drivers/net/ |
| D | eql.c | 264 static int eql_enslave(struct net_device *dev, slaving_request_t __user *srq); 265 static int eql_emancipate(struct net_device *dev, slaving_request_t __user *srq); 412 slaving_request_t srq; in eql_enslave() local 414 if (copy_from_user(&srq, srqp, sizeof (slaving_request_t))) in eql_enslave() 417 slave_dev = __dev_get_by_name(&init_net, srq.slave_name); in eql_enslave() 433 s->priority = srq.priority; in eql_enslave() 434 s->priority_bps = srq.priority; in eql_enslave() 435 s->priority_Bps = srq.priority / 8; in eql_enslave() 455 slaving_request_t srq; in eql_emancipate() local 458 if (copy_from_user(&srq, srqp, sizeof (slaving_request_t))) in eql_emancipate() [all …]
|
| /linux-4.4.14/include/uapi/rdma/ |
| D | ib_user_cm.h | 150 __u8 srq; member 167 __u8 srq; member 247 __u8 srq; member 263 __u8 srq; member
|
| D | rdma_user_cm.h | 182 __u8 srq; member
|
| /linux-4.4.14/drivers/staging/rdma/ehca/ |
| D | ehca_iverbs.h | 166 int ehca_post_srq_recv(struct ib_srq *srq, 174 int ehca_modify_srq(struct ib_srq *srq, struct ib_srq_attr *attr, 177 int ehca_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr); 179 int ehca_destroy_srq(struct ib_srq *srq);
|
| D | ehca_qp.c | 506 if (init_attr->srq) { in internal_create_qp() 507 my_srq = container_of(init_attr->srq, struct ehca_qp, ib_srq); in internal_create_qp() 809 } else if (init_attr->srq && !is_user) { in internal_create_qp() 832 my_qp->ib_qp.srq = init_attr->srq; in internal_create_qp() 2090 int ehca_query_srq(struct ib_srq *srq, struct ib_srq_attr *srq_attr) in ehca_query_srq() argument 2092 struct ehca_qp *my_qp = container_of(srq, struct ehca_qp, ib_srq); in ehca_query_srq() 2093 struct ehca_shca *shca = container_of(srq->device, struct ehca_shca, in ehca_query_srq() 2102 ehca_err(srq->device, "Out of memory for qpcb " in ehca_query_srq() 2112 ehca_err(srq->device, "hipz_h_query_qp() failed " in ehca_query_srq() 2235 int ehca_destroy_srq(struct ib_srq *srq) in ehca_destroy_srq() argument [all …]
|
| D | ehca_reqs.c | 601 int ehca_post_srq_recv(struct ib_srq *srq, in ehca_post_srq_recv() argument 605 return internal_post_recv(container_of(srq, struct ehca_qp, ib_srq), in ehca_post_srq_recv() 606 srq->device, recv_wr, bad_recv_wr); in ehca_post_srq_recv()
|
| D | ehca_irq.c | 195 event.element.srq = &qp->ib_srq; in dispatch_qp_event()
|
| D | ehca_main.c | 592 qp_init_attr.srq = NULL; in ehca_create_aqp1()
|
| /linux-4.4.14/include/rdma/ |
| D | ib_cm.h | 136 unsigned int srq:1; member 150 unsigned int srq:1; member 383 u8 srq; member 406 u8 srq; member
|
| D | ib_verbs.h | 486 struct ib_srq *srq; member 887 struct ib_srq *srq; member 1343 struct ib_srq *srq; member 1681 int (*modify_srq)(struct ib_srq *srq, 1685 int (*query_srq)(struct ib_srq *srq, 1687 int (*destroy_srq)(struct ib_srq *srq); 1688 int (*post_srq_recv)(struct ib_srq *srq, 2324 int ib_modify_srq(struct ib_srq *srq, 2334 int ib_query_srq(struct ib_srq *srq, 2341 int ib_destroy_srq(struct ib_srq *srq); [all …]
|
| D | rdma_cm.h | 101 u8 srq; member
|
| /linux-4.4.14/drivers/infiniband/ulp/ipoib/ |
| D | ipoib_cm.c | 102 ret = ib_post_srq_recv(priv->cm.srq, &priv->cm.rx_wr, &bad_wr); in ipoib_cm_post_receive_srq() 256 .srq = priv->cm.srq, in ipoib_cm_create_rx_qp() 434 rep.srq = ipoib_cm_has_srq(dev); in ipoib_cm_send_rep() 1022 .srq = priv->cm.srq, in ipoib_cm_create_tx_qp() 1079 req.srq = ipoib_cm_has_srq(dev); in ipoib_cm_send_req() 1502 priv->cm.srq = ib_create_srq(priv->pd, &srq_init_attr); in ipoib_cm_create_srq() 1503 if (IS_ERR(priv->cm.srq)) { in ipoib_cm_create_srq() 1504 if (PTR_ERR(priv->cm.srq) != -ENOSYS) in ipoib_cm_create_srq() 1506 priv->ca->name, PTR_ERR(priv->cm.srq)); in ipoib_cm_create_srq() 1507 priv->cm.srq = NULL; in ipoib_cm_create_srq() [all …]
|
| D | ipoib.h | 255 struct ib_srq *srq; member 625 return !!priv->cm.srq; in ipoib_cm_has_srq()
|
| /linux-4.4.14/include/linux/mlx5/ |
| D | driver.h | 715 int mlx5_core_create_srq(struct mlx5_core_dev *dev, struct mlx5_core_srq *srq, 718 int mlx5_core_destroy_srq(struct mlx5_core_dev *dev, struct mlx5_core_srq *srq); 719 int mlx5_core_query_srq(struct mlx5_core_dev *dev, struct mlx5_core_srq *srq, 721 int mlx5_core_arm_srq(struct mlx5_core_dev *dev, struct mlx5_core_srq *srq,
|
| /linux-4.4.14/include/linux/mlx4/ |
| D | device.h | 813 int srq; member 887 } __packed srq; member 1107 struct mlx4_mtt *mtt, u64 db_rec, struct mlx4_srq *srq); 1108 void mlx4_srq_free(struct mlx4_dev *dev, struct mlx4_srq *srq); 1109 int mlx4_srq_arm(struct mlx4_dev *dev, struct mlx4_srq *srq, int limit_watermark); 1110 int mlx4_srq_query(struct mlx4_dev *dev, struct mlx4_srq *srq, int *limit_watermark);
|
| /linux-4.4.14/drivers/net/wireless/orinoco/ |
| D | wext.c | 491 struct iw_param *srq, in orinoco_ioctl_getsens() argument 512 srq->value = val; in orinoco_ioctl_getsens() 513 srq->fixed = 0; /* auto */ in orinoco_ioctl_getsens() 520 struct iw_param *srq, in orinoco_ioctl_setsens() argument 524 int val = srq->value; in orinoco_ioctl_setsens()
|
| /linux-4.4.14/drivers/infiniband/ulp/srpt/ |
| D | ib_srpt.h | 396 struct ib_srq *srq; member
|
| D | ib_srpt.c | 791 return ib_post_srq_recv(sdev->srq, &wr, &bad_wr); in srpt_post_recv() 2047 qp_init->srq = sdev->srq; in srpt_create_ch_ib() 2561 rep_param->srq = 1; in srpt_cm_req_recv() 3165 sdev->srq = ib_create_srq(sdev->pd, &srq_attr); in srpt_add_one() 3166 if (IS_ERR(sdev->srq)) in srpt_add_one() 3251 ib_destroy_srq(sdev->srq); in srpt_add_one() 3295 ib_destroy_srq(sdev->srq); in srpt_remove_one()
|
| /linux-4.4.14/drivers/staging/lustre/lnet/selftest/ |
| D | conrpc.c | 690 srpc_stat_reqst_t *srq; in lstcon_statrpc_prep() local 697 srq = &(*crpc)->crp_rpc->crpc_reqstmsg.msg_body.stat_reqst; in lstcon_statrpc_prep() 699 srq->str_sid = console_session.ses_id; in lstcon_statrpc_prep() 700 srq->str_type = 0; /* XXX remove it */ in lstcon_statrpc_prep()
|
| /linux-4.4.14/drivers/net/wireless/ |
| D | zd1201.c | 1105 struct iw_request_info *info, struct iw_point *srq, char *extra) in zd1201_set_scan() argument 1112 struct iw_request_info *info, struct iw_point *srq, char *extra) in zd1201_get_scan() argument 1197 srq->length = cev - extra; in zd1201_get_scan() 1198 srq->flags = 0; in zd1201_get_scan()
|
| /linux-4.4.14/net/sunrpc/xprtrdma/ |
| D | verbs.c | 594 ep->rep_attr.srq = NULL; in rpcrdma_ep_create()
|
| /linux-4.4.14/drivers/infiniband/hw/nes/ |
| D | nes_verbs.c | 2831 init_attr->srq = nesqp->ibqp.srq; in nes_query_qp()
|