Home
last modified time | relevance | path

Searched refs:wr_id (Results 1 – 87 of 87) sorted by relevance

/linux-4.4.14/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c108 priv->rx_wr.wr_id = id | IPOIB_OP_RECV; in ipoib_ib_post_receive()
179 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc() local
185 wr_id, wc->status); in ipoib_ib_handle_rx_wc()
187 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_ib_handle_rx_wc()
189 wr_id, ipoib_recvq_size); in ipoib_ib_handle_rx_wc()
193 skb = priv->rx_ring[wr_id].skb; in ipoib_ib_handle_rx_wc()
199 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc()
200 ipoib_ud_dma_unmap_rx(priv, priv->rx_ring[wr_id].mapping); in ipoib_ib_handle_rx_wc()
202 priv->rx_ring[wr_id].skb = NULL; in ipoib_ib_handle_rx_wc()
213 memcpy(mapping, priv->rx_ring[wr_id].mapping, in ipoib_ib_handle_rx_wc()
[all …]
Dipoib_cm.c73 .wr_id = IPOIB_CM_RX_DRAIN_WRID,
97 priv->cm.rx_wr.wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_srq()
123 wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_nonsrq()
562 unsigned int wr_id = wc->wr_id & ~(IPOIB_OP_CM | IPOIB_OP_RECV); in ipoib_cm_handle_rx_wc() local
572 wr_id, wc->status); in ipoib_cm_handle_rx_wc()
574 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_cm_handle_rx_wc()
575 if (wr_id == (IPOIB_CM_RX_DRAIN_WRID & ~(IPOIB_OP_CM | IPOIB_OP_RECV))) { in ipoib_cm_handle_rx_wc()
583 wr_id, ipoib_recvq_size); in ipoib_cm_handle_rx_wc()
592 skb = rx_ring[wr_id].skb; in ipoib_cm_handle_rx_wc()
597 wc->status, wr_id, wc->vendor_err); in ipoib_cm_handle_rx_wc()
[all …]
/linux-4.4.14/net/rds/
Diw_ring.c157 u32 rds_iw_ring_completed(struct rds_iw_work_ring *ring, u32 wr_id, u32 oldest) in rds_iw_ring_completed() argument
161 if (oldest <= (unsigned long long)wr_id) in rds_iw_ring_completed()
162 ret = (unsigned long long)wr_id - oldest + 1; in rds_iw_ring_completed()
164 ret = ring->w_nr - oldest + (unsigned long long)wr_id + 1; in rds_iw_ring_completed()
167 wr_id, oldest); in rds_iw_ring_completed()
Dib_ring.c156 u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest) in rds_ib_ring_completed() argument
160 if (oldest <= (unsigned long long)wr_id) in rds_ib_ring_completed()
161 ret = (unsigned long long)wr_id - oldest + 1; in rds_ib_ring_completed()
163 ret = ring->w_nr - oldest + (unsigned long long)wr_id + 1; in rds_ib_ring_completed()
166 wr_id, oldest); in rds_ib_ring_completed()
Diw_send.c141 send->s_send_wr.wr_id = i; in rds_iw_send_init_ring()
207 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_iw_send_cq_comp_handler()
216 if (wc.opcode == IB_WC_LOCAL_INV && wc.wr_id == RDS_IW_LOCAL_INV_WR_ID) { in rds_iw_send_cq_comp_handler()
221 if (wc.opcode == IB_WC_REG_MR && wc.wr_id == RDS_IW_REG_WR_ID) { in rds_iw_send_cq_comp_handler()
226 if (wc.wr_id == RDS_IW_ACK_WR_ID) { in rds_iw_send_cq_comp_handler()
235 completed = rds_iw_ring_completed(&ic->i_send_ring, wc.wr_id, oldest); in rds_iw_send_cq_comp_handler()
775 send->s_reg_wr.wr.wr_id = 0; in rds_iw_build_send_reg()
Diw_recv.c94 recv->r_wr.wr_id = i; in rds_iw_recv_init_ring()
359 wr->wr_id = RDS_IW_ACK_WR_ID; in rds_iw_recv_init_ack()
789 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_poll_cq()
Dib_send.c198 send->s_wr.wr_id = i | RDS_IB_SEND_OP; in rds_ib_send_init_ring()
252 (unsigned long long)wc->wr_id, wc->status, in rds_ib_send_cqe_handler()
257 if (wc->wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cqe_handler()
267 (wc->wr_id & ~RDS_IB_SEND_OP), in rds_ib_send_cqe_handler()
Dib_recv.c58 recv->r_wr.wr_id = i; in rds_ib_recv_init_ring()
572 wr->wr_id = RDS_IB_ACK_WR_ID; in rds_ib_recv_init_ack()
964 (unsigned long long)wc->wr_id, wc->status, in rds_ib_recv_cqe_handler()
Diw_rdma.c675 reg_wr.wr.wr_id = RDS_IW_REG_WR_ID;
710 s_wr.wr_id = RDS_IW_LOCAL_INV_WR_ID;
Diw.h350 u32 rds_iw_ring_completed(struct rds_iw_work_ring *ring, u32 wr_id, u32 oldest);
Dib.h392 u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest);
Dib_cm.c251 (unsigned long long)wc->wr_id, wc->status, in poll_cq()
254 if (wc->wr_id & RDS_IB_SEND_OP) in poll_cq()
/linux-4.4.14/drivers/infiniband/ulp/srpt/
Dib_srpt.h144 static inline enum srpt_opcode opcode_from_wr_id(u64 wr_id) in opcode_from_wr_id() argument
146 return wr_id >> 32; in opcode_from_wr_id()
148 static inline u32 idx_from_wr_id(u64 wr_id) in idx_from_wr_id() argument
150 return (u32)wr_id; in idx_from_wr_id()
Dib_srpt.c781 wr.wr_id = encode_wr_id(SRPT_RECV, ioctx->ioctx.index); in srpt_post_recv()
823 wr.wr_id = encode_wr_id(SRPT_SEND, ioctx->ioctx.index); in srpt_post_send()
1385 static void srpt_handle_send_err_comp(struct srpt_rdma_ch *ch, u64 wr_id) in srpt_handle_send_err_comp() argument
1393 index = idx_from_wr_id(wr_id); in srpt_handle_send_err_comp()
1880 index = idx_from_wr_id(wc->wr_id); in srpt_process_rcv_completion()
1918 index = idx_from_wr_id(wc->wr_id); in srpt_process_send_completion()
1919 opcode = opcode_from_wr_id(wc->wr_id); in srpt_process_send_completion()
1933 srpt_handle_send_err_comp(ch, wc->wr_id); in srpt_process_send_completion()
1965 if (opcode_from_wr_id(wc[i].wr_id) == SRPT_RECV) in srpt_process_completion()
2796 wr.wr.wr_id = encode_wr_id(i == n_rdma - 1 ? in srpt_perform_rdmas()
[all …]
/linux-4.4.14/drivers/infiniband/ulp/iser/
Diser_verbs.c1003 ib_conn->beacon.wr_id = ISER_BEACON_WRID; in iser_connect()
1057 rx_wr.wr_id = (uintptr_t)iser_conn->login_resp_buf; in iser_post_recvl()
1081 rx_wr->wr_id = (uintptr_t)rx_desc; in iser_post_recvm()
1118 wr->wr_id = (uintptr_t)tx_desc; in iser_post_send()
1144 is_iser_tx_desc(struct iser_conn *iser_conn, void *wr_id) in is_iser_tx_desc() argument
1149 if (wr_id >= start && wr_id < start + len) in is_iser_tx_desc()
1169 void *wr_id = (void *)(uintptr_t)wc->wr_id; in iser_handle_comp_error() local
1178 if (wc->wr_id == ISER_FASTREG_LI_WRID) in iser_handle_comp_error()
1181 if (is_iser_tx_desc(iser_conn, wr_id)) { in iser_handle_comp_error()
1182 struct iser_tx_desc *desc = wr_id; in iser_handle_comp_error()
[all …]
Diser_memory.c422 inv_wr->wr_id = ISER_FASTREG_LI_WRID; in iser_inv_rkey()
455 wr->wr.wr_id = ISER_FASTREG_LI_WRID; in iser_reg_sig_mr()
504 wr->wr.wr_id = ISER_FASTREG_LI_WRID; in iser_fast_reg_mr()
/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_reqs.c55 static u64 replace_wr_id(u64 wr_id, u16 idx) in replace_wr_id() argument
59 ret = wr_id & ~QMAP_IDX_MASK; in replace_wr_id()
65 static u16 get_app_wr_id(u64 wr_id) in get_app_wr_id() argument
67 return wr_id & QMAP_IDX_MASK; in get_app_wr_id()
87 wqe_p->work_request_id = replace_wr_id(recv_wr->wr_id, rq_map_idx); in ehca_write_rwqe()
121 "send_flags=%x opcode=%x", idx, ud_wr->wr.wr_id, in trace_ud_wr()
177 wqe_p->work_request_id = replace_wr_id(send_wr->wr_id, sq_map_idx); in ehca_write_swqe()
179 qmap_entry->app_wr_id = get_app_wr_id(send_wr->wr_id); in ehca_write_swqe()
561 qmap_entry->app_wr_id = get_app_wr_id(recv_wr->wr_id); in internal_post_recv()
744 wc->wr_id = replace_wr_id(cqe->work_request_id, qmap_entry->app_wr_id); in ehca_poll_cq_one()
[all …]
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Dmad.c591 wr.wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); in mlx4_ib_send_to_slave()
1153 recv_wr.wr_id = (u64) index | MLX4_TUN_WRID_RECV | in mlx4_ib_post_pv_qp_buf()
1268 wr.wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum); in mlx4_ib_send_to_wire()
1306 struct mlx4_ib_demux_pv_qp *tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc->wr_id)]; in mlx4_ib_multiplex_mad()
1307 int wr_ix = wc->wr_id & (MLX4_NUM_TUNNEL_BUFS - 1); in mlx4_ib_multiplex_mad()
1558 tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc.wr_id)]; in mlx4_ib_tunnel_comp_worker()
1564 wc.wr_id & in mlx4_ib_tunnel_comp_worker()
1568 "buf:%lld\n", wc.wr_id); in mlx4_ib_tunnel_comp_worker()
1573 wc.wr_id, wc.status); in mlx4_ib_tunnel_comp_worker()
1574 ib_destroy_ah(tun_qp->tx_ring[wc.wr_id & in mlx4_ib_tunnel_comp_worker()
[all …]
Dcq.c621 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_qp_sw_comp()
754 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one()
759 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one()
764 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one()
769 wc->wr_id = wq->wrid[tail]; in mlx4_ib_poll_one()
Dmr.c378 wr.wr.wr_id = mw_bind->wr_id; in mlx4_ib_bind_mw()
Dsrq.c347 srq->wrid[srq->head] = wr->wr_id; in mlx4_ib_post_srq_recv()
Dqp.c2796 qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id; in mlx4_ib_post_send()
3131 qp->rq.wrid[ind] = wr->wr_id; in mlx4_ib_post_recv()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dsrq.c97 wqe->wr_id = wr->wr_id; in hfi1_post_srq_receive()
303 p->wr_id = wqe->wr_id; in hfi1_modify_srq()
Druc.c136 wc.wr_id = wqe->wr_id; in init_sge()
208 qp->r_wr_id = wqe->wr_id; in hfi1_get_rwqe()
577 wc.wr_id = qp->r_wr_id; in ruc_loopback()
924 wc.wr_id = wqe->wr.wr_id; in hfi1_send_complete()
Dcq.c100 wc->uqueue[head].wr_id = entry->wr_id; in hfi1_cq_enter()
Drc.c1037 wc.wr_id = wqe->wr.wr_id; in hfi1_rc_send_complete()
1094 wc.wr_id = wqe->wr.wr_id; in do_rc_completion()
2112 wc.wr_id = qp->r_wr_id; in hfi1_rc_rcv()
Dud.c229 wc.wr_id = qp->r_wr_id; in ud_loopback()
843 wc.wr_id = qp->r_wr_id; in hfi1_ud_rcv()
Dqp.c507 wc.wr_id = qp->r_wr_id; in hfi1_error_qp()
529 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in hfi1_error_qp()
Duc.c445 wc.wr_id = qp->r_wr_id; in hfi1_uc_rcv()
Dverbs.h370 u64 wr_id; member
Dverbs.c566 wqe->wr_id = wr->wr_id; in post_receive()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_srq.c80 wqe->wr_id = wr->wr_id; in qib_post_srq_receive()
286 p->wr_id = wqe->wr_id; in qib_modify_srq()
Dqib_ruc.c118 wc.wr_id = wqe->wr_id; in qib_init_sge()
190 qp->r_wr_id = wqe->wr_id; in qib_get_rwqe()
555 wc.wr_id = qp->r_wr_id; in qib_ruc_loopback()
797 wc.wr_id = wqe->wr.wr_id; in qib_send_complete()
Dqib_cq.c84 wc->uqueue[head].wr_id = entry->wr_id; in qib_cq_enter()
Dqib_ud.c207 wc.wr_id = qp->r_wr_id; in qib_ud_loopback()
565 wc.wr_id = qp->r_wr_id; in qib_ud_rcv()
Dqib_rc.c1023 wc.wr_id = wqe->wr.wr_id; in qib_rc_send_complete()
1079 wc.wr_id = wqe->wr.wr_id; in do_rc_completion()
2033 wc.wr_id = qp->r_wr_id; in qib_rc_rcv()
Dqib_qp.c523 wc.wr_id = qp->r_wr_id; in qib_error_qp()
545 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in qib_error_qp()
Dqib_uc.c405 wc.wr_id = qp->r_wr_id; in qib_uc_rcv()
Dqib_verbs.h363 u64 wr_id; member
Dqib_verbs.c548 wqe->wr_id = wr->wr_id; in qib_post_receive()
/linux-4.4.14/drivers/staging/rdma/ipath/
Dipath_srq.c80 wqe->wr_id = wr->wr_id; in ipath_post_srq_receive()
290 p->wr_id = wqe->wr_id; in ipath_modify_srq()
Dipath_ruc.c145 wc.wr_id = wqe->wr_id; in ipath_init_sge()
211 qp->r_wr_id = wqe->wr_id; in ipath_get_rwqe()
435 wc.wr_id = qp->r_wr_id; in ipath_ruc_loopback()
713 wc.wr_id = wqe->wr.wr_id; in ipath_send_complete()
Dipath_ud.c148 wc.wr_id = wqe->wr_id; in ipath_ud_loopback()
556 wc.wr_id = qp->r_wr_id; in ipath_ud_rcv()
Dipath_cq.c81 wc->uqueue[head].wr_id = entry->wr_id; in ipath_cq_enter()
Dipath_qp.c403 wc.wr_id = qp->r_wr_id; in ipath_error_qp()
425 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in ipath_error_qp()
Dipath_rc.c952 wc.wr_id = wqe->wr.wr_id; in do_rc_ack()
1733 wc.wr_id = qp->r_wr_id; in ipath_rc_rcv()
Dipath_uc.c413 wc.wr_id = qp->r_wr_id; in ipath_uc_rcv()
Dipath_verbs.h300 u64 wr_id; member
Dipath_verbs.c525 wqe->wr_id = wr->wr_id; in ipath_post_receive()
/linux-4.4.14/drivers/net/ethernet/cisco/enic/
Dvnic_wq.h61 uint64_t wr_id; /* Cookie */ member
146 buf->wr_id = wrid; in vnic_wq_post()
Dvnic_rq.h76 uint64_t wr_id; member
133 buf->wr_id = wrid; in vnic_rq_post()
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.h101 u64 wr_id; member
140 u64 wr_id; /* work request ID */ member
163 u64 wr_id; /* work request ID from WQE */ member
Dehea_main.c501 rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) in ehea_refill_rq_def()
578 int skb_index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in get_skb_by_index()
793 swqe->wr_id = SWQE_RESTART_CHECK; in check_sqs()
832 if (cqe->wr_id == SWQE_RESTART_CHECK) { in ehea_proc_cqes()
855 if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id) in ehea_proc_cqes()
858 index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in ehea_proc_cqes()
864 swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); in ehea_proc_cqes()
2079 swqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, EHEA_SWQE3_TYPE) in ehea_start_xmit()
2082 swqe->wr_id |= EHEA_BMASK_SET(EHEA_WR_ID_REFILL, in ehea_start_xmit()
2089 swqe->wr_id = in ehea_start_xmit()
[all …]
/linux-4.4.14/drivers/infiniband/ulp/isert/
Dib_isert.c987 rx_wr->wr_id = (uintptr_t)rx_desc; in isert_post_recvm()
1009 rx_wr.wr_id = (uintptr_t)rx_desc; in isert_post_recv()
1032 send_wr.wr_id = (uintptr_t)tx_desc; in isert_post_send()
1102 send_wr->wr_id = (uintptr_t)&isert_cmd->tx_desc; in isert_init_send_wr()
1125 rx_wr.wr_id = (uintptr_t)isert_conn->login_req_buf; in isert_rdma_post_recvl()
2035 is_isert_tx_desc(struct isert_conn *isert_conn, void *wr_id) in is_isert_tx_desc() argument
2040 if ((wr_id >= start && wr_id < start + len) || in is_isert_tx_desc()
2041 (wr_id == isert_conn->login_req_buf)) in is_isert_tx_desc()
2050 if (wc->wr_id == ISER_BEACON_WRID) { in isert_cq_comp_err()
2054 } else if (is_isert_tx_desc(isert_conn, (void *)(uintptr_t)wc->wr_id)) { in isert_cq_comp_err()
[all …]
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
Diwch_qp.c282 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_rdma_recv()
345 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_zero_stag_recv()
437 sqp->wr_id = wr->wr_id; in iwch_post_send()
448 __func__, (unsigned long long) wr->wr_id, idx, in iwch_post_send()
512 "wqe %p \n", __func__, (unsigned long long) wr->wr_id, in iwch_post_receive()
591 sqp->wr_id = mw_bind->wr_id; in iwch_bind_mw()
Diwch_cq.c81 wc->wr_id = cookie; in iwch_poll_cq_one()
Dcxio_wr.h674 __u64 wr_id; member
684 __u64 wr_id; member
Dcxio_hal.c1303 *cookie = wq->sq[Q_PTR2IDX(wq->sq_rptr, wq->sq_size_log2)].wr_id; in cxio_poll_cq()
1308 *cookie = wq->rq[Q_PTR2IDX(wq->rq_rptr, wq->rq_size_log2)].wr_id; in cxio_poll_cq()
/linux-4.4.14/net/sunrpc/xprtrdma/
Dfrwr_ops.c258 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in frwr_sendcompletion()
386 reg_wr.wr.wr_id = (uintptr_t)mw; in frwr_op_map()
435 invalidate_wr.wr_id = (unsigned long)(void *)mw; in frwr_op_unmap()
Dsvc_rdma_recvfrom.c183 read_wr.wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_lcl()
303 reg_wr.wr.wr_id = 0; in rdma_read_chunk_frmr()
320 read_wr.wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_frmr()
327 inv_wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_frmr()
Dverbs.c134 if (wc->wr_id == RPCRDMA_IGNORE_COMPLETION) { in rpcrdma_sendcq_process_wc()
142 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in rpcrdma_sendcq_process_wc()
197 (struct rpcrdma_rep *)(unsigned long)wc->wr_id; in rpcrdma_recvcq_process_wc()
1271 send_wr.wr_id = RPCRDMA_IGNORE_COMPLETION; in rpcrdma_ep_post()
1309 recv_wr.wr_id = (u64) (unsigned long) rep; in rpcrdma_ep_post_recv()
Dsvc_rdma_transport.c352 ctxt = (struct svc_rdma_op_ctxt *)(unsigned long)wc.wr_id; in rq_cq_reap()
470 (unsigned long)wc->wr_id; in sq_cq_reap()
581 recv_wr.wr_id = (u64)(unsigned long)ctxt; in svc_rdma_post_recv()
1350 err_wr.wr_id = (unsigned long)ctxt; in svc_rdma_send_error()
Dsvc_rdma_sendto.c285 write_wr.wr.wr_id = (unsigned long)ctxt; in send_write()
547 send_wr.wr_id = (unsigned long)ctxt; in send_reply()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Dcq.c192 wc->wr_id = srq->wrid[wqe_ctr]; in handle_responder()
199 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder()
479 wc->wr_id = wq->wrid[idx]; in mlx5_poll_one()
505 wc->wr_id = wq->wrid[idx]; in mlx5_poll_one()
513 wc->wr_id = srq->wrid[wqe_ctr]; in mlx5_poll_one()
517 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
Dmr.c752 context = (struct mlx5_ib_umr_context *) (unsigned long) wc.wr_id; in mlx5_umr_cq_handler()
815 umrwr.wr.wr_id = (u64)(unsigned long)&umr_context; in reg_umr()
933 wr.wr.wr_id = (u64)(unsigned long)&umr_context; in mlx5_ib_update_mtt()
1143 umrwr.wr.wr_id = (u64)(unsigned long)&umr_context; in unreg_umr()
Dqp.c2565 u8 size, unsigned idx, u64 wr_id, in finish_wqe() argument
2579 qp->sq.wrid[idx] = wr_id; in finish_wqe()
2694 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send()
2721 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send()
2743 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send()
2829 finish_wqe(qp, ctrl, size, idx, wr->wr_id, nreq, in mlx5_ib_post_send()
2933 qp->rq.wrid[ind] = wr->wr_id; in mlx5_ib_post_recv()
Dsrq.c451 srq->wrid[srq->head] = wr->wr_id; in mlx5_ib_post_srq_recv()
/linux-4.4.14/include/uapi/rdma/
Dib_user_verbs.h401 __u64 wr_id; member
683 __u64 wr_id; member
727 __u64 wr_id; member
/linux-4.4.14/net/9p/
Dtrans_rdma.c351 struct p9_rdma_context *c = (void *) (unsigned long) wc.wr_id; in cq_comp_handler()
417 wr.wr_id = (unsigned long) c; in post_recv()
508 wr.wr_id = (unsigned long) c; in rdma_request()
/linux-4.4.14/drivers/infiniband/ulp/srp/
Dib_srp.c460 static struct ib_recv_wr wr = { .wr_id = SRP_LAST_WR_ID }; in srp_destroy_qp()
1049 .wr_id = LOCAL_INV_WR_ID_MASK, in srp_inv_rkey()
1354 wr.wr.wr_id = FAST_REG_WR_ID_MASK; in srp_map_finish_fr()
1753 wr.wr_id = (uintptr_t) iu; in srp_post_send()
1773 wr.wr_id = (uintptr_t) iu; in srp_post_recv()
1909 struct srp_iu *iu = (struct srp_iu *) (uintptr_t) wc->wr_id; in srp_handle_recv()
1975 static void srp_handle_qp_err(u64 wr_id, enum ib_wc_status wc_status, in srp_handle_qp_err() argument
1980 if (wr_id == SRP_LAST_WR_ID) { in srp_handle_qp_err()
1986 if (wr_id & LOCAL_INV_WR_ID_MASK) { in srp_handle_qp_err()
1990 } else if (wr_id & FAST_REG_WR_ID_MASK) { in srp_handle_qp_err()
[all …]
/linux-4.4.14/drivers/infiniband/core/
Dmad.c693 u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, in build_smp_wc() argument
697 wc->wr_id = wr_id; in build_smp_wc()
835 send_wr->wr.wr_id, drslid, in handle_outgoing_dr_smp()
1042 mad_send_wr->send_wr.wr.wr_id = (unsigned long) mad_send_wr; in ib_create_send_mad()
1154 mad_send_wr->send_wr.wr.wr_id = (unsigned long)&mad_send_wr->mad_list; in ib_send_mad()
1985 mad_recv_wc->wc->wr_id = 0; in ib_mad_complete_recv()
2001 mad_recv_wc->wc->wr_id = (unsigned long) &mad_send_wr->send_buf; in ib_mad_complete_recv()
2189 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_recv_done_handler()
2428 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_send_done_handler()
2502 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in mad_error_handler()
[all …]
Duverbs_cmd.c1596 tmp.wr_id = wc->wr_id; in copy_wc_to_user()
2575 next->wr_id = user_wr->wr_id; in ib_uverbs_post_send()
2678 next->wr_id = user_wr->wr_id; in ib_uverbs_unmarshall_recv()
Dsa_query.c1617 mad_buf = (void *) (unsigned long) mad_recv_wc->wc->wr_id; in recv_handler()
/linux-4.4.14/drivers/infiniband/hw/mthca/
Dmthca_cq.c542 entry->wr_id = (*cur_qp)->wrid[wqe_index + in mthca_poll_one()
549 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one()
563 entry->wr_id = (*cur_qp)->wrid[wqe_index]; in mthca_poll_one()
Dmthca_srq.c532 srq->wrid[ind] = wr->wr_id; in mthca_tavor_post_srq_recv()
621 srq->wrid[ind] = wr->wr_id; in mthca_arbel_post_srq_recv()
Dmthca_qp.c1751 qp->wrid[ind + qp->rq.max] = wr->wr_id; in mthca_tavor_post_send()
1867 qp->wrid[ind] = wr->wr_id; in mthca_tavor_post_receive()
2092 qp->wrid[ind + qp->rq.max] = wr->wr_id; in mthca_arbel_post_send()
2205 qp->wrid[ind] = wr->wr_id; in mthca_arbel_post_receive()
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
Dcq.c640 *cookie = wq->sq.sw_sq[wq->sq.cidx].wr_id; in poll_cq()
646 *cookie = wq->rq.sw_rq[wq->rq.cidx].wr_id; in poll_cq()
708 wc->wr_id = cookie; in c4iw_poll_cq_one()
Dt4.h262 u64 wr_id; member
310 u64 wr_id; member
Dqp.c840 swsqe->wr_id = wr->wr_id; in c4iw_post_send()
850 __func__, (unsigned long long)wr->wr_id, qhp->wq.sq.pidx, in c4iw_post_send()
907 qhp->wq.rq.sw_rq[qhp->wq.rq.pidx].wr_id = wr->wr_id; in c4iw_post_receive()
924 (unsigned long long) wr->wr_id, qhp->wq.rq.pidx); in c4iw_post_receive()
Ddevice.c136 le.wr_id = CQE_WRID_SQ_IDX(cqe); in c4iw_log_wr_stats()
142 le.wr_id = CQE_WRID_MSN(cqe); in c4iw_log_wr_stats()
182 lep->wr_id, in wr_log_show()
Diw_cxgb4.h161 u16 wr_id; member
/linux-4.4.14/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c2296 qp->wqe_wr_id_tbl[qp->sq.head].wrid = wr->wr_id; in ocrdma_post_send()
2364 qp->rqe_wr_id_tbl[qp->rq.head] = wr->wr_id; in ocrdma_post_recv()
2435 srq->rqe_wr_id_tbl[tag] = wr->wr_id; in ocrdma_post_srq_recv()
2532 ibwc->wr_id = qp->wqe_wr_id_tbl[wqe_idx].wrid; in ocrdma_update_wc()
2619 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_update_err_rcqe()
2751 ibwc->wr_id = srq->rqe_wr_id_tbl[wqe_idx]; in ocrdma_update_free_srq_cqe()
2819 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_poll_success_rcqe()
2937 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_add_err_cqe()
/linux-4.4.14/drivers/staging/rdma/amso1100/
Dc2_qp.c812 wr.sqwr.sq_hdr.user_hdr.hdr.context = ib_wr->wr_id; in c2_post_send()
973 wr.rqwr.rq_hdr.user_hdr.hdr.context = ib_wr->wr_id; in c2_post_receive()
Dc2_cq.c157 entry->wr_id = ce->hdr.context; in c2_poll_one()
/linux-4.4.14/include/rdma/
Dib_verbs.h761 u64 wr_id; member
1084 u64 wr_id; member
1178 u64 wr_id; member
1221 u64 wr_id; member
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c171 rx->rx_wrq.wr_id = kiblnd_ptr2wreqid(rx, IBLND_WID_RX); in kiblnd_post_rx()
1035 wrq->wr.wr_id = kiblnd_ptr2wreqid(tx, IBLND_WID_TX); in kiblnd_init_tx_msg()
1100 wrq->wr.wr_id = kiblnd_ptr2wreqid(tx, IBLND_WID_RDMA); in kiblnd_init_rdma()
3204 switch (kiblnd_wreqid2type(wc->wr_id)) { in kiblnd_complete()
3216 kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete()
3220 kiblnd_tx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete()
3224 kiblnd_rx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status, in kiblnd_complete()
/linux-4.4.14/drivers/infiniband/hw/nes/
Dnes_verbs.c243 u64temp = ibmw_bind->wr_id; in nes_bind_mw()
3295 u64temp = (u64)(ib_wr->wr_id); in nes_post_send()
3564 u64temp = (u64)(ib_wr->wr_id); in nes_post_recv()
3734 entry->wr_id = wrid; in nes_poll_cq()