/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_ib.c | 108 priv->rx_wr.wr_id = id | IPOIB_OP_RECV; in ipoib_ib_post_receive() 179 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc() local 185 wr_id, wc->status); in ipoib_ib_handle_rx_wc() 187 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_ib_handle_rx_wc() 189 wr_id, ipoib_recvq_size); in ipoib_ib_handle_rx_wc() 193 skb = priv->rx_ring[wr_id].skb; in ipoib_ib_handle_rx_wc() 199 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc() 200 ipoib_ud_dma_unmap_rx(priv, priv->rx_ring[wr_id].mapping); in ipoib_ib_handle_rx_wc() 202 priv->rx_ring[wr_id].skb = NULL; in ipoib_ib_handle_rx_wc() 213 memcpy(mapping, priv->rx_ring[wr_id].mapping, in ipoib_ib_handle_rx_wc() [all …]
|
D | ipoib_cm.c | 73 .wr_id = IPOIB_CM_RX_DRAIN_WRID, 97 priv->cm.rx_wr.wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_srq() 123 wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_nonsrq() 562 unsigned int wr_id = wc->wr_id & ~(IPOIB_OP_CM | IPOIB_OP_RECV); in ipoib_cm_handle_rx_wc() local 572 wr_id, wc->status); in ipoib_cm_handle_rx_wc() 574 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_cm_handle_rx_wc() 575 if (wr_id == (IPOIB_CM_RX_DRAIN_WRID & ~(IPOIB_OP_CM | IPOIB_OP_RECV))) { in ipoib_cm_handle_rx_wc() 583 wr_id, ipoib_recvq_size); in ipoib_cm_handle_rx_wc() 592 skb = rx_ring[wr_id].skb; in ipoib_cm_handle_rx_wc() 597 wc->status, wr_id, wc->vendor_err); in ipoib_cm_handle_rx_wc() [all …]
|
/linux-4.1.27/net/rds/ |
D | ib_ring.c | 156 u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest) in rds_ib_ring_completed() argument 160 if (oldest <= (unsigned long long)wr_id) in rds_ib_ring_completed() 161 ret = (unsigned long long)wr_id - oldest + 1; in rds_ib_ring_completed() 163 ret = ring->w_nr - oldest + (unsigned long long)wr_id + 1; in rds_ib_ring_completed() 166 wr_id, oldest); in rds_ib_ring_completed()
|
D | iw_ring.c | 157 u32 rds_iw_ring_completed(struct rds_iw_work_ring *ring, u32 wr_id, u32 oldest) in rds_iw_ring_completed() argument 161 if (oldest <= (unsigned long long)wr_id) in rds_iw_ring_completed() 162 ret = (unsigned long long)wr_id - oldest + 1; in rds_iw_ring_completed() 164 ret = ring->w_nr - oldest + (unsigned long long)wr_id + 1; in rds_iw_ring_completed() 167 wr_id, oldest); in rds_iw_ring_completed()
|
D | iw_send.c | 141 send->s_wr.wr_id = i; in rds_iw_send_init_ring() 215 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_iw_send_cq_comp_handler() 224 if (wc.opcode == IB_WC_LOCAL_INV && wc.wr_id == RDS_IW_LOCAL_INV_WR_ID) { in rds_iw_send_cq_comp_handler() 229 if (wc.opcode == IB_WC_FAST_REG_MR && wc.wr_id == RDS_IW_FAST_REG_WR_ID) { in rds_iw_send_cq_comp_handler() 234 if (wc.wr_id == RDS_IW_ACK_WR_ID) { in rds_iw_send_cq_comp_handler() 243 completed = rds_iw_ring_completed(&ic->i_send_ring, wc.wr_id, oldest); in rds_iw_send_cq_comp_handler()
|
D | iw_recv.c | 94 recv->r_wr.wr_id = i; in rds_iw_recv_init_ring() 359 wr->wr_id = RDS_IW_ACK_WR_ID; in rds_iw_recv_init_ack() 789 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_poll_cq()
|
D | ib_send.c | 232 send->s_wr.wr_id = i; in rds_ib_send_init_ring() 295 (unsigned long long)wc.wr_id, wc.status, in rds_ib_send_cq_comp_handler() 300 if (wc.wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cq_comp_handler() 309 completed = rds_ib_ring_completed(&ic->i_send_ring, wc.wr_id, oldest); in rds_ib_send_cq_comp_handler()
|
D | ib_recv.c | 58 recv->r_wr.wr_id = i; in rds_ib_recv_init_ring() 528 wr->wr_id = RDS_IB_ACK_WR_ID; in rds_ib_recv_init_ack() 958 (unsigned long long)wc.wr_id, wc.status, in rds_poll_cq()
|
D | ib.h | 338 u32 rds_ib_ring_completed(struct rds_ib_work_ring *ring, u32 wr_id, u32 oldest);
|
D | iw_rdma.c | 711 f_wr.wr_id = RDS_IW_FAST_REG_WR_ID; 742 s_wr.wr_id = RDS_IW_LOCAL_INV_WR_ID;
|
D | iw.h | 347 u32 rds_iw_ring_completed(struct rds_iw_work_ring *ring, u32 wr_id, u32 oldest);
|
/linux-4.1.27/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.h | 144 static inline enum srpt_opcode opcode_from_wr_id(u64 wr_id) in opcode_from_wr_id() argument 146 return wr_id >> 32; in opcode_from_wr_id() 148 static inline u32 idx_from_wr_id(u64 wr_id) in idx_from_wr_id() argument 150 return (u32)wr_id; in idx_from_wr_id()
|
D | ib_srpt.c | 782 wr.wr_id = encode_wr_id(SRPT_RECV, ioctx->ioctx.index); in srpt_post_recv() 824 wr.wr_id = encode_wr_id(SRPT_SEND, ioctx->ioctx.index); in srpt_post_send() 1386 static void srpt_handle_send_err_comp(struct srpt_rdma_ch *ch, u64 wr_id) in srpt_handle_send_err_comp() argument 1395 index = idx_from_wr_id(wr_id); in srpt_handle_send_err_comp() 1885 index = idx_from_wr_id(wc->wr_id); in srpt_process_rcv_completion() 1923 index = idx_from_wr_id(wc->wr_id); in srpt_process_send_completion() 1924 opcode = opcode_from_wr_id(wc->wr_id); in srpt_process_send_completion() 1938 srpt_handle_send_err_comp(ch, wc->wr_id); in srpt_process_send_completion() 1970 if (opcode_from_wr_id(wc[i].wr_id) == SRPT_RECV) in srpt_process_completion() 2803 wr.wr_id = encode_wr_id(i == n_rdma - 1 ? in srpt_perform_rdmas() [all …]
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
D | iser_verbs.c | 960 ib_conn->beacon.wr_id = ISER_BEACON_WRID; in iser_connect() 1014 rx_wr.wr_id = (uintptr_t)iser_conn->login_resp_buf; in iser_post_recvl() 1038 rx_wr->wr_id = (uintptr_t)rx_desc; in iser_post_recvm() 1075 send_wr.wr_id = (uintptr_t)tx_desc; in iser_post_send() 1100 is_iser_tx_desc(struct iser_conn *iser_conn, void *wr_id) in is_iser_tx_desc() argument 1105 if (wr_id >= start && wr_id < start + len) in is_iser_tx_desc() 1125 void *wr_id = (void *)(uintptr_t)wc->wr_id; in iser_handle_comp_error() local 1134 if (wc->wr_id == ISER_FASTREG_LI_WRID) in iser_handle_comp_error() 1137 if (is_iser_tx_desc(iser_conn, wr_id)) { in iser_handle_comp_error() 1138 struct iser_tx_desc *desc = wr_id; in iser_handle_comp_error() [all …]
|
D | iser_memory.c | 642 inv_wr->wr_id = ISER_FASTREG_LI_WRID; in iser_inv_rkey() 677 sig_wr.wr_id = ISER_FASTREG_LI_WRID; in iser_reg_sig_mr() 752 fastreg_wr.wr_id = ISER_FASTREG_LI_WRID; in iser_fast_reg_mr()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_reqs.c | 55 static u64 replace_wr_id(u64 wr_id, u16 idx) in replace_wr_id() argument 59 ret = wr_id & ~QMAP_IDX_MASK; in replace_wr_id() 65 static u16 get_app_wr_id(u64 wr_id) in get_app_wr_id() argument 67 return wr_id & QMAP_IDX_MASK; in get_app_wr_id() 87 wqe_p->work_request_id = replace_wr_id(recv_wr->wr_id, rq_map_idx); in ehca_write_rwqe() 121 "send_flags=%x opcode=%x", idx, send_wr->wr_id, in trace_send_wr_ud() 177 wqe_p->work_request_id = replace_wr_id(send_wr->wr_id, sq_map_idx); in ehca_write_swqe() 179 qmap_entry->app_wr_id = get_app_wr_id(send_wr->wr_id); in ehca_write_swqe() 560 qmap_entry->app_wr_id = get_app_wr_id(recv_wr->wr_id); in internal_post_recv() 743 wc->wr_id = replace_wr_id(cqe->work_request_id, qmap_entry->app_wr_id); in ehca_poll_cq_one() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | mad.c | 596 wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); in mlx4_ib_send_to_slave() 1135 recv_wr.wr_id = (u64) index | MLX4_TUN_WRID_RECV | in mlx4_ib_post_pv_qp_buf() 1249 wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum); in mlx4_ib_send_to_wire() 1284 struct mlx4_ib_demux_pv_qp *tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc->wr_id)]; in mlx4_ib_multiplex_mad() 1285 int wr_ix = wc->wr_id & (MLX4_NUM_TUNNEL_BUFS - 1); in mlx4_ib_multiplex_mad() 1532 tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc.wr_id)]; in mlx4_ib_tunnel_comp_worker() 1538 wc.wr_id & in mlx4_ib_tunnel_comp_worker() 1542 "buf:%lld\n", wc.wr_id); in mlx4_ib_tunnel_comp_worker() 1547 wc.wr_id, wc.status); in mlx4_ib_tunnel_comp_worker() 1548 ib_destroy_ah(tun_qp->tx_ring[wc.wr_id & in mlx4_ib_tunnel_comp_worker() [all …]
|
D | cq.c | 612 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_qp_sw_comp() 745 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one() 750 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one() 755 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one() 760 wc->wr_id = wq->wrid[tail]; in mlx4_ib_poll_one()
|
D | mr.c | 330 wr.wr_id = mw_bind->wr_id; in mlx4_ib_bind_mw()
|
D | srq.c | 342 srq->wrid[srq->head] = wr->wr_id; in mlx4_ib_post_srq_recv()
|
D | qp.c | 2687 qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id; in mlx4_ib_post_send() 3017 qp->rq.wrid[ind] = wr->wr_id; in mlx4_ib_post_recv()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_srq.c | 80 wqe->wr_id = wr->wr_id; in ipath_post_srq_receive() 290 p->wr_id = wqe->wr_id; in ipath_modify_srq()
|
D | ipath_ruc.c | 146 wc.wr_id = wqe->wr_id; in ipath_init_sge() 212 qp->r_wr_id = wqe->wr_id; in ipath_get_rwqe() 436 wc.wr_id = qp->r_wr_id; in ipath_ruc_loopback() 714 wc.wr_id = wqe->wr.wr_id; in ipath_send_complete()
|
D | ipath_ud.c | 149 wc.wr_id = wqe->wr_id; in ipath_ud_loopback() 557 wc.wr_id = qp->r_wr_id; in ipath_ud_rcv()
|
D | ipath_cq.c | 81 wc->uqueue[head].wr_id = entry->wr_id; in ipath_cq_enter()
|
D | ipath_qp.c | 404 wc.wr_id = qp->r_wr_id; in ipath_error_qp() 426 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in ipath_error_qp()
|
D | ipath_uc.c | 413 wc.wr_id = qp->r_wr_id; in ipath_uc_rcv()
|
D | ipath_rc.c | 952 wc.wr_id = wqe->wr.wr_id; in do_rc_ack() 1733 wc.wr_id = qp->r_wr_id; in ipath_rc_rcv()
|
D | ipath_verbs.h | 294 u64 wr_id; member
|
D | ipath_verbs.c | 512 wqe->wr_id = wr->wr_id; in ipath_post_receive()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_srq.c | 80 wqe->wr_id = wr->wr_id; in qib_post_srq_receive() 286 p->wr_id = wqe->wr_id; in qib_modify_srq()
|
D | qib_ruc.c | 117 wc.wr_id = wqe->wr_id; in qib_init_sge() 189 qp->r_wr_id = wqe->wr_id; in qib_get_rwqe() 554 wc.wr_id = qp->r_wr_id; in qib_ruc_loopback() 796 wc.wr_id = wqe->wr.wr_id; in qib_send_complete()
|
D | qib_cq.c | 84 wc->uqueue[head].wr_id = entry->wr_id; in qib_cq_enter()
|
D | qib_ud.c | 207 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 565 wc.wr_id = qp->r_wr_id; in qib_ud_rcv()
|
D | qib_rc.c | 1021 wc.wr_id = wqe->wr.wr_id; in qib_rc_send_complete() 1077 wc.wr_id = wqe->wr.wr_id; in do_rc_completion() 2031 wc.wr_id = qp->r_wr_id; in qib_rc_rcv()
|
D | qib_qp.c | 523 wc.wr_id = qp->r_wr_id; in qib_error_qp() 545 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in qib_error_qp()
|
D | qib_uc.c | 405 wc.wr_id = qp->r_wr_id; in qib_uc_rcv()
|
D | qib_verbs.h | 355 u64 wr_id; member
|
D | qib_verbs.c | 532 wqe->wr_id = wr->wr_id; in qib_post_receive()
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
D | vnic_wq.h | 61 uint64_t wr_id; /* Cookie */ member 134 buf->wr_id = wrid; in vnic_wq_post()
|
D | vnic_rq.h | 75 uint64_t wr_id; member 138 buf->wr_id = wrid; in vnic_rq_post()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.h | 101 u64 wr_id; member 140 u64 wr_id; /* work request ID */ member 163 u64 wr_id; /* work request ID from WQE */ member
|
D | ehea_main.c | 501 rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) in ehea_refill_rq_def() 578 int skb_index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in get_skb_by_index() 793 swqe->wr_id = SWQE_RESTART_CHECK; in check_sqs() 832 if (cqe->wr_id == SWQE_RESTART_CHECK) { in ehea_proc_cqes() 855 if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id) in ehea_proc_cqes() 858 index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in ehea_proc_cqes() 864 swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); in ehea_proc_cqes() 2079 swqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, EHEA_SWQE3_TYPE) in ehea_start_xmit() 2082 swqe->wr_id |= EHEA_BMASK_SET(EHEA_WR_ID_REFILL, in ehea_start_xmit() 2089 swqe->wr_id = in ehea_start_xmit() [all …]
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
D | frwr_ops.c | 130 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in frwr_sendcompletion() 222 fastreg_wr.wr_id = (unsigned long)(void *)mw; in frwr_op_map() 271 invalidate_wr.wr_id = (unsigned long)(void *)seg1->rl_mw; in frwr_op_unmap()
|
D | verbs.c | 208 if (wc->wr_id == RPCRDMA_IGNORE_COMPLETION) { in rpcrdma_sendcq_process_wc() 216 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in rpcrdma_sendcq_process_wc() 280 (struct rpcrdma_rep *)(unsigned long)wc->wr_id; in rpcrdma_recvcq_process_wc() 1236 invalidate_wr.wr_id = (unsigned long)(void *)r; in rpcrdma_retry_local_inv() 1593 send_wr.wr_id = RPCRDMA_IGNORE_COMPLETION; in rpcrdma_ep_post() 1635 recv_wr.wr_id = (u64) (unsigned long) rep; in rpcrdma_ep_post_recv()
|
D | svc_rdma_recvfrom.c | 193 read_wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_lcl() 324 read_wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_frmr() 331 inv_wr.wr_id = (unsigned long)ctxt; in rdma_read_chunk_frmr()
|
D | svc_rdma_transport.c | 300 ctxt = (struct svc_rdma_op_ctxt *)(unsigned long)wc.wr_id; in rq_cq_reap() 417 (unsigned long)wc->wr_id; in sq_cq_reap() 540 recv_wr.wr_id = (u64)(unsigned long)ctxt; in svc_rdma_post_recv() 1352 err_wr.wr_id = (unsigned long)ctxt; in svc_rdma_send_error()
|
D | svc_rdma_sendto.c | 212 write_wr.wr_id = (unsigned long)ctxt; in send_write() 472 send_wr.wr_id = (unsigned long)ctxt; in send_reply()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
D | iwch_qp.c | 281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_rdma_recv() 344 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_zero_stag_recv() 436 sqp->wr_id = wr->wr_id; in iwch_post_send() 447 __func__, (unsigned long long) wr->wr_id, idx, in iwch_post_send() 511 "wqe %p \n", __func__, (unsigned long long) wr->wr_id, in iwch_post_receive() 590 sqp->wr_id = mw_bind->wr_id; in iwch_bind_mw()
|
D | iwch_cq.c | 81 wc->wr_id = cookie; in iwch_poll_cq_one()
|
D | cxio_wr.h | 674 __u64 wr_id; member 684 __u64 wr_id; member
|
D | cxio_hal.c | 1303 *cookie = wq->sq[Q_PTR2IDX(wq->sq_rptr, wq->sq_size_log2)].wr_id; in cxio_poll_cq() 1308 *cookie = wq->rq[Q_PTR2IDX(wq->rq_rptr, wq->rq_size_log2)].wr_id; in cxio_poll_cq()
|
/linux-4.1.27/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 1018 rx_wr->wr_id = (uintptr_t)rx_desc; in isert_post_recv() 1052 send_wr.wr_id = (uintptr_t)tx_desc; in isert_post_send() 1122 send_wr->wr_id = (uintptr_t)&isert_cmd->tx_desc; in isert_init_send_wr() 1145 rx_wr.wr_id = (uintptr_t)isert_conn->login_req_buf; in isert_rdma_post_recvl() 2057 is_isert_tx_desc(struct isert_conn *isert_conn, void *wr_id) in is_isert_tx_desc() argument 2062 if ((wr_id >= start && wr_id < start + len) || in is_isert_tx_desc() 2063 (wr_id == isert_conn->login_req_buf)) in is_isert_tx_desc() 2072 if (wc->wr_id == ISER_BEACON_WRID) { in isert_cq_comp_err() 2076 } else if (is_isert_tx_desc(isert_conn, (void *)(uintptr_t)wc->wr_id)) { in isert_cq_comp_err() 2081 desc = (struct iser_tx_desc *)(uintptr_t)wc->wr_id; in isert_cq_comp_err() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
D | cq.c | 191 wc->wr_id = srq->wrid[wqe_ctr]; in handle_responder() 198 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder() 471 wc->wr_id = wq->wrid[idx]; in mlx5_poll_one() 497 wc->wr_id = wq->wrid[idx]; in mlx5_poll_one() 505 wc->wr_id = srq->wrid[wqe_ctr]; in mlx5_poll_one() 509 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
|
D | mr.c | 744 context = (struct mlx5_ib_umr_context *) (unsigned long) wc.wr_id; in mlx5_umr_cq_handler() 806 wr.wr_id = (u64)(unsigned long)&umr_context; in reg_umr() 924 wr.wr_id = (u64)(unsigned long)&umr_context; in mlx5_ib_update_mtt() 1134 wr.wr_id = (u64)(unsigned long)&umr_context; in unreg_umr()
|
D | qp.c | 2566 u8 size, unsigned idx, u64 wr_id, in finish_wqe() argument 2580 qp->sq.wrid[idx] = wr_id; in finish_wqe() 2703 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send() 2730 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send() 2752 finish_wqe(qp, ctrl, size, idx, wr->wr_id, in mlx5_ib_post_send() 2838 finish_wqe(qp, ctrl, size, idx, wr->wr_id, nreq, in mlx5_ib_post_send() 2942 qp->rq.wrid[ind] = wr->wr_id; in mlx5_ib_post_recv()
|
D | srq.c | 452 srq->wrid[srq->head] = wr->wr_id; in mlx5_ib_post_srq_recv()
|
/linux-4.1.27/include/uapi/rdma/ |
D | ib_user_verbs.h | 381 __u64 wr_id; member 638 __u64 wr_id; member 682 __u64 wr_id; member
|
/linux-4.1.27/net/9p/ |
D | trans_rdma.c | 353 struct p9_rdma_context *c = (void *) (unsigned long) wc.wr_id; in cq_comp_handler() 422 wr.wr_id = (unsigned long) c; in post_recv() 513 wr.wr_id = (unsigned long) c; in rdma_request()
|
/linux-4.1.27/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 469 static struct ib_recv_wr wr = { .wr_id = SRP_LAST_WR_ID }; in srp_destroy_qp() 1043 .wr_id = LOCAL_INV_WR_ID_MASK, in srp_inv_rkey() 1314 wr.wr_id = FAST_REG_WR_ID_MASK; in srp_map_finish_fr() 1695 wr.wr_id = (uintptr_t) iu; in srp_post_send() 1715 wr.wr_id = (uintptr_t) iu; in srp_post_recv() 1851 struct srp_iu *iu = (struct srp_iu *) (uintptr_t) wc->wr_id; in srp_handle_recv() 1917 static void srp_handle_qp_err(u64 wr_id, enum ib_wc_status wc_status, in srp_handle_qp_err() argument 1922 if (wr_id == SRP_LAST_WR_ID) { in srp_handle_qp_err() 1928 if (wr_id & LOCAL_INV_WR_ID_MASK) { in srp_handle_qp_err() 1932 } else if (wr_id & FAST_REG_WR_ID_MASK) { in srp_handle_qp_err() [all …]
|
/linux-4.1.27/drivers/infiniband/core/ |
D | mad.c | 703 u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, in build_smp_wc() argument 707 wc->wr_id = wr_id; in build_smp_wc() 783 send_wr->wr_id, be16_to_cpu(smp->dr_slid), in handle_outgoing_dr_smp() 960 mad_send_wr->send_wr.wr_id = (unsigned long) mad_send_wr; in ib_create_send_mad() 1072 mad_send_wr->send_wr.wr_id = (unsigned long)&mad_send_wr->mad_list; in ib_send_mad() 1894 mad_recv_wc->wc->wr_id = 0; in ib_mad_complete_recv() 1910 mad_recv_wc->wc->wr_id = (unsigned long) &mad_send_wr->send_buf; in ib_mad_complete_recv() 1958 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_recv_done_handler() 2212 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_send_done_handler() 2286 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in mad_error_handler() [all …]
|
D | uverbs_cmd.c | 1475 tmp.wr_id = wc->wr_id; in copy_wc_to_user() 2241 next->wr_id = user_wr->wr_id; in ib_uverbs_post_send() 2398 next->wr_id = user_wr->wr_id; in ib_uverbs_unmarshall_recv()
|
D | sa_query.c | 1136 mad_buf = (void *) (unsigned long) mad_recv_wc->wc->wr_id; in recv_handler()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 542 entry->wr_id = (*cur_qp)->wrid[wqe_index + in mthca_poll_one() 549 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one() 563 entry->wr_id = (*cur_qp)->wrid[wqe_index]; in mthca_poll_one()
|
D | mthca_srq.c | 532 srq->wrid[ind] = wr->wr_id; in mthca_tavor_post_srq_recv() 621 srq->wrid[ind] = wr->wr_id; in mthca_arbel_post_srq_recv()
|
D | mthca_qp.c | 1751 qp->wrid[ind + qp->rq.max] = wr->wr_id; in mthca_tavor_post_send() 1867 qp->wrid[ind] = wr->wr_id; in mthca_tavor_post_receive() 2092 qp->wrid[ind + qp->rq.max] = wr->wr_id; in mthca_arbel_post_send() 2205 qp->wrid[ind] = wr->wr_id; in mthca_arbel_post_receive()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
D | cq.c | 642 *cookie = wq->sq.sw_sq[wq->sq.cidx].wr_id; in poll_cq() 648 *cookie = wq->rq.sw_rq[wq->rq.cidx].wr_id; in poll_cq() 710 wc->wr_id = cookie; in c4iw_poll_cq_one()
|
D | t4.h | 261 u64 wr_id; member 307 u64 wr_id; member
|
D | qp.c | 825 swsqe->wr_id = wr->wr_id; in c4iw_post_send() 835 __func__, (unsigned long long)wr->wr_id, qhp->wq.sq.pidx, in c4iw_post_send() 893 qhp->wq.rq.sw_rq[qhp->wq.rq.pidx].wr_id = wr->wr_id; in c4iw_post_receive() 910 (unsigned long long) wr->wr_id, qhp->wq.rq.pidx); in c4iw_post_receive()
|
D | device.c | 136 le.wr_id = CQE_WRID_SQ_IDX(cqe); in c4iw_log_wr_stats() 142 le.wr_id = CQE_WRID_MSN(cqe); in c4iw_log_wr_stats() 182 lep->wr_id, in wr_log_show()
|
D | iw_cxgb4.h | 161 u16 wr_id; member
|
/linux-4.1.27/include/rdma/ |
D | ib_verbs.h | 696 u64 wr_id; member 1030 u64 wr_id; member 1089 u64 wr_id; member 1132 u64 wr_id; member
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 2248 qp->wqe_wr_id_tbl[qp->sq.head].wrid = wr->wr_id; in ocrdma_post_send() 2316 qp->rqe_wr_id_tbl[qp->rq.head] = wr->wr_id; in ocrdma_post_recv() 2387 srq->rqe_wr_id_tbl[tag] = wr->wr_id; in ocrdma_post_srq_recv() 2484 ibwc->wr_id = qp->wqe_wr_id_tbl[wqe_idx].wrid; in ocrdma_update_wc() 2571 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_update_err_rcqe() 2703 ibwc->wr_id = srq->rqe_wr_id_tbl[wqe_idx]; in ocrdma_update_free_srq_cqe() 2771 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_poll_success_rcqe() 2894 ibwc->wr_id = qp->rqe_wr_id_tbl[qp->rq.tail]; in ocrdma_add_err_cqe()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd_cb.c | 170 rx->rx_wrq.wr_id = kiblnd_ptr2wreqid(rx, IBLND_WID_RX); in kiblnd_post_rx() 1076 wrq->wr_id = kiblnd_ptr2wreqid(tx, IBLND_WID_TX); in kiblnd_init_tx_msg() 1140 wrq->wr_id = kiblnd_ptr2wreqid(tx, IBLND_WID_RDMA); in kiblnd_init_rdma() 3242 switch (kiblnd_wreqid2type(wc->wr_id)) { in kiblnd_complete() 3254 kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete() 3258 kiblnd_tx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete() 3262 kiblnd_rx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status, in kiblnd_complete()
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2_cq.c | 157 entry->wr_id = ce->hdr.context; in c2_poll_one()
|
D | c2_qp.c | 812 wr.sqwr.sq_hdr.user_hdr.hdr.context = ib_wr->wr_id; in c2_post_send() 973 wr.rqwr.rq_hdr.user_hdr.hdr.context = ib_wr->wr_id; in c2_post_receive()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
D | nes_verbs.c | 242 u64temp = ibmw_bind->wr_id; in nes_bind_mw() 3301 u64temp = (u64)(ib_wr->wr_id); in nes_post_send() 3588 u64temp = (u64)(ib_wr->wr_id); in nes_post_recv() 3758 entry->wr_id = wrid; in nes_poll_cq()
|