Home
last modified time | relevance | path

Searched refs:ib_sge (Results 1 – 46 of 46) sorted by relevance

/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.h64 struct ib_sge rx_sg;
73 struct ib_sge tx_sg[2];
120 struct ib_sge *ib_sge; member
121 struct ib_sge s_ib_sge;
125 struct ib_sge ib_sg[3];
Dib_isert.c214 struct ib_sge *rx_sg; in isert_alloc_rx_descriptors()
1133 struct ib_sge sge; in isert_rdma_post_recvl()
1136 memset(&sge, 0, sizeof(struct ib_sge)); in isert_rdma_post_recvl()
1177 struct ib_sge *tx_dsg = &tx_desc->tx_sg[1]; in isert_put_login_tx()
1712 if (wr->ib_sge) { in isert_unmap_cmd()
1714 kfree(wr->ib_sge); in isert_unmap_cmd()
1715 wr->ib_sge = NULL; in isert_unmap_cmd()
1743 wr->ib_sge = NULL; in isert_unreg_rdma()
2187 struct ib_sge *tx_dsg = &isert_cmd->tx_desc.tx_sg[1]; in isert_put_response()
2318 struct ib_sge *tx_dsg = &isert_cmd->tx_desc.tx_sg[1]; in isert_put_reject()
[all …]
Disert_proto.h40 (ISER_RX_PAYLOAD_SIZE + sizeof(u64) + sizeof(struct ib_sge)))
/linux-4.1.27/net/rds/
Diw.h81 struct ib_sge s_sge[RDS_IW_MAX_SGE];
89 struct ib_sge r_sge[2];
144 struct ib_sge i_ack_sge;
383 static inline struct ib_sge *
384 rds_iw_header_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_header_sge()
389 static inline struct ib_sge *
390 rds_iw_data_sge(struct rds_iw_connection *ic, struct ib_sge *sge) in rds_iw_data_sge()
Dib.h73 struct ib_sge s_sge[RDS_IB_MAX_SGE];
81 struct ib_sge r_sge[2];
138 struct ib_sge i_ack_sge;
Diw_recv.c88 struct ib_sge *sge; in rds_iw_recv_init_ring()
143 struct ib_sge *sge; in rds_iw_recv_refill_one()
350 struct ib_sge *sge = &ic->i_ack_sge; in rds_iw_recv_init_ack()
Dib_recv.c52 struct ib_sge *sge; in rds_ib_recv_init_ring()
303 struct ib_sge *sge; in rds_ib_recv_refill_one()
519 struct ib_sge *sge = &ic->i_ack_sge; in rds_ib_recv_init_ack()
Diw_send.c134 struct ib_sge *sge; in rds_iw_send_init_ring()
453 struct ib_sge *sge; in rds_iw_xmit_populate_wr()
Dib_send.c228 struct ib_sge *sge; in rds_ib_send_init_ring()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Discsi_iser.h251 struct ib_sge sge;
280 struct ib_sge tx_sg[2];
286 sizeof(u64) + sizeof(struct ib_sge)))
302 struct ib_sge rx_sg;
Diser_initiator.c111 struct ib_sge *tx_dsg = &iser_task->desc.tx_sg[1]; in iser_prepare_write_cmd()
261 struct ib_sge *rx_sg; in iser_alloc_rx_descriptors()
458 struct ib_sge *tx_dsg; in iser_send_data_out()
530 struct ib_sge *tx_dsg = &mdesc->tx_sg[1]; in iser_send_control()
Diser_verbs.c1007 struct ib_sge sge; in iser_post_recvl()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_wqe.h117 struct ib_sge *sg) in mthca_set_data_seg()
/linux-4.1.27/net/sunrpc/xprtrdma/
Dxprt_rdma.h123 struct ib_sge rg_iov;
261 struct ib_sge rl_send_iov[4]; /* for active requests */
Dsvc_rdma_sendto.c148 struct ib_sge *sge; in send_write()
Dverbs.c1449 struct ib_mr **mrp, struct ib_sge *iov) in rpcrdma_register_internal()
1500 struct ib_mr *mr, struct ib_sge *iov) in rpcrdma_deregister_internal()
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib.h272 struct ib_sge rx_sge[IPOIB_CM_RX_SG];
363 struct ib_sge tx_sge[MAX_SKB_FRAGS + 1];
369 struct ib_sge rx_sge[IPOIB_UD_RX_SG];
Dipoib_cm.c117 struct ib_sge *sge, int id) in ipoib_cm_post_receive_nonsrq()
329 struct ib_sge *sge) in ipoib_cm_init_rx_wr()
352 struct ib_sge sge[IPOIB_CM_RX_SG]; in ipoib_cm_nonsrq_init_rx()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_srq.c137 sz = sizeof(struct ib_sge) * srq->rq.max_sge + in ipath_create_srq()
237 srq->rq.max_sge * sizeof(struct ib_sge); in ipath_modify_srq()
Dipath_verbs.h296 struct ib_sge sg_list[0];
499 rq->max_sge * sizeof(struct ib_sge)) * n); in get_rwqe_ptr()
786 struct ib_sge *sge, int acc);
Dipath_keys.c122 struct ib_sge *sge, int acc) in ipath_lkey_ok()
Dipath_qp.c830 sz = (sizeof(struct ib_sge) * qp->r_rq.max_sge) + in ipath_create_qp()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_srq.c134 sz = sizeof(struct ib_sge) * srq->rq.max_sge + in qib_create_srq()
233 srq->rq.max_sge * sizeof(struct ib_sge); in qib_modify_srq()
Dqib_verbs.h357 struct ib_sge sg_list[0];
627 rq->max_sge * sizeof(struct ib_sge)) * n); in get_rwqe_ptr()
984 struct qib_sge *isge, struct ib_sge *sge, int acc);
Dqib_keys.c156 struct qib_sge *isge, struct ib_sge *sge, int acc) in qib_lkey_ok()
Dqib_qp.c1069 sz = (sizeof(struct ib_sge) * qp->r_rq.max_sge) + in qib_create_qp()
/linux-4.1.27/drivers/infiniband/core/
Dmad_priv.h129 struct ib_sge sg_list[IB_MAD_SEND_REQ_MAX_SG];
Duverbs_cmd.c2226 next = kmalloc(ALIGN(sizeof *next, sizeof (struct ib_sge)) + in ib_uverbs_post_send()
2227 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_post_send()
2302 ALIGN(sizeof *next, sizeof (struct ib_sge)); in ib_uverbs_post_send()
2306 sg_ind * sizeof (struct ib_sge), in ib_uverbs_post_send()
2307 next->num_sge * sizeof (struct ib_sge))) { in ib_uverbs_post_send()
2383 next = kmalloc(ALIGN(sizeof *next, sizeof (struct ib_sge)) + in ib_uverbs_unmarshall_recv()
2384 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_unmarshall_recv()
2403 ALIGN(sizeof *next, sizeof (struct ib_sge)); in ib_uverbs_unmarshall_recv()
2406 sg_ind * sizeof (struct ib_sge), in ib_uverbs_unmarshall_recv()
2407 next->num_sge * sizeof (struct ib_sge))) { in ib_uverbs_unmarshall_recv()
Dmad.c1066 struct ib_sge *sge; in ib_send_mad()
2647 struct ib_sge sg_list; in ib_mad_post_receive_mads()
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.h156 struct ib_sge *sge;
Dib_srpt.c778 struct ib_sge list; in srpt_post_recv()
803 struct ib_sge list; in srpt_post_send()
1089 struct ib_sge *sge; in srpt_map_sg_to_ib_sge()
/linux-4.1.27/include/linux/sunrpc/
Dsvc_rdma.h85 struct ib_sge sge[RPCSVC_MAXPAGES];
/linux-4.1.27/net/9p/
Dtrans_rdma.c408 struct ib_sge sge; in post_recv()
436 struct ib_sge sge; in rdma_request()
/linux-4.1.27/include/rdma/
Dib_verbs.h999 struct ib_sge { struct
1031 struct ib_sge *sg_list;
1081 struct ib_sge *prot;
1090 struct ib_sge *sg_list;
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd.h521 struct ib_sge rx_sge; /* ...and its memory */
546 struct ib_sge *tx_sge; /* ...and their memory */
Do2iblnd_cb.c1055 struct ib_sge *sge = &tx->tx_sge[tx->tx_nwrq]; in kiblnd_init_tx_msg()
1091 struct ib_sge *sge = &tx->tx_sge[0]; in kiblnd_init_rdma()
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dmr.c688 struct ib_sge *sg, u64 dma, int n, u32 key, in prep_umr_reg_wqe()
761 struct ib_sge sg; in reg_umr()
860 struct ib_sge sg; in mlx5_ib_update_mtt()
Dqp.c1858 static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ib_sge *sg) in set_data_ptr_seg()
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_qp.c200 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, in iwch_sgl2pbl_map()
540 struct ib_sge sgl; in iwch_bind_mw()
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dmad.c465 struct ib_sge list; in mlx4_ib_send_to_slave()
1121 struct ib_sge sg_list; in mlx4_ib_post_pv_qp_buf()
1171 struct ib_sge list; in mlx4_ib_send_to_wire()
Dqp.c2572 static void set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ib_sge *sg) in set_data_seg()
2590 static void __set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ib_sge *sg) in __set_data_seg()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c119 struct ib_sge *sge = send_wr->sg_list; in trace_send_wr_ud()
/linux-4.1.27/drivers/infiniband/hw/amso1100/
Dc2_qp.c658 move_sgl(struct c2_data_addr * dst, struct ib_sge *src, int count, u32 * p_len, in move_sgl()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.c1687 struct ib_sge list; in srp_post_send()
1708 struct ib_sge list; in srp_post_recv()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1944 struct ib_sge *sg_list) in ocrdma_build_sges()
1959 static inline uint32_t ocrdma_sglist_len(struct ib_sge *sg_list, int num_sge) in ocrdma_sglist_len()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dqp.c404 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl()