sges              200 drivers/infiniband/core/rw.c 	ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL);
sges              201 drivers/infiniband/core/rw.c 	if (!ctx->map.sges)
sges              240 drivers/infiniband/core/rw.c 	kfree(ctx->map.sges);
sges              590 drivers/infiniband/core/rw.c 		kfree(ctx->map.sges);
sges              593 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 			  struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges,
sges              603 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		sq_wqe->buf_descs[i].sge = sges[i];
sges              775 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c void hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges,
sges              781 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		sges[i] = sq_wqe->buf_descs[i].sge;
sges              782 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 		hinic_be32_to_cpu(&sges[i], sizeof(sges[i]));
sges              170 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 			  struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
sges              195 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges,
sges              134 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		      struct hinic_sge *sges)
sges              150 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb));
sges              163 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag));
sges              170 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]),
sges              171 drivers/net/ethernet/huawei/hinic/hinic_tx.c 			       sges[j + 1].len, DMA_TO_DEVICE);
sges              173 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[0].len,
sges              185 drivers/net/ethernet/huawei/hinic/hinic_tx.c 			 struct hinic_sge *sges)
sges              193 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[i + 1]),
sges              194 drivers/net/ethernet/huawei/hinic/hinic_tx.c 			       sges[i + 1].len, DMA_TO_DEVICE);
sges              196 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[0].len,
sges              497 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	err = tx_map_skb(nic_dev, skb, txq->sges);
sges              516 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		tx_unmap_skb(nic_dev, skb, txq->sges);
sges              527 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges);
sges              544 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_unmap_skb(nic_dev, skb, txq->sges);
sges              564 drivers/net/ethernet/huawei/hinic/hinic_tx.c 			struct hinic_sge *sges)
sges              566 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_unmap_skb(nic_dev, skb, sges);
sges              773 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	sges_size = txq->max_sges * sizeof(*txq->sges);
sges              774 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL);
sges              775 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	if (!txq->sges)
sges              815 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->sges);
sges              833 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->sges);
sges               36 drivers/net/ethernet/huawei/hinic/hinic_tx.h 	struct hinic_sge        *sges;
sges             2482 drivers/scsi/bfa/bfa_fcpim.c 	sge = &m->sges[0];
sges             2500 drivers/scsi/bfa/bfa_fcpim.c 				sgpge = sgpg->sgpg->sges;
sges              331 drivers/scsi/bfa/bfa_svc.h 	struct bfa_sge_s sges[BFI_SGE_INLINE_MAX];
sges              143 drivers/scsi/bfa/bfi.h 	struct bfi_sge_s sges[BFI_SGPG_SGES_MAX];
sges              677 drivers/scsi/bfa/bfi_ms.h 	struct bfi_sge_s	sges[BFI_SGE_INLINE_MAX];
sges               30 include/rdma/rw.h 			struct ib_sge		*sges;
sges               55 net/sunrpc/xprtrdma/svc_rdma_rw.c svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges)
sges               75 net/sunrpc/xprtrdma/svc_rdma_rw.c 	if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges,