sges 200 drivers/infiniband/core/rw.c ctx->map.sges = sge = kcalloc(sg_cnt, sizeof(*sge), GFP_KERNEL); sges 201 drivers/infiniband/core/rw.c if (!ctx->map.sges) sges 240 drivers/infiniband/core/rw.c kfree(ctx->map.sges); sges 590 drivers/infiniband/core/rw.c kfree(ctx->map.sges); sges 593 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, sges 603 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sq_wqe->buf_descs[i].sge = sges[i]; sges 775 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c void hinic_sq_get_sges(struct hinic_sq_wqe *sq_wqe, struct hinic_sge *sges, sges 781 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sges[i] = sq_wqe->buf_descs[i].sge; sges 782 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c hinic_be32_to_cpu(&sges[i], sizeof(sges[i])); sges 170 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h struct hinic_sq_wqe *wqe, struct hinic_sge *sges, sges 195 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h void hinic_sq_get_sges(struct hinic_sq_wqe *wqe, struct hinic_sge *sges, sges 134 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sge *sges) sges 150 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb)); sges 163 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag)); sges 170 drivers/net/ethernet/huawei/hinic/hinic_tx.c dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[j + 1]), sges 171 drivers/net/ethernet/huawei/hinic/hinic_tx.c sges[j + 1].len, DMA_TO_DEVICE); sges 173 drivers/net/ethernet/huawei/hinic/hinic_tx.c dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[0].len, sges 185 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sge *sges) sges 193 drivers/net/ethernet/huawei/hinic/hinic_tx.c dma_unmap_page(&pdev->dev, hinic_sge_to_dma(&sges[i + 1]), sges 194 drivers/net/ethernet/huawei/hinic/hinic_tx.c sges[i + 1].len, DMA_TO_DEVICE); sges 196 drivers/net/ethernet/huawei/hinic/hinic_tx.c dma_unmap_single(&pdev->dev, hinic_sge_to_dma(&sges[0]), sges[0].len, sges 497 drivers/net/ethernet/huawei/hinic/hinic_tx.c err = tx_map_skb(nic_dev, skb, txq->sges); sges 516 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_unmap_skb(nic_dev, skb, txq->sges); sges 527 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges); sges 544 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_unmap_skb(nic_dev, skb, txq->sges); sges 564 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sge *sges) sges 566 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_unmap_skb(nic_dev, skb, sges); sges 773 drivers/net/ethernet/huawei/hinic/hinic_tx.c sges_size = txq->max_sges * sizeof(*txq->sges); sges 774 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL); sges 775 drivers/net/ethernet/huawei/hinic/hinic_tx.c if (!txq->sges) sges 815 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->sges); sges 833 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->sges); sges 36 drivers/net/ethernet/huawei/hinic/hinic_tx.h struct hinic_sge *sges; sges 2482 drivers/scsi/bfa/bfa_fcpim.c sge = &m->sges[0]; sges 2500 drivers/scsi/bfa/bfa_fcpim.c sgpge = sgpg->sgpg->sges; sges 331 drivers/scsi/bfa/bfa_svc.h struct bfa_sge_s sges[BFI_SGE_INLINE_MAX]; sges 143 drivers/scsi/bfa/bfi.h struct bfi_sge_s sges[BFI_SGPG_SGES_MAX]; sges 677 drivers/scsi/bfa/bfi_ms.h struct bfi_sge_s sges[BFI_SGE_INLINE_MAX]; sges 30 include/rdma/rw.h struct ib_sge *sges; sges 55 net/sunrpc/xprtrdma/svc_rdma_rw.c svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) sges 75 net/sunrpc/xprtrdma/svc_rdma_rw.c if (sg_alloc_table_chained(&ctxt->rw_sg_table, sges,