sgl 117 arch/arm/mm/dma-mapping-nommu.c static int arm_nommu_dma_map_sg(struct device *dev, struct scatterlist *sgl, sgl 124 arch/arm/mm/dma-mapping-nommu.c for_each_sg(sgl, sg, nents, i) { sgl 133 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_unmap_sg(struct device *dev, struct scatterlist *sgl, sgl 140 arch/arm/mm/dma-mapping-nommu.c for_each_sg(sgl, sg, nents, i) sgl 156 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, sgl 162 arch/arm/mm/dma-mapping-nommu.c for_each_sg(sgl, sg, nents, i) sgl 166 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, sgl 172 arch/arm/mm/dma-mapping-nommu.c for_each_sg(sgl, sg, nents, i) sgl 883 arch/arm/mm/dma-mapping.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgl 653 arch/mips/jazz/jazzdma.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 658 arch/mips/jazz/jazzdma.c for_each_sg(sgl, sg, nents, i) sgl 663 arch/mips/jazz/jazzdma.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 668 arch/mips/jazz/jazzdma.c for_each_sg(sgl, sg, nents, i) sgl 184 arch/powerpc/kernel/dma-iommu.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 187 arch/powerpc/kernel/dma-iommu.c dma_direct_sync_sg_for_cpu(dev, sgl, nents, dir); sgl 191 arch/powerpc/kernel/dma-iommu.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 194 arch/powerpc/kernel/dma-iommu.c dma_direct_sync_sg_for_device(dev, sgl, nents, dir); sgl 632 arch/powerpc/platforms/ps3/system-bus.c static int ps3_sb_map_sg(struct device *_dev, struct scatterlist *sgl, sgl 643 arch/powerpc/platforms/ps3/system-bus.c for_each_sg(sgl, sg, nents, i) { sgl 106 arch/powerpc/platforms/pseries/ibmebus.c struct scatterlist *sgl, sgl 113 arch/powerpc/platforms/pseries/ibmebus.c for_each_sg(sgl, sg, nents, i) { sgl 554 arch/powerpc/platforms/pseries/vio.c struct scatterlist *sgl; sgl 558 arch/powerpc/platforms/pseries/vio.c for_each_sg(sglist, sgl, nelems, count) sgl 559 arch/powerpc/platforms/pseries/vio.c alloc_size += roundup(sgl->length, IOMMU_PAGE_SIZE(tbl)); sgl 568 arch/powerpc/platforms/pseries/vio.c for_each_sg(sglist, sgl, ret, count) sgl 569 arch/powerpc/platforms/pseries/vio.c alloc_size -= roundup(sgl->dma_length, IOMMU_PAGE_SIZE(tbl)); sgl 588 arch/powerpc/platforms/pseries/vio.c struct scatterlist *sgl; sgl 592 arch/powerpc/platforms/pseries/vio.c for_each_sg(sglist, sgl, nelems, count) sgl 593 arch/powerpc/platforms/pseries/vio.c alloc_size += roundup(sgl->dma_length, IOMMU_PAGE_SIZE(tbl)); sgl 161 arch/sparc/mm/io-unit.c static int iounit_map_sg(struct device *dev, struct scatterlist *sgl, int nents, sgl 171 arch/sparc/mm/io-unit.c for_each_sg(sgl, sg, nents, i) { sgl 194 arch/sparc/mm/io-unit.c static void iounit_unmap_sg(struct device *dev, struct scatterlist *sgl, sgl 203 arch/sparc/mm/io-unit.c for_each_sg(sgl, sg, nents, i) { sgl 243 arch/sparc/mm/iommu.c static int __sbus_iommu_map_sg(struct device *dev, struct scatterlist *sgl, sgl 250 arch/sparc/mm/iommu.c for_each_sg(sgl, sg, nents, j) { sgl 261 arch/sparc/mm/iommu.c static int sbus_iommu_map_sg_gflush(struct device *dev, struct scatterlist *sgl, sgl 265 arch/sparc/mm/iommu.c return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, false); sgl 268 arch/sparc/mm/iommu.c static int sbus_iommu_map_sg_pflush(struct device *dev, struct scatterlist *sgl, sgl 271 arch/sparc/mm/iommu.c return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, true); sgl 293 arch/sparc/mm/iommu.c static void sbus_iommu_unmap_sg(struct device *dev, struct scatterlist *sgl, sgl 299 arch/sparc/mm/iommu.c for_each_sg(sgl, sg, nents, i) { sgl 256 crypto/adiantum.c struct scatterlist *sgl, le128 *digest) sgl 273 crypto/adiantum.c sg_miter_start(&miter, sgl, sg_nents(sgl), sgl 393 crypto/af_alg.c int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len) sgl 399 crypto/af_alg.c n = iov_iter_get_pages(iter, sgl->pages, len, ALG_MAX_PAGES, &off); sgl 407 crypto/af_alg.c sg_init_table(sgl->sg, npages + 1); sgl 412 crypto/af_alg.c sg_set_page(sgl->sg + i, sgl->pages[i], plen, off); sgl 417 crypto/af_alg.c sg_mark_end(sgl->sg + npages - 1); sgl 418 crypto/af_alg.c sgl->npages = npages; sgl 431 crypto/af_alg.c void af_alg_free_sg(struct af_alg_sgl *sgl) sgl 435 crypto/af_alg.c for (i = 0; i < sgl->npages; i++) sgl 436 crypto/af_alg.c put_page(sgl->pages[i]); sgl 490 crypto/af_alg.c struct af_alg_tsgl *sgl; sgl 493 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list); sgl 495 crypto/af_alg.c sg = sgl->sg; sgl 497 crypto/af_alg.c if (!sg || sgl->cur >= MAX_SGL_ENTS) { sgl 498 crypto/af_alg.c sgl = sock_kmalloc(sk, sgl 499 crypto/af_alg.c struct_size(sgl, sg, (MAX_SGL_ENTS + 1)), sgl 501 crypto/af_alg.c if (!sgl) sgl 504 crypto/af_alg.c sg_init_table(sgl->sg, MAX_SGL_ENTS + 1); sgl 505 crypto/af_alg.c sgl->cur = 0; sgl 508 crypto/af_alg.c sg_chain(sg, MAX_SGL_ENTS + 1, sgl->sg); sgl 510 crypto/af_alg.c list_add_tail(&sgl->list, &ctx->tsgl_list); sgl 531 crypto/af_alg.c const struct af_alg_tsgl *sgl; sgl 538 crypto/af_alg.c list_for_each_entry(sgl, &ctx->tsgl_list, list) { sgl 539 crypto/af_alg.c const struct scatterlist *sg = sgl->sg; sgl 541 crypto/af_alg.c for (i = 0; i < sgl->cur; i++) { sgl 588 crypto/af_alg.c struct af_alg_tsgl *sgl; sgl 593 crypto/af_alg.c sgl = list_first_entry(&ctx->tsgl_list, struct af_alg_tsgl, sgl 595 crypto/af_alg.c sg = sgl->sg; sgl 597 crypto/af_alg.c for (i = 0; i < sgl->cur; i++) { sgl 636 crypto/af_alg.c list_del(&sgl->list); sgl 637 crypto/af_alg.c sock_kfree_s(sk, sgl, struct_size(sgl, sg, MAX_SGL_ENTS + 1)); sgl 662 crypto/af_alg.c af_alg_free_sg(&rsgl->sgl); sgl 821 crypto/af_alg.c struct af_alg_tsgl *sgl; sgl 870 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, sgl 872 crypto/af_alg.c sg = sgl->sg + sgl->cur - 1; sgl 905 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, sgl 907 crypto/af_alg.c sg = sgl->sg; sgl 908 crypto/af_alg.c if (sgl->cur) sgl 909 crypto/af_alg.c sg_unmark_end(sg + sgl->cur - 1); sgl 912 crypto/af_alg.c unsigned int i = sgl->cur; sgl 935 crypto/af_alg.c sgl->cur++; sgl 936 crypto/af_alg.c } while (len && sgl->cur < MAX_SGL_ENTS); sgl 939 crypto/af_alg.c sg_mark_end(sg + sgl->cur - 1); sgl 967 crypto/af_alg.c struct af_alg_tsgl *sgl; sgl 991 crypto/af_alg.c sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list); sgl 993 crypto/af_alg.c if (sgl->cur) sgl 994 crypto/af_alg.c sg_unmark_end(sgl->sg + sgl->cur - 1); sgl 996 crypto/af_alg.c sg_mark_end(sgl->sg + sgl->cur); sgl 999 crypto/af_alg.c sg_set_page(sgl->sg + sgl->cur, page, size, offset); sgl 1000 crypto/af_alg.c sgl->cur++; sgl 1142 crypto/af_alg.c rsgl->sgl.npages = 0; sgl 1146 crypto/af_alg.c err = af_alg_make_sg(&rsgl->sgl, &msg->msg_iter, seglen); sgl 1154 crypto/af_alg.c af_alg_link_sg(&areq->last_rsgl->sgl, &rsgl->sgl); sgl 213 crypto/algif_aead.c rsgl_src = areq->first_rsgl.sgl.sg; sgl 227 crypto/algif_aead.c areq->first_rsgl.sgl.sg, processed); sgl 245 crypto/algif_aead.c areq->first_rsgl.sgl.sg, outlen); sgl 269 crypto/algif_aead.c struct af_alg_sgl *sgl_prev = &areq->last_rsgl->sgl; sgl 281 crypto/algif_aead.c areq->first_rsgl.sgl.sg, used, ctx->iv); sgl 20 crypto/algif_hash.c struct af_alg_sgl sgl; sgl 94 crypto/algif_hash.c len = af_alg_make_sg(&ctx->sgl, &msg->msg_iter, len); sgl 100 crypto/algif_hash.c ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, NULL, len); sgl 104 crypto/algif_hash.c af_alg_free_sg(&ctx->sgl); sgl 143 crypto/algif_hash.c sg_init_table(ctx->sgl.sg, 1); sgl 144 crypto/algif_hash.c sg_set_page(ctx->sgl.sg, page, size, offset); sgl 153 crypto/algif_hash.c ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, ctx->result, size); sgl 112 crypto/algif_skcipher.c areq->first_rsgl.sgl.sg, len, ctx->iv); sgl 493 crypto/testmgr.c struct scatterlist sgl[XBUFSIZE]; sgl 542 crypto/testmgr.c BUILD_BUG_ON(ARRAY_SIZE(partitions) != ARRAY_SIZE(tsgl->sgl)); sgl 569 crypto/testmgr.c sg_init_table(tsgl->sgl, tsgl->nents); sgl 585 crypto/testmgr.c sg_set_buf(&tsgl->sgl[i], addr, partitions[i].length); sgl 605 crypto/testmgr.c sg_mark_end(&tsgl->sgl[tsgl->nents - 1]); sgl 606 crypto/testmgr.c tsgl->sgl_ptr = tsgl->sgl; sgl 607 crypto/testmgr.c memcpy(tsgl->sgl_saved, tsgl->sgl, tsgl->nents * sizeof(tsgl->sgl[0])); sgl 665 crypto/testmgr.c if (tsgl->sgl[i].page_link != tsgl->sgl_saved[i].page_link) sgl 667 crypto/testmgr.c if (tsgl->sgl[i].offset != tsgl->sgl_saved[i].offset) sgl 669 crypto/testmgr.c if (tsgl->sgl[i].length != tsgl->sgl_saved[i].length) sgl 733 crypto/testmgr.c tsgls->dst.sgl_ptr = tsgls->src.sgl; sgl 1142 crypto/testmgr.c err = crypto_shash_digest(desc, sg_data(&tsgl->sgl[0]), sgl 1143 crypto/testmgr.c tsgl->sgl[0].length, result); sgl 1178 crypto/testmgr.c err = crypto_shash_finup(desc, sg_data(&tsgl->sgl[i]), sgl 1179 crypto/testmgr.c tsgl->sgl[i].length, result); sgl 1190 crypto/testmgr.c err = crypto_shash_update(desc, sg_data(&tsgl->sgl[i]), sgl 1191 crypto/testmgr.c tsgl->sgl[i].length); sgl 1327 crypto/testmgr.c ahash_request_set_crypt(req, tsgl->sgl, result, vec->psize); sgl 1401 crypto/testmgr.c pending_sgl = &tsgl->sgl[i]; sgl 1402 crypto/testmgr.c pending_len += tsgl->sgl[i].length; sgl 1968 crypto/testmgr.c if (tsgls->dst.sgl_ptr != tsgls->src.sgl && sgl 2554 crypto/testmgr.c if (tsgls->dst.sgl_ptr != tsgls->src.sgl && sgl 1549 drivers/ata/libata-core.c int dma_dir, struct scatterlist *sgl, sgl 1609 drivers/ata/libata-core.c for_each_sg(sgl, sg, n_elem, i) sgl 1612 drivers/ata/libata-core.c ata_sg_init(qc, sgl, n_elem); sgl 623 drivers/block/skd_main.c struct scatterlist *sgl = &skreq->sg[0], *sg; sgl 632 drivers/block/skd_main.c n_sg = blk_rq_map_sg(skdev->queue, req, sgl); sgl 640 drivers/block/skd_main.c n_sg = dma_map_sg(&skdev->pdev->dev, sgl, n_sg, skreq->data_dir); sgl 648 drivers/block/skd_main.c for_each_sg(sgl, sg, n_sg, i) { sgl 110 drivers/char/agp/intel-gtt.c for_each_sg(st->sgl, sg, num_entries, i) sgl 114 drivers/char/agp/intel-gtt.c st->sgl, st->nents, PCI_DMA_BIDIRECTIONAL)) sgl 132 drivers/char/agp/intel-gtt.c st.sgl = sg_list; sgl 867 drivers/char/agp/intel-gtt.c for_each_sg(st->sgl, sg, st->nents, i) { sgl 935 drivers/char/agp/intel-gtt.c mem->sg_list = st.sgl; sgl 867 drivers/char/virtio_console.c struct sg_list *sgl = sd->u.data; sgl 870 drivers/char/virtio_console.c if (sgl->n == sgl->size) sgl 880 drivers/char/virtio_console.c sg_set_page(&(sgl->sg[sgl->n]), buf->page, len, buf->offset); sgl 899 drivers/char/virtio_console.c sg_set_page(&(sgl->sg[sgl->n]), page, len, offset); sgl 901 drivers/char/virtio_console.c sgl->n++; sgl 902 drivers/char/virtio_console.c sgl->len += len; sgl 913 drivers/char/virtio_console.c struct sg_list sgl; sgl 920 drivers/char/virtio_console.c .u.data = &sgl, sgl 952 drivers/char/virtio_console.c sgl.n = 0; sgl 953 drivers/char/virtio_console.c sgl.len = 0; sgl 954 drivers/char/virtio_console.c sgl.size = pipe->nrbufs; sgl 955 drivers/char/virtio_console.c sgl.sg = buf->sg; sgl 956 drivers/char/virtio_console.c sg_init_table(sgl.sg, sgl.size); sgl 960 drivers/char/virtio_console.c ret = __send_to_port(port, buf->sg, sgl.n, sgl.len, buf, true); sgl 199 drivers/crypto/caam/caampkc.c static int caam_rsa_count_leading_zeros(struct scatterlist *sgl, sgl 209 drivers/crypto/caam/caampkc.c ents = sg_nents_for_len(sgl, nbytes); sgl 213 drivers/crypto/caam/caampkc.c sg_miter_start(&miter, sgl, ents, SG_MITER_FROM_SG | flags); sgl 148 drivers/crypto/ccp/ccp-crypto-aes-cmac.c sg = rctx->data_sg.sgl; sgl 311 drivers/crypto/ccp/ccp-crypto-main.c for (sg = table->sgl; sg; sg = sg_next(sg)) sgl 120 drivers/crypto/ccp/ccp-crypto-sha.c sg = rctx->data_sg.sgl; sgl 27 drivers/crypto/ccree/cc_buffer_mgr.c struct scatterlist *sgl; sgl 164 drivers/crypto/ccree/cc_buffer_mgr.c static int cc_render_sg_to_mlli(struct device *dev, struct scatterlist *sgl, sgl 168 drivers/crypto/ccree/cc_buffer_mgr.c struct scatterlist *curr_sgl = sgl; sgl 218 drivers/crypto/ccree/cc_buffer_mgr.c rc = cc_render_sg_to_mlli(dev, entry->sgl, tot_len, sgl 272 drivers/crypto/ccree/cc_buffer_mgr.c unsigned int nents, struct scatterlist *sgl, sgl 279 drivers/crypto/ccree/cc_buffer_mgr.c index, nents, sgl, data_len, is_last_table); sgl 281 drivers/crypto/ccree/cc_buffer_mgr.c sgl_data->entry[index].sgl = sgl; sgl 854 drivers/crypto/ccree/cc_buffer_mgr.c struct scatterlist *sgl; sgl 879 drivers/crypto/ccree/cc_buffer_mgr.c sgl = sg_next(areq_ctx->src_sgl); sgl 880 drivers/crypto/ccree/cc_buffer_mgr.c if (!sgl) sgl 882 drivers/crypto/ccree/cc_buffer_mgr.c areq_ctx->src_sgl = sgl; sgl 923 drivers/crypto/ccree/cc_buffer_mgr.c sgl = sg_next(areq_ctx->dst_sgl); sgl 924 drivers/crypto/ccree/cc_buffer_mgr.c if (!sgl) sgl 926 drivers/crypto/ccree/cc_buffer_mgr.c areq_ctx->dst_sgl = sgl; sgl 477 drivers/crypto/chelsio/chcr_algo.c walk->sgl = ulp; sgl 487 drivers/crypto/chelsio/chcr_algo.c walk->sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | sgl 500 drivers/crypto/chelsio/chcr_algo.c walk->sgl->len0 = cpu_to_be32(size); sgl 501 drivers/crypto/chelsio/chcr_algo.c walk->sgl->addr0 = cpu_to_be64(addr); sgl 537 drivers/crypto/chelsio/chcr_algo.c walk->sgl->len0 = cpu_to_be32(sgmin); sgl 538 drivers/crypto/chelsio/chcr_algo.c walk->sgl->addr0 = cpu_to_be64(sg_dma_address(sg) + skip_len); sgl 195 drivers/crypto/chelsio/chcr_crypto.h struct ulptx_sgl *sgl; sgl 677 drivers/crypto/chelsio/chcr_ipsec.c u64 *pos, *end, *before, *sgl; sgl 740 drivers/crypto/chelsio/chcr_ipsec.c sgl = (void *)pos; sgl 742 drivers/crypto/chelsio/chcr_ipsec.c cxgb4_inline_tx_skb(skb, &q->q, sgl); sgl 747 drivers/crypto/chelsio/chcr_ipsec.c cxgb4_write_sgl(skb, &q->q, (void *)sgl, end, sgl 755 drivers/crypto/chelsio/chcr_ipsec.c q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)sgl; sgl 176 drivers/crypto/hisilicon/sec/sec_algs.c struct scatterlist *sgl, sgl 189 drivers/crypto/hisilicon/sec/sec_algs.c for_each_sg(sgl, sg, count, i) { sgl 567 drivers/crypto/hisilicon/sec/sec_algs.c static int sec_map_and_split_sg(struct scatterlist *sgl, size_t *split_sizes, sgl 575 drivers/crypto/hisilicon/sec/sec_algs.c count = dma_map_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); sgl 591 drivers/crypto/hisilicon/sec/sec_algs.c ret = sg_split(sgl, count, 0, steps, split_sizes, sgl 605 drivers/crypto/hisilicon/sec/sec_algs.c dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); sgl 614 drivers/crypto/hisilicon/sec/sec_algs.c static void sec_unmap_sg_on_err(struct scatterlist *sgl, int steps, sgl 625 drivers/crypto/hisilicon/sec/sec_algs.c dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); sgl 80 drivers/crypto/hisilicon/sgl.c pool->sgl = dma_alloc_coherent(dev, size, &pool->sgl_dma, GFP_KERNEL); sgl 81 drivers/crypto/hisilicon/sgl.c if (!pool->sgl) sgl 101 drivers/crypto/hisilicon/sgl.c dma_free_coherent(dev, pool->size, pool->sgl, pool->sgl_dma); sgl 109 drivers/crypto/hisilicon/sgl.c if (!pool || !hw_sgl_dma || index >= pool->count || !pool->sgl) sgl 113 drivers/crypto/hisilicon/sgl.c return (void *)pool->sgl + pool->sgl_size * index; sgl 118 drivers/crypto/hisilicon/sgl.c static void sg_map_to_hw_sg(struct scatterlist *sgl, sgl 121 drivers/crypto/hisilicon/sgl.c hw_sge->buf = sgl->dma_address; sgl 122 drivers/crypto/hisilicon/sgl.c hw_sge->len = sgl->dma_length; sgl 148 drivers/crypto/hisilicon/sgl.c struct scatterlist *sgl, sgl 156 drivers/crypto/hisilicon/sgl.c int sg_n = sg_nents(sgl); sgl 159 drivers/crypto/hisilicon/sgl.c if (!dev || !sgl || !pool || !hw_sgl_dma || sg_n > acc_sgl_sge_nr) sgl 162 drivers/crypto/hisilicon/sgl.c ret = dma_map_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL); sgl 174 drivers/crypto/hisilicon/sgl.c for_each_sg(sgl, sg, sg_n, i) { sgl 186 drivers/crypto/hisilicon/sgl.c dma_unmap_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL); sgl 201 drivers/crypto/hisilicon/sgl.c void hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl, sgl 204 drivers/crypto/hisilicon/sgl.c dma_unmap_sg(dev, sgl, sg_nents(sgl), DMA_BIDIRECTIONAL); sgl 7 drivers/crypto/hisilicon/sgl.h struct hisi_acc_hw_sgl *sgl; sgl 16 drivers/crypto/hisilicon/sgl.h struct scatterlist *sgl, sgl 19 drivers/crypto/hisilicon/sgl.h void hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl, sgl 404 drivers/crypto/hisilicon/zip/zip_crypto.c static size_t get_gzip_head_size(struct scatterlist *sgl) sgl 408 drivers/crypto/hisilicon/zip/zip_crypto.c sg_copy_to_buffer(sgl, sg_nents(sgl), buf, sizeof(buf)); sgl 153 drivers/crypto/omap-sham.c struct scatterlist sgl[2]; sgl 719 drivers/crypto/omap-sham.c sg_init_table(ctx->sgl, 1); sgl 720 drivers/crypto/omap-sham.c sg_set_buf(ctx->sgl, buf, len); sgl 721 drivers/crypto/omap-sham.c ctx->sg = ctx->sgl; sgl 865 drivers/crypto/omap-sham.c sg_init_table(rctx->sgl, 2); sgl 866 drivers/crypto/omap-sham.c sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, rctx->bufcnt); sgl 868 drivers/crypto/omap-sham.c sg_chain(rctx->sgl, 2, req->src); sgl 870 drivers/crypto/omap-sham.c rctx->sg = rctx->sgl; sgl 875 drivers/crypto/omap-sham.c sg_init_table(rctx->sgl, 1); sgl 876 drivers/crypto/omap-sham.c sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, xmit_len); sgl 878 drivers/crypto/omap-sham.c rctx->sg = rctx->sgl; sgl 708 drivers/crypto/qat/qat_common/qat_algs.c struct scatterlist *sgl, sgl 714 drivers/crypto/qat/qat_common/qat_algs.c int n = sg_nents(sgl); sgl 734 drivers/crypto/qat/qat_common/qat_algs.c for_each_sg(sgl, sg, n, i) { sgl 753 drivers/crypto/qat/qat_common/qat_algs.c if (sgl != sglout) { sgl 805 drivers/crypto/qat/qat_common/qat_algs.c n = sg_nents(sgl); sgl 111 drivers/crypto/qce/ablkcipher.c rctx->dst_sg = rctx->dst_tbl.sgl; sgl 52 drivers/crypto/qce/dma.c struct scatterlist *sg = sgt->sgl, *sg_last = NULL; sgl 365 drivers/crypto/s5p-sss.c struct scatterlist sgl[2]; sgl 1045 drivers/crypto/s5p-sss.c sg_init_table(ctx->sgl, 1); sgl 1046 drivers/crypto/s5p-sss.c sg_set_buf(ctx->sgl, buf, len); sgl 1047 drivers/crypto/s5p-sss.c ctx->sg = ctx->sgl; sgl 1191 drivers/crypto/s5p-sss.c sg_init_table(ctx->sgl, 2); sgl 1192 drivers/crypto/s5p-sss.c sg_set_buf(ctx->sgl, ctx->dd->xmit_buf, ctx->bufcnt); sgl 1193 drivers/crypto/s5p-sss.c sg_chain(ctx->sgl, 2, sg); sgl 1194 drivers/crypto/s5p-sss.c ctx->sg = ctx->sgl; sgl 1279 drivers/crypto/s5p-sss.c sg_init_table(ctx->sgl, 1); sgl 1280 drivers/crypto/s5p-sss.c sg_set_buf(ctx->sgl, ctx->dd->xmit_buf, xmit_len); sgl 1282 drivers/crypto/s5p-sss.c ctx->sg = ctx->sgl; sgl 64 drivers/dma-buf/udmabuf.c if (!dma_map_sg(at->dev, sg->sgl, sg->nents, direction)) { sgl 80 drivers/dma-buf/udmabuf.c dma_unmap_sg(at->dev, sg->sgl, sg->nents, direction); sgl 385 drivers/dma/altera-msgdma.c msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, sgl 401 drivers/dma/altera-msgdma.c for_each_sg(sgl, sg, sg_len, i) sgl 413 drivers/dma/altera-msgdma.c avail = sg_dma_len(sgl); sgl 424 drivers/dma/altera-msgdma.c dma_src = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; sgl 429 drivers/dma/altera-msgdma.c dma_dst = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; sgl 444 drivers/dma/altera-msgdma.c sgl = sg_next(sgl); sgl 445 drivers/dma/altera-msgdma.c if (sgl == NULL) sgl 448 drivers/dma/altera-msgdma.c avail = sg_dma_len(sgl); sgl 2050 drivers/dma/amba-pl08x.c struct dma_chan *chan, struct scatterlist *sgl, sgl 2062 drivers/dma/amba-pl08x.c __func__, sg_dma_len(sgl), plchan->name); sgl 2068 drivers/dma/amba-pl08x.c for_each_sg(sgl, sg, sg_len, tmp) { sgl 717 drivers/dma/at_hdmac.c first = xt->sgl; sgl 731 drivers/dma/at_hdmac.c struct data_chunk *chunk = xt->sgl + i; sgl 733 drivers/dma/at_hdmac.c if ((chunk->size != xt->sgl->size) || sgl 979 drivers/dma/at_hdmac.c struct scatterlist *sgl, sgl 995 drivers/dma/at_hdmac.c if (unlikely(!sgl || !sg_len)) { sgl 1009 drivers/dma/at_hdmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 1064 drivers/dma/at_hdmac.c atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 1105 drivers/dma/at_hdmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 1146 drivers/dma/at_hdmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 633 drivers/dma/at_xdmac.c at_xdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 645 drivers/dma/at_xdmac.c if (!sgl) sgl 665 drivers/dma/at_xdmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 971 drivers/dma/at_xdmac.c xt, xt->sgl); sgl 985 drivers/dma/at_xdmac.c chunk = xt->sgl + i; sgl 1228 drivers/dma/at_xdmac.c at_xdmac_prep_dma_memset_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 1239 drivers/dma/at_xdmac.c if (!sgl) sgl 1246 drivers/dma/at_xdmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 267 drivers/dma/bcm2835-dma.c struct scatterlist *sgl, sgl 275 drivers/dma/bcm2835-dma.c for_each_sg(sgl, sgent, sg_len, i) sgl 389 drivers/dma/bcm2835-dma.c struct scatterlist *sgl, sgl 399 drivers/dma/bcm2835-dma.c for_each_sg(sgl, sgent, sg_len, i) { sgl 644 drivers/dma/bcm2835-dma.c struct scatterlist *sgl, unsigned int sg_len, sgl 677 drivers/dma/bcm2835-dma.c frames = bcm2835_dma_count_frames_for_sg(c, sgl, sg_len); sgl 689 drivers/dma/bcm2835-dma.c sgl, sg_len); sgl 2268 drivers/dma/coh901318.c coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 2287 drivers/dma/coh901318.c if (!sgl) sgl 2289 drivers/dma/coh901318.c if (sg_dma_len(sgl) == 0) sgl 2335 drivers/dma/coh901318.c for_each_sg(sgl, sg, sg_len, i) { sgl 2360 drivers/dma/coh901318.c ret = coh901318_lli_fill_sg(&cohc->base->pool, lli, sgl, sg_len, sgl 233 drivers/dma/coh901318_lli.c struct scatterlist *sgl, unsigned int nents, sgl 258 drivers/dma/coh901318_lli.c for_each_sg(sgl, sg, nents, i) { sgl 527 drivers/dma/dma-axi-dmac.c struct dma_chan *c, struct scatterlist *sgl, sgl 542 drivers/dma/dma-axi-dmac.c for_each_sg(sgl, sg, sg_len, i) sgl 551 drivers/dma/dma-axi-dmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 625 drivers/dma/dma-axi-dmac.c dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]); sgl 626 drivers/dma/dma-axi-dmac.c src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]); sgl 629 drivers/dma/dma-axi-dmac.c if (!axi_dmac_check_len(chan, xt->sgl[0].size) || sgl 632 drivers/dma/dma-axi-dmac.c if (xt->sgl[0].size + dst_icg > chan->max_length || sgl 633 drivers/dma/dma-axi-dmac.c xt->sgl[0].size + src_icg > chan->max_length) sgl 638 drivers/dma/dma-axi-dmac.c if (chan->max_length / xt->sgl[0].size < xt->numf) sgl 640 drivers/dma/dma-axi-dmac.c if (!axi_dmac_check_len(chan, xt->sgl[0].size * xt->numf)) sgl 650 drivers/dma/dma-axi-dmac.c desc->sg[0].src_stride = xt->sgl[0].size + src_icg; sgl 655 drivers/dma/dma-axi-dmac.c desc->sg[0].dest_stride = xt->sgl[0].size + dst_icg; sgl 659 drivers/dma/dma-axi-dmac.c desc->sg[0].x_len = xt->sgl[0].size; sgl 662 drivers/dma/dma-axi-dmac.c desc->sg[0].x_len = xt->sgl[0].size * xt->numf; sgl 354 drivers/dma/dma-jz4780.c struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, sgl 370 drivers/dma/dma-jz4780.c sg_dma_address(&sgl[i]), sgl 371 drivers/dma/dma-jz4780.c sg_dma_len(&sgl[i]), sgl 364 drivers/dma/dw-edma/dw-edma-core.c sg = xfer->xfer.sg.sgl; sgl 435 drivers/dma/dw-edma/dw-edma-core.c dw_edma_device_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, sgl 444 drivers/dma/dw-edma/dw-edma-core.c xfer.xfer.sg.sgl = sgl; sgl 132 drivers/dma/dw-edma/dw-edma-core.h struct scatterlist *sgl; sgl 622 drivers/dma/dw/core.c dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 663 drivers/dma/dw/core.c for_each_sg(sgl, sg, sg_len, i) { sgl 713 drivers/dma/dw/core.c for_each_sg(sgl, sg, sg_len, i) { sgl 1033 drivers/dma/ep93xx_dma.c ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 1057 drivers/dma/ep93xx_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 512 drivers/dma/fsl-edma-common.c struct dma_chan *chan, struct scatterlist *sgl, sgl 547 drivers/dma/fsl-edma-common.c for_each_sg(sgl, sg, sg_len, i) { sgl 239 drivers/dma/fsl-edma-common.h struct dma_chan *chan, struct scatterlist *sgl, sgl 259 drivers/dma/hsu/hsu.c struct dma_chan *chan, struct scatterlist *sgl, sgl 272 drivers/dma/hsu/hsu.c for_each_sg(sgl, sg, sg_len, i) { sgl 293 drivers/dma/idma64.c struct dma_chan *chan, struct scatterlist *sgl, sgl 306 drivers/dma/idma64.c for_each_sg(sgl, sg, sg_len, i) { sgl 448 drivers/dma/img-mdc-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 460 drivers/dma/img-mdc-dma.c if (!sgl) sgl 474 drivers/dma/img-mdc-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 816 drivers/dma/imx-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 831 drivers/dma/imx-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 837 drivers/dma/imx-dma.c if (sg_dma_len(sgl) & 3 || sgl->dma_address & 3) sgl 841 drivers/dma/imx-dma.c if (sg_dma_len(sgl) & 1 || sgl->dma_address & 1) sgl 851 drivers/dma/imx-dma.c desc->sg = sgl; sgl 982 drivers/dma/imx-dma.c desc->x = xt->sgl[0].size; sgl 984 drivers/dma/imx-dma.c desc->w = xt->sgl[0].icg + desc->x; sgl 1441 drivers/dma/imx-sdma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 1461 drivers/dma/imx-sdma.c for_each_sg(sgl, sg, sg_len, i) { sgl 1334 drivers/dma/ipu/ipu_idmac.c struct scatterlist *sgl, unsigned int sg_len, sgl 1363 drivers/dma/ipu/ipu_idmac.c desc->sg = sgl; sgl 533 drivers/dma/k3dma.c struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen, sgl 543 drivers/dma/k3dma.c if (sgl == NULL) sgl 548 drivers/dma/k3dma.c for_each_sg(sgl, sg, sglen, i) { sgl 560 drivers/dma/k3dma.c for_each_sg(sgl, sg, sglen, i) { sgl 340 drivers/dma/mediatek/mtk-uart-apdma.c (struct dma_chan *chan, struct scatterlist *sgl, sgl 355 drivers/dma/mediatek/mtk-uart-apdma.c d->avail_len = sg_dma_len(sgl); sgl 356 drivers/dma/mediatek/mtk-uart-apdma.c d->addr = sg_dma_address(sgl); sgl 526 drivers/dma/mmp_pdma.c mmp_pdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, sgl 537 drivers/dma/mmp_pdma.c if ((sgl == NULL) || (sg_len == 0)) sgl 544 drivers/dma/mmp_pdma.c for_each_sg(sgl, sg, sg_len, i) { sgl 546 drivers/dma/mmp_pdma.c avail = sg_dma_len(sgl); sgl 270 drivers/dma/moxart-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 320 drivers/dma/moxart-dma.c for_each_sg(sgl, sgent, sg_len, i) { sgl 684 drivers/dma/mpc512x_dma.c mpc_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 706 drivers/dma/mpc512x_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 495 drivers/dma/mxs-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 537 drivers/dma/mxs-dma.c pio = (u32 *) sgl; sgl 554 drivers/dma/mxs-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 1007 drivers/dma/nbpfaxi.c struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, sgl 1020 drivers/dma/nbpfaxi.c return nbpf_prep_sg(chan, sgl, &slave_sg, sg_len, sgl 1025 drivers/dma/nbpfaxi.c return nbpf_prep_sg(chan, &slave_sg, sgl, sg_len, sgl 876 drivers/dma/owl-dma.c struct scatterlist *sgl, sgl 897 drivers/dma/owl-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 567 drivers/dma/pch_dma.c struct scatterlist *sgl, unsigned int sg_len, sgl 595 drivers/dma/pch_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 2829 drivers/dma/pl330.c pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 2838 drivers/dma/pl330.c if (unlikely(!pch || !sgl || !sg_len)) sgl 2848 drivers/dma/pl330.c for_each_sg(sgl, sg, sg_len, i) { sgl 971 drivers/dma/pxa_dma.c pxad_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, sgl 983 drivers/dma/pxa_dma.c if ((sgl == NULL) || (sg_len == 0)) sgl 990 drivers/dma/pxa_dma.c for_each_sg(sgl, sg, sg_len, i) sgl 996 drivers/dma/pxa_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 607 drivers/dma/qcom/bam_dma.c struct scatterlist *sgl, unsigned int sg_len, sgl 626 drivers/dma/qcom/bam_dma.c for_each_sg(sgl, sg, sg_len, i) sgl 648 drivers/dma/qcom/bam_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 975 drivers/dma/s3c24xx-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 991 drivers/dma/s3c24xx-dma.c sg_dma_len(sgl), s3cchan->name); sgl 1042 drivers/dma/s3c24xx-dma.c for_each_sg(sgl, sg, sg_len, tmp) { sgl 894 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_prep_sg(struct rcar_dmac_chan *chan, struct scatterlist *sgl, sgl 930 drivers/dma/sh/rcar-dmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 1097 drivers/dma/sh/rcar-dmac.c struct scatterlist sgl; sgl 1102 drivers/dma/sh/rcar-dmac.c sg_init_table(&sgl, 1); sgl 1103 drivers/dma/sh/rcar-dmac.c sg_set_page(&sgl, pfn_to_page(PFN_DOWN(dma_src)), len, sgl 1105 drivers/dma/sh/rcar-dmac.c sg_dma_address(&sgl) = dma_src; sgl 1106 drivers/dma/sh/rcar-dmac.c sg_dma_len(&sgl) = len; sgl 1108 drivers/dma/sh/rcar-dmac.c return rcar_dmac_chan_prep_sg(rchan, &sgl, 1, dma_dest, sgl 1166 drivers/dma/sh/rcar-dmac.c rcar_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 1173 drivers/dma/sh/rcar-dmac.c if (rchan->mid_rid < 0 || !sg_len || !sg_dma_len(sgl)) { sgl 1183 drivers/dma/sh/rcar-dmac.c return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr, sgl 1196 drivers/dma/sh/rcar-dmac.c struct scatterlist *sgl; sgl 1223 drivers/dma/sh/rcar-dmac.c sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT); sgl 1224 drivers/dma/sh/rcar-dmac.c if (!sgl) sgl 1227 drivers/dma/sh/rcar-dmac.c sg_init_table(sgl, sg_len); sgl 1232 drivers/dma/sh/rcar-dmac.c sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len, sgl 1234 drivers/dma/sh/rcar-dmac.c sg_dma_address(&sgl[i]) = src; sgl 1235 drivers/dma/sh/rcar-dmac.c sg_dma_len(&sgl[i]) = period_len; sgl 1238 drivers/dma/sh/rcar-dmac.c desc = rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr, sgl 1241 drivers/dma/sh/rcar-dmac.c kfree(sgl); sgl 561 drivers/dma/sh/shdma-base.c struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr, sgl 571 drivers/dma/sh/shdma-base.c for_each_sg(sgl, sg, sg_len, i) sgl 588 drivers/dma/sh/shdma-base.c for_each_sg(sgl, sg, sg_len, i) { sgl 662 drivers/dma/sh/shdma-base.c struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, sgl 685 drivers/dma/sh/shdma-base.c return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, sgl 703 drivers/dma/sh/shdma-base.c struct scatterlist *sgl; sgl 731 drivers/dma/sh/shdma-base.c sgl = kcalloc(sg_len, sizeof(*sgl), GFP_KERNEL); sgl 732 drivers/dma/sh/shdma-base.c if (!sgl) sgl 735 drivers/dma/sh/shdma-base.c sg_init_table(sgl, sg_len); sgl 740 drivers/dma/sh/shdma-base.c sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len, sgl 742 drivers/dma/sh/shdma-base.c sg_dma_address(&sgl[i]) = src; sgl 743 drivers/dma/sh/shdma-base.c sg_dma_len(&sgl[i]) = period_len; sgl 746 drivers/dma/sh/shdma-base.c desc = shdma_prep_sg(schan, sgl, sg_len, &slave_addr, sgl 749 drivers/dma/sh/shdma-base.c kfree(sgl); sgl 416 drivers/dma/sh/usb-dmac.c usb_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 437 drivers/dma/sh/usb-dmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 728 drivers/dma/sirf-dma.c sdesc->xlen = xt->sgl[0].size / SIRFSOC_DMA_WORD_LEN; sgl 729 drivers/dma/sirf-dma.c sdesc->width = (xt->sgl[0].size + xt->sgl[0].icg) / sgl 920 drivers/dma/sprd-dma.c sprd_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 963 drivers/dma/sprd-dma.c for_each_sg(sgl, sg, sglen, i) { sgl 506 drivers/dma/st_fdma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 520 drivers/dma/st_fdma.c if (!sgl) sgl 531 drivers/dma/st_fdma.c for_each_sg(sgl, sg, sg_len, i) { sgl 1020 drivers/dma/ste_dma40.c static int d40_sg_2_dmalen(struct scatterlist *sgl, int sg_len, sgl 1028 drivers/dma/ste_dma40.c for_each_sg(sgl, sg, sg_len, i) { sgl 2498 drivers/dma/ste_dma40.c d40_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 2505 drivers/dma/ste_dma40.c return d40_prep_sg(chan, sgl, sgl, sg_len, direction, dma_flags); sgl 829 drivers/dma/stm32-dma.c struct dma_chan *c, struct scatterlist *sgl, sgl 860 drivers/dma/stm32-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 728 drivers/dma/stm32-mdma.c struct scatterlist *sgl, u32 sg_len, sgl 738 drivers/dma/stm32-mdma.c for_each_sg(sgl, sg, sg_len, i) { sgl 781 drivers/dma/stm32-mdma.c stm32_mdma_prep_slave_sg(struct dma_chan *c, struct scatterlist *sgl, sgl 804 drivers/dma/stm32-mdma.c ret = stm32_mdma_setup_xfer(chan, desc, sgl, sg_len, direction); sgl 767 drivers/dma/sun4i-dma.c sun4i_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 781 drivers/dma/sun4i-dma.c if (!sgl) sgl 814 drivers/dma/sun4i-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 683 drivers/dma/sun6i-dma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 697 drivers/dma/sun6i-dma.c if (!sgl) sgl 710 drivers/dma/sun6i-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 1001 drivers/dma/tegra20-apb-dma.c struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len, sgl 1064 drivers/dma/tegra20-apb-dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 585 drivers/dma/ti/cppi41.c struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len, sgl 607 drivers/dma/ti/cppi41.c for_each_sg(sgl, sg, sg_len, i) { sgl 1054 drivers/dma/ti/edma.c struct dma_chan *chan, struct scatterlist *sgl, sgl 1067 drivers/dma/ti/edma.c if (unlikely(!echan || !sgl || !sg_len)) sgl 1114 drivers/dma/ti/edma.c for_each_sg(sgl, sg, sg_len, i) { sgl 886 drivers/dma/ti/omap-dma.c struct dma_chan *chan, struct scatterlist *sgl, unsigned sglen, sgl 1027 drivers/dma/ti/omap-dma.c for_each_sg(sgl, sgent, sglen, i) { sgl 1235 drivers/dma/ti/omap-dma.c data_type = __ffs((xt->src_start | xt->dst_start | xt->sgl[0].size)); sgl 1243 drivers/dma/ti/omap-dma.c sg->en = xt->sgl[0].size / BIT(data_type); sgl 1249 drivers/dma/ti/omap-dma.c src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]); sgl 1250 drivers/dma/ti/omap-dma.c dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]); sgl 490 drivers/dma/timb_dma.c struct scatterlist *sgl, unsigned int sg_len, sgl 501 drivers/dma/timb_dma.c if (!sgl || !sg_len) { sgl 521 drivers/dma/timb_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 803 drivers/dma/txx9dmac.c txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 827 drivers/dma/txx9dmac.c for_each_sg(sgl, sg, sg_len, i) { sgl 53 drivers/dma/uniphier-mdmac.c struct scatterlist *sgl; sgl 114 drivers/dma/uniphier-mdmac.c sg = &md->sgl[md->sg_cur]; sgl 230 drivers/dma/uniphier-mdmac.c uniphier_mdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, sgl 245 drivers/dma/uniphier-mdmac.c md->sgl = sgl; sgl 320 drivers/dma/uniphier-mdmac.c txstate->residue += sg_dma_len(&md->sgl[i]); sgl 1648 drivers/dma/xilinx/xilinx_dma.c if (!xt->numf || !xt->sgl[0].size) sgl 1671 drivers/dma/xilinx/xilinx_dma.c hw->hsize = xt->sgl[0].size; sgl 1672 drivers/dma/xilinx/xilinx_dma.c hw->stride = (xt->sgl[0].icg + xt->sgl[0].size) << sgl 1776 drivers/dma/xilinx/xilinx_dma.c struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, sgl 1801 drivers/dma/xilinx/xilinx_dma.c for_each_sg(sgl, sg, sg_len, i) { sgl 1992 drivers/dma/xilinx/xilinx_dma.c if (!xt->numf || !xt->sgl[0].size) sgl 2023 drivers/dma/xilinx/xilinx_dma.c hw->vsize_stride |= (xt->sgl[0].icg + xt->sgl[0].size) & sgl 2025 drivers/dma/xilinx/xilinx_dma.c hw->control = xt->sgl[0].size & XILINX_DMA_BD_HSIZE_MASK; sgl 546 drivers/dma/zx_dma.c struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen, sgl 556 drivers/dma/zx_dma.c if (!sgl) sgl 562 drivers/dma/zx_dma.c for_each_sg(sgl, sg, sglen, i) { sgl 574 drivers/dma/zx_dma.c for_each_sg(sgl, sg, sglen, i) { sgl 112 drivers/fpga/fpga-mgr.c sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); sgl 127 drivers/fpga/fpga-mgr.c len = sg_copy_to_buffer(sgt->sgl, sgt->nents, buf, sgl 191 drivers/fpga/fpga-mgr.c sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); sgl 400 drivers/fpga/zynq-fpga.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 409 drivers/fpga/zynq-fpga.c dma_map_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE); sgl 426 drivers/fpga/zynq-fpga.c priv->cur_sg = sgt->sgl; sgl 481 drivers/fpga/zynq-fpga.c dma_unmap_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE); sgl 781 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sg->sgl->dma_address = addr; sgl 782 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sg->sgl->length = size; sgl 784 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sg->sgl->dma_length = size; sgl 958 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c nents = dma_map_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction); sgl 986 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (!ttm->sg->sgl) sgl 990 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c dma_unmap_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction); sgl 341 drivers/gpu/drm/arm/malidp_planes.c struct scatterlist *sgl; sgl 354 drivers/gpu/drm/arm/malidp_planes.c sgl = sgt->sgl; sgl 356 drivers/gpu/drm/arm/malidp_planes.c while (sgl) { sgl 357 drivers/gpu/drm/arm/malidp_planes.c if (sgl->length < pgsize) { sgl 363 drivers/gpu/drm/arm/malidp_planes.c sgl = sg_next(sgl); sgl 398 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, count, i) { sgl 410 drivers/gpu/drm/armada/armada_gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) { sgl 419 drivers/gpu/drm/armada/armada_gem.c sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); sgl 421 drivers/gpu/drm/armada/armada_gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) sgl 427 drivers/gpu/drm/armada/armada_gem.c sg_dma_address(sgt->sgl) = dobj->dev_addr; sgl 428 drivers/gpu/drm/armada/armada_gem.c sg_dma_len(sgt->sgl) = dobj->obj.size; sgl 435 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, num, i) sgl 452 drivers/gpu/drm/armada/armada_gem.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); sgl 456 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, sgt->nents, i) sgl 557 drivers/gpu/drm/armada/armada_gem.c if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) { sgl 561 drivers/gpu/drm/armada/armada_gem.c dobj->dev_addr = sg_dma_address(dobj->sgt->sgl); sgl 129 drivers/gpu/drm/drm_cache.c for_each_sg_page(st->sgl, &sg_iter, st->nents, 0) sgl 476 drivers/gpu/drm/drm_gem_cma_helper.c dma_addr_t next_addr = sg_dma_address(sgt->sgl); sgl 480 drivers/gpu/drm/drm_gem_cma_helper.c for_each_sg(sgt->sgl, s, sgt->nents, i) { sgl 500 drivers/gpu/drm/drm_gem_cma_helper.c cma_obj->paddr = sg_dma_address(sgt->sgl); sgl 120 drivers/gpu/drm/drm_gem_shmem_helper.c dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl, sgl 393 drivers/gpu/drm/drm_gem_shmem_helper.c dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl, sgl 628 drivers/gpu/drm/drm_gem_shmem_helper.c dma_map_sg(obj->dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgl 627 drivers/gpu/drm/drm_prime.c if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgl 653 drivers/gpu/drm/drm_prime.c dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgl 957 drivers/gpu/drm/drm_prime.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgl 30 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_map_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgl 54 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_unmap_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgl 407 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_sync_sg_for_cpu(dev->dev, etnaviv_obj->sgt->sgl, sgl 424 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_sync_sg_for_device(dev->dev, etnaviv_obj->sgt->sgl, sgl 82 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 100 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, i, j) { sgl 116 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 247 drivers/gpu/drm/etnaviv/etnaviv_mmu.c iova = sg_dma_address(sgt->sgl) - memory_base; sgl 248 drivers/gpu/drm/etnaviv/etnaviv_mmu.c if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) { sgl 398 drivers/gpu/drm/exynos/exynos_drm_g2d.c dma_unmap_sg(to_dma_dev(g2d->drm_dev), g2d_userptr->sgt->sgl, sgl 514 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (!dma_map_sg(to_dma_dev(g2d->drm_dev), sgt->sgl, sgt->nents, sgl 521 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_userptr->dma_addr = sgt->sgl[0].dma_address; sgl 487 drivers/gpu/drm/exynos/exynos_drm_gem.c exynos_gem->dma_addr = sg_dma_address(sgt->sgl); sgl 10513 drivers/gpu/drm/i915/display/intel_display.c base = sg_dma_address(obj->mm.pages->sgl); sgl 1323 drivers/gpu/drm/i915/display/intel_overlay.c overlay->flip_addr = sg_dma_address(obj->mm.pages->sgl); sgl 43 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c src = obj->mm.pages->sgl; sgl 44 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c dst = st->sgl; sgl 51 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { sgl 74 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir); sgl 237 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c sg_page_sizes = i915_sg_page_sizes(pages->sgl); sgl 26 drivers/gpu/drm/i915/gem/i915_gem_internal.c for (sg = st->sgl; sg; sg = __sg_next(sg)) { sgl 77 drivers/gpu/drm/i915/gem/i915_gem_internal.c sg = st->sgl; sgl 112 drivers/gpu/drm/i915/gem/i915_gem_internal.c if (get_order(st->sgl->length)) { sgl 29 drivers/gpu/drm/i915/gem/i915_gem_pages.c obj->mm.get_page.sg_pos = pages->sgl; sgl 242 drivers/gpu/drm/i915/gem/i915_gem_pages.c return kmap(sg_page(sgt->sgl)); sgl 52 drivers/gpu/drm/i915/gem/i915_gem_phys.c sg = st->sgl; sgl 97 drivers/gpu/drm/i915/gem/i915_gem_phys.c dma_addr_t dma = sg_dma_address(pages->sgl); sgl 98 drivers/gpu/drm/i915/gem/i915_gem_phys.c void *vaddr = sg_page(pages->sgl); sgl 204 drivers/gpu/drm/i915/gem/i915_gem_phys.c unsigned int sg_page_sizes = i915_sg_page_sizes(pages->sgl); sgl 79 drivers/gpu/drm/i915/gem/i915_gem_shmem.c sg = st->sgl; sgl 497 drivers/gpu/drm/i915/gem/i915_gem_stolen.c sg = st->sgl; sgl 464 drivers/gpu/drm/i915/gem/i915_gem_userptr.c sg_page_sizes = i915_sg_page_sizes(st->sgl); sgl 17 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c for (sg = pages->sgl; sg && nreal--; sg = __sg_next(sg)) sgl 42 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c sg = pages->sgl; sgl 56 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { sgl 60 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c src = pages->sgl; sgl 47 drivers/gpu/drm/i915/gem/selftests/huge_pages.c for (sg = st->sgl; sg; sg = __sg_next(sg)) { sgl 75 drivers/gpu/drm/i915/gem/selftests/huge_pages.c sg = st->sgl; sgl 200 drivers/gpu/drm/i915/gem/selftests/huge_pages.c sg = st->sgl; sgl 253 drivers/gpu/drm/i915/gem/selftests/huge_pages.c sg = st->sgl; sgl 25 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c sg = st->sgl; sgl 31 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { sgl 49 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c dma_unmap_sg(attachment->dev, st->sgl, st->nents, dir); sgl 522 drivers/gpu/drm/i915/gt/intel_ringbuffer.c return sg_page(obj->mm.pages->sgl); sgl 24 drivers/gpu/drm/i915/gt/selftest_timeline.c return sg_page(obj->mm.pages->sgl); sgl 370 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c desc->db_trigger_phy = sg_dma_address(client->vma->pages->sgl) + sgl 595 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c return sg_pcopy_to_buffer(pages->sgl, pages->nents, dst, size, offset); sgl 66 drivers/gpu/drm/i915/gvt/dmabuf.c for_each_sg(st->sgl, sg, page_num, i) { sgl 139 drivers/gpu/drm/i915/i915_gem.c void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset; sgl 1161 drivers/gpu/drm/i915/i915_gem_gtt.c struct scatterlist *sg = vma->pages->sgl; sgl 2148 drivers/gpu/drm/i915/i915_gem_gtt.c pages->sgl, pages->nents, sgl 2542 drivers/gpu/drm/i915/i915_gem_gtt.c dma_unmap_sg(kdev, pages->sgl, pages->nents, PCI_DMA_BIDIRECTIONAL); sgl 3405 drivers/gpu/drm/i915/i915_gem_gtt.c sg = st->sgl; sgl 3486 drivers/gpu/drm/i915/i915_gem_gtt.c sg = st->sgl; sgl 3529 drivers/gpu/drm/i915/i915_gem_gtt.c sg = st->sgl; sgl 78 drivers/gpu/drm/i915/i915_gpu_error.c struct scatterlist *sgl; sgl 80 drivers/gpu/drm/i915/i915_gpu_error.c sgl = (typeof(sgl))__get_free_page(ALLOW_FAIL); sgl 81 drivers/gpu/drm/i915/i915_gpu_error.c if (!sgl) { sgl 90 drivers/gpu/drm/i915/i915_gpu_error.c (unsigned long)sgl | SG_CHAIN; sgl 92 drivers/gpu/drm/i915/i915_gpu_error.c e->sgl = sgl; sgl 95 drivers/gpu/drm/i915/i915_gpu_error.c e->cur = sgl; sgl 96 drivers/gpu/drm/i915/i915_gpu_error.c e->end = sgl + SG_MAX_SINGLE_ALLOC - 1; sgl 639 drivers/gpu/drm/i915/i915_gpu_error.c static void err_free_sgl(struct scatterlist *sgl) sgl 641 drivers/gpu/drm/i915/i915_gpu_error.c while (sgl) { sgl 644 drivers/gpu/drm/i915/i915_gpu_error.c for (sg = sgl; !sg_is_chain(sg); sg++) { sgl 651 drivers/gpu/drm/i915/i915_gpu_error.c free_page((unsigned long)sgl); sgl 652 drivers/gpu/drm/i915/i915_gpu_error.c sgl = sg; sgl 798 drivers/gpu/drm/i915/i915_gpu_error.c if (READ_ONCE(error->sgl)) sgl 815 drivers/gpu/drm/i915/i915_gpu_error.c GEM_BUG_ON(m.sgl && !m.cur); sgl 818 drivers/gpu/drm/i915/i915_gpu_error.c err_free_sgl(m.sgl); sgl 822 drivers/gpu/drm/i915/i915_gpu_error.c if (cmpxchg(&error->sgl, NULL, m.sgl)) sgl 823 drivers/gpu/drm/i915/i915_gpu_error.c err_free_sgl(m.sgl); sgl 845 drivers/gpu/drm/i915/i915_gpu_error.c sg = error->sgl; sgl 952 drivers/gpu/drm/i915/i915_gpu_error.c err_free_sgl(error->sgl); sgl 166 drivers/gpu/drm/i915/i915_gpu_error.h struct scatterlist *sgl, *fit; sgl 186 drivers/gpu/drm/i915/i915_gpu_error.h struct scatterlist *sgl, *cur, *end; sgl 21 drivers/gpu/drm/i915/i915_scatterlist.c new_sg = new_st.sgl; sgl 22 drivers/gpu/drm/i915/i915_scatterlist.c for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { sgl 27 drivers/gpu/drm/i915/i915_scatterlist.h } __sgt_iter(struct scatterlist *sgl, bool dma) { sgl 28 drivers/gpu/drm/i915/i915_scatterlist.h struct sgt_iter s = { .sgp = sgl }; sgl 77 drivers/gpu/drm/i915/i915_scatterlist.h for ((__iter) = __sgt_iter((__sgt)->sgl, true); \ sgl 89 drivers/gpu/drm/i915/i915_scatterlist.h for ((__iter) = __sgt_iter((__sgt)->sgl, false); \ sgl 406 drivers/gpu/drm/i915/i915_vma.h return sg_page(vma->pages->sgl); sgl 78 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c for (sg = pages->sgl; sg; sg = sg_next(sg)) { sgl 580 drivers/gpu/drm/i915/selftests/i915_vma.c sg = vma->pages->sgl; sgl 51 drivers/gpu/drm/i915/selftests/scatterlist.c for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { sgl 89 drivers/gpu/drm/i915/selftests/scatterlist.c for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) { sgl 236 drivers/gpu/drm/i915/selftests/scatterlist.c sg = pt->st.sgl; sgl 225 drivers/gpu/drm/mediatek/mtk_drm_gem.c expected = sg_dma_address(sg->sgl); sgl 226 drivers/gpu/drm/mediatek/mtk_drm_gem.c for_each_sg(sg->sgl, s, sg->nents, i) { sgl 235 drivers/gpu/drm/mediatek/mtk_drm_gem.c mtk_gem->dma_addr = sg_dma_address(sg->sgl); sgl 265 drivers/gpu/drm/mediatek/mtk_drm_gem.c for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) { sgl 56 drivers/gpu/drm/msm/msm_gem.c dma_sync_sg_for_device(dev, msm_obj->sgt->sgl, sgl 59 drivers/gpu/drm/msm/msm_gem.c dma_map_sg(dev, msm_obj->sgt->sgl, sgl 69 drivers/gpu/drm/msm/msm_gem.c dma_sync_sg_for_cpu(dev, msm_obj->sgt->sgl, sgl 72 drivers/gpu/drm/msm/msm_gem.c dma_unmap_sg(dev, msm_obj->sgt->sgl, sgl 49 drivers/gpu/drm/msm/msm_gpummu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 48 drivers/gpu/drm/msm/msm_iommu.c ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot); sgl 20 drivers/gpu/drm/nouveau/include/nvif/if000a.h struct scatterlist *sgl; sgl 67 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h struct scatterlist *sgl; sgl 119 drivers/gpu/drm/nouveau/nouveau_mem.c if (tt->ttm.sg) args.sgl = tt->ttm.sg->sgl; sgl 37 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c struct scatterlist *sgl; sgl 117 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c .sgl = mem->sgl, sgl 178 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c mem->sgl = args->v0.sgl; sgl 1476 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (map->sgl) { sgl 1477 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c for (; map->off; map->sgl = sg_next(map->sgl)) { sgl 1478 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u64 size = sg_dma_len(map->sgl); sgl 1483 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c func = map->page->desc->func->sgl; sgl 69 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h nvkm_vmm_pte_func sgl; sgl 321 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h sg_dma_address(MAP->sgl), sg_dma_len(MAP->sgl), \ sgl 322 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h (MAP->sgl = sg_next(MAP->sgl))) sgl 102 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c .sgl = gf100_vmm_pgt_sgl, sgl 41 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c .sgl = gf100_vmm_pgt_sgl, sgl 169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c .sgl = gp100_vmm_pgt_sgl, sgl 70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c .sgl = nv04_vmm_pgt_sgl, sgl 71 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c .sgl = nv41_vmm_pgt_sgl, sgl 175 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c .sgl = nv44_vmm_pgt_sgl, sgl 102 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c .sgl = nv50_vmm_pgt_sgl, sgl 1252 drivers/gpu/drm/omapdrm/omap_gem.c omap_obj->dma_addr = sg_dma_address(sgt->sgl); sgl 1270 drivers/gpu/drm/omapdrm/omap_gem.c for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) { sgl 42 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c sg_init_table(sg->sgl, 1); sgl 43 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c sg_dma_len(sg->sgl) = obj->size; sgl 44 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c sg_set_page(sg->sgl, pfn_to_page(PFN_DOWN(dma_addr)), obj->size, 0); sgl 45 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c sg_dma_address(sg->sgl) = dma_addr; sgl 43 drivers/gpu/drm/panfrost/panfrost_gem.c if (bo->sgts[i].sgl) { sgl 44 drivers/gpu/drm/panfrost/panfrost_gem.c dma_unmap_sg(pfdev->dev, bo->sgts[i].sgl, sgl 252 drivers/gpu/drm/panfrost/panfrost_mmu.c struct scatterlist *sgl; sgl 256 drivers/gpu/drm/panfrost/panfrost_mmu.c for_each_sg(sgt->sgl, sgl, sgt->nents, count) { sgl 257 drivers/gpu/drm/panfrost/panfrost_mmu.c unsigned long paddr = sg_dma_address(sgl); sgl 258 drivers/gpu/drm/panfrost/panfrost_mmu.c size_t len = sg_dma_len(sgl); sgl 520 drivers/gpu/drm/panfrost/panfrost_mmu.c if (!dma_map_sg(pfdev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL)) { sgl 530 drivers/gpu/drm/radeon/radeon_ttm.c nents = dma_map_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); sgl 558 drivers/gpu/drm/radeon/radeon_ttm.c if (!ttm->sg->sgl) sgl 562 drivers/gpu/drm/radeon/radeon_ttm.c dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); sgl 564 drivers/gpu/drm/radeon/radeon_ttm.c for_each_sg_page(ttm->sg->sgl, &sg_iter, ttm->sg->nents, 0) { sgl 173 drivers/gpu/drm/rcar-du/rcar_du_vsp.c cfg.mem[i] = sg_dma_address(state->sg_tables[i].sgl) sgl 235 drivers/gpu/drm/rcar-du/rcar_du_writeback.c cfg->mem[i] = sg_dma_address(rjob->sg_tables[i].sgl) sgl 38 drivers/gpu/drm/rockchip/rockchip_drm_gem.c ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl, sgl 100 drivers/gpu/drm/rockchip/rockchip_drm_gem.c for_each_sg(rk_obj->sgt->sgl, s, rk_obj->sgt->nents, i) sgl 103 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, sgl 352 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_unmap_sg(drm->dev, rk_obj->sgt->sgl, sgl 466 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_addr_t expected = sg_dma_address(sgt->sgl); sgl 470 drivers/gpu/drm/rockchip/rockchip_drm_gem.c for_each_sg(sgt->sgl, s, count, i) { sgl 495 drivers/gpu/drm/rockchip/rockchip_drm_gem.c int count = dma_map_sg(drm->dev, sg->sgl, sg->nents, sgl 502 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_unmap_sg(drm->dev, sg->sgl, sg->nents, sgl 507 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->dma_addr = sg_dma_address(sg->sgl); sgl 138 drivers/gpu/drm/tegra/gem.c bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl, sgl 206 drivers/gpu/drm/tegra/gem.c dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgl 232 drivers/gpu/drm/tegra/gem.c err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgl 368 drivers/gpu/drm/tegra/gem.c bo->paddr = sg_dma_address(bo->sgt->sgl); sgl 517 drivers/gpu/drm/tegra/gem.c for_each_sg(sgt->sgl, sg, bo->num_pages, i) sgl 520 drivers/gpu/drm/tegra/gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) sgl 526 drivers/gpu/drm/tegra/gem.c sg_dma_address(sgt->sgl) = bo->paddr; sgl 527 drivers/gpu/drm/tegra/gem.c sg_dma_len(sgt->sgl) = gem->size; sgl 546 drivers/gpu/drm/tegra/gem.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); sgl 565 drivers/gpu/drm/tegra/gem.c dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgl 579 drivers/gpu/drm/tegra/gem.c dma_sync_sg_for_device(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgl 54 drivers/gpu/drm/udl/udl_dmabuf.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, sgl 107 drivers/gpu/drm/udl/udl_dmabuf.c rd = obj->sg->sgl; sgl 108 drivers/gpu/drm/udl/udl_dmabuf.c wr = sgt->sgl; sgl 116 drivers/gpu/drm/udl/udl_dmabuf.c nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir); sgl 94 drivers/gpu/drm/v3d/v3d_mmu.c struct scatterlist *sgl; sgl 96 drivers/gpu/drm/v3d/v3d_mmu.c for_each_sg(shmem_obj->sgt->sgl, sgl, shmem_obj->sgt->nents, count) { sgl 97 drivers/gpu/drm/v3d/v3d_mmu.c u32 page_address = sg_dma_address(sgl) >> V3D_MMU_PAGE_SHIFT; sgl 101 drivers/gpu/drm/v3d/v3d_mmu.c BUG_ON(page_address + (sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT) >= sgl 104 drivers/gpu/drm/v3d/v3d_mmu.c for (i = 0; i < sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT; i++) sgl 496 drivers/gpu/drm/virtio/virtgpu_vq.c bo->pages->sgl, bo->pages->nents, sgl 903 drivers/gpu/drm/virtio/virtgpu_vq.c bo->pages->sgl, bo->pages->nents, sgl 983 drivers/gpu/drm/virtio/virtgpu_vq.c obj->pages->sgl, obj->pages->nents, sgl 998 drivers/gpu/drm/virtio/virtgpu_vq.c for_each_sg(obj->pages->sgl, sg, nents, si) { sgl 1026 drivers/gpu/drm/virtio/virtgpu_vq.c obj->pages->sgl, obj->mapped, sgl 345 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c __sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl, sgl 365 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents, sgl 388 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c ret = dma_map_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.orig_nents, sgl 145 drivers/gpu/host1x/job.c for_each_sg(sgt->sgl, sg, sgt->nents, j) sgl 159 drivers/gpu/host1x/job.c sgt->sgl, sgt->nents, IOMMU_READ); sgl 199 drivers/hsi/clients/cmt_speech.c u32 *data = sg_virt(msg->sgt.sgl); sgl 205 drivers/hsi/clients/cmt_speech.c u32 *data = sg_virt(msg->sgt.sgl); sgl 259 drivers/hsi/clients/cmt_speech.c kfree(sg_virt(msg->sgt.sgl)); sgl 281 drivers/hsi/clients/cmt_speech.c sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); sgl 624 drivers/hsi/clients/cmt_speech.c sg_init_one(msg->sgt.sgl, address, hi->buf_size); sgl 667 drivers/hsi/clients/cmt_speech.c sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0); sgl 719 drivers/hsi/clients/cmt_speech.c sg_init_one(txmsg->sgt.sgl, address, hi->buf_size); sgl 147 drivers/hsi/clients/hsi_char.c kfree(sg_virt(msg->sgt.sgl)); sgl 186 drivers/hsi/clients/hsi_char.c sg_init_one(msg->sgt.sgl, buf, alloc_size); sgl 217 drivers/hsi/clients/hsi_char.c return msg->sgt.sgl->length; sgl 222 drivers/hsi/clients/hsi_char.c msg->sgt.sgl->length = len; sgl 455 drivers/hsi/clients/hsi_char.c sg_virt(msg->sgt.sgl), hsc_msg_len_get(msg)); sgl 491 drivers/hsi/clients/hsi_char.c if (copy_from_user(sg_virt(msg->sgt.sgl), (void __user *)buf, len)) { sgl 157 drivers/hsi/clients/ssi_protocol.c data = sg_virt(msg->sgt.sgl); sgl 165 drivers/hsi/clients/ssi_protocol.c data = sg_virt(msg->sgt.sgl); sgl 178 drivers/hsi/clients/ssi_protocol.c sg = msg->sgt.sgl; sgl 249 drivers/hsi/clients/ssi_protocol.c kfree(sg_virt(msg->sgt.sgl)); sgl 269 drivers/hsi/clients/ssi_protocol.c sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); sgl 192 drivers/hsi/controllers/omap_ssi_core.c dma_unmap_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, dir); sgl 215 drivers/hsi/controllers/omap_ssi_core.c msg->actual_len = sg_dma_len(msg->sgt.sgl); sgl 200 drivers/hsi/controllers/omap_ssi_port.c omap_ssi->gdd_trn[lch].sg = msg->sgt.sgl; sgl 231 drivers/hsi/controllers/omap_ssi_port.c err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, sgl 246 drivers/hsi/controllers/omap_ssi_port.c d_addr = sg_dma_address(msg->sgt.sgl); sgl 248 drivers/hsi/controllers/omap_ssi_port.c err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, sgl 261 drivers/hsi/controllers/omap_ssi_port.c s_addr = sg_dma_address(msg->sgt.sgl); sgl 272 drivers/hsi/controllers/omap_ssi_port.c writew_relaxed(SSI_BYTES_TO_FRAMES(msg->sgt.sgl->length), sgl 330 drivers/hsi/controllers/omap_ssi_port.c if ((msg->sgt.nents) && (msg->sgt.sgl->length > sizeof(u32))) sgl 445 drivers/hsi/controllers/omap_ssi_port.c msg->channel, msg, msg->sgt.sgl->length, sgl 905 drivers/hsi/controllers/omap_ssi_port.c if ((!msg->sgt.nents) || (!msg->sgt.sgl->length)) { sgl 914 drivers/hsi/controllers/omap_ssi_port.c buf = sg_virt(msg->sgt.sgl) + msg->actual_len; sgl 924 drivers/hsi/controllers/omap_ssi_port.c if (msg->actual_len >= msg->sgt.sgl->length) sgl 70 drivers/hwtracing/intel_th/msu-sink.c for_each_sg((*sgt)->sgl, sg_ptr, nents, i) { sgl 87 drivers/hwtracing/intel_th/msu-sink.c for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) { sgl 276 drivers/hwtracing/intel_th/msu.c return win->sgt->sgl; sgl 326 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgl 421 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgl 660 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgl 928 drivers/hwtracing/intel_th/msu.c sg_set_buf(msc->single_sgt.sgl, page_address(page), size); sgl 930 drivers/hwtracing/intel_th/msu.c ret = dma_map_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl, 1, sgl 937 drivers/hwtracing/intel_th/msu.c msc->base_addr = sg_dma_address(msc->single_sgt.sgl); sgl 959 drivers/hwtracing/intel_th/msu.c dma_unmap_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl, sgl 1000 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { sgl 1013 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, i, ret) sgl 1028 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { sgl 1040 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) { sgl 1123 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, i) { sgl 1190 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgl 1385 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgl 52 drivers/infiniband/core/umem.c ib_dma_unmap_sg(dev, umem->sg_head.sgl, umem->sg_nents, sgl 55 drivers/infiniband/core/umem.c for_each_sg_page(umem->sg_head.sgl, &sg_iter, umem->sg_nents, 0) { sgl 164 drivers/infiniband/core/umem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) { sgl 274 drivers/infiniband/core/umem.c sg = umem->sg_head.sgl; sgl 301 drivers/infiniband/core/umem.c umem->sg_head.sgl, sgl 353 drivers/infiniband/core/umem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) sgl 382 drivers/infiniband/core/umem.c ret = sg_pcopy_to_buffer(umem->sg_head.sgl, umem->sg_nents, dst, length, sgl 2546 drivers/infiniband/core/verbs.c int ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents, sgl 2556 drivers/infiniband/core/verbs.c if (unlikely(sg_nents <= 0 || sg_offset > sg_dma_len(&sgl[0]))) sgl 2559 drivers/infiniband/core/verbs.c mr->iova = sg_dma_address(&sgl[0]) + sg_offset; sgl 2562 drivers/infiniband/core/verbs.c for_each_sg(sgl, sg, sg_nents, i) { sgl 863 drivers/infiniband/hw/bnxt_re/ib_verbs.c qplib_qp->sq.sg_info.sglist = umem->sg_head.sgl; sgl 876 drivers/infiniband/hw/bnxt_re/ib_verbs.c qplib_qp->rq.sg_info.sglist = umem->sg_head.sgl; sgl 1330 drivers/infiniband/hw/bnxt_re/ib_verbs.c qplib_srq->sg_info.sglist = umem->sg_head.sgl; sgl 2573 drivers/infiniband/hw/bnxt_re/ib_verbs.c cq->qplib_cq.sg_info.sglist = cq->umem->sg_head.sgl; sgl 3494 drivers/infiniband/hw/bnxt_re/ib_verbs.c rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap, page_size) sgl 176 drivers/infiniband/hw/cxgb3/cxio_wr.h struct t3_sge sgl[T3_MAX_SGE]; /* 4+ */ sgl 238 drivers/infiniband/hw/cxgb3/cxio_wr.h struct t3_sge sgl[T3_MAX_SGE]; /* 5+ */ sgl 274 drivers/infiniband/hw/cxgb3/cxio_wr.h struct t3_sge sgl[T3_MAX_SGE]; /* 3+ */ sgl 477 drivers/infiniband/hw/cxgb3/iwch_provider.c for_each_sg_dma_page(mhp->umem->sg_head.sgl, &sg_iter, mhp->umem->nmap, 0) { sgl 77 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); sgl 78 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); sgl 79 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); sgl 103 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[0].stag = wr->ex.imm_data; sgl 104 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[0].len = cpu_to_be32(0); sgl 114 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[i].stag = sgl 116 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[i].len = sgl 118 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[i].to = sgl 265 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); sgl 266 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); sgl 269 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = cpu_to_be64(((u32)wr->sg_list[i].addr) & sgl 276 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].stag = 0; sgl 277 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = 0; sgl 278 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = 0; sgl 331 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].stag = 0; sgl 332 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); sgl 333 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); sgl 339 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].stag = 0; sgl 340 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = 0; sgl 341 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = 0; sgl 711 drivers/infiniband/hw/cxgb3/iwch_qp.c term = (struct terminate_message *)wqe->send.sgl; sgl 67 drivers/infiniband/hw/cxgb4/mem.c struct ulptx_sgl *sgl; sgl 75 drivers/infiniband/hw/cxgb4/mem.c wr_len = roundup(sizeof(*req) + sizeof(*sgl), 16); sgl 97 drivers/infiniband/hw/cxgb4/mem.c sgl = (struct ulptx_sgl *)(req + 1); sgl 98 drivers/infiniband/hw/cxgb4/mem.c sgl->cmd_nsge = cpu_to_be32(ULPTX_CMD_V(ULP_TX_SC_DSGL) | sgl 100 drivers/infiniband/hw/cxgb4/mem.c sgl->len0 = cpu_to_be32(len); sgl 101 drivers/infiniband/hw/cxgb4/mem.c sgl->addr0 = cpu_to_be64(data); sgl 565 drivers/infiniband/hw/cxgb4/mem.c for_each_sg_dma_page(mhp->umem->sg_head.sgl, &sg_iter, mhp->umem->nmap, 0) { sgl 105 drivers/infiniband/hw/efa/efa_verbs.c struct scatterlist *sgl; sgl 1058 drivers/infiniband/hw/efa/efa_verbs.c rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap, sgl 1097 drivers/infiniband/hw/efa/efa_verbs.c struct scatterlist *pages_sgl = pbl->phys.indirect.sgl; sgl 1236 drivers/infiniband/hw/efa/efa_verbs.c struct scatterlist *sgl; sgl 1240 drivers/infiniband/hw/efa/efa_verbs.c sgl = efa_vmalloc_buf_to_sg(pbl->pbl_buf, size_in_pages); sgl 1241 drivers/infiniband/hw/efa/efa_verbs.c if (!sgl) sgl 1244 drivers/infiniband/hw/efa/efa_verbs.c sg_dma_cnt = dma_map_sg(&dev->pdev->dev, sgl, size_in_pages, DMA_TO_DEVICE); sgl 1251 drivers/infiniband/hw/efa/efa_verbs.c pbl->phys.indirect.sgl = sgl; sgl 1268 drivers/infiniband/hw/efa/efa_verbs.c dma_unmap_sg(&dev->pdev->dev, sgl, size_in_pages, DMA_TO_DEVICE); sgl 1270 drivers/infiniband/hw/efa/efa_verbs.c kfree(sgl); sgl 1277 drivers/infiniband/hw/efa/efa_verbs.c dma_unmap_sg(&dev->pdev->dev, pbl->phys.indirect.sgl, sgl 1279 drivers/infiniband/hw/efa/efa_verbs.c kfree(pbl->phys.indirect.sgl); sgl 280 drivers/infiniband/hw/hns/hns_roce_alloc.c rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap, sgl 45 drivers/infiniband/hw/hns/hns_roce_db.c db->dma = sg_dma_address(page->umem->sg_head.sgl) + offset; sgl 46 drivers/infiniband/hw/hns/hns_roce_db.c db->virt_addr = sg_virt(page->umem->sg_head.sgl) + offset; sgl 1890 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) { sgl 2241 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) { sgl 1060 drivers/infiniband/hw/hns/hns_roce_mr.c for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 1103 drivers/infiniband/hw/hns/hns_roce_mr.c for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 1324 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwpbl->qp_mr.sq_page = sg_page(region->sg_head.sgl); sgl 1326 drivers/infiniband/hw/i40iw/i40iw_verbs.c rdma_for_each_block(region->sg_head.sgl, &biter, region->nmap, sgl 2098 drivers/infiniband/hw/i40iw/i40iw_verbs.c static void i40iw_copy_sg_list(struct i40iw_sge *sg_list, struct ib_sge *sgl, int num_sges) sgl 2103 drivers/infiniband/hw/i40iw/i40iw_verbs.c sg_list[i].tag_off = sgl[i].addr; sgl 2104 drivers/infiniband/hw/i40iw/i40iw_verbs.c sg_list[i].len = sgl[i].length; sgl 2105 drivers/infiniband/hw/i40iw/i40iw_verbs.c sg_list[i].stag = sgl[i].lkey; sgl 77 drivers/infiniband/hw/mlx4/doorbell.c db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); sgl 203 drivers/infiniband/hw/mlx4/mr.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) { sgl 274 drivers/infiniband/hw/mlx4/mr.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) { sgl 77 drivers/infiniband/hw/mlx5/doorbell.c db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); sgl 65 drivers/infiniband/hw/mlx5/mem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { sgl 155 drivers/infiniband/hw/mlx5/mem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { sgl 2063 drivers/infiniband/hw/mlx5/mr.c struct scatterlist *sgl, sgl 2070 drivers/infiniband/hw/mlx5/mr.c struct scatterlist *sg = sgl; sgl 2079 drivers/infiniband/hw/mlx5/mr.c for_each_sg(sgl, sg, sg_nents, i) { sgl 909 drivers/infiniband/hw/mthca/mthca_provider.c for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) { sgl 837 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c for_each_sg_dma_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 633 drivers/infiniband/hw/qedr/verbs.c for_each_sg_dma_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 1308 drivers/infiniband/hw/qedr/verbs.c sg = srq->prod_umem->sg_head.sgl; sgl 192 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 415 drivers/infiniband/sw/rdmavt/mr.c for_each_sg_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 200 drivers/infiniband/sw/rxe/rxe_mr.c for_each_sg_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) { sgl 585 drivers/input/touchscreen/sur40.c sgt->sgl, sgt->nents, sur40->pix_fmt.sizeimage, 0); sgl 609 drivers/iommu/dma-iommu.c for_each_sg(sgt.sgl, sg, sgt.orig_nents, i) sgl 613 drivers/iommu/dma-iommu.c if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, ioprot) sgl 677 drivers/iommu/dma-iommu.c struct scatterlist *sgl, int nelems, sgl 686 drivers/iommu/dma-iommu.c for_each_sg(sgl, sg, nelems, i) sgl 691 drivers/iommu/dma-iommu.c struct scatterlist *sgl, int nelems, sgl 700 drivers/iommu/dma-iommu.c for_each_sg(sgl, sg, nelems, i) sgl 1082 drivers/iommu/dma-iommu.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgl 52 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_addr_t expected = sg_dma_address(sgt->sgl); sgl 56 drivers/media/common/videobuf2/videobuf2-dma-contig.c for_each_sg(sgt->sgl, s, sgt->nents, i) { sgl 102 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, sgl 115 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); sgl 243 drivers/media/common/videobuf2/videobuf2-dma-contig.c rd = buf->sgt_base->sgl; sgl 244 drivers/media/common/videobuf2/videobuf2-dma-contig.c wr = sgt->sgl; sgl 276 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 302 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 311 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt->nents = dma_map_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 434 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 532 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 548 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_addr = sg_dma_address(sgt->sgl); sgl 556 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 610 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_addr = sg_dma_address(sgt->sgl); sgl 145 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 183 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 205 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, sgl 218 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); sgl 261 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgl 289 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir, sgl 387 drivers/media/common/videobuf2/videobuf2-dma-sg.c rd = buf->dma_sgt->sgl; sgl 388 drivers/media/common/videobuf2/videobuf2-dma-sg.c wr = sgt->sgl; sgl 414 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 440 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 446 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 233 drivers/media/common/videobuf2/videobuf2-vmalloc.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 263 drivers/media/common/videobuf2/videobuf2-vmalloc.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 289 drivers/media/common/videobuf2/videobuf2-vmalloc.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 295 drivers/media/common/videobuf2/videobuf2-vmalloc.c sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgl 80 drivers/media/pci/cobalt/cobalt-v4l2.c ret = descriptor_list_create(cobalt, sg_desc->sgl, sgl 1586 drivers/media/pci/cx23885/cx23885-core.c sgt->sgl, sgl 145 drivers/media/pci/cx23885/cx23885-vbi.c sgt->sgl, sgl 359 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, 0, UNSET, sgl 364 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, UNSET, 0, sgl 392 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, line0_offset, sgl 399 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, sgl 406 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, sgl 180 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, 0, UNSET, sgl 185 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, UNSET, 0, sgl 194 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, line0_offset, sgl 200 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, sgl 206 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, sgl 225 drivers/media/pci/cx88/cx88-mpeg.c rc = cx88_risc_databuffer(dev->pci, risc, sgt->sgl, sgl 147 drivers/media/pci/cx88/cx88-vbi.c cx88_risc_buffer(dev->pci, &buf->risc, sgt->sgl, sgl 451 drivers/media/pci/cx88/cx88-video.c sgt->sgl, 0, UNSET, sgl 456 drivers/media/pci/cx88/cx88-video.c sgt->sgl, UNSET, 0, sgl 461 drivers/media/pci/cx88/cx88-video.c sgt->sgl, sgl 468 drivers/media/pci/cx88/cx88-video.c sgt->sgl, sgl 476 drivers/media/pci/cx88/cx88-video.c sgt->sgl, 0, buf->bpl, sgl 872 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (sg->nents && sg->sgl) sgl 873 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->offset = sg->sgl->offset; sgl 876 drivers/media/pci/intel/ipu3/ipu3-cio2.c for_each_sg_dma_page (sg->sgl, &sg_iter, sg->nents, 0) { sgl 215 drivers/media/pci/saa7134/saa7134-core.c base += dma->sgl[0].offset; sgl 101 drivers/media/pci/saa7134/saa7134-ts.c return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents, sgl 114 drivers/media/pci/saa7134/saa7134-vbi.c if (dma->sgl->offset) { sgl 124 drivers/media/pci/saa7134/saa7134-vbi.c return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents, sgl 922 drivers/media/pci/saa7134/saa7134-video.c if (dma->sgl->offset) { sgl 933 drivers/media/pci/saa7134/saa7134-video.c return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents, sgl 322 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c for_each_sg(vbuf->sgl, sg, vbuf->nents, i) { sgl 743 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c sg_copy_from_buffer(sgt->sgl, sgt->nents, sgl 747 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c sg_copy_from_buffer(sgt->sgl, sgt->nents, sgl 455 drivers/media/pci/tw68/tw68-video.c tw68_risc_buffer(dev->pci, buf, dma->sgl, sgl 459 drivers/media/pci/tw68/tw68-video.c tw68_risc_buffer(dev->pci, buf, dma->sgl, sgl 463 drivers/media/pci/tw68/tw68-video.c tw68_risc_buffer(dev->pci, buf, dma->sgl, sgl 468 drivers/media/pci/tw68/tw68-video.c tw68_risc_buffer(dev->pci, buf, dma->sgl, sgl 474 drivers/media/pci/tw68/tw68-video.c tw68_risc_buffer(dev->pci, buf, dma->sgl, sgl 197 drivers/media/pci/tw686x/tw686x-video.c for_each_sg(vbuf->sgl, sg, vbuf->nents, i) { sgl 236 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width; sgl 237 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width; sgl 243 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width; sgl 244 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width; sgl 250 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 251 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 257 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 258 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 264 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 265 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 271 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 272 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 278 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width; sgl 279 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width; sgl 285 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 286 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 292 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width / 2; sgl 293 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width / 2; sgl 299 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width * 2; sgl 300 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width * 2; sgl 306 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width * 2; sgl 307 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width * 2; sgl 314 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].size = s_width * 2; sgl 315 drivers/media/platform/m2m-deinterlace.c ctx->xt->sgl[0].icg = s_width * 2; sgl 1260 drivers/media/platform/marvell-ccic/mcam-core.c for_each_sg(sg_table->sgl, sg, sg_table->nents, i) { sgl 261 drivers/media/platform/omap/omap_vout_vrfb.c xt->sgl[0].size = vout->pix.width * vout->bpp; sgl 262 drivers/media/platform/omap/omap_vout_vrfb.c xt->sgl[0].icg = dst_icg; sgl 449 drivers/media/platform/omap3isp/ispccdc.c dma_sync_sg_for_cpu(isp->dev, req->table.sgt.sgl, sgl 458 drivers/media/platform/omap3isp/ispccdc.c dma_sync_sg_for_device(isp->dev, req->table.sgt.sgl, sgl 164 drivers/media/platform/omap3isp/ispstat.c dma_sync_sg_for_device(stat->isp->dev, buf->sgt.sgl, sgl 174 drivers/media/platform/omap3isp/ispstat.c dma_sync_sg_for_cpu(stat->isp->dev, buf->sgt.sgl, sgl 1394 drivers/media/platform/pxa_camera.c ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels, sgl 364 drivers/media/platform/qcom/camss/camss-video.c buffer->addr[i] = sg_dma_address(sgt->sgl); sgl 1042 drivers/media/platform/qcom/venus/helpers.c buf->dma_addr = sg_dma_address(sgt->sgl); sgl 118 drivers/media/platform/rockchip/rga/rga-buf.c struct scatterlist *sgl; sgl 131 drivers/media/platform/rockchip/rga/rga-buf.c for_each_sg(sgt->sgl, sgl, sgt->nents, i) { sgl 132 drivers/media/platform/rockchip/rga/rga-buf.c len = sg_dma_len(sgl) >> PAGE_SHIFT; sgl 133 drivers/media/platform/rockchip/rga/rga-buf.c address = sg_phys(sgl); sgl 369 drivers/media/platform/via-camera.c viafb_dma_copy_out_sg(cam->cb_offsets[bufn], sgt->sgl, sgt->nents); sgl 915 drivers/media/platform/vsp1/vsp1_drm.c return dma_map_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents, sgl 924 drivers/media/platform/vsp1/vsp1_drm.c dma_unmap_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents, sgl 367 drivers/media/platform/xilinx/xilinx-dma.c dma->sgl[0].size = dma->format.width * dma->fmtinfo->bpp; sgl 368 drivers/media/platform/xilinx/xilinx-dma.c dma->sgl[0].icg = dma->format.bytesperline - dma->sgl[0].size; sgl 95 drivers/media/platform/xilinx/xilinx-dma.h struct data_chunk sgl[1]; sgl 130 drivers/message/fusion/mptctl.c static void kfree_sgl(MptSge_t *sgl, dma_addr_t sgl_dma, sgl 785 drivers/message/fusion/mptctl.c MptSge_t *sgl, *sgIn; sgl 861 drivers/message/fusion/mptctl.c if ((sgl = kbuf_alloc_2_sgl(fwlen, sgdir, sge_offset, sgl 885 drivers/message/fusion/mptctl.c iocp->name, sgl, numfrags)); sgl 892 drivers/message/fusion/mptctl.c sgIn = sgl; sgl 958 drivers/message/fusion/mptctl.c if (sgl) sgl 959 drivers/message/fusion/mptctl.c kfree_sgl(sgl, sgl_dma, buflist, iocp); sgl 988 drivers/message/fusion/mptctl.c kfree_sgl(sgl, sgl_dma, buflist, iocp); sgl 1014 drivers/message/fusion/mptctl.c MptSge_t *sgl; sgl 1062 drivers/message/fusion/mptctl.c sgl = sglbuf; sgl 1084 drivers/message/fusion/mptctl.c sgl->FlagsLength = (0x10000000|sgdir|this_alloc); sgl 1087 drivers/message/fusion/mptctl.c sgl->Address = dma_addr; sgl 1091 drivers/message/fusion/mptctl.c sgl++; sgl 1118 drivers/message/fusion/mptctl.c sgl[-1].FlagsLength |= 0xC1000000; sgl 1158 drivers/message/fusion/mptctl.c kfree_sgl(MptSge_t *sgl, dma_addr_t sgl_dma, struct buflist *buflist, MPT_ADAPTER *ioc) sgl 1160 drivers/message/fusion/mptctl.c MptSge_t *sg = sgl; sgl 1207 drivers/message/fusion/mptctl.c pci_free_consistent(ioc->pcidev, MAX_SGL_BYTES, sgl, sgl_dma); sgl 495 drivers/misc/fastrpc.c if (!dma_map_sg(attachment->dev, table->sgl, table->nents, dir)) sgl 505 drivers/misc/fastrpc.c dma_unmap_sg(attach->dev, table->sgl, table->nents, dir); sgl 629 drivers/misc/fastrpc.c map->phys = sg_dma_address(map->table->sgl); sgl 632 drivers/misc/fastrpc.c map->va = sg_virt(map->table->sgl); sgl 348 drivers/misc/genwqe/card_base.h struct sg_entry *sgl; sgl 368 drivers/misc/genwqe/card_base.h int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, sgl 371 drivers/misc/genwqe/card_base.h int genwqe_setup_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, sgl 374 drivers/misc/genwqe/card_base.h int genwqe_free_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl); sgl 839 drivers/misc/genwqe/card_dev.c if (req->sgls[i].sgl != NULL) sgl 288 drivers/misc/genwqe/card_utils.c int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, sgl 294 drivers/misc/genwqe/card_utils.c sgl->fpage_offs = offset_in_page((unsigned long)user_addr); sgl 295 drivers/misc/genwqe/card_utils.c sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size); sgl 296 drivers/misc/genwqe/card_utils.c sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); sgl 297 drivers/misc/genwqe/card_utils.c sgl->lpage_size = (user_size - sgl->fpage_size) % PAGE_SIZE; sgl 300 drivers/misc/genwqe/card_utils.c __func__, user_addr, user_size, sgl->nr_pages, sgl 301 drivers/misc/genwqe/card_utils.c sgl->fpage_offs, sgl->fpage_size, sgl->lpage_size); sgl 303 drivers/misc/genwqe/card_utils.c sgl->user_addr = user_addr; sgl 304 drivers/misc/genwqe/card_utils.c sgl->user_size = user_size; sgl 305 drivers/misc/genwqe/card_utils.c sgl->write = write; sgl 306 drivers/misc/genwqe/card_utils.c sgl->sgl_size = genwqe_sgl_size(sgl->nr_pages); sgl 308 drivers/misc/genwqe/card_utils.c if (get_order(sgl->sgl_size) > MAX_ORDER) { sgl 314 drivers/misc/genwqe/card_utils.c sgl->sgl = __genwqe_alloc_consistent(cd, sgl->sgl_size, sgl 315 drivers/misc/genwqe/card_utils.c &sgl->sgl_dma_addr); sgl 316 drivers/misc/genwqe/card_utils.c if (sgl->sgl == NULL) { sgl 323 drivers/misc/genwqe/card_utils.c if ((sgl->fpage_size != 0) && (sgl->fpage_size != PAGE_SIZE)) { sgl 324 drivers/misc/genwqe/card_utils.c sgl->fpage = __genwqe_alloc_consistent(cd, PAGE_SIZE, sgl 325 drivers/misc/genwqe/card_utils.c &sgl->fpage_dma_addr); sgl 326 drivers/misc/genwqe/card_utils.c if (sgl->fpage == NULL) sgl 330 drivers/misc/genwqe/card_utils.c if (copy_from_user(sgl->fpage + sgl->fpage_offs, sgl 331 drivers/misc/genwqe/card_utils.c user_addr, sgl->fpage_size)) { sgl 336 drivers/misc/genwqe/card_utils.c if (sgl->lpage_size != 0) { sgl 337 drivers/misc/genwqe/card_utils.c sgl->lpage = __genwqe_alloc_consistent(cd, PAGE_SIZE, sgl 338 drivers/misc/genwqe/card_utils.c &sgl->lpage_dma_addr); sgl 339 drivers/misc/genwqe/card_utils.c if (sgl->lpage == NULL) sgl 343 drivers/misc/genwqe/card_utils.c if (copy_from_user(sgl->lpage, user_addr + user_size - sgl 344 drivers/misc/genwqe/card_utils.c sgl->lpage_size, sgl->lpage_size)) { sgl 352 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, PAGE_SIZE, sgl->lpage, sgl 353 drivers/misc/genwqe/card_utils.c sgl->lpage_dma_addr); sgl 354 drivers/misc/genwqe/card_utils.c sgl->lpage = NULL; sgl 355 drivers/misc/genwqe/card_utils.c sgl->lpage_dma_addr = 0; sgl 357 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, PAGE_SIZE, sgl->fpage, sgl 358 drivers/misc/genwqe/card_utils.c sgl->fpage_dma_addr); sgl 359 drivers/misc/genwqe/card_utils.c sgl->fpage = NULL; sgl 360 drivers/misc/genwqe/card_utils.c sgl->fpage_dma_addr = 0; sgl 362 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, sgl->sgl_size, sgl->sgl, sgl 363 drivers/misc/genwqe/card_utils.c sgl->sgl_dma_addr); sgl 364 drivers/misc/genwqe/card_utils.c sgl->sgl = NULL; sgl 365 drivers/misc/genwqe/card_utils.c sgl->sgl_dma_addr = 0; sgl 366 drivers/misc/genwqe/card_utils.c sgl->sgl_size = 0; sgl 371 drivers/misc/genwqe/card_utils.c int genwqe_setup_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, sgl 378 drivers/misc/genwqe/card_utils.c size_t size = sgl->user_size; sgl 381 drivers/misc/genwqe/card_utils.c map_offs = sgl->fpage_offs; /* offset in first page */ sgl 383 drivers/misc/genwqe/card_utils.c s = &sgl->sgl[0]; /* first set of 8 entries */ sgl 385 drivers/misc/genwqe/card_utils.c while (p < sgl->nr_pages) { sgl 391 drivers/misc/genwqe/card_utils.c s[j].target_addr = cpu_to_be64(sgl->sgl_dma_addr + dma_offs); sgl 400 drivers/misc/genwqe/card_utils.c if ((p == 0) && (sgl->fpage != NULL)) { sgl 401 drivers/misc/genwqe/card_utils.c daddr = sgl->fpage_dma_addr + map_offs; sgl 403 drivers/misc/genwqe/card_utils.c } else if ((p == sgl->nr_pages - 1) && sgl 404 drivers/misc/genwqe/card_utils.c (sgl->lpage != NULL)) { sgl 405 drivers/misc/genwqe/card_utils.c daddr = sgl->lpage_dma_addr; sgl 424 drivers/misc/genwqe/card_utils.c if (p == sgl->nr_pages) sgl 440 drivers/misc/genwqe/card_utils.c if (p == sgl->nr_pages) sgl 468 drivers/misc/genwqe/card_utils.c int genwqe_free_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl) sgl 475 drivers/misc/genwqe/card_utils.c if (sgl->fpage) { sgl 476 drivers/misc/genwqe/card_utils.c if (sgl->write) { sgl 477 drivers/misc/genwqe/card_utils.c res = copy_to_user(sgl->user_addr, sgl 478 drivers/misc/genwqe/card_utils.c sgl->fpage + sgl->fpage_offs, sgl->fpage_size); sgl 486 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, PAGE_SIZE, sgl->fpage, sgl 487 drivers/misc/genwqe/card_utils.c sgl->fpage_dma_addr); sgl 488 drivers/misc/genwqe/card_utils.c sgl->fpage = NULL; sgl 489 drivers/misc/genwqe/card_utils.c sgl->fpage_dma_addr = 0; sgl 491 drivers/misc/genwqe/card_utils.c if (sgl->lpage) { sgl 492 drivers/misc/genwqe/card_utils.c if (sgl->write) { sgl 493 drivers/misc/genwqe/card_utils.c offset = sgl->user_size - sgl->lpage_size; sgl 494 drivers/misc/genwqe/card_utils.c res = copy_to_user(sgl->user_addr + offset, sgl->lpage, sgl 495 drivers/misc/genwqe/card_utils.c sgl->lpage_size); sgl 503 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, PAGE_SIZE, sgl->lpage, sgl 504 drivers/misc/genwqe/card_utils.c sgl->lpage_dma_addr); sgl 505 drivers/misc/genwqe/card_utils.c sgl->lpage = NULL; sgl 506 drivers/misc/genwqe/card_utils.c sgl->lpage_dma_addr = 0; sgl 508 drivers/misc/genwqe/card_utils.c __genwqe_free_consistent(cd, sgl->sgl_size, sgl->sgl, sgl 509 drivers/misc/genwqe/card_utils.c sgl->sgl_dma_addr); sgl 511 drivers/misc/genwqe/card_utils.c sgl->sgl = NULL; sgl 512 drivers/misc/genwqe/card_utils.c sgl->sgl_dma_addr = 0x0; sgl 513 drivers/misc/genwqe/card_utils.c sgl->sgl_size = 0; sgl 3107 drivers/misc/habanalabs/goya/goya.c static int goya_dma_map_sg(struct hl_device *hdev, struct scatterlist *sgl, sgl 3113 drivers/misc/habanalabs/goya/goya.c if (!dma_map_sg(&hdev->pdev->dev, sgl, nents, dir)) sgl 3117 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgl, sg, nents, i) sgl 3123 drivers/misc/habanalabs/goya/goya.c static void goya_dma_unmap_sg(struct hl_device *hdev, struct scatterlist *sgl, sgl 3130 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgl, sg, nents, i) sgl 3133 drivers/misc/habanalabs/goya/goya.c dma_unmap_sg(&hdev->pdev->dev, sgl, nents, dir); sgl 3145 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgl 3200 drivers/misc/habanalabs/goya/goya.c rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl, sgl 3657 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgl 539 drivers/misc/habanalabs/habanalabs.h struct scatterlist *sgl, int nents, sgl 543 drivers/misc/habanalabs/habanalabs.h struct scatterlist *sgl, int nents, sgl 193 drivers/misc/habanalabs/memory.c rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl, sgl 673 drivers/misc/habanalabs/memory.c for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) { sgl 702 drivers/misc/habanalabs/memory.c for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) { sgl 1337 drivers/misc/habanalabs/memory.c userptr->sgt->sgl, sgl 66 drivers/misc/mic/scif/scif_debugfs.c for_each_sg(window->st->sgl, sg, window->st->nents, j) sgl 232 drivers/misc/mic/scif/scif_rma.c window->st->sgl, window->st->nents, sgl 539 drivers/misc/mic/scif/scif_rma.c for_each_sg(window->st->sgl, sg, window->st->nents, i) sgl 542 drivers/misc/mic/scif/scif_rma.c err = dma_map_sg(&remote_dev->sdev->dev, window->st->sgl, sgl 547 drivers/misc/mic/scif/scif_rma.c sg = window->st->sgl; sgl 156 drivers/mmc/core/sdio_ops.c data.sg = sgtable.sgl; sgl 777 drivers/mmc/host/dw_mmc.c struct scatterlist *sgl = host->data->sg; sgl 806 drivers/mmc/host/dw_mmc.c desc = dmaengine_prep_slave_sg(host->dms->ch, sgl, sgl 821 drivers/mmc/host/dw_mmc.c dma_sync_sg_for_device(mmc_dev(host->slot->mmc), sgl, sgl 220 drivers/mmc/host/mxs-mmc.c struct scatterlist * sgl; sgl 227 drivers/mmc/host/mxs-mmc.c sgl = data->sg; sgl 231 drivers/mmc/host/mxs-mmc.c sgl = (struct scatterlist *) ssp->ssp_pio_words; sgl 236 drivers/mmc/host/mxs-mmc.c sgl, sg_len, ssp->slave_dirn, flags); sgl 349 drivers/mmc/host/mxs-mmc.c struct scatterlist *sgl = data->sg, *sg; sgl 395 drivers/mmc/host/mxs-mmc.c for_each_sg(sgl, sg, sg_len, i) sgl 1257 drivers/mmc/host/omap_hsmmc.c struct scatterlist *sgl; sgl 1259 drivers/mmc/host/omap_hsmmc.c sgl = data->sg + i; sgl 1260 drivers/mmc/host/omap_hsmmc.c if (sgl->length % data->blksz) sgl 809 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c int raw_len, struct scatterlist *sgl, sgl 817 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c sg_init_one(sgl, buf, len); sgl 818 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c ret = dma_map_sg(this->dev, sgl, 1, dr); sgl 827 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c sg_init_one(sgl, this->data_buffer_dma, len); sgl 832 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c dma_map_sg(this->dev, sgl, 1, dr); sgl 2273 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c sg_init_one(&transfer->sgl, transfer->cmdbuf, naddr + 1); sgl 2274 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c dma_map_sg(this->dev, &transfer->sgl, 1, DMA_TO_DEVICE); sgl 2278 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1, DMA_MEM_TO_DEV, sgl 2315 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c *direct = prepare_data_dma(this, buf, raw_len, &transfer->sgl, sgl 2331 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c pio[4] = transfer->sgl.dma_address; sgl 2341 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1, sgl 2362 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c prepare_data_dma(this, buf, raw_len, &transfer->sgl, DMA_TO_DEVICE); sgl 2377 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c pio[4] = transfer->sgl.dma_address; sgl 2388 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1, sgl 2533 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c dma_unmap_sg(this->dev, &transfer->sgl, 1, sgl 110 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.h struct scatterlist sgl; sgl 201 drivers/mtd/nand/raw/lpc32xx_mlc.c struct scatterlist sgl; sgl 401 drivers/mtd/nand/raw/lpc32xx_mlc.c sg_init_one(&host->sgl, mem, len); sgl 403 drivers/mtd/nand/raw/lpc32xx_mlc.c res = dma_map_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 409 drivers/mtd/nand/raw/lpc32xx_mlc.c desc = dmaengine_prep_slave_sg(host->dma_chan, &host->sgl, 1, dir, sgl 425 drivers/mtd/nand/raw/lpc32xx_mlc.c dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 429 drivers/mtd/nand/raw/lpc32xx_mlc.c dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 228 drivers/mtd/nand/raw/lpc32xx_slc.c struct scatterlist sgl; sgl 450 drivers/mtd/nand/raw/lpc32xx_slc.c sg_init_one(&host->sgl, mem, len); sgl 452 drivers/mtd/nand/raw/lpc32xx_slc.c res = dma_map_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 458 drivers/mtd/nand/raw/lpc32xx_slc.c desc = dmaengine_prep_slave_sg(host->dma_chan, &host->sgl, 1, dir, sgl 474 drivers/mtd/nand/raw/lpc32xx_slc.c dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 479 drivers/mtd/nand/raw/lpc32xx_slc.c dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1, sgl 732 drivers/mtd/nand/raw/qcom_nandc.c struct scatterlist *sgl; sgl 744 drivers/mtd/nand/raw/qcom_nandc.c sgl = &bam_txn->cmd_sgl[bam_txn->cmd_sgl_start]; sgl 750 drivers/mtd/nand/raw/qcom_nandc.c sgl = &bam_txn->data_sgl[bam_txn->tx_sgl_start]; sgl 756 drivers/mtd/nand/raw/qcom_nandc.c sgl = &bam_txn->data_sgl[bam_txn->rx_sgl_start]; sgl 763 drivers/mtd/nand/raw/qcom_nandc.c sg_mark_end(sgl + sgl_cnt - 1); sgl 764 drivers/mtd/nand/raw/qcom_nandc.c ret = dma_map_sg(nandc->dev, sgl, sgl_cnt, desc->dir); sgl 772 drivers/mtd/nand/raw/qcom_nandc.c desc->bam_sgl = sgl; sgl 774 drivers/mtd/nand/raw/qcom_nandc.c dma_desc = dmaengine_prep_slave_sg(chan, sgl, sgl_cnt, dir_eng, sgl 779 drivers/mtd/nand/raw/qcom_nandc.c dma_unmap_sg(nandc->dev, sgl, sgl_cnt, desc->dir); sgl 898 drivers/mtd/nand/raw/qcom_nandc.c struct scatterlist *sgl; sgl 907 drivers/mtd/nand/raw/qcom_nandc.c sgl = &desc->adm_sgl; sgl 909 drivers/mtd/nand/raw/qcom_nandc.c sg_init_one(sgl, vaddr, size); sgl 919 drivers/mtd/nand/raw/qcom_nandc.c ret = dma_map_sg(nandc->dev, sgl, 1, desc->dir); sgl 944 drivers/mtd/nand/raw/qcom_nandc.c dma_desc = dmaengine_prep_slave_sg(nandc->chan, sgl, 1, dir_eng, 0); sgl 915 drivers/mtd/nand/raw/stm32_fmc2_nand.c for_each_sg(fmc2->dma_data_sg.sgl, sg, eccsteps, s) { sgl 920 drivers/mtd/nand/raw/stm32_fmc2_nand.c ret = dma_map_sg(fmc2->dev, fmc2->dma_data_sg.sgl, sgl 925 drivers/mtd/nand/raw/stm32_fmc2_nand.c desc_data = dmaengine_prep_slave_sg(dma_ch, fmc2->dma_data_sg.sgl, sgl 946 drivers/mtd/nand/raw/stm32_fmc2_nand.c for_each_sg(fmc2->dma_ecc_sg.sgl, sg, eccsteps, s) { sgl 951 drivers/mtd/nand/raw/stm32_fmc2_nand.c ret = dma_map_sg(fmc2->dev, fmc2->dma_ecc_sg.sgl, sgl 957 drivers/mtd/nand/raw/stm32_fmc2_nand.c fmc2->dma_ecc_sg.sgl, sgl 1014 drivers/mtd/nand/raw/stm32_fmc2_nand.c dma_unmap_sg(fmc2->dev, fmc2->dma_ecc_sg.sgl, sgl 1018 drivers/mtd/nand/raw/stm32_fmc2_nand.c dma_unmap_sg(fmc2->dev, fmc2->dma_data_sg.sgl, eccsteps, dma_data_dir); sgl 749 drivers/mtd/ubi/eba.c struct ubi_sgl *sgl, int lnum, int offset, int len, sgl 757 drivers/mtd/ubi/eba.c ubi_assert(sgl->list_pos < UBI_MAX_SG_COUNT); sgl 758 drivers/mtd/ubi/eba.c sg = &sgl->sg[sgl->list_pos]; sgl 759 drivers/mtd/ubi/eba.c if (len < sg->length - sgl->page_pos) sgl 762 drivers/mtd/ubi/eba.c to_read = sg->length - sgl->page_pos; sgl 765 drivers/mtd/ubi/eba.c sg_virt(sg) + sgl->page_pos, offset, sgl 773 drivers/mtd/ubi/eba.c sgl->page_pos += to_read; sgl 774 drivers/mtd/ubi/eba.c if (sgl->page_pos == sg->length) { sgl 775 drivers/mtd/ubi/eba.c sgl->list_pos++; sgl 776 drivers/mtd/ubi/eba.c sgl->page_pos = 0; sgl 782 drivers/mtd/ubi/eba.c sgl->list_pos++; sgl 783 drivers/mtd/ubi/eba.c sgl->page_pos = 0; sgl 462 drivers/mtd/ubi/kapi.c int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl, sgl 478 drivers/mtd/ubi/kapi.c err = ubi_eba_read_leb_sg(ubi, vol, sgl, lnum, offset, len, check); sgl 889 drivers/mtd/ubi/ubi.h struct ubi_sgl *sgl, int lnum, int offset, int len, sgl 369 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[i]))); sgl 372 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c sge_len - 1, le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1])); sgl 377 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1])); sgl 623 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c u16 sge_idx = RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[j])); sgl 3965 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h __le16 sgl[8]; sgl 2881 drivers/net/ethernet/broadcom/cnic_defs.h struct fcoe_ext_mul_sges_ctx sgl; sgl 1094 drivers/net/ethernet/chelsio/cxgb3/sge.c const struct sg_ent *sgl, sgl 1119 drivers/net/ethernet/chelsio/cxgb3/sge.c const u64 *fp = (const u64 *)sgl; sgl 1186 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sg_ent *sgp, sgl[MAX_SKB_FRAGS / 2 + 1]; sgl 1240 drivers/net/ethernet/chelsio/cxgb3/sge.c sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl; sgl 1243 drivers/net/ethernet/chelsio/cxgb3/sge.c write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits, gen, sgl 1602 drivers/net/ethernet/chelsio/cxgb3/sge.c const struct sg_ent *sgl, int sgl_flits) sgl 1609 drivers/net/ethernet/chelsio/cxgb3/sge.c for (p = dui->addr; sgl_flits >= 3; sgl++, sgl_flits -= 3) { sgl 1610 drivers/net/ethernet/chelsio/cxgb3/sge.c *p++ = be64_to_cpu(sgl->addr[0]); sgl 1611 drivers/net/ethernet/chelsio/cxgb3/sge.c *p++ = be64_to_cpu(sgl->addr[1]); sgl 1614 drivers/net/ethernet/chelsio/cxgb3/sge.c *p = be64_to_cpu(sgl->addr[0]); sgl 1636 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sg_ent *sgp, sgl[MAX_SKB_FRAGS / 2 + 1]; sgl 1652 drivers/net/ethernet/chelsio/cxgb3/sge.c sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl; sgl 1661 drivers/net/ethernet/chelsio/cxgb3/sge.c write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits, sgl 1093 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct ulptx_sgl *sgl; sgl 1902 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct ulptx_sgl *sgl, u64 *end, unsigned int start, sgl 3536 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c dma_unmap_sg(adapter->pdev_dev, adapter->hma.sgt->sgl, sgl 3541 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(adapter->hma.sgt->sgl, iter, sgl 3556 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct scatterlist *sgl, *iter; sgl 3610 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgl = adapter->hma.sgt->sgl; sgl 3612 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(sgl, iter, sgt->orig_nents, i) { sgl 3624 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgt->nents = dma_map_sg(adapter->pdev_dev, sgl, sgt->orig_nents, sgl 3639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(sgl, iter, sgt->nents, i) { sgl 302 drivers/net/ethernet/chelsio/cxgb4/sge.c const struct ulptx_sgl *sgl, const struct sge_txq *q) sgl 308 drivers/net/ethernet/chelsio/cxgb4/sge.c dma_unmap_single(dev, be64_to_cpu(sgl->addr0), ntohl(sgl->len0), sgl 311 drivers/net/ethernet/chelsio/cxgb4/sge.c dma_unmap_page(dev, be64_to_cpu(sgl->addr0), ntohl(sgl->len0), sgl 320 drivers/net/ethernet/chelsio/cxgb4/sge.c for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { sgl 381 drivers/net/ethernet/chelsio/cxgb4/sge.c unmap_sgl(dev, d->skb, d->sgl, q); sgl 881 drivers/net/ethernet/chelsio/cxgb4/sge.c struct ulptx_sgl *sgl, u64 *end, unsigned int start, sgl 892 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl->len0 = htonl(len); sgl 893 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl->addr0 = cpu_to_be64(addr[0] + start); sgl 896 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl->len0 = htonl(skb_frag_size(&si->frags[0])); sgl 897 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl->addr0 = cpu_to_be64(addr[1]); sgl 900 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | sgl 909 drivers/net/ethernet/chelsio/cxgb4/sge.c to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge; sgl 923 drivers/net/ethernet/chelsio/cxgb4/sge.c unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1; sgl 926 drivers/net/ethernet/chelsio/cxgb4/sge.c memcpy(sgl->sge, buf, part0); sgl 1366 drivers/net/ethernet/chelsio/cxgb4/sge.c u64 cntrl, *end, *sgl; sgl 1544 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl = (u64 *)(cpl + 1); /* sgl start here */ sgl 1545 drivers/net/ethernet/chelsio/cxgb4/sge.c if (unlikely((u8 *)sgl >= (u8 *)q->q.stat)) { sgl 1550 drivers/net/ethernet/chelsio/cxgb4/sge.c if (sgl == (u64 *)q->q.stat) { sgl 1554 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl = (void *)q->q.desc; sgl 1567 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl = (u64 *)(cpl + 1); sgl 1601 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_inline_tx_skb(skb, &q->q, sgl); sgl 1606 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_write_sgl(skb, &q->q, (void *)sgl, end, 0, addr); sgl 1613 drivers/net/ethernet/chelsio/cxgb4/sge.c q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)sgl; sgl 1940 drivers/net/ethernet/chelsio/cxgb4/sge.c struct ulptx_sgl *sgl = (struct ulptx_sgl *)(cpl + 1); sgl 1950 drivers/net/ethernet/chelsio/cxgb4/sge.c if (unlikely((void *)sgl == (void *)tq->stat)) { sgl 1951 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl = (void *)tq->desc; sgl 1956 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_write_sgl(skb, tq, sgl, end, 0, addr); sgl 1963 drivers/net/ethernet/chelsio/cxgb4/sge.c tq->sdesc[last_desc].sgl = sgl; sgl 159 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct ulptx_sgl *sgl; /* scatter/gather list in TX Queue */ sgl 307 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const struct ulptx_sgl *sgl, const struct sge_txq *tq) sgl 313 drivers/net/ethernet/chelsio/cxgb4vf/sge.c dma_unmap_single(dev, be64_to_cpu(sgl->addr0), sgl 314 drivers/net/ethernet/chelsio/cxgb4vf/sge.c be32_to_cpu(sgl->len0), DMA_TO_DEVICE); sgl 316 drivers/net/ethernet/chelsio/cxgb4vf/sge.c dma_unmap_page(dev, be64_to_cpu(sgl->addr0), sgl 317 drivers/net/ethernet/chelsio/cxgb4vf/sge.c be32_to_cpu(sgl->len0), DMA_TO_DEVICE); sgl 325 drivers/net/ethernet/chelsio/cxgb4vf/sge.c for (p = sgl->sge; nfrags >= 2; nfrags -= 2) { sgl 394 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq); sgl 902 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct ulptx_sgl *sgl, u64 *end, unsigned int start, sgl 913 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl->len0 = htonl(len); sgl 914 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl->addr0 = cpu_to_be64(addr[0] + start); sgl 917 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl->len0 = htonl(skb_frag_size(&si->frags[0])); sgl 918 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl->addr0 = cpu_to_be64(addr[1]); sgl 921 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | sgl 930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; sgl 944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; sgl 947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c memcpy(sgl->sge, buf, part0); sgl 1419 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct ulptx_sgl *sgl = (struct ulptx_sgl *)(cpl + 1); sgl 1430 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (unlikely((void *)sgl == (void *)tq->stat)) { sgl 1431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl = (void *)tq->desc; sgl 1435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c write_sgl(skb, tq, sgl, end, 0, addr); sgl 1442 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->sdesc[last_desc].sgl = sgl; sgl 95 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h struct scatterlist *sgl; sgl 781 drivers/net/ethernet/emulex/benet/be_cmds.c return &wrb->payload.sgl[0]; sgl 39 drivers/net/ethernet/emulex/benet/be_cmds.h struct be_sge sgl[19]; /* used by non-embedded cmds */ sgl 929 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct scatterlist *sgl, unsigned int sgc); sgl 931 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct scatterlist *sgl, unsigned int sgc); sgl 28 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->sgl = NULL; sgl 109 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (ddp->sgl) sgl 110 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, sgl 133 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c struct scatterlist *sgl, unsigned int sgc, sgl 152 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (!netdev || !sgl) sgl 168 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c if (ddp->sgl) { sgl 170 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c xid, ddp->sgl, ddp->sgc); sgl 188 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); sgl 201 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->sgl = sgl; sgl 205 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c for_each_sg(sgl, sg, dmacount, i) { sgl 323 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dma_unmap_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); sgl 344 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c struct scatterlist *sgl, unsigned int sgc) sgl 346 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 0); sgl 365 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c struct scatterlist *sgl, unsigned int sgc) sgl 367 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 1); sgl 441 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, sgl 444 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c ddp->sgl = NULL; sgl 39 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.h struct scatterlist *sgl; sgl 391 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c struct scatterlist *sgl; sgl 432 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c mmc_dat.sg = sdiodev->sgtable.sgl; sgl 447 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c sgl = sdiodev->sgtable.sgl; sgl 458 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c sg_set_buf(sgl, pkt_data, sg_data_sz); sgl 461 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c sgl = sg_next(sgl); sgl 472 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c sgl = sdiodev->sgtable.sgl; sgl 512 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c sg_init_table(sdiodev->sgtable.sgl, sdiodev->sgtable.orig_nents); sgl 98 drivers/nvme/host/fc.c struct scatterlist sgl[SG_CHUNK_SIZE]; sgl 1774 drivers/nvme/host/fc.c op->op.fcp_req.first_sgl = &op->sgl[0]; sgl 2141 drivers/nvme/host/fc.c freq->sg_table.sgl = freq->first_sgl; sgl 2143 drivers/nvme/host/fc.c blk_rq_nr_phys_segments(rq), freq->sg_table.sgl, sgl 2148 drivers/nvme/host/fc.c op->nents = blk_rq_map_sg(rq->q, rq, freq->sg_table.sgl); sgl 2150 drivers/nvme/host/fc.c freq->sg_cnt = fc_dma_map_sg(ctrl->lport->dev, freq->sg_table.sgl, sgl 2173 drivers/nvme/host/fc.c fc_dma_unmap_sg(ctrl->lport->dev, freq->sg_table.sgl, op->nents, sgl 2260 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) | sgl 2262 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.length = cpu_to_le32(data_len); sgl 2263 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.addr = 0; sgl 571 drivers/nvme/host/pci.c static void nvme_print_sgl(struct scatterlist *sgl, int nents) sgl 576 drivers/nvme/host/pci.c for_each_sg(sgl, sg, nents, i) { sgl 713 drivers/nvme/host/pci.c nvme_pci_sgl_set_data(&cmd->dptr.sgl, sg); sgl 734 drivers/nvme/host/pci.c nvme_pci_sgl_set_seg(&cmd->dptr.sgl, sgl_dma, entries); sgl 789 drivers/nvme/host/pci.c cmnd->dptr.sgl.addr = cpu_to_le64(iod->first_dma); sgl 790 drivers/nvme/host/pci.c cmnd->dptr.sgl.length = cpu_to_le32(iod->dma_len); sgl 791 drivers/nvme/host/pci.c cmnd->dptr.sgl.type = NVME_SGL_FMT_DATA_DESC << 4; sgl 1164 drivers/nvme/host/rdma.c ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq)); sgl 1185 drivers/nvme/host/rdma.c struct nvme_sgl_desc *sg = &c->common.dptr.sgl; sgl 1186 drivers/nvme/host/rdma.c struct scatterlist *sgl = req->sg_table.sgl; sgl 1191 drivers/nvme/host/rdma.c for (i = 0; i < count; i++, sgl++, sge++) { sgl 1192 drivers/nvme/host/rdma.c sge->addr = sg_dma_address(sgl); sgl 1193 drivers/nvme/host/rdma.c sge->length = sg_dma_len(sgl); sgl 1211 drivers/nvme/host/rdma.c sg->addr = cpu_to_le64(sg_dma_address(req->sg_table.sgl)); sgl 1212 drivers/nvme/host/rdma.c put_unaligned_le24(sg_dma_len(req->sg_table.sgl), sg->length); sgl 1233 drivers/nvme/host/rdma.c nr = ib_map_mr_sg(req->mr, req->sg_table.sgl, count, NULL, SZ_4K); sgl 1280 drivers/nvme/host/rdma.c req->sg_table.sgl = req->first_sgl; sgl 1282 drivers/nvme/host/rdma.c blk_rq_nr_phys_segments(rq), req->sg_table.sgl, sgl 1287 drivers/nvme/host/rdma.c req->nents = blk_rq_map_sg(rq->q, rq, req->sg_table.sgl); sgl 1289 drivers/nvme/host/rdma.c count = ib_dma_map_sg(ibdev, req->sg_table.sgl, req->nents, sgl 1319 drivers/nvme/host/rdma.c ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq)); sgl 1988 drivers/nvme/host/tcp.c struct nvme_sgl_desc *sg = &c->common.dptr.sgl; sgl 1999 drivers/nvme/host/tcp.c struct nvme_sgl_desc *sg = &c->common.dptr.sgl; sgl 2009 drivers/nvme/host/tcp.c struct nvme_sgl_desc *sg = &c->common.dptr.sgl; sgl 157 drivers/nvme/target/loop.c iod->sg_table.sgl = iod->first_sgl; sgl 160 drivers/nvme/target/loop.c iod->sg_table.sgl, SG_CHUNK_SIZE)) { sgl 165 drivers/nvme/target/loop.c iod->req.sg = iod->sg_table.sgl; sgl 166 drivers/nvme/target/loop.c iod->req.sg_cnt = blk_rq_map_sg(req->q, req, iod->sg_table.sgl); sgl 635 drivers/nvme/target/rdma.c struct nvme_sgl_desc *sgl = &rsp->req.cmd->common.dptr.sgl; sgl 636 drivers/nvme/target/rdma.c u64 off = le64_to_cpu(sgl->addr); sgl 637 drivers/nvme/target/rdma.c u32 len = le32_to_cpu(sgl->length); sgl 661 drivers/nvme/target/rdma.c struct nvme_keyed_sgl_desc *sgl, bool invalidate) sgl 664 drivers/nvme/target/rdma.c u64 addr = le64_to_cpu(sgl->addr); sgl 665 drivers/nvme/target/rdma.c u32 key = get_unaligned_le32(sgl->key); sgl 668 drivers/nvme/target/rdma.c rsp->req.transfer_len = get_unaligned_le24(sgl->length); sgl 699 drivers/nvme/target/rdma.c struct nvme_keyed_sgl_desc *sgl = &rsp->req.cmd->common.dptr.ksgl; sgl 701 drivers/nvme/target/rdma.c switch (sgl->type >> 4) { sgl 703 drivers/nvme/target/rdma.c switch (sgl->type & 0xf) { sgl 707 drivers/nvme/target/rdma.c pr_err("invalid SGL subtype: %#x\n", sgl->type); sgl 713 drivers/nvme/target/rdma.c switch (sgl->type & 0xf) { sgl 715 drivers/nvme/target/rdma.c return nvmet_rdma_map_sgl_keyed(rsp, sgl, true); sgl 717 drivers/nvme/target/rdma.c return nvmet_rdma_map_sgl_keyed(rsp, sgl, false); sgl 719 drivers/nvme/target/rdma.c pr_err("invalid SGL subtype: %#x\n", sgl->type); sgl 725 drivers/nvme/target/rdma.c pr_err("invalid SGL type: %#x\n", sgl->type); sgl 320 drivers/nvme/target/tcp.c struct nvme_sgl_desc *sgl = &cmd->req.cmd->common.dptr.sgl; sgl 321 drivers/nvme/target/tcp.c u32 len = le32_to_cpu(sgl->length); sgl 326 drivers/nvme/target/tcp.c if (sgl->type == ((NVME_SGL_FMT_DATA_DESC << 4) | sgl 825 drivers/nvme/target/tcp.c req->data_len = le32_to_cpu(req->cmd->common.dptr.sgl.length); sgl 913 drivers/nvme/target/tcp.c le32_to_cpu(req->cmd->common.dptr.sgl.length)); sgl 772 drivers/pci/p2pdma.c void pci_p2pmem_free_sgl(struct pci_dev *pdev, struct scatterlist *sgl) sgl 777 drivers/pci/p2pdma.c for_each_sg(sgl, sg, INT_MAX, count) { sgl 783 drivers/pci/p2pdma.c kfree(sgl); sgl 578 drivers/rapidio/devices/rio_mport_cdev.c req->sgt.sgl, req->sgt.nents, req->dir); sgl 622 drivers/rapidio/devices/rio_mport_cdev.c tx_data.sg = sgt->sgl; sgl 930 drivers/rapidio/devices/rio_mport_cdev.c sg_set_buf(req->sgt.sgl, sgl 936 drivers/rapidio/devices/rio_mport_cdev.c req->sgt.sgl, req->sgt.nents, dir); sgl 825 drivers/rapidio/devices/tsi721_dma.c struct scatterlist *sgl, unsigned int sg_len, sgl 835 drivers/rapidio/devices/tsi721_dma.c if (!sgl || !sg_len) { sgl 877 drivers/rapidio/devices/tsi721_dma.c desc->sg = sgl; sgl 488 drivers/scsi/3w-9xxx.c command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); sgl 489 drivers/scsi/3w-9xxx.c command_packet->byte8_offset.param.sgl[0].length = cpu_to_le32(TW_SECTOR_SIZE); sgl 1081 drivers/scsi/3w-9xxx.c command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); sgl 1082 drivers/scsi/3w-9xxx.c command_packet->byte8_offset.param.sgl[0].length = cpu_to_le32(TW_SECTOR_SIZE); sgl 1384 drivers/scsi/3w-9xxx.c TW_SG_Entry *sgl; sgl 1407 drivers/scsi/3w-9xxx.c sgl = (TW_SG_Entry *)((u32 *)oldcommand+oldcommand->size - (sizeof(TW_SG_Entry)/4) + pae); sgl 1409 drivers/scsi/3w-9xxx.c sgl = (TW_SG_Entry *)((u32 *)oldcommand+TW_SGL_OUT(oldcommand->opcode__sgloffset)); sgl 1410 drivers/scsi/3w-9xxx.c sgl->address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); sgl 1411 drivers/scsi/3w-9xxx.c sgl->length = cpu_to_le32(length); sgl 498 drivers/scsi/3w-9xxx.h TW_SG_Entry sgl[TW_ESCALADE_MAX_SGL_LENGTH]; sgl 502 drivers/scsi/3w-9xxx.h TW_SG_Entry sgl[TW_ESCALADE_MAX_SGL_LENGTH]; sgl 423 drivers/scsi/3w-sas.c command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); sgl 424 drivers/scsi/3w-sas.c command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE); sgl 683 drivers/scsi/3w-sas.c TW_SG_Entry_ISO *sgl; sgl 705 drivers/scsi/3w-sas.c sgl = (TW_SG_Entry_ISO *)((u32 *)oldcommand+oldcommand->size - (sizeof(TW_SG_Entry_ISO)/4) + pae + (sizeof(dma_addr_t) > 4 ? 1 : 0)); sgl 706 drivers/scsi/3w-sas.c sgl->address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1); sgl 707 drivers/scsi/3w-sas.c sgl->length = TW_CPU_TO_SGL(length); sgl 951 drivers/scsi/3w-sas.c command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]); sgl 952 drivers/scsi/3w-sas.c command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE); sgl 229 drivers/scsi/3w-sas.h TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; sgl 233 drivers/scsi/3w-sas.h TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; sgl 589 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 590 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 731 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 732 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 967 drivers/scsi/3w-xxxx.c tw_ioctl->firmware_command.byte8.param.sgl[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1; sgl 968 drivers/scsi/3w-xxxx.c tw_ioctl->firmware_command.byte8.param.sgl[0].length = data_buffer_length_adjusted; sgl 971 drivers/scsi/3w-xxxx.c tw_ioctl->firmware_command.byte8.io.sgl[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1; sgl 972 drivers/scsi/3w-xxxx.c tw_ioctl->firmware_command.byte8.io.sgl[0].length = data_buffer_length_adjusted; sgl 1167 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 1168 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 1439 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 1440 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 1549 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 1550 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 1639 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 1640 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 1762 drivers/scsi/3w-xxxx.c command_packet->byte8.io.sgl[i].address = sg_dma_address(sg); sgl 1763 drivers/scsi/3w-xxxx.c command_packet->byte8.io.sgl[i].length = sg_dma_len(sg); sgl 1884 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].address = param_value; sgl 1885 drivers/scsi/3w-xxxx.c command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector); sgl 307 drivers/scsi/3w-xxxx.h TW_SG_Entry sgl[TW_MAX_SGL_LENGTH]; sgl 311 drivers/scsi/3w-xxxx.h TW_SG_Entry sgl[TW_MAX_SGL_LENGTH]; sgl 43 drivers/scsi/be2iscsi/be_cmds.h struct be_sge sgl[19]; /* used by non-embedded cmds */ sgl 698 drivers/scsi/be2iscsi/be_cmds.h return &wrb->payload.sgl[0]; sgl 2935 drivers/scsi/be2iscsi/be_main.c struct be_dma_mem *sgl) sgl 2940 drivers/scsi/be2iscsi/be_main.c WARN_ON(!sgl); sgl 2942 drivers/scsi/be2iscsi/be_main.c sgl->va = virtual_address; sgl 2943 drivers/scsi/be2iscsi/be_main.c sgl->dma = (unsigned long)physical_address; sgl 2944 drivers/scsi/be2iscsi/be_main.c sgl->size = length; sgl 2949 drivers/scsi/be2iscsi/be_main.c static void be_sgl_destroy_contiguous(struct be_dma_mem *sgl) sgl 2951 drivers/scsi/be2iscsi/be_main.c memset(sgl, 0, sizeof(*sgl)); sgl 2956 drivers/scsi/be2iscsi/be_main.c struct mem_array *pmem, struct be_dma_mem *sgl) sgl 2958 drivers/scsi/be2iscsi/be_main.c if (sgl->va) sgl 2959 drivers/scsi/be2iscsi/be_main.c be_sgl_destroy_contiguous(sgl); sgl 2963 drivers/scsi/be2iscsi/be_main.c pmem->size, sgl); sgl 2968 drivers/scsi/be2iscsi/be_main.c struct mem_array *pmem, struct be_dma_mem *sgl) sgl 2970 drivers/scsi/be2iscsi/be_main.c if (sgl->va) sgl 2971 drivers/scsi/be2iscsi/be_main.c be_sgl_destroy_contiguous(sgl); sgl 2975 drivers/scsi/be2iscsi/be_main.c pmem->size, sgl); sgl 3240 drivers/scsi/be2iscsi/be_main.c struct be_dma_mem sgl; sgl 3250 drivers/scsi/be2iscsi/be_main.c hwi_build_be_sgl_arr(phba, pm_arr, &sgl); sgl 3252 drivers/scsi/be2iscsi/be_main.c &phba->ctrl, &sgl); sgl 3275 drivers/scsi/be2iscsi/be_main.c struct be_dma_mem sgl; sgl 3289 drivers/scsi/be2iscsi/be_main.c hwi_build_be_sgl_arr(phba, pm_arr, &sgl); sgl 3290 drivers/scsi/be2iscsi/be_main.c status = be_cmd_iscsi_post_sgl_pages(&phba->ctrl, &sgl, sgl 3342 drivers/scsi/be2iscsi/be_main.c struct be_dma_mem sgl; sgl 3414 drivers/scsi/be2iscsi/be_main.c hwi_build_be_sgl_by_offset(phba, &pwrb_arr[i], &sgl); sgl 3415 drivers/scsi/be2iscsi/be_main.c status = be_cmd_wrbq_create(&phba->ctrl, &sgl, sgl 579 drivers/scsi/bnx2fc/57xx_hsi_bnx2fc.h struct fcoe_ext_mul_sges_ctx sgl; sgl 1458 drivers/scsi/bnx2fc/bnx2fc_hwi.c struct fcoe_ext_mul_sges_ctx *sgl; sgl 1506 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo = sgl 1508 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi = sgl 1510 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size = sgl 1512 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_off = sgl 1514 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_idx = i; sgl 1520 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; sgl 1521 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.lo = (u32)phys_addr; sgl 1522 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.hi = (u32)((u64)phys_addr >> 32); sgl 1523 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.sgl_size = bd_count; sgl 1524 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_off = offset; /*adjusted offset */ sgl 1525 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_idx = i; sgl 1577 drivers/scsi/bnx2fc/bnx2fc_hwi.c struct fcoe_ext_mul_sges_ctx *sgl; sgl 1603 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo = sgl 1605 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi = sgl 1607 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size = 1; sgl 1658 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; sgl 1660 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.lo = (u32)mp_req->mp_resp_bd_dma; sgl 1661 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.hi = sgl 1663 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.sgl_size = 1; sgl 1675 drivers/scsi/bnx2fc/bnx2fc_hwi.c struct fcoe_ext_mul_sges_ctx *sgl; sgl 1713 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo = sgl 1715 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi = sgl 1717 drivers/scsi/bnx2fc/bnx2fc_hwi.c task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size = sgl 1776 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; sgl 1806 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.lo = (u32)bd_tbl->bd_tbl_dma; sgl 1807 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.hi = sgl 1809 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.sgl_size = bd_count; sgl 1812 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.lo = (u32)bd_tbl->bd_tbl_dma; sgl 1813 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.cur_sge_addr.hi = sgl 1815 drivers/scsi/bnx2fc/bnx2fc_hwi.c sgl->mul_sgl.sgl_size = bd_count; sgl 291 drivers/scsi/csiostor/csio_scsi.c struct ulptx_sgl *sgl) sgl 301 drivers/scsi/csiostor/csio_scsi.c sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | ULPTX_MORE_F | sgl 307 drivers/scsi/csiostor/csio_scsi.c sgl->addr0 = cpu_to_be64(sg_dma_address(sgel)); sgl 308 drivers/scsi/csiostor/csio_scsi.c sgl->len0 = cpu_to_be32(sg_dma_len(sgel)); sgl 309 drivers/scsi/csiostor/csio_scsi.c sge_pair = (struct ulptx_sge_pair *)(sgl + 1); sgl 331 drivers/scsi/csiostor/csio_scsi.c sgl->addr0 = cpu_to_be64(dma_buf->paddr); sgl 332 drivers/scsi/csiostor/csio_scsi.c sgl->len0 = cpu_to_be32( sgl 334 drivers/scsi/csiostor/csio_scsi.c sge_pair = (struct ulptx_sge_pair *)(sgl + 1); sgl 365 drivers/scsi/csiostor/csio_scsi.c struct ulptx_sgl *sgl; sgl 397 drivers/scsi/csiostor/csio_scsi.c sgl = (struct ulptx_sgl *)((uintptr_t)wrp + sgl 401 drivers/scsi/csiostor/csio_scsi.c csio_scsi_init_ultptx_dsgl(hw, req, sgl); sgl 418 drivers/scsi/csiostor/csio_scsi.c struct ulptx_sgl *sgl; sgl 450 drivers/scsi/csiostor/csio_scsi.c sgl = (struct ulptx_sgl *)((uintptr_t)wrp + sgl 454 drivers/scsi/csiostor/csio_scsi.c csio_scsi_init_ultptx_dsgl(hw, req, sgl); sgl 1092 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c struct scatterlist *sg = ttinfo->sgl; sgl 1977 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c struct scatterlist *sg = ttinfo->sgl; sgl 1216 drivers/scsi/cxgbi/libcxgbi.c scmd_get_params(struct scsi_cmnd *sc, struct scatterlist **sgl, sgl 1222 drivers/scsi/cxgbi/libcxgbi.c *sgl = sdb->table.sgl; sgl 1314 drivers/scsi/cxgbi/libcxgbi.c static int cxgbi_ddp_sgl_check(struct scatterlist *sgl, int nents) sgl 1318 drivers/scsi/cxgbi/libcxgbi.c struct scatterlist *sg = sgl; sgl 1344 drivers/scsi/cxgbi/libcxgbi.c struct scatterlist *sgl = ttinfo->sgl; sgl 1346 drivers/scsi/cxgbi/libcxgbi.c unsigned int sg_offset = sgl->offset; sgl 1365 drivers/scsi/cxgbi/libcxgbi.c if (cxgbi_ddp_sgl_check(sgl, sgcnt) < 0) sgl 1368 drivers/scsi/cxgbi/libcxgbi.c ttinfo->nr_pages = (xferlen + sgl->offset + (1 << PAGE_SHIFT) - 1) >> sgl 1385 drivers/scsi/cxgbi/libcxgbi.c sgl->offset = 0; sgl 1386 drivers/scsi/cxgbi/libcxgbi.c err = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); sgl 1387 drivers/scsi/cxgbi/libcxgbi.c sgl->offset = sg_offset; sgl 1402 drivers/scsi/cxgbi/libcxgbi.c cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset, sgl 1422 drivers/scsi/cxgbi/libcxgbi.c dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); sgl 1447 drivers/scsi/cxgbi/libcxgbi.c dma_unmap_sg(&ppm->pdev->dev, ttinfo->sgl, ttinfo->nents, sgl 1475 drivers/scsi/cxgbi/libcxgbi.c scmd_get_params(sc, &ttinfo->sgl, &ttinfo->nents, sgl 1818 drivers/scsi/cxgbi/libcxgbi.c static int sgl_seek_offset(struct scatterlist *sgl, unsigned int sgcnt, sgl 1825 drivers/scsi/cxgbi/libcxgbi.c for_each_sg(sgl, sg, sgcnt, i) { sgl 1987 drivers/scsi/cxgbi/libcxgbi.c sdb->table.sgl, sdb->table.nents, sgl 1151 drivers/scsi/esas2r/esas2r.h struct list_head *sgl; sgl 1156 drivers/scsi/esas2r/esas2r.h sgl = a->free_sg_list_head.next; sgl 1157 drivers/scsi/esas2r/esas2r.h result = list_entry(sgl, struct esas2r_mem_desc, next_desc); sgl 1158 drivers/scsi/esas2r/esas2r.h list_del_init(sgl); sgl 794 drivers/scsi/esas2r/esas2r_init.c struct esas2r_mem_desc *sgl; sgl 863 drivers/scsi/esas2r/esas2r_init.c for (i = 0, sgl = a->sg_list_mds; i < num_sg_lists; i++, sgl++) { sgl 864 drivers/scsi/esas2r/esas2r_init.c sgl->size = sgl_page_size; sgl 866 drivers/scsi/esas2r/esas2r_init.c list_add_tail(&sgl->next_desc, &a->free_sg_list_head); sgl 868 drivers/scsi/esas2r/esas2r_init.c if (!esas2r_initmem_alloc(a, sgl, ESAS2R_SGL_ALIGN)) { sgl 224 drivers/scsi/esas2r/esas2r_io.c struct esas2r_mem_desc *sgl; sgl 231 drivers/scsi/esas2r/esas2r_io.c sgl = esas2r_alloc_sgl(a); sgl 233 drivers/scsi/esas2r/esas2r_io.c if (unlikely(sgl == NULL)) sgl 244 drivers/scsi/esas2r/esas2r_io.c memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); sgl 248 drivers/scsi/esas2r/esas2r_io.c (struct atto_vda_sge *)((u8 *)sgl->virt_addr + sgl 253 drivers/scsi/esas2r/esas2r_io.c (struct atto_vda_sge *)((u8 *)sgl->virt_addr sgl 260 drivers/scsi/esas2r/esas2r_io.c cpu_to_le64(sgl->phys_addr); sgl 302 drivers/scsi/esas2r/esas2r_io.c list_add(&sgl->next_desc, &rq->sg_table_head); sgl 376 drivers/scsi/esas2r/esas2r_io.c struct esas2r_mem_desc *sgl; sgl 449 drivers/scsi/esas2r/esas2r_io.c sgl = esas2r_alloc_sgl(a); sgl 451 drivers/scsi/esas2r/esas2r_io.c if (unlikely(sgl == NULL)) sgl 458 drivers/scsi/esas2r/esas2r_io.c list_add(&sgl->next_desc, &rq->sg_table_head); sgl 471 drivers/scsi/esas2r/esas2r_io.c cpu_to_le64(sgl->phys_addr); sgl 479 drivers/scsi/esas2r/esas2r_io.c (struct atto_physical_region_description *)sgl sgl 1063 drivers/scsi/fcoe/fcoe.c struct scatterlist *sgl, unsigned int sgc) sgl 1069 drivers/scsi/fcoe/fcoe.c xid, sgl, sgl 1085 drivers/scsi/fcoe/fcoe.c struct scatterlist *sgl, unsigned int sgc) sgl 1091 drivers/scsi/fcoe/fcoe.c sgl, sgc); sgl 582 drivers/scsi/isci/request.c stp_req->sgl.offset = 0; sgl 583 drivers/scsi/isci/request.c stp_req->sgl.set = SCU_SGL_ELEMENT_PAIR_A; sgl 587 drivers/scsi/isci/request.c stp_req->sgl.index = 0; sgl 590 drivers/scsi/isci/request.c stp_req->sgl.index = -1; sgl 1261 drivers/scsi/isci/request.c struct scu_sgl_element *sgl; sgl 1264 drivers/scsi/isci/request.c struct isci_stp_pio_sgl *pio_sgl = &stp_req->sgl; sgl 1268 drivers/scsi/isci/request.c sgl = NULL; sgl 1272 drivers/scsi/isci/request.c sgl = NULL; sgl 1275 drivers/scsi/isci/request.c sgl = &sgl_pair->B; sgl 1280 drivers/scsi/isci/request.c sgl = NULL; sgl 1285 drivers/scsi/isci/request.c sgl = &sgl_pair->A; sgl 1289 drivers/scsi/isci/request.c return sgl; sgl 1334 drivers/scsi/isci/request.c sgl_pair = to_sgl_element_pair(ireq, stp_req->sgl.index); sgl 1335 drivers/scsi/isci/request.c if (stp_req->sgl.set == SCU_SGL_ELEMENT_PAIR_A) sgl 1355 drivers/scsi/isci/request.c struct scu_sgl_element *sgl; sgl 1359 drivers/scsi/isci/request.c offset = stp_req->sgl.offset; sgl 1360 drivers/scsi/isci/request.c sgl_pair = to_sgl_element_pair(ireq, stp_req->sgl.index); sgl 1364 drivers/scsi/isci/request.c if (stp_req->sgl.set == SCU_SGL_ELEMENT_PAIR_A) { sgl 1365 drivers/scsi/isci/request.c sgl = &sgl_pair->A; sgl 1368 drivers/scsi/isci/request.c sgl = &sgl_pair->B; sgl 1382 drivers/scsi/isci/request.c sgl = pio_sgl_next(stp_req); sgl 1389 drivers/scsi/isci/request.c sgl->address_lower += stp_req->pio_len; sgl 1393 drivers/scsi/isci/request.c stp_req->sgl.offset = offset; sgl 2033 drivers/scsi/isci/request.c if (stp_req->sgl.index < 0) { sgl 79 drivers/scsi/isci/request.h } sgl; sgl 522 drivers/scsi/iscsi_tcp.c err = iscsi_sw_tcp_send_data_prep(conn, sdb->table.sgl, sgl 713 drivers/scsi/libiscsi_tcp.c sdb->table.sgl, sgl 3711 drivers/scsi/lpfc/lpfc_init.c sglq_entry->sgl = sglq_entry->virt; sgl 3712 drivers/scsi/lpfc/lpfc_init.c memset(sglq_entry->sgl, 0, LPFC_BPL_SIZE); sgl 3832 drivers/scsi/lpfc/lpfc_init.c sglq_entry->sgl = sglq_entry->virt; sgl 3833 drivers/scsi/lpfc/lpfc_init.c memset(sglq_entry->sgl, 0, sgl 772 drivers/scsi/lpfc/lpfc_nvme.c struct sli4_sge *sgl; sgl 791 drivers/scsi/lpfc/lpfc_nvme.c sgl = lpfc_ncmd->dma_sgl; sgl 792 drivers/scsi/lpfc/lpfc_nvme.c sgl->sge_len = cpu_to_le32(nCmd->cmdlen); sgl 794 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_hi = 0; sgl 795 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_lo = 0; sgl 834 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->cmddma)); sgl 835 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->cmddma)); sgl 840 drivers/scsi/lpfc/lpfc_nvme.c wqe->generic.bde.addrHigh = sgl->addr_hi; sgl 841 drivers/scsi/lpfc/lpfc_nvme.c wqe->generic.bde.addrLow = sgl->addr_lo; sgl 848 drivers/scsi/lpfc/lpfc_nvme.c sgl++; sgl 851 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->rspdma)); sgl 852 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->rspdma)); sgl 853 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 855 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 857 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 858 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 859 drivers/scsi/lpfc/lpfc_nvme.c sgl->sge_len = cpu_to_le32(nCmd->rsplen); sgl 1313 drivers/scsi/lpfc/lpfc_nvme.c struct sli4_sge *sgl = lpfc_ncmd->dma_sgl; sgl 1337 drivers/scsi/lpfc/lpfc_nvme.c sgl += 2; sgl 1339 drivers/scsi/lpfc/lpfc_nvme.c first_data_sgl = sgl; sgl 1372 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = 0; sgl 1374 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 1375 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_type, sgl, sgl 1378 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 1385 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_type, sgl, sgl 1395 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 1397 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 1401 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_type, sgl, sgl 1406 drivers/scsi/lpfc/lpfc_nvme.c if (!(bf_get(lpfc_sli4_sge_type, sgl) & sgl 1409 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 1413 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_lo = cpu_to_le32( sgl 1415 drivers/scsi/lpfc/lpfc_nvme.c sgl->addr_hi = cpu_to_le32( sgl 1418 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_offset, sgl, dma_offset); sgl 1419 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 1420 drivers/scsi/lpfc/lpfc_nvme.c sgl->sge_len = cpu_to_le32(dma_len); sgl 1425 drivers/scsi/lpfc/lpfc_nvme.c sgl++; sgl 1429 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 1431 drivers/scsi/lpfc/lpfc_nvme.c sgl->sge_len = cpu_to_le32( sgl 1434 drivers/scsi/lpfc/lpfc_nvme.c sgl = (struct sli4_sge *)sgl_xtra->dma_sgl; sgl 2022 drivers/scsi/lpfc/lpfc_nvme.c struct sli4_sge *sgl; sgl 2044 drivers/scsi/lpfc/lpfc_nvme.c sgl = lpfc_ncmd->dma_sgl; sgl 2045 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP); sgl 2046 drivers/scsi/lpfc/lpfc_nvme.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 2047 drivers/scsi/lpfc/lpfc_nvme.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2592 drivers/scsi/lpfc/lpfc_nvmet.c struct sli4_sge *sgl; sgl 2662 drivers/scsi/lpfc/lpfc_nvmet.c sgl = (struct sli4_sge *)ctxp->ctxbuf->sglq->sgl; sgl 2715 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_hi = 0; sgl 2716 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_lo = 0; sgl 2717 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2718 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP); sgl 2719 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2720 drivers/scsi/lpfc/lpfc_nvmet.c sgl->sge_len = 0; sgl 2721 drivers/scsi/lpfc/lpfc_nvmet.c sgl++; sgl 2722 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_hi = 0; sgl 2723 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_lo = 0; sgl 2724 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2725 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP); sgl 2726 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2727 drivers/scsi/lpfc/lpfc_nvmet.c sgl->sge_len = 0; sgl 2728 drivers/scsi/lpfc/lpfc_nvmet.c sgl++; sgl 2816 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_hi = putPaddrHigh(physaddr); sgl 2817 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_lo = putPaddrLow(physaddr); sgl 2818 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2819 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA); sgl 2820 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2821 drivers/scsi/lpfc/lpfc_nvmet.c sgl->sge_len = cpu_to_le32(TXRDY_PAYLOAD_LEN); sgl 2822 drivers/scsi/lpfc/lpfc_nvmet.c sgl++; sgl 2823 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_hi = 0; sgl 2824 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_lo = 0; sgl 2825 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2826 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP); sgl 2827 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2828 drivers/scsi/lpfc/lpfc_nvmet.c sgl->sge_len = 0; sgl 2829 drivers/scsi/lpfc/lpfc_nvmet.c sgl++; sgl 2887 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2906 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_hi = putPaddrHigh(physaddr); sgl 2907 drivers/scsi/lpfc/lpfc_nvmet.c sgl->addr_lo = putPaddrLow(physaddr); sgl 2908 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = 0; sgl 2909 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA); sgl 2910 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_offset, sgl, ctxp->offset); sgl 2912 drivers/scsi/lpfc/lpfc_nvmet.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 2913 drivers/scsi/lpfc/lpfc_nvmet.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2914 drivers/scsi/lpfc/lpfc_nvmet.c sgl->sge_len = cpu_to_le32(cnt); sgl 2919 drivers/scsi/lpfc/lpfc_nvmet.c bde->addrLow = sgl->addr_lo; sgl 2920 drivers/scsi/lpfc/lpfc_nvmet.c bde->addrHigh = sgl->addr_hi; sgl 2922 drivers/scsi/lpfc/lpfc_nvmet.c le32_to_cpu(sgl->sge_len); sgl 2929 drivers/scsi/lpfc/lpfc_nvmet.c sgl++; sgl 126 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 127 drivers/scsi/lpfc/lpfc_scsi.c if (sgl) { sgl 128 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 129 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 130 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 131 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 635 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl; sgl 687 drivers/scsi/lpfc/lpfc_scsi.c sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 689 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_cmd)); sgl 690 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_cmd)); sgl 691 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 692 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 693 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 694 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(sizeof(struct fcp_cmnd)); sgl 695 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 699 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_rsp)); sgl 700 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_rsp)); sgl 701 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 702 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 703 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 704 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(sizeof(struct fcp_rsp)); sgl 1947 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl, int datasegcnt, sgl 1984 drivers/scsi/lpfc/lpfc_scsi.c diseed = (struct sli4_sge_diseed *) sgl; sgl 1986 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DISEED); sgl 2021 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2028 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 2034 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_LSP); sgl 2042 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 2044 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 2048 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA); sgl 2051 drivers/scsi/lpfc/lpfc_scsi.c if (!(bf_get(lpfc_sli4_sge_type, sgl) & LPFC_SGE_TYPE_LSP)) { sgl 2053 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 2056 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow(physaddr)); sgl 2057 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh(physaddr)); sgl 2059 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_offset, sgl, dma_offset); sgl 2060 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2061 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(dma_len); sgl 2066 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2071 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2072 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(phba->cfg_sg_dma_buf_size); sgl 2074 drivers/scsi/lpfc/lpfc_scsi.c sgl = (struct sli4_sge *)sgl_xtra->dma_sgl; sgl 2127 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl, int datacnt, int protcnt, sgl 2192 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 2195 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_LSP); sgl 2202 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 2204 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 2208 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2209 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(phba->cfg_sg_dma_buf_size); sgl 2211 drivers/scsi/lpfc/lpfc_scsi.c sgl = (struct sli4_sge *)sgl_xtra->dma_sgl; sgl 2216 drivers/scsi/lpfc/lpfc_scsi.c diseed = (struct sli4_sge_diseed *) sgl; sgl 2218 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DISEED); sgl 2262 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2273 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 2274 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DIF); sgl 2275 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = le32_to_cpu(putPaddrHigh(protphysaddr)); sgl 2276 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = le32_to_cpu(putPaddrLow(protphysaddr)); sgl 2277 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2278 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = 0; sgl 2284 drivers/scsi/lpfc/lpfc_scsi.c if ((sgl->addr_lo & 0xfff) + protgroup_len > 0x1000) { sgl 2285 drivers/scsi/lpfc/lpfc_scsi.c protgroup_remainder = 0x1000 - (sgl->addr_lo & 0xfff); sgl 2300 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2317 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 2320 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, sgl 2329 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32( sgl 2331 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32( sgl 2335 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 2336 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32( sgl 2339 drivers/scsi/lpfc/lpfc_scsi.c sgl = (struct sli4_sge *)sgl_xtra->dma_sgl; sgl 2364 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 2365 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 2367 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 2369 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 2370 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_offset, sgl, dma_offset); sgl 2371 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, sgl 2374 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(dma_len); sgl 2381 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2389 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 2404 drivers/scsi/lpfc/lpfc_scsi.c sgl--; sgl 2405 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 3049 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 3078 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 3080 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 3081 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 3082 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 3083 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 3084 drivers/scsi/lpfc/lpfc_scsi.c first_data_sgl = sgl; sgl 3113 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = 0; sgl 3115 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 3116 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, sgl 3119 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 3126 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, sgl 3137 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 3139 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 3143 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_type, sgl, sgl 3148 drivers/scsi/lpfc/lpfc_scsi.c if (!(bf_get(lpfc_sli4_sge_type, sgl) & sgl 3151 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 3155 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_lo = cpu_to_le32(putPaddrLow( sgl 3157 drivers/scsi/lpfc/lpfc_scsi.c sgl->addr_hi = cpu_to_le32(putPaddrHigh( sgl 3160 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_offset, sgl, dma_offset); sgl 3161 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 3162 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32(dma_len); sgl 3167 drivers/scsi/lpfc/lpfc_scsi.c sgl++; sgl 3171 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 3172 drivers/scsi/lpfc/lpfc_scsi.c sgl->sge_len = cpu_to_le32( sgl 3175 drivers/scsi/lpfc/lpfc_scsi.c sgl = (struct sli4_sge *)sgl_xtra->dma_sgl; sgl 3200 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 3202 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 3203 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 3204 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 3261 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_sge *sgl = (struct sli4_sge *)(lpfc_cmd->dma_sgl); sgl 3287 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 3289 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 3290 drivers/scsi/lpfc/lpfc_scsi.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 3291 drivers/scsi/lpfc/lpfc_scsi.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 3293 drivers/scsi/lpfc/lpfc_scsi.c sgl += 1; sgl 3316 drivers/scsi/lpfc/lpfc_scsi.c num_sge = lpfc_bg_setup_sgl(phba, scsi_cmnd, sgl, sgl 3352 drivers/scsi/lpfc/lpfc_scsi.c num_sge = lpfc_bg_setup_sgl_prot(phba, scsi_cmnd, sgl, sgl 9182 drivers/scsi/lpfc/lpfc_sli.c struct sli4_sge *sgl = NULL; sgl 9193 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct sli4_sge *)sglq->sgl; sgl 9215 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_hi = bpl->addrHigh; sgl 9216 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_lo = bpl->addrLow; sgl 9218 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 9220 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 9222 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 9227 drivers/scsi/lpfc/lpfc_sli.c sgl->sge_len = cpu_to_le32(bde.tus.f.bdeSize); sgl 9239 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_offset, sgl, offset); sgl 9240 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_type, sgl, sgl 9244 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 9246 drivers/scsi/lpfc/lpfc_sli.c sgl++; sgl 9253 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_hi = sgl 9255 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_lo = sgl 9257 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 9258 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 9259 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 9260 drivers/scsi/lpfc/lpfc_sli.c sgl->sge_len = sgl 9496 drivers/scsi/lpfc/lpfc_sli.c struct sli4_sge *sgl; sgl 9503 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 9509 drivers/scsi/lpfc/lpfc_sli.c wqe->generic.bde.tus.f.bdeSize = sgl->sge_len; sgl 9560 drivers/scsi/lpfc/lpfc_sli.c struct sli4_sge *sgl; sgl 9567 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 9573 drivers/scsi/lpfc/lpfc_sli.c wqe->generic.bde.tus.f.bdeSize = sgl->sge_len; sgl 9617 drivers/scsi/lpfc/lpfc_sli.c struct sli4_sge *sgl; sgl 9624 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; sgl 9630 drivers/scsi/lpfc/lpfc_sli.c wqe->generic.bde.tus.f.bdeSize = sgl->sge_len; sgl 16700 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_mbx_post_uembed_sgl_page1 *sgl; sgl 16739 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct lpfc_mbx_post_uembed_sgl_page1 *)viraddr; sgl 16740 drivers/scsi/lpfc/lpfc_sli.c sgl_pg_pairs = &sgl->sgl_pg_pairs; sgl 16762 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_post_sgl_pages_xri, sgl, xritag_start); sgl 16763 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_post_sgl_pages_xricnt, sgl, post_cnt); sgl 16764 drivers/scsi/lpfc/lpfc_sli.c sgl->word0 = cpu_to_le32(sgl->word0); sgl 16772 drivers/scsi/lpfc/lpfc_sli.c shdr = (union lpfc_sli4_cfg_shdr *) &sgl->cfg_shdr; sgl 16803 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_mbx_post_uembed_sgl_page1 *sgl; sgl 16849 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct lpfc_mbx_post_uembed_sgl_page1 *)viraddr; sgl 16850 drivers/scsi/lpfc/lpfc_sli.c sgl_pg_pairs = &sgl->sgl_pg_pairs; sgl 16874 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_post_sgl_pages_xri, sgl, xritag_start); sgl 16875 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_post_sgl_pages_xricnt, sgl, pg_pairs); sgl 16877 drivers/scsi/lpfc/lpfc_sli.c sgl->word0 = cpu_to_le32(sgl->word0); sgl 16885 drivers/scsi/lpfc/lpfc_sli.c shdr = (union lpfc_sli4_cfg_shdr *)&sgl->cfg_shdr; sgl 19711 drivers/scsi/lpfc/lpfc_sli.c struct sli4_sge *sgl = NULL; sgl 19723 drivers/scsi/lpfc/lpfc_sli.c sgl = (struct sli4_sge *)sglq->sgl; sgl 19747 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_hi = bpl->addrHigh; sgl 19748 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_lo = bpl->addrLow; sgl 19750 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 19752 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 19754 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 0); sgl 19759 drivers/scsi/lpfc/lpfc_sli.c sgl->sge_len = cpu_to_le32(bde.tus.f.bdeSize); sgl 19772 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_offset, sgl, offset); sgl 19773 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_type, sgl, sgl 19778 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_offset, sgl, 0); sgl 19779 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_type, sgl, sgl 19784 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_type, sgl, sgl 19790 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_offset, sgl, offset); sgl 19793 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 19795 drivers/scsi/lpfc/lpfc_sli.c sgl++; sgl 19802 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_hi = cpu_to_le32(wqe->gen_req.bde.addrHigh); sgl 19803 drivers/scsi/lpfc/lpfc_sli.c sgl->addr_lo = cpu_to_le32(wqe->gen_req.bde.addrLow); sgl 19804 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = le32_to_cpu(sgl->word2); sgl 19805 drivers/scsi/lpfc/lpfc_sli.c bf_set(lpfc_sli4_sge_last, sgl, 1); sgl 19806 drivers/scsi/lpfc/lpfc_sli.c sgl->word2 = cpu_to_le32(sgl->word2); sgl 19807 drivers/scsi/lpfc/lpfc_sli.c sgl->sge_len = cpu_to_le32(wqe->gen_req.bde.tus.f.bdeSize); sgl 960 drivers/scsi/lpfc/lpfc_sli4.h struct sli4_sge *sgl; /* pre-assigned SGL */ sgl 1415 drivers/scsi/megaraid.c struct scatterlist *sgl; sgl 1549 drivers/scsi/megaraid.c sgl = scsi_sglist(cmd); sgl 1550 drivers/scsi/megaraid.c if( sg_page(sgl) ) { sgl 1551 drivers/scsi/megaraid.c c = *(unsigned char *) sg_virt(&sgl[0]); sgl 1753 drivers/scsi/megaraid.c scb->sgl[idx].address = sg_dma_address(sg); sgl 1754 drivers/scsi/megaraid.c *len += scb->sgl[idx].length = sg_dma_len(sg); sgl 2883 drivers/scsi/megaraid.c scb->sgl = NULL; sgl 2898 drivers/scsi/megaraid.c scb->sgl = (mega_sglist *)scb->sgl64; sgl 2900 drivers/scsi/megaraid.c if( !scb->sgl ) { sgl 198 drivers/scsi/megaraid.h mega_sglist *sgl; sgl 1337 drivers/scsi/megaraid/megaraid_mbox.c struct scatterlist *sgl; sgl 1356 drivers/scsi/megaraid/megaraid_mbox.c scsi_for_each_sg(scp, sgl, sgcnt, i) { sgl 1357 drivers/scsi/megaraid/megaraid_mbox.c ccb->sgl64[i].address = sg_dma_address(sgl); sgl 1358 drivers/scsi/megaraid/megaraid_mbox.c ccb->sgl64[i].length = sg_dma_len(sgl); sgl 1540 drivers/scsi/megaraid/megaraid_mbox.c struct scatterlist *sgl; sgl 1543 drivers/scsi/megaraid/megaraid_mbox.c sgl = scsi_sglist(scp); sgl 1544 drivers/scsi/megaraid/megaraid_mbox.c if (sg_page(sgl)) { sgl 1545 drivers/scsi/megaraid/megaraid_mbox.c vaddr = (caddr_t) sg_virt(&sgl[0]); sgl 2176 drivers/scsi/megaraid/megaraid_mbox.c struct scatterlist *sgl; sgl 2260 drivers/scsi/megaraid/megaraid_mbox.c sgl = scsi_sglist(scp); sgl 2261 drivers/scsi/megaraid/megaraid_mbox.c if (sg_page(sgl)) { sgl 2262 drivers/scsi/megaraid/megaraid_mbox.c c = *(unsigned char *) sg_virt(&sgl[0]); sgl 1863 drivers/scsi/megaraid/megaraid_sas.h union megasas_sgl sgl; /*28h */ sgl 1890 drivers/scsi/megaraid/megaraid_sas.h union megasas_sgl sgl; /*30h */ sgl 1917 drivers/scsi/megaraid/megaraid_sas.h union megasas_sgl sgl; /*28h */ sgl 1969 drivers/scsi/megaraid/megaraid_sas.h } sgl; sgl 1998 drivers/scsi/megaraid/megaraid_sas.h } sgl; sgl 2594 drivers/scsi/megaraid/megaraid_sas.h struct iovec sgl[MAX_IOCTL_SGE]; sgl 2617 drivers/scsi/megaraid/megaraid_sas.h struct compat_iovec sgl[MAX_IOCTL_SGE]; sgl 279 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge64[0].phys_addr = cpu_to_le64(dma_addr); sgl 280 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge64[0].length = cpu_to_le32(dma_len); sgl 284 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].phys_addr = sgl 286 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].length = cpu_to_le32(dma_len); sgl 1438 drivers/scsi/megaraid/megaraid_sas_base.c &pthru->sgl); sgl 1442 drivers/scsi/megaraid/megaraid_sas_base.c &pthru->sgl); sgl 1445 drivers/scsi/megaraid/megaraid_sas_base.c &pthru->sgl); sgl 1581 drivers/scsi/megaraid/megaraid_sas_base.c &ldio->sgl); sgl 1584 drivers/scsi/megaraid/megaraid_sas_base.c ldio->sge_count = megasas_make_sgl64(instance, scp, &ldio->sgl); sgl 1586 drivers/scsi/megaraid/megaraid_sas_base.c ldio->sge_count = megasas_make_sgl32(instance, scp, &ldio->sgl); sgl 1671 drivers/scsi/megaraid/megaraid_sas_base.c mfi_sgl = &ldio->sgl; sgl 1680 drivers/scsi/megaraid/megaraid_sas_base.c mfi_sgl = &pthru->sgl; sgl 2348 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].phys_addr = sgl 2351 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].phys_addr = sgl 2354 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].length = cpu_to_le32( sgl 2455 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].phys_addr = sgl 2458 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].phys_addr = sgl 2461 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->sgl.sge32[0].length = cpu_to_le32((MAX_LOGICAL_DRIVES + 1) * sgl 8119 drivers/scsi/megaraid/megaraid_sas_base.c if (!ioc->sgl[i].iov_len) sgl 8123 drivers/scsi/megaraid/megaraid_sas_base.c ioc->sgl[i].iov_len, sgl 8138 drivers/scsi/megaraid/megaraid_sas_base.c kern_sge64[i].length = cpu_to_le32(ioc->sgl[i].iov_len); sgl 8141 drivers/scsi/megaraid/megaraid_sas_base.c kern_sge32[i].length = cpu_to_le32(ioc->sgl[i].iov_len); sgl 8148 drivers/scsi/megaraid/megaraid_sas_base.c if (copy_from_user(kbuff_arr[i], ioc->sgl[i].iov_base, sgl 8149 drivers/scsi/megaraid/megaraid_sas_base.c (u32) (ioc->sgl[i].iov_len))) { sgl 8196 drivers/scsi/megaraid/megaraid_sas_base.c if (copy_to_user(ioc->sgl[i].iov_base, kbuff_arr[i], sgl 8197 drivers/scsi/megaraid/megaraid_sas_base.c ioc->sgl[i].iov_len)) { sgl 8417 drivers/scsi/megaraid/megaraid_sas_base.c if (get_user(ptr, &cioc->sgl[i].iov_base) || sgl 8418 drivers/scsi/megaraid/megaraid_sas_base.c put_user(compat_ptr(ptr), &ioc->sgl[i].iov_base) || sgl 8419 drivers/scsi/megaraid/megaraid_sas_base.c copy_in_user(&ioc->sgl[i].iov_len, sgl 8420 drivers/scsi/megaraid/megaraid_sas_base.c &cioc->sgl[i].iov_len, sizeof(compat_size_t))) sgl 1274 drivers/scsi/myrb.c struct scatterlist *sgl; sgl 1320 drivers/scsi/myrb.c sgl = scsi_sglist(scmd); sgl 1321 drivers/scsi/myrb.c dcdb->dma_addr = sg_dma_address(sgl); sgl 1322 drivers/scsi/myrb.c if (sg_dma_len(sgl) > USHRT_MAX) { sgl 1323 drivers/scsi/myrb.c dcdb->xfer_len_lo = sg_dma_len(sgl) & 0xffff; sgl 1324 drivers/scsi/myrb.c dcdb->xfer_len_hi4 = sg_dma_len(sgl) >> 16; sgl 1326 drivers/scsi/myrb.c dcdb->xfer_len_lo = sg_dma_len(sgl); sgl 1429 drivers/scsi/myrb.c struct scatterlist *sgl; sgl 1587 drivers/scsi/myrb.c sgl = scsi_sglist(scmd); sgl 1596 drivers/scsi/myrb.c mbox->type5.addr = (u32)sg_dma_address(sgl); sgl 1606 drivers/scsi/myrb.c cmd_blk->sgl = hw_sgl; sgl 1620 drivers/scsi/myrb.c scsi_for_each_sg(scmd, sgl, nsge, i) { sgl 1621 drivers/scsi/myrb.c hw_sgl->sge_addr = (u32)sg_dma_address(sgl); sgl 1622 drivers/scsi/myrb.c hw_sgl->sge_count = (u32)sg_dma_len(sgl); sgl 2341 drivers/scsi/myrb.c if (cmd_blk->sgl) { sgl 2342 drivers/scsi/myrb.c dma_pool_free(cb->sg_pool, cmd_blk->sgl, cmd_blk->sgl_addr); sgl 2343 drivers/scsi/myrb.c cmd_blk->sgl = NULL; sgl 688 drivers/scsi/myrb.h struct myrb_sge *sgl; sgl 164 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 187 drivers/scsi/myrs.c sgl = &mbox->ctlr_info.dma_addr; sgl 188 drivers/scsi/myrs.c sgl->sge[0].sge_addr = ctlr_info_addr; sgl 189 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ctlr_info.dma_size; sgl 227 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 246 drivers/scsi/myrs.c sgl = &mbox->ldev_info.dma_addr; sgl 247 drivers/scsi/myrs.c sgl->sge[0].sge_addr = ldev_info_addr; sgl 248 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ldev_info.dma_size; sgl 314 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 334 drivers/scsi/myrs.c sgl = &mbox->pdev_info.dma_addr; sgl 335 drivers/scsi/myrs.c sgl->sge[0].sge_addr = pdev_info_addr; sgl 336 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->pdev_info.dma_size; sgl 384 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 405 drivers/scsi/myrs.c sgl = &mbox->pdev_info.dma_addr; sgl 406 drivers/scsi/myrs.c sgl->sge[0].sge_addr = devmap_addr; sgl 407 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->pdev_info.dma_size; sgl 427 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 441 drivers/scsi/myrs.c sgl = &mbox->get_event.dma_addr; sgl 442 drivers/scsi/myrs.c sgl->sge[0].sge_addr = event_addr; sgl 443 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->get_event.dma_size; sgl 459 drivers/scsi/myrs.c union myrs_sgl *sgl; sgl 469 drivers/scsi/myrs.c sgl = &mbox->common.dma_addr; sgl 470 drivers/scsi/myrs.c sgl->sge[0].sge_addr = cs->fwstat_addr; sgl 471 drivers/scsi/myrs.c sgl->sge[0].sge_count = mbox->ctlr_info.dma_size; sgl 1593 drivers/scsi/myrs.c struct scatterlist *sgl; sgl 1724 drivers/scsi/myrs.c sgl = scsi_sglist(scmd); sgl 1725 drivers/scsi/myrs.c hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl); sgl 1726 drivers/scsi/myrs.c hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl); sgl 1750 drivers/scsi/myrs.c cmd_blk->sgl = hw_sgl; sgl 1761 drivers/scsi/myrs.c scsi_for_each_sg(scmd, sgl, nsge, i) { sgl 1768 drivers/scsi/myrs.c hw_sgl->sge_addr = (u64)sg_dma_address(sgl); sgl 1769 drivers/scsi/myrs.c hw_sgl->sge_count = (u64)sg_dma_len(sgl); sgl 2080 drivers/scsi/myrs.c if (cmd_blk->sgl) { sgl 2081 drivers/scsi/myrs.c dma_pool_free(cs->sg_pool, cmd_blk->sgl, sgl 2083 drivers/scsi/myrs.c cmd_blk->sgl = NULL; sgl 876 drivers/scsi/myrs.h struct myrs_sge *sgl; sgl 4936 drivers/scsi/pm8001/pm8001_hwi.c payload.len = info->sgl.im_len.len ; sgl 4938 drivers/scsi/pm8001/pm8001_hwi.c cpu_to_le32(lower_32_bits(le64_to_cpu(info->sgl.addr))); sgl 4940 drivers/scsi/pm8001/pm8001_hwi.c cpu_to_le32(upper_32_bits(le64_to_cpu(info->sgl.addr))); sgl 4964 drivers/scsi/pm8001/pm8001_hwi.c flash_update_info.sgl.addr = cpu_to_le64(phys_addr); sgl 4965 drivers/scsi/pm8001/pm8001_hwi.c flash_update_info.sgl.im_len.len = cpu_to_le32(fw_control->len); sgl 4966 drivers/scsi/pm8001/pm8001_hwi.c flash_update_info.sgl.im_len.e = 0; sgl 597 drivers/scsi/pm8001/pm8001_sas.h struct pm8001_prd sgl; sgl 34 drivers/scsi/qedf/drv_scsi_fw_funcs.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo); sgl 36 drivers/scsi/qedf/drv_scsi_fw_funcs.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi); sgl 38 drivers/scsi/qedf/drv_scsi_fw_funcs.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len); sgl 12 drivers/scsi/qedf/drv_scsi_fw_funcs.h struct scsi_sge *sgl; sgl 644 drivers/scsi/qedf/qedf_io.c io_req->sgl_task_params->sgl = bd_tbl->bd_tbl; sgl 737 drivers/scsi/qedf/qedf_io.c tx_sgl_task_params.sgl = mp_req->mp_req_bd; sgl 746 drivers/scsi/qedf/qedf_io.c rx_sgl_task_params.sgl = mp_req->mp_resp_bd; sgl 1062 drivers/scsi/qedi/qedi_fw.c tx_sgl_task_params.sgl = sgl 1071 drivers/scsi/qedi/qedi_fw.c rx_sgl_task_params.sgl = sgl 1647 drivers/scsi/qedi/qedi_fw.c tx_sgl_task_params.sgl = sgl 1656 drivers/scsi/qedi/qedi_fw.c rx_sgl_task_params.sgl = sgl 1763 drivers/scsi/qedi/qedi_fw.c tx_sgl_task_params.sgl = sgl 1772 drivers/scsi/qedi/qedi_fw.c rx_sgl_task_params.sgl = sgl 2110 drivers/scsi/qedi/qedi_fw.c tx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl; sgl 2120 drivers/scsi/qedi/qedi_fw.c rx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl; sgl 43 drivers/scsi/qedi/qedi_fw_api.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo); sgl 45 drivers/scsi/qedi/qedi_fw_api.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi); sgl 47 drivers/scsi/qedi/qedi_fw_api.c val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len); sgl 16 drivers/scsi/qedi/qedi_fw_scsi.h struct scsi_sge *sgl; sgl 1008 drivers/scsi/qla2xxx/qla_iocb.c struct scatterlist *sg, *sgl; sgl 1016 drivers/scsi/qla2xxx/qla_iocb.c sgl = scsi_sglist(cmd); sgl 1018 drivers/scsi/qla2xxx/qla_iocb.c sgl = tc->sg; sgl 1025 drivers/scsi/qla2xxx/qla_iocb.c for_each_sg(sgl, sg, tot_dsds, i) { sgl 1085 drivers/scsi/qla2xxx/qla_iocb.c struct scatterlist *sg, *sgl; sgl 1098 drivers/scsi/qla2xxx/qla_iocb.c sgl = scsi_prot_sglist(cmd); sgl 1107 drivers/scsi/qla2xxx/qla_iocb.c sgl = tc->prot_sg; sgl 1120 drivers/scsi/qla2xxx/qla_iocb.c for_each_sg(sgl, sg, tot_dsds, i) { sgl 1162 drivers/scsi/qla2xxx/qla_iocb.c for_each_sg(sgl, sg, tot_dsds, i) { sgl 1219 drivers/scsi/qla2xxx/qla_iocb.c sg_pcopy_to_buffer(sgl, sg_nents(sgl), sgl 1321 drivers/scsi/qla2xxx/qla_iocb.c for_each_sg(sgl, sg, tot_dsds, i) { sgl 368 drivers/scsi/qla2xxx/qla_nvme.c struct scatterlist *sgl, *sg; sgl 470 drivers/scsi/qla2xxx/qla_nvme.c sgl = fd->first_sgl; sgl 473 drivers/scsi/qla2xxx/qla_nvme.c for_each_sg(sgl, sg, tot_dsds, i) { sgl 1013 drivers/scsi/scsi_debug.c act_len = sg_copy_from_buffer(sdb->table.sgl, sdb->table.nents, sgl 1037 drivers/scsi/scsi_debug.c act_len = sg_pcopy_from_buffer(sdb->table.sgl, sdb->table.nents, sgl 2503 drivers/scsi/scsi_debug.c ret = sg_copy_buffer(sdb->table.sgl, sdb->table.nents, sgl 2510 drivers/scsi/scsi_debug.c ret += sg_copy_buffer(sdb->table.sgl, sdb->table.nents, sgl 988 drivers/scsi/scsi_error.c scmd->sdb.table.sgl = &ses->sense_sgl; sgl 994 drivers/scsi/scsi_lib.c blk_rq_nr_phys_segments(req), sdb->table.sgl, sgl 1002 drivers/scsi/scsi_lib.c count = blk_rq_map_sg(req->q, req, sdb->table.sgl); sgl 1049 drivers/scsi/scsi_lib.c prot_sdb->table.sgl, sgl 1056 drivers/scsi/scsi_lib.c prot_sdb->table.sgl); sgl 1593 drivers/scsi/scsi_lib.c cmd->sdb.table.sgl = sg; sgl 1598 drivers/scsi/scsi_lib.c cmd->prot_sdb->table.sgl = sgl 2856 drivers/scsi/scsi_lib.c void *scsi_kmap_atomic_sg(struct scatterlist *sgl, int sg_count, sgl 2866 drivers/scsi/scsi_lib.c for_each_sg(sgl, sg, sg_count, i) { sgl 1565 drivers/scsi/storvsc_drv.c struct scatterlist *sgl; sgl 1638 drivers/scsi/storvsc_drv.c sgl = (struct scatterlist *)scsi_sglist(scmnd); sgl 1656 drivers/scsi/storvsc_drv.c payload->range.offset = sgl[0].offset; sgl 1658 drivers/scsi/storvsc_drv.c cur_sgl = sgl; sgl 432 drivers/scsi/virtio_scsi.c sgs[out_num++] = out->sgl; sgl 444 drivers/scsi/virtio_scsi.c sgs[out_num + in_num++] = in->sgl; sgl 61 drivers/scsi/vmw_pvscsi.c struct pvscsi_sg_list *sgl; sgl 337 drivers/scsi/vmw_pvscsi.c sge = &ctx->sgl->sge[0]; sgl 376 drivers/scsi/vmw_pvscsi.c ctx->sgl, SGL_SIZE, DMA_TO_DEVICE); sgl 1195 drivers/scsi/vmw_pvscsi.c free_pages((unsigned long)ctx->sgl, get_order(SGL_SIZE)); sgl 1261 drivers/scsi/vmw_pvscsi.c ctx->sgl = (void *)__get_free_pages(GFP_KERNEL, sgl 1264 drivers/scsi/vmw_pvscsi.c BUG_ON(!IS_ALIGNED(((unsigned long)ctx->sgl), PAGE_SIZE)); sgl 1265 drivers/scsi/vmw_pvscsi.c if (!ctx->sgl) { sgl 1267 drivers/scsi/vmw_pvscsi.c free_pages((unsigned long)ctx->sgl, sgl 1269 drivers/scsi/vmw_pvscsi.c ctx->sgl = NULL; sgl 230 drivers/spi/spi-at91-usart.c xfer->rx_sg.sgl, sgl 239 drivers/spi/spi-at91-usart.c xfer->tx_sg.sgl, sgl 766 drivers/spi/spi-atmel.c xfer->rx_sg.sgl, sgl 785 drivers/spi/spi-atmel.c xfer->tx_sg.sgl, sgl 484 drivers/spi/spi-bcm2835.c if (bs->tx_buf && !sg_is_last(&tfr->tx_sg.sgl[0])) sgl 485 drivers/spi/spi-bcm2835.c bs->tx_prologue = sg_dma_len(&tfr->tx_sg.sgl[0]) & 3; sgl 487 drivers/spi/spi-bcm2835.c if (bs->rx_buf && !sg_is_last(&tfr->rx_sg.sgl[0])) { sgl 488 drivers/spi/spi-bcm2835.c bs->rx_prologue = sg_dma_len(&tfr->rx_sg.sgl[0]) & 3; sgl 491 drivers/spi/spi-bcm2835.c if (!bs->tx_buf || sg_is_last(&tfr->tx_sg.sgl[0])) { sgl 496 drivers/spi/spi-bcm2835.c !(sg_dma_len(&tfr->tx_sg.sgl[0]) & ~3); sgl 518 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->rx_sg.sgl[0]), sgl 521 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->rx_sg.sgl[0]) += bs->rx_prologue; sgl 522 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue; sgl 544 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->tx_sg.sgl[0]) += bs->tx_prologue; sgl 545 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[0]) -= bs->tx_prologue; sgl 547 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[0]) = 0; sgl 548 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->tx_sg.sgl[1]) += 4; sgl 549 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[1]) -= 4; sgl 569 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue; sgl 570 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->rx_sg.sgl[0]) += bs->rx_prologue; sgl 577 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->tx_sg.sgl[0]) -= bs->tx_prologue; sgl 578 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[0]) += bs->tx_prologue; sgl 580 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[0]) = bs->tx_prologue - 4; sgl 581 drivers/spi/spi-bcm2835.c sg_dma_address(&tfr->tx_sg.sgl[1]) -= 4; sgl 582 drivers/spi/spi-bcm2835.c sg_dma_len(&tfr->tx_sg.sgl[1]) += 4; sgl 666 drivers/spi/spi-bcm2835.c struct scatterlist *sgl; sgl 678 drivers/spi/spi-bcm2835.c sgl = tfr->tx_sg.sgl; sgl 684 drivers/spi/spi-bcm2835.c sgl = tfr->rx_sg.sgl; sgl 688 drivers/spi/spi-bcm2835.c desc = dmaengine_prep_slave_sg(chan, sgl, nents, dir, flags); sgl 631 drivers/spi/spi-davinci.c t->rx_sg.sgl, t->rx_sg.nents, DMA_DEV_TO_MEM, sgl 642 drivers/spi/spi-davinci.c t->tx_sg.sgl = t->rx_sg.sgl; sgl 647 drivers/spi/spi-davinci.c t->tx_sg.sgl, t->tx_sg.nents, DMA_MEM_TO_DEV, sgl 160 drivers/spi/spi-dw-mid.c xfer->tx_sg.sgl, sgl 206 drivers/spi/spi-dw-mid.c xfer->rx_sg.sgl, sgl 331 drivers/spi/spi-ep93xx.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgl 351 drivers/spi/spi-ep93xx.c nents = dma_map_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgl 355 drivers/spi/spi-ep93xx.c txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction, sgl 358 drivers/spi/spi-ep93xx.c dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgl 387 drivers/spi/spi-ep93xx.c dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgl 580 drivers/spi/spi-fsl-lpspi.c rx->sgl, rx->nents, DMA_DEV_TO_MEM, sgl 592 drivers/spi/spi-fsl-lpspi.c tx->sgl, tx->nents, DMA_MEM_TO_DEV, sgl 335 drivers/spi/spi-img-spfi.c rxdesc = dmaengine_prep_slave_sg(spfi->rx_ch, xfer->rx_sg.sgl, sgl 359 drivers/spi/spi-img-spfi.c txdesc = dmaengine_prep_slave_sg(spfi->tx_ch, xfer->tx_sg.sgl, sgl 1341 drivers/spi/spi-imx.c struct scatterlist *last_sg = sg_last(rx->sgl, rx->nents); sgl 1372 drivers/spi/spi-imx.c rx->sgl, rx->nents, DMA_DEV_TO_MEM, sgl 1384 drivers/spi/spi-imx.c tx->sgl, tx->nents, DMA_MEM_TO_DEV, sgl 467 drivers/spi/spi-mt65xx.c mdata->tx_sgl = xfer->tx_sg.sgl; sgl 469 drivers/spi/spi-mt65xx.c mdata->rx_sgl = xfer->rx_sg.sgl; sgl 409 drivers/spi/spi-omap2-mcspi.c tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, sgl 491 drivers/spi/spi-omap2-mcspi.c ret = sg_split(xfer->rx_sg.sgl, xfer->rx_sg.nents, sgl 262 drivers/spi/spi-pic32-sqi.c struct scatterlist *sg, *sgl; sgl 276 drivers/spi/spi-pic32-sqi.c sgl = xfer->rx_sg.sgl; sgl 280 drivers/spi/spi-pic32-sqi.c sgl = xfer->tx_sg.sgl; sgl 296 drivers/spi/spi-pic32-sqi.c for_each_sg(sgl, sg, nents, i) { sgl 313 drivers/spi/spi-pic32.c xfer->rx_sg.sgl, sgl 323 drivers/spi/spi-pic32.c xfer->tx_sg.sgl, sgl 798 drivers/spi/spi-pl022.c dma_unmap_sg(pl022->dma_tx_channel->device->dev, pl022->sgt_tx.sgl, sgl 800 drivers/spi/spi-pl022.c dma_unmap_sg(pl022->dma_rx_channel->device->dev, pl022->sgt_rx.sgl, sgl 811 drivers/spi/spi-pl022.c BUG_ON(!pl022->sgt_rx.sgl); sgl 825 drivers/spi/spi-pl022.c pl022->sgt_rx.sgl, sgl 829 drivers/spi/spi-pl022.c for_each_sg(pl022->sgt_rx.sgl, sg, pl022->sgt_rx.nents, i) { sgl 839 drivers/spi/spi-pl022.c for_each_sg(pl022->sgt_tx.sgl, sg, pl022->sgt_tx.nents, i) { sgl 875 drivers/spi/spi-pl022.c for_each_sg(sgtab->sgl, sg, sgtab->nents, i) { sgl 896 drivers/spi/spi-pl022.c for_each_sg(sgtab->sgl, sg, sgtab->nents, i) { sgl 1050 drivers/spi/spi-pl022.c rx_sglen = dma_map_sg(rxchan->device->dev, pl022->sgt_rx.sgl, sgl 1055 drivers/spi/spi-pl022.c tx_sglen = dma_map_sg(txchan->device->dev, pl022->sgt_tx.sgl, sgl 1062 drivers/spi/spi-pl022.c pl022->sgt_rx.sgl, sgl 1070 drivers/spi/spi-pl022.c pl022->sgt_tx.sgl, sgl 1094 drivers/spi/spi-pl022.c dma_unmap_sg(txchan->device->dev, pl022->sgt_tx.sgl, sgl 1097 drivers/spi/spi-pl022.c dma_unmap_sg(rxchan->device->dev, pl022->sgt_rx.sgl, sgl 118 drivers/spi/spi-pxa2xx-dma.c return dmaengine_prep_slave_sg(chan, sgt->sgl, sgt->nents, dir, sgl 389 drivers/spi/spi-qup.c static int spi_qup_prep_sg(struct spi_master *master, struct scatterlist *sgl, sgl 404 drivers/spi/spi-qup.c desc = dmaengine_prep_slave_sg(chan, sgl, nents, dir, flags); sgl 425 drivers/spi/spi-qup.c static u32 spi_qup_sgl_get_nents_len(struct scatterlist *sgl, u32 max, sgl 431 drivers/spi/spi-qup.c for (sg = sgl; sg; sg = sg_next(sg)) { sgl 459 drivers/spi/spi-qup.c rx_sgl = xfer->rx_sg.sgl; sgl 460 drivers/spi/spi-qup.c tx_sgl = xfer->tx_sg.sgl; sgl 405 drivers/spi/spi-rockchip.c xfer->rx_sg.sgl, xfer->rx_sg.nents, sgl 427 drivers/spi/spi-rockchip.c xfer->tx_sg.sgl, xfer->tx_sg.nents, sgl 532 drivers/spi/spi-rspi.c desc_rx = dmaengine_prep_slave_sg(rspi->ctlr->dma_rx, rx->sgl, sgl 552 drivers/spi/spi-rspi.c desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl, sgl 294 drivers/spi/spi-s3c64xx.c desc = dmaengine_prep_slave_sg(dma->ch, sgt->sgl, sgt->nents, sgl 504 drivers/spi/spi-sprd.c desc = dmaengine_prep_slave_sg(dma_chan, sg->sgl, sg->nents, dir, flags); sgl 226 drivers/spi/spi-stm32-qspi.c desc = dmaengine_prep_slave_sg(dma_ch, sgt.sgl, sgt.nents, sgl 1293 drivers/spi/spi-stm32.c spi->dma_rx, xfer->rx_sg.sgl, sgl 1305 drivers/spi/spi-stm32.c spi->dma_tx, xfer->tx_sg.sgl, sgl 468 drivers/spi/spi-ti-qspi.c for_each_sg(rx_sg.sgl, sg, rx_sg.nents, i) { sgl 841 drivers/spi/spi.c sg = &sgt->sgl[0]; sgl 874 drivers/spi/spi.c ret = dma_map_sg(dev, sgt->sgl, sgt->nents, dir); sgl 891 drivers/spi/spi.c dma_unmap_sg(dev, sgt->sgl, sgt->orig_nents, dir); sgl 156 drivers/staging/android/ion/ion.c new_sg = new_table->sgl; sgl 157 drivers/staging/android/ion/ion.c for_each_sg(table->sgl, sg, table->nents, i) { sgl 230 drivers/staging/android/ion/ion.c if (!dma_map_sg(attachment->dev, table->sgl, table->nents, sgl 241 drivers/staging/android/ion/ion.c dma_unmap_sg(attachment->dev, table->sgl, table->nents, direction); sgl 312 drivers/staging/android/ion/ion.c dma_sync_sg_for_cpu(a->dev, a->table->sgl, a->table->nents, sgl 335 drivers/staging/android/ion/ion.c dma_sync_sg_for_device(a->dev, a->table->sgl, a->table->nents, sgl 70 drivers/staging/android/ion/ion_cma_heap.c sg_set_page(table->sgl, pages, size, 0); sgl 41 drivers/staging/android/ion/ion_heap.c for_each_sg(table->sgl, sg, table->nents, i) { sgl 74 drivers/staging/android/ion/ion_heap.c for_each_sg(table->sgl, sg, table->nents, i) { sgl 112 drivers/staging/android/ion/ion_heap.c static int ion_heap_sglist_zero(struct scatterlist *sgl, unsigned int nents, sgl 120 drivers/staging/android/ion/ion_heap.c for_each_sg_page(sgl, &piter, nents, 0) { sgl 145 drivers/staging/android/ion/ion_heap.c return ion_heap_sglist_zero(table->sgl, table->nents, pgprot); sgl 134 drivers/staging/android/ion/ion_system_heap.c sg = table->sgl; sgl 165 drivers/staging/android/ion/ion_system_heap.c for_each_sg(table->sgl, sg, table->nents, i) sgl 315 drivers/staging/android/ion/ion_system_heap.c sg_set_page(table->sgl, page, len, 0); sgl 333 drivers/staging/android/ion/ion_system_heap.c struct page *page = sg_page(table->sgl); sgl 95 drivers/staging/kpc2000/kpc_dma/fileops.c acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir); sgl 102 drivers/staging/kpc2000/kpc_dma/fileops.c for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) { sgl 125 drivers/staging/kpc2000/kpc_dma/fileops.c for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) { sgl 189 drivers/staging/kpc2000/kpc_dma/fileops.c dma_unmap_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir); sgl 210 drivers/staging/kpc2000/kpc_dma/fileops.c BUG_ON(acd->sgt.sgl == NULL); sgl 220 drivers/staging/kpc2000/kpc_dma/fileops.c dma_unmap_sg(&acd->ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, acd->ldev->dir); sgl 318 drivers/staging/media/ipu3/ipu3-v4l2.c return imgu_dmamap_map_sg(imgu, sg->sgl, sg->nents, &buf->map); sgl 90 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = sg_dma_address(entry->sgt->sgl); sgl 128 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = sg_dma_address(sgt->sgl); sgl 39 drivers/staging/media/tegra-vde/iommu.c size = iommu_map_sg(vde->domain, addr, sgt->sgl, sgt->nents, sgl 514 drivers/staging/ralink-gdma/ralink-gdma.c struct dma_chan *c, struct scatterlist *sgl, sgl 530 drivers/staging/ralink-gdma/ralink-gdma.c for_each_sg(sgl, sg, sg_len, i) { sgl 133 drivers/target/iscsi/cxgbit/cxgbit_ddp.c struct scatterlist *sg = ttinfo->sgl; sgl 176 drivers/target/iscsi/cxgbit/cxgbit_ddp.c struct scatterlist *sgl = ttinfo->sgl; sgl 178 drivers/target/iscsi/cxgbit/cxgbit_ddp.c unsigned int sg_offset = sgl->offset; sgl 188 drivers/target/iscsi/cxgbit/cxgbit_ddp.c if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0) sgl 191 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ttinfo->nr_pages = (xferlen + sgl->offset + sgl 203 drivers/target/iscsi/cxgbit/cxgbit_ddp.c sgl->offset = 0; sgl 204 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); sgl 205 drivers/target/iscsi/cxgbit/cxgbit_ddp.c sgl->offset = sg_offset; sgl 212 drivers/target/iscsi/cxgbit/cxgbit_ddp.c cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset, sgl 218 drivers/target/iscsi/cxgbit/cxgbit_ddp.c dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); sgl 245 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ttinfo->sgl = cmd->se_cmd.t_data_sg; sgl 253 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ttinfo->sgl = NULL; sgl 270 drivers/target/iscsi/cxgbit/cxgbit_ddp.c if (ttinfo->sgl) { sgl 285 drivers/target/iscsi/cxgbit/cxgbit_ddp.c dma_unmap_sg(&ppm->pdev->dev, ttinfo->sgl, sgl 261 drivers/target/target_core_file.c fd_execute_rw_aio(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 285 drivers/target/target_core_file.c for_each_sg(sgl, sg, sgl_nents, i) { sgl 319 drivers/target/target_core_file.c u32 block_size, struct scatterlist *sgl, sgl 335 drivers/target/target_core_file.c for_each_sg(sgl, sg, sgl_nents, i) { sgl 595 drivers/target/target_core_file.c fd_execute_rw_buffered(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 618 drivers/target/target_core_file.c sgl, sgl_nents, cmd->data_length, 0); sgl 643 drivers/target/target_core_file.c sgl, sgl_nents, cmd->data_length, 1); sgl 679 drivers/target/target_core_file.c fd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 697 drivers/target/target_core_file.c return fd_execute_rw_aio(cmd, sgl, sgl_nents, data_direction); sgl 698 drivers/target/target_core_file.c return fd_execute_rw_buffered(cmd, sgl, sgl_nents, data_direction); sgl 681 drivers/target/target_core_iblock.c iblock_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 739 drivers/target/target_core_iblock.c for_each_sg(sgl, sg, sgl_nents, i) { sgl 846 drivers/target/target_core_pscsi.c pscsi_map_sg(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 854 drivers/target/target_core_pscsi.c int nr_pages = (cmd->data_length + sgl[0].offset + sgl 863 drivers/target/target_core_pscsi.c for_each_sg(sgl, sg, sgl_nents, i) { sgl 957 drivers/target/target_core_pscsi.c struct scatterlist *sgl = cmd->t_data_sg; sgl 986 drivers/target/target_core_pscsi.c if (sgl) { sgl 987 drivers/target/target_core_pscsi.c ret = pscsi_map_sg(cmd, sgl, sgl_nents, req); sgl 417 drivers/target/target_core_rd.c rd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, sgl 461 drivers/target/target_core_rd.c sg_miter_start(&m, sgl, sgl_nents, sgl 488 drivers/target/target_core_sbc.c write_sg = write_tbl.sgl; sgl 1502 drivers/target/target_core_transport.c transport_generic_map_mem_to_cmd(struct se_cmd *cmd, struct scatterlist *sgl, sgl 1505 drivers/target/target_core_transport.c if (!sgl || !sgl_count) sgl 1519 drivers/target/target_core_transport.c cmd->t_data_sg = sgl; sgl 1560 drivers/target/target_core_transport.c struct scatterlist *sgl, u32 sgl_count, sgl 1633 drivers/target/target_core_transport.c BUG_ON(!sgl); sgl 1647 drivers/target/target_core_transport.c if (sgl) sgl 1648 drivers/target/target_core_transport.c buf = kmap(sg_page(sgl)) + sgl->offset; sgl 1651 drivers/target/target_core_transport.c memset(buf, 0, sgl->length); sgl 1652 drivers/target/target_core_transport.c kunmap(sg_page(sgl)); sgl 1656 drivers/target/target_core_transport.c rc = transport_generic_map_mem_to_cmd(se_cmd, sgl, sgl_count, sgl 2409 drivers/target/target_core_transport.c void target_free_sgl(struct scatterlist *sgl, int nents) sgl 2411 drivers/target/target_core_transport.c sgl_free_n_order(sgl, nents, 0); sgl 2516 drivers/target/target_core_transport.c target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents, u32 length, sgl 2521 drivers/target/target_core_transport.c *sgl = sgl_alloc_order(length, 0, chainable, gfp, nents); sgl 2522 drivers/target/target_core_transport.c return *sgl ? 0 : -ENOMEM; sgl 933 drivers/tty/serial/atmel_serial.c struct scatterlist sgl[2], *sg, *sg_tx = &atmel_port->sg_tx; sgl 965 drivers/tty/serial/atmel_serial.c sg_init_table(sgl, 2); sgl 969 drivers/tty/serial/atmel_serial.c sg = &sgl[sg_len++]; sgl 977 drivers/tty/serial/atmel_serial.c sg = &sgl[sg_len++]; sgl 989 drivers/tty/serial/atmel_serial.c sgl, sgl 414 drivers/tty/serial/fsl_lpuart.c struct scatterlist *sgl = sport->tx_sgl; sgl 425 drivers/tty/serial/fsl_lpuart.c sg_init_one(sgl, xmit->buf + xmit->tail, sport->dma_tx_bytes); sgl 428 drivers/tty/serial/fsl_lpuart.c sg_init_table(sgl, 2); sgl 429 drivers/tty/serial/fsl_lpuart.c sg_set_buf(sgl, xmit->buf + xmit->tail, sgl 431 drivers/tty/serial/fsl_lpuart.c sg_set_buf(sgl + 1, xmit->buf, xmit->head); sgl 434 drivers/tty/serial/fsl_lpuart.c ret = dma_map_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE); sgl 440 drivers/tty/serial/fsl_lpuart.c sport->dma_tx_desc = dmaengine_prep_slave_sg(sport->dma_tx_chan, sgl, sgl 444 drivers/tty/serial/fsl_lpuart.c dma_unmap_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE); sgl 464 drivers/tty/serial/fsl_lpuart.c struct scatterlist *sgl = &sport->tx_sgl[0]; sgl 470 drivers/tty/serial/fsl_lpuart.c dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE); sgl 551 drivers/tty/serial/imx.c struct scatterlist *sgl = &sport->tx_sgl[0]; sgl 558 drivers/tty/serial/imx.c dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE); sgl 590 drivers/tty/serial/imx.c struct scatterlist *sgl = sport->tx_sgl; sgl 608 drivers/tty/serial/imx.c sg_init_one(sgl, xmit->buf + xmit->tail, sport->tx_bytes); sgl 611 drivers/tty/serial/imx.c sg_init_table(sgl, 2); sgl 612 drivers/tty/serial/imx.c sg_set_buf(sgl, xmit->buf + xmit->tail, sgl 614 drivers/tty/serial/imx.c sg_set_buf(sgl + 1, xmit->buf, xmit->head); sgl 617 drivers/tty/serial/imx.c ret = dma_map_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE); sgl 622 drivers/tty/serial/imx.c desc = dmaengine_prep_slave_sg(chan, sgl, ret, sgl 625 drivers/tty/serial/imx.c dma_unmap_sg(dev, sgl, sport->dma_tx_nents, sgl 1076 drivers/tty/serial/imx.c struct scatterlist *sgl = &sport->rx_sgl; sgl 1106 drivers/tty/serial/imx.c rx_ring->head = sg_dma_len(sgl) - state.residue; sgl 1109 drivers/tty/serial/imx.c bd_size = sg_dma_len(sgl) / sport->rx_periods; sgl 1112 drivers/tty/serial/imx.c if (rx_ring->head <= sg_dma_len(sgl) && sgl 1119 drivers/tty/serial/imx.c dma_sync_sg_for_cpu(sport->port.dev, sgl, 1, sgl 1126 drivers/tty/serial/imx.c dma_sync_sg_for_device(sport->port.dev, sgl, 1, sgl 1134 drivers/tty/serial/imx.c WARN_ON(rx_ring->head > sg_dma_len(sgl)); sgl 1150 drivers/tty/serial/imx.c struct scatterlist *sgl = &sport->rx_sgl; sgl 1160 drivers/tty/serial/imx.c sg_init_one(sgl, sport->rx_buf, RX_BUF_SIZE); sgl 1161 drivers/tty/serial/imx.c ret = dma_map_sg(dev, sgl, 1, DMA_FROM_DEVICE); sgl 1167 drivers/tty/serial/imx.c desc = dmaengine_prep_dma_cyclic(chan, sg_dma_address(sgl), sgl 1168 drivers/tty/serial/imx.c sg_dma_len(sgl), sg_dma_len(sgl) / sport->rx_periods, sgl 1172 drivers/tty/serial/imx.c dma_unmap_sg(dev, sgl, 1, DMA_FROM_DEVICE); sgl 1515 drivers/tty/serial/imx.c struct scatterlist *sgl = &sport->tx_sgl[0]; sgl 1527 drivers/tty/serial/imx.c dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, sgl 552 drivers/tty/serial/mxs-auart.c struct scatterlist *sgl = &s->tx_sgl; sgl 566 drivers/tty/serial/mxs-auart.c sg_init_one(sgl, s->tx_dma_buf, size); sgl 567 drivers/tty/serial/mxs-auart.c dma_map_sg(s->dev, sgl, 1, DMA_TO_DEVICE); sgl 568 drivers/tty/serial/mxs-auart.c desc = dmaengine_prep_slave_sg(channel, sgl, sgl 874 drivers/tty/serial/mxs-auart.c struct scatterlist *sgl = &s->rx_sgl; sgl 890 drivers/tty/serial/mxs-auart.c sg_init_one(sgl, s->rx_dma_buf, UART_XMIT_SIZE); sgl 891 drivers/tty/serial/mxs-auart.c dma_map_sg(s->dev, sgl, 1, DMA_FROM_DEVICE); sgl 892 drivers/tty/serial/mxs-auart.c desc = dmaengine_prep_slave_sg(channel, sgl, 1, DMA_DEV_TO_MEM, sgl 1061 drivers/usb/gadget/function/f_fs.c req->sg = io_data->sgt.sgl; sgl 1105 drivers/usb/gadget/function/f_fs.c req->sg = io_data->sgt.sgl; sgl 458 drivers/usb/storage/isd200.c srb->sdb.table.sgl = buff ? &info->sg : NULL; sgl 462 drivers/usb/storage/uas.c urb->sg = sdb->table.sgl; sgl 56 drivers/usb/usbip/stub.h struct scatterlist *sgl; sgl 307 drivers/usb/usbip/stub_main.c if (urb->transfer_buffer && !priv->sgl) { sgl 322 drivers/usb/usbip/stub_main.c if (priv->sgl) sgl 323 drivers/usb/usbip/stub_main.c sgl_free(priv->sgl); sgl 459 drivers/usb/usbip/stub_rx.c struct scatterlist *sgl = NULL, *sg; sgl 492 drivers/usb/usbip/stub_rx.c sgl = sgl_alloc(buf_len, GFP_KERNEL, &nents); sgl 493 drivers/usb/usbip/stub_rx.c if (!sgl) sgl 541 drivers/usb/usbip/stub_rx.c priv->urbs[0]->sg = sgl; sgl 559 drivers/usb/usbip/stub_rx.c for_each_sg(sgl, sg, nents, i) { sgl 571 drivers/usb/usbip/stub_rx.c priv->sgl = sgl; sgl 621 drivers/usb/usbip/stub_rx.c sgl_free(sgl); sgl 79 drivers/usb/usbip/stub_tx.c if (priv->sgl) { sgl 188 drivers/usb/usbip/stub_tx.c else if (usb_pipein(urb->pipe) && priv->sgl) sgl 207 drivers/usb/usbip/stub_tx.c if (priv->sgl) { sgl 223 drivers/usb/usbip/stub_tx.c if (usb_pipein(urb->pipe) && priv->sgl) { sgl 624 drivers/vhost/scsi.c struct scatterlist *sgl, sgl 628 drivers/vhost/scsi.c struct scatterlist *sg = sgl; sgl 250 drivers/xen/gntdev-dmabuf.c dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgl 291 drivers/xen/gntdev-dmabuf.c if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgl 656 drivers/xen/gntdev-dmabuf.c for_each_sg_page(sgt->sgl, &sg_iter, sgt->nents, 0) { sgl 466 drivers/xen/swiotlb-xen.c xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, sgl 474 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) sgl 481 drivers/xen/swiotlb-xen.c xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems, sgl 489 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { sgl 499 drivers/xen/swiotlb-xen.c xen_swiotlb_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC); sgl 500 drivers/xen/swiotlb-xen.c sg_dma_len(sgl) = 0; sgl 505 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, sgl 511 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { sgl 518 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, sgl 524 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { sgl 129 drivers/xen/xen-scsiback.c struct scatterlist *sgl; sgl 277 drivers/xen/xen-scsiback.c kfree(req->sgl); sgl 278 drivers/xen/xen-scsiback.c req->sgl = NULL; sgl 408 drivers/xen/xen-scsiback.c pending_req->sgl, pending_req->n_sg, sgl 524 drivers/xen/xen-scsiback.c pending_req->sgl = kmalloc_array(nr_segments, sgl 526 drivers/xen/xen-scsiback.c if (!pending_req->sgl) sgl 529 drivers/xen/xen-scsiback.c sg_init_table(pending_req->sgl, nr_segments); sgl 563 drivers/xen/xen-scsiback.c for_each_sg(pending_req->sgl, sg, nr_segments, i) { sgl 983 fs/cifs/smbdirect.c struct scatterlist *sgl, int data_length, int remaining_data_length) sgl 995 fs/cifs/smbdirect.c num_sgs = sgl ? sg_nents(sgl) : 0; sgl 996 fs/cifs/smbdirect.c for_each_sg(sgl, sg, num_sgs, i) { sgl 1036 fs/cifs/smbdirect.c struct scatterlist sgl; sgl 1038 fs/cifs/smbdirect.c sg_init_table(&sgl, 1); sgl 1039 fs/cifs/smbdirect.c sg_set_page(&sgl, page, size, offset); sgl 1041 fs/cifs/smbdirect.c return smbd_post_send_sgl(info, &sgl, size, remaining_data_length); sgl 1068 fs/cifs/smbdirect.c struct scatterlist sgl[SMBDIRECT_MAX_SGE]; sgl 1075 fs/cifs/smbdirect.c sg_init_table(sgl, n_vec); sgl 1078 fs/cifs/smbdirect.c sg_set_buf(&sgl[i], iov[i].iov_base, iov[i].iov_len); sgl 1081 fs/cifs/smbdirect.c return smbd_post_send_sgl(info, sgl, data_length, remaining_data_length); sgl 2323 fs/cifs/smbdirect.c ib_dma_unmap_sg(info->id->device, mr->sgl, sgl 2326 fs/cifs/smbdirect.c kfree(mr->sgl); sgl 2362 fs/cifs/smbdirect.c smbdirect_mr->sgl = kcalloc( sgl 2366 fs/cifs/smbdirect.c if (!smbdirect_mr->sgl) { sgl 2385 fs/cifs/smbdirect.c kfree(smbdirect_mr->sgl); sgl 2468 fs/cifs/smbdirect.c sg_init_table(smbdirect_mr->sgl, num_pages); sgl 2474 fs/cifs/smbdirect.c sg_set_page(&smbdirect_mr->sgl[0], pages[0], tailsz, offset); sgl 2480 fs/cifs/smbdirect.c &smbdirect_mr->sgl[0], pages[0], PAGE_SIZE - offset, offset); sgl 2483 fs/cifs/smbdirect.c sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0); sgl 2486 fs/cifs/smbdirect.c sg_set_page(&smbdirect_mr->sgl[i], pages[i], sgl 2492 fs/cifs/smbdirect.c rc = ib_dma_map_sg(info->id->device, smbdirect_mr->sgl, num_pages, dir); sgl 2499 fs/cifs/smbdirect.c rc = ib_map_mr_sg(smbdirect_mr->mr, smbdirect_mr->sgl, num_pages, sgl 2536 fs/cifs/smbdirect.c ib_dma_unmap_sg(info->id->device, smbdirect_mr->sgl, sgl 2604 fs/cifs/smbdirect.c info->id->device, smbdirect_mr->sgl, sgl 293 fs/cifs/smbdirect.h struct scatterlist *sgl; sgl 77 include/crypto/if_alg.h struct af_alg_sgl sgl; sgl 165 include/crypto/if_alg.h int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len); sgl 166 include/crypto/if_alg.h void af_alg_free_sg(struct af_alg_sgl *sgl); sgl 206 include/linux/dma-mapping.h int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents, sgl 216 include/linux/dma-mapping.h struct scatterlist *sgl, int nents, enum dma_data_direction dir); sgl 223 include/linux/dma-mapping.h struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 233 include/linux/dma-mapping.h void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl, sgl 238 include/linux/dma-mapping.h struct scatterlist *sgl, int nents, enum dma_data_direction dir); sgl 245 include/linux/dma-mapping.h struct scatterlist *sgl, int nents, enum dma_data_direction dir, sgl 254 include/linux/dma-mapping.h struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 156 include/linux/dmaengine.h struct data_chunk sgl[0]; sgl 780 include/linux/dmaengine.h struct dma_chan *chan, struct scatterlist *sgl, sgl 838 include/linux/dmaengine.h struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, sgl 844 include/linux/dmaengine.h return chan->device->device_prep_slave_sg(chan, sgl, sg_len, sgl 851 include/linux/dmaengine.h struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, sgl 858 include/linux/dmaengine.h return chan->device->device_prep_slave_sg(chan, sgl, sg_len, sgl 1122 include/linux/dmaengine.h static inline size_t dmaengine_get_icg(bool inc, bool sgl, size_t icg, sgl 1128 include/linux/dmaengine.h else if (sgl) sgl 645 include/linux/kfifo.h #define kfifo_dma_in_prepare(fifo, sgl, nents, len) \ sgl 648 include/linux/kfifo.h struct scatterlist *__sgl = (sgl); \ sgl 696 include/linux/kfifo.h #define kfifo_dma_out_prepare(fifo, sgl, nents, len) \ sgl 699 include/linux/kfifo.h struct scatterlist *__sgl = (sgl); \ sgl 779 include/linux/kfifo.h struct scatterlist *sgl, int nents, unsigned int len); sgl 782 include/linux/kfifo.h struct scatterlist *sgl, int nents, unsigned int len); sgl 801 include/linux/kfifo.h struct scatterlist *sgl, int nents, unsigned int len, size_t recsize); sgl 807 include/linux/kfifo.h struct scatterlist *sgl, int nents, unsigned int len, size_t recsize); sgl 52 include/linux/mpi.h MPI mpi_read_raw_from_sgl(struct scatterlist *sgl, unsigned int len); sgl 239 include/linux/mtd/ubi.h int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl, sgl 267 include/linux/mtd/ubi.h struct ubi_sgl *sgl, int offset, int len) sgl 269 include/linux/mtd/ubi.h return ubi_leb_read_sg(desc, lnum, sgl, offset, len, 0); sgl 1336 include/linux/netdevice.h struct scatterlist *sgl, sgl 1342 include/linux/netdevice.h struct scatterlist *sgl, sgl 645 include/linux/nvme.h struct nvme_sgl_desc sgl; sgl 31 include/linux/pci-p2pdma.h void pci_p2pmem_free_sgl(struct pci_dev *pdev, struct scatterlist *sgl); sgl 80 include/linux/pci-p2pdma.h struct scatterlist *sgl) sgl 43 include/linux/scatterlist.h struct scatterlist *sgl; /* the list */ sgl 165 include/linux/scatterlist.h struct scatterlist *sgl) sgl 177 include/linux/scatterlist.h prv[prv_nents - 1].page_link = ((unsigned long) sgl | SG_CHAIN) sgl 248 include/linux/scatterlist.h static inline void sg_init_marker(struct scatterlist *sgl, sgl 251 include/linux/scatterlist.h sg_mark_end(&sgl[nents - 1]); sgl 289 include/linux/scatterlist.h void sgl_free_n_order(struct scatterlist *sgl, int nents, int order); sgl 290 include/linux/scatterlist.h void sgl_free_order(struct scatterlist *sgl, int order); sgl 291 include/linux/scatterlist.h void sgl_free(struct scatterlist *sgl); sgl 294 include/linux/scatterlist.h size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, sgl 297 include/linux/scatterlist.h size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents, sgl 299 include/linux/scatterlist.h size_t sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents, sgl 302 include/linux/scatterlist.h size_t sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents, sgl 304 include/linux/scatterlist.h size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents, sgl 306 include/linux/scatterlist.h size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents, sgl 461 include/linux/scatterlist.h void sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl, sgl 4352 include/rdma/ib_verbs.h int ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents, sgl 185 include/scsi/scsi_cmnd.h return cmd->sdb.table.sgl; sgl 299 include/scsi/scsi_cmnd.h return cmd->prot_sdb ? cmd->prot_sdb->table.sgl : NULL; sgl 205 include/target/target_core_fabric.h int target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents, sgl 207 include/target/target_core_fabric.h void target_free_sgl(struct scatterlist *sgl, int nents); sgl 240 kernel/dma/direct.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 245 kernel/dma/direct.c for_each_sg(sgl, sg, nents, i) { sgl 279 kernel/dma/direct.c struct scatterlist *sgl, int nents, enum dma_data_direction dir) sgl 284 kernel/dma/direct.c for_each_sg(sgl, sg, nents, i) { sgl 313 kernel/dma/direct.c void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl, sgl 319 kernel/dma/direct.c for_each_sg(sgl, sg, nents, i) sgl 352 kernel/dma/direct.c int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents, sgl 358 kernel/dma/direct.c for_each_sg(sgl, sg, nents, i) { sgl 369 kernel/dma/direct.c dma_direct_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC); sgl 21 kernel/dma/dummy.c static int dma_dummy_map_sg(struct device *dev, struct scatterlist *sgl, sgl 135 kernel/dma/mapping.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgl 37 kernel/dma/virt.c static int dma_virt_map_sg(struct device *dev, struct scatterlist *sgl, sgl 44 kernel/dma/virt.c for_each_sg(sgl, sg, nents, i) { sgl 295 lib/kfifo.c static int setup_sgl_buf(struct scatterlist *sgl, void *buf, sgl 321 lib/kfifo.c sg_set_page(sgl, page, l - off, off); sgl 322 lib/kfifo.c sgl = sg_next(sgl); sgl 323 lib/kfifo.c if (++n == nents || sgl == NULL) sgl 330 lib/kfifo.c sg_set_page(sgl, page, len, off); sgl 334 lib/kfifo.c static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl, sgl 350 lib/kfifo.c n = setup_sgl_buf(sgl, fifo->data + off, nents, l); sgl 351 lib/kfifo.c n += setup_sgl_buf(sgl + n, fifo->data, nents - n, len - l); sgl 357 lib/kfifo.c struct scatterlist *sgl, int nents, unsigned int len) sgl 365 lib/kfifo.c return setup_sgl(fifo, sgl, nents, len, fifo->in); sgl 370 lib/kfifo.c struct scatterlist *sgl, int nents, unsigned int len) sgl 378 lib/kfifo.c return setup_sgl(fifo, sgl, nents, len, fifo->out); sgl 549 lib/kfifo.c struct scatterlist *sgl, int nents, unsigned int len, size_t recsize) sgl 558 lib/kfifo.c return setup_sgl(fifo, sgl, nents, len, fifo->in + recsize); sgl 572 lib/kfifo.c struct scatterlist *sgl, int nents, unsigned int len, size_t recsize) sgl 581 lib/kfifo.c return setup_sgl(fifo, sgl, nents, len, fifo->out + recsize); sgl 247 lib/mpi/mpicoder.c int mpi_write_to_sgl(MPI a, struct scatterlist *sgl, unsigned nbytes, sgl 269 lib/mpi/mpicoder.c nents = sg_nents_for_len(sgl, nbytes); sgl 273 lib/mpi/mpicoder.c sg_miter_start(&miter, sgl, nents, SG_MITER_ATOMIC | SG_MITER_TO_SG); sgl 330 lib/mpi/mpicoder.c MPI mpi_read_raw_from_sgl(struct scatterlist *sgl, unsigned int nbytes) sgl 340 lib/mpi/mpicoder.c ents = sg_nents_for_len(sgl, nbytes); sgl 344 lib/mpi/mpicoder.c sg_miter_start(&miter, sgl, ents, SG_MITER_ATOMIC | SG_MITER_FROM_SG); sgl 101 lib/scatterlist.c struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents) sgl 106 lib/scatterlist.c for_each_sg(sgl, sg, nents, i) sgl 124 lib/scatterlist.c void sg_init_table(struct scatterlist *sgl, unsigned int nents) sgl 126 lib/scatterlist.c memset(sgl, 0, sizeof(*sgl) * nents); sgl 127 lib/scatterlist.c sg_init_marker(sgl, nents); sgl 195 lib/scatterlist.c struct scatterlist *sgl, *next; sgl 198 lib/scatterlist.c if (unlikely(!table->sgl)) sgl 201 lib/scatterlist.c sgl = table->sgl; sgl 213 lib/scatterlist.c next = sg_chain_ptr(&sgl[curr_max_ents - 1]); sgl 225 lib/scatterlist.c free_fn(sgl, alloc_size); sgl 226 lib/scatterlist.c sgl = next; sgl 230 lib/scatterlist.c table->sgl = NULL; sgl 327 lib/scatterlist.c table->sgl = sg; sgl 419 lib/scatterlist.c for_each_sg(sgt->sgl, s, sgt->orig_nents, i) { sgl 490 lib/scatterlist.c struct scatterlist *sgl, *sg; sgl 506 lib/scatterlist.c sgl = kmalloc_array(nalloc, sizeof(struct scatterlist), sgl 508 lib/scatterlist.c if (!sgl) sgl 511 lib/scatterlist.c sg_init_table(sgl, nalloc); sgl 512 lib/scatterlist.c sg = sgl; sgl 517 lib/scatterlist.c sgl_free(sgl); sgl 528 lib/scatterlist.c return sgl; sgl 560 lib/scatterlist.c void sgl_free_n_order(struct scatterlist *sgl, int nents, int order) sgl 566 lib/scatterlist.c for_each_sg(sgl, sg, nents, i) { sgl 573 lib/scatterlist.c kfree(sgl); sgl 582 lib/scatterlist.c void sgl_free_order(struct scatterlist *sgl, int order) sgl 584 lib/scatterlist.c sgl_free_n_order(sgl, INT_MAX, order); sgl 592 lib/scatterlist.c void sgl_free(struct scatterlist *sgl) sgl 594 lib/scatterlist.c sgl_free_order(sgl, 0); sgl 674 lib/scatterlist.c void sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl, sgl 679 lib/scatterlist.c __sg_page_iter_start(&miter->piter, sgl, nents, 0); sgl 840 lib/scatterlist.c size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, sgl 852 lib/scatterlist.c sg_miter_start(&miter, sgl, nents, sg_flags); sgl 886 lib/scatterlist.c size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents, sgl 889 lib/scatterlist.c return sg_copy_buffer(sgl, nents, (void *)buf, buflen, 0, false); sgl 903 lib/scatterlist.c size_t sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents, sgl 906 lib/scatterlist.c return sg_copy_buffer(sgl, nents, buf, buflen, 0, true); sgl 921 lib/scatterlist.c size_t sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents, sgl 924 lib/scatterlist.c return sg_copy_buffer(sgl, nents, (void *)buf, buflen, skip, false); sgl 939 lib/scatterlist.c size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents, sgl 942 lib/scatterlist.c return sg_copy_buffer(sgl, nents, buf, buflen, skip, true); sgl 955 lib/scatterlist.c size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents, sgl 962 lib/scatterlist.c sg_miter_start(&miter, sgl, nents, sg_flags); sgl 54 lib/sg_pool.c static void sg_pool_free(struct scatterlist *sgl, unsigned int nents) sgl 59 lib/sg_pool.c mempool_free(sgl, sgp->pool); sgl 121 lib/sg_pool.c sg_init_table(table->sgl, nents); sgl 401 net/8021q/vlan_dev.c struct scatterlist *sgl, unsigned int sgc) sgl 408 net/8021q/vlan_dev.c rc = ops->ndo_fcoe_ddp_setup(real_dev, xid, sgl, sgc); sgl 448 net/8021q/vlan_dev.c struct scatterlist *sgl, unsigned int sgc) sgl 455 net/8021q/vlan_dev.c rc = ops->ndo_fcoe_ddp_target(real_dev, xid, sgl, sgc); sgl 188 net/ceph/crypto.c sgt->sgl = prealloc_sg; sgl 192 net/ceph/crypto.c for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) { sgl 239 net/ceph/crypto.c skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv); sgl 516 net/smc/smc_clc.c (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgl 588 net/smc/smc_clc.c (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgl 770 net/smc/smc_core.c sg_set_buf(buf_desc->sgt[SMC_SINGLE_LINK].sgl, sgl 369 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgl 408 net/smc/smc_ib.c for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg, sgl 428 net/smc/smc_ib.c for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg, sgl 447 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgl 460 net/smc/smc_ib.c if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address) sgl 464 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgl 467 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0; sgl 236 net/smc/smc_llc.c (u64)sg_dma_address(rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgl 315 net/smc/smc_tx.c sg_dma_address(conn->sndbuf_desc->sgt[SMC_SINGLE_LINK].sgl); sgl 74 net/sunrpc/xprtrdma/svc_rdma_rw.c ctxt->rw_sg_table.sgl = ctxt->rw_first_sgl; sgl 76 net/sunrpc/xprtrdma/svc_rdma_rw.c ctxt->rw_sg_table.sgl, sgl 145 net/sunrpc/xprtrdma/svc_rdma_rw.c rdma->sc_port_num, ctxt->rw_sg_table.sgl, sgl 356 net/sunrpc/xprtrdma/svc_rdma_rw.c struct scatterlist *sg = ctxt->rw_sg_table.sgl; sgl 380 net/sunrpc/xprtrdma/svc_rdma_rw.c sg = ctxt->rw_sg_table.sgl; sgl 435 net/sunrpc/xprtrdma/svc_rdma_rw.c rdma->sc_port_num, ctxt->rw_sg_table.sgl, sgl 615 net/sunrpc/xprtrdma/svc_rdma_rw.c sg = ctxt->rw_sg_table.sgl; sgl 644 net/sunrpc/xprtrdma/svc_rdma_rw.c ctxt->rw_sg_table.sgl, ctxt->rw_nents, sgl 849 samples/vfio-mdev/mbochs.c if (!dma_map_sg(at->dev, sg->sgl, sg->nents, direction)) sgl 91 tools/virtio/linux/scatterlist.h struct scatterlist *sgl) sgl 103 tools/virtio/linux/scatterlist.h prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02; sgl 149 tools/virtio/linux/scatterlist.h static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents) sgl 151 tools/virtio/linux/scatterlist.h memset(sgl, 0, sizeof(*sgl) * nents); sgl 152 tools/virtio/linux/scatterlist.h sg_mark_end(&sgl[nents - 1]);