sgl               117 arch/arm/mm/dma-mapping-nommu.c static int arm_nommu_dma_map_sg(struct device *dev, struct scatterlist *sgl,
sgl               124 arch/arm/mm/dma-mapping-nommu.c 	for_each_sg(sgl, sg, nents, i) {
sgl               133 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_unmap_sg(struct device *dev, struct scatterlist *sgl,
sgl               140 arch/arm/mm/dma-mapping-nommu.c 	for_each_sg(sgl, sg, nents, i)
sgl               156 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl,
sgl               162 arch/arm/mm/dma-mapping-nommu.c 	for_each_sg(sgl, sg, nents, i)
sgl               166 arch/arm/mm/dma-mapping-nommu.c static void arm_nommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl,
sgl               172 arch/arm/mm/dma-mapping-nommu.c 	for_each_sg(sgl, sg, nents, i)
sgl               883 arch/arm/mm/dma-mapping.c 	sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgl               653 arch/mips/jazz/jazzdma.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               658 arch/mips/jazz/jazzdma.c 	for_each_sg(sgl, sg, nents, i)
sgl               663 arch/mips/jazz/jazzdma.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               668 arch/mips/jazz/jazzdma.c 	for_each_sg(sgl, sg, nents, i)
sgl               184 arch/powerpc/kernel/dma-iommu.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               187 arch/powerpc/kernel/dma-iommu.c 		dma_direct_sync_sg_for_cpu(dev, sgl, nents, dir);
sgl               191 arch/powerpc/kernel/dma-iommu.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               194 arch/powerpc/kernel/dma-iommu.c 		dma_direct_sync_sg_for_device(dev, sgl, nents, dir);
sgl               632 arch/powerpc/platforms/ps3/system-bus.c static int ps3_sb_map_sg(struct device *_dev, struct scatterlist *sgl,
sgl               643 arch/powerpc/platforms/ps3/system-bus.c 	for_each_sg(sgl, sg, nents, i) {
sgl               106 arch/powerpc/platforms/pseries/ibmebus.c 			  struct scatterlist *sgl,
sgl               113 arch/powerpc/platforms/pseries/ibmebus.c 	for_each_sg(sgl, sg, nents, i) {
sgl               554 arch/powerpc/platforms/pseries/vio.c 	struct scatterlist *sgl;
sgl               558 arch/powerpc/platforms/pseries/vio.c 	for_each_sg(sglist, sgl, nelems, count)
sgl               559 arch/powerpc/platforms/pseries/vio.c 		alloc_size += roundup(sgl->length, IOMMU_PAGE_SIZE(tbl));
sgl               568 arch/powerpc/platforms/pseries/vio.c 	for_each_sg(sglist, sgl, ret, count)
sgl               569 arch/powerpc/platforms/pseries/vio.c 		alloc_size -= roundup(sgl->dma_length, IOMMU_PAGE_SIZE(tbl));
sgl               588 arch/powerpc/platforms/pseries/vio.c 	struct scatterlist *sgl;
sgl               592 arch/powerpc/platforms/pseries/vio.c 	for_each_sg(sglist, sgl, nelems, count)
sgl               593 arch/powerpc/platforms/pseries/vio.c 		alloc_size += roundup(sgl->dma_length, IOMMU_PAGE_SIZE(tbl));
sgl               161 arch/sparc/mm/io-unit.c static int iounit_map_sg(struct device *dev, struct scatterlist *sgl, int nents,
sgl               171 arch/sparc/mm/io-unit.c 	for_each_sg(sgl, sg, nents, i) {
sgl               194 arch/sparc/mm/io-unit.c static void iounit_unmap_sg(struct device *dev, struct scatterlist *sgl,
sgl               203 arch/sparc/mm/io-unit.c 	for_each_sg(sgl, sg, nents, i) {
sgl               243 arch/sparc/mm/iommu.c static int __sbus_iommu_map_sg(struct device *dev, struct scatterlist *sgl,
sgl               250 arch/sparc/mm/iommu.c 	for_each_sg(sgl, sg, nents, j) {
sgl               261 arch/sparc/mm/iommu.c static int sbus_iommu_map_sg_gflush(struct device *dev, struct scatterlist *sgl,
sgl               265 arch/sparc/mm/iommu.c 	return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, false);
sgl               268 arch/sparc/mm/iommu.c static int sbus_iommu_map_sg_pflush(struct device *dev, struct scatterlist *sgl,
sgl               271 arch/sparc/mm/iommu.c 	return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, true);
sgl               293 arch/sparc/mm/iommu.c static void sbus_iommu_unmap_sg(struct device *dev, struct scatterlist *sgl,
sgl               299 arch/sparc/mm/iommu.c 	for_each_sg(sgl, sg, nents, i) {
sgl               256 crypto/adiantum.c 				 struct scatterlist *sgl, le128 *digest)
sgl               273 crypto/adiantum.c 	sg_miter_start(&miter, sgl, sg_nents(sgl),
sgl               393 crypto/af_alg.c int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len)
sgl               399 crypto/af_alg.c 	n = iov_iter_get_pages(iter, sgl->pages, len, ALG_MAX_PAGES, &off);
sgl               407 crypto/af_alg.c 	sg_init_table(sgl->sg, npages + 1);
sgl               412 crypto/af_alg.c 		sg_set_page(sgl->sg + i, sgl->pages[i], plen, off);
sgl               417 crypto/af_alg.c 	sg_mark_end(sgl->sg + npages - 1);
sgl               418 crypto/af_alg.c 	sgl->npages = npages;
sgl               431 crypto/af_alg.c void af_alg_free_sg(struct af_alg_sgl *sgl)
sgl               435 crypto/af_alg.c 	for (i = 0; i < sgl->npages; i++)
sgl               436 crypto/af_alg.c 		put_page(sgl->pages[i]);
sgl               490 crypto/af_alg.c 	struct af_alg_tsgl *sgl;
sgl               493 crypto/af_alg.c 	sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list);
sgl               495 crypto/af_alg.c 		sg = sgl->sg;
sgl               497 crypto/af_alg.c 	if (!sg || sgl->cur >= MAX_SGL_ENTS) {
sgl               498 crypto/af_alg.c 		sgl = sock_kmalloc(sk,
sgl               499 crypto/af_alg.c 				   struct_size(sgl, sg, (MAX_SGL_ENTS + 1)),
sgl               501 crypto/af_alg.c 		if (!sgl)
sgl               504 crypto/af_alg.c 		sg_init_table(sgl->sg, MAX_SGL_ENTS + 1);
sgl               505 crypto/af_alg.c 		sgl->cur = 0;
sgl               508 crypto/af_alg.c 			sg_chain(sg, MAX_SGL_ENTS + 1, sgl->sg);
sgl               510 crypto/af_alg.c 		list_add_tail(&sgl->list, &ctx->tsgl_list);
sgl               531 crypto/af_alg.c 	const struct af_alg_tsgl *sgl;
sgl               538 crypto/af_alg.c 	list_for_each_entry(sgl, &ctx->tsgl_list, list) {
sgl               539 crypto/af_alg.c 		const struct scatterlist *sg = sgl->sg;
sgl               541 crypto/af_alg.c 		for (i = 0; i < sgl->cur; i++) {
sgl               588 crypto/af_alg.c 	struct af_alg_tsgl *sgl;
sgl               593 crypto/af_alg.c 		sgl = list_first_entry(&ctx->tsgl_list, struct af_alg_tsgl,
sgl               595 crypto/af_alg.c 		sg = sgl->sg;
sgl               597 crypto/af_alg.c 		for (i = 0; i < sgl->cur; i++) {
sgl               636 crypto/af_alg.c 		list_del(&sgl->list);
sgl               637 crypto/af_alg.c 		sock_kfree_s(sk, sgl, struct_size(sgl, sg, MAX_SGL_ENTS + 1));
sgl               662 crypto/af_alg.c 		af_alg_free_sg(&rsgl->sgl);
sgl               821 crypto/af_alg.c 	struct af_alg_tsgl *sgl;
sgl               870 crypto/af_alg.c 			sgl = list_entry(ctx->tsgl_list.prev,
sgl               872 crypto/af_alg.c 			sg = sgl->sg + sgl->cur - 1;
sgl               905 crypto/af_alg.c 		sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl,
sgl               907 crypto/af_alg.c 		sg = sgl->sg;
sgl               908 crypto/af_alg.c 		if (sgl->cur)
sgl               909 crypto/af_alg.c 			sg_unmark_end(sg + sgl->cur - 1);
sgl               912 crypto/af_alg.c 			unsigned int i = sgl->cur;
sgl               935 crypto/af_alg.c 			sgl->cur++;
sgl               936 crypto/af_alg.c 		} while (len && sgl->cur < MAX_SGL_ENTS);
sgl               939 crypto/af_alg.c 			sg_mark_end(sg + sgl->cur - 1);
sgl               967 crypto/af_alg.c 	struct af_alg_tsgl *sgl;
sgl               991 crypto/af_alg.c 	sgl = list_entry(ctx->tsgl_list.prev, struct af_alg_tsgl, list);
sgl               993 crypto/af_alg.c 	if (sgl->cur)
sgl               994 crypto/af_alg.c 		sg_unmark_end(sgl->sg + sgl->cur - 1);
sgl               996 crypto/af_alg.c 	sg_mark_end(sgl->sg + sgl->cur);
sgl               999 crypto/af_alg.c 	sg_set_page(sgl->sg + sgl->cur, page, size, offset);
sgl              1000 crypto/af_alg.c 	sgl->cur++;
sgl              1142 crypto/af_alg.c 		rsgl->sgl.npages = 0;
sgl              1146 crypto/af_alg.c 		err = af_alg_make_sg(&rsgl->sgl, &msg->msg_iter, seglen);
sgl              1154 crypto/af_alg.c 			af_alg_link_sg(&areq->last_rsgl->sgl, &rsgl->sgl);
sgl               213 crypto/algif_aead.c 	rsgl_src = areq->first_rsgl.sgl.sg;
sgl               227 crypto/algif_aead.c 					   areq->first_rsgl.sgl.sg, processed);
sgl               245 crypto/algif_aead.c 					   areq->first_rsgl.sgl.sg, outlen);
sgl               269 crypto/algif_aead.c 			struct af_alg_sgl *sgl_prev = &areq->last_rsgl->sgl;
sgl               281 crypto/algif_aead.c 			       areq->first_rsgl.sgl.sg, used, ctx->iv);
sgl                20 crypto/algif_hash.c 	struct af_alg_sgl sgl;
sgl                94 crypto/algif_hash.c 		len = af_alg_make_sg(&ctx->sgl, &msg->msg_iter, len);
sgl               100 crypto/algif_hash.c 		ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, NULL, len);
sgl               104 crypto/algif_hash.c 		af_alg_free_sg(&ctx->sgl);
sgl               143 crypto/algif_hash.c 	sg_init_table(ctx->sgl.sg, 1);
sgl               144 crypto/algif_hash.c 	sg_set_page(ctx->sgl.sg, page, size, offset);
sgl               153 crypto/algif_hash.c 	ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, ctx->result, size);
sgl               112 crypto/algif_skcipher.c 				   areq->first_rsgl.sgl.sg, len, ctx->iv);
sgl               493 crypto/testmgr.c 	struct scatterlist sgl[XBUFSIZE];
sgl               542 crypto/testmgr.c 	BUILD_BUG_ON(ARRAY_SIZE(partitions) != ARRAY_SIZE(tsgl->sgl));
sgl               569 crypto/testmgr.c 	sg_init_table(tsgl->sgl, tsgl->nents);
sgl               585 crypto/testmgr.c 		sg_set_buf(&tsgl->sgl[i], addr, partitions[i].length);
sgl               605 crypto/testmgr.c 	sg_mark_end(&tsgl->sgl[tsgl->nents - 1]);
sgl               606 crypto/testmgr.c 	tsgl->sgl_ptr = tsgl->sgl;
sgl               607 crypto/testmgr.c 	memcpy(tsgl->sgl_saved, tsgl->sgl, tsgl->nents * sizeof(tsgl->sgl[0]));
sgl               665 crypto/testmgr.c 		if (tsgl->sgl[i].page_link != tsgl->sgl_saved[i].page_link)
sgl               667 crypto/testmgr.c 		if (tsgl->sgl[i].offset != tsgl->sgl_saved[i].offset)
sgl               669 crypto/testmgr.c 		if (tsgl->sgl[i].length != tsgl->sgl_saved[i].length)
sgl               733 crypto/testmgr.c 		tsgls->dst.sgl_ptr = tsgls->src.sgl;
sgl              1142 crypto/testmgr.c 		err = crypto_shash_digest(desc, sg_data(&tsgl->sgl[0]),
sgl              1143 crypto/testmgr.c 					  tsgl->sgl[0].length, result);
sgl              1178 crypto/testmgr.c 			err = crypto_shash_finup(desc, sg_data(&tsgl->sgl[i]),
sgl              1179 crypto/testmgr.c 						 tsgl->sgl[i].length, result);
sgl              1190 crypto/testmgr.c 		err = crypto_shash_update(desc, sg_data(&tsgl->sgl[i]),
sgl              1191 crypto/testmgr.c 					  tsgl->sgl[i].length);
sgl              1327 crypto/testmgr.c 		ahash_request_set_crypt(req, tsgl->sgl, result, vec->psize);
sgl              1401 crypto/testmgr.c 			pending_sgl = &tsgl->sgl[i];
sgl              1402 crypto/testmgr.c 		pending_len += tsgl->sgl[i].length;
sgl              1968 crypto/testmgr.c 	if (tsgls->dst.sgl_ptr != tsgls->src.sgl &&
sgl              2554 crypto/testmgr.c 	if (tsgls->dst.sgl_ptr != tsgls->src.sgl &&
sgl              1549 drivers/ata/libata-core.c 			      int dma_dir, struct scatterlist *sgl,
sgl              1609 drivers/ata/libata-core.c 		for_each_sg(sgl, sg, n_elem, i)
sgl              1612 drivers/ata/libata-core.c 		ata_sg_init(qc, sgl, n_elem);
sgl               623 drivers/block/skd_main.c 	struct scatterlist *sgl = &skreq->sg[0], *sg;
sgl               632 drivers/block/skd_main.c 	n_sg = blk_rq_map_sg(skdev->queue, req, sgl);
sgl               640 drivers/block/skd_main.c 	n_sg = dma_map_sg(&skdev->pdev->dev, sgl, n_sg, skreq->data_dir);
sgl               648 drivers/block/skd_main.c 	for_each_sg(sgl, sg, n_sg, i) {
sgl               110 drivers/char/agp/intel-gtt.c 	for_each_sg(st->sgl, sg, num_entries, i)
sgl               114 drivers/char/agp/intel-gtt.c 			st->sgl, st->nents, PCI_DMA_BIDIRECTIONAL))
sgl               132 drivers/char/agp/intel-gtt.c 	st.sgl = sg_list;
sgl               867 drivers/char/agp/intel-gtt.c 	for_each_sg(st->sgl, sg, st->nents, i) {
sgl               935 drivers/char/agp/intel-gtt.c 		mem->sg_list = st.sgl;
sgl               867 drivers/char/virtio_console.c 	struct sg_list *sgl = sd->u.data;
sgl               870 drivers/char/virtio_console.c 	if (sgl->n == sgl->size)
sgl               880 drivers/char/virtio_console.c 		sg_set_page(&(sgl->sg[sgl->n]), buf->page, len, buf->offset);
sgl               899 drivers/char/virtio_console.c 		sg_set_page(&(sgl->sg[sgl->n]), page, len, offset);
sgl               901 drivers/char/virtio_console.c 	sgl->n++;
sgl               902 drivers/char/virtio_console.c 	sgl->len += len;
sgl               913 drivers/char/virtio_console.c 	struct sg_list sgl;
sgl               920 drivers/char/virtio_console.c 		.u.data = &sgl,
sgl               952 drivers/char/virtio_console.c 	sgl.n = 0;
sgl               953 drivers/char/virtio_console.c 	sgl.len = 0;
sgl               954 drivers/char/virtio_console.c 	sgl.size = pipe->nrbufs;
sgl               955 drivers/char/virtio_console.c 	sgl.sg = buf->sg;
sgl               956 drivers/char/virtio_console.c 	sg_init_table(sgl.sg, sgl.size);
sgl               960 drivers/char/virtio_console.c 		ret = __send_to_port(port, buf->sg, sgl.n, sgl.len, buf, true);
sgl               199 drivers/crypto/caam/caampkc.c static int caam_rsa_count_leading_zeros(struct scatterlist *sgl,
sgl               209 drivers/crypto/caam/caampkc.c 	ents = sg_nents_for_len(sgl, nbytes);
sgl               213 drivers/crypto/caam/caampkc.c 	sg_miter_start(&miter, sgl, ents, SG_MITER_FROM_SG | flags);
sgl               148 drivers/crypto/ccp/ccp-crypto-aes-cmac.c 		sg = rctx->data_sg.sgl;
sgl               311 drivers/crypto/ccp/ccp-crypto-main.c 	for (sg = table->sgl; sg; sg = sg_next(sg))
sgl               120 drivers/crypto/ccp/ccp-crypto-sha.c 		sg = rctx->data_sg.sgl;
sgl                27 drivers/crypto/ccree/cc_buffer_mgr.c 	struct scatterlist *sgl;
sgl               164 drivers/crypto/ccree/cc_buffer_mgr.c static int cc_render_sg_to_mlli(struct device *dev, struct scatterlist *sgl,
sgl               168 drivers/crypto/ccree/cc_buffer_mgr.c 	struct scatterlist *curr_sgl = sgl;
sgl               218 drivers/crypto/ccree/cc_buffer_mgr.c 			rc = cc_render_sg_to_mlli(dev, entry->sgl, tot_len,
sgl               272 drivers/crypto/ccree/cc_buffer_mgr.c 			    unsigned int nents, struct scatterlist *sgl,
sgl               279 drivers/crypto/ccree/cc_buffer_mgr.c 		index, nents, sgl, data_len, is_last_table);
sgl               281 drivers/crypto/ccree/cc_buffer_mgr.c 	sgl_data->entry[index].sgl = sgl;
sgl               854 drivers/crypto/ccree/cc_buffer_mgr.c 	struct scatterlist *sgl;
sgl               879 drivers/crypto/ccree/cc_buffer_mgr.c 		sgl = sg_next(areq_ctx->src_sgl);
sgl               880 drivers/crypto/ccree/cc_buffer_mgr.c 		if (!sgl)
sgl               882 drivers/crypto/ccree/cc_buffer_mgr.c 		areq_ctx->src_sgl = sgl;
sgl               923 drivers/crypto/ccree/cc_buffer_mgr.c 		sgl = sg_next(areq_ctx->dst_sgl);
sgl               924 drivers/crypto/ccree/cc_buffer_mgr.c 		if (!sgl)
sgl               926 drivers/crypto/ccree/cc_buffer_mgr.c 		areq_ctx->dst_sgl = sgl;
sgl               477 drivers/crypto/chelsio/chcr_algo.c 	walk->sgl = ulp;
sgl               487 drivers/crypto/chelsio/chcr_algo.c 	walk->sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) |
sgl               500 drivers/crypto/chelsio/chcr_algo.c 		walk->sgl->len0 = cpu_to_be32(size);
sgl               501 drivers/crypto/chelsio/chcr_algo.c 		walk->sgl->addr0 = cpu_to_be64(addr);
sgl               537 drivers/crypto/chelsio/chcr_algo.c 		walk->sgl->len0 = cpu_to_be32(sgmin);
sgl               538 drivers/crypto/chelsio/chcr_algo.c 		walk->sgl->addr0 = cpu_to_be64(sg_dma_address(sg) + skip_len);
sgl               195 drivers/crypto/chelsio/chcr_crypto.h 	struct ulptx_sgl *sgl;
sgl               677 drivers/crypto/chelsio/chcr_ipsec.c 	u64 *pos, *end, *before, *sgl;
sgl               740 drivers/crypto/chelsio/chcr_ipsec.c 	sgl = (void *)pos;
sgl               742 drivers/crypto/chelsio/chcr_ipsec.c 		cxgb4_inline_tx_skb(skb, &q->q, sgl);
sgl               747 drivers/crypto/chelsio/chcr_ipsec.c 		cxgb4_write_sgl(skb, &q->q, (void *)sgl, end,
sgl               755 drivers/crypto/chelsio/chcr_ipsec.c 		q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)sgl;
sgl               176 drivers/crypto/hisilicon/sec/sec_algs.c 				     struct scatterlist *sgl,
sgl               189 drivers/crypto/hisilicon/sec/sec_algs.c 	for_each_sg(sgl, sg, count, i) {
sgl               567 drivers/crypto/hisilicon/sec/sec_algs.c static int sec_map_and_split_sg(struct scatterlist *sgl, size_t *split_sizes,
sgl               575 drivers/crypto/hisilicon/sec/sec_algs.c 	count = dma_map_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL);
sgl               591 drivers/crypto/hisilicon/sec/sec_algs.c 	ret = sg_split(sgl, count, 0, steps, split_sizes,
sgl               605 drivers/crypto/hisilicon/sec/sec_algs.c 	dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL);
sgl               614 drivers/crypto/hisilicon/sec/sec_algs.c static void sec_unmap_sg_on_err(struct scatterlist *sgl, int steps,
sgl               625 drivers/crypto/hisilicon/sec/sec_algs.c 	dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL);
sgl                80 drivers/crypto/hisilicon/sgl.c 	pool->sgl = dma_alloc_coherent(dev, size, &pool->sgl_dma, GFP_KERNEL);
sgl                81 drivers/crypto/hisilicon/sgl.c 	if (!pool->sgl)
sgl               101 drivers/crypto/hisilicon/sgl.c 	dma_free_coherent(dev, pool->size, pool->sgl, pool->sgl_dma);
sgl               109 drivers/crypto/hisilicon/sgl.c 	if (!pool || !hw_sgl_dma || index >= pool->count || !pool->sgl)
sgl               113 drivers/crypto/hisilicon/sgl.c 	return (void *)pool->sgl + pool->sgl_size * index;
sgl               118 drivers/crypto/hisilicon/sgl.c static void sg_map_to_hw_sg(struct scatterlist *sgl,
sgl               121 drivers/crypto/hisilicon/sgl.c 	hw_sge->buf = sgl->dma_address;
sgl               122 drivers/crypto/hisilicon/sgl.c 	hw_sge->len = sgl->dma_length;
sgl               148 drivers/crypto/hisilicon/sgl.c 			      struct scatterlist *sgl,
sgl               156 drivers/crypto/hisilicon/sgl.c 	int sg_n = sg_nents(sgl);
sgl               159 drivers/crypto/hisilicon/sgl.c 	if (!dev || !sgl || !pool || !hw_sgl_dma || sg_n > acc_sgl_sge_nr)
sgl               162 drivers/crypto/hisilicon/sgl.c 	ret = dma_map_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL);
sgl               174 drivers/crypto/hisilicon/sgl.c 	for_each_sg(sgl, sg, sg_n, i) {
sgl               186 drivers/crypto/hisilicon/sgl.c 	dma_unmap_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL);
sgl               201 drivers/crypto/hisilicon/sgl.c void hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl,
sgl               204 drivers/crypto/hisilicon/sgl.c 	dma_unmap_sg(dev, sgl, sg_nents(sgl), DMA_BIDIRECTIONAL);
sgl                 7 drivers/crypto/hisilicon/sgl.h 	struct hisi_acc_hw_sgl *sgl;
sgl                16 drivers/crypto/hisilicon/sgl.h 			      struct scatterlist *sgl,
sgl                19 drivers/crypto/hisilicon/sgl.h void hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl,
sgl               404 drivers/crypto/hisilicon/zip/zip_crypto.c static size_t get_gzip_head_size(struct scatterlist *sgl)
sgl               408 drivers/crypto/hisilicon/zip/zip_crypto.c 	sg_copy_to_buffer(sgl, sg_nents(sgl), buf, sizeof(buf));
sgl               153 drivers/crypto/omap-sham.c 	struct scatterlist	sgl[2];
sgl               719 drivers/crypto/omap-sham.c 	sg_init_table(ctx->sgl, 1);
sgl               720 drivers/crypto/omap-sham.c 	sg_set_buf(ctx->sgl, buf, len);
sgl               721 drivers/crypto/omap-sham.c 	ctx->sg = ctx->sgl;
sgl               865 drivers/crypto/omap-sham.c 		sg_init_table(rctx->sgl, 2);
sgl               866 drivers/crypto/omap-sham.c 		sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, rctx->bufcnt);
sgl               868 drivers/crypto/omap-sham.c 		sg_chain(rctx->sgl, 2, req->src);
sgl               870 drivers/crypto/omap-sham.c 		rctx->sg = rctx->sgl;
sgl               875 drivers/crypto/omap-sham.c 		sg_init_table(rctx->sgl, 1);
sgl               876 drivers/crypto/omap-sham.c 		sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, xmit_len);
sgl               878 drivers/crypto/omap-sham.c 		rctx->sg = rctx->sgl;
sgl               708 drivers/crypto/qat/qat_common/qat_algs.c 			       struct scatterlist *sgl,
sgl               714 drivers/crypto/qat/qat_common/qat_algs.c 	int n = sg_nents(sgl);
sgl               734 drivers/crypto/qat/qat_common/qat_algs.c 	for_each_sg(sgl, sg, n, i) {
sgl               753 drivers/crypto/qat/qat_common/qat_algs.c 	if (sgl != sglout) {
sgl               805 drivers/crypto/qat/qat_common/qat_algs.c 	n = sg_nents(sgl);
sgl               111 drivers/crypto/qce/ablkcipher.c 	rctx->dst_sg = rctx->dst_tbl.sgl;
sgl                52 drivers/crypto/qce/dma.c 	struct scatterlist *sg = sgt->sgl, *sg_last = NULL;
sgl               365 drivers/crypto/s5p-sss.c 	struct scatterlist	sgl[2];
sgl              1045 drivers/crypto/s5p-sss.c 	sg_init_table(ctx->sgl, 1);
sgl              1046 drivers/crypto/s5p-sss.c 	sg_set_buf(ctx->sgl, buf, len);
sgl              1047 drivers/crypto/s5p-sss.c 	ctx->sg = ctx->sgl;
sgl              1191 drivers/crypto/s5p-sss.c 		sg_init_table(ctx->sgl, 2);
sgl              1192 drivers/crypto/s5p-sss.c 		sg_set_buf(ctx->sgl, ctx->dd->xmit_buf, ctx->bufcnt);
sgl              1193 drivers/crypto/s5p-sss.c 		sg_chain(ctx->sgl, 2, sg);
sgl              1194 drivers/crypto/s5p-sss.c 		ctx->sg = ctx->sgl;
sgl              1279 drivers/crypto/s5p-sss.c 		sg_init_table(ctx->sgl, 1);
sgl              1280 drivers/crypto/s5p-sss.c 		sg_set_buf(ctx->sgl, ctx->dd->xmit_buf, xmit_len);
sgl              1282 drivers/crypto/s5p-sss.c 		ctx->sg = ctx->sgl;
sgl                64 drivers/dma-buf/udmabuf.c 	if (!dma_map_sg(at->dev, sg->sgl, sg->nents, direction)) {
sgl                80 drivers/dma-buf/udmabuf.c 	dma_unmap_sg(at->dev, sg->sgl, sg->nents, direction);
sgl               385 drivers/dma/altera-msgdma.c msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl,
sgl               401 drivers/dma/altera-msgdma.c 	for_each_sg(sgl, sg, sg_len, i)
sgl               413 drivers/dma/altera-msgdma.c 	avail = sg_dma_len(sgl);
sgl               424 drivers/dma/altera-msgdma.c 			dma_src = sg_dma_address(sgl) + sg_dma_len(sgl) - avail;
sgl               429 drivers/dma/altera-msgdma.c 			dma_dst = sg_dma_address(sgl) + sg_dma_len(sgl) - avail;
sgl               444 drivers/dma/altera-msgdma.c 			sgl = sg_next(sgl);
sgl               445 drivers/dma/altera-msgdma.c 			if (sgl == NULL)
sgl               448 drivers/dma/altera-msgdma.c 			avail = sg_dma_len(sgl);
sgl              2050 drivers/dma/amba-pl08x.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl              2062 drivers/dma/amba-pl08x.c 			__func__, sg_dma_len(sgl), plchan->name);
sgl              2068 drivers/dma/amba-pl08x.c 	for_each_sg(sgl, sg, sg_len, tmp) {
sgl               717 drivers/dma/at_hdmac.c 	first = xt->sgl;
sgl               731 drivers/dma/at_hdmac.c 		struct data_chunk *chunk = xt->sgl + i;
sgl               733 drivers/dma/at_hdmac.c 		if ((chunk->size != xt->sgl->size) ||
sgl               979 drivers/dma/at_hdmac.c 		       struct scatterlist *sgl,
sgl               995 drivers/dma/at_hdmac.c 	if (unlikely(!sgl || !sg_len)) {
sgl              1009 drivers/dma/at_hdmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1064 drivers/dma/at_hdmac.c atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              1105 drivers/dma/at_hdmac.c 		for_each_sg(sgl, sg, sg_len, i) {
sgl              1146 drivers/dma/at_hdmac.c 		for_each_sg(sgl, sg, sg_len, i) {
sgl               633 drivers/dma/at_xdmac.c at_xdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               645 drivers/dma/at_xdmac.c 	if (!sgl)
sgl               665 drivers/dma/at_xdmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               971 drivers/dma/at_xdmac.c 							xt, xt->sgl);
sgl               985 drivers/dma/at_xdmac.c 			chunk = xt->sgl + i;
sgl              1228 drivers/dma/at_xdmac.c at_xdmac_prep_dma_memset_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              1239 drivers/dma/at_xdmac.c 	if (!sgl)
sgl              1246 drivers/dma/at_xdmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               267 drivers/dma/bcm2835-dma.c 	struct scatterlist *sgl,
sgl               275 drivers/dma/bcm2835-dma.c 	for_each_sg(sgl, sgent, sg_len, i)
sgl               389 drivers/dma/bcm2835-dma.c 	struct scatterlist *sgl,
sgl               399 drivers/dma/bcm2835-dma.c 	for_each_sg(sgl, sgent, sg_len, i) {
sgl               644 drivers/dma/bcm2835-dma.c 	struct scatterlist *sgl, unsigned int sg_len,
sgl               677 drivers/dma/bcm2835-dma.c 	frames = bcm2835_dma_count_frames_for_sg(c, sgl, sg_len);
sgl               689 drivers/dma/bcm2835-dma.c 					  sgl, sg_len);
sgl              2268 drivers/dma/coh901318.c coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              2287 drivers/dma/coh901318.c 	if (!sgl)
sgl              2289 drivers/dma/coh901318.c 	if (sg_dma_len(sgl) == 0)
sgl              2335 drivers/dma/coh901318.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              2360 drivers/dma/coh901318.c 	ret = coh901318_lli_fill_sg(&cohc->base->pool, lli, sgl, sg_len,
sgl               233 drivers/dma/coh901318_lli.c 		      struct scatterlist *sgl, unsigned int nents,
sgl               258 drivers/dma/coh901318_lli.c 	for_each_sg(sgl, sg, nents, i) {
sgl               527 drivers/dma/dma-axi-dmac.c 	struct dma_chan *c, struct scatterlist *sgl,
sgl               542 drivers/dma/dma-axi-dmac.c 	for_each_sg(sgl, sg, sg_len, i)
sgl               551 drivers/dma/dma-axi-dmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               625 drivers/dma/dma-axi-dmac.c 	dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]);
sgl               626 drivers/dma/dma-axi-dmac.c 	src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]);
sgl               629 drivers/dma/dma-axi-dmac.c 		if (!axi_dmac_check_len(chan, xt->sgl[0].size) ||
sgl               632 drivers/dma/dma-axi-dmac.c 		if (xt->sgl[0].size + dst_icg > chan->max_length ||
sgl               633 drivers/dma/dma-axi-dmac.c 		    xt->sgl[0].size + src_icg > chan->max_length)
sgl               638 drivers/dma/dma-axi-dmac.c 		if (chan->max_length / xt->sgl[0].size < xt->numf)
sgl               640 drivers/dma/dma-axi-dmac.c 		if (!axi_dmac_check_len(chan, xt->sgl[0].size * xt->numf))
sgl               650 drivers/dma/dma-axi-dmac.c 		desc->sg[0].src_stride = xt->sgl[0].size + src_icg;
sgl               655 drivers/dma/dma-axi-dmac.c 		desc->sg[0].dest_stride = xt->sgl[0].size + dst_icg;
sgl               659 drivers/dma/dma-axi-dmac.c 		desc->sg[0].x_len = xt->sgl[0].size;
sgl               662 drivers/dma/dma-axi-dmac.c 		desc->sg[0].x_len = xt->sgl[0].size * xt->numf;
sgl               354 drivers/dma/dma-jz4780.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
sgl               370 drivers/dma/dma-jz4780.c 					      sg_dma_address(&sgl[i]),
sgl               371 drivers/dma/dma-jz4780.c 					      sg_dma_len(&sgl[i]),
sgl               364 drivers/dma/dw-edma/dw-edma-core.c 		sg = xfer->xfer.sg.sgl;
sgl               435 drivers/dma/dw-edma/dw-edma-core.c dw_edma_device_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl,
sgl               444 drivers/dma/dw-edma/dw-edma-core.c 	xfer.xfer.sg.sgl = sgl;
sgl               132 drivers/dma/dw-edma/dw-edma-core.h 	struct scatterlist		*sgl;
sgl               622 drivers/dma/dw/core.c dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               663 drivers/dma/dw/core.c 		for_each_sg(sgl, sg, sg_len, i) {
sgl               713 drivers/dma/dw/core.c 		for_each_sg(sgl, sg, sg_len, i) {
sgl              1033 drivers/dma/ep93xx_dma.c ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              1057 drivers/dma/ep93xx_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               512 drivers/dma/fsl-edma-common.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               547 drivers/dma/fsl-edma-common.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               239 drivers/dma/fsl-edma-common.h 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               259 drivers/dma/hsu/hsu.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               272 drivers/dma/hsu/hsu.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               293 drivers/dma/idma64.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               306 drivers/dma/idma64.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               448 drivers/dma/img-mdc-dma.c 	struct dma_chan *chan, struct scatterlist *sgl,
sgl               460 drivers/dma/img-mdc-dma.c 	if (!sgl)
sgl               474 drivers/dma/img-mdc-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               816 drivers/dma/imx-dma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               831 drivers/dma/imx-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               837 drivers/dma/imx-dma.c 		if (sg_dma_len(sgl) & 3 || sgl->dma_address & 3)
sgl               841 drivers/dma/imx-dma.c 		if (sg_dma_len(sgl) & 1 || sgl->dma_address & 1)
sgl               851 drivers/dma/imx-dma.c 	desc->sg = sgl;
sgl               982 drivers/dma/imx-dma.c 	desc->x = xt->sgl[0].size;
sgl               984 drivers/dma/imx-dma.c 	desc->w = xt->sgl[0].icg + desc->x;
sgl              1441 drivers/dma/imx-sdma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl              1461 drivers/dma/imx-sdma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1334 drivers/dma/ipu/ipu_idmac.c 		struct scatterlist *sgl, unsigned int sg_len,
sgl              1363 drivers/dma/ipu/ipu_idmac.c 		desc->sg	= sgl;
sgl               533 drivers/dma/k3dma.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen,
sgl               543 drivers/dma/k3dma.c 	if (sgl == NULL)
sgl               548 drivers/dma/k3dma.c 	for_each_sg(sgl, sg, sglen, i) {
sgl               560 drivers/dma/k3dma.c 	for_each_sg(sgl, sg, sglen, i) {
sgl               340 drivers/dma/mediatek/mtk-uart-apdma.c 	(struct dma_chan *chan, struct scatterlist *sgl,
sgl               355 drivers/dma/mediatek/mtk-uart-apdma.c 	d->avail_len = sg_dma_len(sgl);
sgl               356 drivers/dma/mediatek/mtk-uart-apdma.c 	d->addr = sg_dma_address(sgl);
sgl               526 drivers/dma/mmp_pdma.c mmp_pdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl,
sgl               537 drivers/dma/mmp_pdma.c 	if ((sgl == NULL) || (sg_len == 0))
sgl               544 drivers/dma/mmp_pdma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               546 drivers/dma/mmp_pdma.c 		avail = sg_dma_len(sgl);
sgl               270 drivers/dma/moxart-dma.c 	struct dma_chan *chan, struct scatterlist *sgl,
sgl               320 drivers/dma/moxart-dma.c 	for_each_sg(sgl, sgent, sg_len, i) {
sgl               684 drivers/dma/mpc512x_dma.c mpc_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               706 drivers/dma/mpc512x_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               495 drivers/dma/mxs-dma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               537 drivers/dma/mxs-dma.c 		pio = (u32 *) sgl;
sgl               554 drivers/dma/mxs-dma.c 		for_each_sg(sgl, sg, sg_len, i) {
sgl              1007 drivers/dma/nbpfaxi.c 	struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len,
sgl              1020 drivers/dma/nbpfaxi.c 		return nbpf_prep_sg(chan, sgl, &slave_sg, sg_len,
sgl              1025 drivers/dma/nbpfaxi.c 		return nbpf_prep_sg(chan, &slave_sg, sgl, sg_len,
sgl               876 drivers/dma/owl-dma.c 				       struct scatterlist *sgl,
sgl               897 drivers/dma/owl-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               567 drivers/dma/pch_dma.c 			struct scatterlist *sgl, unsigned int sg_len,
sgl               595 drivers/dma/pch_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              2829 drivers/dma/pl330.c pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              2838 drivers/dma/pl330.c 	if (unlikely(!pch || !sgl || !sg_len))
sgl              2848 drivers/dma/pl330.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               971 drivers/dma/pxa_dma.c pxad_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl,
sgl               983 drivers/dma/pxa_dma.c 	if ((sgl == NULL) || (sg_len == 0))
sgl               990 drivers/dma/pxa_dma.c 	for_each_sg(sgl, sg, sg_len, i)
sgl               996 drivers/dma/pxa_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               607 drivers/dma/qcom/bam_dma.c 	struct scatterlist *sgl, unsigned int sg_len,
sgl               626 drivers/dma/qcom/bam_dma.c 	for_each_sg(sgl, sg, sg_len, i)
sgl               648 drivers/dma/qcom/bam_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               975 drivers/dma/s3c24xx-dma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               991 drivers/dma/s3c24xx-dma.c 			sg_dma_len(sgl), s3cchan->name);
sgl              1042 drivers/dma/s3c24xx-dma.c 	for_each_sg(sgl, sg, sg_len, tmp) {
sgl               894 drivers/dma/sh/rcar-dmac.c rcar_dmac_chan_prep_sg(struct rcar_dmac_chan *chan, struct scatterlist *sgl,
sgl               930 drivers/dma/sh/rcar-dmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1097 drivers/dma/sh/rcar-dmac.c 	struct scatterlist sgl;
sgl              1102 drivers/dma/sh/rcar-dmac.c 	sg_init_table(&sgl, 1);
sgl              1103 drivers/dma/sh/rcar-dmac.c 	sg_set_page(&sgl, pfn_to_page(PFN_DOWN(dma_src)), len,
sgl              1105 drivers/dma/sh/rcar-dmac.c 	sg_dma_address(&sgl) = dma_src;
sgl              1106 drivers/dma/sh/rcar-dmac.c 	sg_dma_len(&sgl) = len;
sgl              1108 drivers/dma/sh/rcar-dmac.c 	return rcar_dmac_chan_prep_sg(rchan, &sgl, 1, dma_dest,
sgl              1166 drivers/dma/sh/rcar-dmac.c rcar_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              1173 drivers/dma/sh/rcar-dmac.c 	if (rchan->mid_rid < 0 || !sg_len || !sg_dma_len(sgl)) {
sgl              1183 drivers/dma/sh/rcar-dmac.c 	return rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr,
sgl              1196 drivers/dma/sh/rcar-dmac.c 	struct scatterlist *sgl;
sgl              1223 drivers/dma/sh/rcar-dmac.c 	sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT);
sgl              1224 drivers/dma/sh/rcar-dmac.c 	if (!sgl)
sgl              1227 drivers/dma/sh/rcar-dmac.c 	sg_init_table(sgl, sg_len);
sgl              1232 drivers/dma/sh/rcar-dmac.c 		sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
sgl              1234 drivers/dma/sh/rcar-dmac.c 		sg_dma_address(&sgl[i]) = src;
sgl              1235 drivers/dma/sh/rcar-dmac.c 		sg_dma_len(&sgl[i]) = period_len;
sgl              1238 drivers/dma/sh/rcar-dmac.c 	desc = rcar_dmac_chan_prep_sg(rchan, sgl, sg_len, rchan->map.addr,
sgl              1241 drivers/dma/sh/rcar-dmac.c 	kfree(sgl);
sgl               561 drivers/dma/sh/shdma-base.c 	struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr,
sgl               571 drivers/dma/sh/shdma-base.c 	for_each_sg(sgl, sg, sg_len, i)
sgl               588 drivers/dma/sh/shdma-base.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               662 drivers/dma/sh/shdma-base.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
sgl               685 drivers/dma/sh/shdma-base.c 	return shdma_prep_sg(schan, sgl, sg_len, &slave_addr,
sgl               703 drivers/dma/sh/shdma-base.c 	struct scatterlist *sgl;
sgl               731 drivers/dma/sh/shdma-base.c 	sgl = kcalloc(sg_len, sizeof(*sgl), GFP_KERNEL);
sgl               732 drivers/dma/sh/shdma-base.c 	if (!sgl)
sgl               735 drivers/dma/sh/shdma-base.c 	sg_init_table(sgl, sg_len);
sgl               740 drivers/dma/sh/shdma-base.c 		sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
sgl               742 drivers/dma/sh/shdma-base.c 		sg_dma_address(&sgl[i]) = src;
sgl               743 drivers/dma/sh/shdma-base.c 		sg_dma_len(&sgl[i]) = period_len;
sgl               746 drivers/dma/sh/shdma-base.c 	desc = shdma_prep_sg(schan, sgl, sg_len, &slave_addr,
sgl               749 drivers/dma/sh/shdma-base.c 	kfree(sgl);
sgl               416 drivers/dma/sh/usb-dmac.c usb_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               437 drivers/dma/sh/usb-dmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               728 drivers/dma/sirf-dma.c 		sdesc->xlen = xt->sgl[0].size / SIRFSOC_DMA_WORD_LEN;
sgl               729 drivers/dma/sirf-dma.c 		sdesc->width = (xt->sgl[0].size + xt->sgl[0].icg) /
sgl               920 drivers/dma/sprd-dma.c sprd_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               963 drivers/dma/sprd-dma.c 	for_each_sg(sgl, sg, sglen, i) {
sgl               506 drivers/dma/st_fdma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               520 drivers/dma/st_fdma.c 	if (!sgl)
sgl               531 drivers/dma/st_fdma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1020 drivers/dma/ste_dma40.c static int d40_sg_2_dmalen(struct scatterlist *sgl, int sg_len,
sgl              1028 drivers/dma/ste_dma40.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              2498 drivers/dma/ste_dma40.c d40_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl              2505 drivers/dma/ste_dma40.c 	return d40_prep_sg(chan, sgl, sgl, sg_len, direction, dma_flags);
sgl               829 drivers/dma/stm32-dma.c 	struct dma_chan *c, struct scatterlist *sgl,
sgl               860 drivers/dma/stm32-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               728 drivers/dma/stm32-mdma.c 				 struct scatterlist *sgl, u32 sg_len,
sgl               738 drivers/dma/stm32-mdma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               781 drivers/dma/stm32-mdma.c stm32_mdma_prep_slave_sg(struct dma_chan *c, struct scatterlist *sgl,
sgl               804 drivers/dma/stm32-mdma.c 	ret = stm32_mdma_setup_xfer(chan, desc, sgl, sg_len, direction);
sgl               767 drivers/dma/sun4i-dma.c sun4i_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               781 drivers/dma/sun4i-dma.c 	if (!sgl)
sgl               814 drivers/dma/sun4i-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               683 drivers/dma/sun6i-dma.c 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               697 drivers/dma/sun6i-dma.c 	if (!sgl)
sgl               710 drivers/dma/sun6i-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1001 drivers/dma/tegra20-apb-dma.c 	struct dma_chan *dc, struct scatterlist *sgl, unsigned int sg_len,
sgl              1064 drivers/dma/tegra20-apb-dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               585 drivers/dma/ti/cppi41.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned sg_len,
sgl               607 drivers/dma/ti/cppi41.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1054 drivers/dma/ti/edma.c 	struct dma_chan *chan, struct scatterlist *sgl,
sgl              1067 drivers/dma/ti/edma.c 	if (unlikely(!echan || !sgl || !sg_len))
sgl              1114 drivers/dma/ti/edma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               886 drivers/dma/ti/omap-dma.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned sglen,
sgl              1027 drivers/dma/ti/omap-dma.c 	for_each_sg(sgl, sgent, sglen, i) {
sgl              1235 drivers/dma/ti/omap-dma.c 	data_type = __ffs((xt->src_start | xt->dst_start | xt->sgl[0].size));
sgl              1243 drivers/dma/ti/omap-dma.c 	sg->en = xt->sgl[0].size / BIT(data_type);
sgl              1249 drivers/dma/ti/omap-dma.c 	src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]);
sgl              1250 drivers/dma/ti/omap-dma.c 	dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]);
sgl               490 drivers/dma/timb_dma.c 	struct scatterlist *sgl, unsigned int sg_len,
sgl               501 drivers/dma/timb_dma.c 	if (!sgl || !sg_len) {
sgl               521 drivers/dma/timb_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               803 drivers/dma/txx9dmac.c txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               827 drivers/dma/txx9dmac.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl                53 drivers/dma/uniphier-mdmac.c 	struct scatterlist *sgl;
sgl               114 drivers/dma/uniphier-mdmac.c 	sg = &md->sgl[md->sg_cur];
sgl               230 drivers/dma/uniphier-mdmac.c uniphier_mdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
sgl               245 drivers/dma/uniphier-mdmac.c 	md->sgl = sgl;
sgl               320 drivers/dma/uniphier-mdmac.c 			txstate->residue += sg_dma_len(&md->sgl[i]);
sgl              1648 drivers/dma/xilinx/xilinx_dma.c 	if (!xt->numf || !xt->sgl[0].size)
sgl              1671 drivers/dma/xilinx/xilinx_dma.c 	hw->hsize = xt->sgl[0].size;
sgl              1672 drivers/dma/xilinx/xilinx_dma.c 	hw->stride = (xt->sgl[0].icg + xt->sgl[0].size) <<
sgl              1776 drivers/dma/xilinx/xilinx_dma.c 	struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len,
sgl              1801 drivers/dma/xilinx/xilinx_dma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl              1992 drivers/dma/xilinx/xilinx_dma.c 	if (!xt->numf || !xt->sgl[0].size)
sgl              2023 drivers/dma/xilinx/xilinx_dma.c 	hw->vsize_stride |= (xt->sgl[0].icg + xt->sgl[0].size) &
sgl              2025 drivers/dma/xilinx/xilinx_dma.c 	hw->control = xt->sgl[0].size & XILINX_DMA_BD_HSIZE_MASK;
sgl               546 drivers/dma/zx_dma.c 	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen,
sgl               556 drivers/dma/zx_dma.c 	if (!sgl)
sgl               562 drivers/dma/zx_dma.c 	for_each_sg(sgl, sg, sglen, i) {
sgl               574 drivers/dma/zx_dma.c 	for_each_sg(sgl, sg, sglen, i) {
sgl               112 drivers/fpga/fpga-mgr.c 	sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG);
sgl               127 drivers/fpga/fpga-mgr.c 	len = sg_copy_to_buffer(sgt->sgl, sgt->nents, buf,
sgl               191 drivers/fpga/fpga-mgr.c 		sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG);
sgl               400 drivers/fpga/zynq-fpga.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl               409 drivers/fpga/zynq-fpga.c 	    dma_map_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE);
sgl               426 drivers/fpga/zynq-fpga.c 	priv->cur_sg = sgt->sgl;
sgl               481 drivers/fpga/zynq-fpga.c 	dma_unmap_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE);
sgl               781 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	sg->sgl->dma_address = addr;
sgl               782 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	sg->sgl->length = size;
sgl               784 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	sg->sgl->dma_length = size;
sgl               958 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	nents = dma_map_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
sgl               986 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	if (!ttm->sg->sgl)
sgl               990 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	dma_unmap_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
sgl               341 drivers/gpu/drm/arm/malidp_planes.c 		struct scatterlist *sgl;
sgl               354 drivers/gpu/drm/arm/malidp_planes.c 		sgl = sgt->sgl;
sgl               356 drivers/gpu/drm/arm/malidp_planes.c 		while (sgl) {
sgl               357 drivers/gpu/drm/arm/malidp_planes.c 			if (sgl->length < pgsize) {
sgl               363 drivers/gpu/drm/arm/malidp_planes.c 			sgl = sg_next(sgl);
sgl               398 drivers/gpu/drm/armada/armada_gem.c 		for_each_sg(sgt->sgl, sg, count, i) {
sgl               410 drivers/gpu/drm/armada/armada_gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) {
sgl               419 drivers/gpu/drm/armada/armada_gem.c 		sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0);
sgl               421 drivers/gpu/drm/armada/armada_gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0)
sgl               427 drivers/gpu/drm/armada/armada_gem.c 		sg_dma_address(sgt->sgl) = dobj->dev_addr;
sgl               428 drivers/gpu/drm/armada/armada_gem.c 		sg_dma_len(sgt->sgl) = dobj->obj.size;
sgl               435 drivers/gpu/drm/armada/armada_gem.c 	for_each_sg(sgt->sgl, sg, num, i)
sgl               452 drivers/gpu/drm/armada/armada_gem.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
sgl               456 drivers/gpu/drm/armada/armada_gem.c 		for_each_sg(sgt->sgl, sg, sgt->nents, i)
sgl               557 drivers/gpu/drm/armada/armada_gem.c 	if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) {
sgl               561 drivers/gpu/drm/armada/armada_gem.c 	dobj->dev_addr = sg_dma_address(dobj->sgt->sgl);
sgl               129 drivers/gpu/drm/drm_cache.c 		for_each_sg_page(st->sgl, &sg_iter, st->nents, 0)
sgl               476 drivers/gpu/drm/drm_gem_cma_helper.c 		dma_addr_t next_addr = sg_dma_address(sgt->sgl);
sgl               480 drivers/gpu/drm/drm_gem_cma_helper.c 		for_each_sg(sgt->sgl, s, sgt->nents, i) {
sgl               500 drivers/gpu/drm/drm_gem_cma_helper.c 	cma_obj->paddr = sg_dma_address(sgt->sgl);
sgl               120 drivers/gpu/drm/drm_gem_shmem_helper.c 			dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl,
sgl               393 drivers/gpu/drm/drm_gem_shmem_helper.c 	dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl,
sgl               628 drivers/gpu/drm/drm_gem_shmem_helper.c 	dma_map_sg(obj->dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgl               627 drivers/gpu/drm/drm_prime.c 	if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgl               653 drivers/gpu/drm/drm_prime.c 	dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgl               957 drivers/gpu/drm/drm_prime.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgl                30 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_map_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgl                54 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_unmap_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgl               407 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_sync_sg_for_cpu(dev->dev, etnaviv_obj->sgt->sgl,
sgl               424 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_sync_sg_for_device(dev->dev, etnaviv_obj->sgt->sgl,
sgl                82 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl               100 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, i, j) {
sgl               116 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl               247 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		iova = sg_dma_address(sgt->sgl) - memory_base;
sgl               248 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) {
sgl               398 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	dma_unmap_sg(to_dma_dev(g2d->drm_dev), g2d_userptr->sgt->sgl,
sgl               514 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (!dma_map_sg(to_dma_dev(g2d->drm_dev), sgt->sgl, sgt->nents,
sgl               521 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_userptr->dma_addr = sgt->sgl[0].dma_address;
sgl               487 drivers/gpu/drm/exynos/exynos_drm_gem.c 	exynos_gem->dma_addr = sg_dma_address(sgt->sgl);
sgl              10513 drivers/gpu/drm/i915/display/intel_display.c 		base = sg_dma_address(obj->mm.pages->sgl);
sgl              1323 drivers/gpu/drm/i915/display/intel_overlay.c 		overlay->flip_addr = sg_dma_address(obj->mm.pages->sgl);
sgl                43 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c 	src = obj->mm.pages->sgl;
sgl                44 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c 	dst = st->sgl;
sgl                51 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c 	if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) {
sgl                74 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c 	dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir);
sgl               237 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c 	sg_page_sizes = i915_sg_page_sizes(pages->sgl);
sgl                26 drivers/gpu/drm/i915/gem/i915_gem_internal.c 	for (sg = st->sgl; sg; sg = __sg_next(sg)) {
sgl                77 drivers/gpu/drm/i915/gem/i915_gem_internal.c 	sg = st->sgl;
sgl               112 drivers/gpu/drm/i915/gem/i915_gem_internal.c 		if (get_order(st->sgl->length)) {
sgl                29 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	obj->mm.get_page.sg_pos = pages->sgl;
sgl               242 drivers/gpu/drm/i915/gem/i915_gem_pages.c 		return kmap(sg_page(sgt->sgl));
sgl                52 drivers/gpu/drm/i915/gem/i915_gem_phys.c 	sg = st->sgl;
sgl                97 drivers/gpu/drm/i915/gem/i915_gem_phys.c 	dma_addr_t dma = sg_dma_address(pages->sgl);
sgl                98 drivers/gpu/drm/i915/gem/i915_gem_phys.c 	void *vaddr = sg_page(pages->sgl);
sgl               204 drivers/gpu/drm/i915/gem/i915_gem_phys.c 		unsigned int sg_page_sizes = i915_sg_page_sizes(pages->sgl);
sgl                79 drivers/gpu/drm/i915/gem/i915_gem_shmem.c 	sg = st->sgl;
sgl               497 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	sg = st->sgl;
sgl               464 drivers/gpu/drm/i915/gem/i915_gem_userptr.c 	sg_page_sizes = i915_sg_page_sizes(st->sgl);
sgl                17 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c 	for (sg = pages->sgl; sg && nreal--; sg = __sg_next(sg))
sgl                42 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c 	sg = pages->sgl;
sgl                56 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c 		for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) {
sgl                60 drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c 				src = pages->sgl;
sgl                47 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	for (sg = st->sgl; sg; sg = __sg_next(sg)) {
sgl                75 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	sg = st->sgl;
sgl               200 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	sg = st->sgl;
sgl               253 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	sg = st->sgl;
sgl                25 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c 	sg = st->sgl;
sgl                31 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c 	if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) {
sgl                49 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c 	dma_unmap_sg(attachment->dev, st->sgl, st->nents, dir);
sgl               522 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	return sg_page(obj->mm.pages->sgl);
sgl                24 drivers/gpu/drm/i915/gt/selftest_timeline.c 	return sg_page(obj->mm.pages->sgl);
sgl               370 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	desc->db_trigger_phy = sg_dma_address(client->vma->pages->sgl) +
sgl               595 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	return sg_pcopy_to_buffer(pages->sgl, pages->nents, dst, size, offset);
sgl                66 drivers/gpu/drm/i915/gvt/dmabuf.c 	for_each_sg(st->sgl, sg, page_num, i) {
sgl               139 drivers/gpu/drm/i915/i915_gem.c 	void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset;
sgl              1161 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct scatterlist *sg = vma->pages->sgl;
sgl              2148 drivers/gpu/drm/i915/i915_gem_gtt.c 				     pages->sgl, pages->nents,
sgl              2542 drivers/gpu/drm/i915/i915_gem_gtt.c 	dma_unmap_sg(kdev, pages->sgl, pages->nents, PCI_DMA_BIDIRECTIONAL);
sgl              3405 drivers/gpu/drm/i915/i915_gem_gtt.c 	sg = st->sgl;
sgl              3486 drivers/gpu/drm/i915/i915_gem_gtt.c 	sg = st->sgl;
sgl              3529 drivers/gpu/drm/i915/i915_gem_gtt.c 	sg = st->sgl;
sgl                78 drivers/gpu/drm/i915/i915_gpu_error.c 		struct scatterlist *sgl;
sgl                80 drivers/gpu/drm/i915/i915_gpu_error.c 		sgl = (typeof(sgl))__get_free_page(ALLOW_FAIL);
sgl                81 drivers/gpu/drm/i915/i915_gpu_error.c 		if (!sgl) {
sgl                90 drivers/gpu/drm/i915/i915_gpu_error.c 				(unsigned long)sgl | SG_CHAIN;
sgl                92 drivers/gpu/drm/i915/i915_gpu_error.c 			e->sgl = sgl;
sgl                95 drivers/gpu/drm/i915/i915_gpu_error.c 		e->cur = sgl;
sgl                96 drivers/gpu/drm/i915/i915_gpu_error.c 		e->end = sgl + SG_MAX_SINGLE_ALLOC - 1;
sgl               639 drivers/gpu/drm/i915/i915_gpu_error.c static void err_free_sgl(struct scatterlist *sgl)
sgl               641 drivers/gpu/drm/i915/i915_gpu_error.c 	while (sgl) {
sgl               644 drivers/gpu/drm/i915/i915_gpu_error.c 		for (sg = sgl; !sg_is_chain(sg); sg++) {
sgl               651 drivers/gpu/drm/i915/i915_gpu_error.c 		free_page((unsigned long)sgl);
sgl               652 drivers/gpu/drm/i915/i915_gpu_error.c 		sgl = sg;
sgl               798 drivers/gpu/drm/i915/i915_gpu_error.c 	if (READ_ONCE(error->sgl))
sgl               815 drivers/gpu/drm/i915/i915_gpu_error.c 	GEM_BUG_ON(m.sgl && !m.cur);
sgl               818 drivers/gpu/drm/i915/i915_gpu_error.c 		err_free_sgl(m.sgl);
sgl               822 drivers/gpu/drm/i915/i915_gpu_error.c 	if (cmpxchg(&error->sgl, NULL, m.sgl))
sgl               823 drivers/gpu/drm/i915/i915_gpu_error.c 		err_free_sgl(m.sgl);
sgl               845 drivers/gpu/drm/i915/i915_gpu_error.c 		sg = error->sgl;
sgl               952 drivers/gpu/drm/i915/i915_gpu_error.c 	err_free_sgl(error->sgl);
sgl               166 drivers/gpu/drm/i915/i915_gpu_error.h 	struct scatterlist *sgl, *fit;
sgl               186 drivers/gpu/drm/i915/i915_gpu_error.h 	struct scatterlist *sgl, *cur, *end;
sgl                21 drivers/gpu/drm/i915/i915_scatterlist.c 	new_sg = new_st.sgl;
sgl                22 drivers/gpu/drm/i915/i915_scatterlist.c 	for_each_sg(orig_st->sgl, sg, orig_st->nents, i) {
sgl                27 drivers/gpu/drm/i915/i915_scatterlist.h } __sgt_iter(struct scatterlist *sgl, bool dma) {
sgl                28 drivers/gpu/drm/i915/i915_scatterlist.h 	struct sgt_iter s = { .sgp = sgl };
sgl                77 drivers/gpu/drm/i915/i915_scatterlist.h 	for ((__iter) = __sgt_iter((__sgt)->sgl, true);			\
sgl                89 drivers/gpu/drm/i915/i915_scatterlist.h 	for ((__iter) = __sgt_iter((__sgt)->sgl, false);		\
sgl               406 drivers/gpu/drm/i915/i915_vma.h 	return sg_page(vma->pages->sgl);
sgl                78 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	for (sg = pages->sgl; sg; sg = sg_next(sg)) {
sgl               580 drivers/gpu/drm/i915/selftests/i915_vma.c 					sg = vma->pages->sgl;
sgl                51 drivers/gpu/drm/i915/selftests/scatterlist.c 	for_each_sg(pt->st.sgl, sg, pt->st.nents, n) {
sgl                89 drivers/gpu/drm/i915/selftests/scatterlist.c 	for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) {
sgl               236 drivers/gpu/drm/i915/selftests/scatterlist.c 	sg = pt->st.sgl;
sgl               225 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	expected = sg_dma_address(sg->sgl);
sgl               226 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	for_each_sg(sg->sgl, s, sg->nents, i) {
sgl               235 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	mtk_gem->dma_addr = sg_dma_address(sg->sgl);
sgl               265 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) {
sgl                56 drivers/gpu/drm/msm/msm_gem.c 		dma_sync_sg_for_device(dev, msm_obj->sgt->sgl,
sgl                59 drivers/gpu/drm/msm/msm_gem.c 		dma_map_sg(dev, msm_obj->sgt->sgl,
sgl                69 drivers/gpu/drm/msm/msm_gem.c 		dma_sync_sg_for_cpu(dev, msm_obj->sgt->sgl,
sgl                72 drivers/gpu/drm/msm/msm_gem.c 		dma_unmap_sg(dev, msm_obj->sgt->sgl,
sgl                49 drivers/gpu/drm/msm/msm_gpummu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl                48 drivers/gpu/drm/msm/msm_iommu.c 	ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot);
sgl                20 drivers/gpu/drm/nouveau/include/nvif/if000a.h 	struct scatterlist *sgl;
sgl                67 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h 	struct scatterlist *sgl;
sgl               119 drivers/gpu/drm/nouveau/nouveau_mem.c 	if (tt->ttm.sg) args.sgl = tt->ttm.sg->sgl;
sgl                37 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c 		struct scatterlist *sgl;
sgl               117 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c 		.sgl = mem->sgl,
sgl               178 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c 			mem->sgl = args->v0.sgl;
sgl              1476 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (map->sgl) {
sgl              1477 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		for (; map->off; map->sgl = sg_next(map->sgl)) {
sgl              1478 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			u64 size = sg_dma_len(map->sgl);
sgl              1483 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		func = map->page->desc->func->sgl;
sgl                69 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 	nvkm_vmm_pte_func sgl;
sgl               321 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 		     sg_dma_address(MAP->sgl), sg_dma_len(MAP->sgl),           \
sgl               322 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 		     (MAP->sgl = sg_next(MAP->sgl)))
sgl               102 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	.sgl = gf100_vmm_pgt_sgl,
sgl                41 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	.sgl = gf100_vmm_pgt_sgl,
sgl               169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	.sgl = gp100_vmm_pgt_sgl,
sgl                70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	.sgl = nv04_vmm_pgt_sgl,
sgl                71 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c 	.sgl = nv41_vmm_pgt_sgl,
sgl               175 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c 	.sgl = nv44_vmm_pgt_sgl,
sgl               102 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	.sgl = nv50_vmm_pgt_sgl,
sgl              1252 drivers/gpu/drm/omapdrm/omap_gem.c 		omap_obj->dma_addr = sg_dma_address(sgt->sgl);
sgl              1270 drivers/gpu/drm/omapdrm/omap_gem.c 		for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) {
sgl                42 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	sg_init_table(sg->sgl, 1);
sgl                43 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	sg_dma_len(sg->sgl) = obj->size;
sgl                44 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	sg_set_page(sg->sgl, pfn_to_page(PFN_DOWN(dma_addr)), obj->size, 0);
sgl                45 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	sg_dma_address(sg->sgl) = dma_addr;
sgl                43 drivers/gpu/drm/panfrost/panfrost_gem.c 			if (bo->sgts[i].sgl) {
sgl                44 drivers/gpu/drm/panfrost/panfrost_gem.c 				dma_unmap_sg(pfdev->dev, bo->sgts[i].sgl,
sgl               252 drivers/gpu/drm/panfrost/panfrost_mmu.c 	struct scatterlist *sgl;
sgl               256 drivers/gpu/drm/panfrost/panfrost_mmu.c 	for_each_sg(sgt->sgl, sgl, sgt->nents, count) {
sgl               257 drivers/gpu/drm/panfrost/panfrost_mmu.c 		unsigned long paddr = sg_dma_address(sgl);
sgl               258 drivers/gpu/drm/panfrost/panfrost_mmu.c 		size_t len = sg_dma_len(sgl);
sgl               520 drivers/gpu/drm/panfrost/panfrost_mmu.c 	if (!dma_map_sg(pfdev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL)) {
sgl               530 drivers/gpu/drm/radeon/radeon_ttm.c 	nents = dma_map_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
sgl               558 drivers/gpu/drm/radeon/radeon_ttm.c 	if (!ttm->sg->sgl)
sgl               562 drivers/gpu/drm/radeon/radeon_ttm.c 	dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
sgl               564 drivers/gpu/drm/radeon/radeon_ttm.c 	for_each_sg_page(ttm->sg->sgl, &sg_iter, ttm->sg->nents, 0) {
sgl               173 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		cfg.mem[i] = sg_dma_address(state->sg_tables[i].sgl)
sgl               235 drivers/gpu/drm/rcar-du/rcar_du_writeback.c 		cfg->mem[i] = sg_dma_address(rjob->sg_tables[i].sgl)
sgl                38 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl,
sgl               100 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	for_each_sg(rk_obj->sgt->sgl, s, rk_obj->sgt->nents, i)
sgl               103 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents,
sgl               352 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 			dma_unmap_sg(drm->dev, rk_obj->sgt->sgl,
sgl               466 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	dma_addr_t expected = sg_dma_address(sgt->sgl);
sgl               470 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	for_each_sg(sgt->sgl, s, count, i) {
sgl               495 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	int count = dma_map_sg(drm->dev, sg->sgl, sg->nents,
sgl               502 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 		dma_unmap_sg(drm->dev, sg->sgl, sg->nents,
sgl               507 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	rk_obj->dma_addr = sg_dma_address(sg->sgl);
sgl               138 drivers/gpu/drm/tegra/gem.c 	bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl,
sgl               206 drivers/gpu/drm/tegra/gem.c 		dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgl               232 drivers/gpu/drm/tegra/gem.c 	err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgl               368 drivers/gpu/drm/tegra/gem.c 		bo->paddr = sg_dma_address(bo->sgt->sgl);
sgl               517 drivers/gpu/drm/tegra/gem.c 		for_each_sg(sgt->sgl, sg, bo->num_pages, i)
sgl               520 drivers/gpu/drm/tegra/gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0)
sgl               526 drivers/gpu/drm/tegra/gem.c 		sg_dma_address(sgt->sgl) = bo->paddr;
sgl               527 drivers/gpu/drm/tegra/gem.c 		sg_dma_len(sgt->sgl) = gem->size;
sgl               546 drivers/gpu/drm/tegra/gem.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
sgl               565 drivers/gpu/drm/tegra/gem.c 		dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgl               579 drivers/gpu/drm/tegra/gem.c 		dma_sync_sg_for_device(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgl                54 drivers/gpu/drm/udl/udl_dmabuf.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents,
sgl               107 drivers/gpu/drm/udl/udl_dmabuf.c 	rd = obj->sg->sgl;
sgl               108 drivers/gpu/drm/udl/udl_dmabuf.c 	wr = sgt->sgl;
sgl               116 drivers/gpu/drm/udl/udl_dmabuf.c 		nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir);
sgl                94 drivers/gpu/drm/v3d/v3d_mmu.c 	struct scatterlist *sgl;
sgl                96 drivers/gpu/drm/v3d/v3d_mmu.c 	for_each_sg(shmem_obj->sgt->sgl, sgl, shmem_obj->sgt->nents, count) {
sgl                97 drivers/gpu/drm/v3d/v3d_mmu.c 		u32 page_address = sg_dma_address(sgl) >> V3D_MMU_PAGE_SHIFT;
sgl               101 drivers/gpu/drm/v3d/v3d_mmu.c 		BUG_ON(page_address + (sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT) >=
sgl               104 drivers/gpu/drm/v3d/v3d_mmu.c 		for (i = 0; i < sg_dma_len(sgl) >> V3D_MMU_PAGE_SHIFT; i++)
sgl               496 drivers/gpu/drm/virtio/virtgpu_vq.c 				       bo->pages->sgl, bo->pages->nents,
sgl               903 drivers/gpu/drm/virtio/virtgpu_vq.c 				       bo->pages->sgl, bo->pages->nents,
sgl               983 drivers/gpu/drm/virtio/virtgpu_vq.c 					 obj->pages->sgl, obj->pages->nents,
sgl               998 drivers/gpu/drm/virtio/virtgpu_vq.c 	for_each_sg(obj->pages->sgl, sg, nents, si) {
sgl              1026 drivers/gpu/drm/virtio/virtgpu_vq.c 			     obj->pages->sgl, obj->mapped,
sgl               345 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		__sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl,
sgl               365 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents,
sgl               388 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	ret = dma_map_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.orig_nents,
sgl               145 drivers/gpu/host1x/job.c 			for_each_sg(sgt->sgl, sg, sgt->nents, j)
sgl               159 drivers/gpu/host1x/job.c 					sgt->sgl, sgt->nents, IOMMU_READ);
sgl               199 drivers/hsi/clients/cmt_speech.c 	u32 *data = sg_virt(msg->sgt.sgl);
sgl               205 drivers/hsi/clients/cmt_speech.c 	u32 *data = sg_virt(msg->sgt.sgl);
sgl               259 drivers/hsi/clients/cmt_speech.c 		kfree(sg_virt(msg->sgt.sgl));
sgl               281 drivers/hsi/clients/cmt_speech.c 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
sgl               624 drivers/hsi/clients/cmt_speech.c 	sg_init_one(msg->sgt.sgl, address, hi->buf_size);
sgl               667 drivers/hsi/clients/cmt_speech.c 	sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0);
sgl               719 drivers/hsi/clients/cmt_speech.c 	sg_init_one(txmsg->sgt.sgl, address, hi->buf_size);
sgl               147 drivers/hsi/clients/hsi_char.c 	kfree(sg_virt(msg->sgt.sgl));
sgl               186 drivers/hsi/clients/hsi_char.c 	sg_init_one(msg->sgt.sgl, buf, alloc_size);
sgl               217 drivers/hsi/clients/hsi_char.c 	return msg->sgt.sgl->length;
sgl               222 drivers/hsi/clients/hsi_char.c 	msg->sgt.sgl->length = len;
sgl               455 drivers/hsi/clients/hsi_char.c 			sg_virt(msg->sgt.sgl), hsc_msg_len_get(msg));
sgl               491 drivers/hsi/clients/hsi_char.c 	if (copy_from_user(sg_virt(msg->sgt.sgl), (void __user *)buf, len)) {
sgl               157 drivers/hsi/clients/ssi_protocol.c 	data = sg_virt(msg->sgt.sgl);
sgl               165 drivers/hsi/clients/ssi_protocol.c 	data = sg_virt(msg->sgt.sgl);
sgl               178 drivers/hsi/clients/ssi_protocol.c 	sg = msg->sgt.sgl;
sgl               249 drivers/hsi/clients/ssi_protocol.c 		kfree(sg_virt(msg->sgt.sgl));
sgl               269 drivers/hsi/clients/ssi_protocol.c 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
sgl               192 drivers/hsi/controllers/omap_ssi_core.c 	dma_unmap_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, dir);
sgl               215 drivers/hsi/controllers/omap_ssi_core.c 	msg->actual_len = sg_dma_len(msg->sgt.sgl);
sgl               200 drivers/hsi/controllers/omap_ssi_port.c 			omap_ssi->gdd_trn[lch].sg = msg->sgt.sgl;
sgl               231 drivers/hsi/controllers/omap_ssi_port.c 		err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents,
sgl               246 drivers/hsi/controllers/omap_ssi_port.c 		d_addr = sg_dma_address(msg->sgt.sgl);
sgl               248 drivers/hsi/controllers/omap_ssi_port.c 		err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents,
sgl               261 drivers/hsi/controllers/omap_ssi_port.c 		s_addr = sg_dma_address(msg->sgt.sgl);
sgl               272 drivers/hsi/controllers/omap_ssi_port.c 	writew_relaxed(SSI_BYTES_TO_FRAMES(msg->sgt.sgl->length),
sgl               330 drivers/hsi/controllers/omap_ssi_port.c 	if ((msg->sgt.nents) && (msg->sgt.sgl->length > sizeof(u32)))
sgl               445 drivers/hsi/controllers/omap_ssi_port.c 			msg->channel, msg, msg->sgt.sgl->length,
sgl               905 drivers/hsi/controllers/omap_ssi_port.c 	if ((!msg->sgt.nents) || (!msg->sgt.sgl->length)) {
sgl               914 drivers/hsi/controllers/omap_ssi_port.c 		buf = sg_virt(msg->sgt.sgl) + msg->actual_len;
sgl               924 drivers/hsi/controllers/omap_ssi_port.c 		if (msg->actual_len >= msg->sgt.sgl->length)
sgl                70 drivers/hwtracing/intel_th/msu-sink.c 	for_each_sg((*sgt)->sgl, sg_ptr, nents, i) {
sgl                87 drivers/hwtracing/intel_th/msu-sink.c 	for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) {
sgl               276 drivers/hwtracing/intel_th/msu.c 	return win->sgt->sgl;
sgl               326 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgl               421 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgl               660 drivers/hwtracing/intel_th/msu.c 		for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgl               928 drivers/hwtracing/intel_th/msu.c 	sg_set_buf(msc->single_sgt.sgl, page_address(page), size);
sgl               930 drivers/hwtracing/intel_th/msu.c 	ret = dma_map_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl, 1,
sgl               937 drivers/hwtracing/intel_th/msu.c 	msc->base_addr = sg_dma_address(msc->single_sgt.sgl);
sgl               959 drivers/hwtracing/intel_th/msu.c 	dma_unmap_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl,
sgl              1000 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) {
sgl              1013 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, i, ret)
sgl              1028 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) {
sgl              1040 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) {
sgl              1123 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, i) {
sgl              1190 drivers/hwtracing/intel_th/msu.c 		for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgl              1385 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgl                52 drivers/infiniband/core/umem.c 		ib_dma_unmap_sg(dev, umem->sg_head.sgl, umem->sg_nents,
sgl                55 drivers/infiniband/core/umem.c 	for_each_sg_page(umem->sg_head.sgl, &sg_iter, umem->sg_nents, 0) {
sgl               164 drivers/infiniband/core/umem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) {
sgl               274 drivers/infiniband/core/umem.c 	sg = umem->sg_head.sgl;
sgl               301 drivers/infiniband/core/umem.c 				  umem->sg_head.sgl,
sgl               353 drivers/infiniband/core/umem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i)
sgl               382 drivers/infiniband/core/umem.c 	ret = sg_pcopy_to_buffer(umem->sg_head.sgl, umem->sg_nents, dst, length,
sgl              2546 drivers/infiniband/core/verbs.c int ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents,
sgl              2556 drivers/infiniband/core/verbs.c 	if (unlikely(sg_nents <= 0 || sg_offset > sg_dma_len(&sgl[0])))
sgl              2559 drivers/infiniband/core/verbs.c 	mr->iova = sg_dma_address(&sgl[0]) + sg_offset;
sgl              2562 drivers/infiniband/core/verbs.c 	for_each_sg(sgl, sg, sg_nents, i) {
sgl               863 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	qplib_qp->sq.sg_info.sglist = umem->sg_head.sgl;
sgl               876 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		qplib_qp->rq.sg_info.sglist = umem->sg_head.sgl;
sgl              1330 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	qplib_srq->sg_info.sglist = umem->sg_head.sgl;
sgl              2573 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		cq->qplib_cq.sg_info.sglist = cq->umem->sg_head.sgl;
sgl              3494 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap, page_size)
sgl               176 drivers/infiniband/hw/cxgb3/cxio_wr.h 	struct t3_sge sgl[T3_MAX_SGE];	/* 4+ */
sgl               238 drivers/infiniband/hw/cxgb3/cxio_wr.h 	struct t3_sge sgl[T3_MAX_SGE];	/* 5+ */
sgl               274 drivers/infiniband/hw/cxgb3/cxio_wr.h 	struct t3_sge sgl[T3_MAX_SGE];	/* 3+ */
sgl               477 drivers/infiniband/hw/cxgb3/iwch_provider.c 	for_each_sg_dma_page(mhp->umem->sg_head.sgl, &sg_iter, mhp->umem->nmap, 0) {
sgl                77 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
sgl                78 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
sgl                79 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
sgl               103 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->write.sgl[0].stag = wr->ex.imm_data;
sgl               104 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->write.sgl[0].len = cpu_to_be32(0);
sgl               114 drivers/infiniband/hw/cxgb3/iwch_qp.c 			wqe->write.sgl[i].stag =
sgl               116 drivers/infiniband/hw/cxgb3/iwch_qp.c 			wqe->write.sgl[i].len =
sgl               118 drivers/infiniband/hw/cxgb3/iwch_qp.c 			wqe->write.sgl[i].to =
sgl               265 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
sgl               266 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
sgl               269 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = cpu_to_be64(((u32)wr->sg_list[i].addr) &
sgl               276 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].stag = 0;
sgl               277 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = 0;
sgl               278 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = 0;
sgl               331 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].stag = 0;
sgl               332 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
sgl               333 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
sgl               339 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].stag = 0;
sgl               340 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = 0;
sgl               341 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = 0;
sgl               711 drivers/infiniband/hw/cxgb3/iwch_qp.c 	term = (struct terminate_message *)wqe->send.sgl;
sgl                67 drivers/infiniband/hw/cxgb4/mem.c 	struct ulptx_sgl *sgl;
sgl                75 drivers/infiniband/hw/cxgb4/mem.c 	wr_len = roundup(sizeof(*req) + sizeof(*sgl), 16);
sgl                97 drivers/infiniband/hw/cxgb4/mem.c 	sgl = (struct ulptx_sgl *)(req + 1);
sgl                98 drivers/infiniband/hw/cxgb4/mem.c 	sgl->cmd_nsge = cpu_to_be32(ULPTX_CMD_V(ULP_TX_SC_DSGL) |
sgl               100 drivers/infiniband/hw/cxgb4/mem.c 	sgl->len0 = cpu_to_be32(len);
sgl               101 drivers/infiniband/hw/cxgb4/mem.c 	sgl->addr0 = cpu_to_be64(data);
sgl               565 drivers/infiniband/hw/cxgb4/mem.c 	for_each_sg_dma_page(mhp->umem->sg_head.sgl, &sg_iter, mhp->umem->nmap, 0) {
sgl               105 drivers/infiniband/hw/efa/efa_verbs.c 			struct scatterlist *sgl;
sgl              1058 drivers/infiniband/hw/efa/efa_verbs.c 	rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap,
sgl              1097 drivers/infiniband/hw/efa/efa_verbs.c 	struct scatterlist *pages_sgl = pbl->phys.indirect.sgl;
sgl              1236 drivers/infiniband/hw/efa/efa_verbs.c 	struct scatterlist *sgl;
sgl              1240 drivers/infiniband/hw/efa/efa_verbs.c 	sgl = efa_vmalloc_buf_to_sg(pbl->pbl_buf, size_in_pages);
sgl              1241 drivers/infiniband/hw/efa/efa_verbs.c 	if (!sgl)
sgl              1244 drivers/infiniband/hw/efa/efa_verbs.c 	sg_dma_cnt = dma_map_sg(&dev->pdev->dev, sgl, size_in_pages, DMA_TO_DEVICE);
sgl              1251 drivers/infiniband/hw/efa/efa_verbs.c 	pbl->phys.indirect.sgl = sgl;
sgl              1268 drivers/infiniband/hw/efa/efa_verbs.c 	dma_unmap_sg(&dev->pdev->dev, sgl, size_in_pages, DMA_TO_DEVICE);
sgl              1270 drivers/infiniband/hw/efa/efa_verbs.c 	kfree(sgl);
sgl              1277 drivers/infiniband/hw/efa/efa_verbs.c 	dma_unmap_sg(&dev->pdev->dev, pbl->phys.indirect.sgl,
sgl              1279 drivers/infiniband/hw/efa/efa_verbs.c 	kfree(pbl->phys.indirect.sgl);
sgl               280 drivers/infiniband/hw/hns/hns_roce_alloc.c 	rdma_for_each_block(umem->sg_head.sgl, &biter, umem->nmap,
sgl                45 drivers/infiniband/hw/hns/hns_roce_db.c 	db->dma = sg_dma_address(page->umem->sg_head.sgl) + offset;
sgl                46 drivers/infiniband/hw/hns/hns_roce_db.c 	db->virt_addr = sg_virt(page->umem->sg_head.sgl) + offset;
sgl              1890 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) {
sgl              2241 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) {
sgl              1060 drivers/infiniband/hw/hns/hns_roce_mr.c 	for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl              1103 drivers/infiniband/hw/hns/hns_roce_mr.c 	for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl              1324 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		iwpbl->qp_mr.sq_page = sg_page(region->sg_head.sgl);
sgl              1326 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	rdma_for_each_block(region->sg_head.sgl, &biter, region->nmap,
sgl              2098 drivers/infiniband/hw/i40iw/i40iw_verbs.c static void i40iw_copy_sg_list(struct i40iw_sge *sg_list, struct ib_sge *sgl, int num_sges)
sgl              2103 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		sg_list[i].tag_off = sgl[i].addr;
sgl              2104 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		sg_list[i].len = sgl[i].length;
sgl              2105 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		sg_list[i].stag = sgl[i].lkey;
sgl                77 drivers/infiniband/hw/mlx4/doorbell.c 	db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK);
sgl               203 drivers/infiniband/hw/mlx4/mr.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) {
sgl               274 drivers/infiniband/hw/mlx4/mr.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, i) {
sgl                77 drivers/infiniband/hw/mlx5/doorbell.c 	db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK);
sgl                65 drivers/infiniband/hw/mlx5/mem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) {
sgl               155 drivers/infiniband/hw/mlx5/mem.c 	for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) {
sgl              2063 drivers/infiniband/hw/mlx5/mr.c 		   struct scatterlist *sgl,
sgl              2070 drivers/infiniband/hw/mlx5/mr.c 	struct scatterlist *sg = sgl;
sgl              2079 drivers/infiniband/hw/mlx5/mr.c 	for_each_sg(sgl, sg, sg_nents, i) {
sgl               909 drivers/infiniband/hw/mthca/mthca_provider.c 	for_each_sg_dma_page(mr->umem->sg_head.sgl, &sg_iter, mr->umem->nmap, 0) {
sgl               837 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	for_each_sg_dma_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl               633 drivers/infiniband/hw/qedr/verbs.c 	for_each_sg_dma_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl              1308 drivers/infiniband/hw/qedr/verbs.c 	sg = srq->prod_umem->sg_head.sgl;
sgl               192 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c 	for_each_sg_dma_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl               415 drivers/infiniband/sw/rdmavt/mr.c 	for_each_sg_page (umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl               200 drivers/infiniband/sw/rxe/rxe_mr.c 		for_each_sg_page(umem->sg_head.sgl, &sg_iter, umem->nmap, 0) {
sgl               585 drivers/input/touchscreen/sur40.c 		sgt->sgl, sgt->nents, sur40->pix_fmt.sizeimage, 0);
sgl               609 drivers/iommu/dma-iommu.c 		for_each_sg(sgt.sgl, sg, sgt.orig_nents, i)
sgl               613 drivers/iommu/dma-iommu.c 	if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, ioprot)
sgl               677 drivers/iommu/dma-iommu.c 		struct scatterlist *sgl, int nelems,
sgl               686 drivers/iommu/dma-iommu.c 	for_each_sg(sgl, sg, nelems, i)
sgl               691 drivers/iommu/dma-iommu.c 		struct scatterlist *sgl, int nelems,
sgl               700 drivers/iommu/dma-iommu.c 	for_each_sg(sgl, sg, nelems, i)
sgl              1082 drivers/iommu/dma-iommu.c 		sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgl                52 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_addr_t expected = sg_dma_address(sgt->sgl);
sgl                56 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	for_each_sg(sgt->sgl, s, sgt->nents, i) {
sgl               102 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               115 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir);
sgl               243 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	rd = buf->sgt_base->sgl;
sgl               244 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	wr = sgt->sgl;
sgl               276 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               302 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               311 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt->nents = dma_map_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               434 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               532 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               548 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgl               556 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               610 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgl               145 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               183 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               205 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               218 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir);
sgl               261 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgl               289 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir,
sgl               387 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	rd = buf->dma_sgt->sgl;
sgl               388 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	wr = sgt->sgl;
sgl               414 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               440 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               446 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               233 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl               263 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               289 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl               295 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgl                80 drivers/media/pci/cobalt/cobalt-v4l2.c 	ret = descriptor_list_create(cobalt, sg_desc->sgl,
sgl              1586 drivers/media/pci/cx23885/cx23885-core.c 				sgt->sgl,
sgl               145 drivers/media/pci/cx23885/cx23885-vbi.c 			 sgt->sgl,
sgl               359 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, 0, UNSET,
sgl               364 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, UNSET, 0,
sgl               392 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, line0_offset,
sgl               399 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl,
sgl               406 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl,
sgl               180 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, 0, UNSET,
sgl               185 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, UNSET, 0,
sgl               194 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, line0_offset,
sgl               200 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl,
sgl               206 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl,
sgl               225 drivers/media/pci/cx88/cx88-mpeg.c 	rc = cx88_risc_databuffer(dev->pci, risc, sgt->sgl,
sgl               147 drivers/media/pci/cx88/cx88-vbi.c 	cx88_risc_buffer(dev->pci, &buf->risc, sgt->sgl,
sgl               451 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, 0, UNSET,
sgl               456 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, UNSET, 0,
sgl               461 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl,
sgl               468 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl,
sgl               476 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, 0, buf->bpl,
sgl               872 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	if (sg->nents && sg->sgl)
sgl               873 drivers/media/pci/intel/ipu3/ipu3-cio2.c 		b->offset = sg->sgl->offset;
sgl               876 drivers/media/pci/intel/ipu3/ipu3-cio2.c 	for_each_sg_dma_page (sg->sgl, &sg_iter, sg->nents, 0) {
sgl               215 drivers/media/pci/saa7134/saa7134-core.c 	base += dma->sgl[0].offset;
sgl               101 drivers/media/pci/saa7134/saa7134-ts.c 	return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
sgl               114 drivers/media/pci/saa7134/saa7134-vbi.c 	if (dma->sgl->offset) {
sgl               124 drivers/media/pci/saa7134/saa7134-vbi.c 	return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
sgl               922 drivers/media/pci/saa7134/saa7134-video.c 	if (dma->sgl->offset) {
sgl               933 drivers/media/pci/saa7134/saa7134-video.c 	return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
sgl               322 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	for_each_sg(vbuf->sgl, sg, vbuf->nents, i) {
sgl               743 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 			sg_copy_from_buffer(sgt->sgl, sgt->nents,
sgl               747 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 		sg_copy_from_buffer(sgt->sgl, sgt->nents,
sgl               455 drivers/media/pci/tw68/tw68-video.c 		tw68_risc_buffer(dev->pci, buf, dma->sgl,
sgl               459 drivers/media/pci/tw68/tw68-video.c 		tw68_risc_buffer(dev->pci, buf, dma->sgl,
sgl               463 drivers/media/pci/tw68/tw68-video.c 		tw68_risc_buffer(dev->pci, buf, dma->sgl,
sgl               468 drivers/media/pci/tw68/tw68-video.c 		tw68_risc_buffer(dev->pci, buf, dma->sgl,
sgl               474 drivers/media/pci/tw68/tw68-video.c 		tw68_risc_buffer(dev->pci, buf, dma->sgl,
sgl               197 drivers/media/pci/tw686x/tw686x-video.c 	for_each_sg(vbuf->sgl, sg, vbuf->nents, i) {
sgl               236 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width;
sgl               237 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width;
sgl               243 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width;
sgl               244 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width;
sgl               250 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               251 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               257 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               258 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               264 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               265 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               271 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               272 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               278 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width;
sgl               279 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width;
sgl               285 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               286 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               292 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width / 2;
sgl               293 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width / 2;
sgl               299 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width * 2;
sgl               300 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width * 2;
sgl               306 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width * 2;
sgl               307 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width * 2;
sgl               314 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].size = s_width * 2;
sgl               315 drivers/media/platform/m2m-deinterlace.c 		ctx->xt->sgl[0].icg = s_width * 2;
sgl              1260 drivers/media/platform/marvell-ccic/mcam-core.c 	for_each_sg(sg_table->sgl, sg, sg_table->nents, i) {
sgl               261 drivers/media/platform/omap/omap_vout_vrfb.c 	xt->sgl[0].size = vout->pix.width * vout->bpp;
sgl               262 drivers/media/platform/omap/omap_vout_vrfb.c 	xt->sgl[0].icg = dst_icg;
sgl               449 drivers/media/platform/omap3isp/ispccdc.c 		dma_sync_sg_for_cpu(isp->dev, req->table.sgt.sgl,
sgl               458 drivers/media/platform/omap3isp/ispccdc.c 		dma_sync_sg_for_device(isp->dev, req->table.sgt.sgl,
sgl               164 drivers/media/platform/omap3isp/ispstat.c 	dma_sync_sg_for_device(stat->isp->dev, buf->sgt.sgl,
sgl               174 drivers/media/platform/omap3isp/ispstat.c 	dma_sync_sg_for_cpu(stat->isp->dev, buf->sgt.sgl,
sgl              1394 drivers/media/platform/pxa_camera.c 	ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels,
sgl               364 drivers/media/platform/qcom/camss/camss-video.c 		buffer->addr[i] = sg_dma_address(sgt->sgl);
sgl              1042 drivers/media/platform/qcom/venus/helpers.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgl               118 drivers/media/platform/rockchip/rga/rga-buf.c 	struct scatterlist *sgl;
sgl               131 drivers/media/platform/rockchip/rga/rga-buf.c 	for_each_sg(sgt->sgl, sgl, sgt->nents, i) {
sgl               132 drivers/media/platform/rockchip/rga/rga-buf.c 		len = sg_dma_len(sgl) >> PAGE_SHIFT;
sgl               133 drivers/media/platform/rockchip/rga/rga-buf.c 		address = sg_phys(sgl);
sgl               369 drivers/media/platform/via-camera.c 	viafb_dma_copy_out_sg(cam->cb_offsets[bufn], sgt->sgl, sgt->nents);
sgl               915 drivers/media/platform/vsp1/vsp1_drm.c 	return dma_map_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents,
sgl               924 drivers/media/platform/vsp1/vsp1_drm.c 	dma_unmap_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents,
sgl               367 drivers/media/platform/xilinx/xilinx-dma.c 	dma->sgl[0].size = dma->format.width * dma->fmtinfo->bpp;
sgl               368 drivers/media/platform/xilinx/xilinx-dma.c 	dma->sgl[0].icg = dma->format.bytesperline - dma->sgl[0].size;
sgl                95 drivers/media/platform/xilinx/xilinx-dma.h 	struct data_chunk sgl[1];
sgl               130 drivers/message/fusion/mptctl.c static void kfree_sgl(MptSge_t *sgl, dma_addr_t sgl_dma,
sgl               785 drivers/message/fusion/mptctl.c 	MptSge_t		*sgl, *sgIn;
sgl               861 drivers/message/fusion/mptctl.c 	if ((sgl = kbuf_alloc_2_sgl(fwlen, sgdir, sge_offset,
sgl               885 drivers/message/fusion/mptctl.c 	    iocp->name, sgl, numfrags));
sgl               892 drivers/message/fusion/mptctl.c 	sgIn = sgl;
sgl               958 drivers/message/fusion/mptctl.c 	if (sgl)
sgl               959 drivers/message/fusion/mptctl.c 		kfree_sgl(sgl, sgl_dma, buflist, iocp);
sgl               988 drivers/message/fusion/mptctl.c         kfree_sgl(sgl, sgl_dma, buflist, iocp);
sgl              1014 drivers/message/fusion/mptctl.c 	MptSge_t	*sgl;
sgl              1062 drivers/message/fusion/mptctl.c 	sgl = sglbuf;
sgl              1084 drivers/message/fusion/mptctl.c 			sgl->FlagsLength = (0x10000000|sgdir|this_alloc);
sgl              1087 drivers/message/fusion/mptctl.c 			sgl->Address = dma_addr;
sgl              1091 drivers/message/fusion/mptctl.c 			sgl++;
sgl              1118 drivers/message/fusion/mptctl.c 	sgl[-1].FlagsLength |= 0xC1000000;
sgl              1158 drivers/message/fusion/mptctl.c kfree_sgl(MptSge_t *sgl, dma_addr_t sgl_dma, struct buflist *buflist, MPT_ADAPTER *ioc)
sgl              1160 drivers/message/fusion/mptctl.c 	MptSge_t	*sg = sgl;
sgl              1207 drivers/message/fusion/mptctl.c 	pci_free_consistent(ioc->pcidev, MAX_SGL_BYTES, sgl, sgl_dma);
sgl               495 drivers/misc/fastrpc.c 	if (!dma_map_sg(attachment->dev, table->sgl, table->nents, dir))
sgl               505 drivers/misc/fastrpc.c 	dma_unmap_sg(attach->dev, table->sgl, table->nents, dir);
sgl               629 drivers/misc/fastrpc.c 	map->phys = sg_dma_address(map->table->sgl);
sgl               632 drivers/misc/fastrpc.c 	map->va = sg_virt(map->table->sgl);
sgl               348 drivers/misc/genwqe/card_base.h 	struct sg_entry *sgl;
sgl               368 drivers/misc/genwqe/card_base.h int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl,
sgl               371 drivers/misc/genwqe/card_base.h int genwqe_setup_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl,
sgl               374 drivers/misc/genwqe/card_base.h int genwqe_free_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl);
sgl               839 drivers/misc/genwqe/card_dev.c 		if (req->sgls[i].sgl != NULL)
sgl               288 drivers/misc/genwqe/card_utils.c int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl,
sgl               294 drivers/misc/genwqe/card_utils.c 	sgl->fpage_offs = offset_in_page((unsigned long)user_addr);
sgl               295 drivers/misc/genwqe/card_utils.c 	sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size);
sgl               296 drivers/misc/genwqe/card_utils.c 	sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE);
sgl               297 drivers/misc/genwqe/card_utils.c 	sgl->lpage_size = (user_size - sgl->fpage_size) % PAGE_SIZE;
sgl               300 drivers/misc/genwqe/card_utils.c 		__func__, user_addr, user_size, sgl->nr_pages,
sgl               301 drivers/misc/genwqe/card_utils.c 		sgl->fpage_offs, sgl->fpage_size, sgl->lpage_size);
sgl               303 drivers/misc/genwqe/card_utils.c 	sgl->user_addr = user_addr;
sgl               304 drivers/misc/genwqe/card_utils.c 	sgl->user_size = user_size;
sgl               305 drivers/misc/genwqe/card_utils.c 	sgl->write = write;
sgl               306 drivers/misc/genwqe/card_utils.c 	sgl->sgl_size = genwqe_sgl_size(sgl->nr_pages);
sgl               308 drivers/misc/genwqe/card_utils.c 	if (get_order(sgl->sgl_size) > MAX_ORDER) {
sgl               314 drivers/misc/genwqe/card_utils.c 	sgl->sgl = __genwqe_alloc_consistent(cd, sgl->sgl_size,
sgl               315 drivers/misc/genwqe/card_utils.c 					     &sgl->sgl_dma_addr);
sgl               316 drivers/misc/genwqe/card_utils.c 	if (sgl->sgl == NULL) {
sgl               323 drivers/misc/genwqe/card_utils.c 	if ((sgl->fpage_size != 0) && (sgl->fpage_size != PAGE_SIZE)) {
sgl               324 drivers/misc/genwqe/card_utils.c 		sgl->fpage = __genwqe_alloc_consistent(cd, PAGE_SIZE,
sgl               325 drivers/misc/genwqe/card_utils.c 						       &sgl->fpage_dma_addr);
sgl               326 drivers/misc/genwqe/card_utils.c 		if (sgl->fpage == NULL)
sgl               330 drivers/misc/genwqe/card_utils.c 		if (copy_from_user(sgl->fpage + sgl->fpage_offs,
sgl               331 drivers/misc/genwqe/card_utils.c 				   user_addr, sgl->fpage_size)) {
sgl               336 drivers/misc/genwqe/card_utils.c 	if (sgl->lpage_size != 0) {
sgl               337 drivers/misc/genwqe/card_utils.c 		sgl->lpage = __genwqe_alloc_consistent(cd, PAGE_SIZE,
sgl               338 drivers/misc/genwqe/card_utils.c 						       &sgl->lpage_dma_addr);
sgl               339 drivers/misc/genwqe/card_utils.c 		if (sgl->lpage == NULL)
sgl               343 drivers/misc/genwqe/card_utils.c 		if (copy_from_user(sgl->lpage, user_addr + user_size -
sgl               344 drivers/misc/genwqe/card_utils.c 				   sgl->lpage_size, sgl->lpage_size)) {
sgl               352 drivers/misc/genwqe/card_utils.c 	__genwqe_free_consistent(cd, PAGE_SIZE, sgl->lpage,
sgl               353 drivers/misc/genwqe/card_utils.c 				 sgl->lpage_dma_addr);
sgl               354 drivers/misc/genwqe/card_utils.c 	sgl->lpage = NULL;
sgl               355 drivers/misc/genwqe/card_utils.c 	sgl->lpage_dma_addr = 0;
sgl               357 drivers/misc/genwqe/card_utils.c 	__genwqe_free_consistent(cd, PAGE_SIZE, sgl->fpage,
sgl               358 drivers/misc/genwqe/card_utils.c 				 sgl->fpage_dma_addr);
sgl               359 drivers/misc/genwqe/card_utils.c 	sgl->fpage = NULL;
sgl               360 drivers/misc/genwqe/card_utils.c 	sgl->fpage_dma_addr = 0;
sgl               362 drivers/misc/genwqe/card_utils.c 	__genwqe_free_consistent(cd, sgl->sgl_size, sgl->sgl,
sgl               363 drivers/misc/genwqe/card_utils.c 				 sgl->sgl_dma_addr);
sgl               364 drivers/misc/genwqe/card_utils.c 	sgl->sgl = NULL;
sgl               365 drivers/misc/genwqe/card_utils.c 	sgl->sgl_dma_addr = 0;
sgl               366 drivers/misc/genwqe/card_utils.c 	sgl->sgl_size = 0;
sgl               371 drivers/misc/genwqe/card_utils.c int genwqe_setup_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl,
sgl               378 drivers/misc/genwqe/card_utils.c 	size_t size = sgl->user_size;
sgl               381 drivers/misc/genwqe/card_utils.c 	map_offs = sgl->fpage_offs; /* offset in first page */
sgl               383 drivers/misc/genwqe/card_utils.c 	s = &sgl->sgl[0];	/* first set of 8 entries */
sgl               385 drivers/misc/genwqe/card_utils.c 	while (p < sgl->nr_pages) {
sgl               391 drivers/misc/genwqe/card_utils.c 		s[j].target_addr = cpu_to_be64(sgl->sgl_dma_addr + dma_offs);
sgl               400 drivers/misc/genwqe/card_utils.c 			if ((p == 0) && (sgl->fpage != NULL)) {
sgl               401 drivers/misc/genwqe/card_utils.c 				daddr = sgl->fpage_dma_addr + map_offs;
sgl               403 drivers/misc/genwqe/card_utils.c 			} else if ((p == sgl->nr_pages - 1) &&
sgl               404 drivers/misc/genwqe/card_utils.c 				   (sgl->lpage != NULL)) {
sgl               405 drivers/misc/genwqe/card_utils.c 				daddr = sgl->lpage_dma_addr;
sgl               424 drivers/misc/genwqe/card_utils.c 				if (p == sgl->nr_pages)
sgl               440 drivers/misc/genwqe/card_utils.c 			if (p == sgl->nr_pages)
sgl               468 drivers/misc/genwqe/card_utils.c int genwqe_free_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl)
sgl               475 drivers/misc/genwqe/card_utils.c 	if (sgl->fpage) {
sgl               476 drivers/misc/genwqe/card_utils.c 		if (sgl->write) {
sgl               477 drivers/misc/genwqe/card_utils.c 			res = copy_to_user(sgl->user_addr,
sgl               478 drivers/misc/genwqe/card_utils.c 				sgl->fpage + sgl->fpage_offs, sgl->fpage_size);
sgl               486 drivers/misc/genwqe/card_utils.c 		__genwqe_free_consistent(cd, PAGE_SIZE, sgl->fpage,
sgl               487 drivers/misc/genwqe/card_utils.c 					 sgl->fpage_dma_addr);
sgl               488 drivers/misc/genwqe/card_utils.c 		sgl->fpage = NULL;
sgl               489 drivers/misc/genwqe/card_utils.c 		sgl->fpage_dma_addr = 0;
sgl               491 drivers/misc/genwqe/card_utils.c 	if (sgl->lpage) {
sgl               492 drivers/misc/genwqe/card_utils.c 		if (sgl->write) {
sgl               493 drivers/misc/genwqe/card_utils.c 			offset = sgl->user_size - sgl->lpage_size;
sgl               494 drivers/misc/genwqe/card_utils.c 			res = copy_to_user(sgl->user_addr + offset, sgl->lpage,
sgl               495 drivers/misc/genwqe/card_utils.c 					   sgl->lpage_size);
sgl               503 drivers/misc/genwqe/card_utils.c 		__genwqe_free_consistent(cd, PAGE_SIZE, sgl->lpage,
sgl               504 drivers/misc/genwqe/card_utils.c 					 sgl->lpage_dma_addr);
sgl               505 drivers/misc/genwqe/card_utils.c 		sgl->lpage = NULL;
sgl               506 drivers/misc/genwqe/card_utils.c 		sgl->lpage_dma_addr = 0;
sgl               508 drivers/misc/genwqe/card_utils.c 	__genwqe_free_consistent(cd, sgl->sgl_size, sgl->sgl,
sgl               509 drivers/misc/genwqe/card_utils.c 				 sgl->sgl_dma_addr);
sgl               511 drivers/misc/genwqe/card_utils.c 	sgl->sgl = NULL;
sgl               512 drivers/misc/genwqe/card_utils.c 	sgl->sgl_dma_addr = 0x0;
sgl               513 drivers/misc/genwqe/card_utils.c 	sgl->sgl_size = 0;
sgl              3107 drivers/misc/habanalabs/goya/goya.c static int goya_dma_map_sg(struct hl_device *hdev, struct scatterlist *sgl,
sgl              3113 drivers/misc/habanalabs/goya/goya.c 	if (!dma_map_sg(&hdev->pdev->dev, sgl, nents, dir))
sgl              3117 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgl, sg, nents, i)
sgl              3123 drivers/misc/habanalabs/goya/goya.c static void goya_dma_unmap_sg(struct hl_device *hdev, struct scatterlist *sgl,
sgl              3130 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgl, sg, nents, i)
sgl              3133 drivers/misc/habanalabs/goya/goya.c 	dma_unmap_sg(&hdev->pdev->dev, sgl, nents, dir);
sgl              3145 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgl              3200 drivers/misc/habanalabs/goya/goya.c 	rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl,
sgl              3657 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgl               539 drivers/misc/habanalabs/habanalabs.h 				struct scatterlist *sgl, int nents,
sgl               543 drivers/misc/habanalabs/habanalabs.h 				struct scatterlist *sgl, int nents,
sgl               193 drivers/misc/habanalabs/memory.c 	rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl,
sgl               673 drivers/misc/habanalabs/memory.c 	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
sgl               702 drivers/misc/habanalabs/memory.c 	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
sgl              1337 drivers/misc/habanalabs/memory.c 				userptr->sgt->sgl,
sgl                66 drivers/misc/mic/scif/scif_debugfs.c 		for_each_sg(window->st->sgl, sg, window->st->nents, j)
sgl               232 drivers/misc/mic/scif/scif_rma.c 				     window->st->sgl, window->st->nents,
sgl               539 drivers/misc/mic/scif/scif_rma.c 	for_each_sg(window->st->sgl, sg, window->st->nents, i)
sgl               542 drivers/misc/mic/scif/scif_rma.c 	err = dma_map_sg(&remote_dev->sdev->dev, window->st->sgl,
sgl               547 drivers/misc/mic/scif/scif_rma.c 	sg = window->st->sgl;
sgl               156 drivers/mmc/core/sdio_ops.c 		data.sg = sgtable.sgl;
sgl               777 drivers/mmc/host/dw_mmc.c 	struct scatterlist *sgl = host->data->sg;
sgl               806 drivers/mmc/host/dw_mmc.c 	desc = dmaengine_prep_slave_sg(host->dms->ch, sgl,
sgl               821 drivers/mmc/host/dw_mmc.c 		dma_sync_sg_for_device(mmc_dev(host->slot->mmc), sgl,
sgl               220 drivers/mmc/host/mxs-mmc.c 	struct scatterlist * sgl;
sgl               227 drivers/mmc/host/mxs-mmc.c 		sgl = data->sg;
sgl               231 drivers/mmc/host/mxs-mmc.c 		sgl = (struct scatterlist *) ssp->ssp_pio_words;
sgl               236 drivers/mmc/host/mxs-mmc.c 				sgl, sg_len, ssp->slave_dirn, flags);
sgl               349 drivers/mmc/host/mxs-mmc.c 	struct scatterlist *sgl = data->sg, *sg;
sgl               395 drivers/mmc/host/mxs-mmc.c 	for_each_sg(sgl, sg, sg_len, i)
sgl              1257 drivers/mmc/host/omap_hsmmc.c 		struct scatterlist *sgl;
sgl              1259 drivers/mmc/host/omap_hsmmc.c 		sgl = data->sg + i;
sgl              1260 drivers/mmc/host/omap_hsmmc.c 		if (sgl->length % data->blksz)
sgl               809 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 			     int raw_len, struct scatterlist *sgl,
sgl               817 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		sg_init_one(sgl, buf, len);
sgl               818 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		ret = dma_map_sg(this->dev, sgl, 1, dr);
sgl               827 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	sg_init_one(sgl, this->data_buffer_dma, len);
sgl               832 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	dma_map_sg(this->dev, sgl, 1, dr);
sgl              2273 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	sg_init_one(&transfer->sgl, transfer->cmdbuf, naddr + 1);
sgl              2274 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	dma_map_sg(this->dev, &transfer->sgl, 1, DMA_TO_DEVICE);
sgl              2278 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1, DMA_MEM_TO_DEV,
sgl              2315 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	*direct = prepare_data_dma(this, buf, raw_len, &transfer->sgl,
sgl              2331 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		pio[4] = transfer->sgl.dma_address;
sgl              2341 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1,
sgl              2362 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 	prepare_data_dma(this, buf, raw_len, &transfer->sgl, DMA_TO_DEVICE);
sgl              2377 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		pio[4] = transfer->sgl.dma_address;
sgl              2388 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		desc = dmaengine_prep_slave_sg(channel, &transfer->sgl, 1,
sgl              2533 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 			dma_unmap_sg(this->dev, &transfer->sgl, 1,
sgl               110 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.h 	struct scatterlist sgl;
sgl               201 drivers/mtd/nand/raw/lpc32xx_mlc.c 	struct scatterlist	sgl;
sgl               401 drivers/mtd/nand/raw/lpc32xx_mlc.c 	sg_init_one(&host->sgl, mem, len);
sgl               403 drivers/mtd/nand/raw/lpc32xx_mlc.c 	res = dma_map_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               409 drivers/mtd/nand/raw/lpc32xx_mlc.c 	desc = dmaengine_prep_slave_sg(host->dma_chan, &host->sgl, 1, dir,
sgl               425 drivers/mtd/nand/raw/lpc32xx_mlc.c 	dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               429 drivers/mtd/nand/raw/lpc32xx_mlc.c 	dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               228 drivers/mtd/nand/raw/lpc32xx_slc.c 	struct scatterlist	sgl;
sgl               450 drivers/mtd/nand/raw/lpc32xx_slc.c 	sg_init_one(&host->sgl, mem, len);
sgl               452 drivers/mtd/nand/raw/lpc32xx_slc.c 	res = dma_map_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               458 drivers/mtd/nand/raw/lpc32xx_slc.c 	desc = dmaengine_prep_slave_sg(host->dma_chan, &host->sgl, 1, dir,
sgl               474 drivers/mtd/nand/raw/lpc32xx_slc.c 	dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               479 drivers/mtd/nand/raw/lpc32xx_slc.c 	dma_unmap_sg(host->dma_chan->device->dev, &host->sgl, 1,
sgl               732 drivers/mtd/nand/raw/qcom_nandc.c 	struct scatterlist *sgl;
sgl               744 drivers/mtd/nand/raw/qcom_nandc.c 		sgl = &bam_txn->cmd_sgl[bam_txn->cmd_sgl_start];
sgl               750 drivers/mtd/nand/raw/qcom_nandc.c 		sgl = &bam_txn->data_sgl[bam_txn->tx_sgl_start];
sgl               756 drivers/mtd/nand/raw/qcom_nandc.c 		sgl = &bam_txn->data_sgl[bam_txn->rx_sgl_start];
sgl               763 drivers/mtd/nand/raw/qcom_nandc.c 	sg_mark_end(sgl + sgl_cnt - 1);
sgl               764 drivers/mtd/nand/raw/qcom_nandc.c 	ret = dma_map_sg(nandc->dev, sgl, sgl_cnt, desc->dir);
sgl               772 drivers/mtd/nand/raw/qcom_nandc.c 	desc->bam_sgl = sgl;
sgl               774 drivers/mtd/nand/raw/qcom_nandc.c 	dma_desc = dmaengine_prep_slave_sg(chan, sgl, sgl_cnt, dir_eng,
sgl               779 drivers/mtd/nand/raw/qcom_nandc.c 		dma_unmap_sg(nandc->dev, sgl, sgl_cnt, desc->dir);
sgl               898 drivers/mtd/nand/raw/qcom_nandc.c 	struct scatterlist *sgl;
sgl               907 drivers/mtd/nand/raw/qcom_nandc.c 	sgl = &desc->adm_sgl;
sgl               909 drivers/mtd/nand/raw/qcom_nandc.c 	sg_init_one(sgl, vaddr, size);
sgl               919 drivers/mtd/nand/raw/qcom_nandc.c 	ret = dma_map_sg(nandc->dev, sgl, 1, desc->dir);
sgl               944 drivers/mtd/nand/raw/qcom_nandc.c 	dma_desc = dmaengine_prep_slave_sg(nandc->chan, sgl, 1, dir_eng, 0);
sgl               915 drivers/mtd/nand/raw/stm32_fmc2_nand.c 	for_each_sg(fmc2->dma_data_sg.sgl, sg, eccsteps, s) {
sgl               920 drivers/mtd/nand/raw/stm32_fmc2_nand.c 	ret = dma_map_sg(fmc2->dev, fmc2->dma_data_sg.sgl,
sgl               925 drivers/mtd/nand/raw/stm32_fmc2_nand.c 	desc_data = dmaengine_prep_slave_sg(dma_ch, fmc2->dma_data_sg.sgl,
sgl               946 drivers/mtd/nand/raw/stm32_fmc2_nand.c 		for_each_sg(fmc2->dma_ecc_sg.sgl, sg, eccsteps, s) {
sgl               951 drivers/mtd/nand/raw/stm32_fmc2_nand.c 		ret = dma_map_sg(fmc2->dev, fmc2->dma_ecc_sg.sgl,
sgl               957 drivers/mtd/nand/raw/stm32_fmc2_nand.c 						   fmc2->dma_ecc_sg.sgl,
sgl              1014 drivers/mtd/nand/raw/stm32_fmc2_nand.c 		dma_unmap_sg(fmc2->dev, fmc2->dma_ecc_sg.sgl,
sgl              1018 drivers/mtd/nand/raw/stm32_fmc2_nand.c 	dma_unmap_sg(fmc2->dev, fmc2->dma_data_sg.sgl, eccsteps, dma_data_dir);
sgl               749 drivers/mtd/ubi/eba.c 			struct ubi_sgl *sgl, int lnum, int offset, int len,
sgl               757 drivers/mtd/ubi/eba.c 		ubi_assert(sgl->list_pos < UBI_MAX_SG_COUNT);
sgl               758 drivers/mtd/ubi/eba.c 		sg = &sgl->sg[sgl->list_pos];
sgl               759 drivers/mtd/ubi/eba.c 		if (len < sg->length - sgl->page_pos)
sgl               762 drivers/mtd/ubi/eba.c 			to_read = sg->length - sgl->page_pos;
sgl               765 drivers/mtd/ubi/eba.c 				       sg_virt(sg) + sgl->page_pos, offset,
sgl               773 drivers/mtd/ubi/eba.c 			sgl->page_pos += to_read;
sgl               774 drivers/mtd/ubi/eba.c 			if (sgl->page_pos == sg->length) {
sgl               775 drivers/mtd/ubi/eba.c 				sgl->list_pos++;
sgl               776 drivers/mtd/ubi/eba.c 				sgl->page_pos = 0;
sgl               782 drivers/mtd/ubi/eba.c 		sgl->list_pos++;
sgl               783 drivers/mtd/ubi/eba.c 		sgl->page_pos = 0;
sgl               462 drivers/mtd/ubi/kapi.c int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl,
sgl               478 drivers/mtd/ubi/kapi.c 	err = ubi_eba_read_leb_sg(ubi, vol, sgl, lnum, offset, len, check);
sgl               889 drivers/mtd/ubi/ubi.h 			struct ubi_sgl *sgl, int lnum, int offset, int len,
sgl               369 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[i])));
sgl               372 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	   sge_len - 1, le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1]));
sgl               377 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1]));
sgl               623 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		u16 sge_idx = RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[j]));
sgl              3965 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h 	__le16 sgl[8];
sgl              2881 drivers/net/ethernet/broadcom/cnic_defs.h 	struct fcoe_ext_mul_sges_ctx sgl;
sgl              1094 drivers/net/ethernet/chelsio/cxgb3/sge.c 			     const struct sg_ent *sgl,
sgl              1119 drivers/net/ethernet/chelsio/cxgb3/sge.c 		const u64 *fp = (const u64 *)sgl;
sgl              1186 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sg_ent *sgp, sgl[MAX_SKB_FRAGS / 2 + 1];
sgl              1240 drivers/net/ethernet/chelsio/cxgb3/sge.c 	sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl;
sgl              1243 drivers/net/ethernet/chelsio/cxgb3/sge.c 	write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits, gen,
sgl              1602 drivers/net/ethernet/chelsio/cxgb3/sge.c 				     const struct sg_ent *sgl, int sgl_flits)
sgl              1609 drivers/net/ethernet/chelsio/cxgb3/sge.c 	for (p = dui->addr; sgl_flits >= 3; sgl++, sgl_flits -= 3) {
sgl              1610 drivers/net/ethernet/chelsio/cxgb3/sge.c 		*p++ = be64_to_cpu(sgl->addr[0]);
sgl              1611 drivers/net/ethernet/chelsio/cxgb3/sge.c 		*p++ = be64_to_cpu(sgl->addr[1]);
sgl              1614 drivers/net/ethernet/chelsio/cxgb3/sge.c 		*p = be64_to_cpu(sgl->addr[0]);
sgl              1636 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sg_ent *sgp, sgl[MAX_SKB_FRAGS / 2 + 1];
sgl              1652 drivers/net/ethernet/chelsio/cxgb3/sge.c 	sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl;
sgl              1661 drivers/net/ethernet/chelsio/cxgb3/sge.c 	write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits,
sgl              1093 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	struct ulptx_sgl *sgl;
sgl              1902 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 		     struct ulptx_sgl *sgl, u64 *end, unsigned int start,
sgl              3536 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		dma_unmap_sg(adapter->pdev_dev, adapter->hma.sgt->sgl,
sgl              3541 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(adapter->hma.sgt->sgl, iter,
sgl              3556 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct scatterlist *sgl, *iter;
sgl              3610 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgl = adapter->hma.sgt->sgl;
sgl              3612 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(sgl, iter, sgt->orig_nents, i) {
sgl              3624 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgt->nents = dma_map_sg(adapter->pdev_dev, sgl, sgt->orig_nents,
sgl              3639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(sgl, iter, sgt->nents, i) {
sgl               302 drivers/net/ethernet/chelsio/cxgb4/sge.c 		      const struct ulptx_sgl *sgl, const struct sge_txq *q)
sgl               308 drivers/net/ethernet/chelsio/cxgb4/sge.c 		dma_unmap_single(dev, be64_to_cpu(sgl->addr0), ntohl(sgl->len0),
sgl               311 drivers/net/ethernet/chelsio/cxgb4/sge.c 		dma_unmap_page(dev, be64_to_cpu(sgl->addr0), ntohl(sgl->len0),
sgl               320 drivers/net/ethernet/chelsio/cxgb4/sge.c 	for (p = sgl->sge; nfrags >= 2; nfrags -= 2) {
sgl               381 drivers/net/ethernet/chelsio/cxgb4/sge.c 				unmap_sgl(dev, d->skb, d->sgl, q);
sgl               881 drivers/net/ethernet/chelsio/cxgb4/sge.c 		     struct ulptx_sgl *sgl, u64 *end, unsigned int start,
sgl               892 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl->len0 = htonl(len);
sgl               893 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl->addr0 = cpu_to_be64(addr[0] + start);
sgl               896 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl->len0 = htonl(skb_frag_size(&si->frags[0]));
sgl               897 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl->addr0 = cpu_to_be64(addr[1]);
sgl               900 drivers/net/ethernet/chelsio/cxgb4/sge.c 	sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) |
sgl               909 drivers/net/ethernet/chelsio/cxgb4/sge.c 	to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge;
sgl               923 drivers/net/ethernet/chelsio/cxgb4/sge.c 		unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1;
sgl               926 drivers/net/ethernet/chelsio/cxgb4/sge.c 			memcpy(sgl->sge, buf, part0);
sgl              1366 drivers/net/ethernet/chelsio/cxgb4/sge.c 	u64 cntrl, *end, *sgl;
sgl              1544 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl = (u64 *)(cpl + 1); /* sgl start here */
sgl              1545 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (unlikely((u8 *)sgl >= (u8 *)q->q.stat)) {
sgl              1550 drivers/net/ethernet/chelsio/cxgb4/sge.c 			if (sgl == (u64 *)q->q.stat) {
sgl              1554 drivers/net/ethernet/chelsio/cxgb4/sge.c 				sgl = (void *)q->q.desc;
sgl              1567 drivers/net/ethernet/chelsio/cxgb4/sge.c 		sgl = (u64 *)(cpl + 1);
sgl              1601 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_inline_tx_skb(skb, &q->q, sgl);
sgl              1606 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_write_sgl(skb, &q->q, (void *)sgl, end, 0, addr);
sgl              1613 drivers/net/ethernet/chelsio/cxgb4/sge.c 		q->q.sdesc[last_desc].sgl = (struct ulptx_sgl *)sgl;
sgl              1940 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct ulptx_sgl *sgl = (struct ulptx_sgl *)(cpl + 1);
sgl              1950 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (unlikely((void *)sgl == (void *)tq->stat)) {
sgl              1951 drivers/net/ethernet/chelsio/cxgb4/sge.c 			sgl = (void *)tq->desc;
sgl              1956 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_write_sgl(skb, tq, sgl, end, 0, addr);
sgl              1963 drivers/net/ethernet/chelsio/cxgb4/sge.c 		tq->sdesc[last_desc].sgl = sgl;
sgl               159 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct ulptx_sgl *sgl;		/* scatter/gather list in TX Queue */
sgl               307 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		      const struct ulptx_sgl *sgl, const struct sge_txq *tq)
sgl               313 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		dma_unmap_single(dev, be64_to_cpu(sgl->addr0),
sgl               314 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				 be32_to_cpu(sgl->len0), DMA_TO_DEVICE);
sgl               316 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		dma_unmap_page(dev, be64_to_cpu(sgl->addr0),
sgl               317 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			       be32_to_cpu(sgl->len0), DMA_TO_DEVICE);
sgl               325 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	for (p = sgl->sge; nfrags >= 2; nfrags -= 2) {
sgl               394 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq);
sgl               902 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		      struct ulptx_sgl *sgl, u64 *end, unsigned int start,
sgl               913 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		sgl->len0 = htonl(len);
sgl               914 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		sgl->addr0 = cpu_to_be64(addr[0] + start);
sgl               917 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		sgl->len0 = htonl(skb_frag_size(&si->frags[0]));
sgl               918 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		sgl->addr0 = cpu_to_be64(addr[1]);
sgl               921 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) |
sgl               930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge;
sgl               944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1;
sgl               947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			memcpy(sgl->sge, buf, part0);
sgl              1419 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		struct ulptx_sgl *sgl = (struct ulptx_sgl *)(cpl + 1);
sgl              1430 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (unlikely((void *)sgl == (void *)tq->stat)) {
sgl              1431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			sgl = (void *)tq->desc;
sgl              1435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		write_sgl(skb, tq, sgl, end, 0, addr);
sgl              1442 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		tq->sdesc[last_desc].sgl = sgl;
sgl                95 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h 	struct scatterlist *sgl;
sgl               781 drivers/net/ethernet/emulex/benet/be_cmds.c 	return &wrb->payload.sgl[0];
sgl                39 drivers/net/ethernet/emulex/benet/be_cmds.h 		struct be_sge sgl[19];    /* used by non-embedded cmds */
sgl               929 drivers/net/ethernet/intel/ixgbe/ixgbe.h 		       struct scatterlist *sgl, unsigned int sgc);
sgl               931 drivers/net/ethernet/intel/ixgbe/ixgbe.h 			  struct scatterlist *sgl, unsigned int sgc);
sgl                28 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	ddp->sgl = NULL;
sgl               109 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (ddp->sgl)
sgl               110 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc,
sgl               133 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 				struct scatterlist *sgl, unsigned int sgc,
sgl               152 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (!netdev || !sgl)
sgl               168 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	if (ddp->sgl) {
sgl               170 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		      xid, ddp->sgl, ddp->sgc);
sgl               188 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE);
sgl               201 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	ddp->sgl = sgl;
sgl               205 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	for_each_sg(sgl, sg, dmacount, i) {
sgl               323 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	dma_unmap_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE);
sgl               344 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		       struct scatterlist *sgl, unsigned int sgc)
sgl               346 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 0);
sgl               365 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 			    struct scatterlist *sgl, unsigned int sgc)
sgl               367 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 	return ixgbe_fcoe_ddp_setup(netdev, xid, sgl, sgc, 1);
sgl               441 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		dma_unmap_sg(&adapter->pdev->dev, ddp->sgl,
sgl               444 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c 		ddp->sgl = NULL;
sgl                39 drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.h 	struct scatterlist *sgl;
sgl               391 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 	struct scatterlist *sgl;
sgl               432 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 	mmc_dat.sg = sdiodev->sgtable.sgl;
sgl               447 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 	sgl = sdiodev->sgtable.sgl;
sgl               458 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 			sg_set_buf(sgl, pkt_data, sg_data_sz);
sgl               461 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 			sgl = sg_next(sgl);
sgl               472 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 				sgl = sdiodev->sgtable.sgl;
sgl               512 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c 	sg_init_table(sdiodev->sgtable.sgl, sdiodev->sgtable.orig_nents);
sgl                98 drivers/nvme/host/fc.c 	struct scatterlist	sgl[SG_CHUNK_SIZE];
sgl              1774 drivers/nvme/host/fc.c 	op->op.fcp_req.first_sgl = &op->sgl[0];
sgl              2141 drivers/nvme/host/fc.c 	freq->sg_table.sgl = freq->first_sgl;
sgl              2143 drivers/nvme/host/fc.c 			blk_rq_nr_phys_segments(rq), freq->sg_table.sgl,
sgl              2148 drivers/nvme/host/fc.c 	op->nents = blk_rq_map_sg(rq->q, rq, freq->sg_table.sgl);
sgl              2150 drivers/nvme/host/fc.c 	freq->sg_cnt = fc_dma_map_sg(ctrl->lport->dev, freq->sg_table.sgl,
sgl              2173 drivers/nvme/host/fc.c 	fc_dma_unmap_sg(ctrl->lport->dev, freq->sg_table.sgl, op->nents,
sgl              2260 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) |
sgl              2262 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.length = cpu_to_le32(data_len);
sgl              2263 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.addr = 0;
sgl               571 drivers/nvme/host/pci.c static void nvme_print_sgl(struct scatterlist *sgl, int nents)
sgl               576 drivers/nvme/host/pci.c 	for_each_sg(sgl, sg, nents, i) {
sgl               713 drivers/nvme/host/pci.c 		nvme_pci_sgl_set_data(&cmd->dptr.sgl, sg);
sgl               734 drivers/nvme/host/pci.c 	nvme_pci_sgl_set_seg(&cmd->dptr.sgl, sgl_dma, entries);
sgl               789 drivers/nvme/host/pci.c 	cmnd->dptr.sgl.addr = cpu_to_le64(iod->first_dma);
sgl               790 drivers/nvme/host/pci.c 	cmnd->dptr.sgl.length = cpu_to_le32(iod->dma_len);
sgl               791 drivers/nvme/host/pci.c 	cmnd->dptr.sgl.type = NVME_SGL_FMT_DATA_DESC << 4;
sgl              1164 drivers/nvme/host/rdma.c 	ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq));
sgl              1185 drivers/nvme/host/rdma.c 	struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
sgl              1186 drivers/nvme/host/rdma.c 	struct scatterlist *sgl = req->sg_table.sgl;
sgl              1191 drivers/nvme/host/rdma.c 	for (i = 0; i < count; i++, sgl++, sge++) {
sgl              1192 drivers/nvme/host/rdma.c 		sge->addr = sg_dma_address(sgl);
sgl              1193 drivers/nvme/host/rdma.c 		sge->length = sg_dma_len(sgl);
sgl              1211 drivers/nvme/host/rdma.c 	sg->addr = cpu_to_le64(sg_dma_address(req->sg_table.sgl));
sgl              1212 drivers/nvme/host/rdma.c 	put_unaligned_le24(sg_dma_len(req->sg_table.sgl), sg->length);
sgl              1233 drivers/nvme/host/rdma.c 	nr = ib_map_mr_sg(req->mr, req->sg_table.sgl, count, NULL, SZ_4K);
sgl              1280 drivers/nvme/host/rdma.c 	req->sg_table.sgl = req->first_sgl;
sgl              1282 drivers/nvme/host/rdma.c 			blk_rq_nr_phys_segments(rq), req->sg_table.sgl,
sgl              1287 drivers/nvme/host/rdma.c 	req->nents = blk_rq_map_sg(rq->q, rq, req->sg_table.sgl);
sgl              1289 drivers/nvme/host/rdma.c 	count = ib_dma_map_sg(ibdev, req->sg_table.sgl, req->nents,
sgl              1319 drivers/nvme/host/rdma.c 	ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq));
sgl              1988 drivers/nvme/host/tcp.c 	struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
sgl              1999 drivers/nvme/host/tcp.c 	struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
sgl              2009 drivers/nvme/host/tcp.c 	struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
sgl               157 drivers/nvme/target/loop.c 		iod->sg_table.sgl = iod->first_sgl;
sgl               160 drivers/nvme/target/loop.c 				iod->sg_table.sgl, SG_CHUNK_SIZE)) {
sgl               165 drivers/nvme/target/loop.c 		iod->req.sg = iod->sg_table.sgl;
sgl               166 drivers/nvme/target/loop.c 		iod->req.sg_cnt = blk_rq_map_sg(req->q, req, iod->sg_table.sgl);
sgl               635 drivers/nvme/target/rdma.c 	struct nvme_sgl_desc *sgl = &rsp->req.cmd->common.dptr.sgl;
sgl               636 drivers/nvme/target/rdma.c 	u64 off = le64_to_cpu(sgl->addr);
sgl               637 drivers/nvme/target/rdma.c 	u32 len = le32_to_cpu(sgl->length);
sgl               661 drivers/nvme/target/rdma.c 		struct nvme_keyed_sgl_desc *sgl, bool invalidate)
sgl               664 drivers/nvme/target/rdma.c 	u64 addr = le64_to_cpu(sgl->addr);
sgl               665 drivers/nvme/target/rdma.c 	u32 key = get_unaligned_le32(sgl->key);
sgl               668 drivers/nvme/target/rdma.c 	rsp->req.transfer_len = get_unaligned_le24(sgl->length);
sgl               699 drivers/nvme/target/rdma.c 	struct nvme_keyed_sgl_desc *sgl = &rsp->req.cmd->common.dptr.ksgl;
sgl               701 drivers/nvme/target/rdma.c 	switch (sgl->type >> 4) {
sgl               703 drivers/nvme/target/rdma.c 		switch (sgl->type & 0xf) {
sgl               707 drivers/nvme/target/rdma.c 			pr_err("invalid SGL subtype: %#x\n", sgl->type);
sgl               713 drivers/nvme/target/rdma.c 		switch (sgl->type & 0xf) {
sgl               715 drivers/nvme/target/rdma.c 			return nvmet_rdma_map_sgl_keyed(rsp, sgl, true);
sgl               717 drivers/nvme/target/rdma.c 			return nvmet_rdma_map_sgl_keyed(rsp, sgl, false);
sgl               719 drivers/nvme/target/rdma.c 			pr_err("invalid SGL subtype: %#x\n", sgl->type);
sgl               725 drivers/nvme/target/rdma.c 		pr_err("invalid SGL type: %#x\n", sgl->type);
sgl               320 drivers/nvme/target/tcp.c 	struct nvme_sgl_desc *sgl = &cmd->req.cmd->common.dptr.sgl;
sgl               321 drivers/nvme/target/tcp.c 	u32 len = le32_to_cpu(sgl->length);
sgl               326 drivers/nvme/target/tcp.c 	if (sgl->type == ((NVME_SGL_FMT_DATA_DESC << 4) |
sgl               825 drivers/nvme/target/tcp.c 	req->data_len = le32_to_cpu(req->cmd->common.dptr.sgl.length);
sgl               913 drivers/nvme/target/tcp.c 			le32_to_cpu(req->cmd->common.dptr.sgl.length));
sgl               772 drivers/pci/p2pdma.c void pci_p2pmem_free_sgl(struct pci_dev *pdev, struct scatterlist *sgl)
sgl               777 drivers/pci/p2pdma.c 	for_each_sg(sgl, sg, INT_MAX, count) {
sgl               783 drivers/pci/p2pdma.c 	kfree(sgl);
sgl               578 drivers/rapidio/devices/rio_mport_cdev.c 		     req->sgt.sgl, req->sgt.nents, req->dir);
sgl               622 drivers/rapidio/devices/rio_mport_cdev.c 	tx_data.sg = sgt->sgl;
sgl               930 drivers/rapidio/devices/rio_mport_cdev.c 		sg_set_buf(req->sgt.sgl,
sgl               936 drivers/rapidio/devices/rio_mport_cdev.c 			   req->sgt.sgl, req->sgt.nents, dir);
sgl               825 drivers/rapidio/devices/tsi721_dma.c 			struct scatterlist *sgl, unsigned int sg_len,
sgl               835 drivers/rapidio/devices/tsi721_dma.c 	if (!sgl || !sg_len) {
sgl               877 drivers/rapidio/devices/tsi721_dma.c 		desc->sg	= sgl;
sgl               488 drivers/scsi/3w-9xxx.c 	command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]);
sgl               489 drivers/scsi/3w-9xxx.c 	command_packet->byte8_offset.param.sgl[0].length = cpu_to_le32(TW_SECTOR_SIZE);
sgl              1081 drivers/scsi/3w-9xxx.c 	command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]);
sgl              1082 drivers/scsi/3w-9xxx.c 	command_packet->byte8_offset.param.sgl[0].length = cpu_to_le32(TW_SECTOR_SIZE);
sgl              1384 drivers/scsi/3w-9xxx.c 	TW_SG_Entry *sgl;
sgl              1407 drivers/scsi/3w-9xxx.c 				sgl = (TW_SG_Entry *)((u32 *)oldcommand+oldcommand->size - (sizeof(TW_SG_Entry)/4) + pae);
sgl              1409 drivers/scsi/3w-9xxx.c 				sgl = (TW_SG_Entry *)((u32 *)oldcommand+TW_SGL_OUT(oldcommand->opcode__sgloffset));
sgl              1410 drivers/scsi/3w-9xxx.c 			sgl->address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1);
sgl              1411 drivers/scsi/3w-9xxx.c 			sgl->length = cpu_to_le32(length);
sgl               498 drivers/scsi/3w-9xxx.h 			TW_SG_Entry sgl[TW_ESCALADE_MAX_SGL_LENGTH];
sgl               502 drivers/scsi/3w-9xxx.h 			TW_SG_Entry sgl[TW_ESCALADE_MAX_SGL_LENGTH];
sgl               423 drivers/scsi/3w-sas.c 	command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]);
sgl               424 drivers/scsi/3w-sas.c 	command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE);
sgl               683 drivers/scsi/3w-sas.c 	TW_SG_Entry_ISO *sgl;
sgl               705 drivers/scsi/3w-sas.c 			sgl = (TW_SG_Entry_ISO *)((u32 *)oldcommand+oldcommand->size - (sizeof(TW_SG_Entry_ISO)/4) + pae + (sizeof(dma_addr_t) > 4 ? 1 : 0));
sgl               706 drivers/scsi/3w-sas.c 			sgl->address = TW_CPU_TO_SGL(dma_handle + sizeof(TW_Ioctl_Buf_Apache) - 1);
sgl               707 drivers/scsi/3w-sas.c 			sgl->length = TW_CPU_TO_SGL(length);
sgl               951 drivers/scsi/3w-sas.c 	command_packet->byte8_offset.param.sgl[0].address = TW_CPU_TO_SGL(tw_dev->generic_buffer_phys[request_id]);
sgl               952 drivers/scsi/3w-sas.c 	command_packet->byte8_offset.param.sgl[0].length = TW_CPU_TO_SGL(TW_SECTOR_SIZE);
sgl               229 drivers/scsi/3w-sas.h 			TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD];
sgl               233 drivers/scsi/3w-sas.h 			TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD];
sgl               589 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl               590 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl               731 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl               732 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl               967 drivers/scsi/3w-xxxx.c 				tw_ioctl->firmware_command.byte8.param.sgl[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1;
sgl               968 drivers/scsi/3w-xxxx.c 				tw_ioctl->firmware_command.byte8.param.sgl[0].length = data_buffer_length_adjusted;
sgl               971 drivers/scsi/3w-xxxx.c 				tw_ioctl->firmware_command.byte8.io.sgl[0].address = dma_handle + sizeof(TW_New_Ioctl) - 1;
sgl               972 drivers/scsi/3w-xxxx.c 				tw_ioctl->firmware_command.byte8.io.sgl[0].length = data_buffer_length_adjusted;
sgl              1167 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl              1168 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl              1439 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl              1440 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl              1549 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl              1550 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl              1639 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl              1640 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl              1762 drivers/scsi/3w-xxxx.c 		command_packet->byte8.io.sgl[i].address = sg_dma_address(sg);
sgl              1763 drivers/scsi/3w-xxxx.c 		command_packet->byte8.io.sgl[i].length = sg_dma_len(sg);
sgl              1884 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].address = param_value;
sgl              1885 drivers/scsi/3w-xxxx.c 	command_packet->byte8.param.sgl[0].length = sizeof(TW_Sector);
sgl               307 drivers/scsi/3w-xxxx.h 			TW_SG_Entry sgl[TW_MAX_SGL_LENGTH];
sgl               311 drivers/scsi/3w-xxxx.h 			TW_SG_Entry sgl[TW_MAX_SGL_LENGTH];
sgl                43 drivers/scsi/be2iscsi/be_cmds.h 		struct be_sge sgl[19];	/* used by non-embedded cmds */
sgl               698 drivers/scsi/be2iscsi/be_cmds.h 	return &wrb->payload.sgl[0];
sgl              2935 drivers/scsi/be2iscsi/be_main.c 			 struct be_dma_mem *sgl)
sgl              2940 drivers/scsi/be2iscsi/be_main.c 	WARN_ON(!sgl);
sgl              2942 drivers/scsi/be2iscsi/be_main.c 	sgl->va = virtual_address;
sgl              2943 drivers/scsi/be2iscsi/be_main.c 	sgl->dma = (unsigned long)physical_address;
sgl              2944 drivers/scsi/be2iscsi/be_main.c 	sgl->size = length;
sgl              2949 drivers/scsi/be2iscsi/be_main.c static void be_sgl_destroy_contiguous(struct be_dma_mem *sgl)
sgl              2951 drivers/scsi/be2iscsi/be_main.c 	memset(sgl, 0, sizeof(*sgl));
sgl              2956 drivers/scsi/be2iscsi/be_main.c 		     struct mem_array *pmem, struct be_dma_mem *sgl)
sgl              2958 drivers/scsi/be2iscsi/be_main.c 	if (sgl->va)
sgl              2959 drivers/scsi/be2iscsi/be_main.c 		be_sgl_destroy_contiguous(sgl);
sgl              2963 drivers/scsi/be2iscsi/be_main.c 				 pmem->size, sgl);
sgl              2968 drivers/scsi/be2iscsi/be_main.c 			   struct mem_array *pmem, struct be_dma_mem *sgl)
sgl              2970 drivers/scsi/be2iscsi/be_main.c 	if (sgl->va)
sgl              2971 drivers/scsi/be2iscsi/be_main.c 		be_sgl_destroy_contiguous(sgl);
sgl              2975 drivers/scsi/be2iscsi/be_main.c 				 pmem->size, sgl);
sgl              3240 drivers/scsi/be2iscsi/be_main.c 	struct be_dma_mem sgl;
sgl              3250 drivers/scsi/be2iscsi/be_main.c 			hwi_build_be_sgl_arr(phba, pm_arr, &sgl);
sgl              3252 drivers/scsi/be2iscsi/be_main.c 				 &phba->ctrl, &sgl);
sgl              3275 drivers/scsi/be2iscsi/be_main.c 	struct be_dma_mem sgl;
sgl              3289 drivers/scsi/be2iscsi/be_main.c 		hwi_build_be_sgl_arr(phba, pm_arr, &sgl);
sgl              3290 drivers/scsi/be2iscsi/be_main.c 		status = be_cmd_iscsi_post_sgl_pages(&phba->ctrl, &sgl,
sgl              3342 drivers/scsi/be2iscsi/be_main.c 	struct be_dma_mem sgl;
sgl              3414 drivers/scsi/be2iscsi/be_main.c 		hwi_build_be_sgl_by_offset(phba, &pwrb_arr[i], &sgl);
sgl              3415 drivers/scsi/be2iscsi/be_main.c 		status = be_cmd_wrbq_create(&phba->ctrl, &sgl,
sgl               579 drivers/scsi/bnx2fc/57xx_hsi_bnx2fc.h 	struct fcoe_ext_mul_sges_ctx sgl;
sgl              1458 drivers/scsi/bnx2fc/bnx2fc_hwi.c 	struct fcoe_ext_mul_sges_ctx *sgl;
sgl              1506 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo =
sgl              1508 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi =
sgl              1510 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size =
sgl              1512 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_off =
sgl              1514 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_idx = i;
sgl              1520 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl;
sgl              1521 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.lo = (u32)phys_addr;
sgl              1522 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.hi = (u32)((u64)phys_addr >> 32);
sgl              1523 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.sgl_size = bd_count;
sgl              1524 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_off = offset; /*adjusted offset */
sgl              1525 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_idx = i;
sgl              1577 drivers/scsi/bnx2fc/bnx2fc_hwi.c 	struct fcoe_ext_mul_sges_ctx *sgl;
sgl              1603 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo =
sgl              1605 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi =
sgl              1607 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size = 1;
sgl              1658 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl;
sgl              1660 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.lo = (u32)mp_req->mp_resp_bd_dma;
sgl              1661 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.hi =
sgl              1663 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.sgl_size = 1;
sgl              1675 drivers/scsi/bnx2fc/bnx2fc_hwi.c 	struct fcoe_ext_mul_sges_ctx *sgl;
sgl              1713 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.lo =
sgl              1715 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			task->txwr_only.sgl_ctx.sgl.mul_sgl.cur_sge_addr.hi =
sgl              1717 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			task->txwr_only.sgl_ctx.sgl.mul_sgl.sgl_size =
sgl              1776 drivers/scsi/bnx2fc/bnx2fc_hwi.c 	sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl;
sgl              1806 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			sgl->mul_sgl.cur_sge_addr.lo = (u32)bd_tbl->bd_tbl_dma;
sgl              1807 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			sgl->mul_sgl.cur_sge_addr.hi =
sgl              1809 drivers/scsi/bnx2fc/bnx2fc_hwi.c 			sgl->mul_sgl.sgl_size = bd_count;
sgl              1812 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.lo = (u32)bd_tbl->bd_tbl_dma;
sgl              1813 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.cur_sge_addr.hi =
sgl              1815 drivers/scsi/bnx2fc/bnx2fc_hwi.c 		sgl->mul_sgl.sgl_size = bd_count;
sgl               291 drivers/scsi/csiostor/csio_scsi.c 			   struct ulptx_sgl *sgl)
sgl               301 drivers/scsi/csiostor/csio_scsi.c 	sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | ULPTX_MORE_F |
sgl               307 drivers/scsi/csiostor/csio_scsi.c 				sgl->addr0 = cpu_to_be64(sg_dma_address(sgel));
sgl               308 drivers/scsi/csiostor/csio_scsi.c 				sgl->len0 = cpu_to_be32(sg_dma_len(sgel));
sgl               309 drivers/scsi/csiostor/csio_scsi.c 				sge_pair = (struct ulptx_sge_pair *)(sgl + 1);
sgl               331 drivers/scsi/csiostor/csio_scsi.c 				sgl->addr0 = cpu_to_be64(dma_buf->paddr);
sgl               332 drivers/scsi/csiostor/csio_scsi.c 				sgl->len0 = cpu_to_be32(
sgl               334 drivers/scsi/csiostor/csio_scsi.c 				sge_pair = (struct ulptx_sge_pair *)(sgl + 1);
sgl               365 drivers/scsi/csiostor/csio_scsi.c 	struct ulptx_sgl *sgl;
sgl               397 drivers/scsi/csiostor/csio_scsi.c 	sgl = (struct ulptx_sgl *)((uintptr_t)wrp +
sgl               401 drivers/scsi/csiostor/csio_scsi.c 	csio_scsi_init_ultptx_dsgl(hw, req, sgl);
sgl               418 drivers/scsi/csiostor/csio_scsi.c 	struct ulptx_sgl *sgl;
sgl               450 drivers/scsi/csiostor/csio_scsi.c 	sgl = (struct ulptx_sgl *)((uintptr_t)wrp +
sgl               454 drivers/scsi/csiostor/csio_scsi.c 	csio_scsi_init_ultptx_dsgl(hw, req, sgl);
sgl              1092 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	struct scatterlist *sg = ttinfo->sgl;
sgl              1977 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c 	struct scatterlist *sg = ttinfo->sgl;
sgl              1216 drivers/scsi/cxgbi/libcxgbi.c scmd_get_params(struct scsi_cmnd *sc, struct scatterlist **sgl,
sgl              1222 drivers/scsi/cxgbi/libcxgbi.c 	*sgl = sdb->table.sgl;
sgl              1314 drivers/scsi/cxgbi/libcxgbi.c static int cxgbi_ddp_sgl_check(struct scatterlist *sgl, int nents)
sgl              1318 drivers/scsi/cxgbi/libcxgbi.c 	struct scatterlist *sg = sgl;
sgl              1344 drivers/scsi/cxgbi/libcxgbi.c 	struct scatterlist *sgl = ttinfo->sgl;
sgl              1346 drivers/scsi/cxgbi/libcxgbi.c 	unsigned int sg_offset = sgl->offset;
sgl              1365 drivers/scsi/cxgbi/libcxgbi.c 	if (cxgbi_ddp_sgl_check(sgl, sgcnt) < 0)
sgl              1368 drivers/scsi/cxgbi/libcxgbi.c 	ttinfo->nr_pages = (xferlen + sgl->offset + (1 << PAGE_SHIFT) - 1) >>
sgl              1385 drivers/scsi/cxgbi/libcxgbi.c 	sgl->offset = 0;
sgl              1386 drivers/scsi/cxgbi/libcxgbi.c 	err = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE);
sgl              1387 drivers/scsi/cxgbi/libcxgbi.c 	sgl->offset = sg_offset;
sgl              1402 drivers/scsi/cxgbi/libcxgbi.c 	cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset,
sgl              1422 drivers/scsi/cxgbi/libcxgbi.c 		dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE);
sgl              1447 drivers/scsi/cxgbi/libcxgbi.c 		dma_unmap_sg(&ppm->pdev->dev, ttinfo->sgl, ttinfo->nents,
sgl              1475 drivers/scsi/cxgbi/libcxgbi.c 		scmd_get_params(sc, &ttinfo->sgl, &ttinfo->nents,
sgl              1818 drivers/scsi/cxgbi/libcxgbi.c static int sgl_seek_offset(struct scatterlist *sgl, unsigned int sgcnt,
sgl              1825 drivers/scsi/cxgbi/libcxgbi.c 	for_each_sg(sgl, sg, sgcnt, i) {
sgl              1987 drivers/scsi/cxgbi/libcxgbi.c 					sdb->table.sgl, sdb->table.nents,
sgl              1151 drivers/scsi/esas2r/esas2r.h 	struct list_head *sgl;
sgl              1156 drivers/scsi/esas2r/esas2r.h 		sgl = a->free_sg_list_head.next;
sgl              1157 drivers/scsi/esas2r/esas2r.h 		result = list_entry(sgl, struct esas2r_mem_desc, next_desc);
sgl              1158 drivers/scsi/esas2r/esas2r.h 		list_del_init(sgl);
sgl               794 drivers/scsi/esas2r/esas2r_init.c 	struct esas2r_mem_desc *sgl;
sgl               863 drivers/scsi/esas2r/esas2r_init.c 	for (i = 0, sgl = a->sg_list_mds; i < num_sg_lists; i++, sgl++) {
sgl               864 drivers/scsi/esas2r/esas2r_init.c 		sgl->size = sgl_page_size;
sgl               866 drivers/scsi/esas2r/esas2r_init.c 		list_add_tail(&sgl->next_desc, &a->free_sg_list_head);
sgl               868 drivers/scsi/esas2r/esas2r_init.c 		if (!esas2r_initmem_alloc(a, sgl, ESAS2R_SGL_ALIGN)) {
sgl               224 drivers/scsi/esas2r/esas2r_io.c 			struct esas2r_mem_desc *sgl;
sgl               231 drivers/scsi/esas2r/esas2r_io.c 			sgl = esas2r_alloc_sgl(a);
sgl               233 drivers/scsi/esas2r/esas2r_io.c 			if (unlikely(sgl == NULL))
sgl               244 drivers/scsi/esas2r/esas2r_io.c 			memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen);
sgl               248 drivers/scsi/esas2r/esas2r_io.c 				(struct atto_vda_sge *)((u8 *)sgl->virt_addr +
sgl               253 drivers/scsi/esas2r/esas2r_io.c 				(struct atto_vda_sge *)((u8 *)sgl->virt_addr
sgl               260 drivers/scsi/esas2r/esas2r_io.c 				cpu_to_le64(sgl->phys_addr);
sgl               302 drivers/scsi/esas2r/esas2r_io.c 			list_add(&sgl->next_desc, &rq->sg_table_head);
sgl               376 drivers/scsi/esas2r/esas2r_io.c 	struct esas2r_mem_desc *sgl;
sgl               449 drivers/scsi/esas2r/esas2r_io.c 			sgl = esas2r_alloc_sgl(a);
sgl               451 drivers/scsi/esas2r/esas2r_io.c 			if (unlikely(sgl == NULL))
sgl               458 drivers/scsi/esas2r/esas2r_io.c 			list_add(&sgl->next_desc, &rq->sg_table_head);
sgl               471 drivers/scsi/esas2r/esas2r_io.c 				cpu_to_le64(sgl->phys_addr);
sgl               479 drivers/scsi/esas2r/esas2r_io.c 				(struct atto_physical_region_description *)sgl
sgl              1063 drivers/scsi/fcoe/fcoe.c 			  struct scatterlist *sgl, unsigned int sgc)
sgl              1069 drivers/scsi/fcoe/fcoe.c 							      xid, sgl,
sgl              1085 drivers/scsi/fcoe/fcoe.c 			   struct scatterlist *sgl, unsigned int sgc)
sgl              1091 drivers/scsi/fcoe/fcoe.c 							       sgl, sgc);
sgl               582 drivers/scsi/isci/request.c 	stp_req->sgl.offset = 0;
sgl               583 drivers/scsi/isci/request.c 	stp_req->sgl.set = SCU_SGL_ELEMENT_PAIR_A;
sgl               587 drivers/scsi/isci/request.c 		stp_req->sgl.index = 0;
sgl               590 drivers/scsi/isci/request.c 		stp_req->sgl.index = -1;
sgl              1261 drivers/scsi/isci/request.c 	struct scu_sgl_element *sgl;
sgl              1264 drivers/scsi/isci/request.c 	struct isci_stp_pio_sgl *pio_sgl = &stp_req->sgl;
sgl              1268 drivers/scsi/isci/request.c 		sgl = NULL;
sgl              1272 drivers/scsi/isci/request.c 			sgl = NULL;
sgl              1275 drivers/scsi/isci/request.c 			sgl = &sgl_pair->B;
sgl              1280 drivers/scsi/isci/request.c 			sgl = NULL;
sgl              1285 drivers/scsi/isci/request.c 			sgl = &sgl_pair->A;
sgl              1289 drivers/scsi/isci/request.c 	return sgl;
sgl              1334 drivers/scsi/isci/request.c 	sgl_pair = to_sgl_element_pair(ireq, stp_req->sgl.index);
sgl              1335 drivers/scsi/isci/request.c 	if (stp_req->sgl.set == SCU_SGL_ELEMENT_PAIR_A)
sgl              1355 drivers/scsi/isci/request.c 	struct scu_sgl_element *sgl;
sgl              1359 drivers/scsi/isci/request.c 	offset = stp_req->sgl.offset;
sgl              1360 drivers/scsi/isci/request.c 	sgl_pair = to_sgl_element_pair(ireq, stp_req->sgl.index);
sgl              1364 drivers/scsi/isci/request.c 	if (stp_req->sgl.set == SCU_SGL_ELEMENT_PAIR_A) {
sgl              1365 drivers/scsi/isci/request.c 		sgl = &sgl_pair->A;
sgl              1368 drivers/scsi/isci/request.c 		sgl = &sgl_pair->B;
sgl              1382 drivers/scsi/isci/request.c 		sgl = pio_sgl_next(stp_req);
sgl              1389 drivers/scsi/isci/request.c 		sgl->address_lower += stp_req->pio_len;
sgl              1393 drivers/scsi/isci/request.c 	stp_req->sgl.offset = offset;
sgl              2033 drivers/scsi/isci/request.c 		if (stp_req->sgl.index < 0) {
sgl                79 drivers/scsi/isci/request.h 	} sgl;
sgl               522 drivers/scsi/iscsi_tcp.c 		err = iscsi_sw_tcp_send_data_prep(conn, sdb->table.sgl,
sgl               713 drivers/scsi/libiscsi_tcp.c 						   sdb->table.sgl,
sgl              3711 drivers/scsi/lpfc/lpfc_init.c 			sglq_entry->sgl = sglq_entry->virt;
sgl              3712 drivers/scsi/lpfc/lpfc_init.c 			memset(sglq_entry->sgl, 0, LPFC_BPL_SIZE);
sgl              3832 drivers/scsi/lpfc/lpfc_init.c 			sglq_entry->sgl = sglq_entry->virt;
sgl              3833 drivers/scsi/lpfc/lpfc_init.c 			memset(sglq_entry->sgl, 0,
sgl               772 drivers/scsi/lpfc/lpfc_nvme.c 	struct sli4_sge *sgl;
sgl               791 drivers/scsi/lpfc/lpfc_nvme.c 	sgl = lpfc_ncmd->dma_sgl;
sgl               792 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->sge_len = cpu_to_le32(nCmd->cmdlen);
sgl               794 drivers/scsi/lpfc/lpfc_nvme.c 		sgl->addr_hi = 0;
sgl               795 drivers/scsi/lpfc/lpfc_nvme.c 		sgl->addr_lo = 0;
sgl               834 drivers/scsi/lpfc/lpfc_nvme.c 		sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->cmddma));
sgl               835 drivers/scsi/lpfc/lpfc_nvme.c 		sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->cmddma));
sgl               840 drivers/scsi/lpfc/lpfc_nvme.c 		wqe->generic.bde.addrHigh = sgl->addr_hi;
sgl               841 drivers/scsi/lpfc/lpfc_nvme.c 		wqe->generic.bde.addrLow =  sgl->addr_lo;
sgl               848 drivers/scsi/lpfc/lpfc_nvme.c 	sgl++;
sgl               851 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->rspdma));
sgl               852 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->rspdma));
sgl               853 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->word2 = le32_to_cpu(sgl->word2);
sgl               855 drivers/scsi/lpfc/lpfc_nvme.c 		bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl               857 drivers/scsi/lpfc/lpfc_nvme.c 		bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl               858 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->word2 = cpu_to_le32(sgl->word2);
sgl               859 drivers/scsi/lpfc/lpfc_nvme.c 	sgl->sge_len = cpu_to_le32(nCmd->rsplen);
sgl              1313 drivers/scsi/lpfc/lpfc_nvme.c 	struct sli4_sge *sgl = lpfc_ncmd->dma_sgl;
sgl              1337 drivers/scsi/lpfc/lpfc_nvme.c 		sgl += 2;
sgl              1339 drivers/scsi/lpfc/lpfc_nvme.c 		first_data_sgl = sgl;
sgl              1372 drivers/scsi/lpfc/lpfc_nvme.c 			sgl->word2 = 0;
sgl              1374 drivers/scsi/lpfc/lpfc_nvme.c 				bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              1375 drivers/scsi/lpfc/lpfc_nvme.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              1378 drivers/scsi/lpfc/lpfc_nvme.c 				bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              1385 drivers/scsi/lpfc/lpfc_nvme.c 					bf_set(lpfc_sli4_sge_type, sgl,
sgl              1395 drivers/scsi/lpfc/lpfc_nvme.c 					sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              1397 drivers/scsi/lpfc/lpfc_nvme.c 					sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              1401 drivers/scsi/lpfc/lpfc_nvme.c 					bf_set(lpfc_sli4_sge_type, sgl,
sgl              1406 drivers/scsi/lpfc/lpfc_nvme.c 			if (!(bf_get(lpfc_sli4_sge_type, sgl) &
sgl              1409 drivers/scsi/lpfc/lpfc_nvme.c 					bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              1413 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->addr_lo = cpu_to_le32(
sgl              1415 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->addr_hi = cpu_to_le32(
sgl              1418 drivers/scsi/lpfc/lpfc_nvme.c 				bf_set(lpfc_sli4_sge_offset, sgl, dma_offset);
sgl              1419 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->word2 = cpu_to_le32(sgl->word2);
sgl              1420 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->sge_len = cpu_to_le32(dma_len);
sgl              1425 drivers/scsi/lpfc/lpfc_nvme.c 				sgl++;
sgl              1429 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->word2 = cpu_to_le32(sgl->word2);
sgl              1431 drivers/scsi/lpfc/lpfc_nvme.c 				sgl->sge_len = cpu_to_le32(
sgl              1434 drivers/scsi/lpfc/lpfc_nvme.c 				sgl = (struct sli4_sge *)sgl_xtra->dma_sgl;
sgl              2022 drivers/scsi/lpfc/lpfc_nvme.c 	struct sli4_sge *sgl;
sgl              2044 drivers/scsi/lpfc/lpfc_nvme.c 		sgl = lpfc_ncmd->dma_sgl;
sgl              2045 drivers/scsi/lpfc/lpfc_nvme.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
sgl              2046 drivers/scsi/lpfc/lpfc_nvme.c 		bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              2047 drivers/scsi/lpfc/lpfc_nvme.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2592 drivers/scsi/lpfc/lpfc_nvmet.c 	struct sli4_sge *sgl;
sgl              2662 drivers/scsi/lpfc/lpfc_nvmet.c 	sgl  = (struct sli4_sge *)ctxp->ctxbuf->sglq->sgl;
sgl              2715 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_hi = 0;
sgl              2716 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_lo = 0;
sgl              2717 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2718 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
sgl              2719 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2720 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->sge_len = 0;
sgl              2721 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl++;
sgl              2722 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_hi = 0;
sgl              2723 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_lo = 0;
sgl              2724 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2725 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
sgl              2726 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2727 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->sge_len = 0;
sgl              2728 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl++;
sgl              2816 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_hi = putPaddrHigh(physaddr);
sgl              2817 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_lo = putPaddrLow(physaddr);
sgl              2818 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2819 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA);
sgl              2820 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2821 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->sge_len = cpu_to_le32(TXRDY_PAYLOAD_LEN);
sgl              2822 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl++;
sgl              2823 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_hi = 0;
sgl              2824 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_lo = 0;
sgl              2825 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2826 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
sgl              2827 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2828 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->sge_len = 0;
sgl              2829 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl++;
sgl              2887 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2906 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_hi = putPaddrHigh(physaddr);
sgl              2907 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->addr_lo = putPaddrLow(physaddr);
sgl              2908 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = 0;
sgl              2909 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA);
sgl              2910 drivers/scsi/lpfc/lpfc_nvmet.c 		bf_set(lpfc_sli4_sge_offset, sgl, ctxp->offset);
sgl              2912 drivers/scsi/lpfc/lpfc_nvmet.c 			bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              2913 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2914 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl->sge_len = cpu_to_le32(cnt);
sgl              2919 drivers/scsi/lpfc/lpfc_nvmet.c 				bde->addrLow = sgl->addr_lo;
sgl              2920 drivers/scsi/lpfc/lpfc_nvmet.c 				bde->addrHigh = sgl->addr_hi;
sgl              2922 drivers/scsi/lpfc/lpfc_nvmet.c 					le32_to_cpu(sgl->sge_len);
sgl              2929 drivers/scsi/lpfc/lpfc_nvmet.c 		sgl++;
sgl               126 drivers/scsi/lpfc/lpfc_scsi.c 	struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl               127 drivers/scsi/lpfc/lpfc_scsi.c 	if (sgl) {
sgl               128 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl               129 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = le32_to_cpu(sgl->word2);
sgl               130 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl               131 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl               635 drivers/scsi/lpfc/lpfc_scsi.c 	struct sli4_sge *sgl;
sgl               687 drivers/scsi/lpfc/lpfc_scsi.c 	sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl               689 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_cmd));
sgl               690 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_cmd));
sgl               691 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->word2 = le32_to_cpu(sgl->word2);
sgl               692 drivers/scsi/lpfc/lpfc_scsi.c 	bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl               693 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->word2 = cpu_to_le32(sgl->word2);
sgl               694 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->sge_len = cpu_to_le32(sizeof(struct fcp_cmnd));
sgl               695 drivers/scsi/lpfc/lpfc_scsi.c 	sgl++;
sgl               699 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_rsp));
sgl               700 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_rsp));
sgl               701 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->word2 = le32_to_cpu(sgl->word2);
sgl               702 drivers/scsi/lpfc/lpfc_scsi.c 	bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl               703 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->word2 = cpu_to_le32(sgl->word2);
sgl               704 drivers/scsi/lpfc/lpfc_scsi.c 	sgl->sge_len = cpu_to_le32(sizeof(struct fcp_rsp));
sgl              1947 drivers/scsi/lpfc/lpfc_scsi.c 		struct sli4_sge *sgl, int datasegcnt,
sgl              1984 drivers/scsi/lpfc/lpfc_scsi.c 	diseed = (struct sli4_sge_diseed *) sgl;
sgl              1986 drivers/scsi/lpfc/lpfc_scsi.c 	bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DISEED);
sgl              2021 drivers/scsi/lpfc/lpfc_scsi.c 	sgl++;
sgl              2028 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = 0;
sgl              2034 drivers/scsi/lpfc/lpfc_scsi.c 			bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_LSP);
sgl              2042 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              2044 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              2048 drivers/scsi/lpfc/lpfc_scsi.c 			bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA);
sgl              2051 drivers/scsi/lpfc/lpfc_scsi.c 		if (!(bf_get(lpfc_sli4_sge_type, sgl) & LPFC_SGE_TYPE_LSP)) {
sgl              2053 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              2056 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->addr_lo = cpu_to_le32(putPaddrLow(physaddr));
sgl              2057 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->addr_hi = cpu_to_le32(putPaddrHigh(physaddr));
sgl              2059 drivers/scsi/lpfc/lpfc_scsi.c 			bf_set(lpfc_sli4_sge_offset, sgl, dma_offset);
sgl              2060 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2061 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->sge_len = cpu_to_le32(dma_len);
sgl              2066 drivers/scsi/lpfc/lpfc_scsi.c 			sgl++;
sgl              2071 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2072 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->sge_len = cpu_to_le32(phba->cfg_sg_dma_buf_size);
sgl              2074 drivers/scsi/lpfc/lpfc_scsi.c 			sgl = (struct sli4_sge *)sgl_xtra->dma_sgl;
sgl              2127 drivers/scsi/lpfc/lpfc_scsi.c 		struct sli4_sge *sgl, int datacnt, int protcnt,
sgl              2192 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->word2 = 0;
sgl              2195 drivers/scsi/lpfc/lpfc_scsi.c 			bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_LSP);
sgl              2202 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              2204 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              2208 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2209 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->sge_len = cpu_to_le32(phba->cfg_sg_dma_buf_size);
sgl              2211 drivers/scsi/lpfc/lpfc_scsi.c 			sgl = (struct sli4_sge *)sgl_xtra->dma_sgl;
sgl              2216 drivers/scsi/lpfc/lpfc_scsi.c 		diseed = (struct sli4_sge_diseed *) sgl;
sgl              2218 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DISEED);
sgl              2262 drivers/scsi/lpfc/lpfc_scsi.c 		sgl++;
sgl              2273 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = 0;
sgl              2274 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DIF);
sgl              2275 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->addr_hi = le32_to_cpu(putPaddrHigh(protphysaddr));
sgl              2276 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->addr_lo = le32_to_cpu(putPaddrLow(protphysaddr));
sgl              2277 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2278 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->sge_len = 0;
sgl              2284 drivers/scsi/lpfc/lpfc_scsi.c 		if ((sgl->addr_lo & 0xfff) + protgroup_len > 0x1000) {
sgl              2285 drivers/scsi/lpfc/lpfc_scsi.c 			protgroup_remainder = 0x1000 - (sgl->addr_lo & 0xfff);
sgl              2300 drivers/scsi/lpfc/lpfc_scsi.c 		sgl++;
sgl              2317 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->word2 = 0;
sgl              2320 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              2329 drivers/scsi/lpfc/lpfc_scsi.c 					sgl->addr_lo = cpu_to_le32(
sgl              2331 drivers/scsi/lpfc/lpfc_scsi.c 					sgl->addr_hi = cpu_to_le32(
sgl              2335 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->word2 = cpu_to_le32(sgl->word2);
sgl              2336 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->sge_len = cpu_to_le32(
sgl              2339 drivers/scsi/lpfc/lpfc_scsi.c 				sgl = (struct sli4_sge *)sgl_xtra->dma_sgl;
sgl              2364 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->word2 = 0;
sgl              2365 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              2367 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              2369 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              2370 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_offset, sgl, dma_offset);
sgl              2371 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              2374 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->sge_len = cpu_to_le32(dma_len);
sgl              2381 drivers/scsi/lpfc/lpfc_scsi.c 					sgl++;
sgl              2389 drivers/scsi/lpfc/lpfc_scsi.c 				sgl++;
sgl              2404 drivers/scsi/lpfc/lpfc_scsi.c 			sgl--;
sgl              2405 drivers/scsi/lpfc/lpfc_scsi.c 			bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              3049 drivers/scsi/lpfc/lpfc_scsi.c 	struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl              3078 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl              3080 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = le32_to_cpu(sgl->word2);
sgl              3081 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              3082 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              3083 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl              3084 drivers/scsi/lpfc/lpfc_scsi.c 		first_data_sgl = sgl;
sgl              3113 drivers/scsi/lpfc/lpfc_scsi.c 			sgl->word2 = 0;
sgl              3115 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              3116 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              3119 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              3126 drivers/scsi/lpfc/lpfc_scsi.c 					bf_set(lpfc_sli4_sge_type, sgl,
sgl              3137 drivers/scsi/lpfc/lpfc_scsi.c 					sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              3139 drivers/scsi/lpfc/lpfc_scsi.c 					sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              3143 drivers/scsi/lpfc/lpfc_scsi.c 					bf_set(lpfc_sli4_sge_type, sgl,
sgl              3148 drivers/scsi/lpfc/lpfc_scsi.c 			if (!(bf_get(lpfc_sli4_sge_type, sgl) &
sgl              3151 drivers/scsi/lpfc/lpfc_scsi.c 					bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              3155 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_lo = cpu_to_le32(putPaddrLow(
sgl              3157 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->addr_hi = cpu_to_le32(putPaddrHigh(
sgl              3160 drivers/scsi/lpfc/lpfc_scsi.c 				bf_set(lpfc_sli4_sge_offset, sgl, dma_offset);
sgl              3161 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->word2 = cpu_to_le32(sgl->word2);
sgl              3162 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->sge_len = cpu_to_le32(dma_len);
sgl              3167 drivers/scsi/lpfc/lpfc_scsi.c 				sgl++;
sgl              3171 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->word2 = cpu_to_le32(sgl->word2);
sgl              3172 drivers/scsi/lpfc/lpfc_scsi.c 				sgl->sge_len = cpu_to_le32(
sgl              3175 drivers/scsi/lpfc/lpfc_scsi.c 				sgl = (struct sli4_sge *)sgl_xtra->dma_sgl;
sgl              3200 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl              3202 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = le32_to_cpu(sgl->word2);
sgl              3203 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              3204 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              3261 drivers/scsi/lpfc/lpfc_scsi.c 	struct sli4_sge *sgl = (struct sli4_sge *)(lpfc_cmd->dma_sgl);
sgl              3287 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl              3289 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = le32_to_cpu(sgl->word2);
sgl              3290 drivers/scsi/lpfc/lpfc_scsi.c 		bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              3291 drivers/scsi/lpfc/lpfc_scsi.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              3293 drivers/scsi/lpfc/lpfc_scsi.c 		sgl += 1;
sgl              3316 drivers/scsi/lpfc/lpfc_scsi.c 			num_sge = lpfc_bg_setup_sgl(phba, scsi_cmnd, sgl,
sgl              3352 drivers/scsi/lpfc/lpfc_scsi.c 			num_sge = lpfc_bg_setup_sgl_prot(phba, scsi_cmnd, sgl,
sgl              9182 drivers/scsi/lpfc/lpfc_sli.c 	struct sli4_sge *sgl  = NULL;
sgl              9193 drivers/scsi/lpfc/lpfc_sli.c 	sgl  = (struct sli4_sge *)sglq->sgl;
sgl              9215 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_hi = bpl->addrHigh;
sgl              9216 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_lo = bpl->addrLow;
sgl              9218 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = le32_to_cpu(sgl->word2);
sgl              9220 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              9222 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              9227 drivers/scsi/lpfc/lpfc_sli.c 			sgl->sge_len = cpu_to_le32(bde.tus.f.bdeSize);
sgl              9239 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_offset, sgl, offset);
sgl              9240 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              9244 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              9246 drivers/scsi/lpfc/lpfc_sli.c 			sgl++;
sgl              9253 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_hi =
sgl              9255 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_lo =
sgl              9257 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = le32_to_cpu(sgl->word2);
sgl              9258 drivers/scsi/lpfc/lpfc_sli.c 			bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              9259 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              9260 drivers/scsi/lpfc/lpfc_sli.c 			sgl->sge_len =
sgl              9496 drivers/scsi/lpfc/lpfc_sli.c 			struct sli4_sge *sgl;
sgl              9503 drivers/scsi/lpfc/lpfc_sli.c 			sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl              9509 drivers/scsi/lpfc/lpfc_sli.c 			wqe->generic.bde.tus.f.bdeSize = sgl->sge_len;
sgl              9560 drivers/scsi/lpfc/lpfc_sli.c 			struct sli4_sge *sgl;
sgl              9567 drivers/scsi/lpfc/lpfc_sli.c 			sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl              9573 drivers/scsi/lpfc/lpfc_sli.c 			wqe->generic.bde.tus.f.bdeSize = sgl->sge_len;
sgl              9617 drivers/scsi/lpfc/lpfc_sli.c 			struct sli4_sge *sgl;
sgl              9624 drivers/scsi/lpfc/lpfc_sli.c 			sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl;
sgl              9630 drivers/scsi/lpfc/lpfc_sli.c 			wqe->generic.bde.tus.f.bdeSize = sgl->sge_len;
sgl              16700 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_mbx_post_uembed_sgl_page1 *sgl;
sgl              16739 drivers/scsi/lpfc/lpfc_sli.c 	sgl = (struct lpfc_mbx_post_uembed_sgl_page1 *)viraddr;
sgl              16740 drivers/scsi/lpfc/lpfc_sli.c 	sgl_pg_pairs = &sgl->sgl_pg_pairs;
sgl              16762 drivers/scsi/lpfc/lpfc_sli.c 	bf_set(lpfc_post_sgl_pages_xri, sgl, xritag_start);
sgl              16763 drivers/scsi/lpfc/lpfc_sli.c 	bf_set(lpfc_post_sgl_pages_xricnt, sgl, post_cnt);
sgl              16764 drivers/scsi/lpfc/lpfc_sli.c 	sgl->word0 = cpu_to_le32(sgl->word0);
sgl              16772 drivers/scsi/lpfc/lpfc_sli.c 	shdr = (union lpfc_sli4_cfg_shdr *) &sgl->cfg_shdr;
sgl              16803 drivers/scsi/lpfc/lpfc_sli.c 	struct lpfc_mbx_post_uembed_sgl_page1 *sgl;
sgl              16849 drivers/scsi/lpfc/lpfc_sli.c 	sgl = (struct lpfc_mbx_post_uembed_sgl_page1 *)viraddr;
sgl              16850 drivers/scsi/lpfc/lpfc_sli.c 	sgl_pg_pairs = &sgl->sgl_pg_pairs;
sgl              16874 drivers/scsi/lpfc/lpfc_sli.c 	bf_set(lpfc_post_sgl_pages_xri, sgl, xritag_start);
sgl              16875 drivers/scsi/lpfc/lpfc_sli.c 	bf_set(lpfc_post_sgl_pages_xricnt, sgl, pg_pairs);
sgl              16877 drivers/scsi/lpfc/lpfc_sli.c 	sgl->word0 = cpu_to_le32(sgl->word0);
sgl              16885 drivers/scsi/lpfc/lpfc_sli.c 	shdr = (union lpfc_sli4_cfg_shdr *)&sgl->cfg_shdr;
sgl              19711 drivers/scsi/lpfc/lpfc_sli.c 	struct sli4_sge *sgl  = NULL;
sgl              19723 drivers/scsi/lpfc/lpfc_sli.c 	sgl  = (struct sli4_sge *)sglq->sgl;
sgl              19747 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_hi = bpl->addrHigh;
sgl              19748 drivers/scsi/lpfc/lpfc_sli.c 			sgl->addr_lo = bpl->addrLow;
sgl              19750 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = le32_to_cpu(sgl->word2);
sgl              19752 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              19754 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_last, sgl, 0);
sgl              19759 drivers/scsi/lpfc/lpfc_sli.c 			sgl->sge_len = cpu_to_le32(bde.tus.f.bdeSize);
sgl              19772 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_offset, sgl, offset);
sgl              19773 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              19778 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_offset, sgl, 0);
sgl              19779 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              19784 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_type, sgl,
sgl              19790 drivers/scsi/lpfc/lpfc_sli.c 				bf_set(lpfc_sli4_sge_offset, sgl, offset);
sgl              19793 drivers/scsi/lpfc/lpfc_sli.c 			sgl->word2 = cpu_to_le32(sgl->word2);
sgl              19795 drivers/scsi/lpfc/lpfc_sli.c 			sgl++;
sgl              19802 drivers/scsi/lpfc/lpfc_sli.c 		sgl->addr_hi = cpu_to_le32(wqe->gen_req.bde.addrHigh);
sgl              19803 drivers/scsi/lpfc/lpfc_sli.c 		sgl->addr_lo = cpu_to_le32(wqe->gen_req.bde.addrLow);
sgl              19804 drivers/scsi/lpfc/lpfc_sli.c 		sgl->word2 = le32_to_cpu(sgl->word2);
sgl              19805 drivers/scsi/lpfc/lpfc_sli.c 		bf_set(lpfc_sli4_sge_last, sgl, 1);
sgl              19806 drivers/scsi/lpfc/lpfc_sli.c 		sgl->word2 = cpu_to_le32(sgl->word2);
sgl              19807 drivers/scsi/lpfc/lpfc_sli.c 		sgl->sge_len = cpu_to_le32(wqe->gen_req.bde.tus.f.bdeSize);
sgl               960 drivers/scsi/lpfc/lpfc_sli4.h 	struct sli4_sge *sgl;	/* pre-assigned SGL */
sgl              1415 drivers/scsi/megaraid.c 	struct scatterlist	*sgl;
sgl              1549 drivers/scsi/megaraid.c 			sgl = scsi_sglist(cmd);
sgl              1550 drivers/scsi/megaraid.c 			if( sg_page(sgl) ) {
sgl              1551 drivers/scsi/megaraid.c 				c = *(unsigned char *) sg_virt(&sgl[0]);
sgl              1753 drivers/scsi/megaraid.c 			scb->sgl[idx].address = sg_dma_address(sg);
sgl              1754 drivers/scsi/megaraid.c 			*len += scb->sgl[idx].length = sg_dma_len(sg);
sgl              2883 drivers/scsi/megaraid.c 		scb->sgl = NULL;
sgl              2898 drivers/scsi/megaraid.c 		scb->sgl = (mega_sglist *)scb->sgl64;
sgl              2900 drivers/scsi/megaraid.c 		if( !scb->sgl ) {
sgl               198 drivers/scsi/megaraid.h 	mega_sglist	*sgl;
sgl              1337 drivers/scsi/megaraid/megaraid_mbox.c 	struct scatterlist	*sgl;
sgl              1356 drivers/scsi/megaraid/megaraid_mbox.c 	scsi_for_each_sg(scp, sgl, sgcnt, i) {
sgl              1357 drivers/scsi/megaraid/megaraid_mbox.c 		ccb->sgl64[i].address	= sg_dma_address(sgl);
sgl              1358 drivers/scsi/megaraid/megaraid_mbox.c 		ccb->sgl64[i].length	= sg_dma_len(sgl);
sgl              1540 drivers/scsi/megaraid/megaraid_mbox.c 			struct scatterlist	*sgl;
sgl              1543 drivers/scsi/megaraid/megaraid_mbox.c 			sgl = scsi_sglist(scp);
sgl              1544 drivers/scsi/megaraid/megaraid_mbox.c 			if (sg_page(sgl)) {
sgl              1545 drivers/scsi/megaraid/megaraid_mbox.c 				vaddr = (caddr_t) sg_virt(&sgl[0]);
sgl              2176 drivers/scsi/megaraid/megaraid_mbox.c 	struct scatterlist	*sgl;
sgl              2260 drivers/scsi/megaraid/megaraid_mbox.c 			sgl = scsi_sglist(scp);
sgl              2261 drivers/scsi/megaraid/megaraid_mbox.c 			if (sg_page(sgl)) {
sgl              2262 drivers/scsi/megaraid/megaraid_mbox.c 				c = *(unsigned char *) sg_virt(&sgl[0]);
sgl              1863 drivers/scsi/megaraid/megaraid_sas.h 	union megasas_sgl sgl;	/*28h */
sgl              1890 drivers/scsi/megaraid/megaraid_sas.h 	union megasas_sgl sgl;	/*30h */
sgl              1917 drivers/scsi/megaraid/megaraid_sas.h 	union megasas_sgl sgl;	/*28h */
sgl              1969 drivers/scsi/megaraid/megaraid_sas.h 	} sgl;
sgl              1998 drivers/scsi/megaraid/megaraid_sas.h 	} sgl;
sgl              2594 drivers/scsi/megaraid/megaraid_sas.h 	struct iovec sgl[MAX_IOCTL_SGE];
sgl              2617 drivers/scsi/megaraid/megaraid_sas.h 	struct compat_iovec sgl[MAX_IOCTL_SGE];
sgl               279 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge64[0].phys_addr = cpu_to_le64(dma_addr);
sgl               280 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge64[0].length = cpu_to_le32(dma_len);
sgl               284 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].phys_addr =
sgl               286 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].length = cpu_to_le32(dma_len);
sgl              1438 drivers/scsi/megaraid/megaraid_sas_base.c 						      &pthru->sgl);
sgl              1442 drivers/scsi/megaraid/megaraid_sas_base.c 						      &pthru->sgl);
sgl              1445 drivers/scsi/megaraid/megaraid_sas_base.c 						      &pthru->sgl);
sgl              1581 drivers/scsi/megaraid/megaraid_sas_base.c 					      &ldio->sgl);
sgl              1584 drivers/scsi/megaraid/megaraid_sas_base.c 		ldio->sge_count = megasas_make_sgl64(instance, scp, &ldio->sgl);
sgl              1586 drivers/scsi/megaraid/megaraid_sas_base.c 		ldio->sge_count = megasas_make_sgl32(instance, scp, &ldio->sgl);
sgl              1671 drivers/scsi/megaraid/megaraid_sas_base.c 			mfi_sgl = &ldio->sgl;
sgl              1680 drivers/scsi/megaraid/megaraid_sas_base.c 			mfi_sgl = &pthru->sgl;
sgl              2348 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].phys_addr =
sgl              2351 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].phys_addr =
sgl              2354 drivers/scsi/megaraid/megaraid_sas_base.c 	dcmd->sgl.sge32[0].length = cpu_to_le32(
sgl              2455 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].phys_addr =
sgl              2458 drivers/scsi/megaraid/megaraid_sas_base.c 		dcmd->sgl.sge32[0].phys_addr =
sgl              2461 drivers/scsi/megaraid/megaraid_sas_base.c 	dcmd->sgl.sge32[0].length = cpu_to_le32((MAX_LOGICAL_DRIVES + 1) *
sgl              8119 drivers/scsi/megaraid/megaraid_sas_base.c 		if (!ioc->sgl[i].iov_len)
sgl              8123 drivers/scsi/megaraid/megaraid_sas_base.c 						    ioc->sgl[i].iov_len,
sgl              8138 drivers/scsi/megaraid/megaraid_sas_base.c 			kern_sge64[i].length = cpu_to_le32(ioc->sgl[i].iov_len);
sgl              8141 drivers/scsi/megaraid/megaraid_sas_base.c 			kern_sge32[i].length = cpu_to_le32(ioc->sgl[i].iov_len);
sgl              8148 drivers/scsi/megaraid/megaraid_sas_base.c 		if (copy_from_user(kbuff_arr[i], ioc->sgl[i].iov_base,
sgl              8149 drivers/scsi/megaraid/megaraid_sas_base.c 				   (u32) (ioc->sgl[i].iov_len))) {
sgl              8196 drivers/scsi/megaraid/megaraid_sas_base.c 		if (copy_to_user(ioc->sgl[i].iov_base, kbuff_arr[i],
sgl              8197 drivers/scsi/megaraid/megaraid_sas_base.c 				 ioc->sgl[i].iov_len)) {
sgl              8417 drivers/scsi/megaraid/megaraid_sas_base.c 		if (get_user(ptr, &cioc->sgl[i].iov_base) ||
sgl              8418 drivers/scsi/megaraid/megaraid_sas_base.c 		    put_user(compat_ptr(ptr), &ioc->sgl[i].iov_base) ||
sgl              8419 drivers/scsi/megaraid/megaraid_sas_base.c 		    copy_in_user(&ioc->sgl[i].iov_len,
sgl              8420 drivers/scsi/megaraid/megaraid_sas_base.c 				 &cioc->sgl[i].iov_len, sizeof(compat_size_t)))
sgl              1274 drivers/scsi/myrb.c 	struct scatterlist *sgl;
sgl              1320 drivers/scsi/myrb.c 	sgl = scsi_sglist(scmd);
sgl              1321 drivers/scsi/myrb.c 	dcdb->dma_addr = sg_dma_address(sgl);
sgl              1322 drivers/scsi/myrb.c 	if (sg_dma_len(sgl) > USHRT_MAX) {
sgl              1323 drivers/scsi/myrb.c 		dcdb->xfer_len_lo = sg_dma_len(sgl) & 0xffff;
sgl              1324 drivers/scsi/myrb.c 		dcdb->xfer_len_hi4 = sg_dma_len(sgl) >> 16;
sgl              1326 drivers/scsi/myrb.c 		dcdb->xfer_len_lo = sg_dma_len(sgl);
sgl              1429 drivers/scsi/myrb.c 	struct scatterlist *sgl;
sgl              1587 drivers/scsi/myrb.c 		sgl = scsi_sglist(scmd);
sgl              1596 drivers/scsi/myrb.c 		mbox->type5.addr = (u32)sg_dma_address(sgl);
sgl              1606 drivers/scsi/myrb.c 		cmd_blk->sgl = hw_sgl;
sgl              1620 drivers/scsi/myrb.c 		scsi_for_each_sg(scmd, sgl, nsge, i) {
sgl              1621 drivers/scsi/myrb.c 			hw_sgl->sge_addr = (u32)sg_dma_address(sgl);
sgl              1622 drivers/scsi/myrb.c 			hw_sgl->sge_count = (u32)sg_dma_len(sgl);
sgl              2341 drivers/scsi/myrb.c 	if (cmd_blk->sgl) {
sgl              2342 drivers/scsi/myrb.c 		dma_pool_free(cb->sg_pool, cmd_blk->sgl, cmd_blk->sgl_addr);
sgl              2343 drivers/scsi/myrb.c 		cmd_blk->sgl = NULL;
sgl               688 drivers/scsi/myrb.h 	struct myrb_sge *sgl;
sgl               164 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               187 drivers/scsi/myrs.c 	sgl = &mbox->ctlr_info.dma_addr;
sgl               188 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = ctlr_info_addr;
sgl               189 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
sgl               227 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               246 drivers/scsi/myrs.c 	sgl = &mbox->ldev_info.dma_addr;
sgl               247 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = ldev_info_addr;
sgl               248 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ldev_info.dma_size;
sgl               314 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               334 drivers/scsi/myrs.c 	sgl = &mbox->pdev_info.dma_addr;
sgl               335 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = pdev_info_addr;
sgl               336 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
sgl               384 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               405 drivers/scsi/myrs.c 	sgl = &mbox->pdev_info.dma_addr;
sgl               406 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = devmap_addr;
sgl               407 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
sgl               427 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               441 drivers/scsi/myrs.c 	sgl = &mbox->get_event.dma_addr;
sgl               442 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = event_addr;
sgl               443 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->get_event.dma_size;
sgl               459 drivers/scsi/myrs.c 	union myrs_sgl *sgl;
sgl               469 drivers/scsi/myrs.c 	sgl = &mbox->common.dma_addr;
sgl               470 drivers/scsi/myrs.c 	sgl->sge[0].sge_addr = cs->fwstat_addr;
sgl               471 drivers/scsi/myrs.c 	sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
sgl              1593 drivers/scsi/myrs.c 	struct scatterlist *sgl;
sgl              1724 drivers/scsi/myrs.c 		sgl = scsi_sglist(scmd);
sgl              1725 drivers/scsi/myrs.c 		hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl);
sgl              1726 drivers/scsi/myrs.c 		hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl);
sgl              1750 drivers/scsi/myrs.c 			cmd_blk->sgl = hw_sgl;
sgl              1761 drivers/scsi/myrs.c 		scsi_for_each_sg(scmd, sgl, nsge, i) {
sgl              1768 drivers/scsi/myrs.c 			hw_sgl->sge_addr = (u64)sg_dma_address(sgl);
sgl              1769 drivers/scsi/myrs.c 			hw_sgl->sge_count = (u64)sg_dma_len(sgl);
sgl              2080 drivers/scsi/myrs.c 	if (cmd_blk->sgl) {
sgl              2081 drivers/scsi/myrs.c 		dma_pool_free(cs->sg_pool, cmd_blk->sgl,
sgl              2083 drivers/scsi/myrs.c 		cmd_blk->sgl = NULL;
sgl               876 drivers/scsi/myrs.h 	struct myrs_sge *sgl;
sgl              4936 drivers/scsi/pm8001/pm8001_hwi.c 	payload.len = info->sgl.im_len.len ;
sgl              4938 drivers/scsi/pm8001/pm8001_hwi.c 		cpu_to_le32(lower_32_bits(le64_to_cpu(info->sgl.addr)));
sgl              4940 drivers/scsi/pm8001/pm8001_hwi.c 		cpu_to_le32(upper_32_bits(le64_to_cpu(info->sgl.addr)));
sgl              4964 drivers/scsi/pm8001/pm8001_hwi.c 	flash_update_info.sgl.addr = cpu_to_le64(phys_addr);
sgl              4965 drivers/scsi/pm8001/pm8001_hwi.c 	flash_update_info.sgl.im_len.len = cpu_to_le32(fw_control->len);
sgl              4966 drivers/scsi/pm8001/pm8001_hwi.c 	flash_update_info.sgl.im_len.e = 0;
sgl               597 drivers/scsi/pm8001/pm8001_sas.h 	struct pm8001_prd	sgl;
sgl                34 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo);
sgl                36 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi);
sgl                38 drivers/scsi/qedf/drv_scsi_fw_funcs.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len);
sgl                12 drivers/scsi/qedf/drv_scsi_fw_funcs.h 	struct scsi_sge *sgl;
sgl               644 drivers/scsi/qedf/qedf_io.c 		io_req->sgl_task_params->sgl = bd_tbl->bd_tbl;
sgl               737 drivers/scsi/qedf/qedf_io.c 	tx_sgl_task_params.sgl = mp_req->mp_req_bd;
sgl               746 drivers/scsi/qedf/qedf_io.c 	rx_sgl_task_params.sgl = mp_req->mp_resp_bd;
sgl              1062 drivers/scsi/qedi/qedi_fw.c 	tx_sgl_task_params.sgl =
sgl              1071 drivers/scsi/qedi/qedi_fw.c 	rx_sgl_task_params.sgl =
sgl              1647 drivers/scsi/qedi/qedi_fw.c 	tx_sgl_task_params.sgl =
sgl              1656 drivers/scsi/qedi/qedi_fw.c 	rx_sgl_task_params.sgl =
sgl              1763 drivers/scsi/qedi/qedi_fw.c 		tx_sgl_task_params.sgl =
sgl              1772 drivers/scsi/qedi/qedi_fw.c 		rx_sgl_task_params.sgl =
sgl              2110 drivers/scsi/qedi/qedi_fw.c 		tx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
sgl              2120 drivers/scsi/qedi/qedi_fw.c 		rx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
sgl                43 drivers/scsi/qedi/qedi_fw_api.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo);
sgl                45 drivers/scsi/qedi/qedi_fw_api.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi);
sgl                47 drivers/scsi/qedi/qedi_fw_api.c 		val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len);
sgl                16 drivers/scsi/qedi/qedi_fw_scsi.h 	struct scsi_sge	*sgl;
sgl              1008 drivers/scsi/qla2xxx/qla_iocb.c 	struct scatterlist *sg, *sgl;
sgl              1016 drivers/scsi/qla2xxx/qla_iocb.c 		sgl = scsi_sglist(cmd);
sgl              1018 drivers/scsi/qla2xxx/qla_iocb.c 		sgl = tc->sg;
sgl              1025 drivers/scsi/qla2xxx/qla_iocb.c 	for_each_sg(sgl, sg, tot_dsds, i) {
sgl              1085 drivers/scsi/qla2xxx/qla_iocb.c 	struct scatterlist *sg, *sgl;
sgl              1098 drivers/scsi/qla2xxx/qla_iocb.c 		sgl = scsi_prot_sglist(cmd);
sgl              1107 drivers/scsi/qla2xxx/qla_iocb.c 		sgl = tc->prot_sg;
sgl              1120 drivers/scsi/qla2xxx/qla_iocb.c 		for_each_sg(sgl, sg, tot_dsds, i) {
sgl              1162 drivers/scsi/qla2xxx/qla_iocb.c 		for_each_sg(sgl, sg, tot_dsds, i) {
sgl              1219 drivers/scsi/qla2xxx/qla_iocb.c 				sg_pcopy_to_buffer(sgl, sg_nents(sgl),
sgl              1321 drivers/scsi/qla2xxx/qla_iocb.c 		for_each_sg(sgl, sg, tot_dsds, i) {
sgl               368 drivers/scsi/qla2xxx/qla_nvme.c 	struct scatterlist *sgl, *sg;
sgl               470 drivers/scsi/qla2xxx/qla_nvme.c 	sgl = fd->first_sgl;
sgl               473 drivers/scsi/qla2xxx/qla_nvme.c 	for_each_sg(sgl, sg, tot_dsds, i) {
sgl              1013 drivers/scsi/scsi_debug.c 	act_len = sg_copy_from_buffer(sdb->table.sgl, sdb->table.nents,
sgl              1037 drivers/scsi/scsi_debug.c 	act_len = sg_pcopy_from_buffer(sdb->table.sgl, sdb->table.nents,
sgl              2503 drivers/scsi/scsi_debug.c 	ret = sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
sgl              2510 drivers/scsi/scsi_debug.c 		ret += sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
sgl               988 drivers/scsi/scsi_error.c 		scmd->sdb.table.sgl = &ses->sense_sgl;
sgl               994 drivers/scsi/scsi_lib.c 			blk_rq_nr_phys_segments(req), sdb->table.sgl,
sgl              1002 drivers/scsi/scsi_lib.c 	count = blk_rq_map_sg(req->q, req, sdb->table.sgl);
sgl              1049 drivers/scsi/scsi_lib.c 				prot_sdb->table.sgl,
sgl              1056 drivers/scsi/scsi_lib.c 						prot_sdb->table.sgl);
sgl              1593 drivers/scsi/scsi_lib.c 	cmd->sdb.table.sgl = sg;
sgl              1598 drivers/scsi/scsi_lib.c 		cmd->prot_sdb->table.sgl =
sgl              2856 drivers/scsi/scsi_lib.c void *scsi_kmap_atomic_sg(struct scatterlist *sgl, int sg_count,
sgl              2866 drivers/scsi/scsi_lib.c 	for_each_sg(sgl, sg, sg_count, i) {
sgl              1565 drivers/scsi/storvsc_drv.c 	struct scatterlist *sgl;
sgl              1638 drivers/scsi/storvsc_drv.c 	sgl = (struct scatterlist *)scsi_sglist(scmnd);
sgl              1656 drivers/scsi/storvsc_drv.c 		payload->range.offset = sgl[0].offset;
sgl              1658 drivers/scsi/storvsc_drv.c 		cur_sgl = sgl;
sgl               432 drivers/scsi/virtio_scsi.c 		sgs[out_num++] = out->sgl;
sgl               444 drivers/scsi/virtio_scsi.c 		sgs[out_num + in_num++] = in->sgl;
sgl                61 drivers/scsi/vmw_pvscsi.c 	struct pvscsi_sg_list	*sgl;
sgl               337 drivers/scsi/vmw_pvscsi.c 	sge = &ctx->sgl->sge[0];
sgl               376 drivers/scsi/vmw_pvscsi.c 					ctx->sgl, SGL_SIZE, DMA_TO_DEVICE);
sgl              1195 drivers/scsi/vmw_pvscsi.c 		free_pages((unsigned long)ctx->sgl, get_order(SGL_SIZE));
sgl              1261 drivers/scsi/vmw_pvscsi.c 		ctx->sgl = (void *)__get_free_pages(GFP_KERNEL,
sgl              1264 drivers/scsi/vmw_pvscsi.c 		BUG_ON(!IS_ALIGNED(((unsigned long)ctx->sgl), PAGE_SIZE));
sgl              1265 drivers/scsi/vmw_pvscsi.c 		if (!ctx->sgl) {
sgl              1267 drivers/scsi/vmw_pvscsi.c 				free_pages((unsigned long)ctx->sgl,
sgl              1269 drivers/scsi/vmw_pvscsi.c 				ctx->sgl = NULL;
sgl               230 drivers/spi/spi-at91-usart.c 					 xfer->rx_sg.sgl,
sgl               239 drivers/spi/spi-at91-usart.c 					 xfer->tx_sg.sgl,
sgl               766 drivers/spi/spi-atmel.c 						 xfer->rx_sg.sgl,
sgl               785 drivers/spi/spi-atmel.c 						 xfer->tx_sg.sgl,
sgl               484 drivers/spi/spi-bcm2835.c 	if (bs->tx_buf && !sg_is_last(&tfr->tx_sg.sgl[0]))
sgl               485 drivers/spi/spi-bcm2835.c 		bs->tx_prologue = sg_dma_len(&tfr->tx_sg.sgl[0]) & 3;
sgl               487 drivers/spi/spi-bcm2835.c 	if (bs->rx_buf && !sg_is_last(&tfr->rx_sg.sgl[0])) {
sgl               488 drivers/spi/spi-bcm2835.c 		bs->rx_prologue = sg_dma_len(&tfr->rx_sg.sgl[0]) & 3;
sgl               491 drivers/spi/spi-bcm2835.c 			if (!bs->tx_buf || sg_is_last(&tfr->tx_sg.sgl[0])) {
sgl               496 drivers/spi/spi-bcm2835.c 					!(sg_dma_len(&tfr->tx_sg.sgl[0]) & ~3);
sgl               518 drivers/spi/spi-bcm2835.c 					   sg_dma_address(&tfr->rx_sg.sgl[0]),
sgl               521 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->rx_sg.sgl[0]) += bs->rx_prologue;
sgl               522 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->rx_sg.sgl[0])     -= bs->rx_prologue;
sgl               544 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->tx_sg.sgl[0]) += bs->tx_prologue;
sgl               545 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[0])     -= bs->tx_prologue;
sgl               547 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[0])      = 0;
sgl               548 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->tx_sg.sgl[1]) += 4;
sgl               549 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[1])     -= 4;
sgl               569 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue;
sgl               570 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->rx_sg.sgl[0])     += bs->rx_prologue;
sgl               577 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->tx_sg.sgl[0]) -= bs->tx_prologue;
sgl               578 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[0])     += bs->tx_prologue;
sgl               580 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[0])      = bs->tx_prologue - 4;
sgl               581 drivers/spi/spi-bcm2835.c 		sg_dma_address(&tfr->tx_sg.sgl[1]) -= 4;
sgl               582 drivers/spi/spi-bcm2835.c 		sg_dma_len(&tfr->tx_sg.sgl[1])     += 4;
sgl               666 drivers/spi/spi-bcm2835.c 	struct scatterlist *sgl;
sgl               678 drivers/spi/spi-bcm2835.c 		sgl   = tfr->tx_sg.sgl;
sgl               684 drivers/spi/spi-bcm2835.c 		sgl   = tfr->rx_sg.sgl;
sgl               688 drivers/spi/spi-bcm2835.c 	desc = dmaengine_prep_slave_sg(chan, sgl, nents, dir, flags);
sgl               631 drivers/spi/spi-davinci.c 				t->rx_sg.sgl, t->rx_sg.nents, DMA_DEV_TO_MEM,
sgl               642 drivers/spi/spi-davinci.c 			t->tx_sg.sgl = t->rx_sg.sgl;
sgl               647 drivers/spi/spi-davinci.c 				t->tx_sg.sgl, t->tx_sg.nents, DMA_MEM_TO_DEV,
sgl               160 drivers/spi/spi-dw-mid.c 				xfer->tx_sg.sgl,
sgl               206 drivers/spi/spi-dw-mid.c 				xfer->rx_sg.sgl,
sgl               331 drivers/spi/spi-ep93xx.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgl               351 drivers/spi/spi-ep93xx.c 	nents = dma_map_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgl               355 drivers/spi/spi-ep93xx.c 	txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction,
sgl               358 drivers/spi/spi-ep93xx.c 		dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgl               387 drivers/spi/spi-ep93xx.c 	dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgl               580 drivers/spi/spi-fsl-lpspi.c 				rx->sgl, rx->nents, DMA_DEV_TO_MEM,
sgl               592 drivers/spi/spi-fsl-lpspi.c 				tx->sgl, tx->nents, DMA_MEM_TO_DEV,
sgl               335 drivers/spi/spi-img-spfi.c 		rxdesc = dmaengine_prep_slave_sg(spfi->rx_ch, xfer->rx_sg.sgl,
sgl               359 drivers/spi/spi-img-spfi.c 		txdesc = dmaengine_prep_slave_sg(spfi->tx_ch, xfer->tx_sg.sgl,
sgl              1341 drivers/spi/spi-imx.c 	struct scatterlist *last_sg = sg_last(rx->sgl, rx->nents);
sgl              1372 drivers/spi/spi-imx.c 				rx->sgl, rx->nents, DMA_DEV_TO_MEM,
sgl              1384 drivers/spi/spi-imx.c 				tx->sgl, tx->nents, DMA_MEM_TO_DEV,
sgl               467 drivers/spi/spi-mt65xx.c 		mdata->tx_sgl = xfer->tx_sg.sgl;
sgl               469 drivers/spi/spi-mt65xx.c 		mdata->rx_sgl = xfer->rx_sg.sgl;
sgl               409 drivers/spi/spi-omap2-mcspi.c 		tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl,
sgl               491 drivers/spi/spi-omap2-mcspi.c 		ret = sg_split(xfer->rx_sg.sgl, xfer->rx_sg.nents,
sgl               262 drivers/spi/spi-pic32-sqi.c 	struct scatterlist *sg, *sgl;
sgl               276 drivers/spi/spi-pic32-sqi.c 		sgl = xfer->rx_sg.sgl;
sgl               280 drivers/spi/spi-pic32-sqi.c 		sgl = xfer->tx_sg.sgl;
sgl               296 drivers/spi/spi-pic32-sqi.c 	for_each_sg(sgl, sg, nents, i) {
sgl               313 drivers/spi/spi-pic32.c 					  xfer->rx_sg.sgl,
sgl               323 drivers/spi/spi-pic32.c 					  xfer->tx_sg.sgl,
sgl               798 drivers/spi/spi-pl022.c 	dma_unmap_sg(pl022->dma_tx_channel->device->dev, pl022->sgt_tx.sgl,
sgl               800 drivers/spi/spi-pl022.c 	dma_unmap_sg(pl022->dma_rx_channel->device->dev, pl022->sgt_rx.sgl,
sgl               811 drivers/spi/spi-pl022.c 	BUG_ON(!pl022->sgt_rx.sgl);
sgl               825 drivers/spi/spi-pl022.c 				    pl022->sgt_rx.sgl,
sgl               829 drivers/spi/spi-pl022.c 		for_each_sg(pl022->sgt_rx.sgl, sg, pl022->sgt_rx.nents, i) {
sgl               839 drivers/spi/spi-pl022.c 		for_each_sg(pl022->sgt_tx.sgl, sg, pl022->sgt_tx.nents, i) {
sgl               875 drivers/spi/spi-pl022.c 		for_each_sg(sgtab->sgl, sg, sgtab->nents, i) {
sgl               896 drivers/spi/spi-pl022.c 		for_each_sg(sgtab->sgl, sg, sgtab->nents, i) {
sgl              1050 drivers/spi/spi-pl022.c 	rx_sglen = dma_map_sg(rxchan->device->dev, pl022->sgt_rx.sgl,
sgl              1055 drivers/spi/spi-pl022.c 	tx_sglen = dma_map_sg(txchan->device->dev, pl022->sgt_tx.sgl,
sgl              1062 drivers/spi/spi-pl022.c 				      pl022->sgt_rx.sgl,
sgl              1070 drivers/spi/spi-pl022.c 				      pl022->sgt_tx.sgl,
sgl              1094 drivers/spi/spi-pl022.c 	dma_unmap_sg(txchan->device->dev, pl022->sgt_tx.sgl,
sgl              1097 drivers/spi/spi-pl022.c 	dma_unmap_sg(rxchan->device->dev, pl022->sgt_rx.sgl,
sgl               118 drivers/spi/spi-pxa2xx-dma.c 	return dmaengine_prep_slave_sg(chan, sgt->sgl, sgt->nents, dir,
sgl               389 drivers/spi/spi-qup.c static int spi_qup_prep_sg(struct spi_master *master, struct scatterlist *sgl,
sgl               404 drivers/spi/spi-qup.c 	desc = dmaengine_prep_slave_sg(chan, sgl, nents, dir, flags);
sgl               425 drivers/spi/spi-qup.c static u32 spi_qup_sgl_get_nents_len(struct scatterlist *sgl, u32 max,
sgl               431 drivers/spi/spi-qup.c 	for (sg = sgl; sg; sg = sg_next(sg)) {
sgl               459 drivers/spi/spi-qup.c 	rx_sgl = xfer->rx_sg.sgl;
sgl               460 drivers/spi/spi-qup.c 	tx_sgl = xfer->tx_sg.sgl;
sgl               405 drivers/spi/spi-rockchip.c 				xfer->rx_sg.sgl, xfer->rx_sg.nents,
sgl               427 drivers/spi/spi-rockchip.c 				xfer->tx_sg.sgl, xfer->tx_sg.nents,
sgl               532 drivers/spi/spi-rspi.c 		desc_rx = dmaengine_prep_slave_sg(rspi->ctlr->dma_rx, rx->sgl,
sgl               552 drivers/spi/spi-rspi.c 		desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl,
sgl               294 drivers/spi/spi-s3c64xx.c 	desc = dmaengine_prep_slave_sg(dma->ch, sgt->sgl, sgt->nents,
sgl               504 drivers/spi/spi-sprd.c 	desc = dmaengine_prep_slave_sg(dma_chan, sg->sgl, sg->nents, dir, flags);
sgl               226 drivers/spi/spi-stm32-qspi.c 	desc = dmaengine_prep_slave_sg(dma_ch, sgt.sgl, sgt.nents,
sgl              1293 drivers/spi/spi-stm32.c 					spi->dma_rx, xfer->rx_sg.sgl,
sgl              1305 drivers/spi/spi-stm32.c 					spi->dma_tx, xfer->tx_sg.sgl,
sgl               468 drivers/spi/spi-ti-qspi.c 	for_each_sg(rx_sg.sgl, sg, rx_sg.nents, i) {
sgl               841 drivers/spi/spi.c 	sg = &sgt->sgl[0];
sgl               874 drivers/spi/spi.c 	ret = dma_map_sg(dev, sgt->sgl, sgt->nents, dir);
sgl               891 drivers/spi/spi.c 		dma_unmap_sg(dev, sgt->sgl, sgt->orig_nents, dir);
sgl               156 drivers/staging/android/ion/ion.c 	new_sg = new_table->sgl;
sgl               157 drivers/staging/android/ion/ion.c 	for_each_sg(table->sgl, sg, table->nents, i) {
sgl               230 drivers/staging/android/ion/ion.c 	if (!dma_map_sg(attachment->dev, table->sgl, table->nents,
sgl               241 drivers/staging/android/ion/ion.c 	dma_unmap_sg(attachment->dev, table->sgl, table->nents, direction);
sgl               312 drivers/staging/android/ion/ion.c 		dma_sync_sg_for_cpu(a->dev, a->table->sgl, a->table->nents,
sgl               335 drivers/staging/android/ion/ion.c 		dma_sync_sg_for_device(a->dev, a->table->sgl, a->table->nents,
sgl                70 drivers/staging/android/ion/ion_cma_heap.c 	sg_set_page(table->sgl, pages, size, 0);
sgl                41 drivers/staging/android/ion/ion_heap.c 	for_each_sg(table->sgl, sg, table->nents, i) {
sgl                74 drivers/staging/android/ion/ion_heap.c 	for_each_sg(table->sgl, sg, table->nents, i) {
sgl               112 drivers/staging/android/ion/ion_heap.c static int ion_heap_sglist_zero(struct scatterlist *sgl, unsigned int nents,
sgl               120 drivers/staging/android/ion/ion_heap.c 	for_each_sg_page(sgl, &piter, nents, 0) {
sgl               145 drivers/staging/android/ion/ion_heap.c 	return ion_heap_sglist_zero(table->sgl, table->nents, pgprot);
sgl               134 drivers/staging/android/ion/ion_system_heap.c 	sg = table->sgl;
sgl               165 drivers/staging/android/ion/ion_system_heap.c 	for_each_sg(table->sgl, sg, table->nents, i)
sgl               315 drivers/staging/android/ion/ion_system_heap.c 	sg_set_page(table->sgl, page, len, 0);
sgl               333 drivers/staging/android/ion/ion_system_heap.c 	struct page *page = sg_page(table->sgl);
sgl                95 drivers/staging/kpc2000/kpc_dma/fileops.c 	acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
sgl               102 drivers/staging/kpc2000/kpc_dma/fileops.c 	for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) {
sgl               125 drivers/staging/kpc2000/kpc_dma/fileops.c 	for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) {
sgl               189 drivers/staging/kpc2000/kpc_dma/fileops.c 	dma_unmap_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
sgl               210 drivers/staging/kpc2000/kpc_dma/fileops.c 	BUG_ON(acd->sgt.sgl == NULL);
sgl               220 drivers/staging/kpc2000/kpc_dma/fileops.c 	dma_unmap_sg(&acd->ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, acd->ldev->dir);
sgl               318 drivers/staging/media/ipu3/ipu3-v4l2.c 	return imgu_dmamap_map_sg(imgu, sg->sgl, sg->nents, &buf->map);
sgl                90 drivers/staging/media/tegra-vde/dmabuf-cache.c 			*addrp = sg_dma_address(entry->sgt->sgl);
sgl               128 drivers/staging/media/tegra-vde/dmabuf-cache.c 		*addrp = sg_dma_address(sgt->sgl);
sgl                39 drivers/staging/media/tegra-vde/iommu.c 	size = iommu_map_sg(vde->domain, addr, sgt->sgl, sgt->nents,
sgl               514 drivers/staging/ralink-gdma/ralink-gdma.c 		struct dma_chan *c, struct scatterlist *sgl,
sgl               530 drivers/staging/ralink-gdma/ralink-gdma.c 	for_each_sg(sgl, sg, sg_len, i) {
sgl               133 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	struct scatterlist *sg = ttinfo->sgl;
sgl               176 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	struct scatterlist *sgl = ttinfo->sgl;
sgl               178 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	unsigned int sg_offset = sgl->offset;
sgl               188 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0)
sgl               191 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	ttinfo->nr_pages = (xferlen + sgl->offset +
sgl               203 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	sgl->offset = 0;
sgl               204 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE);
sgl               205 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	sgl->offset = sg_offset;
sgl               212 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset,
sgl               218 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 		dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE);
sgl               245 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	ttinfo->sgl = cmd->se_cmd.t_data_sg;
sgl               253 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 		ttinfo->sgl = NULL;
sgl               270 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 		if (ttinfo->sgl) {
sgl               285 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 			dma_unmap_sg(&ppm->pdev->dev, ttinfo->sgl,
sgl               261 drivers/target/target_core_file.c fd_execute_rw_aio(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               285 drivers/target/target_core_file.c 	for_each_sg(sgl, sg, sgl_nents, i) {
sgl               319 drivers/target/target_core_file.c 		    u32 block_size, struct scatterlist *sgl,
sgl               335 drivers/target/target_core_file.c 	for_each_sg(sgl, sg, sgl_nents, i) {
sgl               595 drivers/target/target_core_file.c fd_execute_rw_buffered(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               618 drivers/target/target_core_file.c 			       sgl, sgl_nents, cmd->data_length, 0);
sgl               643 drivers/target/target_core_file.c 			       sgl, sgl_nents, cmd->data_length, 1);
sgl               679 drivers/target/target_core_file.c fd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               697 drivers/target/target_core_file.c 		return fd_execute_rw_aio(cmd, sgl, sgl_nents, data_direction);
sgl               698 drivers/target/target_core_file.c 	return fd_execute_rw_buffered(cmd, sgl, sgl_nents, data_direction);
sgl               681 drivers/target/target_core_iblock.c iblock_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               739 drivers/target/target_core_iblock.c 	for_each_sg(sgl, sg, sgl_nents, i) {
sgl               846 drivers/target/target_core_pscsi.c pscsi_map_sg(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               854 drivers/target/target_core_pscsi.c 	int nr_pages = (cmd->data_length + sgl[0].offset +
sgl               863 drivers/target/target_core_pscsi.c 	for_each_sg(sgl, sg, sgl_nents, i) {
sgl               957 drivers/target/target_core_pscsi.c 	struct scatterlist *sgl = cmd->t_data_sg;
sgl               986 drivers/target/target_core_pscsi.c 	if (sgl) {
sgl               987 drivers/target/target_core_pscsi.c 		ret = pscsi_map_sg(cmd, sgl, sgl_nents, req);
sgl               417 drivers/target/target_core_rd.c rd_execute_rw(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents,
sgl               461 drivers/target/target_core_rd.c 	sg_miter_start(&m, sgl, sgl_nents,
sgl               488 drivers/target/target_core_sbc.c 	write_sg = write_tbl.sgl;
sgl              1502 drivers/target/target_core_transport.c transport_generic_map_mem_to_cmd(struct se_cmd *cmd, struct scatterlist *sgl,
sgl              1505 drivers/target/target_core_transport.c 	if (!sgl || !sgl_count)
sgl              1519 drivers/target/target_core_transport.c 	cmd->t_data_sg = sgl;
sgl              1560 drivers/target/target_core_transport.c 		struct scatterlist *sgl, u32 sgl_count,
sgl              1633 drivers/target/target_core_transport.c 		BUG_ON(!sgl);
sgl              1647 drivers/target/target_core_transport.c 			if (sgl)
sgl              1648 drivers/target/target_core_transport.c 				buf = kmap(sg_page(sgl)) + sgl->offset;
sgl              1651 drivers/target/target_core_transport.c 				memset(buf, 0, sgl->length);
sgl              1652 drivers/target/target_core_transport.c 				kunmap(sg_page(sgl));
sgl              1656 drivers/target/target_core_transport.c 		rc = transport_generic_map_mem_to_cmd(se_cmd, sgl, sgl_count,
sgl              2409 drivers/target/target_core_transport.c void target_free_sgl(struct scatterlist *sgl, int nents)
sgl              2411 drivers/target/target_core_transport.c 	sgl_free_n_order(sgl, nents, 0);
sgl              2516 drivers/target/target_core_transport.c target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents, u32 length,
sgl              2521 drivers/target/target_core_transport.c 	*sgl = sgl_alloc_order(length, 0, chainable, gfp, nents);
sgl              2522 drivers/target/target_core_transport.c 	return *sgl ? 0 : -ENOMEM;
sgl               933 drivers/tty/serial/atmel_serial.c 	struct scatterlist sgl[2], *sg, *sg_tx = &atmel_port->sg_tx;
sgl               965 drivers/tty/serial/atmel_serial.c 		sg_init_table(sgl, 2);
sgl               969 drivers/tty/serial/atmel_serial.c 			sg = &sgl[sg_len++];
sgl               977 drivers/tty/serial/atmel_serial.c 			sg = &sgl[sg_len++];
sgl               989 drivers/tty/serial/atmel_serial.c 					       sgl,
sgl               414 drivers/tty/serial/fsl_lpuart.c 	struct scatterlist *sgl = sport->tx_sgl;
sgl               425 drivers/tty/serial/fsl_lpuart.c 		sg_init_one(sgl, xmit->buf + xmit->tail, sport->dma_tx_bytes);
sgl               428 drivers/tty/serial/fsl_lpuart.c 		sg_init_table(sgl, 2);
sgl               429 drivers/tty/serial/fsl_lpuart.c 		sg_set_buf(sgl, xmit->buf + xmit->tail,
sgl               431 drivers/tty/serial/fsl_lpuart.c 		sg_set_buf(sgl + 1, xmit->buf, xmit->head);
sgl               434 drivers/tty/serial/fsl_lpuart.c 	ret = dma_map_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
sgl               440 drivers/tty/serial/fsl_lpuart.c 	sport->dma_tx_desc = dmaengine_prep_slave_sg(sport->dma_tx_chan, sgl,
sgl               444 drivers/tty/serial/fsl_lpuart.c 		dma_unmap_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
sgl               464 drivers/tty/serial/fsl_lpuart.c 	struct scatterlist *sgl = &sport->tx_sgl[0];
sgl               470 drivers/tty/serial/fsl_lpuart.c 	dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
sgl               551 drivers/tty/serial/imx.c 	struct scatterlist *sgl = &sport->tx_sgl[0];
sgl               558 drivers/tty/serial/imx.c 	dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
sgl               590 drivers/tty/serial/imx.c 	struct scatterlist *sgl = sport->tx_sgl;
sgl               608 drivers/tty/serial/imx.c 		sg_init_one(sgl, xmit->buf + xmit->tail, sport->tx_bytes);
sgl               611 drivers/tty/serial/imx.c 		sg_init_table(sgl, 2);
sgl               612 drivers/tty/serial/imx.c 		sg_set_buf(sgl, xmit->buf + xmit->tail,
sgl               614 drivers/tty/serial/imx.c 		sg_set_buf(sgl + 1, xmit->buf, xmit->head);
sgl               617 drivers/tty/serial/imx.c 	ret = dma_map_sg(dev, sgl, sport->dma_tx_nents, DMA_TO_DEVICE);
sgl               622 drivers/tty/serial/imx.c 	desc = dmaengine_prep_slave_sg(chan, sgl, ret,
sgl               625 drivers/tty/serial/imx.c 		dma_unmap_sg(dev, sgl, sport->dma_tx_nents,
sgl              1076 drivers/tty/serial/imx.c 	struct scatterlist *sgl = &sport->rx_sgl;
sgl              1106 drivers/tty/serial/imx.c 		rx_ring->head = sg_dma_len(sgl) - state.residue;
sgl              1109 drivers/tty/serial/imx.c 		bd_size = sg_dma_len(sgl) / sport->rx_periods;
sgl              1112 drivers/tty/serial/imx.c 		if (rx_ring->head <= sg_dma_len(sgl) &&
sgl              1119 drivers/tty/serial/imx.c 			dma_sync_sg_for_cpu(sport->port.dev, sgl, 1,
sgl              1126 drivers/tty/serial/imx.c 			dma_sync_sg_for_device(sport->port.dev, sgl, 1,
sgl              1134 drivers/tty/serial/imx.c 			WARN_ON(rx_ring->head > sg_dma_len(sgl));
sgl              1150 drivers/tty/serial/imx.c 	struct scatterlist *sgl = &sport->rx_sgl;
sgl              1160 drivers/tty/serial/imx.c 	sg_init_one(sgl, sport->rx_buf, RX_BUF_SIZE);
sgl              1161 drivers/tty/serial/imx.c 	ret = dma_map_sg(dev, sgl, 1, DMA_FROM_DEVICE);
sgl              1167 drivers/tty/serial/imx.c 	desc = dmaengine_prep_dma_cyclic(chan, sg_dma_address(sgl),
sgl              1168 drivers/tty/serial/imx.c 		sg_dma_len(sgl), sg_dma_len(sgl) / sport->rx_periods,
sgl              1172 drivers/tty/serial/imx.c 		dma_unmap_sg(dev, sgl, 1, DMA_FROM_DEVICE);
sgl              1515 drivers/tty/serial/imx.c 	struct scatterlist *sgl = &sport->tx_sgl[0];
sgl              1527 drivers/tty/serial/imx.c 		dma_unmap_sg(sport->port.dev, sgl, sport->dma_tx_nents,
sgl               552 drivers/tty/serial/mxs-auart.c 	struct scatterlist *sgl = &s->tx_sgl;
sgl               566 drivers/tty/serial/mxs-auart.c 	sg_init_one(sgl, s->tx_dma_buf, size);
sgl               567 drivers/tty/serial/mxs-auart.c 	dma_map_sg(s->dev, sgl, 1, DMA_TO_DEVICE);
sgl               568 drivers/tty/serial/mxs-auart.c 	desc = dmaengine_prep_slave_sg(channel, sgl,
sgl               874 drivers/tty/serial/mxs-auart.c 	struct scatterlist *sgl = &s->rx_sgl;
sgl               890 drivers/tty/serial/mxs-auart.c 	sg_init_one(sgl, s->rx_dma_buf, UART_XMIT_SIZE);
sgl               891 drivers/tty/serial/mxs-auart.c 	dma_map_sg(s->dev, sgl, 1, DMA_FROM_DEVICE);
sgl               892 drivers/tty/serial/mxs-auart.c 	desc = dmaengine_prep_slave_sg(channel, sgl, 1, DMA_DEV_TO_MEM,
sgl              1061 drivers/usb/gadget/function/f_fs.c 			req->sg	= io_data->sgt.sgl;
sgl              1105 drivers/usb/gadget/function/f_fs.c 			req->sg	= io_data->sgt.sgl;
sgl               458 drivers/usb/storage/isd200.c 	srb->sdb.table.sgl = buff ? &info->sg : NULL;
sgl               462 drivers/usb/storage/uas.c 	urb->sg = sdb->table.sgl;
sgl                56 drivers/usb/usbip/stub.h 	struct scatterlist *sgl;
sgl               307 drivers/usb/usbip/stub_main.c 		if (urb->transfer_buffer && !priv->sgl) {
sgl               322 drivers/usb/usbip/stub_main.c 	if (priv->sgl)
sgl               323 drivers/usb/usbip/stub_main.c 		sgl_free(priv->sgl);
sgl               459 drivers/usb/usbip/stub_rx.c 	struct scatterlist *sgl = NULL, *sg;
sgl               492 drivers/usb/usbip/stub_rx.c 			sgl = sgl_alloc(buf_len, GFP_KERNEL, &nents);
sgl               493 drivers/usb/usbip/stub_rx.c 			if (!sgl)
sgl               541 drivers/usb/usbip/stub_rx.c 				priv->urbs[0]->sg = sgl;
sgl               559 drivers/usb/usbip/stub_rx.c 		for_each_sg(sgl, sg, nents, i) {
sgl               571 drivers/usb/usbip/stub_rx.c 		priv->sgl = sgl;
sgl               621 drivers/usb/usbip/stub_rx.c 	sgl_free(sgl);
sgl                79 drivers/usb/usbip/stub_tx.c 	if (priv->sgl) {
sgl               188 drivers/usb/usbip/stub_tx.c 		else if (usb_pipein(urb->pipe) && priv->sgl)
sgl               207 drivers/usb/usbip/stub_tx.c 		if (priv->sgl) {
sgl               223 drivers/usb/usbip/stub_tx.c 		if (usb_pipein(urb->pipe) && priv->sgl) {
sgl               624 drivers/vhost/scsi.c 		      struct scatterlist *sgl,
sgl               628 drivers/vhost/scsi.c 	struct scatterlist *sg = sgl;
sgl               250 drivers/xen/gntdev-dmabuf.c 				dma_unmap_sg_attrs(attach->dev, sgt->sgl,
sgl               291 drivers/xen/gntdev-dmabuf.c 		if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgl               656 drivers/xen/gntdev-dmabuf.c 	for_each_sg_page(sgt->sgl, &sg_iter, sgt->nents, 0) {
sgl               466 drivers/xen/swiotlb-xen.c xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems,
sgl               474 drivers/xen/swiotlb-xen.c 	for_each_sg(sgl, sg, nelems, i)
sgl               481 drivers/xen/swiotlb-xen.c xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems,
sgl               489 drivers/xen/swiotlb-xen.c 	for_each_sg(sgl, sg, nelems, i) {
sgl               499 drivers/xen/swiotlb-xen.c 	xen_swiotlb_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC);
sgl               500 drivers/xen/swiotlb-xen.c 	sg_dma_len(sgl) = 0;
sgl               505 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl,
sgl               511 drivers/xen/swiotlb-xen.c 	for_each_sg(sgl, sg, nelems, i) {
sgl               518 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_sg_for_device(struct device *dev, struct scatterlist *sgl,
sgl               524 drivers/xen/swiotlb-xen.c 	for_each_sg(sgl, sg, nelems, i) {
sgl               129 drivers/xen/xen-scsiback.c 	struct scatterlist *sgl;
sgl               277 drivers/xen/xen-scsiback.c 	kfree(req->sgl);
sgl               278 drivers/xen/xen-scsiback.c 	req->sgl = NULL;
sgl               408 drivers/xen/xen-scsiback.c 			pending_req->sgl, pending_req->n_sg,
sgl               524 drivers/xen/xen-scsiback.c 	pending_req->sgl = kmalloc_array(nr_segments,
sgl               526 drivers/xen/xen-scsiback.c 	if (!pending_req->sgl)
sgl               529 drivers/xen/xen-scsiback.c 	sg_init_table(pending_req->sgl, nr_segments);
sgl               563 drivers/xen/xen-scsiback.c 	for_each_sg(pending_req->sgl, sg, nr_segments, i) {
sgl               983 fs/cifs/smbdirect.c 	struct scatterlist *sgl, int data_length, int remaining_data_length)
sgl               995 fs/cifs/smbdirect.c 	num_sgs = sgl ? sg_nents(sgl) : 0;
sgl               996 fs/cifs/smbdirect.c 	for_each_sg(sgl, sg, num_sgs, i) {
sgl              1036 fs/cifs/smbdirect.c 	struct scatterlist sgl;
sgl              1038 fs/cifs/smbdirect.c 	sg_init_table(&sgl, 1);
sgl              1039 fs/cifs/smbdirect.c 	sg_set_page(&sgl, page, size, offset);
sgl              1041 fs/cifs/smbdirect.c 	return smbd_post_send_sgl(info, &sgl, size, remaining_data_length);
sgl              1068 fs/cifs/smbdirect.c 	struct scatterlist sgl[SMBDIRECT_MAX_SGE];
sgl              1075 fs/cifs/smbdirect.c 	sg_init_table(sgl, n_vec);
sgl              1078 fs/cifs/smbdirect.c 		sg_set_buf(&sgl[i], iov[i].iov_base, iov[i].iov_len);
sgl              1081 fs/cifs/smbdirect.c 	return smbd_post_send_sgl(info, sgl, data_length, remaining_data_length);
sgl              2323 fs/cifs/smbdirect.c 			ib_dma_unmap_sg(info->id->device, mr->sgl,
sgl              2326 fs/cifs/smbdirect.c 		kfree(mr->sgl);
sgl              2362 fs/cifs/smbdirect.c 		smbdirect_mr->sgl = kcalloc(
sgl              2366 fs/cifs/smbdirect.c 		if (!smbdirect_mr->sgl) {
sgl              2385 fs/cifs/smbdirect.c 		kfree(smbdirect_mr->sgl);
sgl              2468 fs/cifs/smbdirect.c 	sg_init_table(smbdirect_mr->sgl, num_pages);
sgl              2474 fs/cifs/smbdirect.c 		sg_set_page(&smbdirect_mr->sgl[0], pages[0], tailsz, offset);
sgl              2480 fs/cifs/smbdirect.c 		&smbdirect_mr->sgl[0], pages[0], PAGE_SIZE - offset, offset);
sgl              2483 fs/cifs/smbdirect.c 		sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0);
sgl              2486 fs/cifs/smbdirect.c 	sg_set_page(&smbdirect_mr->sgl[i], pages[i],
sgl              2492 fs/cifs/smbdirect.c 	rc = ib_dma_map_sg(info->id->device, smbdirect_mr->sgl, num_pages, dir);
sgl              2499 fs/cifs/smbdirect.c 	rc = ib_map_mr_sg(smbdirect_mr->mr, smbdirect_mr->sgl, num_pages,
sgl              2536 fs/cifs/smbdirect.c 	ib_dma_unmap_sg(info->id->device, smbdirect_mr->sgl,
sgl              2604 fs/cifs/smbdirect.c 			info->id->device, smbdirect_mr->sgl,
sgl               293 fs/cifs/smbdirect.h 	struct scatterlist	*sgl;
sgl                77 include/crypto/if_alg.h 	struct af_alg_sgl sgl;
sgl               165 include/crypto/if_alg.h int af_alg_make_sg(struct af_alg_sgl *sgl, struct iov_iter *iter, int len);
sgl               166 include/crypto/if_alg.h void af_alg_free_sg(struct af_alg_sgl *sgl);
sgl               206 include/linux/dma-mapping.h int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents,
sgl               216 include/linux/dma-mapping.h 		struct scatterlist *sgl, int nents, enum dma_data_direction dir);
sgl               223 include/linux/dma-mapping.h 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               233 include/linux/dma-mapping.h void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl,
sgl               238 include/linux/dma-mapping.h 		struct scatterlist *sgl, int nents, enum dma_data_direction dir);
sgl               245 include/linux/dma-mapping.h 		struct scatterlist *sgl, int nents, enum dma_data_direction dir,
sgl               254 include/linux/dma-mapping.h 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               156 include/linux/dmaengine.h 	struct data_chunk sgl[0];
sgl               780 include/linux/dmaengine.h 		struct dma_chan *chan, struct scatterlist *sgl,
sgl               838 include/linux/dmaengine.h 	struct dma_chan *chan, struct scatterlist *sgl,	unsigned int sg_len,
sgl               844 include/linux/dmaengine.h 	return chan->device->device_prep_slave_sg(chan, sgl, sg_len,
sgl               851 include/linux/dmaengine.h 	struct dma_chan *chan, struct scatterlist *sgl,	unsigned int sg_len,
sgl               858 include/linux/dmaengine.h 	return chan->device->device_prep_slave_sg(chan, sgl, sg_len,
sgl              1122 include/linux/dmaengine.h static inline size_t dmaengine_get_icg(bool inc, bool sgl, size_t icg,
sgl              1128 include/linux/dmaengine.h 		else if (sgl)
sgl               645 include/linux/kfifo.h #define	kfifo_dma_in_prepare(fifo, sgl, nents, len) \
sgl               648 include/linux/kfifo.h 	struct scatterlist *__sgl = (sgl); \
sgl               696 include/linux/kfifo.h #define	kfifo_dma_out_prepare(fifo, sgl, nents, len) \
sgl               699 include/linux/kfifo.h 	struct scatterlist *__sgl = (sgl); \
sgl               779 include/linux/kfifo.h 	struct scatterlist *sgl, int nents, unsigned int len);
sgl               782 include/linux/kfifo.h 	struct scatterlist *sgl, int nents, unsigned int len);
sgl               801 include/linux/kfifo.h 	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize);
sgl               807 include/linux/kfifo.h 	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize);
sgl                52 include/linux/mpi.h MPI mpi_read_raw_from_sgl(struct scatterlist *sgl, unsigned int len);
sgl               239 include/linux/mtd/ubi.h int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl,
sgl               267 include/linux/mtd/ubi.h 			      struct ubi_sgl *sgl, int offset, int len)
sgl               269 include/linux/mtd/ubi.h 	return ubi_leb_read_sg(desc, lnum, sgl, offset, len, 0);
sgl              1336 include/linux/netdevice.h 						      struct scatterlist *sgl,
sgl              1342 include/linux/netdevice.h 						       struct scatterlist *sgl,
sgl               645 include/linux/nvme.h 	struct nvme_sgl_desc	sgl;
sgl                31 include/linux/pci-p2pdma.h void pci_p2pmem_free_sgl(struct pci_dev *pdev, struct scatterlist *sgl);
sgl                80 include/linux/pci-p2pdma.h 		struct scatterlist *sgl)
sgl                43 include/linux/scatterlist.h 	struct scatterlist *sgl;	/* the list */
sgl               165 include/linux/scatterlist.h 			    struct scatterlist *sgl)
sgl               177 include/linux/scatterlist.h 	prv[prv_nents - 1].page_link = ((unsigned long) sgl | SG_CHAIN)
sgl               248 include/linux/scatterlist.h static inline void sg_init_marker(struct scatterlist *sgl,
sgl               251 include/linux/scatterlist.h 	sg_mark_end(&sgl[nents - 1]);
sgl               289 include/linux/scatterlist.h void sgl_free_n_order(struct scatterlist *sgl, int nents, int order);
sgl               290 include/linux/scatterlist.h void sgl_free_order(struct scatterlist *sgl, int order);
sgl               291 include/linux/scatterlist.h void sgl_free(struct scatterlist *sgl);
sgl               294 include/linux/scatterlist.h size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf,
sgl               297 include/linux/scatterlist.h size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               299 include/linux/scatterlist.h size_t sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               302 include/linux/scatterlist.h size_t sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               304 include/linux/scatterlist.h size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               306 include/linux/scatterlist.h size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               461 include/linux/scatterlist.h void sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl,
sgl              4352 include/rdma/ib_verbs.h int ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents,
sgl               185 include/scsi/scsi_cmnd.h 	return cmd->sdb.table.sgl;
sgl               299 include/scsi/scsi_cmnd.h 	return cmd->prot_sdb ? cmd->prot_sdb->table.sgl : NULL;
sgl               205 include/target/target_core_fabric.h int	target_alloc_sgl(struct scatterlist **sgl, unsigned int *nents,
sgl               207 include/target/target_core_fabric.h void	target_free_sgl(struct scatterlist *sgl, int nents);
sgl               240 kernel/dma/direct.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               245 kernel/dma/direct.c 	for_each_sg(sgl, sg, nents, i) {
sgl               279 kernel/dma/direct.c 		struct scatterlist *sgl, int nents, enum dma_data_direction dir)
sgl               284 kernel/dma/direct.c 	for_each_sg(sgl, sg, nents, i) {
sgl               313 kernel/dma/direct.c void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl,
sgl               319 kernel/dma/direct.c 	for_each_sg(sgl, sg, nents, i)
sgl               352 kernel/dma/direct.c int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents,
sgl               358 kernel/dma/direct.c 	for_each_sg(sgl, sg, nents, i) {
sgl               369 kernel/dma/direct.c 	dma_direct_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC);
sgl                21 kernel/dma/dummy.c static int dma_dummy_map_sg(struct device *dev, struct scatterlist *sgl,
sgl               135 kernel/dma/mapping.c 		sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgl                37 kernel/dma/virt.c static int dma_virt_map_sg(struct device *dev, struct scatterlist *sgl,
sgl                44 kernel/dma/virt.c 	for_each_sg(sgl, sg, nents, i) {
sgl               295 lib/kfifo.c    static int setup_sgl_buf(struct scatterlist *sgl, void *buf,
sgl               321 lib/kfifo.c    			sg_set_page(sgl, page, l - off, off);
sgl               322 lib/kfifo.c    			sgl = sg_next(sgl);
sgl               323 lib/kfifo.c    			if (++n == nents || sgl == NULL)
sgl               330 lib/kfifo.c    	sg_set_page(sgl, page, len, off);
sgl               334 lib/kfifo.c    static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl,
sgl               350 lib/kfifo.c    	n = setup_sgl_buf(sgl, fifo->data + off, nents, l);
sgl               351 lib/kfifo.c    	n += setup_sgl_buf(sgl + n, fifo->data, nents - n, len - l);
sgl               357 lib/kfifo.c    		struct scatterlist *sgl, int nents, unsigned int len)
sgl               365 lib/kfifo.c    	return setup_sgl(fifo, sgl, nents, len, fifo->in);
sgl               370 lib/kfifo.c    		struct scatterlist *sgl, int nents, unsigned int len)
sgl               378 lib/kfifo.c    	return setup_sgl(fifo, sgl, nents, len, fifo->out);
sgl               549 lib/kfifo.c    	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize)
sgl               558 lib/kfifo.c    	return setup_sgl(fifo, sgl, nents, len, fifo->in + recsize);
sgl               572 lib/kfifo.c    	struct scatterlist *sgl, int nents, unsigned int len, size_t recsize)
sgl               581 lib/kfifo.c    	return setup_sgl(fifo, sgl, nents, len, fifo->out + recsize);
sgl               247 lib/mpi/mpicoder.c int mpi_write_to_sgl(MPI a, struct scatterlist *sgl, unsigned nbytes,
sgl               269 lib/mpi/mpicoder.c 	nents = sg_nents_for_len(sgl, nbytes);
sgl               273 lib/mpi/mpicoder.c 	sg_miter_start(&miter, sgl, nents, SG_MITER_ATOMIC | SG_MITER_TO_SG);
sgl               330 lib/mpi/mpicoder.c MPI mpi_read_raw_from_sgl(struct scatterlist *sgl, unsigned int nbytes)
sgl               340 lib/mpi/mpicoder.c 	ents = sg_nents_for_len(sgl, nbytes);
sgl               344 lib/mpi/mpicoder.c 	sg_miter_start(&miter, sgl, ents, SG_MITER_ATOMIC | SG_MITER_FROM_SG);
sgl               101 lib/scatterlist.c struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents)
sgl               106 lib/scatterlist.c 	for_each_sg(sgl, sg, nents, i)
sgl               124 lib/scatterlist.c void sg_init_table(struct scatterlist *sgl, unsigned int nents)
sgl               126 lib/scatterlist.c 	memset(sgl, 0, sizeof(*sgl) * nents);
sgl               127 lib/scatterlist.c 	sg_init_marker(sgl, nents);
sgl               195 lib/scatterlist.c 	struct scatterlist *sgl, *next;
sgl               198 lib/scatterlist.c 	if (unlikely(!table->sgl))
sgl               201 lib/scatterlist.c 	sgl = table->sgl;
sgl               213 lib/scatterlist.c 			next = sg_chain_ptr(&sgl[curr_max_ents - 1]);
sgl               225 lib/scatterlist.c 			free_fn(sgl, alloc_size);
sgl               226 lib/scatterlist.c 		sgl = next;
sgl               230 lib/scatterlist.c 	table->sgl = NULL;
sgl               327 lib/scatterlist.c 			table->sgl = sg;
sgl               419 lib/scatterlist.c 	for_each_sg(sgt->sgl, s, sgt->orig_nents, i) {
sgl               490 lib/scatterlist.c 	struct scatterlist *sgl, *sg;
sgl               506 lib/scatterlist.c 	sgl = kmalloc_array(nalloc, sizeof(struct scatterlist),
sgl               508 lib/scatterlist.c 	if (!sgl)
sgl               511 lib/scatterlist.c 	sg_init_table(sgl, nalloc);
sgl               512 lib/scatterlist.c 	sg = sgl;
sgl               517 lib/scatterlist.c 			sgl_free(sgl);
sgl               528 lib/scatterlist.c 	return sgl;
sgl               560 lib/scatterlist.c void sgl_free_n_order(struct scatterlist *sgl, int nents, int order)
sgl               566 lib/scatterlist.c 	for_each_sg(sgl, sg, nents, i) {
sgl               573 lib/scatterlist.c 	kfree(sgl);
sgl               582 lib/scatterlist.c void sgl_free_order(struct scatterlist *sgl, int order)
sgl               584 lib/scatterlist.c 	sgl_free_n_order(sgl, INT_MAX, order);
sgl               592 lib/scatterlist.c void sgl_free(struct scatterlist *sgl)
sgl               594 lib/scatterlist.c 	sgl_free_order(sgl, 0);
sgl               674 lib/scatterlist.c void sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl,
sgl               679 lib/scatterlist.c 	__sg_page_iter_start(&miter->piter, sgl, nents, 0);
sgl               840 lib/scatterlist.c size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf,
sgl               852 lib/scatterlist.c 	sg_miter_start(&miter, sgl, nents, sg_flags);
sgl               886 lib/scatterlist.c size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               889 lib/scatterlist.c 	return sg_copy_buffer(sgl, nents, (void *)buf, buflen, 0, false);
sgl               903 lib/scatterlist.c size_t sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               906 lib/scatterlist.c 	return sg_copy_buffer(sgl, nents, buf, buflen, 0, true);
sgl               921 lib/scatterlist.c size_t sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               924 lib/scatterlist.c 	return sg_copy_buffer(sgl, nents, (void *)buf, buflen, skip, false);
sgl               939 lib/scatterlist.c size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               942 lib/scatterlist.c 	return sg_copy_buffer(sgl, nents, buf, buflen, skip, true);
sgl               955 lib/scatterlist.c size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents,
sgl               962 lib/scatterlist.c 	sg_miter_start(&miter, sgl, nents, sg_flags);
sgl                54 lib/sg_pool.c  static void sg_pool_free(struct scatterlist *sgl, unsigned int nents)
sgl                59 lib/sg_pool.c  	mempool_free(sgl, sgp->pool);
sgl               121 lib/sg_pool.c  			sg_init_table(table->sgl, nents);
sgl               401 net/8021q/vlan_dev.c 				   struct scatterlist *sgl, unsigned int sgc)
sgl               408 net/8021q/vlan_dev.c 		rc = ops->ndo_fcoe_ddp_setup(real_dev, xid, sgl, sgc);
sgl               448 net/8021q/vlan_dev.c 				    struct scatterlist *sgl, unsigned int sgc)
sgl               455 net/8021q/vlan_dev.c 		rc = ops->ndo_fcoe_ddp_target(real_dev, xid, sgl, sgc);
sgl               188 net/ceph/crypto.c 		sgt->sgl = prealloc_sg;
sgl               192 net/ceph/crypto.c 	for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) {
sgl               239 net/ceph/crypto.c 	skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv);
sgl               516 net/smc/smc_clc.c 				(conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgl               588 net/smc/smc_clc.c 				(conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgl               770 net/smc/smc_core.c 	sg_set_buf(buf_desc->sgt[SMC_SINGLE_LINK].sgl,
sgl               369 net/smc/smc_ib.c 			      buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgl               408 net/smc/smc_ib.c 	for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg,
sgl               428 net/smc/smc_ib.c 	for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg,
sgl               447 net/smc/smc_ib.c 				     buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgl               460 net/smc/smc_ib.c 	if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address)
sgl               464 net/smc/smc_ib.c 			buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgl               467 net/smc/smc_ib.c 	buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0;
sgl               236 net/smc/smc_llc.c 		(u64)sg_dma_address(rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgl               315 net/smc/smc_tx.c 		sg_dma_address(conn->sndbuf_desc->sgt[SMC_SINGLE_LINK].sgl);
sgl                74 net/sunrpc/xprtrdma/svc_rdma_rw.c 	ctxt->rw_sg_table.sgl = ctxt->rw_first_sgl;
sgl                76 net/sunrpc/xprtrdma/svc_rdma_rw.c 				   ctxt->rw_sg_table.sgl,
sgl               145 net/sunrpc/xprtrdma/svc_rdma_rw.c 				    rdma->sc_port_num, ctxt->rw_sg_table.sgl,
sgl               356 net/sunrpc/xprtrdma/svc_rdma_rw.c 	struct scatterlist *sg = ctxt->rw_sg_table.sgl;
sgl               380 net/sunrpc/xprtrdma/svc_rdma_rw.c 	sg = ctxt->rw_sg_table.sgl;
sgl               435 net/sunrpc/xprtrdma/svc_rdma_rw.c 				       rdma->sc_port_num, ctxt->rw_sg_table.sgl,
sgl               615 net/sunrpc/xprtrdma/svc_rdma_rw.c 	sg = ctxt->rw_sg_table.sgl;
sgl               644 net/sunrpc/xprtrdma/svc_rdma_rw.c 			       ctxt->rw_sg_table.sgl, ctxt->rw_nents,
sgl               849 samples/vfio-mdev/mbochs.c 	if (!dma_map_sg(at->dev, sg->sgl, sg->nents, direction))
sgl                91 tools/virtio/linux/scatterlist.h 			    struct scatterlist *sgl)
sgl               103 tools/virtio/linux/scatterlist.h 	prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02;
sgl               149 tools/virtio/linux/scatterlist.h static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents)
sgl               151 tools/virtio/linux/scatterlist.h 	memset(sgl, 0, sizeof(*sgl) * nents);
sgl               152 tools/virtio/linux/scatterlist.h 	sg_mark_end(&sgl[nents - 1]);