sgt               194 arch/arm/include/asm/dma-mapping.h extern int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt               865 arch/arm/mm/dma-mapping.c int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt               879 arch/arm/mm/dma-mapping.c 	ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
sgt               883 arch/arm/mm/dma-mapping.c 	sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgt              1598 arch/arm/mm/dma-mapping.c static int arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt              1608 arch/arm/mm/dma-mapping.c 	return sg_alloc_table_from_pages(sgt, pages, count, 0, size,
sgt               129 arch/s390/include/asm/gmap.h int gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt,
sgt              1353 arch/s390/mm/gmap.c 				unsigned long *sgt)
sgt              1361 arch/s390/mm/gmap.c 		if (!(sgt[i] & _SEGMENT_ENTRY_ORIGIN))
sgt              1363 arch/s390/mm/gmap.c 		pgt = (unsigned long *)(sgt[i] & _REGION_ENTRY_ORIGIN);
sgt              1364 arch/s390/mm/gmap.c 		sgt[i] = _SEGMENT_ENTRY_EMPTY;
sgt              1382 arch/s390/mm/gmap.c 	unsigned long r3o, *r3e, *sgt;
sgt              1392 arch/s390/mm/gmap.c 	sgt = (unsigned long *)(*r3e & _REGION_ENTRY_ORIGIN);
sgt              1394 arch/s390/mm/gmap.c 	__gmap_unshadow_sgt(sg, raddr, sgt);
sgt              1396 arch/s390/mm/gmap.c 	page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT);
sgt              1412 arch/s390/mm/gmap.c 	unsigned long *sgt;
sgt              1420 arch/s390/mm/gmap.c 		sgt = (unsigned long *)(r3t[i] & _REGION_ENTRY_ORIGIN);
sgt              1422 arch/s390/mm/gmap.c 		__gmap_unshadow_sgt(sg, raddr, sgt);
sgt              1424 arch/s390/mm/gmap.c 		page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT);
sgt              1902 arch/s390/mm/gmap.c int gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt,
sgt              1910 arch/s390/mm/gmap.c 	BUG_ON(!gmap_is_shadow(sg) || (sgt & _REGION3_ENTRY_LARGE));
sgt              1915 arch/s390/mm/gmap.c 	page->index = sgt & _REGION_ENTRY_ORIGIN;
sgt              1938 arch/s390/mm/gmap.c 		*table |= sgt & _REGION_ENTRY_PROTECT;
sgt              1949 arch/s390/mm/gmap.c 	origin = sgt & _REGION_ENTRY_ORIGIN;
sgt              1950 arch/s390/mm/gmap.c 	offset = ((sgt & _REGION_ENTRY_OFFSET) >> 6) * PAGE_SIZE;
sgt              1951 arch/s390/mm/gmap.c 	len = ((sgt & _REGION_ENTRY_LENGTH) + 1) * PAGE_SIZE - offset;
sgt               808 drivers/crypto/caam/caamalg_qi.c 	struct qm_sg_entry sgt[0];
sgt               828 drivers/crypto/caam/caamalg_qi.c 	struct qm_sg_entry sgt[0];
sgt              1061 drivers/crypto/caam/caamalg_qi.c 	sg_table = &edesc->sgt[0];
sgt              1063 drivers/crypto/caam/caamalg_qi.c 	if (unlikely(offsetof(struct aead_edesc, sgt) + qm_sg_bytes + ivsize >
sgt              1239 drivers/crypto/caam/caamalg_qi.c 		memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes,
sgt              1322 drivers/crypto/caam/caamalg_qi.c 	if (unlikely(offsetof(struct skcipher_edesc, sgt) + qm_sg_bytes +
sgt              1341 drivers/crypto/caam/caamalg_qi.c 	sg_table = &edesc->sgt[0];
sgt               468 drivers/crypto/caam/caamalg_qi2.c 	sg_table = &edesc->sgt[0];
sgt               470 drivers/crypto/caam/caamalg_qi2.c 	if (unlikely(offsetof(struct aead_edesc, sgt) + qm_sg_bytes + ivsize >
sgt              1189 drivers/crypto/caam/caamalg_qi2.c 	if (unlikely(offsetof(struct skcipher_edesc, sgt) + qm_sg_bytes +
sgt              1208 drivers/crypto/caam/caamalg_qi2.c 	sg_table = &edesc->sgt[0];
sgt              1427 drivers/crypto/caam/caamalg_qi2.c 		memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes,
sgt              1465 drivers/crypto/caam/caamalg_qi2.c 		memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes,
sgt              3513 drivers/crypto/caam/caamalg_qi2.c 		sg_table = &edesc->sgt[0];
sgt              3608 drivers/crypto/caam/caamalg_qi2.c 	sg_table = &edesc->sgt[0];
sgt              3702 drivers/crypto/caam/caamalg_qi2.c 	sg_table = &edesc->sgt[0];
sgt              3796 drivers/crypto/caam/caamalg_qi2.c 		struct dpaa2_sg_entry *sg_table = &edesc->sgt[0];
sgt              3972 drivers/crypto/caam/caamalg_qi2.c 		sg_table = &edesc->sgt[0];
sgt              4091 drivers/crypto/caam/caamalg_qi2.c 	sg_table = &edesc->sgt[0];
sgt              4195 drivers/crypto/caam/caamalg_qi2.c 		sg_table = &edesc->sgt[0];
sgt               117 drivers/crypto/caam/caamalg_qi2.h 	struct dpaa2_sg_entry sgt[0];
sgt               135 drivers/crypto/caam/caamalg_qi2.h 	struct dpaa2_sg_entry sgt[0];
sgt               149 drivers/crypto/caam/caamalg_qi2.h 	struct dpaa2_sg_entry sgt[0];
sgt                50 drivers/crypto/qce/dma.c qce_sgtable_add(struct sg_table *sgt, struct scatterlist *new_sgl)
sgt                52 drivers/crypto/qce/dma.c 	struct scatterlist *sg = sgt->sgl, *sg_last = NULL;
sgt                45 drivers/crypto/qce/dma.h qce_sgtable_add(struct sg_table *sgt, struct scatterlist *sg_add);
sgt               716 drivers/dma-buf/dma-buf.c 	if (attach->sgt)
sgt               717 drivers/dma-buf/dma-buf.c 		dmabuf->ops->unmap_dma_buf(attach, attach->sgt, attach->dir);
sgt               754 drivers/dma-buf/dma-buf.c 	if (attach->sgt) {
sgt               763 drivers/dma-buf/dma-buf.c 		return attach->sgt;
sgt               771 drivers/dma-buf/dma-buf.c 		attach->sgt = sg_table;
sgt               798 drivers/dma-buf/dma-buf.c 	if (attach->sgt == sg_table)
sgt                98 drivers/fpga/fpga-mgr.c 				  struct sg_table *sgt)
sgt               112 drivers/fpga/fpga-mgr.c 	sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG);
sgt               127 drivers/fpga/fpga-mgr.c 	len = sg_copy_to_buffer(sgt->sgl, sgt->nents, buf,
sgt               176 drivers/fpga/fpga-mgr.c 				struct sg_table *sgt)
sgt               180 drivers/fpga/fpga-mgr.c 	ret = fpga_mgr_write_init_sg(mgr, info, sgt);
sgt               187 drivers/fpga/fpga-mgr.c 		ret = mgr->mops->write_sg(mgr, sgt);
sgt               191 drivers/fpga/fpga-mgr.c 		sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG);
sgt               252 drivers/fpga/fpga-mgr.c 	struct sg_table sgt;
sgt               293 drivers/fpga/fpga-mgr.c 	rc = sg_alloc_table_from_pages(&sgt, pages, index, offset_in_page(buf),
sgt               299 drivers/fpga/fpga-mgr.c 	rc = fpga_mgr_buf_load_sg(mgr, info, &sgt);
sgt               300 drivers/fpga/fpga-mgr.c 	sg_free_table(&sgt);
sgt               357 drivers/fpga/fpga-mgr.c 	if (info->sgt)
sgt               358 drivers/fpga/fpga-mgr.c 		return fpga_mgr_buf_load_sg(mgr, info, info->sgt);
sgt               384 drivers/fpga/zynq-fpga.c static int zynq_fpga_ops_write(struct fpga_manager *mgr, struct sg_table *sgt)
sgt               400 drivers/fpga/zynq-fpga.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt               409 drivers/fpga/zynq-fpga.c 	    dma_map_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE);
sgt               426 drivers/fpga/zynq-fpga.c 	priv->cur_sg = sgt->sgl;
sgt               481 drivers/fpga/zynq-fpga.c 	dma_unmap_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE);
sgt               340 drivers/gpu/drm/arm/malidp_planes.c 		struct sg_table *sgt;
sgt               346 drivers/gpu/drm/arm/malidp_planes.c 		if (cma_obj->sgt)
sgt               347 drivers/gpu/drm/arm/malidp_planes.c 			sgt = cma_obj->sgt;
sgt               349 drivers/gpu/drm/arm/malidp_planes.c 			sgt = obj->dev->driver->gem_prime_get_sg_table(obj);
sgt               351 drivers/gpu/drm/arm/malidp_planes.c 		if (!sgt)
sgt               354 drivers/gpu/drm/arm/malidp_planes.c 		sgl = sgt->sgl;
sgt               358 drivers/gpu/drm/arm/malidp_planes.c 				if (!cma_obj->sgt)
sgt               359 drivers/gpu/drm/arm/malidp_planes.c 					kfree(sgt);
sgt               365 drivers/gpu/drm/arm/malidp_planes.c 		if (!cma_obj->sgt)
sgt               366 drivers/gpu/drm/arm/malidp_planes.c 			kfree(sgt);
sgt               114 drivers/gpu/drm/armada/armada_fb.c 	if (obj->obj.import_attach && !obj->sgt) {
sgt                66 drivers/gpu/drm/armada/armada_gem.c 		if (dobj->sgt)
sgt                68 drivers/gpu/drm/armada/armada_gem.c 						 dobj->sgt, DMA_TO_DEVICE);
sgt               381 drivers/gpu/drm/armada/armada_gem.c 	struct sg_table *sgt;
sgt               384 drivers/gpu/drm/armada/armada_gem.c 	sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
sgt               385 drivers/gpu/drm/armada/armada_gem.c 	if (!sgt)
sgt               393 drivers/gpu/drm/armada/armada_gem.c 		if (sg_alloc_table(sgt, count, GFP_KERNEL))
sgt               398 drivers/gpu/drm/armada/armada_gem.c 		for_each_sg(sgt->sgl, sg, count, i) {
sgt               410 drivers/gpu/drm/armada/armada_gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) {
sgt               411 drivers/gpu/drm/armada/armada_gem.c 			num = sgt->nents;
sgt               416 drivers/gpu/drm/armada/armada_gem.c 		if (sg_alloc_table(sgt, 1, GFP_KERNEL))
sgt               419 drivers/gpu/drm/armada/armada_gem.c 		sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0);
sgt               421 drivers/gpu/drm/armada/armada_gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0)
sgt               425 drivers/gpu/drm/armada/armada_gem.c 		if (sg_alloc_table(sgt, 1, GFP_KERNEL))
sgt               427 drivers/gpu/drm/armada/armada_gem.c 		sg_dma_address(sgt->sgl) = dobj->dev_addr;
sgt               428 drivers/gpu/drm/armada/armada_gem.c 		sg_dma_len(sgt->sgl) = dobj->obj.size;
sgt               432 drivers/gpu/drm/armada/armada_gem.c 	return sgt;
sgt               435 drivers/gpu/drm/armada/armada_gem.c 	for_each_sg(sgt->sgl, sg, num, i)
sgt               438 drivers/gpu/drm/armada/armada_gem.c 	sg_free_table(sgt);
sgt               440 drivers/gpu/drm/armada/armada_gem.c 	kfree(sgt);
sgt               445 drivers/gpu/drm/armada/armada_gem.c 	struct sg_table *sgt, enum dma_data_direction dir)
sgt               452 drivers/gpu/drm/armada/armada_gem.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
sgt               456 drivers/gpu/drm/armada/armada_gem.c 		for_each_sg(sgt->sgl, sg, sgt->nents, i)
sgt               460 drivers/gpu/drm/armada/armada_gem.c 	sg_free_table(sgt);
sgt               461 drivers/gpu/drm/armada/armada_gem.c 	kfree(sgt);
sgt               545 drivers/gpu/drm/armada/armada_gem.c 	dobj->sgt = dma_buf_map_attachment(dobj->obj.import_attach,
sgt               547 drivers/gpu/drm/armada/armada_gem.c 	if (IS_ERR(dobj->sgt)) {
sgt               548 drivers/gpu/drm/armada/armada_gem.c 		ret = PTR_ERR(dobj->sgt);
sgt               549 drivers/gpu/drm/armada/armada_gem.c 		dobj->sgt = NULL;
sgt               553 drivers/gpu/drm/armada/armada_gem.c 	if (dobj->sgt->nents > 1) {
sgt               557 drivers/gpu/drm/armada/armada_gem.c 	if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) {
sgt               561 drivers/gpu/drm/armada/armada_gem.c 	dobj->dev_addr = sg_dma_address(dobj->sgt->sgl);
sgt                19 drivers/gpu/drm/armada/armada_gem.h 	struct sg_table		*sgt;		/* for imported */
sgt               185 drivers/gpu/drm/drm_gem_cma_helper.c 		drm_prime_gem_destroy(gem_obj, cma_obj->sgt);
sgt               430 drivers/gpu/drm/drm_gem_cma_helper.c 	struct sg_table *sgt;
sgt               433 drivers/gpu/drm/drm_gem_cma_helper.c 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
sgt               434 drivers/gpu/drm/drm_gem_cma_helper.c 	if (!sgt)
sgt               437 drivers/gpu/drm/drm_gem_cma_helper.c 	ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr,
sgt               442 drivers/gpu/drm/drm_gem_cma_helper.c 	return sgt;
sgt               445 drivers/gpu/drm/drm_gem_cma_helper.c 	kfree(sgt);
sgt               470 drivers/gpu/drm/drm_gem_cma_helper.c 				  struct sg_table *sgt)
sgt               474 drivers/gpu/drm/drm_gem_cma_helper.c 	if (sgt->nents != 1) {
sgt               476 drivers/gpu/drm/drm_gem_cma_helper.c 		dma_addr_t next_addr = sg_dma_address(sgt->sgl);
sgt               480 drivers/gpu/drm/drm_gem_cma_helper.c 		for_each_sg(sgt->sgl, s, sgt->nents, i) {
sgt               500 drivers/gpu/drm/drm_gem_cma_helper.c 	cma_obj->paddr = sg_dma_address(sgt->sgl);
sgt               501 drivers/gpu/drm/drm_gem_cma_helper.c 	cma_obj->sgt = sgt;
sgt               633 drivers/gpu/drm/drm_gem_cma_helper.c 				       struct sg_table *sgt)
sgt               645 drivers/gpu/drm/drm_gem_cma_helper.c 	obj = drm_gem_cma_prime_import_sg_table(dev, attach, sgt);
sgt               116 drivers/gpu/drm/drm_gem_shmem_helper.c 		drm_prime_gem_destroy(obj, shmem->sgt);
sgt               119 drivers/gpu/drm/drm_gem_shmem_helper.c 		if (shmem->sgt) {
sgt               120 drivers/gpu/drm/drm_gem_shmem_helper.c 			dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl,
sgt               121 drivers/gpu/drm/drm_gem_shmem_helper.c 				     shmem->sgt->nents, DMA_BIDIRECTIONAL);
sgt               122 drivers/gpu/drm/drm_gem_shmem_helper.c 			sg_free_table(shmem->sgt);
sgt               123 drivers/gpu/drm/drm_gem_shmem_helper.c 			kfree(shmem->sgt);
sgt               393 drivers/gpu/drm/drm_gem_shmem_helper.c 	dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl,
sgt               394 drivers/gpu/drm/drm_gem_shmem_helper.c 		     shmem->sgt->nents, DMA_BIDIRECTIONAL);
sgt               395 drivers/gpu/drm/drm_gem_shmem_helper.c 	sg_free_table(shmem->sgt);
sgt               396 drivers/gpu/drm/drm_gem_shmem_helper.c 	kfree(shmem->sgt);
sgt               397 drivers/gpu/drm/drm_gem_shmem_helper.c 	shmem->sgt = NULL;
sgt               611 drivers/gpu/drm/drm_gem_shmem_helper.c 	struct sg_table *sgt;
sgt               613 drivers/gpu/drm/drm_gem_shmem_helper.c 	if (shmem->sgt)
sgt               614 drivers/gpu/drm/drm_gem_shmem_helper.c 		return shmem->sgt;
sgt               622 drivers/gpu/drm/drm_gem_shmem_helper.c 	sgt = drm_gem_shmem_get_sg_table(&shmem->base);
sgt               623 drivers/gpu/drm/drm_gem_shmem_helper.c 	if (IS_ERR(sgt)) {
sgt               624 drivers/gpu/drm/drm_gem_shmem_helper.c 		ret = PTR_ERR(sgt);
sgt               628 drivers/gpu/drm/drm_gem_shmem_helper.c 	dma_map_sg(obj->dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgt               630 drivers/gpu/drm/drm_gem_shmem_helper.c 	shmem->sgt = sgt;
sgt               632 drivers/gpu/drm/drm_gem_shmem_helper.c 	return sgt;
sgt               658 drivers/gpu/drm/drm_gem_shmem_helper.c 				    struct sg_table *sgt)
sgt               675 drivers/gpu/drm/drm_gem_shmem_helper.c 	ret = drm_prime_sg_to_page_addr_arrays(sgt, shmem->pages, NULL, npages);
sgt               679 drivers/gpu/drm/drm_gem_shmem_helper.c 	shmem->sgt = sgt;
sgt               617 drivers/gpu/drm/drm_prime.c 	struct sg_table *sgt;
sgt               623 drivers/gpu/drm/drm_prime.c 		sgt = obj->funcs->get_sg_table(obj);
sgt               625 drivers/gpu/drm/drm_prime.c 		sgt = obj->dev->driver->gem_prime_get_sg_table(obj);
sgt               627 drivers/gpu/drm/drm_prime.c 	if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgt               629 drivers/gpu/drm/drm_prime.c 		sg_free_table(sgt);
sgt               630 drivers/gpu/drm/drm_prime.c 		kfree(sgt);
sgt               631 drivers/gpu/drm/drm_prime.c 		sgt = ERR_PTR(-ENOMEM);
sgt               634 drivers/gpu/drm/drm_prime.c 	return sgt;
sgt               647 drivers/gpu/drm/drm_prime.c 			   struct sg_table *sgt,
sgt               650 drivers/gpu/drm/drm_prime.c 	if (!sgt)
sgt               653 drivers/gpu/drm/drm_prime.c 	dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgt               655 drivers/gpu/drm/drm_prime.c 	sg_free_table(sgt);
sgt               656 drivers/gpu/drm/drm_prime.c 	kfree(sgt);
sgt               862 drivers/gpu/drm/drm_prime.c 	struct sg_table *sgt;
sgt               887 drivers/gpu/drm/drm_prime.c 	sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL);
sgt               888 drivers/gpu/drm/drm_prime.c 	if (IS_ERR(sgt)) {
sgt               889 drivers/gpu/drm/drm_prime.c 		ret = PTR_ERR(sgt);
sgt               893 drivers/gpu/drm/drm_prime.c 	obj = dev->driver->gem_prime_import_sg_table(dev, attach, sgt);
sgt               905 drivers/gpu/drm/drm_prime.c 	dma_buf_unmap_attachment(attach, sgt, DMA_BIDIRECTIONAL);
sgt               947 drivers/gpu/drm/drm_prime.c int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages,
sgt               957 drivers/gpu/drm/drm_prime.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgt                23 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	struct sg_table *sgt = etnaviv_obj->sgt;
sgt                30 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_map_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgt                36 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	struct sg_table *sgt = etnaviv_obj->sgt;
sgt                54 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_unmap_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
sgt                75 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	if (etnaviv_obj->sgt) {
sgt                77 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		sg_free_table(etnaviv_obj->sgt);
sgt                78 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		kfree(etnaviv_obj->sgt);
sgt                79 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		etnaviv_obj->sgt = NULL;
sgt               101 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	if (!etnaviv_obj->sgt) {
sgt               104 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		struct sg_table *sgt;
sgt               106 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		sgt = drm_prime_pages_to_sg(etnaviv_obj->pages, npages);
sgt               107 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		if (IS_ERR(sgt)) {
sgt               109 drivers/gpu/drm/etnaviv/etnaviv_gem.c 				PTR_ERR(sgt));
sgt               110 drivers/gpu/drm/etnaviv/etnaviv_gem.c 			return ERR_CAST(sgt);
sgt               113 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		etnaviv_obj->sgt = sgt;
sgt               383 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	if (!etnaviv_obj->sgt) {
sgt               407 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_sync_sg_for_cpu(dev->dev, etnaviv_obj->sgt->sgl,
sgt               408 drivers/gpu/drm/etnaviv/etnaviv_gem.c 				    etnaviv_obj->sgt->nents,
sgt               424 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		dma_sync_sg_for_device(dev->dev, etnaviv_obj->sgt->sgl,
sgt               425 drivers/gpu/drm/etnaviv/etnaviv_gem.c 			etnaviv_obj->sgt->nents,
sgt               697 drivers/gpu/drm/etnaviv/etnaviv_gem.c 	if (etnaviv_obj->sgt) {
sgt               699 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		sg_free_table(etnaviv_obj->sgt);
sgt               700 drivers/gpu/drm/etnaviv/etnaviv_gem.c 		kfree(etnaviv_obj->sgt);
sgt                47 drivers/gpu/drm/etnaviv/etnaviv_gem.h 	struct sg_table *sgt;
sgt                83 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c 	drm_prime_gem_destroy(&etnaviv_obj->base, etnaviv_obj->sgt);
sgt               107 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c 	struct dma_buf_attachment *attach, struct sg_table *sgt)
sgt               122 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c 	etnaviv_obj->sgt = sgt;
sgt               129 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c 	ret = drm_prime_sg_to_page_addr_arrays(sgt, etnaviv_obj->pages,
sgt                73 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 			     struct sg_table *sgt, unsigned len, int prot)
sgt                79 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	if (!context || !sgt)
sgt                82 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt               100 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, i, j) {
sgt               110 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 				struct sg_table *sgt, unsigned len)
sgt               116 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt               135 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 			    etnaviv_obj->sgt, etnaviv_obj->base.size);
sgt               234 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	struct sg_table *sgt = etnaviv_obj->sgt;
sgt               244 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	    sgt->nents == 1 && !(etnaviv_obj->flags & ETNA_BO_FORCE_MMU)) {
sgt               247 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		iova = sg_dma_address(sgt->sgl) - memory_base;
sgt               248 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) {
sgt               268 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 	ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size,
sgt               209 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct sg_table		*sgt;
sgt               398 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	dma_unmap_sg(to_dma_dev(g2d->drm_dev), g2d_userptr->sgt->sgl,
sgt               399 drivers/gpu/drm/exynos/exynos_drm_g2d.c 			g2d_userptr->sgt->nents, DMA_BIDIRECTIONAL);
sgt               414 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	sg_free_table(g2d_userptr->sgt);
sgt               415 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	kfree(g2d_userptr->sgt);
sgt               427 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	struct sg_table	*sgt;
sgt               498 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
sgt               499 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (!sgt) {
sgt               504 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	ret = sg_alloc_table_from_pages(sgt,
sgt               512 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_userptr->sgt = sgt;
sgt               514 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	if (!dma_map_sg(to_dma_dev(g2d->drm_dev), sgt->sgl, sgt->nents,
sgt               521 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_userptr->dma_addr = sgt->sgl[0].dma_address;
sgt               536 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	sg_free_table(sgt);
sgt               539 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	kfree(sgt);
sgt                25 drivers/gpu/drm/exynos/exynos_drm_gem.c 	struct sg_table sgt;
sgt                73 drivers/gpu/drm/exynos/exynos_drm_gem.c 	ret = dma_get_sgtable_attrs(to_dma_dev(dev), &sgt, exynos_gem->cookie,
sgt                81 drivers/gpu/drm/exynos/exynos_drm_gem.c 	if (drm_prime_sg_to_page_addr_arrays(&sgt, exynos_gem->pages, NULL,
sgt                88 drivers/gpu/drm/exynos/exynos_drm_gem.c 	sg_free_table(&sgt);
sgt                96 drivers/gpu/drm/exynos/exynos_drm_gem.c 	sg_free_table(&sgt);
sgt               161 drivers/gpu/drm/exynos/exynos_drm_gem.c 		drm_prime_gem_destroy(obj, exynos_gem->sgt);
sgt               475 drivers/gpu/drm/exynos/exynos_drm_gem.c 				     struct sg_table *sgt)
sgt               487 drivers/gpu/drm/exynos/exynos_drm_gem.c 	exynos_gem->dma_addr = sg_dma_address(sgt->sgl);
sgt               496 drivers/gpu/drm/exynos/exynos_drm_gem.c 	ret = drm_prime_sg_to_page_addr_arrays(sgt, exynos_gem->pages, NULL,
sgt               501 drivers/gpu/drm/exynos/exynos_drm_gem.c 	exynos_gem->sgt = sgt;
sgt               503 drivers/gpu/drm/exynos/exynos_drm_gem.c 	if (sgt->nents == 1) {
sgt                52 drivers/gpu/drm/exynos/exynos_drm_gem.h 	struct sg_table		*sgt;
sgt               117 drivers/gpu/drm/exynos/exynos_drm_gem.h 				     struct sg_table *sgt);
sgt               231 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	struct sg_table *sgt = obj->mm.pages;
sgt               242 drivers/gpu/drm/i915/gem/i915_gem_pages.c 		return kmap(sg_page(sgt->sgl));
sgt               251 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	for_each_sgt_page(page, sgt_iter, sgt)
sgt               623 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct sgt_iter sgt;
sgt               626 drivers/gpu/drm/i915/selftests/i915_vma.c 	for_each_sgt_dma(dma, sgt, vma->pages) {
sgt               116 drivers/gpu/drm/i915/selftests/scatterlist.c 	struct sgt_iter sgt;
sgt               121 drivers/gpu/drm/i915/selftests/scatterlist.c 	for_each_sgt_page(page, sgt, &pt->st) {
sgt                16 drivers/gpu/drm/lima/lima_gem_prime.c 	struct sg_table *sgt)
sgt                21 drivers/gpu/drm/lima/lima_gem_prime.c 	bo = lima_bo_create(ldev, attach->dmabuf->size, 0, sgt);
sgt                 9 drivers/gpu/drm/lima/lima_gem_prime.h 	struct sg_table *sgt);
sgt                12 drivers/gpu/drm/lima/lima_object.c 	if (bo->sgt) {
sgt                14 drivers/gpu/drm/lima/lima_object.c 		drm_prime_gem_destroy(&bo->gem, bo->sgt);
sgt                60 drivers/gpu/drm/lima/lima_object.c 			       u32 flags, struct sg_table *sgt)
sgt                78 drivers/gpu/drm/lima/lima_object.c 	if (sgt) {
sgt                79 drivers/gpu/drm/lima/lima_object.c 		bo->sgt = sgt;
sgt                88 drivers/gpu/drm/lima/lima_object.c 			sgt, bo->pages, bo->pages_dma_addr, npages);
sgt                16 drivers/gpu/drm/lima/lima_object.h 	struct sg_table *sgt;
sgt                30 drivers/gpu/drm/lima/lima_object.h 			       u32 flags, struct sg_table *sgt);
sgt               192 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	struct sg_table *sgt;
sgt               195 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
sgt               196 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	if (!sgt)
sgt               199 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie,
sgt               204 drivers/gpu/drm/mediatek/mtk_drm_gem.c 		kfree(sgt);
sgt               208 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	return sgt;
sgt               248 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	struct sg_table *sgt;
sgt               256 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	sgt = mtk_gem_prime_get_sg_table(obj);
sgt               257 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	if (IS_ERR(sgt))
sgt               265 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) {
sgt               274 drivers/gpu/drm/mediatek/mtk_drm_gem.c 	kfree((void *)sgt);
sgt               243 drivers/gpu/drm/msm/msm_drv.h 		struct sg_table *sgt, int npages);
sgt               323 drivers/gpu/drm/msm/msm_drv.h 		struct dma_buf *dmabuf, struct sg_table *sgt);
sgt                56 drivers/gpu/drm/msm/msm_gem.c 		dma_sync_sg_for_device(dev, msm_obj->sgt->sgl,
sgt                57 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
sgt                59 drivers/gpu/drm/msm/msm_gem.c 		dma_map_sg(dev, msm_obj->sgt->sgl,
sgt                60 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
sgt                69 drivers/gpu/drm/msm/msm_gem.c 		dma_sync_sg_for_cpu(dev, msm_obj->sgt->sgl,
sgt                70 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
sgt                72 drivers/gpu/drm/msm/msm_gem.c 		dma_unmap_sg(dev, msm_obj->sgt->sgl,
sgt                73 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt->nents, DMA_BIDIRECTIONAL);
sgt               129 drivers/gpu/drm/msm/msm_gem.c 		msm_obj->sgt = drm_prime_pages_to_sg(p, npages);
sgt               130 drivers/gpu/drm/msm/msm_gem.c 		if (IS_ERR(msm_obj->sgt)) {
sgt               131 drivers/gpu/drm/msm/msm_gem.c 			void *ptr = ERR_CAST(msm_obj->sgt);
sgt               134 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt = NULL;
sgt               165 drivers/gpu/drm/msm/msm_gem.c 		if (msm_obj->sgt) {
sgt               173 drivers/gpu/drm/msm/msm_gem.c 			sg_free_table(msm_obj->sgt);
sgt               174 drivers/gpu/drm/msm/msm_gem.c 			kfree(msm_obj->sgt);
sgt               443 drivers/gpu/drm/msm/msm_gem.c 			msm_obj->sgt, obj->size >> PAGE_SHIFT);
sgt               919 drivers/gpu/drm/msm/msm_gem.c 		drm_prime_gem_destroy(obj, msm_obj->sgt);
sgt              1109 drivers/gpu/drm/msm/msm_gem.c 		struct dma_buf *dmabuf, struct sg_table *sgt)
sgt              1134 drivers/gpu/drm/msm/msm_gem.c 	msm_obj->sgt = sgt;
sgt              1142 drivers/gpu/drm/msm/msm_gem.c 	ret = drm_prime_sg_to_page_addr_arrays(sgt, msm_obj->pages, NULL, npages);
sgt                71 drivers/gpu/drm/msm/msm_gem.h 	struct sg_table *sgt;
sgt                96 drivers/gpu/drm/msm/msm_gem.h 	return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt &&
sgt                61 drivers/gpu/drm/msm/msm_gem_vma.c 		struct sg_table *sgt, int npages)
sgt                78 drivers/gpu/drm/msm/msm_gem_vma.c 		ret = aspace->mmu->funcs->map(aspace->mmu, vma->iova, sgt,
sgt                36 drivers/gpu/drm/msm/msm_gpummu.c 		struct sg_table *sgt, unsigned len, int prot)
sgt                49 drivers/gpu/drm/msm/msm_gpummu.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt                43 drivers/gpu/drm/msm/msm_iommu.c 		struct sg_table *sgt, unsigned len, int prot)
sgt                48 drivers/gpu/drm/msm/msm_iommu.c 	ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot);
sgt                15 drivers/gpu/drm/msm/msm_mmu.h 	int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt,
sgt                76 drivers/gpu/drm/omapdrm/omap_gem.c 	struct sg_table *sgt;
sgt               156 drivers/gpu/drm/omapdrm/omap_gem.c 	if ((omap_obj->flags & OMAP_BO_MEM_DMABUF) && omap_obj->sgt->nents == 1)
sgt              1111 drivers/gpu/drm/omapdrm/omap_gem.c 		drm_prime_gem_destroy(obj, omap_obj->sgt);
sgt              1229 drivers/gpu/drm/omapdrm/omap_gem.c 					   struct sg_table *sgt)
sgt              1237 drivers/gpu/drm/omapdrm/omap_gem.c 	if (sgt->orig_nents != 1 && !priv->has_dmm)
sgt              1249 drivers/gpu/drm/omapdrm/omap_gem.c 	omap_obj->sgt = sgt;
sgt              1251 drivers/gpu/drm/omapdrm/omap_gem.c 	if (sgt->orig_nents == 1) {
sgt              1252 drivers/gpu/drm/omapdrm/omap_gem.c 		omap_obj->dma_addr = sg_dma_address(sgt->sgl);
sgt              1270 drivers/gpu/drm/omapdrm/omap_gem.c 		for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) {
sgt                48 drivers/gpu/drm/omapdrm/omap_gem.h 		struct sg_table *sgt);
sgt               152 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	struct sg_table *sgt;
sgt               173 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	sgt = dma_buf_map_attachment(attach, DMA_TO_DEVICE);
sgt               174 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	if (IS_ERR(sgt)) {
sgt               175 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 		ret = PTR_ERR(sgt);
sgt               179 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	obj = omap_gem_new_dmabuf(dev, dma_buf->size, sgt);
sgt               190 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c 	dma_buf_unmap_attachment(attach, sgt, DMA_TO_DEVICE);
sgt               274 drivers/gpu/drm/panfrost/panfrost_gem.c 				   struct sg_table *sgt)
sgt               279 drivers/gpu/drm/panfrost/panfrost_gem.c 	obj = drm_gem_shmem_prime_import_sg_table(dev, attach, sgt);
sgt                69 drivers/gpu/drm/panfrost/panfrost_gem.h 				   struct sg_table *sgt);
sgt               249 drivers/gpu/drm/panfrost/panfrost_mmu.c 		      u64 iova, int prot, struct sg_table *sgt)
sgt               256 drivers/gpu/drm/panfrost/panfrost_mmu.c 	for_each_sg(sgt->sgl, sgl, sgt->nents, count) {
sgt               282 drivers/gpu/drm/panfrost/panfrost_mmu.c 	struct sg_table *sgt;
sgt               291 drivers/gpu/drm/panfrost/panfrost_mmu.c 	sgt = drm_gem_shmem_get_pages_sgt(obj);
sgt               292 drivers/gpu/drm/panfrost/panfrost_mmu.c 	if (WARN_ON(IS_ERR(sgt)))
sgt               293 drivers/gpu/drm/panfrost/panfrost_mmu.c 		return PTR_ERR(sgt);
sgt               296 drivers/gpu/drm/panfrost/panfrost_mmu.c 		   prot, sgt);
sgt               454 drivers/gpu/drm/panfrost/panfrost_mmu.c 	struct sg_table *sgt;
sgt               514 drivers/gpu/drm/panfrost/panfrost_mmu.c 	sgt = &bo->sgts[page_offset / (SZ_2M / PAGE_SIZE)];
sgt               515 drivers/gpu/drm/panfrost/panfrost_mmu.c 	ret = sg_alloc_table_from_pages(sgt, pages + page_offset,
sgt               520 drivers/gpu/drm/panfrost/panfrost_mmu.c 	if (!dma_map_sg(pfdev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL)) {
sgt               526 drivers/gpu/drm/panfrost/panfrost_mmu.c 		   IOMMU_WRITE | IOMMU_READ | IOMMU_NOEXEC, sgt);
sgt               537 drivers/gpu/drm/panfrost/panfrost_mmu.c 	sg_free_table(sgt);
sgt               209 drivers/gpu/drm/pl111/pl111_drv.c 			  struct sg_table *sgt)
sgt               221 drivers/gpu/drm/pl111/pl111_drv.c 	return drm_gem_cma_prime_import_sg_table(dev, attach, sgt);
sgt               452 drivers/gpu/drm/qxl/qxl_drv.h 	struct sg_table *sgt);
sgt               192 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		struct sg_table *sgt = &sg_tables[i];
sgt               194 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		ret = dma_get_sgtable(rcdu->dev, sgt, gem->vaddr, gem->paddr,
sgt               199 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		ret = vsp1_du_map_sg(vsp->vsp, sgt);
sgt               201 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 			sg_free_table(sgt);
sgt               211 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		struct sg_table *sgt = &sg_tables[i];
sgt               213 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		vsp1_du_unmap_sg(vsp->vsp, sgt);
sgt               214 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		sg_free_table(sgt);
sgt               247 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		struct sg_table *sgt = &sg_tables[i];
sgt               249 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		vsp1_du_unmap_sg(vsp->vsp, sgt);
sgt               250 drivers/gpu/drm/rcar-du/rcar_du_vsp.c 		sg_free_table(sgt);
sgt                38 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl,
sgt                39 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 			   rk_obj->sgt->nents, prot);
sgt                87 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages);
sgt                88 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	if (IS_ERR(rk_obj->sgt)) {
sgt                89 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 		ret = PTR_ERR(rk_obj->sgt);
sgt               100 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	for_each_sg(rk_obj->sgt->sgl, s, rk_obj->sgt->nents, i)
sgt               103 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents,
sgt               115 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	sg_free_table(rk_obj->sgt);
sgt               116 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	kfree(rk_obj->sgt);
sgt               352 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 			dma_unmap_sg(drm->dev, rk_obj->sgt->sgl,
sgt               353 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 				     rk_obj->sgt->nents, DMA_BIDIRECTIONAL);
sgt               355 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 		drm_prime_gem_destroy(obj, rk_obj->sgt);
sgt               440 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	struct sg_table *sgt;
sgt               446 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
sgt               447 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	if (!sgt)
sgt               450 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	ret = dma_get_sgtable_attrs(drm->dev, sgt, rk_obj->kvaddr,
sgt               455 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 		kfree(sgt);
sgt               459 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	return sgt;
sgt               462 drivers/gpu/drm/rockchip/rockchip_drm_gem.c static unsigned long rockchip_sg_get_contiguous_size(struct sg_table *sgt,
sgt               466 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	dma_addr_t expected = sg_dma_address(sgt->sgl);
sgt               470 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	for_each_sg(sgt->sgl, s, count, i) {
sgt               485 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	rk_obj->sgt = sg;
sgt               508 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	rk_obj->sgt = sg;
sgt                25 drivers/gpu/drm/rockchip/rockchip_drm_gem.h 	struct sg_table *sgt;
sgt                30 drivers/gpu/drm/tegra/gem.c static dma_addr_t tegra_bo_pin(struct host1x_bo *bo, struct sg_table **sgt)
sgt                34 drivers/gpu/drm/tegra/gem.c 	*sgt = obj->sgt;
sgt                39 drivers/gpu/drm/tegra/gem.c static void tegra_bo_unpin(struct host1x_bo *bo, struct sg_table *sgt)
sgt               138 drivers/gpu/drm/tegra/gem.c 	bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl,
sgt               139 drivers/gpu/drm/tegra/gem.c 				bo->sgt->nents, prot);
sgt               206 drivers/gpu/drm/tegra/gem.c 		dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgt               209 drivers/gpu/drm/tegra/gem.c 		sg_free_table(bo->sgt);
sgt               210 drivers/gpu/drm/tegra/gem.c 		kfree(bo->sgt);
sgt               226 drivers/gpu/drm/tegra/gem.c 	bo->sgt = drm_prime_pages_to_sg(bo->pages, bo->num_pages);
sgt               227 drivers/gpu/drm/tegra/gem.c 	if (IS_ERR(bo->sgt)) {
sgt               228 drivers/gpu/drm/tegra/gem.c 		err = PTR_ERR(bo->sgt);
sgt               232 drivers/gpu/drm/tegra/gem.c 	err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgt               242 drivers/gpu/drm/tegra/gem.c 	sg_free_table(bo->sgt);
sgt               243 drivers/gpu/drm/tegra/gem.c 	kfree(bo->sgt);
sgt               352 drivers/gpu/drm/tegra/gem.c 	bo->sgt = dma_buf_map_attachment(attach, DMA_TO_DEVICE);
sgt               353 drivers/gpu/drm/tegra/gem.c 	if (IS_ERR(bo->sgt)) {
sgt               354 drivers/gpu/drm/tegra/gem.c 		err = PTR_ERR(bo->sgt);
sgt               363 drivers/gpu/drm/tegra/gem.c 		if (bo->sgt->nents > 1) {
sgt               368 drivers/gpu/drm/tegra/gem.c 		bo->paddr = sg_dma_address(bo->sgt->sgl);
sgt               376 drivers/gpu/drm/tegra/gem.c 	if (!IS_ERR_OR_NULL(bo->sgt))
sgt               377 drivers/gpu/drm/tegra/gem.c 		dma_buf_unmap_attachment(attach, bo->sgt, DMA_TO_DEVICE);
sgt               396 drivers/gpu/drm/tegra/gem.c 		dma_buf_unmap_attachment(gem->import_attach, bo->sgt,
sgt               504 drivers/gpu/drm/tegra/gem.c 	struct sg_table *sgt;
sgt               506 drivers/gpu/drm/tegra/gem.c 	sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
sgt               507 drivers/gpu/drm/tegra/gem.c 	if (!sgt)
sgt               514 drivers/gpu/drm/tegra/gem.c 		if (sg_alloc_table(sgt, bo->num_pages, GFP_KERNEL))
sgt               517 drivers/gpu/drm/tegra/gem.c 		for_each_sg(sgt->sgl, sg, bo->num_pages, i)
sgt               520 drivers/gpu/drm/tegra/gem.c 		if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0)
sgt               523 drivers/gpu/drm/tegra/gem.c 		if (sg_alloc_table(sgt, 1, GFP_KERNEL))
sgt               526 drivers/gpu/drm/tegra/gem.c 		sg_dma_address(sgt->sgl) = bo->paddr;
sgt               527 drivers/gpu/drm/tegra/gem.c 		sg_dma_len(sgt->sgl) = gem->size;
sgt               530 drivers/gpu/drm/tegra/gem.c 	return sgt;
sgt               533 drivers/gpu/drm/tegra/gem.c 	sg_free_table(sgt);
sgt               534 drivers/gpu/drm/tegra/gem.c 	kfree(sgt);
sgt               539 drivers/gpu/drm/tegra/gem.c 					  struct sg_table *sgt,
sgt               546 drivers/gpu/drm/tegra/gem.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir);
sgt               548 drivers/gpu/drm/tegra/gem.c 	sg_free_table(sgt);
sgt               549 drivers/gpu/drm/tegra/gem.c 	kfree(sgt);
sgt               565 drivers/gpu/drm/tegra/gem.c 		dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgt               579 drivers/gpu/drm/tegra/gem.c 		dma_sync_sg_for_device(drm->dev, bo->sgt->sgl, bo->sgt->nents,
sgt                33 drivers/gpu/drm/tegra/gem.h 	struct sg_table *sgt;
sgt                16 drivers/gpu/drm/udl/udl_dmabuf.c 	struct sg_table sgt;
sgt                43 drivers/gpu/drm/udl/udl_dmabuf.c 	struct sg_table *sgt;
sgt                51 drivers/gpu/drm/udl/udl_dmabuf.c 	sgt = &udl_attach->sgt;
sgt                54 drivers/gpu/drm/udl/udl_dmabuf.c 		dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents,
sgt                57 drivers/gpu/drm/udl/udl_dmabuf.c 	sg_free_table(sgt);
sgt                70 drivers/gpu/drm/udl/udl_dmabuf.c 	struct sg_table *sgt = NULL;
sgt                80 drivers/gpu/drm/udl/udl_dmabuf.c 		return &udl_attach->sgt;
sgt                97 drivers/gpu/drm/udl/udl_dmabuf.c 	sgt = &udl_attach->sgt;
sgt                99 drivers/gpu/drm/udl/udl_dmabuf.c 	ret = sg_alloc_table(sgt, obj->sg->orig_nents, GFP_KERNEL);
sgt               108 drivers/gpu/drm/udl/udl_dmabuf.c 	wr = sgt->sgl;
sgt               109 drivers/gpu/drm/udl/udl_dmabuf.c 	for (i = 0; i < sgt->orig_nents; ++i) {
sgt               116 drivers/gpu/drm/udl/udl_dmabuf.c 		nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir);
sgt               119 drivers/gpu/drm/udl/udl_dmabuf.c 			sg_free_table(sgt);
sgt               120 drivers/gpu/drm/udl/udl_dmabuf.c 			sgt = ERR_PTR(-EIO);
sgt               131 drivers/gpu/drm/udl/udl_dmabuf.c 	return sgt;
sgt               135 drivers/gpu/drm/udl/udl_dmabuf.c 			      struct sg_table *sgt,
sgt                92 drivers/gpu/drm/v3d/v3d_bo.c 	struct sg_table *sgt;
sgt                98 drivers/gpu/drm/v3d/v3d_bo.c 	sgt = drm_gem_shmem_get_pages_sgt(&bo->base.base);
sgt                99 drivers/gpu/drm/v3d/v3d_bo.c 	if (IS_ERR(sgt))
sgt               100 drivers/gpu/drm/v3d/v3d_bo.c 		return PTR_ERR(sgt);
sgt               151 drivers/gpu/drm/v3d/v3d_bo.c 			  struct sg_table *sgt)
sgt               156 drivers/gpu/drm/v3d/v3d_bo.c 	obj = drm_gem_shmem_prime_import_sg_table(dev, attach, sgt);
sgt               301 drivers/gpu/drm/v3d/v3d_drv.h 						 struct sg_table *sgt);
sgt                96 drivers/gpu/drm/v3d/v3d_mmu.c 	for_each_sg(shmem_obj->sgt->sgl, sgl, shmem_obj->sgt->nents, count) {
sgt               785 drivers/gpu/drm/vc4/vc4_bo.c 			  struct sg_table *sgt)
sgt               789 drivers/gpu/drm/vc4/vc4_bo.c 	obj = drm_gem_cma_prime_import_sg_table(dev, attach, sgt);
sgt               735 drivers/gpu/drm/vc4/vc4_drv.h 						 struct sg_table *sgt);
sgt               373 drivers/gpu/drm/virtio/virtgpu_drv.h 	struct sg_table *sgt);
sgt               299 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	struct sg_table *sgt;
sgt               245 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	struct sg_table sgt;
sgt               345 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		__sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl,
sgt               346 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 				     vsgt->sgt->orig_nents, p_offset);
sgt               365 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents,
sgt               367 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	vmw_tt->sgt.nents = vmw_tt->sgt.orig_nents;
sgt               388 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	ret = dma_map_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.orig_nents,
sgt               393 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	vmw_tt->sgt.nents = ret;
sgt               430 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	vsgt->sgt = &vmw_tt->sgt;
sgt               445 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 			(&vmw_tt->sgt, vsgt->pages, vsgt->num_pages, 0,
sgt               452 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		if (vsgt->num_pages > vmw_tt->sgt.nents) {
sgt               455 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 					    vmw_tt->sgt.nents);
sgt               484 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	sg_free_table(vmw_tt->vsgt.sgt);
sgt               485 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	vmw_tt->vsgt.sgt = NULL;
sgt               504 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	if (!vmw_tt->vsgt.sgt)
sgt               511 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		sg_free_table(vmw_tt->vsgt.sgt);
sgt               512 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		vmw_tt->vsgt.sgt = NULL;
sgt               188 drivers/gpu/drm/xen/xen_drm_front_gem.c 				  struct sg_table *sgt)
sgt               204 drivers/gpu/drm/xen/xen_drm_front_gem.c 	xen_obj->sgt_imported = sgt;
sgt               206 drivers/gpu/drm/xen/xen_drm_front_gem.c 	ret = drm_prime_sg_to_page_addr_arrays(sgt, xen_obj->pages,
sgt               218 drivers/gpu/drm/xen/xen_drm_front_gem.c 		  size, sgt->nents);
sgt                27 drivers/gpu/drm/xen/xen_drm_front_gem.h 				  struct sg_table *sgt);
sgt               109 drivers/gpu/host1x/job.c 		struct sg_table *sgt;
sgt               118 drivers/gpu/host1x/job.c 		phys_addr = host1x_bo_pin(reloc->target.bo, &sgt);
sgt               122 drivers/gpu/host1x/job.c 		job->unpins[job->num_unpins].sgt = sgt;
sgt               130 drivers/gpu/host1x/job.c 		struct sg_table *sgt;
sgt               142 drivers/gpu/host1x/job.c 		phys_addr = host1x_bo_pin(g->bo, &sgt);
sgt               145 drivers/gpu/host1x/job.c 			for_each_sg(sgt->sgl, sg, sgt->nents, j)
sgt               159 drivers/gpu/host1x/job.c 					sgt->sgl, sgt->nents, IOMMU_READ);
sgt               176 drivers/gpu/host1x/job.c 		job->unpins[job->num_unpins].sgt = sgt;
sgt               570 drivers/gpu/host1x/job.c 		host1x_bo_unpin(unpin->bo, unpin->sgt);
sgt                21 drivers/gpu/host1x/job.h 	struct sg_table *sgt;
sgt               199 drivers/hsi/clients/cmt_speech.c 	u32 *data = sg_virt(msg->sgt.sgl);
sgt               205 drivers/hsi/clients/cmt_speech.c 	u32 *data = sg_virt(msg->sgt.sgl);
sgt               259 drivers/hsi/clients/cmt_speech.c 		kfree(sg_virt(msg->sgt.sgl));
sgt               281 drivers/hsi/clients/cmt_speech.c 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
sgt               472 drivers/hsi/clients/cmt_speech.c 	msg->sgt.nents = 1;
sgt               502 drivers/hsi/clients/cmt_speech.c 	msg->sgt.nents = 0;
sgt               547 drivers/hsi/clients/cmt_speech.c 	msg->sgt.nents = 1;
sgt               624 drivers/hsi/clients/cmt_speech.c 	sg_init_one(msg->sgt.sgl, address, hi->buf_size);
sgt               625 drivers/hsi/clients/cmt_speech.c 	msg->sgt.nents = 1;
sgt               667 drivers/hsi/clients/cmt_speech.c 	sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0);
sgt               668 drivers/hsi/clients/cmt_speech.c 	rxmsg->sgt.nents = 0;
sgt               719 drivers/hsi/clients/cmt_speech.c 	sg_init_one(txmsg->sgt.sgl, address, hi->buf_size);
sgt               147 drivers/hsi/clients/hsi_char.c 	kfree(sg_virt(msg->sgt.sgl));
sgt               186 drivers/hsi/clients/hsi_char.c 	sg_init_one(msg->sgt.sgl, buf, alloc_size);
sgt               217 drivers/hsi/clients/hsi_char.c 	return msg->sgt.sgl->length;
sgt               222 drivers/hsi/clients/hsi_char.c 	msg->sgt.sgl->length = len;
sgt               455 drivers/hsi/clients/hsi_char.c 			sg_virt(msg->sgt.sgl), hsc_msg_len_get(msg));
sgt               491 drivers/hsi/clients/hsi_char.c 	if (copy_from_user(sg_virt(msg->sgt.sgl), (void __user *)buf, len)) {
sgt               157 drivers/hsi/clients/ssi_protocol.c 	data = sg_virt(msg->sgt.sgl);
sgt               165 drivers/hsi/clients/ssi_protocol.c 	data = sg_virt(msg->sgt.sgl);
sgt               176 drivers/hsi/clients/ssi_protocol.c 	BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1));
sgt               178 drivers/hsi/clients/ssi_protocol.c 	sg = msg->sgt.sgl;
sgt               249 drivers/hsi/clients/ssi_protocol.c 		kfree(sg_virt(msg->sgt.sgl));
sgt               269 drivers/hsi/clients/ssi_protocol.c 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
sgt               192 drivers/hsi/controllers/omap_ssi_core.c 	dma_unmap_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, dir);
sgt               215 drivers/hsi/controllers/omap_ssi_core.c 	msg->actual_len = sg_dma_len(msg->sgt.sgl);
sgt               200 drivers/hsi/controllers/omap_ssi_port.c 			omap_ssi->gdd_trn[lch].sg = msg->sgt.sgl;
sgt               231 drivers/hsi/controllers/omap_ssi_port.c 		err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents,
sgt               246 drivers/hsi/controllers/omap_ssi_port.c 		d_addr = sg_dma_address(msg->sgt.sgl);
sgt               248 drivers/hsi/controllers/omap_ssi_port.c 		err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents,
sgt               261 drivers/hsi/controllers/omap_ssi_port.c 		s_addr = sg_dma_address(msg->sgt.sgl);
sgt               272 drivers/hsi/controllers/omap_ssi_port.c 	writew_relaxed(SSI_BYTES_TO_FRAMES(msg->sgt.sgl->length),
sgt               330 drivers/hsi/controllers/omap_ssi_port.c 	if ((msg->sgt.nents) && (msg->sgt.sgl->length > sizeof(u32)))
sgt               386 drivers/hsi/controllers/omap_ssi_port.c 	if (msg->sgt.nents > 1)
sgt               445 drivers/hsi/controllers/omap_ssi_port.c 			msg->channel, msg, msg->sgt.sgl->length,
sgt               905 drivers/hsi/controllers/omap_ssi_port.c 	if ((!msg->sgt.nents) || (!msg->sgt.sgl->length)) {
sgt               914 drivers/hsi/controllers/omap_ssi_port.c 		buf = sg_virt(msg->sgt.sgl) + msg->actual_len;
sgt               924 drivers/hsi/controllers/omap_ssi_port.c 		if (msg->actual_len >= msg->sgt.sgl->length)
sgt               520 drivers/hsi/hsi_core.c 	sg_free_table(&msg->sgt);
sgt               548 drivers/hsi/hsi_core.c 	err = sg_alloc_table(&msg->sgt, nents, flags);
sgt                51 drivers/hwtracing/intel_th/msu-sink.c static int msu_sink_alloc_window(void *data, struct sg_table **sgt, size_t size)
sgt                64 drivers/hwtracing/intel_th/msu-sink.c 	ret = sg_alloc_table(*sgt, nents, GFP_KERNEL);
sgt                68 drivers/hwtracing/intel_th/msu-sink.c 	priv->sgts[priv->nr_sgts++] = *sgt;
sgt                70 drivers/hwtracing/intel_th/msu-sink.c 	for_each_sg((*sgt)->sgl, sg_ptr, nents, i) {
sgt                81 drivers/hwtracing/intel_th/msu-sink.c static void msu_sink_free_window(void *data, struct sg_table *sgt)
sgt                87 drivers/hwtracing/intel_th/msu-sink.c 	for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) {
sgt                92 drivers/hwtracing/intel_th/msu-sink.c 	sg_free_table(sgt);
sgt                96 drivers/hwtracing/intel_th/msu-sink.c static int msu_sink_ready(void *data, struct sg_table *sgt, size_t bytes)
sgt               100 drivers/hwtracing/intel_th/msu-sink.c 	intel_th_msc_window_unlock(priv->dev, sgt);
sgt                76 drivers/hwtracing/intel_th/msu.c 	struct sg_table		*sgt;
sgt               276 drivers/hwtracing/intel_th/msu.c 	return win->sgt->sgl;
sgt               326 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgt               350 drivers/hwtracing/intel_th/msu.c msc_find_window(struct msc *msc, struct sg_table *sgt, bool nonempty)
sgt               364 drivers/hwtracing/intel_th/msu.c 		if (win->sgt == sgt)
sgt               394 drivers/hwtracing/intel_th/msu.c 	win = msc_find_window(msc, msc_next_window(msc->cur_win)->sgt, true);
sgt               421 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgt               660 drivers/hwtracing/intel_th/msu.c 		for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgt               845 drivers/hwtracing/intel_th/msu.c 		msc->mbuf->ready(msc->mbuf_priv, win->sgt,
sgt               996 drivers/hwtracing/intel_th/msu.c 	ret = sg_alloc_table(win->sgt, nr_segs, GFP_KERNEL);
sgt              1000 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) {
sgt              1013 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, i, ret)
sgt              1017 drivers/hwtracing/intel_th/msu.c 	sg_free_table(win->sgt);
sgt              1028 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) {
sgt              1040 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) {
sgt              1075 drivers/hwtracing/intel_th/msu.c 	win->sgt = &win->_sgt;
sgt              1088 drivers/hwtracing/intel_th/msu.c 		ret = msc->mbuf->alloc_window(msc->mbuf_priv, &win->sgt,
sgt              1123 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, i) {
sgt              1130 drivers/hwtracing/intel_th/msu.c 	sg_free_table(win->sgt);
sgt              1154 drivers/hwtracing/intel_th/msu.c 		msc->mbuf->free_window(msc->mbuf_priv, win->sgt);
sgt              1190 drivers/hwtracing/intel_th/msu.c 		for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgt              1385 drivers/hwtracing/intel_th/msu.c 	for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
sgt              1707 drivers/hwtracing/intel_th/msu.c void intel_th_msc_window_unlock(struct device *dev, struct sg_table *sgt)
sgt              1712 drivers/hwtracing/intel_th/msu.c 	if (!sgt)
sgt              1715 drivers/hwtracing/intel_th/msu.c 	win = msc_find_window(msc, sgt, false);
sgt              1770 drivers/hwtracing/intel_th/msu.c 		msc->mbuf->ready(msc->mbuf_priv, win->sgt,
sgt               533 drivers/input/touchscreen/sur40.c 	struct sg_table *sgt;
sgt               581 drivers/input/touchscreen/sur40.c 	sgt = vb2_dma_sg_plane_desc(&new_buf->vb.vb2_buf, 0);
sgt               585 drivers/input/touchscreen/sur40.c 		sgt->sgl, sgt->nents, sur40->pix_fmt.sizeimage, 0);
sgt               575 drivers/iommu/dma-iommu.c 	struct sg_table sgt;
sgt               602 drivers/iommu/dma-iommu.c 	if (sg_alloc_table_from_pages(&sgt, pages, count, 0, size, GFP_KERNEL))
sgt               609 drivers/iommu/dma-iommu.c 		for_each_sg(sgt.sgl, sg, sgt.orig_nents, i)
sgt               613 drivers/iommu/dma-iommu.c 	if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, ioprot)
sgt               623 drivers/iommu/dma-iommu.c 	sg_free_table(&sgt);
sgt               629 drivers/iommu/dma-iommu.c 	sg_free_table(&sgt);
sgt              1059 drivers/iommu/dma-iommu.c static int iommu_dma_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt              1070 drivers/iommu/dma-iommu.c 			return sg_alloc_table_from_pages(sgt, pages,
sgt              1080 drivers/iommu/dma-iommu.c 	ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
sgt              1082 drivers/iommu/dma-iommu.c 		sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgt                49 drivers/media/common/videobuf2/videobuf2-dma-contig.c static unsigned long vb2_dc_get_contiguous_size(struct sg_table *sgt)
sgt                52 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_addr_t expected = sg_dma_address(sgt->sgl);
sgt                56 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	for_each_sg(sgt->sgl, s, sgt->nents, i) {
sgt                96 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt = buf->dma_sgt;
sgt                99 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (!sgt || buf->db_attach)
sgt               102 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               109 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               112 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (!sgt || buf->db_attach)
sgt               115 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir);
sgt               215 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table sgt;
sgt               225 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               233 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = &attach->sgt;
sgt               237 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	ret = sg_alloc_table(sgt, buf->sgt_base->orig_nents, GFP_KERNEL);
sgt               244 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	wr = sgt->sgl;
sgt               245 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	for (i = 0; i < sgt->orig_nents; ++i) {
sgt               261 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               266 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = &attach->sgt;
sgt               276 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               278 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sg_free_table(sgt);
sgt               289 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               293 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = &attach->sgt;
sgt               297 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		return sgt;
sgt               302 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               311 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt->nents = dma_map_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               313 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (!sgt->nents) {
sgt               323 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	return sgt;
sgt               327 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt, enum dma_data_direction dma_dir)
sgt               372 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               374 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
sgt               375 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (!sgt) {
sgt               380 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	ret = dma_get_sgtable_attrs(buf->dev, sgt, buf->cookie, buf->dma_addr,
sgt               384 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		kfree(sgt);
sgt               388 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	return sgt;
sgt               425 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               429 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (sgt) {
sgt               434 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               443 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		sg_free_table(sgt);
sgt               444 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		kfree(sgt);
sgt               461 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               514 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
sgt               515 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (!sgt) {
sgt               521 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	ret = sg_alloc_table_from_pages(sgt, frame_vector_pages(vec), n_pages,
sgt               532 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               534 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (sgt->nents <= 0) {
sgt               540 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	contig_size = vb2_dc_get_contiguous_size(sgt);
sgt               548 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgt               549 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_sgt = sgt;
sgt               556 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               560 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sg_free_table(sgt);
sgt               563 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	kfree(sgt);
sgt               581 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt;
sgt               595 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	sgt = dma_buf_map_attachment(buf->db_attach, buf->dma_dir);
sgt               596 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (IS_ERR(sgt)) {
sgt               602 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	contig_size = vb2_dc_get_contiguous_size(sgt);
sgt               606 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir);
sgt               610 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgt               611 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	buf->dma_sgt = sgt;
sgt               620 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               627 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	if (WARN_ON(!sgt)) {
sgt               636 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir);
sgt               104 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               140 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = &buf->sg_table;
sgt               145 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               147 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	if (!sgt->nents)
sgt               177 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt = &buf->sg_table;
sgt               183 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               199 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               205 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               212 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               218 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir);
sgt               226 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               256 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = &buf->sg_table;
sgt               261 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents,
sgt               263 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	if (!sgt->nents)
sgt               284 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt = &buf->sg_table;
sgt               289 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir,
sgt               359 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table sgt;
sgt               369 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               377 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = &attach->sgt;
sgt               381 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	ret = sg_alloc_table(sgt, buf->dma_sgt->orig_nents, GFP_KERNEL);
sgt               388 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	wr = sgt->sgl;
sgt               389 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	for (i = 0; i < sgt->orig_nents; ++i) {
sgt               405 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               410 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = &attach->sgt;
sgt               414 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               416 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sg_free_table(sgt);
sgt               427 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               431 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = &attach->sgt;
sgt               435 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		return sgt;
sgt               440 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               446 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               448 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	if (!sgt->nents) {
sgt               458 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	return sgt;
sgt               462 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt, enum dma_data_direction dma_dir)
sgt               535 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt;
sgt               548 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	sgt = dma_buf_map_attachment(buf->db_attach, buf->dma_dir);
sgt               549 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	if (IS_ERR(sgt)) {
sgt               554 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	buf->dma_sgt = sgt;
sgt               563 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct sg_table *sgt = buf->dma_sgt;
sgt               570 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	if (WARN_ON(!sgt)) {
sgt               579 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir);
sgt               207 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	struct sg_table sgt;
sgt               217 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	struct sg_table *sgt;
sgt               227 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sgt = &attach->sgt;
sgt               228 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL);
sgt               233 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt               237 drivers/media/common/videobuf2/videobuf2-vmalloc.c 			sg_free_table(sgt);
sgt               254 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	struct sg_table *sgt;
sgt               259 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sgt = &attach->sgt;
sgt               263 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               265 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sg_free_table(sgt);
sgt               276 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	struct sg_table *sgt;
sgt               280 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sgt = &attach->sgt;
sgt               284 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		return sgt;
sgt               289 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               295 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents,
sgt               297 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	if (!sgt->nents) {
sgt               307 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	return sgt;
sgt               311 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	struct sg_table *sgt, enum dma_data_direction dma_dir)
sgt              1578 drivers/media/pci/cx23885/cx23885-core.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0);
sgt              1586 drivers/media/pci/cx23885/cx23885-core.c 				sgt->sgl,
sgt               134 drivers/media/pci/cx23885/cx23885-vbi.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               145 drivers/media/pci/cx23885/cx23885-vbi.c 			 sgt->sgl,
sgt               347 drivers/media/pci/cx23885/cx23885-video.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               359 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, 0, UNSET,
sgt               364 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, UNSET, 0,
sgt               392 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl, line0_offset,
sgt               399 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl,
sgt               406 drivers/media/pci/cx23885/cx23885-video.c 				sgt->sgl,
sgt               149 drivers/media/pci/cx25821/cx25821-video.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               180 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, 0, UNSET,
sgt               185 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, UNSET, 0,
sgt               194 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl, line0_offset,
sgt               200 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl,
sgt               206 drivers/media/pci/cx25821/cx25821-video.c 				sgt->sgl,
sgt               217 drivers/media/pci/cx88/cx88-mpeg.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0);
sgt               225 drivers/media/pci/cx88/cx88-mpeg.c 	rc = cx88_risc_databuffer(dev->pci, risc, sgt->sgl,
sgt               134 drivers/media/pci/cx88/cx88-vbi.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               147 drivers/media/pci/cx88/cx88-vbi.c 	cx88_risc_buffer(dev->pci, &buf->risc, sgt->sgl,
sgt               440 drivers/media/pci/cx88/cx88-video.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               451 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, 0, UNSET,
sgt               456 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, UNSET, 0,
sgt               461 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl,
sgt               468 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl,
sgt               476 drivers/media/pci/cx88/cx88-video.c 				 sgt->sgl, 0, buf->bpl,
sgt               454 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               465 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	return solo_send_desc(solo_enc, solo_enc->jpeg_len, sgt,
sgt               476 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               501 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	return solo_send_desc(solo_enc, skip, sgt, frame_off, frame_size,
sgt               737 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt               743 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 			sg_copy_from_buffer(sgt->sgl, sgt->nents,
sgt               747 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c 		sg_copy_from_buffer(sgt->sgl, sgt->nents,
sgt               359 drivers/media/platform/omap3isp/ispccdc.c 		sg_free_table(&req->table.sgt);
sgt               443 drivers/media/platform/omap3isp/ispccdc.c 		ret = dma_get_sgtable(isp->dev, &req->table.sgt,
sgt               449 drivers/media/platform/omap3isp/ispccdc.c 		dma_sync_sg_for_cpu(isp->dev, req->table.sgt.sgl,
sgt               450 drivers/media/platform/omap3isp/ispccdc.c 				    req->table.sgt.nents, DMA_TO_DEVICE);
sgt               458 drivers/media/platform/omap3isp/ispccdc.c 		dma_sync_sg_for_device(isp->dev, req->table.sgt.sgl,
sgt               459 drivers/media/platform/omap3isp/ispccdc.c 				       req->table.sgt.nents, DMA_TO_DEVICE);
sgt                57 drivers/media/platform/omap3isp/ispccdc.h 		struct sg_table sgt;
sgt               164 drivers/media/platform/omap3isp/ispstat.c 	dma_sync_sg_for_device(stat->isp->dev, buf->sgt.sgl,
sgt               165 drivers/media/platform/omap3isp/ispstat.c 			       buf->sgt.nents, DMA_FROM_DEVICE);
sgt               174 drivers/media/platform/omap3isp/ispstat.c 	dma_sync_sg_for_cpu(stat->isp->dev, buf->sgt.sgl,
sgt               175 drivers/media/platform/omap3isp/ispstat.c 			    buf->sgt.nents, DMA_FROM_DEVICE);
sgt               347 drivers/media/platform/omap3isp/ispstat.c 		sg_free_table(&buf->sgt);
sgt               375 drivers/media/platform/omap3isp/ispstat.c 	ret = dma_get_sgtable(dev, &buf->sgt, buf->virt_addr, buf->dma_addr,
sgt                36 drivers/media/platform/omap3isp/ispstat.h 	struct sg_table sgt;
sgt              1375 drivers/media/platform/pxa_camera.c 	struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt              1394 drivers/media/platform/pxa_camera.c 	ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels,
sgt               356 drivers/media/platform/qcom/camss/camss-video.c 	struct sg_table *sgt;
sgt               360 drivers/media/platform/qcom/camss/camss-video.c 		sgt = vb2_dma_sg_plane_desc(vb, i);
sgt               361 drivers/media/platform/qcom/camss/camss-video.c 		if (!sgt)
sgt               364 drivers/media/platform/qcom/camss/camss-video.c 		buffer->addr[i] = sg_dma_address(sgt->sgl);
sgt              1035 drivers/media/platform/qcom/venus/helpers.c 	struct sg_table *sgt;
sgt              1037 drivers/media/platform/qcom/venus/helpers.c 	sgt = vb2_dma_sg_plane_desc(vb, 0);
sgt              1038 drivers/media/platform/qcom/venus/helpers.c 	if (!sgt)
sgt              1042 drivers/media/platform/qcom/venus/helpers.c 	buf->dma_addr = sg_dma_address(sgt->sgl);
sgt               117 drivers/media/platform/rockchip/rga/rga-buf.c 	struct sg_table *sgt;
sgt               129 drivers/media/platform/rockchip/rga/rga-buf.c 	sgt = vb2_plane_cookie(vb, 0);
sgt               131 drivers/media/platform/rockchip/rga/rga-buf.c 	for_each_sg(sgt->sgl, sgl, sgt->nents, i) {
sgt               347 drivers/media/platform/via-camera.c 	struct sg_table *sgt;
sgt               367 drivers/media/platform/via-camera.c 	sgt = vb2_dma_sg_plane_desc(&vb->vbuf.vb2_buf, 0);
sgt               369 drivers/media/platform/via-camera.c 	viafb_dma_copy_out_sg(cam->cb_offsets[bufn], sgt->sgl, sgt->nents);
sgt               906 drivers/media/platform/vsp1/vsp1_drm.c int vsp1_du_map_sg(struct device *dev, struct sg_table *sgt)
sgt               915 drivers/media/platform/vsp1/vsp1_drm.c 	return dma_map_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents,
sgt               920 drivers/media/platform/vsp1/vsp1_drm.c void vsp1_du_unmap_sg(struct device *dev, struct sg_table *sgt)
sgt               924 drivers/media/platform/vsp1/vsp1_drm.c 	dma_unmap_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents,
sgt               130 drivers/misc/fastrpc.c 	struct sg_table sgt;
sgt               493 drivers/misc/fastrpc.c 	table = &a->sgt;
sgt               526 drivers/misc/fastrpc.c 	ret = dma_get_sgtable(buffer->dev, &a->sgt, buffer->virt,
sgt               554 drivers/misc/fastrpc.c 	sg_free_table(&a->sgt);
sgt              3136 drivers/misc/habanalabs/goya/goya.c u32 goya_get_dma_desc_list_size(struct hl_device *hdev, struct sg_table *sgt)
sgt              3145 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgt              3153 drivers/misc/habanalabs/goya/goya.c 		while ((count + 1) < sgt->nents) {
sgt              3200 drivers/misc/habanalabs/goya/goya.c 	rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl,
sgt              3201 drivers/misc/habanalabs/goya/goya.c 					userptr->sgt->nents, dir);
sgt              3212 drivers/misc/habanalabs/goya/goya.c 			goya_get_dma_desc_list_size(hdev, userptr->sgt);
sgt              3603 drivers/misc/habanalabs/goya/goya.c 	struct sg_table *sgt;
sgt              3654 drivers/misc/habanalabs/goya/goya.c 	sgt = userptr->sgt;
sgt              3657 drivers/misc/habanalabs/goya/goya.c 	for_each_sg(sgt->sgl, sg, sgt->nents, count) {
sgt              3664 drivers/misc/habanalabs/goya/goya.c 		while ((count + 1) < sgt->nents) {
sgt               228 drivers/misc/habanalabs/goya/goyaP.h u32 goya_get_dma_desc_list_size(struct hl_device *hdev, struct sg_table *sgt);
sgt               546 drivers/misc/habanalabs/habanalabs.h 					struct sg_table *sgt);
sgt               699 drivers/misc/habanalabs/habanalabs.h 	struct sg_table		*sgt;
sgt               193 drivers/misc/habanalabs/memory.c 	rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl,
sgt               194 drivers/misc/habanalabs/memory.c 					userptr->sgt->nents, DMA_BIDIRECTIONAL);
sgt               673 drivers/misc/habanalabs/memory.c 	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
sgt               702 drivers/misc/habanalabs/memory.c 	for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) {
sgt              1292 drivers/misc/habanalabs/memory.c 	userptr->sgt = kzalloc(sizeof(*userptr->sgt), GFP_ATOMIC);
sgt              1293 drivers/misc/habanalabs/memory.c 	if (!userptr->sgt) {
sgt              1298 drivers/misc/habanalabs/memory.c 	rc = sg_alloc_table_from_pages(userptr->sgt,
sgt              1311 drivers/misc/habanalabs/memory.c 	kfree(userptr->sgt);
sgt              1337 drivers/misc/habanalabs/memory.c 				userptr->sgt->sgl,
sgt              1338 drivers/misc/habanalabs/memory.c 				userptr->sgt->nents,
sgt              1353 drivers/misc/habanalabs/memory.c 	sg_free_table(userptr->sgt);
sgt              1354 drivers/misc/habanalabs/memory.c 	kfree(userptr->sgt);
sgt               901 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	struct sg_table *sgt;
sgt              3532 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!adapter->hma.sgt)
sgt              3536 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		dma_unmap_sg(adapter->pdev_dev, adapter->hma.sgt->sgl,
sgt              3537 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			     adapter->hma.sgt->nents, PCI_DMA_BIDIRECTIONAL);
sgt              3541 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(adapter->hma.sgt->sgl, iter,
sgt              3542 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		    adapter->hma.sgt->orig_nents, i) {
sgt              3549 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sg_free_table(adapter->hma.sgt);
sgt              3550 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	kfree(adapter->hma.sgt);
sgt              3551 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adapter->hma.sgt = NULL;
sgt              3557 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sg_table *sgt;
sgt              3594 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adapter->hma.sgt = kzalloc(sizeof(*adapter->hma.sgt), GFP_KERNEL);
sgt              3595 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (unlikely(!adapter->hma.sgt)) {
sgt              3599 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgt = adapter->hma.sgt;
sgt              3602 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgt->orig_nents = (hma_size << 20) / (page_size << page_order);
sgt              3603 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (sg_alloc_table(sgt, sgt->orig_nents, GFP_KERNEL)) {
sgt              3605 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		kfree(adapter->hma.sgt);
sgt              3606 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		adapter->hma.sgt = NULL;
sgt              3610 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgl = adapter->hma.sgt->sgl;
sgt              3612 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(sgl, iter, sgt->orig_nents, i) {
sgt              3624 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	sgt->nents = dma_map_sg(adapter->pdev_dev, sgl, sgt->orig_nents,
sgt              3626 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	if (!sgt->nents) {
sgt              3634 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	adapter->hma.phy_addr = kcalloc(sgt->nents, sizeof(dma_addr_t),
sgt              3639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for_each_sg(sgl, iter, sgt->nents, i) {
sgt              3644 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	ncmds = DIV_ROUND_UP(sgt->nents, HMA_MAX_ADDR_IN_CMD);
sgt              3659 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			naddr = sgt->nents % HMA_MAX_ADDR_IN_CMD;
sgt              1290 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c static void dpaa_release_sgt_members(struct qm_sg_entry *sgt)
sgt              1299 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dpaa_bp = dpaa_bpid2pool(sgt[i].bpid);
sgt              1305 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			WARN_ON(qm_sg_entry_is_ext(&sgt[i]));
sgt              1307 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			bm_buffer_set64(&bmb[j], qm_sg_entry_get64(&sgt[i]));
sgt              1311 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 				!qm_sg_entry_is_final(&sgt[i - 1]) &&
sgt              1312 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 				sgt[i - 1].bpid == sgt[i].bpid);
sgt              1315 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	} while (!qm_sg_entry_is_final(&sgt[i - 1]));
sgt              1321 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	struct qm_sg_entry *sgt;
sgt              1336 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sgt = vaddr + qm_fd_get_offset(fd);
sgt              1341 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dpaa_release_sgt_members(sgt);
sgt              1617 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	const struct qm_sg_entry *sgt;
sgt              1634 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sgt = phys_to_virt(addr + qm_fd_get_offset(fd));
sgt              1637 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dma_unmap_single(dev, qm_sg_addr(&sgt[0]),
sgt              1638 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 				 qm_sg_entry_get_len(&sgt[0]), dma_dir);
sgt              1642 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			WARN_ON(qm_sg_entry_is_ext(&sgt[i]));
sgt              1644 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			dma_unmap_page(dev, qm_sg_addr(&sgt[i]),
sgt              1645 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 				       qm_sg_entry_get_len(&sgt[i]), dma_dir);
sgt              1737 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	const struct qm_sg_entry *sgt;
sgt              1753 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	sgt = vaddr + fd_off;
sgt              1757 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		WARN_ON(qm_sg_entry_is_ext(&sgt[i]));
sgt              1759 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sg_addr = qm_sg_addr(&sgt[i]);
sgt              1765 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dpaa_bp = dpaa_bpid2pool(sgt[i].bpid);
sgt              1786 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			skb_put(skb, qm_sg_entry_get_len(&sgt[i]));
sgt              1805 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			frag_off = qm_sg_entry_get_off(&sgt[i]) + page_offset;
sgt              1806 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			frag_len = qm_sg_entry_get_len(&sgt[i]);
sgt              1817 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		if (qm_sg_entry_is_final(&sgt[i]))
sgt              1830 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dpaa_bp = dpaa_bpid2pool(sgt[i].bpid);
sgt              1838 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sg_addr = qm_sg_addr(&sgt[i]);
sgt              1841 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dpaa_bp = dpaa_bpid2pool(sgt[i].bpid);
sgt              1847 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		if (qm_sg_entry_is_final(&sgt[i]))
sgt              1916 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	struct qm_sg_entry *sgt;
sgt              1949 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	sgt = (struct qm_sg_entry *)(sgt_buf + priv->tx_headroom);
sgt              1951 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	qm_sg_entry_set_len(&sgt[0], frag_len);
sgt              1952 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	sgt[0].bpid = FSL_DPAA_BPID_INV;
sgt              1953 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	sgt[0].offset = 0;
sgt              1961 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	qm_sg_entry_set64(&sgt[0], addr);
sgt              1976 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		qm_sg_entry_set_len(&sgt[i + 1], frag_len);
sgt              1977 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sgt[i + 1].bpid = FSL_DPAA_BPID_INV;
sgt              1978 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		sgt[i + 1].offset = 0;
sgt              1981 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		qm_sg_entry_set64(&sgt[i + 1], addr);
sgt              1985 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	qm_sg_entry_set_f(&sgt[nr_frags], frag_len);
sgt              1990 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	buffer_start = (void *)sgt - priv->tx_headroom;
sgt              2011 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 		dma_unmap_page(dev, qm_sg_addr(&sgt[j]),
sgt              2012 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 			       qm_sg_entry_get_len(&sgt[j]), dma_dir);
sgt                71 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	struct dpaa2_sg_entry *sgt;
sgt                85 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	sgt = vaddr + dpaa2_fd_get_offset(fd);
sgt                87 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		addr = dpaa2_sg_get_addr(&sgt[i]);
sgt                93 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		if (dpaa2_sg_is_final(&sgt[i]))
sgt               125 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 				      struct dpaa2_sg_entry *sgt)
sgt               138 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		struct dpaa2_sg_entry *sge = &sgt[i];
sgt               164 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 				while (!dpaa2_sg_is_final(&sgt[i]) &&
sgt               527 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	struct dpaa2_sg_entry *sgt;
sgt               566 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	sgt = (struct dpaa2_sg_entry *)(sgt_buf + priv->tx_data_offset);
sgt               576 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		dpaa2_sg_set_addr(&sgt[i], sg_dma_address(crt_scl));
sgt               577 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		dpaa2_sg_set_len(&sgt[i], sg_dma_len(crt_scl));
sgt               579 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	dpaa2_sg_set_final(&sgt[i - 1], true);
sgt               331 drivers/pci/controller/vmd.c static int vmd_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt               335 drivers/pci/controller/vmd.c 	return dma_get_sgtable_attrs(to_vmd_dev(dev), sgt, cpu_addr, addr, size,
sgt               540 drivers/rapidio/devices/rio_mport_cdev.c 	struct sg_table sgt;
sgt               578 drivers/rapidio/devices/rio_mport_cdev.c 		     req->sgt.sgl, req->sgt.nents, req->dir);
sgt               579 drivers/rapidio/devices/rio_mport_cdev.c 	sg_free_table(&req->sgt);
sgt               617 drivers/rapidio/devices/rio_mport_cdev.c 	struct sg_table *sgt, int nents, enum dma_transfer_direction dir,
sgt               622 drivers/rapidio/devices/rio_mport_cdev.c 	tx_data.sg = sgt->sgl;
sgt               698 drivers/rapidio/devices/rio_mport_cdev.c 	struct sg_table *sgt;
sgt               708 drivers/rapidio/devices/rio_mport_cdev.c 	sgt = &req->sgt;
sgt               719 drivers/rapidio/devices/rio_mport_cdev.c 	tx = prep_dma_xfer(chan, xfer, sgt, nents, dir,
sgt               888 drivers/rapidio/devices/rio_mport_cdev.c 		ret = sg_alloc_table_from_pages(&req->sgt, page_list, nr_pages,
sgt               924 drivers/rapidio/devices/rio_mport_cdev.c 		ret = sg_alloc_table(&req->sgt, 1, GFP_KERNEL);
sgt               930 drivers/rapidio/devices/rio_mport_cdev.c 		sg_set_buf(req->sgt.sgl,
sgt               936 drivers/rapidio/devices/rio_mport_cdev.c 			   req->sgt.sgl, req->sgt.nents, dir);
sgt               855 drivers/scsi/nsp32.c 	nsp32_sgtable *sgt = data->cur_lunt->sglun->sgt;
sgt               859 drivers/scsi/nsp32.c 	if (sgt == NULL) {
sgt               874 drivers/scsi/nsp32.c 			sgt[i].addr = cpu_to_le32(sg_dma_address(sg));
sgt               875 drivers/scsi/nsp32.c 			sgt[i].len  = cpu_to_le32(sg_dma_len(sg));
sgt               877 drivers/scsi/nsp32.c 			if (le32_to_cpu(sgt[i].len) > 0x10000) {
sgt               879 drivers/scsi/nsp32.c 					"can't transfer over 64KB at a time, size=0x%lx", le32_to_cpu(sgt[i].len));
sgt               885 drivers/scsi/nsp32.c 				  le32_to_cpu(sgt[i].addr),
sgt               886 drivers/scsi/nsp32.c 				  le32_to_cpu(sgt[i].len ));
sgt               890 drivers/scsi/nsp32.c 		l = le32_to_cpu(sgt[num-1].len);
sgt               891 drivers/scsi/nsp32.c 		sgt[num-1].len = cpu_to_le32(l | SGTEND);
sgt              1712 drivers/scsi/nsp32.c 	nsp32_sgtable *sgt    = data->cur_lunt->sglun->sgt;
sgt              1719 drivers/scsi/nsp32.c 	s_sacklen -= le32_to_cpu(sgt[old_entry].addr) & 3;
sgt              1727 drivers/scsi/nsp32.c 		sentlen += (le32_to_cpu(sgt[new_entry].len) & ~SGTEND);
sgt              1748 drivers/scsi/nsp32.c 	len  = le32_to_cpu(sgt[new_entry].len);
sgt              1749 drivers/scsi/nsp32.c 	addr = le32_to_cpu(sgt[new_entry].addr);
sgt              1751 drivers/scsi/nsp32.c 	sgt[new_entry].addr = cpu_to_le32(addr);
sgt              1752 drivers/scsi/nsp32.c 	sgt[new_entry].len  = cpu_to_le32(restlen);
sgt               446 drivers/scsi/nsp32.h 	nsp32_sgtable sgt[NSP32_SG_SIZE+1];	/* SG table */
sgt                24 drivers/spi/internals.h 		struct sg_table *sgt, void *buf, size_t len,
sgt                27 drivers/spi/internals.h 		   struct sg_table *sgt, enum dma_data_direction dir);
sgt                30 drivers/spi/internals.h 			      struct sg_table *sgt, void *buf, size_t len,
sgt                37 drivers/spi/internals.h 				 struct device *dev, struct sg_table *sgt,
sgt               277 drivers/spi/spi-ep93xx.c 	struct sg_table *sgt;
sgt               294 drivers/spi/spi-ep93xx.c 		sgt = &espi->rx_sgt;
sgt               301 drivers/spi/spi-ep93xx.c 		sgt = &espi->tx_sgt;
sgt               322 drivers/spi/spi-ep93xx.c 	if (nents != sgt->nents) {
sgt               323 drivers/spi/spi-ep93xx.c 		sg_free_table(sgt);
sgt               325 drivers/spi/spi-ep93xx.c 		ret = sg_alloc_table(sgt, nents, GFP_KERNEL);
sgt               331 drivers/spi/spi-ep93xx.c 	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
sgt               351 drivers/spi/spi-ep93xx.c 	nents = dma_map_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgt               355 drivers/spi/spi-ep93xx.c 	txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction,
sgt               358 drivers/spi/spi-ep93xx.c 		dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgt               377 drivers/spi/spi-ep93xx.c 	struct sg_table *sgt;
sgt               381 drivers/spi/spi-ep93xx.c 		sgt = &espi->rx_sgt;
sgt               384 drivers/spi/spi-ep93xx.c 		sgt = &espi->tx_sgt;
sgt               387 drivers/spi/spi-ep93xx.c 	dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
sgt                36 drivers/spi/spi-mem.c 				       struct sg_table *sgt)
sgt                53 drivers/spi/spi-mem.c 	return spi_map_buf(ctlr, dmadev, sgt, op->data.buf.in, op->data.nbytes,
sgt                82 drivers/spi/spi-mem.c 					  struct sg_table *sgt)
sgt                96 drivers/spi/spi-mem.c 	spi_unmap_buf(ctlr, dmadev, sgt,
sgt                78 drivers/spi/spi-pxa2xx-dma.c 	struct sg_table *sgt;
sgt               101 drivers/spi/spi-pxa2xx-dma.c 		sgt = &xfer->tx_sg;
sgt               108 drivers/spi/spi-pxa2xx-dma.c 		sgt = &xfer->rx_sg;
sgt               118 drivers/spi/spi-pxa2xx-dma.c 	return dmaengine_prep_slave_sg(chan, sgt->sgl, sgt->nents, dir,
sgt               268 drivers/spi/spi-s3c64xx.c 			struct sg_table *sgt)
sgt               294 drivers/spi/spi-s3c64xx.c 	desc = dmaengine_prep_slave_sg(dma->ch, sgt->sgl, sgt->nents,
sgt               205 drivers/spi/spi-stm32-qspi.c 	struct sg_table sgt;
sgt               222 drivers/spi/spi-stm32-qspi.c 	err = spi_controller_dma_map_mem_op_data(qspi->ctrl, op, &sgt);
sgt               226 drivers/spi/spi-stm32-qspi.c 	desc = dmaengine_prep_slave_sg(dma_ch, sgt.sgl, sgt.nents,
sgt               247 drivers/spi/spi-stm32-qspi.c 	t_out = sgt.nents * STM32_COMP_TIMEOUT_MS;
sgt               258 drivers/spi/spi-stm32-qspi.c 	spi_controller_dma_unmap_mem_op_data(qspi->ctrl, op, &sgt);
sgt               555 drivers/spi/spi-ti-qspi.c 		struct sg_table sgt;
sgt               559 drivers/spi/spi-ti-qspi.c 							&sgt)) {
sgt               560 drivers/spi/spi-ti-qspi.c 			ret = ti_qspi_dma_xfer_sg(qspi, sgt, from);
sgt               562 drivers/spi/spi-ti-qspi.c 							     op, &sgt);
sgt               807 drivers/spi/spi.c 		struct sg_table *sgt, void *buf, size_t len,
sgt               837 drivers/spi/spi.c 	ret = sg_alloc_table(sgt, sgs, GFP_KERNEL);
sgt               841 drivers/spi/spi.c 	sg = &sgt->sgl[0];
sgt               858 drivers/spi/spi.c 				sg_free_table(sgt);
sgt               874 drivers/spi/spi.c 	ret = dma_map_sg(dev, sgt->sgl, sgt->nents, dir);
sgt               878 drivers/spi/spi.c 		sg_free_table(sgt);
sgt               882 drivers/spi/spi.c 	sgt->nents = ret;
sgt               888 drivers/spi/spi.c 		   struct sg_table *sgt, enum dma_data_direction dir)
sgt               890 drivers/spi/spi.c 	if (sgt->orig_nents) {
sgt               891 drivers/spi/spi.c 		dma_unmap_sg(dev, sgt->sgl, sgt->orig_nents, dir);
sgt               892 drivers/spi/spi.c 		sg_free_table(sgt);
sgt                88 drivers/staging/kpc2000/kpc_dma/fileops.c 	rv = sg_alloc_table_from_pages(&acd->sgt, acd->user_pages, acd->page_count, iov_base & (PAGE_SIZE-1), iov_len, GFP_KERNEL);
sgt                95 drivers/staging/kpc2000/kpc_dma/fileops.c 	acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
sgt               102 drivers/staging/kpc2000/kpc_dma/fileops.c 	for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) {
sgt               125 drivers/staging/kpc2000/kpc_dma/fileops.c 	for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) {
sgt               189 drivers/staging/kpc2000/kpc_dma/fileops.c 	dma_unmap_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
sgt               190 drivers/staging/kpc2000/kpc_dma/fileops.c 	sg_free_table(&acd->sgt);
sgt               210 drivers/staging/kpc2000/kpc_dma/fileops.c 	BUG_ON(acd->sgt.sgl == NULL);
sgt               220 drivers/staging/kpc2000/kpc_dma/fileops.c 	dma_unmap_sg(&acd->ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, acd->ldev->dir);
sgt               226 drivers/staging/kpc2000/kpc_dma/fileops.c 	sg_free_table(&acd->sgt);
sgt                93 drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.h 	struct sg_table     sgt;
sgt                24 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct sg_table *sgt;
sgt                38 drivers/staging/media/tegra-vde/dmabuf-cache.c 	dma_buf_unmap_attachment(entry->a, entry->sgt, entry->dma_dir);
sgt                69 drivers/staging/media/tegra-vde/dmabuf-cache.c 	struct sg_table *sgt;
sgt                90 drivers/staging/media/tegra-vde/dmabuf-cache.c 			*addrp = sg_dma_address(entry->sgt->sgl);
sgt               102 drivers/staging/media/tegra-vde/dmabuf-cache.c 	sgt = dma_buf_map_attachment(attachment, dma_dir);
sgt               103 drivers/staging/media/tegra-vde/dmabuf-cache.c 	if (IS_ERR(sgt)) {
sgt               105 drivers/staging/media/tegra-vde/dmabuf-cache.c 		err = PTR_ERR(sgt);
sgt               109 drivers/staging/media/tegra-vde/dmabuf-cache.c 	if (!vde->domain && sgt->nents > 1) {
sgt               122 drivers/staging/media/tegra-vde/dmabuf-cache.c 		err = tegra_vde_iommu_map(vde, sgt, &iova, dmabuf->size);
sgt               128 drivers/staging/media/tegra-vde/dmabuf-cache.c 		*addrp = sg_dma_address(sgt->sgl);
sgt               138 drivers/staging/media/tegra-vde/dmabuf-cache.c 	entry->sgt = sgt;
sgt               152 drivers/staging/media/tegra-vde/dmabuf-cache.c 	dma_buf_unmap_attachment(attachment, sgt, dma_dir);
sgt                20 drivers/staging/media/tegra-vde/iommu.c 			struct sg_table *sgt,
sgt                39 drivers/staging/media/tegra-vde/iommu.c 	size = iommu_map_sg(vde->domain, addr, sgt->sgl, sgt->nents,
sgt                58 drivers/staging/media/tegra-vde/vde.h 			struct sg_table *sgt,
sgt               226 drivers/usb/gadget/function/f_fs.c 	struct sg_table sgt;
sgt               764 drivers/usb/gadget/function/f_fs.c static void *ffs_build_sg_list(struct sg_table *sgt, size_t sz)
sgt               785 drivers/usb/gadget/function/f_fs.c 	if (sg_alloc_table_from_pages(sgt, pages, n_pages, 0, sz, GFP_KERNEL)) {
sgt               800 drivers/usb/gadget/function/f_fs.c 		return ffs_build_sg_list(&io_data->sgt, data_len);
sgt               811 drivers/usb/gadget/function/f_fs.c 		sg_free_table(&io_data->sgt);
sgt              1061 drivers/usb/gadget/function/f_fs.c 			req->sg	= io_data->sgt.sgl;
sgt              1062 drivers/usb/gadget/function/f_fs.c 			req->num_sgs = io_data->sgt.nents;
sgt              1105 drivers/usb/gadget/function/f_fs.c 			req->sg	= io_data->sgt.sgl;
sgt              1106 drivers/usb/gadget/function/f_fs.c 			req->num_sgs = io_data->sgt.nents;
sgt                51 drivers/xen/gntdev-dmabuf.c 			struct sg_table *sgt;
sgt                70 drivers/xen/gntdev-dmabuf.c 	struct sg_table *sgt;
sgt               203 drivers/xen/gntdev-dmabuf.c 	struct sg_table *sgt;
sgt               206 drivers/xen/gntdev-dmabuf.c 	sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
sgt               207 drivers/xen/gntdev-dmabuf.c 	if (!sgt) {
sgt               212 drivers/xen/gntdev-dmabuf.c 	ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0,
sgt               218 drivers/xen/gntdev-dmabuf.c 	return sgt;
sgt               221 drivers/xen/gntdev-dmabuf.c 	kfree(sgt);
sgt               246 drivers/xen/gntdev-dmabuf.c 		struct sg_table *sgt = gntdev_dmabuf_attach->sgt;
sgt               248 drivers/xen/gntdev-dmabuf.c 		if (sgt) {
sgt               250 drivers/xen/gntdev-dmabuf.c 				dma_unmap_sg_attrs(attach->dev, sgt->sgl,
sgt               251 drivers/xen/gntdev-dmabuf.c 						   sgt->nents,
sgt               254 drivers/xen/gntdev-dmabuf.c 			sg_free_table(sgt);
sgt               257 drivers/xen/gntdev-dmabuf.c 		kfree(sgt);
sgt               269 drivers/xen/gntdev-dmabuf.c 	struct sg_table *sgt;
sgt               279 drivers/xen/gntdev-dmabuf.c 		return gntdev_dmabuf_attach->sgt;
sgt               288 drivers/xen/gntdev-dmabuf.c 	sgt = dmabuf_pages_to_sgt(gntdev_dmabuf->pages,
sgt               290 drivers/xen/gntdev-dmabuf.c 	if (!IS_ERR(sgt)) {
sgt               291 drivers/xen/gntdev-dmabuf.c 		if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir,
sgt               293 drivers/xen/gntdev-dmabuf.c 			sg_free_table(sgt);
sgt               294 drivers/xen/gntdev-dmabuf.c 			kfree(sgt);
sgt               295 drivers/xen/gntdev-dmabuf.c 			sgt = ERR_PTR(-ENOMEM);
sgt               297 drivers/xen/gntdev-dmabuf.c 			gntdev_dmabuf_attach->sgt = sgt;
sgt               301 drivers/xen/gntdev-dmabuf.c 	if (IS_ERR(sgt))
sgt               303 drivers/xen/gntdev-dmabuf.c 	return sgt;
sgt               307 drivers/xen/gntdev-dmabuf.c 					 struct sg_table *sgt,
sgt               613 drivers/xen/gntdev-dmabuf.c 	struct sg_table *sgt;
sgt               638 drivers/xen/gntdev-dmabuf.c 	sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL);
sgt               639 drivers/xen/gntdev-dmabuf.c 	if (IS_ERR(sgt)) {
sgt               640 drivers/xen/gntdev-dmabuf.c 		ret = ERR_CAST(sgt);
sgt               652 drivers/xen/gntdev-dmabuf.c 	gntdev_dmabuf->u.imp.sgt = sgt;
sgt               656 drivers/xen/gntdev-dmabuf.c 	for_each_sg_page(sgt->sgl, &sg_iter, sgt->nents, 0) {
sgt               688 drivers/xen/gntdev-dmabuf.c 	dma_buf_unmap_attachment(attach, sgt, DMA_BIDIRECTIONAL);
sgt               737 drivers/xen/gntdev-dmabuf.c 	if (gntdev_dmabuf->u.imp.sgt)
sgt               738 drivers/xen/gntdev-dmabuf.c 		dma_buf_unmap_attachment(attach, gntdev_dmabuf->u.imp.sgt,
sgt               627 include/drm/drm_drv.h 				struct sg_table *sgt);
sgt                23 include/drm/drm_gem_cma_helper.h 	struct sg_table *sgt;
sgt               103 include/drm/drm_gem_cma_helper.h 				  struct sg_table *sgt);
sgt               131 include/drm/drm_gem_cma_helper.h 				       struct sg_table *sgt);
sgt                80 include/drm/drm_gem_shmem_helper.h 	struct sg_table *sgt;
sgt               145 include/drm/drm_gem_shmem_helper.h 		!shmem->vmap_use_count && shmem->sgt &&
sgt               170 include/drm/drm_gem_shmem_helper.h 				    struct sg_table *sgt);
sgt                85 include/drm/drm_prime.h 			   struct sg_table *sgt,
sgt               106 include/drm/drm_prime.h int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages,
sgt               344 include/linux/dma-buf.h 	struct sg_table *sgt;
sgt                90 include/linux/dma-mapping.h 	int (*get_sgtable)(struct device *dev, struct sg_table *sgt, void *,
sgt               455 include/linux/dma-mapping.h int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt,
sgt               542 include/linux/dma-mapping.h 		struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr,
sgt               641 include/linux/dma-mapping.h dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr,
sgt                98 include/linux/fpga/fpga-mgr.h 	struct sg_table *sgt;
sgt               132 include/linux/fpga/fpga-mgr.h 	int (*write_sg)(struct fpga_manager *mgr, struct sg_table *sgt);
sgt                67 include/linux/host1x.h 	dma_addr_t (*pin)(struct host1x_bo *bo, struct sg_table **sgt);
sgt                68 include/linux/host1x.h 	void (*unpin)(struct host1x_bo *bo, struct sg_table *sgt);
sgt                96 include/linux/host1x.h 				       struct sg_table **sgt)
sgt                98 include/linux/host1x.h 	return bo->ops->pin(bo, sgt);
sgt               101 include/linux/host1x.h static inline void host1x_bo_unpin(struct host1x_bo *bo, struct sg_table *sgt)
sgt               103 include/linux/host1x.h 	bo->ops->unpin(bo, sgt);
sgt               188 include/linux/hsi/hsi.h 	struct sg_table		sgt;
sgt                43 include/linux/intel_th.h 	int	(*alloc_window)(void *priv, struct sg_table **sgt,
sgt                45 include/linux/intel_th.h 	void	(*free_window)(void *priv, struct sg_table *sgt);
sgt                59 include/linux/intel_th.h 	int	(*ready)(void *priv, struct sg_table *sgt, size_t bytes);
sgt                65 include/linux/intel_th.h void intel_th_msc_window_unlock(struct device *dev, struct sg_table *sgt);
sgt               275 include/linux/scatterlist.h int __sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages,
sgt               279 include/linux/scatterlist.h int sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages,
sgt               115 include/media/vsp1.h int vsp1_du_map_sg(struct device *dev, struct sg_table *sgt);
sgt               116 include/media/vsp1.h void vsp1_du_unmap_sg(struct device *dev, struct sg_table *sgt);
sgt               263 include/net/erspan.h 	u8 sgt = 0;
sgt               294 include/net/erspan.h 	md2->sgt = htons(sgt);
sgt                22 include/uapi/linux/erspan.h 	__be16 sgt;	/* security group tag */
sgt               111 kernel/dma/mapping.c int dma_common_get_sgtable(struct device *dev, struct sg_table *sgt,
sgt               133 kernel/dma/mapping.c 	ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
sgt               135 kernel/dma/mapping.c 		sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
sgt               150 kernel/dma/mapping.c int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt,
sgt               157 kernel/dma/mapping.c 		return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr,
sgt               161 kernel/dma/mapping.c 	return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs);
sgt               389 lib/scatterlist.c int __sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages,
sgt               413 lib/scatterlist.c 	ret = sg_alloc_table(sgt, chunks, gfp_mask);
sgt               419 lib/scatterlist.c 	for_each_sg(sgt->sgl, s, sgt->orig_nents, i) {
sgt               464 lib/scatterlist.c int sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages,
sgt               468 lib/scatterlist.c 	return __sg_alloc_table_from_pages(sgt, pages, n_pages, offset, size,
sgt               160 net/ceph/crypto.c static int setup_sgtable(struct sg_table *sgt, struct scatterlist *prealloc_sg,
sgt               172 net/ceph/crypto.c 		memset(sgt, 0, sizeof(*sgt));
sgt               182 net/ceph/crypto.c 		ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS);
sgt               188 net/ceph/crypto.c 		sgt->sgl = prealloc_sg;
sgt               189 net/ceph/crypto.c 		sgt->nents = sgt->orig_nents = 1;
sgt               192 net/ceph/crypto.c 	for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) {
sgt               212 net/ceph/crypto.c static void teardown_sgtable(struct sg_table *sgt)
sgt               214 net/ceph/crypto.c 	if (sgt->orig_nents > 1)
sgt               215 net/ceph/crypto.c 		sg_free_table(sgt);
sgt               222 net/ceph/crypto.c 	struct sg_table sgt;
sgt               232 net/ceph/crypto.c 	ret = setup_sgtable(&sgt, &prealloc_sg, buf, crypt_len);
sgt               239 net/ceph/crypto.c 	skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv);
sgt               278 net/ceph/crypto.c 	teardown_sgtable(&sgt);
sgt               516 net/smc/smc_clc.c 				(conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgt               588 net/smc/smc_clc.c 				(conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgt               379 net/smc/smc_core.c 	sg_free_table(&buf_desc->sgt[SMC_SINGLE_LINK]);
sgt               764 net/smc/smc_core.c 	rc = sg_alloc_table(&buf_desc->sgt[SMC_SINGLE_LINK], 1,
sgt               770 net/smc/smc_core.c 	sg_set_buf(buf_desc->sgt[SMC_SINGLE_LINK].sgl,
sgt               157 net/smc/smc_core.h 			struct sg_table		sgt[SMC_LINKS_PER_LGR_MAX];
sgt               369 net/smc/smc_ib.c 			      buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgt               370 net/smc/smc_ib.c 			      buf_slot->sgt[SMC_SINGLE_LINK].orig_nents,
sgt               408 net/smc/smc_ib.c 	for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg,
sgt               409 net/smc/smc_ib.c 		    buf_slot->sgt[SMC_SINGLE_LINK].nents, i) {
sgt               428 net/smc/smc_ib.c 	for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg,
sgt               429 net/smc/smc_ib.c 		    buf_slot->sgt[SMC_SINGLE_LINK].nents, i) {
sgt               447 net/smc/smc_ib.c 				     buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgt               448 net/smc/smc_ib.c 				     buf_slot->sgt[SMC_SINGLE_LINK].orig_nents,
sgt               460 net/smc/smc_ib.c 	if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address)
sgt               464 net/smc/smc_ib.c 			buf_slot->sgt[SMC_SINGLE_LINK].sgl,
sgt               465 net/smc/smc_ib.c 			buf_slot->sgt[SMC_SINGLE_LINK].orig_nents,
sgt               467 net/smc/smc_ib.c 	buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0;
sgt               236 net/smc/smc_llc.c 		(u64)sg_dma_address(rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
sgt               315 net/smc/smc_tx.c 		sg_dma_address(conn->sndbuf_desc->sgt[SMC_SINGLE_LINK].sgl);
sgt               144 sound/soc/sprd/sprd-pcm-compress.c 	struct scatterlist *sg, *sgt;
sgt               163 sound/soc/sprd/sprd-pcm-compress.c 	sgt = sg = devm_kcalloc(dev, sg_num, sizeof(*sg), GFP_KERNEL);
sgt               214 sound/soc/sprd/sprd-pcm-compress.c 	sg_init_table(sgt, sg_num);
sgt               215 sound/soc/sprd/sprd-pcm-compress.c 	for (j = 0; j < sg_num; j++, sgt++) {
sgt               216 sound/soc/sprd/sprd-pcm-compress.c 		sg_dma_len(sgt) = dma->trans_len;
sgt               217 sound/soc/sprd/sprd-pcm-compress.c 		sg_dma_address(sgt) = dst_addr;
sgt               244 sound/soc/sprd/sprd-pcm-dma.c 		struct scatterlist *sgt = sg;
sgt               261 sound/soc/sprd/sprd-pcm-dma.c 		sg_init_table(sgt, sg_num);
sgt               262 sound/soc/sprd/sprd-pcm-dma.c 		for (j = 0; j < sg_num; j++, sgt++) {
sgt               265 sound/soc/sprd/sprd-pcm-dma.c 			sg_dma_len(sgt) = sg_len;
sgt               266 sound/soc/sprd/sprd-pcm-dma.c 			sg_dma_address(sgt) = runtime->dma_addr +
sgt                22 tools/include/uapi/linux/erspan.h 	__be16 sgt;	/* security group tag */