sgt 194 arch/arm/include/asm/dma-mapping.h extern int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 865 arch/arm/mm/dma-mapping.c int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 879 arch/arm/mm/dma-mapping.c ret = sg_alloc_table(sgt, 1, GFP_KERNEL); sgt 883 arch/arm/mm/dma-mapping.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgt 1598 arch/arm/mm/dma-mapping.c static int arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 1608 arch/arm/mm/dma-mapping.c return sg_alloc_table_from_pages(sgt, pages, count, 0, size, sgt 129 arch/s390/include/asm/gmap.h int gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt, sgt 1353 arch/s390/mm/gmap.c unsigned long *sgt) sgt 1361 arch/s390/mm/gmap.c if (!(sgt[i] & _SEGMENT_ENTRY_ORIGIN)) sgt 1363 arch/s390/mm/gmap.c pgt = (unsigned long *)(sgt[i] & _REGION_ENTRY_ORIGIN); sgt 1364 arch/s390/mm/gmap.c sgt[i] = _SEGMENT_ENTRY_EMPTY; sgt 1382 arch/s390/mm/gmap.c unsigned long r3o, *r3e, *sgt; sgt 1392 arch/s390/mm/gmap.c sgt = (unsigned long *)(*r3e & _REGION_ENTRY_ORIGIN); sgt 1394 arch/s390/mm/gmap.c __gmap_unshadow_sgt(sg, raddr, sgt); sgt 1396 arch/s390/mm/gmap.c page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT); sgt 1412 arch/s390/mm/gmap.c unsigned long *sgt; sgt 1420 arch/s390/mm/gmap.c sgt = (unsigned long *)(r3t[i] & _REGION_ENTRY_ORIGIN); sgt 1422 arch/s390/mm/gmap.c __gmap_unshadow_sgt(sg, raddr, sgt); sgt 1424 arch/s390/mm/gmap.c page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT); sgt 1902 arch/s390/mm/gmap.c int gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt, sgt 1910 arch/s390/mm/gmap.c BUG_ON(!gmap_is_shadow(sg) || (sgt & _REGION3_ENTRY_LARGE)); sgt 1915 arch/s390/mm/gmap.c page->index = sgt & _REGION_ENTRY_ORIGIN; sgt 1938 arch/s390/mm/gmap.c *table |= sgt & _REGION_ENTRY_PROTECT; sgt 1949 arch/s390/mm/gmap.c origin = sgt & _REGION_ENTRY_ORIGIN; sgt 1950 arch/s390/mm/gmap.c offset = ((sgt & _REGION_ENTRY_OFFSET) >> 6) * PAGE_SIZE; sgt 1951 arch/s390/mm/gmap.c len = ((sgt & _REGION_ENTRY_LENGTH) + 1) * PAGE_SIZE - offset; sgt 808 drivers/crypto/caam/caamalg_qi.c struct qm_sg_entry sgt[0]; sgt 828 drivers/crypto/caam/caamalg_qi.c struct qm_sg_entry sgt[0]; sgt 1061 drivers/crypto/caam/caamalg_qi.c sg_table = &edesc->sgt[0]; sgt 1063 drivers/crypto/caam/caamalg_qi.c if (unlikely(offsetof(struct aead_edesc, sgt) + qm_sg_bytes + ivsize > sgt 1239 drivers/crypto/caam/caamalg_qi.c memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes, sgt 1322 drivers/crypto/caam/caamalg_qi.c if (unlikely(offsetof(struct skcipher_edesc, sgt) + qm_sg_bytes + sgt 1341 drivers/crypto/caam/caamalg_qi.c sg_table = &edesc->sgt[0]; sgt 468 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 470 drivers/crypto/caam/caamalg_qi2.c if (unlikely(offsetof(struct aead_edesc, sgt) + qm_sg_bytes + ivsize > sgt 1189 drivers/crypto/caam/caamalg_qi2.c if (unlikely(offsetof(struct skcipher_edesc, sgt) + qm_sg_bytes + sgt 1208 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 1427 drivers/crypto/caam/caamalg_qi2.c memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes, sgt 1465 drivers/crypto/caam/caamalg_qi2.c memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes, sgt 3513 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 3608 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 3702 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 3796 drivers/crypto/caam/caamalg_qi2.c struct dpaa2_sg_entry *sg_table = &edesc->sgt[0]; sgt 3972 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 4091 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 4195 drivers/crypto/caam/caamalg_qi2.c sg_table = &edesc->sgt[0]; sgt 117 drivers/crypto/caam/caamalg_qi2.h struct dpaa2_sg_entry sgt[0]; sgt 135 drivers/crypto/caam/caamalg_qi2.h struct dpaa2_sg_entry sgt[0]; sgt 149 drivers/crypto/caam/caamalg_qi2.h struct dpaa2_sg_entry sgt[0]; sgt 50 drivers/crypto/qce/dma.c qce_sgtable_add(struct sg_table *sgt, struct scatterlist *new_sgl) sgt 52 drivers/crypto/qce/dma.c struct scatterlist *sg = sgt->sgl, *sg_last = NULL; sgt 45 drivers/crypto/qce/dma.h qce_sgtable_add(struct sg_table *sgt, struct scatterlist *sg_add); sgt 716 drivers/dma-buf/dma-buf.c if (attach->sgt) sgt 717 drivers/dma-buf/dma-buf.c dmabuf->ops->unmap_dma_buf(attach, attach->sgt, attach->dir); sgt 754 drivers/dma-buf/dma-buf.c if (attach->sgt) { sgt 763 drivers/dma-buf/dma-buf.c return attach->sgt; sgt 771 drivers/dma-buf/dma-buf.c attach->sgt = sg_table; sgt 798 drivers/dma-buf/dma-buf.c if (attach->sgt == sg_table) sgt 98 drivers/fpga/fpga-mgr.c struct sg_table *sgt) sgt 112 drivers/fpga/fpga-mgr.c sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); sgt 127 drivers/fpga/fpga-mgr.c len = sg_copy_to_buffer(sgt->sgl, sgt->nents, buf, sgt 176 drivers/fpga/fpga-mgr.c struct sg_table *sgt) sgt 180 drivers/fpga/fpga-mgr.c ret = fpga_mgr_write_init_sg(mgr, info, sgt); sgt 187 drivers/fpga/fpga-mgr.c ret = mgr->mops->write_sg(mgr, sgt); sgt 191 drivers/fpga/fpga-mgr.c sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); sgt 252 drivers/fpga/fpga-mgr.c struct sg_table sgt; sgt 293 drivers/fpga/fpga-mgr.c rc = sg_alloc_table_from_pages(&sgt, pages, index, offset_in_page(buf), sgt 299 drivers/fpga/fpga-mgr.c rc = fpga_mgr_buf_load_sg(mgr, info, &sgt); sgt 300 drivers/fpga/fpga-mgr.c sg_free_table(&sgt); sgt 357 drivers/fpga/fpga-mgr.c if (info->sgt) sgt 358 drivers/fpga/fpga-mgr.c return fpga_mgr_buf_load_sg(mgr, info, info->sgt); sgt 384 drivers/fpga/zynq-fpga.c static int zynq_fpga_ops_write(struct fpga_manager *mgr, struct sg_table *sgt) sgt 400 drivers/fpga/zynq-fpga.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 409 drivers/fpga/zynq-fpga.c dma_map_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE); sgt 426 drivers/fpga/zynq-fpga.c priv->cur_sg = sgt->sgl; sgt 481 drivers/fpga/zynq-fpga.c dma_unmap_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE); sgt 340 drivers/gpu/drm/arm/malidp_planes.c struct sg_table *sgt; sgt 346 drivers/gpu/drm/arm/malidp_planes.c if (cma_obj->sgt) sgt 347 drivers/gpu/drm/arm/malidp_planes.c sgt = cma_obj->sgt; sgt 349 drivers/gpu/drm/arm/malidp_planes.c sgt = obj->dev->driver->gem_prime_get_sg_table(obj); sgt 351 drivers/gpu/drm/arm/malidp_planes.c if (!sgt) sgt 354 drivers/gpu/drm/arm/malidp_planes.c sgl = sgt->sgl; sgt 358 drivers/gpu/drm/arm/malidp_planes.c if (!cma_obj->sgt) sgt 359 drivers/gpu/drm/arm/malidp_planes.c kfree(sgt); sgt 365 drivers/gpu/drm/arm/malidp_planes.c if (!cma_obj->sgt) sgt 366 drivers/gpu/drm/arm/malidp_planes.c kfree(sgt); sgt 114 drivers/gpu/drm/armada/armada_fb.c if (obj->obj.import_attach && !obj->sgt) { sgt 66 drivers/gpu/drm/armada/armada_gem.c if (dobj->sgt) sgt 68 drivers/gpu/drm/armada/armada_gem.c dobj->sgt, DMA_TO_DEVICE); sgt 381 drivers/gpu/drm/armada/armada_gem.c struct sg_table *sgt; sgt 384 drivers/gpu/drm/armada/armada_gem.c sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); sgt 385 drivers/gpu/drm/armada/armada_gem.c if (!sgt) sgt 393 drivers/gpu/drm/armada/armada_gem.c if (sg_alloc_table(sgt, count, GFP_KERNEL)) sgt 398 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, count, i) { sgt 410 drivers/gpu/drm/armada/armada_gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) { sgt 411 drivers/gpu/drm/armada/armada_gem.c num = sgt->nents; sgt 416 drivers/gpu/drm/armada/armada_gem.c if (sg_alloc_table(sgt, 1, GFP_KERNEL)) sgt 419 drivers/gpu/drm/armada/armada_gem.c sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); sgt 421 drivers/gpu/drm/armada/armada_gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) sgt 425 drivers/gpu/drm/armada/armada_gem.c if (sg_alloc_table(sgt, 1, GFP_KERNEL)) sgt 427 drivers/gpu/drm/armada/armada_gem.c sg_dma_address(sgt->sgl) = dobj->dev_addr; sgt 428 drivers/gpu/drm/armada/armada_gem.c sg_dma_len(sgt->sgl) = dobj->obj.size; sgt 432 drivers/gpu/drm/armada/armada_gem.c return sgt; sgt 435 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, num, i) sgt 438 drivers/gpu/drm/armada/armada_gem.c sg_free_table(sgt); sgt 440 drivers/gpu/drm/armada/armada_gem.c kfree(sgt); sgt 445 drivers/gpu/drm/armada/armada_gem.c struct sg_table *sgt, enum dma_data_direction dir) sgt 452 drivers/gpu/drm/armada/armada_gem.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); sgt 456 drivers/gpu/drm/armada/armada_gem.c for_each_sg(sgt->sgl, sg, sgt->nents, i) sgt 460 drivers/gpu/drm/armada/armada_gem.c sg_free_table(sgt); sgt 461 drivers/gpu/drm/armada/armada_gem.c kfree(sgt); sgt 545 drivers/gpu/drm/armada/armada_gem.c dobj->sgt = dma_buf_map_attachment(dobj->obj.import_attach, sgt 547 drivers/gpu/drm/armada/armada_gem.c if (IS_ERR(dobj->sgt)) { sgt 548 drivers/gpu/drm/armada/armada_gem.c ret = PTR_ERR(dobj->sgt); sgt 549 drivers/gpu/drm/armada/armada_gem.c dobj->sgt = NULL; sgt 553 drivers/gpu/drm/armada/armada_gem.c if (dobj->sgt->nents > 1) { sgt 557 drivers/gpu/drm/armada/armada_gem.c if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) { sgt 561 drivers/gpu/drm/armada/armada_gem.c dobj->dev_addr = sg_dma_address(dobj->sgt->sgl); sgt 19 drivers/gpu/drm/armada/armada_gem.h struct sg_table *sgt; /* for imported */ sgt 185 drivers/gpu/drm/drm_gem_cma_helper.c drm_prime_gem_destroy(gem_obj, cma_obj->sgt); sgt 430 drivers/gpu/drm/drm_gem_cma_helper.c struct sg_table *sgt; sgt 433 drivers/gpu/drm/drm_gem_cma_helper.c sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); sgt 434 drivers/gpu/drm/drm_gem_cma_helper.c if (!sgt) sgt 437 drivers/gpu/drm/drm_gem_cma_helper.c ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, sgt 442 drivers/gpu/drm/drm_gem_cma_helper.c return sgt; sgt 445 drivers/gpu/drm/drm_gem_cma_helper.c kfree(sgt); sgt 470 drivers/gpu/drm/drm_gem_cma_helper.c struct sg_table *sgt) sgt 474 drivers/gpu/drm/drm_gem_cma_helper.c if (sgt->nents != 1) { sgt 476 drivers/gpu/drm/drm_gem_cma_helper.c dma_addr_t next_addr = sg_dma_address(sgt->sgl); sgt 480 drivers/gpu/drm/drm_gem_cma_helper.c for_each_sg(sgt->sgl, s, sgt->nents, i) { sgt 500 drivers/gpu/drm/drm_gem_cma_helper.c cma_obj->paddr = sg_dma_address(sgt->sgl); sgt 501 drivers/gpu/drm/drm_gem_cma_helper.c cma_obj->sgt = sgt; sgt 633 drivers/gpu/drm/drm_gem_cma_helper.c struct sg_table *sgt) sgt 645 drivers/gpu/drm/drm_gem_cma_helper.c obj = drm_gem_cma_prime_import_sg_table(dev, attach, sgt); sgt 116 drivers/gpu/drm/drm_gem_shmem_helper.c drm_prime_gem_destroy(obj, shmem->sgt); sgt 119 drivers/gpu/drm/drm_gem_shmem_helper.c if (shmem->sgt) { sgt 120 drivers/gpu/drm/drm_gem_shmem_helper.c dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl, sgt 121 drivers/gpu/drm/drm_gem_shmem_helper.c shmem->sgt->nents, DMA_BIDIRECTIONAL); sgt 122 drivers/gpu/drm/drm_gem_shmem_helper.c sg_free_table(shmem->sgt); sgt 123 drivers/gpu/drm/drm_gem_shmem_helper.c kfree(shmem->sgt); sgt 393 drivers/gpu/drm/drm_gem_shmem_helper.c dma_unmap_sg(obj->dev->dev, shmem->sgt->sgl, sgt 394 drivers/gpu/drm/drm_gem_shmem_helper.c shmem->sgt->nents, DMA_BIDIRECTIONAL); sgt 395 drivers/gpu/drm/drm_gem_shmem_helper.c sg_free_table(shmem->sgt); sgt 396 drivers/gpu/drm/drm_gem_shmem_helper.c kfree(shmem->sgt); sgt 397 drivers/gpu/drm/drm_gem_shmem_helper.c shmem->sgt = NULL; sgt 611 drivers/gpu/drm/drm_gem_shmem_helper.c struct sg_table *sgt; sgt 613 drivers/gpu/drm/drm_gem_shmem_helper.c if (shmem->sgt) sgt 614 drivers/gpu/drm/drm_gem_shmem_helper.c return shmem->sgt; sgt 622 drivers/gpu/drm/drm_gem_shmem_helper.c sgt = drm_gem_shmem_get_sg_table(&shmem->base); sgt 623 drivers/gpu/drm/drm_gem_shmem_helper.c if (IS_ERR(sgt)) { sgt 624 drivers/gpu/drm/drm_gem_shmem_helper.c ret = PTR_ERR(sgt); sgt 628 drivers/gpu/drm/drm_gem_shmem_helper.c dma_map_sg(obj->dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgt 630 drivers/gpu/drm/drm_gem_shmem_helper.c shmem->sgt = sgt; sgt 632 drivers/gpu/drm/drm_gem_shmem_helper.c return sgt; sgt 658 drivers/gpu/drm/drm_gem_shmem_helper.c struct sg_table *sgt) sgt 675 drivers/gpu/drm/drm_gem_shmem_helper.c ret = drm_prime_sg_to_page_addr_arrays(sgt, shmem->pages, NULL, npages); sgt 679 drivers/gpu/drm/drm_gem_shmem_helper.c shmem->sgt = sgt; sgt 617 drivers/gpu/drm/drm_prime.c struct sg_table *sgt; sgt 623 drivers/gpu/drm/drm_prime.c sgt = obj->funcs->get_sg_table(obj); sgt 625 drivers/gpu/drm/drm_prime.c sgt = obj->dev->driver->gem_prime_get_sg_table(obj); sgt 627 drivers/gpu/drm/drm_prime.c if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgt 629 drivers/gpu/drm/drm_prime.c sg_free_table(sgt); sgt 630 drivers/gpu/drm/drm_prime.c kfree(sgt); sgt 631 drivers/gpu/drm/drm_prime.c sgt = ERR_PTR(-ENOMEM); sgt 634 drivers/gpu/drm/drm_prime.c return sgt; sgt 647 drivers/gpu/drm/drm_prime.c struct sg_table *sgt, sgt 650 drivers/gpu/drm/drm_prime.c if (!sgt) sgt 653 drivers/gpu/drm/drm_prime.c dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgt 655 drivers/gpu/drm/drm_prime.c sg_free_table(sgt); sgt 656 drivers/gpu/drm/drm_prime.c kfree(sgt); sgt 862 drivers/gpu/drm/drm_prime.c struct sg_table *sgt; sgt 887 drivers/gpu/drm/drm_prime.c sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL); sgt 888 drivers/gpu/drm/drm_prime.c if (IS_ERR(sgt)) { sgt 889 drivers/gpu/drm/drm_prime.c ret = PTR_ERR(sgt); sgt 893 drivers/gpu/drm/drm_prime.c obj = dev->driver->gem_prime_import_sg_table(dev, attach, sgt); sgt 905 drivers/gpu/drm/drm_prime.c dma_buf_unmap_attachment(attach, sgt, DMA_BIDIRECTIONAL); sgt 947 drivers/gpu/drm/drm_prime.c int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages, sgt 957 drivers/gpu/drm/drm_prime.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgt 23 drivers/gpu/drm/etnaviv/etnaviv_gem.c struct sg_table *sgt = etnaviv_obj->sgt; sgt 30 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_map_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgt 36 drivers/gpu/drm/etnaviv/etnaviv_gem.c struct sg_table *sgt = etnaviv_obj->sgt; sgt 54 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_unmap_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL); sgt 75 drivers/gpu/drm/etnaviv/etnaviv_gem.c if (etnaviv_obj->sgt) { sgt 77 drivers/gpu/drm/etnaviv/etnaviv_gem.c sg_free_table(etnaviv_obj->sgt); sgt 78 drivers/gpu/drm/etnaviv/etnaviv_gem.c kfree(etnaviv_obj->sgt); sgt 79 drivers/gpu/drm/etnaviv/etnaviv_gem.c etnaviv_obj->sgt = NULL; sgt 101 drivers/gpu/drm/etnaviv/etnaviv_gem.c if (!etnaviv_obj->sgt) { sgt 104 drivers/gpu/drm/etnaviv/etnaviv_gem.c struct sg_table *sgt; sgt 106 drivers/gpu/drm/etnaviv/etnaviv_gem.c sgt = drm_prime_pages_to_sg(etnaviv_obj->pages, npages); sgt 107 drivers/gpu/drm/etnaviv/etnaviv_gem.c if (IS_ERR(sgt)) { sgt 109 drivers/gpu/drm/etnaviv/etnaviv_gem.c PTR_ERR(sgt)); sgt 110 drivers/gpu/drm/etnaviv/etnaviv_gem.c return ERR_CAST(sgt); sgt 113 drivers/gpu/drm/etnaviv/etnaviv_gem.c etnaviv_obj->sgt = sgt; sgt 383 drivers/gpu/drm/etnaviv/etnaviv_gem.c if (!etnaviv_obj->sgt) { sgt 407 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_sync_sg_for_cpu(dev->dev, etnaviv_obj->sgt->sgl, sgt 408 drivers/gpu/drm/etnaviv/etnaviv_gem.c etnaviv_obj->sgt->nents, sgt 424 drivers/gpu/drm/etnaviv/etnaviv_gem.c dma_sync_sg_for_device(dev->dev, etnaviv_obj->sgt->sgl, sgt 425 drivers/gpu/drm/etnaviv/etnaviv_gem.c etnaviv_obj->sgt->nents, sgt 697 drivers/gpu/drm/etnaviv/etnaviv_gem.c if (etnaviv_obj->sgt) { sgt 699 drivers/gpu/drm/etnaviv/etnaviv_gem.c sg_free_table(etnaviv_obj->sgt); sgt 700 drivers/gpu/drm/etnaviv/etnaviv_gem.c kfree(etnaviv_obj->sgt); sgt 47 drivers/gpu/drm/etnaviv/etnaviv_gem.h struct sg_table *sgt; sgt 83 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c drm_prime_gem_destroy(&etnaviv_obj->base, etnaviv_obj->sgt); sgt 107 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c struct dma_buf_attachment *attach, struct sg_table *sgt) sgt 122 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c etnaviv_obj->sgt = sgt; sgt 129 drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c ret = drm_prime_sg_to_page_addr_arrays(sgt, etnaviv_obj->pages, sgt 73 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct sg_table *sgt, unsigned len, int prot) sgt 79 drivers/gpu/drm/etnaviv/etnaviv_mmu.c if (!context || !sgt) sgt 82 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 100 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, i, j) { sgt 110 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct sg_table *sgt, unsigned len) sgt 116 drivers/gpu/drm/etnaviv/etnaviv_mmu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 135 drivers/gpu/drm/etnaviv/etnaviv_mmu.c etnaviv_obj->sgt, etnaviv_obj->base.size); sgt 234 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct sg_table *sgt = etnaviv_obj->sgt; sgt 244 drivers/gpu/drm/etnaviv/etnaviv_mmu.c sgt->nents == 1 && !(etnaviv_obj->flags & ETNA_BO_FORCE_MMU)) { sgt 247 drivers/gpu/drm/etnaviv/etnaviv_mmu.c iova = sg_dma_address(sgt->sgl) - memory_base; sgt 248 drivers/gpu/drm/etnaviv/etnaviv_mmu.c if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) { sgt 268 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size, sgt 209 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct sg_table *sgt; sgt 398 drivers/gpu/drm/exynos/exynos_drm_g2d.c dma_unmap_sg(to_dma_dev(g2d->drm_dev), g2d_userptr->sgt->sgl, sgt 399 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_userptr->sgt->nents, DMA_BIDIRECTIONAL); sgt 414 drivers/gpu/drm/exynos/exynos_drm_g2d.c sg_free_table(g2d_userptr->sgt); sgt 415 drivers/gpu/drm/exynos/exynos_drm_g2d.c kfree(g2d_userptr->sgt); sgt 427 drivers/gpu/drm/exynos/exynos_drm_g2d.c struct sg_table *sgt; sgt 498 drivers/gpu/drm/exynos/exynos_drm_g2d.c sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); sgt 499 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (!sgt) { sgt 504 drivers/gpu/drm/exynos/exynos_drm_g2d.c ret = sg_alloc_table_from_pages(sgt, sgt 512 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_userptr->sgt = sgt; sgt 514 drivers/gpu/drm/exynos/exynos_drm_g2d.c if (!dma_map_sg(to_dma_dev(g2d->drm_dev), sgt->sgl, sgt->nents, sgt 521 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_userptr->dma_addr = sgt->sgl[0].dma_address; sgt 536 drivers/gpu/drm/exynos/exynos_drm_g2d.c sg_free_table(sgt); sgt 539 drivers/gpu/drm/exynos/exynos_drm_g2d.c kfree(sgt); sgt 25 drivers/gpu/drm/exynos/exynos_drm_gem.c struct sg_table sgt; sgt 73 drivers/gpu/drm/exynos/exynos_drm_gem.c ret = dma_get_sgtable_attrs(to_dma_dev(dev), &sgt, exynos_gem->cookie, sgt 81 drivers/gpu/drm/exynos/exynos_drm_gem.c if (drm_prime_sg_to_page_addr_arrays(&sgt, exynos_gem->pages, NULL, sgt 88 drivers/gpu/drm/exynos/exynos_drm_gem.c sg_free_table(&sgt); sgt 96 drivers/gpu/drm/exynos/exynos_drm_gem.c sg_free_table(&sgt); sgt 161 drivers/gpu/drm/exynos/exynos_drm_gem.c drm_prime_gem_destroy(obj, exynos_gem->sgt); sgt 475 drivers/gpu/drm/exynos/exynos_drm_gem.c struct sg_table *sgt) sgt 487 drivers/gpu/drm/exynos/exynos_drm_gem.c exynos_gem->dma_addr = sg_dma_address(sgt->sgl); sgt 496 drivers/gpu/drm/exynos/exynos_drm_gem.c ret = drm_prime_sg_to_page_addr_arrays(sgt, exynos_gem->pages, NULL, sgt 501 drivers/gpu/drm/exynos/exynos_drm_gem.c exynos_gem->sgt = sgt; sgt 503 drivers/gpu/drm/exynos/exynos_drm_gem.c if (sgt->nents == 1) { sgt 52 drivers/gpu/drm/exynos/exynos_drm_gem.h struct sg_table *sgt; sgt 117 drivers/gpu/drm/exynos/exynos_drm_gem.h struct sg_table *sgt); sgt 231 drivers/gpu/drm/i915/gem/i915_gem_pages.c struct sg_table *sgt = obj->mm.pages; sgt 242 drivers/gpu/drm/i915/gem/i915_gem_pages.c return kmap(sg_page(sgt->sgl)); sgt 251 drivers/gpu/drm/i915/gem/i915_gem_pages.c for_each_sgt_page(page, sgt_iter, sgt) sgt 623 drivers/gpu/drm/i915/selftests/i915_vma.c struct sgt_iter sgt; sgt 626 drivers/gpu/drm/i915/selftests/i915_vma.c for_each_sgt_dma(dma, sgt, vma->pages) { sgt 116 drivers/gpu/drm/i915/selftests/scatterlist.c struct sgt_iter sgt; sgt 121 drivers/gpu/drm/i915/selftests/scatterlist.c for_each_sgt_page(page, sgt, &pt->st) { sgt 16 drivers/gpu/drm/lima/lima_gem_prime.c struct sg_table *sgt) sgt 21 drivers/gpu/drm/lima/lima_gem_prime.c bo = lima_bo_create(ldev, attach->dmabuf->size, 0, sgt); sgt 9 drivers/gpu/drm/lima/lima_gem_prime.h struct sg_table *sgt); sgt 12 drivers/gpu/drm/lima/lima_object.c if (bo->sgt) { sgt 14 drivers/gpu/drm/lima/lima_object.c drm_prime_gem_destroy(&bo->gem, bo->sgt); sgt 60 drivers/gpu/drm/lima/lima_object.c u32 flags, struct sg_table *sgt) sgt 78 drivers/gpu/drm/lima/lima_object.c if (sgt) { sgt 79 drivers/gpu/drm/lima/lima_object.c bo->sgt = sgt; sgt 88 drivers/gpu/drm/lima/lima_object.c sgt, bo->pages, bo->pages_dma_addr, npages); sgt 16 drivers/gpu/drm/lima/lima_object.h struct sg_table *sgt; sgt 30 drivers/gpu/drm/lima/lima_object.h u32 flags, struct sg_table *sgt); sgt 192 drivers/gpu/drm/mediatek/mtk_drm_gem.c struct sg_table *sgt; sgt 195 drivers/gpu/drm/mediatek/mtk_drm_gem.c sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); sgt 196 drivers/gpu/drm/mediatek/mtk_drm_gem.c if (!sgt) sgt 199 drivers/gpu/drm/mediatek/mtk_drm_gem.c ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie, sgt 204 drivers/gpu/drm/mediatek/mtk_drm_gem.c kfree(sgt); sgt 208 drivers/gpu/drm/mediatek/mtk_drm_gem.c return sgt; sgt 248 drivers/gpu/drm/mediatek/mtk_drm_gem.c struct sg_table *sgt; sgt 256 drivers/gpu/drm/mediatek/mtk_drm_gem.c sgt = mtk_gem_prime_get_sg_table(obj); sgt 257 drivers/gpu/drm/mediatek/mtk_drm_gem.c if (IS_ERR(sgt)) sgt 265 drivers/gpu/drm/mediatek/mtk_drm_gem.c for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) { sgt 274 drivers/gpu/drm/mediatek/mtk_drm_gem.c kfree((void *)sgt); sgt 243 drivers/gpu/drm/msm/msm_drv.h struct sg_table *sgt, int npages); sgt 323 drivers/gpu/drm/msm/msm_drv.h struct dma_buf *dmabuf, struct sg_table *sgt); sgt 56 drivers/gpu/drm/msm/msm_gem.c dma_sync_sg_for_device(dev, msm_obj->sgt->sgl, sgt 57 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt->nents, DMA_BIDIRECTIONAL); sgt 59 drivers/gpu/drm/msm/msm_gem.c dma_map_sg(dev, msm_obj->sgt->sgl, sgt 60 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt->nents, DMA_BIDIRECTIONAL); sgt 69 drivers/gpu/drm/msm/msm_gem.c dma_sync_sg_for_cpu(dev, msm_obj->sgt->sgl, sgt 70 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt->nents, DMA_BIDIRECTIONAL); sgt 72 drivers/gpu/drm/msm/msm_gem.c dma_unmap_sg(dev, msm_obj->sgt->sgl, sgt 73 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt->nents, DMA_BIDIRECTIONAL); sgt 129 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt = drm_prime_pages_to_sg(p, npages); sgt 130 drivers/gpu/drm/msm/msm_gem.c if (IS_ERR(msm_obj->sgt)) { sgt 131 drivers/gpu/drm/msm/msm_gem.c void *ptr = ERR_CAST(msm_obj->sgt); sgt 134 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt = NULL; sgt 165 drivers/gpu/drm/msm/msm_gem.c if (msm_obj->sgt) { sgt 173 drivers/gpu/drm/msm/msm_gem.c sg_free_table(msm_obj->sgt); sgt 174 drivers/gpu/drm/msm/msm_gem.c kfree(msm_obj->sgt); sgt 443 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt, obj->size >> PAGE_SHIFT); sgt 919 drivers/gpu/drm/msm/msm_gem.c drm_prime_gem_destroy(obj, msm_obj->sgt); sgt 1109 drivers/gpu/drm/msm/msm_gem.c struct dma_buf *dmabuf, struct sg_table *sgt) sgt 1134 drivers/gpu/drm/msm/msm_gem.c msm_obj->sgt = sgt; sgt 1142 drivers/gpu/drm/msm/msm_gem.c ret = drm_prime_sg_to_page_addr_arrays(sgt, msm_obj->pages, NULL, npages); sgt 71 drivers/gpu/drm/msm/msm_gem.h struct sg_table *sgt; sgt 96 drivers/gpu/drm/msm/msm_gem.h return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt && sgt 61 drivers/gpu/drm/msm/msm_gem_vma.c struct sg_table *sgt, int npages) sgt 78 drivers/gpu/drm/msm/msm_gem_vma.c ret = aspace->mmu->funcs->map(aspace->mmu, vma->iova, sgt, sgt 36 drivers/gpu/drm/msm/msm_gpummu.c struct sg_table *sgt, unsigned len, int prot) sgt 49 drivers/gpu/drm/msm/msm_gpummu.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 43 drivers/gpu/drm/msm/msm_iommu.c struct sg_table *sgt, unsigned len, int prot) sgt 48 drivers/gpu/drm/msm/msm_iommu.c ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot); sgt 15 drivers/gpu/drm/msm/msm_mmu.h int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt, sgt 76 drivers/gpu/drm/omapdrm/omap_gem.c struct sg_table *sgt; sgt 156 drivers/gpu/drm/omapdrm/omap_gem.c if ((omap_obj->flags & OMAP_BO_MEM_DMABUF) && omap_obj->sgt->nents == 1) sgt 1111 drivers/gpu/drm/omapdrm/omap_gem.c drm_prime_gem_destroy(obj, omap_obj->sgt); sgt 1229 drivers/gpu/drm/omapdrm/omap_gem.c struct sg_table *sgt) sgt 1237 drivers/gpu/drm/omapdrm/omap_gem.c if (sgt->orig_nents != 1 && !priv->has_dmm) sgt 1249 drivers/gpu/drm/omapdrm/omap_gem.c omap_obj->sgt = sgt; sgt 1251 drivers/gpu/drm/omapdrm/omap_gem.c if (sgt->orig_nents == 1) { sgt 1252 drivers/gpu/drm/omapdrm/omap_gem.c omap_obj->dma_addr = sg_dma_address(sgt->sgl); sgt 1270 drivers/gpu/drm/omapdrm/omap_gem.c for_each_sg_page(sgt->sgl, &iter, sgt->orig_nents, 0) { sgt 48 drivers/gpu/drm/omapdrm/omap_gem.h struct sg_table *sgt); sgt 152 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c struct sg_table *sgt; sgt 173 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c sgt = dma_buf_map_attachment(attach, DMA_TO_DEVICE); sgt 174 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c if (IS_ERR(sgt)) { sgt 175 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c ret = PTR_ERR(sgt); sgt 179 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c obj = omap_gem_new_dmabuf(dev, dma_buf->size, sgt); sgt 190 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c dma_buf_unmap_attachment(attach, sgt, DMA_TO_DEVICE); sgt 274 drivers/gpu/drm/panfrost/panfrost_gem.c struct sg_table *sgt) sgt 279 drivers/gpu/drm/panfrost/panfrost_gem.c obj = drm_gem_shmem_prime_import_sg_table(dev, attach, sgt); sgt 69 drivers/gpu/drm/panfrost/panfrost_gem.h struct sg_table *sgt); sgt 249 drivers/gpu/drm/panfrost/panfrost_mmu.c u64 iova, int prot, struct sg_table *sgt) sgt 256 drivers/gpu/drm/panfrost/panfrost_mmu.c for_each_sg(sgt->sgl, sgl, sgt->nents, count) { sgt 282 drivers/gpu/drm/panfrost/panfrost_mmu.c struct sg_table *sgt; sgt 291 drivers/gpu/drm/panfrost/panfrost_mmu.c sgt = drm_gem_shmem_get_pages_sgt(obj); sgt 292 drivers/gpu/drm/panfrost/panfrost_mmu.c if (WARN_ON(IS_ERR(sgt))) sgt 293 drivers/gpu/drm/panfrost/panfrost_mmu.c return PTR_ERR(sgt); sgt 296 drivers/gpu/drm/panfrost/panfrost_mmu.c prot, sgt); sgt 454 drivers/gpu/drm/panfrost/panfrost_mmu.c struct sg_table *sgt; sgt 514 drivers/gpu/drm/panfrost/panfrost_mmu.c sgt = &bo->sgts[page_offset / (SZ_2M / PAGE_SIZE)]; sgt 515 drivers/gpu/drm/panfrost/panfrost_mmu.c ret = sg_alloc_table_from_pages(sgt, pages + page_offset, sgt 520 drivers/gpu/drm/panfrost/panfrost_mmu.c if (!dma_map_sg(pfdev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL)) { sgt 526 drivers/gpu/drm/panfrost/panfrost_mmu.c IOMMU_WRITE | IOMMU_READ | IOMMU_NOEXEC, sgt); sgt 537 drivers/gpu/drm/panfrost/panfrost_mmu.c sg_free_table(sgt); sgt 209 drivers/gpu/drm/pl111/pl111_drv.c struct sg_table *sgt) sgt 221 drivers/gpu/drm/pl111/pl111_drv.c return drm_gem_cma_prime_import_sg_table(dev, attach, sgt); sgt 452 drivers/gpu/drm/qxl/qxl_drv.h struct sg_table *sgt); sgt 192 drivers/gpu/drm/rcar-du/rcar_du_vsp.c struct sg_table *sgt = &sg_tables[i]; sgt 194 drivers/gpu/drm/rcar-du/rcar_du_vsp.c ret = dma_get_sgtable(rcdu->dev, sgt, gem->vaddr, gem->paddr, sgt 199 drivers/gpu/drm/rcar-du/rcar_du_vsp.c ret = vsp1_du_map_sg(vsp->vsp, sgt); sgt 201 drivers/gpu/drm/rcar-du/rcar_du_vsp.c sg_free_table(sgt); sgt 211 drivers/gpu/drm/rcar-du/rcar_du_vsp.c struct sg_table *sgt = &sg_tables[i]; sgt 213 drivers/gpu/drm/rcar-du/rcar_du_vsp.c vsp1_du_unmap_sg(vsp->vsp, sgt); sgt 214 drivers/gpu/drm/rcar-du/rcar_du_vsp.c sg_free_table(sgt); sgt 247 drivers/gpu/drm/rcar-du/rcar_du_vsp.c struct sg_table *sgt = &sg_tables[i]; sgt 249 drivers/gpu/drm/rcar-du/rcar_du_vsp.c vsp1_du_unmap_sg(vsp->vsp, sgt); sgt 250 drivers/gpu/drm/rcar-du/rcar_du_vsp.c sg_free_table(sgt); sgt 38 drivers/gpu/drm/rockchip/rockchip_drm_gem.c ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl, sgt 39 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt->nents, prot); sgt 87 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages); sgt 88 drivers/gpu/drm/rockchip/rockchip_drm_gem.c if (IS_ERR(rk_obj->sgt)) { sgt 89 drivers/gpu/drm/rockchip/rockchip_drm_gem.c ret = PTR_ERR(rk_obj->sgt); sgt 100 drivers/gpu/drm/rockchip/rockchip_drm_gem.c for_each_sg(rk_obj->sgt->sgl, s, rk_obj->sgt->nents, i) sgt 103 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, sgt 115 drivers/gpu/drm/rockchip/rockchip_drm_gem.c sg_free_table(rk_obj->sgt); sgt 116 drivers/gpu/drm/rockchip/rockchip_drm_gem.c kfree(rk_obj->sgt); sgt 352 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_unmap_sg(drm->dev, rk_obj->sgt->sgl, sgt 353 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt->nents, DMA_BIDIRECTIONAL); sgt 355 drivers/gpu/drm/rockchip/rockchip_drm_gem.c drm_prime_gem_destroy(obj, rk_obj->sgt); sgt 440 drivers/gpu/drm/rockchip/rockchip_drm_gem.c struct sg_table *sgt; sgt 446 drivers/gpu/drm/rockchip/rockchip_drm_gem.c sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); sgt 447 drivers/gpu/drm/rockchip/rockchip_drm_gem.c if (!sgt) sgt 450 drivers/gpu/drm/rockchip/rockchip_drm_gem.c ret = dma_get_sgtable_attrs(drm->dev, sgt, rk_obj->kvaddr, sgt 455 drivers/gpu/drm/rockchip/rockchip_drm_gem.c kfree(sgt); sgt 459 drivers/gpu/drm/rockchip/rockchip_drm_gem.c return sgt; sgt 462 drivers/gpu/drm/rockchip/rockchip_drm_gem.c static unsigned long rockchip_sg_get_contiguous_size(struct sg_table *sgt, sgt 466 drivers/gpu/drm/rockchip/rockchip_drm_gem.c dma_addr_t expected = sg_dma_address(sgt->sgl); sgt 470 drivers/gpu/drm/rockchip/rockchip_drm_gem.c for_each_sg(sgt->sgl, s, count, i) { sgt 485 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt = sg; sgt 508 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt = sg; sgt 25 drivers/gpu/drm/rockchip/rockchip_drm_gem.h struct sg_table *sgt; sgt 30 drivers/gpu/drm/tegra/gem.c static dma_addr_t tegra_bo_pin(struct host1x_bo *bo, struct sg_table **sgt) sgt 34 drivers/gpu/drm/tegra/gem.c *sgt = obj->sgt; sgt 39 drivers/gpu/drm/tegra/gem.c static void tegra_bo_unpin(struct host1x_bo *bo, struct sg_table *sgt) sgt 138 drivers/gpu/drm/tegra/gem.c bo->size = iommu_map_sg(tegra->domain, bo->paddr, bo->sgt->sgl, sgt 139 drivers/gpu/drm/tegra/gem.c bo->sgt->nents, prot); sgt 206 drivers/gpu/drm/tegra/gem.c dma_unmap_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgt 209 drivers/gpu/drm/tegra/gem.c sg_free_table(bo->sgt); sgt 210 drivers/gpu/drm/tegra/gem.c kfree(bo->sgt); sgt 226 drivers/gpu/drm/tegra/gem.c bo->sgt = drm_prime_pages_to_sg(bo->pages, bo->num_pages); sgt 227 drivers/gpu/drm/tegra/gem.c if (IS_ERR(bo->sgt)) { sgt 228 drivers/gpu/drm/tegra/gem.c err = PTR_ERR(bo->sgt); sgt 232 drivers/gpu/drm/tegra/gem.c err = dma_map_sg(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgt 242 drivers/gpu/drm/tegra/gem.c sg_free_table(bo->sgt); sgt 243 drivers/gpu/drm/tegra/gem.c kfree(bo->sgt); sgt 352 drivers/gpu/drm/tegra/gem.c bo->sgt = dma_buf_map_attachment(attach, DMA_TO_DEVICE); sgt 353 drivers/gpu/drm/tegra/gem.c if (IS_ERR(bo->sgt)) { sgt 354 drivers/gpu/drm/tegra/gem.c err = PTR_ERR(bo->sgt); sgt 363 drivers/gpu/drm/tegra/gem.c if (bo->sgt->nents > 1) { sgt 368 drivers/gpu/drm/tegra/gem.c bo->paddr = sg_dma_address(bo->sgt->sgl); sgt 376 drivers/gpu/drm/tegra/gem.c if (!IS_ERR_OR_NULL(bo->sgt)) sgt 377 drivers/gpu/drm/tegra/gem.c dma_buf_unmap_attachment(attach, bo->sgt, DMA_TO_DEVICE); sgt 396 drivers/gpu/drm/tegra/gem.c dma_buf_unmap_attachment(gem->import_attach, bo->sgt, sgt 504 drivers/gpu/drm/tegra/gem.c struct sg_table *sgt; sgt 506 drivers/gpu/drm/tegra/gem.c sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); sgt 507 drivers/gpu/drm/tegra/gem.c if (!sgt) sgt 514 drivers/gpu/drm/tegra/gem.c if (sg_alloc_table(sgt, bo->num_pages, GFP_KERNEL)) sgt 517 drivers/gpu/drm/tegra/gem.c for_each_sg(sgt->sgl, sg, bo->num_pages, i) sgt 520 drivers/gpu/drm/tegra/gem.c if (dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir) == 0) sgt 523 drivers/gpu/drm/tegra/gem.c if (sg_alloc_table(sgt, 1, GFP_KERNEL)) sgt 526 drivers/gpu/drm/tegra/gem.c sg_dma_address(sgt->sgl) = bo->paddr; sgt 527 drivers/gpu/drm/tegra/gem.c sg_dma_len(sgt->sgl) = gem->size; sgt 530 drivers/gpu/drm/tegra/gem.c return sgt; sgt 533 drivers/gpu/drm/tegra/gem.c sg_free_table(sgt); sgt 534 drivers/gpu/drm/tegra/gem.c kfree(sgt); sgt 539 drivers/gpu/drm/tegra/gem.c struct sg_table *sgt, sgt 546 drivers/gpu/drm/tegra/gem.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, dir); sgt 548 drivers/gpu/drm/tegra/gem.c sg_free_table(sgt); sgt 549 drivers/gpu/drm/tegra/gem.c kfree(sgt); sgt 565 drivers/gpu/drm/tegra/gem.c dma_sync_sg_for_cpu(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgt 579 drivers/gpu/drm/tegra/gem.c dma_sync_sg_for_device(drm->dev, bo->sgt->sgl, bo->sgt->nents, sgt 33 drivers/gpu/drm/tegra/gem.h struct sg_table *sgt; sgt 16 drivers/gpu/drm/udl/udl_dmabuf.c struct sg_table sgt; sgt 43 drivers/gpu/drm/udl/udl_dmabuf.c struct sg_table *sgt; sgt 51 drivers/gpu/drm/udl/udl_dmabuf.c sgt = &udl_attach->sgt; sgt 54 drivers/gpu/drm/udl/udl_dmabuf.c dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, sgt 57 drivers/gpu/drm/udl/udl_dmabuf.c sg_free_table(sgt); sgt 70 drivers/gpu/drm/udl/udl_dmabuf.c struct sg_table *sgt = NULL; sgt 80 drivers/gpu/drm/udl/udl_dmabuf.c return &udl_attach->sgt; sgt 97 drivers/gpu/drm/udl/udl_dmabuf.c sgt = &udl_attach->sgt; sgt 99 drivers/gpu/drm/udl/udl_dmabuf.c ret = sg_alloc_table(sgt, obj->sg->orig_nents, GFP_KERNEL); sgt 108 drivers/gpu/drm/udl/udl_dmabuf.c wr = sgt->sgl; sgt 109 drivers/gpu/drm/udl/udl_dmabuf.c for (i = 0; i < sgt->orig_nents; ++i) { sgt 116 drivers/gpu/drm/udl/udl_dmabuf.c nents = dma_map_sg(attach->dev, sgt->sgl, sgt->orig_nents, dir); sgt 119 drivers/gpu/drm/udl/udl_dmabuf.c sg_free_table(sgt); sgt 120 drivers/gpu/drm/udl/udl_dmabuf.c sgt = ERR_PTR(-EIO); sgt 131 drivers/gpu/drm/udl/udl_dmabuf.c return sgt; sgt 135 drivers/gpu/drm/udl/udl_dmabuf.c struct sg_table *sgt, sgt 92 drivers/gpu/drm/v3d/v3d_bo.c struct sg_table *sgt; sgt 98 drivers/gpu/drm/v3d/v3d_bo.c sgt = drm_gem_shmem_get_pages_sgt(&bo->base.base); sgt 99 drivers/gpu/drm/v3d/v3d_bo.c if (IS_ERR(sgt)) sgt 100 drivers/gpu/drm/v3d/v3d_bo.c return PTR_ERR(sgt); sgt 151 drivers/gpu/drm/v3d/v3d_bo.c struct sg_table *sgt) sgt 156 drivers/gpu/drm/v3d/v3d_bo.c obj = drm_gem_shmem_prime_import_sg_table(dev, attach, sgt); sgt 301 drivers/gpu/drm/v3d/v3d_drv.h struct sg_table *sgt); sgt 96 drivers/gpu/drm/v3d/v3d_mmu.c for_each_sg(shmem_obj->sgt->sgl, sgl, shmem_obj->sgt->nents, count) { sgt 785 drivers/gpu/drm/vc4/vc4_bo.c struct sg_table *sgt) sgt 789 drivers/gpu/drm/vc4/vc4_bo.c obj = drm_gem_cma_prime_import_sg_table(dev, attach, sgt); sgt 735 drivers/gpu/drm/vc4/vc4_drv.h struct sg_table *sgt); sgt 373 drivers/gpu/drm/virtio/virtgpu_drv.h struct sg_table *sgt); sgt 299 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h struct sg_table *sgt; sgt 245 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c struct sg_table sgt; sgt 345 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c __sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl, sgt 346 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vsgt->sgt->orig_nents, p_offset); sgt 365 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c dma_unmap_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.nents, sgt 367 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vmw_tt->sgt.nents = vmw_tt->sgt.orig_nents; sgt 388 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c ret = dma_map_sg(dev, vmw_tt->sgt.sgl, vmw_tt->sgt.orig_nents, sgt 393 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vmw_tt->sgt.nents = ret; sgt 430 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vsgt->sgt = &vmw_tt->sgt; sgt 445 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c (&vmw_tt->sgt, vsgt->pages, vsgt->num_pages, 0, sgt 452 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c if (vsgt->num_pages > vmw_tt->sgt.nents) { sgt 455 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vmw_tt->sgt.nents); sgt 484 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c sg_free_table(vmw_tt->vsgt.sgt); sgt 485 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vmw_tt->vsgt.sgt = NULL; sgt 504 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c if (!vmw_tt->vsgt.sgt) sgt 511 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c sg_free_table(vmw_tt->vsgt.sgt); sgt 512 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vmw_tt->vsgt.sgt = NULL; sgt 188 drivers/gpu/drm/xen/xen_drm_front_gem.c struct sg_table *sgt) sgt 204 drivers/gpu/drm/xen/xen_drm_front_gem.c xen_obj->sgt_imported = sgt; sgt 206 drivers/gpu/drm/xen/xen_drm_front_gem.c ret = drm_prime_sg_to_page_addr_arrays(sgt, xen_obj->pages, sgt 218 drivers/gpu/drm/xen/xen_drm_front_gem.c size, sgt->nents); sgt 27 drivers/gpu/drm/xen/xen_drm_front_gem.h struct sg_table *sgt); sgt 109 drivers/gpu/host1x/job.c struct sg_table *sgt; sgt 118 drivers/gpu/host1x/job.c phys_addr = host1x_bo_pin(reloc->target.bo, &sgt); sgt 122 drivers/gpu/host1x/job.c job->unpins[job->num_unpins].sgt = sgt; sgt 130 drivers/gpu/host1x/job.c struct sg_table *sgt; sgt 142 drivers/gpu/host1x/job.c phys_addr = host1x_bo_pin(g->bo, &sgt); sgt 145 drivers/gpu/host1x/job.c for_each_sg(sgt->sgl, sg, sgt->nents, j) sgt 159 drivers/gpu/host1x/job.c sgt->sgl, sgt->nents, IOMMU_READ); sgt 176 drivers/gpu/host1x/job.c job->unpins[job->num_unpins].sgt = sgt; sgt 570 drivers/gpu/host1x/job.c host1x_bo_unpin(unpin->bo, unpin->sgt); sgt 21 drivers/gpu/host1x/job.h struct sg_table *sgt; sgt 199 drivers/hsi/clients/cmt_speech.c u32 *data = sg_virt(msg->sgt.sgl); sgt 205 drivers/hsi/clients/cmt_speech.c u32 *data = sg_virt(msg->sgt.sgl); sgt 259 drivers/hsi/clients/cmt_speech.c kfree(sg_virt(msg->sgt.sgl)); sgt 281 drivers/hsi/clients/cmt_speech.c sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); sgt 472 drivers/hsi/clients/cmt_speech.c msg->sgt.nents = 1; sgt 502 drivers/hsi/clients/cmt_speech.c msg->sgt.nents = 0; sgt 547 drivers/hsi/clients/cmt_speech.c msg->sgt.nents = 1; sgt 624 drivers/hsi/clients/cmt_speech.c sg_init_one(msg->sgt.sgl, address, hi->buf_size); sgt 625 drivers/hsi/clients/cmt_speech.c msg->sgt.nents = 1; sgt 667 drivers/hsi/clients/cmt_speech.c sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0); sgt 668 drivers/hsi/clients/cmt_speech.c rxmsg->sgt.nents = 0; sgt 719 drivers/hsi/clients/cmt_speech.c sg_init_one(txmsg->sgt.sgl, address, hi->buf_size); sgt 147 drivers/hsi/clients/hsi_char.c kfree(sg_virt(msg->sgt.sgl)); sgt 186 drivers/hsi/clients/hsi_char.c sg_init_one(msg->sgt.sgl, buf, alloc_size); sgt 217 drivers/hsi/clients/hsi_char.c return msg->sgt.sgl->length; sgt 222 drivers/hsi/clients/hsi_char.c msg->sgt.sgl->length = len; sgt 455 drivers/hsi/clients/hsi_char.c sg_virt(msg->sgt.sgl), hsc_msg_len_get(msg)); sgt 491 drivers/hsi/clients/hsi_char.c if (copy_from_user(sg_virt(msg->sgt.sgl), (void __user *)buf, len)) { sgt 157 drivers/hsi/clients/ssi_protocol.c data = sg_virt(msg->sgt.sgl); sgt 165 drivers/hsi/clients/ssi_protocol.c data = sg_virt(msg->sgt.sgl); sgt 176 drivers/hsi/clients/ssi_protocol.c BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); sgt 178 drivers/hsi/clients/ssi_protocol.c sg = msg->sgt.sgl; sgt 249 drivers/hsi/clients/ssi_protocol.c kfree(sg_virt(msg->sgt.sgl)); sgt 269 drivers/hsi/clients/ssi_protocol.c sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); sgt 192 drivers/hsi/controllers/omap_ssi_core.c dma_unmap_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, dir); sgt 215 drivers/hsi/controllers/omap_ssi_core.c msg->actual_len = sg_dma_len(msg->sgt.sgl); sgt 200 drivers/hsi/controllers/omap_ssi_port.c omap_ssi->gdd_trn[lch].sg = msg->sgt.sgl; sgt 231 drivers/hsi/controllers/omap_ssi_port.c err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, sgt 246 drivers/hsi/controllers/omap_ssi_port.c d_addr = sg_dma_address(msg->sgt.sgl); sgt 248 drivers/hsi/controllers/omap_ssi_port.c err = dma_map_sg(&ssi->device, msg->sgt.sgl, msg->sgt.nents, sgt 261 drivers/hsi/controllers/omap_ssi_port.c s_addr = sg_dma_address(msg->sgt.sgl); sgt 272 drivers/hsi/controllers/omap_ssi_port.c writew_relaxed(SSI_BYTES_TO_FRAMES(msg->sgt.sgl->length), sgt 330 drivers/hsi/controllers/omap_ssi_port.c if ((msg->sgt.nents) && (msg->sgt.sgl->length > sizeof(u32))) sgt 386 drivers/hsi/controllers/omap_ssi_port.c if (msg->sgt.nents > 1) sgt 445 drivers/hsi/controllers/omap_ssi_port.c msg->channel, msg, msg->sgt.sgl->length, sgt 905 drivers/hsi/controllers/omap_ssi_port.c if ((!msg->sgt.nents) || (!msg->sgt.sgl->length)) { sgt 914 drivers/hsi/controllers/omap_ssi_port.c buf = sg_virt(msg->sgt.sgl) + msg->actual_len; sgt 924 drivers/hsi/controllers/omap_ssi_port.c if (msg->actual_len >= msg->sgt.sgl->length) sgt 520 drivers/hsi/hsi_core.c sg_free_table(&msg->sgt); sgt 548 drivers/hsi/hsi_core.c err = sg_alloc_table(&msg->sgt, nents, flags); sgt 51 drivers/hwtracing/intel_th/msu-sink.c static int msu_sink_alloc_window(void *data, struct sg_table **sgt, size_t size) sgt 64 drivers/hwtracing/intel_th/msu-sink.c ret = sg_alloc_table(*sgt, nents, GFP_KERNEL); sgt 68 drivers/hwtracing/intel_th/msu-sink.c priv->sgts[priv->nr_sgts++] = *sgt; sgt 70 drivers/hwtracing/intel_th/msu-sink.c for_each_sg((*sgt)->sgl, sg_ptr, nents, i) { sgt 81 drivers/hwtracing/intel_th/msu-sink.c static void msu_sink_free_window(void *data, struct sg_table *sgt) sgt 87 drivers/hwtracing/intel_th/msu-sink.c for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) { sgt 92 drivers/hwtracing/intel_th/msu-sink.c sg_free_table(sgt); sgt 96 drivers/hwtracing/intel_th/msu-sink.c static int msu_sink_ready(void *data, struct sg_table *sgt, size_t bytes) sgt 100 drivers/hwtracing/intel_th/msu-sink.c intel_th_msc_window_unlock(priv->dev, sgt); sgt 76 drivers/hwtracing/intel_th/msu.c struct sg_table *sgt; sgt 276 drivers/hwtracing/intel_th/msu.c return win->sgt->sgl; sgt 326 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgt 350 drivers/hwtracing/intel_th/msu.c msc_find_window(struct msc *msc, struct sg_table *sgt, bool nonempty) sgt 364 drivers/hwtracing/intel_th/msu.c if (win->sgt == sgt) sgt 394 drivers/hwtracing/intel_th/msu.c win = msc_find_window(msc, msc_next_window(msc->cur_win)->sgt, true); sgt 421 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgt 660 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgt 845 drivers/hwtracing/intel_th/msu.c msc->mbuf->ready(msc->mbuf_priv, win->sgt, sgt 996 drivers/hwtracing/intel_th/msu.c ret = sg_alloc_table(win->sgt, nr_segs, GFP_KERNEL); sgt 1000 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { sgt 1013 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, i, ret) sgt 1017 drivers/hwtracing/intel_th/msu.c sg_free_table(win->sgt); sgt 1028 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { sgt 1040 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) { sgt 1075 drivers/hwtracing/intel_th/msu.c win->sgt = &win->_sgt; sgt 1088 drivers/hwtracing/intel_th/msu.c ret = msc->mbuf->alloc_window(msc->mbuf_priv, &win->sgt, sgt 1123 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, i) { sgt 1130 drivers/hwtracing/intel_th/msu.c sg_free_table(win->sgt); sgt 1154 drivers/hwtracing/intel_th/msu.c msc->mbuf->free_window(msc->mbuf_priv, win->sgt); sgt 1190 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgt 1385 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { sgt 1707 drivers/hwtracing/intel_th/msu.c void intel_th_msc_window_unlock(struct device *dev, struct sg_table *sgt) sgt 1712 drivers/hwtracing/intel_th/msu.c if (!sgt) sgt 1715 drivers/hwtracing/intel_th/msu.c win = msc_find_window(msc, sgt, false); sgt 1770 drivers/hwtracing/intel_th/msu.c msc->mbuf->ready(msc->mbuf_priv, win->sgt, sgt 533 drivers/input/touchscreen/sur40.c struct sg_table *sgt; sgt 581 drivers/input/touchscreen/sur40.c sgt = vb2_dma_sg_plane_desc(&new_buf->vb.vb2_buf, 0); sgt 585 drivers/input/touchscreen/sur40.c sgt->sgl, sgt->nents, sur40->pix_fmt.sizeimage, 0); sgt 575 drivers/iommu/dma-iommu.c struct sg_table sgt; sgt 602 drivers/iommu/dma-iommu.c if (sg_alloc_table_from_pages(&sgt, pages, count, 0, size, GFP_KERNEL)) sgt 609 drivers/iommu/dma-iommu.c for_each_sg(sgt.sgl, sg, sgt.orig_nents, i) sgt 613 drivers/iommu/dma-iommu.c if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, ioprot) sgt 623 drivers/iommu/dma-iommu.c sg_free_table(&sgt); sgt 629 drivers/iommu/dma-iommu.c sg_free_table(&sgt); sgt 1059 drivers/iommu/dma-iommu.c static int iommu_dma_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 1070 drivers/iommu/dma-iommu.c return sg_alloc_table_from_pages(sgt, pages, sgt 1080 drivers/iommu/dma-iommu.c ret = sg_alloc_table(sgt, 1, GFP_KERNEL); sgt 1082 drivers/iommu/dma-iommu.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgt 49 drivers/media/common/videobuf2/videobuf2-dma-contig.c static unsigned long vb2_dc_get_contiguous_size(struct sg_table *sgt) sgt 52 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_addr_t expected = sg_dma_address(sgt->sgl); sgt 56 drivers/media/common/videobuf2/videobuf2-dma-contig.c for_each_sg(sgt->sgl, s, sgt->nents, i) { sgt 96 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt = buf->dma_sgt; sgt 99 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (!sgt || buf->db_attach) sgt 102 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, sgt 109 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt = buf->dma_sgt; sgt 112 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (!sgt || buf->db_attach) sgt 115 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); sgt 215 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table sgt; sgt 225 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 233 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = &attach->sgt; sgt 237 drivers/media/common/videobuf2/videobuf2-dma-contig.c ret = sg_alloc_table(sgt, buf->sgt_base->orig_nents, GFP_KERNEL); sgt 244 drivers/media/common/videobuf2/videobuf2-dma-contig.c wr = sgt->sgl; sgt 245 drivers/media/common/videobuf2/videobuf2-dma-contig.c for (i = 0; i < sgt->orig_nents; ++i) { sgt 261 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 266 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = &attach->sgt; sgt 276 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 278 drivers/media/common/videobuf2/videobuf2-dma-contig.c sg_free_table(sgt); sgt 289 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 293 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = &attach->sgt; sgt 297 drivers/media/common/videobuf2/videobuf2-dma-contig.c return sgt; sgt 302 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 311 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt->nents = dma_map_sg_attrs(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 313 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (!sgt->nents) { sgt 323 drivers/media/common/videobuf2/videobuf2-dma-contig.c return sgt; sgt 327 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt, enum dma_data_direction dma_dir) sgt 372 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 374 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); sgt 375 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (!sgt) { sgt 380 drivers/media/common/videobuf2/videobuf2-dma-contig.c ret = dma_get_sgtable_attrs(buf->dev, sgt, buf->cookie, buf->dma_addr, sgt 384 drivers/media/common/videobuf2/videobuf2-dma-contig.c kfree(sgt); sgt 388 drivers/media/common/videobuf2/videobuf2-dma-contig.c return sgt; sgt 425 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt = buf->dma_sgt; sgt 429 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (sgt) { sgt 434 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 443 drivers/media/common/videobuf2/videobuf2-dma-contig.c sg_free_table(sgt); sgt 444 drivers/media/common/videobuf2/videobuf2-dma-contig.c kfree(sgt); sgt 461 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 514 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); sgt 515 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (!sgt) { sgt 521 drivers/media/common/videobuf2/videobuf2-dma-contig.c ret = sg_alloc_table_from_pages(sgt, frame_vector_pages(vec), n_pages, sgt 532 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 534 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (sgt->nents <= 0) { sgt 540 drivers/media/common/videobuf2/videobuf2-dma-contig.c contig_size = vb2_dc_get_contiguous_size(sgt); sgt 548 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_addr = sg_dma_address(sgt->sgl); sgt 549 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_sgt = sgt; sgt 556 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 560 drivers/media/common/videobuf2/videobuf2-dma-contig.c sg_free_table(sgt); sgt 563 drivers/media/common/videobuf2/videobuf2-dma-contig.c kfree(sgt); sgt 581 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt; sgt 595 drivers/media/common/videobuf2/videobuf2-dma-contig.c sgt = dma_buf_map_attachment(buf->db_attach, buf->dma_dir); sgt 596 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (IS_ERR(sgt)) { sgt 602 drivers/media/common/videobuf2/videobuf2-dma-contig.c contig_size = vb2_dc_get_contiguous_size(sgt); sgt 606 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir); sgt 610 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_addr = sg_dma_address(sgt->sgl); sgt 611 drivers/media/common/videobuf2/videobuf2-dma-contig.c buf->dma_sgt = sgt; sgt 620 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct sg_table *sgt = buf->dma_sgt; sgt 627 drivers/media/common/videobuf2/videobuf2-dma-contig.c if (WARN_ON(!sgt)) { sgt 636 drivers/media/common/videobuf2/videobuf2-dma-contig.c dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir); sgt 104 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 140 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = &buf->sg_table; sgt 145 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 147 drivers/media/common/videobuf2/videobuf2-dma-sg.c if (!sgt->nents) sgt 177 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt = &buf->sg_table; sgt 183 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 199 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt = buf->dma_sgt; sgt 205 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->orig_nents, sgt 212 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt = buf->dma_sgt; sgt 218 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir); sgt 226 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 256 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = &buf->sg_table; sgt 261 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, sgt 263 drivers/media/common/videobuf2/videobuf2-dma-sg.c if (!sgt->nents) sgt 284 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt = &buf->sg_table; sgt 289 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg_attrs(buf->dev, sgt->sgl, sgt->orig_nents, buf->dma_dir, sgt 359 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table sgt; sgt 369 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 377 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = &attach->sgt; sgt 381 drivers/media/common/videobuf2/videobuf2-dma-sg.c ret = sg_alloc_table(sgt, buf->dma_sgt->orig_nents, GFP_KERNEL); sgt 388 drivers/media/common/videobuf2/videobuf2-dma-sg.c wr = sgt->sgl; sgt 389 drivers/media/common/videobuf2/videobuf2-dma-sg.c for (i = 0; i < sgt->orig_nents; ++i) { sgt 405 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 410 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = &attach->sgt; sgt 414 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 416 drivers/media/common/videobuf2/videobuf2-dma-sg.c sg_free_table(sgt); sgt 427 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 431 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = &attach->sgt; sgt 435 drivers/media/common/videobuf2/videobuf2-dma-sg.c return sgt; sgt 440 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 446 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 448 drivers/media/common/videobuf2/videobuf2-dma-sg.c if (!sgt->nents) { sgt 458 drivers/media/common/videobuf2/videobuf2-dma-sg.c return sgt; sgt 462 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt, enum dma_data_direction dma_dir) sgt 535 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt; sgt 548 drivers/media/common/videobuf2/videobuf2-dma-sg.c sgt = dma_buf_map_attachment(buf->db_attach, buf->dma_dir); sgt 549 drivers/media/common/videobuf2/videobuf2-dma-sg.c if (IS_ERR(sgt)) { sgt 554 drivers/media/common/videobuf2/videobuf2-dma-sg.c buf->dma_sgt = sgt; sgt 563 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct sg_table *sgt = buf->dma_sgt; sgt 570 drivers/media/common/videobuf2/videobuf2-dma-sg.c if (WARN_ON(!sgt)) { sgt 579 drivers/media/common/videobuf2/videobuf2-dma-sg.c dma_buf_unmap_attachment(buf->db_attach, sgt, buf->dma_dir); sgt 207 drivers/media/common/videobuf2/videobuf2-vmalloc.c struct sg_table sgt; sgt 217 drivers/media/common/videobuf2/videobuf2-vmalloc.c struct sg_table *sgt; sgt 227 drivers/media/common/videobuf2/videobuf2-vmalloc.c sgt = &attach->sgt; sgt 228 drivers/media/common/videobuf2/videobuf2-vmalloc.c ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL); sgt 233 drivers/media/common/videobuf2/videobuf2-vmalloc.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 237 drivers/media/common/videobuf2/videobuf2-vmalloc.c sg_free_table(sgt); sgt 254 drivers/media/common/videobuf2/videobuf2-vmalloc.c struct sg_table *sgt; sgt 259 drivers/media/common/videobuf2/videobuf2-vmalloc.c sgt = &attach->sgt; sgt 263 drivers/media/common/videobuf2/videobuf2-vmalloc.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 265 drivers/media/common/videobuf2/videobuf2-vmalloc.c sg_free_table(sgt); sgt 276 drivers/media/common/videobuf2/videobuf2-vmalloc.c struct sg_table *sgt; sgt 280 drivers/media/common/videobuf2/videobuf2-vmalloc.c sgt = &attach->sgt; sgt 284 drivers/media/common/videobuf2/videobuf2-vmalloc.c return sgt; sgt 289 drivers/media/common/videobuf2/videobuf2-vmalloc.c dma_unmap_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 295 drivers/media/common/videobuf2/videobuf2-vmalloc.c sgt->nents = dma_map_sg(db_attach->dev, sgt->sgl, sgt->orig_nents, sgt 297 drivers/media/common/videobuf2/videobuf2-vmalloc.c if (!sgt->nents) { sgt 307 drivers/media/common/videobuf2/videobuf2-vmalloc.c return sgt; sgt 311 drivers/media/common/videobuf2/videobuf2-vmalloc.c struct sg_table *sgt, enum dma_data_direction dma_dir) sgt 1578 drivers/media/pci/cx23885/cx23885-core.c struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0); sgt 1586 drivers/media/pci/cx23885/cx23885-core.c sgt->sgl, sgt 134 drivers/media/pci/cx23885/cx23885-vbi.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 145 drivers/media/pci/cx23885/cx23885-vbi.c sgt->sgl, sgt 347 drivers/media/pci/cx23885/cx23885-video.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 359 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, 0, UNSET, sgt 364 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, UNSET, 0, sgt 392 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, line0_offset, sgt 399 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, sgt 406 drivers/media/pci/cx23885/cx23885-video.c sgt->sgl, sgt 149 drivers/media/pci/cx25821/cx25821-video.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 180 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, 0, UNSET, sgt 185 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, UNSET, 0, sgt 194 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, line0_offset, sgt 200 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, sgt 206 drivers/media/pci/cx25821/cx25821-video.c sgt->sgl, sgt 217 drivers/media/pci/cx88/cx88-mpeg.c struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0); sgt 225 drivers/media/pci/cx88/cx88-mpeg.c rc = cx88_risc_databuffer(dev->pci, risc, sgt->sgl, sgt 134 drivers/media/pci/cx88/cx88-vbi.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 147 drivers/media/pci/cx88/cx88-vbi.c cx88_risc_buffer(dev->pci, &buf->risc, sgt->sgl, sgt 440 drivers/media/pci/cx88/cx88-video.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 451 drivers/media/pci/cx88/cx88-video.c sgt->sgl, 0, UNSET, sgt 456 drivers/media/pci/cx88/cx88-video.c sgt->sgl, UNSET, 0, sgt 461 drivers/media/pci/cx88/cx88-video.c sgt->sgl, sgt 468 drivers/media/pci/cx88/cx88-video.c sgt->sgl, sgt 476 drivers/media/pci/cx88/cx88-video.c sgt->sgl, 0, buf->bpl, sgt 454 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 465 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c return solo_send_desc(solo_enc, solo_enc->jpeg_len, sgt, sgt 476 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 501 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c return solo_send_desc(solo_enc, skip, sgt, frame_off, frame_size, sgt 737 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 743 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c sg_copy_from_buffer(sgt->sgl, sgt->nents, sgt 747 drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c sg_copy_from_buffer(sgt->sgl, sgt->nents, sgt 359 drivers/media/platform/omap3isp/ispccdc.c sg_free_table(&req->table.sgt); sgt 443 drivers/media/platform/omap3isp/ispccdc.c ret = dma_get_sgtable(isp->dev, &req->table.sgt, sgt 449 drivers/media/platform/omap3isp/ispccdc.c dma_sync_sg_for_cpu(isp->dev, req->table.sgt.sgl, sgt 450 drivers/media/platform/omap3isp/ispccdc.c req->table.sgt.nents, DMA_TO_DEVICE); sgt 458 drivers/media/platform/omap3isp/ispccdc.c dma_sync_sg_for_device(isp->dev, req->table.sgt.sgl, sgt 459 drivers/media/platform/omap3isp/ispccdc.c req->table.sgt.nents, DMA_TO_DEVICE); sgt 57 drivers/media/platform/omap3isp/ispccdc.h struct sg_table sgt; sgt 164 drivers/media/platform/omap3isp/ispstat.c dma_sync_sg_for_device(stat->isp->dev, buf->sgt.sgl, sgt 165 drivers/media/platform/omap3isp/ispstat.c buf->sgt.nents, DMA_FROM_DEVICE); sgt 174 drivers/media/platform/omap3isp/ispstat.c dma_sync_sg_for_cpu(stat->isp->dev, buf->sgt.sgl, sgt 175 drivers/media/platform/omap3isp/ispstat.c buf->sgt.nents, DMA_FROM_DEVICE); sgt 347 drivers/media/platform/omap3isp/ispstat.c sg_free_table(&buf->sgt); sgt 375 drivers/media/platform/omap3isp/ispstat.c ret = dma_get_sgtable(dev, &buf->sgt, buf->virt_addr, buf->dma_addr, sgt 36 drivers/media/platform/omap3isp/ispstat.h struct sg_table sgt; sgt 1375 drivers/media/platform/pxa_camera.c struct sg_table *sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 1394 drivers/media/platform/pxa_camera.c ret = sg_split(sgt->sgl, sgt->nents, 0, nb_channels, sgt 356 drivers/media/platform/qcom/camss/camss-video.c struct sg_table *sgt; sgt 360 drivers/media/platform/qcom/camss/camss-video.c sgt = vb2_dma_sg_plane_desc(vb, i); sgt 361 drivers/media/platform/qcom/camss/camss-video.c if (!sgt) sgt 364 drivers/media/platform/qcom/camss/camss-video.c buffer->addr[i] = sg_dma_address(sgt->sgl); sgt 1035 drivers/media/platform/qcom/venus/helpers.c struct sg_table *sgt; sgt 1037 drivers/media/platform/qcom/venus/helpers.c sgt = vb2_dma_sg_plane_desc(vb, 0); sgt 1038 drivers/media/platform/qcom/venus/helpers.c if (!sgt) sgt 1042 drivers/media/platform/qcom/venus/helpers.c buf->dma_addr = sg_dma_address(sgt->sgl); sgt 117 drivers/media/platform/rockchip/rga/rga-buf.c struct sg_table *sgt; sgt 129 drivers/media/platform/rockchip/rga/rga-buf.c sgt = vb2_plane_cookie(vb, 0); sgt 131 drivers/media/platform/rockchip/rga/rga-buf.c for_each_sg(sgt->sgl, sgl, sgt->nents, i) { sgt 347 drivers/media/platform/via-camera.c struct sg_table *sgt; sgt 367 drivers/media/platform/via-camera.c sgt = vb2_dma_sg_plane_desc(&vb->vbuf.vb2_buf, 0); sgt 369 drivers/media/platform/via-camera.c viafb_dma_copy_out_sg(cam->cb_offsets[bufn], sgt->sgl, sgt->nents); sgt 906 drivers/media/platform/vsp1/vsp1_drm.c int vsp1_du_map_sg(struct device *dev, struct sg_table *sgt) sgt 915 drivers/media/platform/vsp1/vsp1_drm.c return dma_map_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents, sgt 920 drivers/media/platform/vsp1/vsp1_drm.c void vsp1_du_unmap_sg(struct device *dev, struct sg_table *sgt) sgt 924 drivers/media/platform/vsp1/vsp1_drm.c dma_unmap_sg_attrs(vsp1->bus_master, sgt->sgl, sgt->nents, sgt 130 drivers/misc/fastrpc.c struct sg_table sgt; sgt 493 drivers/misc/fastrpc.c table = &a->sgt; sgt 526 drivers/misc/fastrpc.c ret = dma_get_sgtable(buffer->dev, &a->sgt, buffer->virt, sgt 554 drivers/misc/fastrpc.c sg_free_table(&a->sgt); sgt 3136 drivers/misc/habanalabs/goya/goya.c u32 goya_get_dma_desc_list_size(struct hl_device *hdev, struct sg_table *sgt) sgt 3145 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgt 3153 drivers/misc/habanalabs/goya/goya.c while ((count + 1) < sgt->nents) { sgt 3200 drivers/misc/habanalabs/goya/goya.c rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl, sgt 3201 drivers/misc/habanalabs/goya/goya.c userptr->sgt->nents, dir); sgt 3212 drivers/misc/habanalabs/goya/goya.c goya_get_dma_desc_list_size(hdev, userptr->sgt); sgt 3603 drivers/misc/habanalabs/goya/goya.c struct sg_table *sgt; sgt 3654 drivers/misc/habanalabs/goya/goya.c sgt = userptr->sgt; sgt 3657 drivers/misc/habanalabs/goya/goya.c for_each_sg(sgt->sgl, sg, sgt->nents, count) { sgt 3664 drivers/misc/habanalabs/goya/goya.c while ((count + 1) < sgt->nents) { sgt 228 drivers/misc/habanalabs/goya/goyaP.h u32 goya_get_dma_desc_list_size(struct hl_device *hdev, struct sg_table *sgt); sgt 546 drivers/misc/habanalabs/habanalabs.h struct sg_table *sgt); sgt 699 drivers/misc/habanalabs/habanalabs.h struct sg_table *sgt; sgt 193 drivers/misc/habanalabs/memory.c rc = hdev->asic_funcs->asic_dma_map_sg(hdev, userptr->sgt->sgl, sgt 194 drivers/misc/habanalabs/memory.c userptr->sgt->nents, DMA_BIDIRECTIONAL); sgt 673 drivers/misc/habanalabs/memory.c for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) { sgt 702 drivers/misc/habanalabs/memory.c for_each_sg(userptr->sgt->sgl, sg, userptr->sgt->nents, i) { sgt 1292 drivers/misc/habanalabs/memory.c userptr->sgt = kzalloc(sizeof(*userptr->sgt), GFP_ATOMIC); sgt 1293 drivers/misc/habanalabs/memory.c if (!userptr->sgt) { sgt 1298 drivers/misc/habanalabs/memory.c rc = sg_alloc_table_from_pages(userptr->sgt, sgt 1311 drivers/misc/habanalabs/memory.c kfree(userptr->sgt); sgt 1337 drivers/misc/habanalabs/memory.c userptr->sgt->sgl, sgt 1338 drivers/misc/habanalabs/memory.c userptr->sgt->nents, sgt 1353 drivers/misc/habanalabs/memory.c sg_free_table(userptr->sgt); sgt 1354 drivers/misc/habanalabs/memory.c kfree(userptr->sgt); sgt 901 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct sg_table *sgt; sgt 3532 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!adapter->hma.sgt) sgt 3536 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c dma_unmap_sg(adapter->pdev_dev, adapter->hma.sgt->sgl, sgt 3537 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.sgt->nents, PCI_DMA_BIDIRECTIONAL); sgt 3541 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(adapter->hma.sgt->sgl, iter, sgt 3542 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.sgt->orig_nents, i) { sgt 3549 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sg_free_table(adapter->hma.sgt); sgt 3550 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->hma.sgt); sgt 3551 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.sgt = NULL; sgt 3557 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sg_table *sgt; sgt 3594 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.sgt = kzalloc(sizeof(*adapter->hma.sgt), GFP_KERNEL); sgt 3595 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (unlikely(!adapter->hma.sgt)) { sgt 3599 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgt = adapter->hma.sgt; sgt 3602 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgt->orig_nents = (hma_size << 20) / (page_size << page_order); sgt 3603 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (sg_alloc_table(sgt, sgt->orig_nents, GFP_KERNEL)) { sgt 3605 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(adapter->hma.sgt); sgt 3606 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.sgt = NULL; sgt 3610 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgl = adapter->hma.sgt->sgl; sgt 3612 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(sgl, iter, sgt->orig_nents, i) { sgt 3624 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sgt->nents = dma_map_sg(adapter->pdev_dev, sgl, sgt->orig_nents, sgt 3626 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (!sgt->nents) { sgt 3634 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adapter->hma.phy_addr = kcalloc(sgt->nents, sizeof(dma_addr_t), sgt 3639 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for_each_sg(sgl, iter, sgt->nents, i) { sgt 3644 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ncmds = DIV_ROUND_UP(sgt->nents, HMA_MAX_ADDR_IN_CMD); sgt 3659 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c naddr = sgt->nents % HMA_MAX_ADDR_IN_CMD; sgt 1290 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c static void dpaa_release_sgt_members(struct qm_sg_entry *sgt) sgt 1299 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dpaa_bp = dpaa_bpid2pool(sgt[i].bpid); sgt 1305 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c WARN_ON(qm_sg_entry_is_ext(&sgt[i])); sgt 1307 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c bm_buffer_set64(&bmb[j], qm_sg_entry_get64(&sgt[i])); sgt 1311 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c !qm_sg_entry_is_final(&sgt[i - 1]) && sgt 1312 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt[i - 1].bpid == sgt[i].bpid); sgt 1315 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c } while (!qm_sg_entry_is_final(&sgt[i - 1])); sgt 1321 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c struct qm_sg_entry *sgt; sgt 1336 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt = vaddr + qm_fd_get_offset(fd); sgt 1341 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dpaa_release_sgt_members(sgt); sgt 1617 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c const struct qm_sg_entry *sgt; sgt 1634 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt = phys_to_virt(addr + qm_fd_get_offset(fd)); sgt 1637 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dma_unmap_single(dev, qm_sg_addr(&sgt[0]), sgt 1638 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_get_len(&sgt[0]), dma_dir); sgt 1642 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c WARN_ON(qm_sg_entry_is_ext(&sgt[i])); sgt 1644 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dma_unmap_page(dev, qm_sg_addr(&sgt[i]), sgt 1645 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_get_len(&sgt[i]), dma_dir); sgt 1737 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c const struct qm_sg_entry *sgt; sgt 1753 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt = vaddr + fd_off; sgt 1757 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c WARN_ON(qm_sg_entry_is_ext(&sgt[i])); sgt 1759 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sg_addr = qm_sg_addr(&sgt[i]); sgt 1765 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dpaa_bp = dpaa_bpid2pool(sgt[i].bpid); sgt 1786 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c skb_put(skb, qm_sg_entry_get_len(&sgt[i])); sgt 1805 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c frag_off = qm_sg_entry_get_off(&sgt[i]) + page_offset; sgt 1806 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c frag_len = qm_sg_entry_get_len(&sgt[i]); sgt 1817 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c if (qm_sg_entry_is_final(&sgt[i])) sgt 1830 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dpaa_bp = dpaa_bpid2pool(sgt[i].bpid); sgt 1838 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sg_addr = qm_sg_addr(&sgt[i]); sgt 1841 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dpaa_bp = dpaa_bpid2pool(sgt[i].bpid); sgt 1847 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c if (qm_sg_entry_is_final(&sgt[i])) sgt 1916 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c struct qm_sg_entry *sgt; sgt 1949 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt = (struct qm_sg_entry *)(sgt_buf + priv->tx_headroom); sgt 1951 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_set_len(&sgt[0], frag_len); sgt 1952 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt[0].bpid = FSL_DPAA_BPID_INV; sgt 1953 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt[0].offset = 0; sgt 1961 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_set64(&sgt[0], addr); sgt 1976 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_set_len(&sgt[i + 1], frag_len); sgt 1977 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt[i + 1].bpid = FSL_DPAA_BPID_INV; sgt 1978 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c sgt[i + 1].offset = 0; sgt 1981 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_set64(&sgt[i + 1], addr); sgt 1985 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_set_f(&sgt[nr_frags], frag_len); sgt 1990 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c buffer_start = (void *)sgt - priv->tx_headroom; sgt 2011 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c dma_unmap_page(dev, qm_sg_addr(&sgt[j]), sgt 2012 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c qm_sg_entry_get_len(&sgt[j]), dma_dir); sgt 71 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_sg_entry *sgt; sgt 85 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c sgt = vaddr + dpaa2_fd_get_offset(fd); sgt 87 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c addr = dpaa2_sg_get_addr(&sgt[i]); sgt 93 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c if (dpaa2_sg_is_final(&sgt[i])) sgt 125 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_sg_entry *sgt) sgt 138 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_sg_entry *sge = &sgt[i]; sgt 164 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c while (!dpaa2_sg_is_final(&sgt[i]) && sgt 527 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpaa2_sg_entry *sgt; sgt 566 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c sgt = (struct dpaa2_sg_entry *)(sgt_buf + priv->tx_data_offset); sgt 576 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c dpaa2_sg_set_addr(&sgt[i], sg_dma_address(crt_scl)); sgt 577 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c dpaa2_sg_set_len(&sgt[i], sg_dma_len(crt_scl)); sgt 579 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c dpaa2_sg_set_final(&sgt[i - 1], true); sgt 331 drivers/pci/controller/vmd.c static int vmd_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 335 drivers/pci/controller/vmd.c return dma_get_sgtable_attrs(to_vmd_dev(dev), sgt, cpu_addr, addr, size, sgt 540 drivers/rapidio/devices/rio_mport_cdev.c struct sg_table sgt; sgt 578 drivers/rapidio/devices/rio_mport_cdev.c req->sgt.sgl, req->sgt.nents, req->dir); sgt 579 drivers/rapidio/devices/rio_mport_cdev.c sg_free_table(&req->sgt); sgt 617 drivers/rapidio/devices/rio_mport_cdev.c struct sg_table *sgt, int nents, enum dma_transfer_direction dir, sgt 622 drivers/rapidio/devices/rio_mport_cdev.c tx_data.sg = sgt->sgl; sgt 698 drivers/rapidio/devices/rio_mport_cdev.c struct sg_table *sgt; sgt 708 drivers/rapidio/devices/rio_mport_cdev.c sgt = &req->sgt; sgt 719 drivers/rapidio/devices/rio_mport_cdev.c tx = prep_dma_xfer(chan, xfer, sgt, nents, dir, sgt 888 drivers/rapidio/devices/rio_mport_cdev.c ret = sg_alloc_table_from_pages(&req->sgt, page_list, nr_pages, sgt 924 drivers/rapidio/devices/rio_mport_cdev.c ret = sg_alloc_table(&req->sgt, 1, GFP_KERNEL); sgt 930 drivers/rapidio/devices/rio_mport_cdev.c sg_set_buf(req->sgt.sgl, sgt 936 drivers/rapidio/devices/rio_mport_cdev.c req->sgt.sgl, req->sgt.nents, dir); sgt 855 drivers/scsi/nsp32.c nsp32_sgtable *sgt = data->cur_lunt->sglun->sgt; sgt 859 drivers/scsi/nsp32.c if (sgt == NULL) { sgt 874 drivers/scsi/nsp32.c sgt[i].addr = cpu_to_le32(sg_dma_address(sg)); sgt 875 drivers/scsi/nsp32.c sgt[i].len = cpu_to_le32(sg_dma_len(sg)); sgt 877 drivers/scsi/nsp32.c if (le32_to_cpu(sgt[i].len) > 0x10000) { sgt 879 drivers/scsi/nsp32.c "can't transfer over 64KB at a time, size=0x%lx", le32_to_cpu(sgt[i].len)); sgt 885 drivers/scsi/nsp32.c le32_to_cpu(sgt[i].addr), sgt 886 drivers/scsi/nsp32.c le32_to_cpu(sgt[i].len )); sgt 890 drivers/scsi/nsp32.c l = le32_to_cpu(sgt[num-1].len); sgt 891 drivers/scsi/nsp32.c sgt[num-1].len = cpu_to_le32(l | SGTEND); sgt 1712 drivers/scsi/nsp32.c nsp32_sgtable *sgt = data->cur_lunt->sglun->sgt; sgt 1719 drivers/scsi/nsp32.c s_sacklen -= le32_to_cpu(sgt[old_entry].addr) & 3; sgt 1727 drivers/scsi/nsp32.c sentlen += (le32_to_cpu(sgt[new_entry].len) & ~SGTEND); sgt 1748 drivers/scsi/nsp32.c len = le32_to_cpu(sgt[new_entry].len); sgt 1749 drivers/scsi/nsp32.c addr = le32_to_cpu(sgt[new_entry].addr); sgt 1751 drivers/scsi/nsp32.c sgt[new_entry].addr = cpu_to_le32(addr); sgt 1752 drivers/scsi/nsp32.c sgt[new_entry].len = cpu_to_le32(restlen); sgt 446 drivers/scsi/nsp32.h nsp32_sgtable sgt[NSP32_SG_SIZE+1]; /* SG table */ sgt 24 drivers/spi/internals.h struct sg_table *sgt, void *buf, size_t len, sgt 27 drivers/spi/internals.h struct sg_table *sgt, enum dma_data_direction dir); sgt 30 drivers/spi/internals.h struct sg_table *sgt, void *buf, size_t len, sgt 37 drivers/spi/internals.h struct device *dev, struct sg_table *sgt, sgt 277 drivers/spi/spi-ep93xx.c struct sg_table *sgt; sgt 294 drivers/spi/spi-ep93xx.c sgt = &espi->rx_sgt; sgt 301 drivers/spi/spi-ep93xx.c sgt = &espi->tx_sgt; sgt 322 drivers/spi/spi-ep93xx.c if (nents != sgt->nents) { sgt 323 drivers/spi/spi-ep93xx.c sg_free_table(sgt); sgt 325 drivers/spi/spi-ep93xx.c ret = sg_alloc_table(sgt, nents, GFP_KERNEL); sgt 331 drivers/spi/spi-ep93xx.c for_each_sg(sgt->sgl, sg, sgt->nents, i) { sgt 351 drivers/spi/spi-ep93xx.c nents = dma_map_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgt 355 drivers/spi/spi-ep93xx.c txd = dmaengine_prep_slave_sg(chan, sgt->sgl, nents, conf.direction, sgt 358 drivers/spi/spi-ep93xx.c dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgt 377 drivers/spi/spi-ep93xx.c struct sg_table *sgt; sgt 381 drivers/spi/spi-ep93xx.c sgt = &espi->rx_sgt; sgt 384 drivers/spi/spi-ep93xx.c sgt = &espi->tx_sgt; sgt 387 drivers/spi/spi-ep93xx.c dma_unmap_sg(chan->device->dev, sgt->sgl, sgt->nents, dir); sgt 36 drivers/spi/spi-mem.c struct sg_table *sgt) sgt 53 drivers/spi/spi-mem.c return spi_map_buf(ctlr, dmadev, sgt, op->data.buf.in, op->data.nbytes, sgt 82 drivers/spi/spi-mem.c struct sg_table *sgt) sgt 96 drivers/spi/spi-mem.c spi_unmap_buf(ctlr, dmadev, sgt, sgt 78 drivers/spi/spi-pxa2xx-dma.c struct sg_table *sgt; sgt 101 drivers/spi/spi-pxa2xx-dma.c sgt = &xfer->tx_sg; sgt 108 drivers/spi/spi-pxa2xx-dma.c sgt = &xfer->rx_sg; sgt 118 drivers/spi/spi-pxa2xx-dma.c return dmaengine_prep_slave_sg(chan, sgt->sgl, sgt->nents, dir, sgt 268 drivers/spi/spi-s3c64xx.c struct sg_table *sgt) sgt 294 drivers/spi/spi-s3c64xx.c desc = dmaengine_prep_slave_sg(dma->ch, sgt->sgl, sgt->nents, sgt 205 drivers/spi/spi-stm32-qspi.c struct sg_table sgt; sgt 222 drivers/spi/spi-stm32-qspi.c err = spi_controller_dma_map_mem_op_data(qspi->ctrl, op, &sgt); sgt 226 drivers/spi/spi-stm32-qspi.c desc = dmaengine_prep_slave_sg(dma_ch, sgt.sgl, sgt.nents, sgt 247 drivers/spi/spi-stm32-qspi.c t_out = sgt.nents * STM32_COMP_TIMEOUT_MS; sgt 258 drivers/spi/spi-stm32-qspi.c spi_controller_dma_unmap_mem_op_data(qspi->ctrl, op, &sgt); sgt 555 drivers/spi/spi-ti-qspi.c struct sg_table sgt; sgt 559 drivers/spi/spi-ti-qspi.c &sgt)) { sgt 560 drivers/spi/spi-ti-qspi.c ret = ti_qspi_dma_xfer_sg(qspi, sgt, from); sgt 562 drivers/spi/spi-ti-qspi.c op, &sgt); sgt 807 drivers/spi/spi.c struct sg_table *sgt, void *buf, size_t len, sgt 837 drivers/spi/spi.c ret = sg_alloc_table(sgt, sgs, GFP_KERNEL); sgt 841 drivers/spi/spi.c sg = &sgt->sgl[0]; sgt 858 drivers/spi/spi.c sg_free_table(sgt); sgt 874 drivers/spi/spi.c ret = dma_map_sg(dev, sgt->sgl, sgt->nents, dir); sgt 878 drivers/spi/spi.c sg_free_table(sgt); sgt 882 drivers/spi/spi.c sgt->nents = ret; sgt 888 drivers/spi/spi.c struct sg_table *sgt, enum dma_data_direction dir) sgt 890 drivers/spi/spi.c if (sgt->orig_nents) { sgt 891 drivers/spi/spi.c dma_unmap_sg(dev, sgt->sgl, sgt->orig_nents, dir); sgt 892 drivers/spi/spi.c sg_free_table(sgt); sgt 88 drivers/staging/kpc2000/kpc_dma/fileops.c rv = sg_alloc_table_from_pages(&acd->sgt, acd->user_pages, acd->page_count, iov_base & (PAGE_SIZE-1), iov_len, GFP_KERNEL); sgt 95 drivers/staging/kpc2000/kpc_dma/fileops.c acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir); sgt 102 drivers/staging/kpc2000/kpc_dma/fileops.c for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) { sgt 125 drivers/staging/kpc2000/kpc_dma/fileops.c for_each_sg(acd->sgt.sgl, sg, acd->mapped_entry_count, i) { sgt 189 drivers/staging/kpc2000/kpc_dma/fileops.c dma_unmap_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir); sgt 190 drivers/staging/kpc2000/kpc_dma/fileops.c sg_free_table(&acd->sgt); sgt 210 drivers/staging/kpc2000/kpc_dma/fileops.c BUG_ON(acd->sgt.sgl == NULL); sgt 220 drivers/staging/kpc2000/kpc_dma/fileops.c dma_unmap_sg(&acd->ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, acd->ldev->dir); sgt 226 drivers/staging/kpc2000/kpc_dma/fileops.c sg_free_table(&acd->sgt); sgt 93 drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.h struct sg_table sgt; sgt 24 drivers/staging/media/tegra-vde/dmabuf-cache.c struct sg_table *sgt; sgt 38 drivers/staging/media/tegra-vde/dmabuf-cache.c dma_buf_unmap_attachment(entry->a, entry->sgt, entry->dma_dir); sgt 69 drivers/staging/media/tegra-vde/dmabuf-cache.c struct sg_table *sgt; sgt 90 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = sg_dma_address(entry->sgt->sgl); sgt 102 drivers/staging/media/tegra-vde/dmabuf-cache.c sgt = dma_buf_map_attachment(attachment, dma_dir); sgt 103 drivers/staging/media/tegra-vde/dmabuf-cache.c if (IS_ERR(sgt)) { sgt 105 drivers/staging/media/tegra-vde/dmabuf-cache.c err = PTR_ERR(sgt); sgt 109 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!vde->domain && sgt->nents > 1) { sgt 122 drivers/staging/media/tegra-vde/dmabuf-cache.c err = tegra_vde_iommu_map(vde, sgt, &iova, dmabuf->size); sgt 128 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = sg_dma_address(sgt->sgl); sgt 138 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->sgt = sgt; sgt 152 drivers/staging/media/tegra-vde/dmabuf-cache.c dma_buf_unmap_attachment(attachment, sgt, dma_dir); sgt 20 drivers/staging/media/tegra-vde/iommu.c struct sg_table *sgt, sgt 39 drivers/staging/media/tegra-vde/iommu.c size = iommu_map_sg(vde->domain, addr, sgt->sgl, sgt->nents, sgt 58 drivers/staging/media/tegra-vde/vde.h struct sg_table *sgt, sgt 226 drivers/usb/gadget/function/f_fs.c struct sg_table sgt; sgt 764 drivers/usb/gadget/function/f_fs.c static void *ffs_build_sg_list(struct sg_table *sgt, size_t sz) sgt 785 drivers/usb/gadget/function/f_fs.c if (sg_alloc_table_from_pages(sgt, pages, n_pages, 0, sz, GFP_KERNEL)) { sgt 800 drivers/usb/gadget/function/f_fs.c return ffs_build_sg_list(&io_data->sgt, data_len); sgt 811 drivers/usb/gadget/function/f_fs.c sg_free_table(&io_data->sgt); sgt 1061 drivers/usb/gadget/function/f_fs.c req->sg = io_data->sgt.sgl; sgt 1062 drivers/usb/gadget/function/f_fs.c req->num_sgs = io_data->sgt.nents; sgt 1105 drivers/usb/gadget/function/f_fs.c req->sg = io_data->sgt.sgl; sgt 1106 drivers/usb/gadget/function/f_fs.c req->num_sgs = io_data->sgt.nents; sgt 51 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt; sgt 70 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt; sgt 203 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt; sgt 206 drivers/xen/gntdev-dmabuf.c sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); sgt 207 drivers/xen/gntdev-dmabuf.c if (!sgt) { sgt 212 drivers/xen/gntdev-dmabuf.c ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0, sgt 218 drivers/xen/gntdev-dmabuf.c return sgt; sgt 221 drivers/xen/gntdev-dmabuf.c kfree(sgt); sgt 246 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt = gntdev_dmabuf_attach->sgt; sgt 248 drivers/xen/gntdev-dmabuf.c if (sgt) { sgt 250 drivers/xen/gntdev-dmabuf.c dma_unmap_sg_attrs(attach->dev, sgt->sgl, sgt 251 drivers/xen/gntdev-dmabuf.c sgt->nents, sgt 254 drivers/xen/gntdev-dmabuf.c sg_free_table(sgt); sgt 257 drivers/xen/gntdev-dmabuf.c kfree(sgt); sgt 269 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt; sgt 279 drivers/xen/gntdev-dmabuf.c return gntdev_dmabuf_attach->sgt; sgt 288 drivers/xen/gntdev-dmabuf.c sgt = dmabuf_pages_to_sgt(gntdev_dmabuf->pages, sgt 290 drivers/xen/gntdev-dmabuf.c if (!IS_ERR(sgt)) { sgt 291 drivers/xen/gntdev-dmabuf.c if (!dma_map_sg_attrs(attach->dev, sgt->sgl, sgt->nents, dir, sgt 293 drivers/xen/gntdev-dmabuf.c sg_free_table(sgt); sgt 294 drivers/xen/gntdev-dmabuf.c kfree(sgt); sgt 295 drivers/xen/gntdev-dmabuf.c sgt = ERR_PTR(-ENOMEM); sgt 297 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf_attach->sgt = sgt; sgt 301 drivers/xen/gntdev-dmabuf.c if (IS_ERR(sgt)) sgt 303 drivers/xen/gntdev-dmabuf.c return sgt; sgt 307 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt, sgt 613 drivers/xen/gntdev-dmabuf.c struct sg_table *sgt; sgt 638 drivers/xen/gntdev-dmabuf.c sgt = dma_buf_map_attachment(attach, DMA_BIDIRECTIONAL); sgt 639 drivers/xen/gntdev-dmabuf.c if (IS_ERR(sgt)) { sgt 640 drivers/xen/gntdev-dmabuf.c ret = ERR_CAST(sgt); sgt 652 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->u.imp.sgt = sgt; sgt 656 drivers/xen/gntdev-dmabuf.c for_each_sg_page(sgt->sgl, &sg_iter, sgt->nents, 0) { sgt 688 drivers/xen/gntdev-dmabuf.c dma_buf_unmap_attachment(attach, sgt, DMA_BIDIRECTIONAL); sgt 737 drivers/xen/gntdev-dmabuf.c if (gntdev_dmabuf->u.imp.sgt) sgt 738 drivers/xen/gntdev-dmabuf.c dma_buf_unmap_attachment(attach, gntdev_dmabuf->u.imp.sgt, sgt 627 include/drm/drm_drv.h struct sg_table *sgt); sgt 23 include/drm/drm_gem_cma_helper.h struct sg_table *sgt; sgt 103 include/drm/drm_gem_cma_helper.h struct sg_table *sgt); sgt 131 include/drm/drm_gem_cma_helper.h struct sg_table *sgt); sgt 80 include/drm/drm_gem_shmem_helper.h struct sg_table *sgt; sgt 145 include/drm/drm_gem_shmem_helper.h !shmem->vmap_use_count && shmem->sgt && sgt 170 include/drm/drm_gem_shmem_helper.h struct sg_table *sgt); sgt 85 include/drm/drm_prime.h struct sg_table *sgt, sgt 106 include/drm/drm_prime.h int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages, sgt 344 include/linux/dma-buf.h struct sg_table *sgt; sgt 90 include/linux/dma-mapping.h int (*get_sgtable)(struct device *dev, struct sg_table *sgt, void *, sgt 455 include/linux/dma-mapping.h int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, sgt 542 include/linux/dma-mapping.h struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, sgt 641 include/linux/dma-mapping.h dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr, sgt 98 include/linux/fpga/fpga-mgr.h struct sg_table *sgt; sgt 132 include/linux/fpga/fpga-mgr.h int (*write_sg)(struct fpga_manager *mgr, struct sg_table *sgt); sgt 67 include/linux/host1x.h dma_addr_t (*pin)(struct host1x_bo *bo, struct sg_table **sgt); sgt 68 include/linux/host1x.h void (*unpin)(struct host1x_bo *bo, struct sg_table *sgt); sgt 96 include/linux/host1x.h struct sg_table **sgt) sgt 98 include/linux/host1x.h return bo->ops->pin(bo, sgt); sgt 101 include/linux/host1x.h static inline void host1x_bo_unpin(struct host1x_bo *bo, struct sg_table *sgt) sgt 103 include/linux/host1x.h bo->ops->unpin(bo, sgt); sgt 188 include/linux/hsi/hsi.h struct sg_table sgt; sgt 43 include/linux/intel_th.h int (*alloc_window)(void *priv, struct sg_table **sgt, sgt 45 include/linux/intel_th.h void (*free_window)(void *priv, struct sg_table *sgt); sgt 59 include/linux/intel_th.h int (*ready)(void *priv, struct sg_table *sgt, size_t bytes); sgt 65 include/linux/intel_th.h void intel_th_msc_window_unlock(struct device *dev, struct sg_table *sgt); sgt 275 include/linux/scatterlist.h int __sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages, sgt 279 include/linux/scatterlist.h int sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages, sgt 115 include/media/vsp1.h int vsp1_du_map_sg(struct device *dev, struct sg_table *sgt); sgt 116 include/media/vsp1.h void vsp1_du_unmap_sg(struct device *dev, struct sg_table *sgt); sgt 263 include/net/erspan.h u8 sgt = 0; sgt 294 include/net/erspan.h md2->sgt = htons(sgt); sgt 22 include/uapi/linux/erspan.h __be16 sgt; /* security group tag */ sgt 111 kernel/dma/mapping.c int dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, sgt 133 kernel/dma/mapping.c ret = sg_alloc_table(sgt, 1, GFP_KERNEL); sgt 135 kernel/dma/mapping.c sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); sgt 150 kernel/dma/mapping.c int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, sgt 157 kernel/dma/mapping.c return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, sgt 161 kernel/dma/mapping.c return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); sgt 389 lib/scatterlist.c int __sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages, sgt 413 lib/scatterlist.c ret = sg_alloc_table(sgt, chunks, gfp_mask); sgt 419 lib/scatterlist.c for_each_sg(sgt->sgl, s, sgt->orig_nents, i) { sgt 464 lib/scatterlist.c int sg_alloc_table_from_pages(struct sg_table *sgt, struct page **pages, sgt 468 lib/scatterlist.c return __sg_alloc_table_from_pages(sgt, pages, n_pages, offset, size, sgt 160 net/ceph/crypto.c static int setup_sgtable(struct sg_table *sgt, struct scatterlist *prealloc_sg, sgt 172 net/ceph/crypto.c memset(sgt, 0, sizeof(*sgt)); sgt 182 net/ceph/crypto.c ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS); sgt 188 net/ceph/crypto.c sgt->sgl = prealloc_sg; sgt 189 net/ceph/crypto.c sgt->nents = sgt->orig_nents = 1; sgt 192 net/ceph/crypto.c for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) { sgt 212 net/ceph/crypto.c static void teardown_sgtable(struct sg_table *sgt) sgt 214 net/ceph/crypto.c if (sgt->orig_nents > 1) sgt 215 net/ceph/crypto.c sg_free_table(sgt); sgt 222 net/ceph/crypto.c struct sg_table sgt; sgt 232 net/ceph/crypto.c ret = setup_sgtable(&sgt, &prealloc_sg, buf, crypt_len); sgt 239 net/ceph/crypto.c skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv); sgt 278 net/ceph/crypto.c teardown_sgtable(&sgt); sgt 516 net/smc/smc_clc.c (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgt 588 net/smc/smc_clc.c (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgt 379 net/smc/smc_core.c sg_free_table(&buf_desc->sgt[SMC_SINGLE_LINK]); sgt 764 net/smc/smc_core.c rc = sg_alloc_table(&buf_desc->sgt[SMC_SINGLE_LINK], 1, sgt 770 net/smc/smc_core.c sg_set_buf(buf_desc->sgt[SMC_SINGLE_LINK].sgl, sgt 157 net/smc/smc_core.h struct sg_table sgt[SMC_LINKS_PER_LGR_MAX]; sgt 369 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgt 370 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].orig_nents, sgt 408 net/smc/smc_ib.c for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg, sgt 409 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].nents, i) { sgt 428 net/smc/smc_ib.c for_each_sg(buf_slot->sgt[SMC_SINGLE_LINK].sgl, sg, sgt 429 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].nents, i) { sgt 447 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgt 448 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].orig_nents, sgt 460 net/smc/smc_ib.c if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address) sgt 464 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl, sgt 465 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].orig_nents, sgt 467 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0; sgt 236 net/smc/smc_llc.c (u64)sg_dma_address(rmb_desc->sgt[SMC_SINGLE_LINK].sgl)); sgt 315 net/smc/smc_tx.c sg_dma_address(conn->sndbuf_desc->sgt[SMC_SINGLE_LINK].sgl); sgt 144 sound/soc/sprd/sprd-pcm-compress.c struct scatterlist *sg, *sgt; sgt 163 sound/soc/sprd/sprd-pcm-compress.c sgt = sg = devm_kcalloc(dev, sg_num, sizeof(*sg), GFP_KERNEL); sgt 214 sound/soc/sprd/sprd-pcm-compress.c sg_init_table(sgt, sg_num); sgt 215 sound/soc/sprd/sprd-pcm-compress.c for (j = 0; j < sg_num; j++, sgt++) { sgt 216 sound/soc/sprd/sprd-pcm-compress.c sg_dma_len(sgt) = dma->trans_len; sgt 217 sound/soc/sprd/sprd-pcm-compress.c sg_dma_address(sgt) = dst_addr; sgt 244 sound/soc/sprd/sprd-pcm-dma.c struct scatterlist *sgt = sg; sgt 261 sound/soc/sprd/sprd-pcm-dma.c sg_init_table(sgt, sg_num); sgt 262 sound/soc/sprd/sprd-pcm-dma.c for (j = 0; j < sg_num; j++, sgt++) { sgt 265 sound/soc/sprd/sprd-pcm-dma.c sg_dma_len(sgt) = sg_len; sgt 266 sound/soc/sprd/sprd-pcm-dma.c sg_dma_address(sgt) = runtime->dma_addr + sgt 22 tools/include/uapi/linux/erspan.h __be16 sgt; /* security group tag */