dma_address       541 arch/alpha/kernel/pci_iommu.c 			sg->dma_address = -1;
dma_address       544 arch/alpha/kernel/pci_iommu.c 			sg->dma_address = -2;
dma_address       549 arch/alpha/kernel/pci_iommu.c 			leader->dma_address = leader_flag;
dma_address       559 arch/alpha/kernel/pci_iommu.c 	leader->dma_address = leader_flag;
dma_address       580 arch/alpha/kernel/pci_iommu.c 	if (leader->dma_address == 0
dma_address       583 arch/alpha/kernel/pci_iommu.c 		out->dma_address = paddr + __direct_map_base;
dma_address       587 arch/alpha/kernel/pci_iommu.c 		     __va(paddr), size, out->dma_address);
dma_address       594 arch/alpha/kernel/pci_iommu.c 	if (leader->dma_address == 0 && dac_allowed) {
dma_address       595 arch/alpha/kernel/pci_iommu.c 		out->dma_address = paddr + alpha_mv.pci_dac_offset;
dma_address       599 arch/alpha/kernel/pci_iommu.c 		     __va(paddr), size, out->dma_address);
dma_address       612 arch/alpha/kernel/pci_iommu.c 		if (leader->dma_address == 0)
dma_address       621 arch/alpha/kernel/pci_iommu.c 	out->dma_address = arena->dma_base + dma_ofs*PAGE_SIZE + paddr;
dma_address       625 arch/alpha/kernel/pci_iommu.c 	     __va(paddr), size, out->dma_address, npages);
dma_address       639 arch/alpha/kernel/pci_iommu.c 		while (sg+1 < end && (int) sg[1].dma_address == -1) {
dma_address       660 arch/alpha/kernel/pci_iommu.c 	} while (++sg < end && (int) sg->dma_address < 0);
dma_address       683 arch/alpha/kernel/pci_iommu.c 		sg->dma_address
dma_address       686 arch/alpha/kernel/pci_iommu.c 		return sg->dma_address != DMA_MAPPING_ERROR;
dma_address       711 arch/alpha/kernel/pci_iommu.c 		if ((int) sg->dma_address < 0)
dma_address       776 arch/alpha/kernel/pci_iommu.c 		addr = sg->dma_address;
dma_address        95 arch/arm/kernel/dma-isa.c 			dma->buf.dma_address = dma_map_single(&isa_dma_dev,
dma_address       100 arch/arm/kernel/dma-isa.c 		address = dma->buf.dma_address;
dma_address        78 arch/arm/mach-rpc/dma.c 				idma->dma_addr = idma->dma.sg->dma_address;
dma_address       172 arch/arm/mach-rpc/dma.c 			idma->dma.buf.dma_address = dma_map_single(&isa_dma_dev,
dma_address       178 arch/arm/mach-rpc/dma.c 		idma->dma_addr = idma->dma.sg->dma_address;
dma_address      1017 arch/arm/mm/dma-mapping.c 		s->dma_address = ops->map_page(dev, sg_page(s), s->offset,
dma_address      1019 arch/arm/mm/dma-mapping.c 		if (dma_mapping_error(dev, s->dma_address))
dma_address      1671 arch/arm/mm/dma-mapping.c 		s->dma_address = DMA_MAPPING_ERROR;
dma_address      1675 arch/arm/mm/dma-mapping.c 			if (__map_sg_chunk(dev, start, size, &dma->dma_address,
dma_address      1679 arch/arm/mm/dma-mapping.c 			dma->dma_address += offset;
dma_address      1689 arch/arm/mm/dma-mapping.c 	if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs,
dma_address      1693 arch/arm/mm/dma-mapping.c 	dma->dma_address += offset;
dma_address       391 arch/ia64/hp/common/sba_iommu.c 		       startsg->dma_address, startsg->dma_length,
dma_address      1215 arch/ia64/hp/common/sba_iommu.c 				nents, startsg->dma_address, cnt,
dma_address      1219 arch/ia64/hp/common/sba_iommu.c 				nents, startsg->dma_address, cnt,
dma_address      1225 arch/ia64/hp/common/sba_iommu.c 		if (startsg->dma_address & PIDE_FLAG) {
dma_address      1226 arch/ia64/hp/common/sba_iommu.c 			u32 pide = startsg->dma_address & ~PIDE_FLAG;
dma_address      1228 arch/ia64/hp/common/sba_iommu.c 			startsg->dma_address = 0;
dma_address      1231 arch/ia64/hp/common/sba_iommu.c 			dma_sg->dma_address = pide | ioc->ibase;
dma_address      1319 arch/ia64/hp/common/sba_iommu.c 		startsg->dma_address = startsg->dma_length = 0;
dma_address      1331 arch/ia64/hp/common/sba_iommu.c 			startsg->dma_address = startsg->dma_length = 0;
dma_address      1409 arch/ia64/hp/common/sba_iommu.c 		dma_sg->dma_address = (dma_addr_t)(PIDE_FLAG | (idx << iovp_shift)
dma_address      1452 arch/ia64/hp/common/sba_iommu.c 			sg->dma_address = virt_to_phys(sba_sg_address(sg));
dma_address      1460 arch/ia64/hp/common/sba_iommu.c 		sglist->dma_address = sba_map_page(dev, sg_page(sglist),
dma_address      1462 arch/ia64/hp/common/sba_iommu.c 		if (dma_mapping_error(dev, sglist->dma_address))
dma_address      1552 arch/ia64/hp/common/sba_iommu.c 		sba_unmap_page(dev, sglist->dma_address, sglist->dma_length,
dma_address       617 arch/mips/jazz/jazzdma.c 		sg->dma_address = vdma_alloc(sg_phys(sg), sg->length);
dma_address       618 arch/mips/jazz/jazzdma.c 		if (sg->dma_address == DMA_MAPPING_ERROR)
dma_address       636 arch/mips/jazz/jazzdma.c 		vdma_free(sg->dma_address);
dma_address       514 arch/powerpc/kernel/iommu.c 			outs->dma_address = dma_addr;
dma_address       535 arch/powerpc/kernel/iommu.c 		outs->dma_address = DMA_MAPPING_ERROR;
dma_address       549 arch/powerpc/kernel/iommu.c 			vaddr = s->dma_address & IOMMU_PAGE_MASK(tbl);
dma_address       550 arch/powerpc/kernel/iommu.c 			npages = iommu_num_pages(s->dma_address, s->dma_length,
dma_address       553 arch/powerpc/kernel/iommu.c 			s->dma_address = DMA_MAPPING_ERROR;
dma_address       577 arch/powerpc/kernel/iommu.c 		dma_addr_t dma_handle = sg->dma_address;
dma_address       645 arch/powerpc/platforms/ps3/system-bus.c 					sg->length, &sg->dma_address, 0);
dma_address       114 arch/powerpc/platforms/pseries/ibmebus.c 		sg->dma_address = (dma_addr_t) sg_virt(sg);
dma_address       497 arch/s390/pci/pci_dma.c 		s->dma_address = DMA_MAPPING_ERROR;
dma_address       503 arch/s390/pci/pci_dma.c 					      &dma->dma_address, dir))
dma_address       506 arch/s390/pci/pci_dma.c 			dma->dma_address += offset;
dma_address       516 arch/s390/pci/pci_dma.c 	if (__s390_dma_map_sg(dev, start, size, &dma->dma_address, dir))
dma_address       519 arch/s390/pci/pci_dma.c 	dma->dma_address += offset;
dma_address       540 arch/s390/pci/pci_dma.c 			s390_dma_unmap_pages(dev, s->dma_address, s->dma_length,
dma_address       542 arch/s390/pci/pci_dma.c 		s->dma_address = 0;
dma_address       537 arch/sparc/kernel/iommu.c 			outs->dma_address = dma_addr;
dma_address       550 arch/sparc/kernel/iommu.c 		outs->dma_address = DMA_MAPPING_ERROR;
dma_address       562 arch/sparc/kernel/iommu.c 			vaddr = s->dma_address & IO_PAGE_MASK;
dma_address       563 arch/sparc/kernel/iommu.c 			npages = iommu_num_pages(s->dma_address, s->dma_length,
dma_address       576 arch/sparc/kernel/iommu.c 			s->dma_address = DMA_MAPPING_ERROR;
dma_address       599 arch/sparc/kernel/iommu.c 		bus_addr = sg->dma_address & IO_PAGE_MASK;
dma_address       628 arch/sparc/kernel/iommu.c 		dma_addr_t dma_handle = sg->dma_address;
dma_address       723 arch/sparc/kernel/iommu.c 		iopte = iommu->page_table + ((sglist[0].dma_address -
dma_address       729 arch/sparc/kernel/iommu.c 	bus_addr = sglist[0].dma_address & IO_PAGE_MASK;
dma_address       737 arch/sparc/kernel/iommu.c 	npages = (IO_PAGE_ALIGN(sgprv->dma_address + sgprv->dma_length)
dma_address       579 arch/sparc/kernel/pci_sun4v.c 			outs->dma_address = dma_addr;
dma_address       597 arch/sparc/kernel/pci_sun4v.c 		outs->dma_address = DMA_MAPPING_ERROR;
dma_address       608 arch/sparc/kernel/pci_sun4v.c 			vaddr = s->dma_address & IO_PAGE_MASK;
dma_address       609 arch/sparc/kernel/pci_sun4v.c 			npages = iommu_num_pages(s->dma_address, s->dma_length,
dma_address       614 arch/sparc/kernel/pci_sun4v.c 			s->dma_address = DMA_MAPPING_ERROR;
dma_address       648 arch/sparc/kernel/pci_sun4v.c 		dma_addr_t dma_handle = sg->dma_address;
dma_address       172 arch/sparc/mm/io-unit.c 		sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length);
dma_address       204 arch/sparc/mm/io-unit.c 		len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT;
dma_address       205 arch/sparc/mm/io-unit.c 		vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
dma_address       251 arch/sparc/mm/iommu.c 		sg->dma_address =__sbus_iommu_map_page(dev, sg_page(sg),
dma_address       253 arch/sparc/mm/iommu.c 		if (sg->dma_address == DMA_MAPPING_ERROR)
dma_address       300 arch/sparc/mm/iommu.c 		sbus_iommu_unmap_page(dev, sg->dma_address, sg->length, dir,
dma_address       302 arch/sparc/mm/iommu.c 		sg->dma_address = 0x21212121;
dma_address       288 arch/x86/kernel/amd_gart_64.c 		gart_unmap_page(dev, s->dma_address, s->dma_length, dir, 0);
dma_address       316 arch/x86/kernel/amd_gart_64.c 		s->dma_address = addr;
dma_address       339 arch/x86/kernel/amd_gart_64.c 		unsigned long phys_addr = s->dma_address;
dma_address       343 arch/x86/kernel/amd_gart_64.c 			sout->dma_address = iommu_bus_base;
dma_address       344 arch/x86/kernel/amd_gart_64.c 			sout->dma_address += iommu_page*PAGE_SIZE + s->offset;
dma_address       369 arch/x86/kernel/amd_gart_64.c 		sout->dma_address = start->dma_address;
dma_address       403 arch/x86/kernel/amd_gart_64.c 		s->dma_address = addr;
dma_address       461 arch/x86/kernel/amd_gart_64.c 		s->dma_address = DMA_MAPPING_ERROR;
dma_address       338 arch/x86/kernel/pci-calgary_64.c 		dma_addr_t dma = s->dma_address;
dma_address       373 arch/x86/kernel/pci-calgary_64.c 		s->dma_address = (entry << PAGE_SHIFT) | s->offset;
dma_address       385 arch/x86/kernel/pci-calgary_64.c 		sg->dma_address = DMA_MAPPING_ERROR;
dma_address       236 drivers/atm/idt77252.h 	u32		dma_address;
dma_address       309 drivers/atm/nicstar.h 	u32 dma_address;
dma_address       346 drivers/atm/nicstar.h 	u32 dma_address;
dma_address       889 drivers/block/skd_main.c 	uint64_t dma_address;
dma_address       898 drivers/block/skd_main.c 	dma_address = skspcl->req.sksg_dma_address;
dma_address       899 drivers/block/skd_main.c 	scsi->hdr.sg_list_dma_address = cpu_to_be64(dma_address);
dma_address      2713 drivers/block/skd_main.c 		uint64_t dma_address = *ret_dma_addr;
dma_address      2720 drivers/block/skd_main.c 			sg_list[i].next_desc_ptr = dma_address + ndp_off;
dma_address       121 drivers/crypto/hisilicon/sgl.c 	hw_sge->buf = sgl->dma_address;
dma_address       504 drivers/crypto/sahara.c 		dev->hw_link[i]->p = sg->dma_address;
dma_address       518 drivers/crypto/sahara.c 		dev->hw_link[j]->p = sg->dma_address;
dma_address       822 drivers/crypto/sahara.c 		dev->hw_link[i]->p = sg->dma_address;
dma_address       293 drivers/dma/imx-dma.c 		imx_dmav1_writel(imxdma, sg->dma_address,
dma_address       296 drivers/dma/imx-dma.c 		imx_dmav1_writel(imxdma, sg->dma_address,
dma_address       837 drivers/dma/imx-dma.c 		if (sg_dma_len(sgl) & 3 || sgl->dma_address & 3)
dma_address       841 drivers/dma/imx-dma.c 		if (sg_dma_len(sgl) & 1 || sgl->dma_address & 1)
dma_address       898 drivers/dma/imx-dma.c 		imxdmac->sg_list[i].dma_address = dma_addr;
dma_address      1465 drivers/dma/imx-sdma.c 		bd->buffer_addr = sg->dma_address;
dma_address      1484 drivers/dma/imx-sdma.c 			if (count & 3 || sg->dma_address & 3)
dma_address      1489 drivers/dma/imx-sdma.c 			if (count & 1 || sg->dma_address & 1)
dma_address      1508 drivers/dma/imx-sdma.c 				i, count, (u64)sg->dma_address,
dma_address       564 drivers/dma/mxs-dma.c 			ccw->bufaddr = sg->dma_address;
dma_address       781 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	sg->sgl->dma_address = addr;
dma_address        50 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 		*addr = ttm->dma_address[0];
dma_address       128 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	if (ttm->dma_address[0] + PAGE_SIZE >= adev->gmc.agp_size)
dma_address       131 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	return adev->gmc.agp_start + ttm->dma_address[0];
dma_address       964 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 					 gtt->ttm.dma_address, ttm->num_pages);
dma_address      1015 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 				ttm->pages, gtt->ttm.dma_address, flags);
dma_address      1027 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 				&(gtt->ttm.dma_address[page_idx]), flags);
dma_address      1030 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 				     ttm->pages, gtt->ttm.dma_address, flags);
dma_address      1083 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		ttm->pages, gtt->ttm.dma_address, flags);
dma_address      1270 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 						 gtt->ttm.dma_address,
dma_address      1908 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	dma_addr_t *dma_address;
dma_address      1942 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	dma_address = &gtt->ttm.dma_address[offset >> PAGE_SHIFT];
dma_address      1944 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	r = amdgpu_gart_map(adev, 0, num_pages, dma_address, flags,
dma_address      1707 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			pages_addr = ttm->dma_address;
dma_address       521 drivers/gpu/drm/exynos/exynos_drm_g2d.c 	g2d_userptr->dma_addr = sgt->sgl[0].dma_address;
dma_address        59 drivers/gpu/drm/i915/i915_gpu_error.c 	sg->dma_address = it;
dma_address       844 drivers/gpu/drm/i915/i915_gpu_error.c 	if (!sg || off < sg->dma_address)
dma_address       849 drivers/gpu/drm/i915/i915_gpu_error.c 	pos = sg->dma_address;
dma_address        50 drivers/gpu/drm/msm/msm_gpummu.c 		dma_addr_t addr = sg->dma_address;
dma_address       554 drivers/gpu/drm/nouveau/nouveau_bo.c 					   ttm_dma->dma_address[i],
dma_address       573 drivers/gpu/drm/nouveau/nouveau_bo.c 		dma_sync_single_for_cpu(drm->dev->dev, ttm_dma->dma_address[i],
dma_address      1599 drivers/gpu/drm/nouveau/nouveau_bo.c 						 ttm_dma->dma_address, ttm->num_pages);
dma_address      1632 drivers/gpu/drm/nouveau/nouveau_bo.c 				dma_unmap_page(dev, ttm_dma->dma_address[i],
dma_address      1634 drivers/gpu/drm/nouveau/nouveau_bo.c 				ttm_dma->dma_address[i] = 0;
dma_address      1640 drivers/gpu/drm/nouveau/nouveau_bo.c 		ttm_dma->dma_address[i] = addr;
dma_address      1675 drivers/gpu/drm/nouveau/nouveau_bo.c 		if (ttm_dma->dma_address[i]) {
dma_address      1676 drivers/gpu/drm/nouveau/nouveau_bo.c 			dma_unmap_page(dev, ttm_dma->dma_address[i], PAGE_SIZE,
dma_address       120 drivers/gpu/drm/nouveau/nouveau_mem.c 	else            args.dma = tt->dma_address;
dma_address       535 drivers/gpu/drm/radeon/radeon_ttm.c 					 gtt->ttm.dma_address, ttm->num_pages);
dma_address       597 drivers/gpu/drm/radeon/radeon_ttm.c 			     ttm->pages, gtt->ttm.dma_address, flags);
dma_address       685 drivers/gpu/drm/radeon/radeon_ttm.c 						 gtt->ttm.dma_address, ttm->num_pages);
dma_address      1118 drivers/gpu/drm/ttm/ttm_page_alloc.c 		tt->dma_address[i] = dma_map_page(dev, tt->ttm.pages[i],
dma_address      1121 drivers/gpu/drm/ttm/ttm_page_alloc.c 		if (dma_mapping_error(dev, tt->dma_address[i])) {
dma_address      1123 drivers/gpu/drm/ttm/ttm_page_alloc.c 				dma_unmap_page(dev, tt->dma_address[i],
dma_address      1125 drivers/gpu/drm/ttm/ttm_page_alloc.c 				tt->dma_address[i] = 0;
dma_address      1132 drivers/gpu/drm/ttm/ttm_page_alloc.c 			tt->dma_address[i + 1] = tt->dma_address[i] + PAGE_SIZE;
dma_address      1148 drivers/gpu/drm/ttm/ttm_page_alloc.c 		if (!tt->dma_address[i] || !tt->ttm.pages[i]) {
dma_address      1160 drivers/gpu/drm/ttm/ttm_page_alloc.c 		dma_unmap_page(dev, tt->dma_address[i], num_pages * PAGE_SIZE,
dma_address       848 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		ttm_dma->dma_address[index] = d_page->dma;
dma_address       939 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			ttm_dma->dma_address[j] = ttm_dma->dma_address[j - 1] +
dma_address      1074 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		ttm_dma->dma_address[i] = 0;
dma_address        98 drivers/gpu/drm/ttm/ttm_tt.c 					  sizeof(*ttm->dma_address),
dma_address       102 drivers/gpu/drm/ttm/ttm_tt.c 	ttm->dma_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages);
dma_address       108 drivers/gpu/drm/ttm/ttm_tt.c 	ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages,
dma_address       109 drivers/gpu/drm/ttm/ttm_tt.c 					  sizeof(*ttm->dma_address),
dma_address       111 drivers/gpu/drm/ttm/ttm_tt.c 	if (!ttm->dma_address)
dma_address       305 drivers/gpu/drm/ttm/ttm_tt.c 		kvfree(ttm_dma->dma_address);
dma_address       307 drivers/gpu/drm/ttm/ttm_tt.c 	ttm_dma->dma_address = NULL;
dma_address       325 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	dma_addr_t (*dma_address)(struct vmw_piter *);
dma_address       965 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 	return viter->dma_address(viter);
dma_address       334 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->dma_address = &__vmw_piter_phys_addr;
dma_address       338 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->dma_address = &__vmw_piter_dma_addr;
dma_address       344 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		viter->dma_address = &__vmw_piter_sg_addr;
dma_address       429 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 	vsgt->addrs = vmw_tt->dma_ttm.dma_address;
dma_address      1741 drivers/infiniband/ulp/srp/ib_srp.c 		idb_sg->dma_address = req->indirect_dma_addr; /* hack! */
dma_address      2560 drivers/iommu/amd_iommu.c 		s->dma_address = npages << PAGE_SHIFT;
dma_address      2609 drivers/iommu/amd_iommu.c 			bus_addr  = address + s->dma_address + (j << PAGE_SHIFT);
dma_address      2628 drivers/iommu/amd_iommu.c 		s->dma_address += address + (s->offset & ~PAGE_MASK);
dma_address      2633 drivers/iommu/amd_iommu.c 		domain_flush_np_cache(domain, s->dma_address, s->dma_length);
dma_address      2647 drivers/iommu/amd_iommu.c 			bus_addr  = address + s->dma_address + (j << PAGE_SHIFT);
dma_address      2252 drivers/iommu/intel-iommu.c 			sg->dma_address = ((dma_addr_t)iov_pfn << VTD_PAGE_SHIFT) + pgoff;
dma_address      3955 drivers/iommu/intel-iommu.c 		bounce_unmap_page(dev, sg->dma_address,
dma_address      3967 drivers/iommu/intel-iommu.c 		sg->dma_address = bounce_map_page(dev, sg_page(sg),
dma_address      3970 drivers/iommu/intel-iommu.c 		if (sg->dma_address == DMA_MAPPING_ERROR)
dma_address      3118 drivers/misc/habanalabs/goya/goya.c 		sg->dma_address += HOST_PHYS_BASE;
dma_address      3131 drivers/misc/habanalabs/goya/goya.c 		sg->dma_address -= HOST_PHYS_BASE;
dma_address       615 drivers/mmc/host/mtk-sd.c 	dma_addr_t dma_address;
dma_address       635 drivers/mmc/host/mtk-sd.c 		dma_address = sg_dma_address(sg);
dma_address       641 drivers/mmc/host/mtk-sd.c 		bd[j].ptr = lower_32_bits(dma_address);
dma_address       644 drivers/mmc/host/mtk-sd.c 			bd[j].bd_info |= (upper_32_bits(dma_address) & 0xf)
dma_address       573 drivers/mmc/host/wmt-sdmmc.c 	u32 dma_address;
dma_address       633 drivers/mmc/host/wmt-sdmmc.c 		dma_address = priv->dma_desc_device_addr + 16;
dma_address       641 drivers/mmc/host/wmt-sdmmc.c 						dma_address, 0);
dma_address       645 drivers/mmc/host/wmt-sdmmc.c 				dma_address += 16;
dma_address      2331 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		pio[4] = transfer->sgl.dma_address;
dma_address      2377 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c 		pio[4] = transfer->sgl.dma_address;
dma_address      9588 drivers/net/ethernet/sun/niu.c static void niu_pci_unmap_page(struct device *dev, u64 dma_address,
dma_address      9591 drivers/net/ethernet/sun/niu.c 	dma_unmap_page(dev, dma_address, size, direction);
dma_address      9601 drivers/net/ethernet/sun/niu.c static void niu_pci_unmap_single(struct device *dev, u64 dma_address,
dma_address      9605 drivers/net/ethernet/sun/niu.c 	dma_unmap_single(dev, dma_address, size, direction);
dma_address      9972 drivers/net/ethernet/sun/niu.c static void niu_phys_unmap_page(struct device *dev, u64 dma_address,
dma_address      9985 drivers/net/ethernet/sun/niu.c static void niu_phys_unmap_single(struct device *dev, u64 dma_address,
dma_address      3136 drivers/net/ethernet/sun/niu.h 	void (*unmap_page)(struct device *dev, u64 dma_address,
dma_address      3141 drivers/net/ethernet/sun/niu.h 	void (*unmap_single)(struct device *dev, u64 dma_address,
dma_address      1107 drivers/net/fddi/skfp/skfddi.c 	dma_addr_t dma_address;
dma_address      1172 drivers/net/fddi/skfp/skfddi.c 		dma_address = pci_map_single(&bp->pdev, skb->data,
dma_address      1176 drivers/net/fddi/skfp/skfddi.c 			txd->txd_os.dma_addr = dma_address;	// save dma mapping
dma_address      1178 drivers/net/fddi/skfp/skfddi.c 		hwm_tx_frag(smc, skb->data, dma_address, skb->len,
dma_address      1182 drivers/net/fddi/skfp/skfddi.c 			pci_unmap_single(&bp->pdev, dma_address,
dma_address       892 drivers/net/wireless/ti/wl1251/acx.c 	mem_conf->rx_queue_config.dma_address = 0;
dma_address      1123 drivers/net/wireless/ti/wl1251/acx.h 	__le32 dma_address;
dma_address       975 drivers/nvme/host/fc.c 		s->dma_address = 0L;
dma_address       297 drivers/nvme/target/fc.c 		s->dma_address = 0L;
dma_address       835 drivers/pci/p2pdma.c 		s->dma_address = paddr - p2p_pgmap->bus_offset;
dma_address      1960 drivers/scsi/bnx2fc/bnx2fc_hwi.c 				dma_addr_t dma_address;
dma_address      1962 drivers/scsi/bnx2fc/bnx2fc_hwi.c 				dma_address = le32_to_cpu(*pbl);
dma_address      1964 drivers/scsi/bnx2fc/bnx2fc_hwi.c 				dma_address += ((u64)le32_to_cpu(*pbl)) << 32;
dma_address      1969 drivers/scsi/bnx2fc/bnx2fc_hwi.c 						  dma_address);
dma_address       387 drivers/scsi/esp_scsi.c 			s->dma_address = (uintptr_t)sg_virt(s);
dma_address      6842 drivers/scsi/ips.c 	dma_addr_t dma_address;
dma_address      6951 drivers/scsi/ips.c 			&dma_address, GFP_KERNEL);
dma_address      6957 drivers/scsi/ips.c 	ha->adapt->hw_status_start = dma_address;
dma_address      6963 drivers/scsi/ips.c 			sizeof (IPS_LD_INFO), &dma_address, GFP_KERNEL);
dma_address      6969 drivers/scsi/ips.c 	ha->logical_drive_info_dma_addr = dma_address;
dma_address       225 drivers/scsi/lpfc/lpfc_bsg.c 	unsigned char *dma_address, *sg_address;
dma_address       247 drivers/scsi/lpfc/lpfc_bsg.c 			dma_address = mp->virt + dma_offset;
dma_address       261 drivers/scsi/lpfc/lpfc_bsg.c 				memcpy(dma_address, sg_address, transfer_bytes);
dma_address       263 drivers/scsi/lpfc/lpfc_bsg.c 				memcpy(sg_address, dma_address, transfer_bytes);
dma_address      1137 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_EQ_PAGE];
dma_address      1210 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address lwpd;
dma_address      1211 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address buff_fwlog[LPFC_MAX_FWLOG_PAGE];
dma_address      1261 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_CQ_PAGE];
dma_address      1371 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[1];
dma_address      1421 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_WQ_PAGE_V0];
dma_address      1455 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_WQ_PAGE-1];
dma_address      1579 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_RQ_PAGE];
dma_address      1631 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[1];
dma_address      1697 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_MQ_PAGE];
dma_address      1748 drivers/scsi/lpfc/lpfc_hw4.h 			struct dma_address page[LPFC_MAX_MQ_PAGE];
dma_address      8168 drivers/scsi/lpfc/lpfc_init.c 	struct dma_address *dma_address;
dma_address      8211 drivers/scsi/lpfc/lpfc_init.c 	dma_address = &phba->sli4_hba.bmbx.dma_address;
dma_address      8214 drivers/scsi/lpfc/lpfc_init.c 	dma_address->addr_hi = (uint32_t) ((pa_addr << 2) |
dma_address      8218 drivers/scsi/lpfc/lpfc_init.c 	dma_address->addr_lo = (uint32_t) ((pa_addr << 2) |
dma_address      1411 drivers/scsi/lpfc/lpfc_nvme.c 				physaddr = data_sg->dma_address;
dma_address      8550 drivers/scsi/lpfc/lpfc_sli.c 	struct dma_address *dma_address;
dma_address      8589 drivers/scsi/lpfc/lpfc_sli.c 	dma_address = &phba->sli4_hba.bmbx.dma_address;
dma_address      8590 drivers/scsi/lpfc/lpfc_sli.c 	writel(dma_address->addr_hi, phba->sli4_hba.BMBXregaddr);
dma_address      8598 drivers/scsi/lpfc/lpfc_sli.c 	writel(dma_address->addr_lo, phba->sli4_hba.BMBXregaddr);
dma_address      15052 drivers/scsi/lpfc/lpfc_sli.c 		   sizeof(struct dma_address));
dma_address      15500 drivers/scsi/lpfc/lpfc_sli.c 	struct dma_address *page;
dma_address      16073 drivers/scsi/lpfc/lpfc_sli.c 		   sizeof(struct dma_address));
dma_address       404 drivers/scsi/lpfc/lpfc_sli4.h 	struct dma_address dma_address;
dma_address      2134 drivers/scsi/ufs/ufshcd.c 				cpu_to_le32(lower_32_bits(sg->dma_address));
dma_address      2136 drivers/scsi/ufs/ufshcd.c 				cpu_to_le32(upper_32_bits(sg->dma_address));
dma_address       313 drivers/spi/spi-pic32-sqi.c 		bd->bd_addr = sg->dma_address;
dma_address       159 drivers/staging/android/ion/ion.c 		new_sg->dma_address = 0;
dma_address        49 drivers/staging/gasket/gasket.h 	u64 dma_address;
dma_address       196 drivers/staging/gasket/gasket_ioctl.c 						     ibuf.dma_address);
dma_address       206 drivers/staging/gasket/gasket_ioctl.c 						  ibuf.dma_address,
dma_address       210 drivers/staging/gasket/gasket_ioctl.c 						   &ibuf.dma_address,
dma_address      1227 drivers/staging/gasket/gasket_page_table.c 			 dma_addr_t dma_address, ulong vma)
dma_address      1253 drivers/staging/gasket/gasket_page_table.c 				 dma_addr_t *dma_address, u64 index)
dma_address      1288 drivers/staging/gasket/gasket_page_table.c 	*dma_address = driver_desc->coherent_buffer_description.base;
dma_address      1315 drivers/staging/gasket/gasket_page_table.c 				dma_addr_t dma_address, u64 index)
dma_address      1324 drivers/staging/gasket/gasket_page_table.c 	if (driver_desc->coherent_buffer_description.base != dma_address)
dma_address       225 drivers/staging/gasket/gasket_page_table.h 				 dma_addr_t *dma_address, uint64_t index);
dma_address       228 drivers/staging/gasket/gasket_page_table.h 				dma_addr_t dma_address, uint64_t index);
dma_address       247 drivers/staging/gasket/gasket_page_table.h 			 dma_addr_t dma_address, ulong vma);
dma_address      1565 drivers/staging/octeon-usb/octeon-hcd.c 		u64 dma_address = transaction->buffer +
dma_address      1569 drivers/staging/octeon-usb/octeon-hcd.c 			dma_address = transaction->buffer +
dma_address      1577 drivers/staging/octeon-usb/octeon-hcd.c 		cvmx_write64_uint64(reg + channel * 8, dma_address);
dma_address       903 drivers/usb/cdns3/gadget.c 				? trb_dma : request->sg[sg_iter].dma_address);
dma_address        93 drivers/usb/serial/io_ti.c 	__u16 dma_address;
dma_address       546 drivers/usb/serial/io_ti.c 	status = read_ram(port->port->serial->dev, port->dma_address,
dma_address      2599 drivers/usb/serial/io_ti.c 		edge_port->dma_address = UMPD_OEDB1_ADDRESS;
dma_address      2603 drivers/usb/serial/io_ti.c 		edge_port->dma_address = UMPD_OEDB2_ADDRESS;
dma_address      2614 drivers/usb/serial/io_ti.c 		edge_port->dma_address);
dma_address       475 drivers/xen/swiotlb-xen.c 		xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg),
dma_address       490 drivers/xen/swiotlb-xen.c 		sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg),
dma_address       492 drivers/xen/swiotlb-xen.c 		if (sg->dma_address == DMA_MAPPING_ERROR)
dma_address       512 drivers/xen/swiotlb-xen.c 		xen_swiotlb_sync_single_for_cpu(dev, sg->dma_address,
dma_address       525 drivers/xen/swiotlb-xen.c 		xen_swiotlb_sync_single_for_device(dev, sg->dma_address,
dma_address       134 include/drm/ttm/ttm_tt.h 	dma_addr_t *dma_address;
dma_address        58 include/linux/pci-dma-compat.h pci_unmap_page(struct pci_dev *hwdev, dma_addr_t dma_address,
dma_address        61 include/linux/pci-dma-compat.h 	dma_unmap_page(&hwdev->dev, dma_address, size, (enum dma_data_direction)direction);
dma_address       186 include/linux/platform_data/dma-ste-dma40.h 	sg.dma_address = addr;
dma_address        15 include/linux/scatterlist.h 	dma_addr_t	dma_address;
dma_address        34 include/linux/scatterlist.h #define sg_dma_address(sg)	((sg)->dma_address)
dma_address       909 include/trace/events/rpcrdma.h 		__entry->addr = mr->mr_sg->dma_address;
dma_address       939 include/trace/events/rpcrdma.h 		__entry->addr = mr->mr_sg->dma_address;
dma_address       320 kernel/dma/direct.c 		dma_direct_unmap_page(dev, sg->dma_address, sg_dma_len(sg), dir,
dma_address       359 kernel/dma/direct.c 		sg->dma_address = dma_direct_map_page(dev, sg_page(sg),
dma_address       361 kernel/dma/direct.c 		if (sg->dma_address == DMA_MAPPING_ERROR)
dma_address       460 net/smc/smc_ib.c 	if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address)
dma_address       467 net/smc/smc_ib.c 	buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0;
dma_address        10 tools/virtio/linux/scatterlist.h 	dma_addr_t	dma_address;