dma_address 541 arch/alpha/kernel/pci_iommu.c sg->dma_address = -1; dma_address 544 arch/alpha/kernel/pci_iommu.c sg->dma_address = -2; dma_address 549 arch/alpha/kernel/pci_iommu.c leader->dma_address = leader_flag; dma_address 559 arch/alpha/kernel/pci_iommu.c leader->dma_address = leader_flag; dma_address 580 arch/alpha/kernel/pci_iommu.c if (leader->dma_address == 0 dma_address 583 arch/alpha/kernel/pci_iommu.c out->dma_address = paddr + __direct_map_base; dma_address 587 arch/alpha/kernel/pci_iommu.c __va(paddr), size, out->dma_address); dma_address 594 arch/alpha/kernel/pci_iommu.c if (leader->dma_address == 0 && dac_allowed) { dma_address 595 arch/alpha/kernel/pci_iommu.c out->dma_address = paddr + alpha_mv.pci_dac_offset; dma_address 599 arch/alpha/kernel/pci_iommu.c __va(paddr), size, out->dma_address); dma_address 612 arch/alpha/kernel/pci_iommu.c if (leader->dma_address == 0) dma_address 621 arch/alpha/kernel/pci_iommu.c out->dma_address = arena->dma_base + dma_ofs*PAGE_SIZE + paddr; dma_address 625 arch/alpha/kernel/pci_iommu.c __va(paddr), size, out->dma_address, npages); dma_address 639 arch/alpha/kernel/pci_iommu.c while (sg+1 < end && (int) sg[1].dma_address == -1) { dma_address 660 arch/alpha/kernel/pci_iommu.c } while (++sg < end && (int) sg->dma_address < 0); dma_address 683 arch/alpha/kernel/pci_iommu.c sg->dma_address dma_address 686 arch/alpha/kernel/pci_iommu.c return sg->dma_address != DMA_MAPPING_ERROR; dma_address 711 arch/alpha/kernel/pci_iommu.c if ((int) sg->dma_address < 0) dma_address 776 arch/alpha/kernel/pci_iommu.c addr = sg->dma_address; dma_address 95 arch/arm/kernel/dma-isa.c dma->buf.dma_address = dma_map_single(&isa_dma_dev, dma_address 100 arch/arm/kernel/dma-isa.c address = dma->buf.dma_address; dma_address 78 arch/arm/mach-rpc/dma.c idma->dma_addr = idma->dma.sg->dma_address; dma_address 172 arch/arm/mach-rpc/dma.c idma->dma.buf.dma_address = dma_map_single(&isa_dma_dev, dma_address 178 arch/arm/mach-rpc/dma.c idma->dma_addr = idma->dma.sg->dma_address; dma_address 1017 arch/arm/mm/dma-mapping.c s->dma_address = ops->map_page(dev, sg_page(s), s->offset, dma_address 1019 arch/arm/mm/dma-mapping.c if (dma_mapping_error(dev, s->dma_address)) dma_address 1671 arch/arm/mm/dma-mapping.c s->dma_address = DMA_MAPPING_ERROR; dma_address 1675 arch/arm/mm/dma-mapping.c if (__map_sg_chunk(dev, start, size, &dma->dma_address, dma_address 1679 arch/arm/mm/dma-mapping.c dma->dma_address += offset; dma_address 1689 arch/arm/mm/dma-mapping.c if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs, dma_address 1693 arch/arm/mm/dma-mapping.c dma->dma_address += offset; dma_address 391 arch/ia64/hp/common/sba_iommu.c startsg->dma_address, startsg->dma_length, dma_address 1215 arch/ia64/hp/common/sba_iommu.c nents, startsg->dma_address, cnt, dma_address 1219 arch/ia64/hp/common/sba_iommu.c nents, startsg->dma_address, cnt, dma_address 1225 arch/ia64/hp/common/sba_iommu.c if (startsg->dma_address & PIDE_FLAG) { dma_address 1226 arch/ia64/hp/common/sba_iommu.c u32 pide = startsg->dma_address & ~PIDE_FLAG; dma_address 1228 arch/ia64/hp/common/sba_iommu.c startsg->dma_address = 0; dma_address 1231 arch/ia64/hp/common/sba_iommu.c dma_sg->dma_address = pide | ioc->ibase; dma_address 1319 arch/ia64/hp/common/sba_iommu.c startsg->dma_address = startsg->dma_length = 0; dma_address 1331 arch/ia64/hp/common/sba_iommu.c startsg->dma_address = startsg->dma_length = 0; dma_address 1409 arch/ia64/hp/common/sba_iommu.c dma_sg->dma_address = (dma_addr_t)(PIDE_FLAG | (idx << iovp_shift) dma_address 1452 arch/ia64/hp/common/sba_iommu.c sg->dma_address = virt_to_phys(sba_sg_address(sg)); dma_address 1460 arch/ia64/hp/common/sba_iommu.c sglist->dma_address = sba_map_page(dev, sg_page(sglist), dma_address 1462 arch/ia64/hp/common/sba_iommu.c if (dma_mapping_error(dev, sglist->dma_address)) dma_address 1552 arch/ia64/hp/common/sba_iommu.c sba_unmap_page(dev, sglist->dma_address, sglist->dma_length, dma_address 617 arch/mips/jazz/jazzdma.c sg->dma_address = vdma_alloc(sg_phys(sg), sg->length); dma_address 618 arch/mips/jazz/jazzdma.c if (sg->dma_address == DMA_MAPPING_ERROR) dma_address 636 arch/mips/jazz/jazzdma.c vdma_free(sg->dma_address); dma_address 514 arch/powerpc/kernel/iommu.c outs->dma_address = dma_addr; dma_address 535 arch/powerpc/kernel/iommu.c outs->dma_address = DMA_MAPPING_ERROR; dma_address 549 arch/powerpc/kernel/iommu.c vaddr = s->dma_address & IOMMU_PAGE_MASK(tbl); dma_address 550 arch/powerpc/kernel/iommu.c npages = iommu_num_pages(s->dma_address, s->dma_length, dma_address 553 arch/powerpc/kernel/iommu.c s->dma_address = DMA_MAPPING_ERROR; dma_address 577 arch/powerpc/kernel/iommu.c dma_addr_t dma_handle = sg->dma_address; dma_address 645 arch/powerpc/platforms/ps3/system-bus.c sg->length, &sg->dma_address, 0); dma_address 114 arch/powerpc/platforms/pseries/ibmebus.c sg->dma_address = (dma_addr_t) sg_virt(sg); dma_address 497 arch/s390/pci/pci_dma.c s->dma_address = DMA_MAPPING_ERROR; dma_address 503 arch/s390/pci/pci_dma.c &dma->dma_address, dir)) dma_address 506 arch/s390/pci/pci_dma.c dma->dma_address += offset; dma_address 516 arch/s390/pci/pci_dma.c if (__s390_dma_map_sg(dev, start, size, &dma->dma_address, dir)) dma_address 519 arch/s390/pci/pci_dma.c dma->dma_address += offset; dma_address 540 arch/s390/pci/pci_dma.c s390_dma_unmap_pages(dev, s->dma_address, s->dma_length, dma_address 542 arch/s390/pci/pci_dma.c s->dma_address = 0; dma_address 537 arch/sparc/kernel/iommu.c outs->dma_address = dma_addr; dma_address 550 arch/sparc/kernel/iommu.c outs->dma_address = DMA_MAPPING_ERROR; dma_address 562 arch/sparc/kernel/iommu.c vaddr = s->dma_address & IO_PAGE_MASK; dma_address 563 arch/sparc/kernel/iommu.c npages = iommu_num_pages(s->dma_address, s->dma_length, dma_address 576 arch/sparc/kernel/iommu.c s->dma_address = DMA_MAPPING_ERROR; dma_address 599 arch/sparc/kernel/iommu.c bus_addr = sg->dma_address & IO_PAGE_MASK; dma_address 628 arch/sparc/kernel/iommu.c dma_addr_t dma_handle = sg->dma_address; dma_address 723 arch/sparc/kernel/iommu.c iopte = iommu->page_table + ((sglist[0].dma_address - dma_address 729 arch/sparc/kernel/iommu.c bus_addr = sglist[0].dma_address & IO_PAGE_MASK; dma_address 737 arch/sparc/kernel/iommu.c npages = (IO_PAGE_ALIGN(sgprv->dma_address + sgprv->dma_length) dma_address 579 arch/sparc/kernel/pci_sun4v.c outs->dma_address = dma_addr; dma_address 597 arch/sparc/kernel/pci_sun4v.c outs->dma_address = DMA_MAPPING_ERROR; dma_address 608 arch/sparc/kernel/pci_sun4v.c vaddr = s->dma_address & IO_PAGE_MASK; dma_address 609 arch/sparc/kernel/pci_sun4v.c npages = iommu_num_pages(s->dma_address, s->dma_length, dma_address 614 arch/sparc/kernel/pci_sun4v.c s->dma_address = DMA_MAPPING_ERROR; dma_address 648 arch/sparc/kernel/pci_sun4v.c dma_addr_t dma_handle = sg->dma_address; dma_address 172 arch/sparc/mm/io-unit.c sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length); dma_address 204 arch/sparc/mm/io-unit.c len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT; dma_address 205 arch/sparc/mm/io-unit.c vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT; dma_address 251 arch/sparc/mm/iommu.c sg->dma_address =__sbus_iommu_map_page(dev, sg_page(sg), dma_address 253 arch/sparc/mm/iommu.c if (sg->dma_address == DMA_MAPPING_ERROR) dma_address 300 arch/sparc/mm/iommu.c sbus_iommu_unmap_page(dev, sg->dma_address, sg->length, dir, dma_address 302 arch/sparc/mm/iommu.c sg->dma_address = 0x21212121; dma_address 288 arch/x86/kernel/amd_gart_64.c gart_unmap_page(dev, s->dma_address, s->dma_length, dir, 0); dma_address 316 arch/x86/kernel/amd_gart_64.c s->dma_address = addr; dma_address 339 arch/x86/kernel/amd_gart_64.c unsigned long phys_addr = s->dma_address; dma_address 343 arch/x86/kernel/amd_gart_64.c sout->dma_address = iommu_bus_base; dma_address 344 arch/x86/kernel/amd_gart_64.c sout->dma_address += iommu_page*PAGE_SIZE + s->offset; dma_address 369 arch/x86/kernel/amd_gart_64.c sout->dma_address = start->dma_address; dma_address 403 arch/x86/kernel/amd_gart_64.c s->dma_address = addr; dma_address 461 arch/x86/kernel/amd_gart_64.c s->dma_address = DMA_MAPPING_ERROR; dma_address 338 arch/x86/kernel/pci-calgary_64.c dma_addr_t dma = s->dma_address; dma_address 373 arch/x86/kernel/pci-calgary_64.c s->dma_address = (entry << PAGE_SHIFT) | s->offset; dma_address 385 arch/x86/kernel/pci-calgary_64.c sg->dma_address = DMA_MAPPING_ERROR; dma_address 236 drivers/atm/idt77252.h u32 dma_address; dma_address 309 drivers/atm/nicstar.h u32 dma_address; dma_address 346 drivers/atm/nicstar.h u32 dma_address; dma_address 889 drivers/block/skd_main.c uint64_t dma_address; dma_address 898 drivers/block/skd_main.c dma_address = skspcl->req.sksg_dma_address; dma_address 899 drivers/block/skd_main.c scsi->hdr.sg_list_dma_address = cpu_to_be64(dma_address); dma_address 2713 drivers/block/skd_main.c uint64_t dma_address = *ret_dma_addr; dma_address 2720 drivers/block/skd_main.c sg_list[i].next_desc_ptr = dma_address + ndp_off; dma_address 121 drivers/crypto/hisilicon/sgl.c hw_sge->buf = sgl->dma_address; dma_address 504 drivers/crypto/sahara.c dev->hw_link[i]->p = sg->dma_address; dma_address 518 drivers/crypto/sahara.c dev->hw_link[j]->p = sg->dma_address; dma_address 822 drivers/crypto/sahara.c dev->hw_link[i]->p = sg->dma_address; dma_address 293 drivers/dma/imx-dma.c imx_dmav1_writel(imxdma, sg->dma_address, dma_address 296 drivers/dma/imx-dma.c imx_dmav1_writel(imxdma, sg->dma_address, dma_address 837 drivers/dma/imx-dma.c if (sg_dma_len(sgl) & 3 || sgl->dma_address & 3) dma_address 841 drivers/dma/imx-dma.c if (sg_dma_len(sgl) & 1 || sgl->dma_address & 1) dma_address 898 drivers/dma/imx-dma.c imxdmac->sg_list[i].dma_address = dma_addr; dma_address 1465 drivers/dma/imx-sdma.c bd->buffer_addr = sg->dma_address; dma_address 1484 drivers/dma/imx-sdma.c if (count & 3 || sg->dma_address & 3) dma_address 1489 drivers/dma/imx-sdma.c if (count & 1 || sg->dma_address & 1) dma_address 1508 drivers/dma/imx-sdma.c i, count, (u64)sg->dma_address, dma_address 564 drivers/dma/mxs-dma.c ccw->bufaddr = sg->dma_address; dma_address 781 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sg->sgl->dma_address = addr; dma_address 50 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c *addr = ttm->dma_address[0]; dma_address 128 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c if (ttm->dma_address[0] + PAGE_SIZE >= adev->gmc.agp_size) dma_address 131 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c return adev->gmc.agp_start + ttm->dma_address[0]; dma_address 964 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c gtt->ttm.dma_address, ttm->num_pages); dma_address 1015 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c ttm->pages, gtt->ttm.dma_address, flags); dma_address 1027 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c &(gtt->ttm.dma_address[page_idx]), flags); dma_address 1030 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c ttm->pages, gtt->ttm.dma_address, flags); dma_address 1083 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c ttm->pages, gtt->ttm.dma_address, flags); dma_address 1270 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c gtt->ttm.dma_address, dma_address 1908 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c dma_addr_t *dma_address; dma_address 1942 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c dma_address = >t->ttm.dma_address[offset >> PAGE_SHIFT]; dma_address 1944 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c r = amdgpu_gart_map(adev, 0, num_pages, dma_address, flags, dma_address 1707 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c pages_addr = ttm->dma_address; dma_address 521 drivers/gpu/drm/exynos/exynos_drm_g2d.c g2d_userptr->dma_addr = sgt->sgl[0].dma_address; dma_address 59 drivers/gpu/drm/i915/i915_gpu_error.c sg->dma_address = it; dma_address 844 drivers/gpu/drm/i915/i915_gpu_error.c if (!sg || off < sg->dma_address) dma_address 849 drivers/gpu/drm/i915/i915_gpu_error.c pos = sg->dma_address; dma_address 50 drivers/gpu/drm/msm/msm_gpummu.c dma_addr_t addr = sg->dma_address; dma_address 554 drivers/gpu/drm/nouveau/nouveau_bo.c ttm_dma->dma_address[i], dma_address 573 drivers/gpu/drm/nouveau/nouveau_bo.c dma_sync_single_for_cpu(drm->dev->dev, ttm_dma->dma_address[i], dma_address 1599 drivers/gpu/drm/nouveau/nouveau_bo.c ttm_dma->dma_address, ttm->num_pages); dma_address 1632 drivers/gpu/drm/nouveau/nouveau_bo.c dma_unmap_page(dev, ttm_dma->dma_address[i], dma_address 1634 drivers/gpu/drm/nouveau/nouveau_bo.c ttm_dma->dma_address[i] = 0; dma_address 1640 drivers/gpu/drm/nouveau/nouveau_bo.c ttm_dma->dma_address[i] = addr; dma_address 1675 drivers/gpu/drm/nouveau/nouveau_bo.c if (ttm_dma->dma_address[i]) { dma_address 1676 drivers/gpu/drm/nouveau/nouveau_bo.c dma_unmap_page(dev, ttm_dma->dma_address[i], PAGE_SIZE, dma_address 120 drivers/gpu/drm/nouveau/nouveau_mem.c else args.dma = tt->dma_address; dma_address 535 drivers/gpu/drm/radeon/radeon_ttm.c gtt->ttm.dma_address, ttm->num_pages); dma_address 597 drivers/gpu/drm/radeon/radeon_ttm.c ttm->pages, gtt->ttm.dma_address, flags); dma_address 685 drivers/gpu/drm/radeon/radeon_ttm.c gtt->ttm.dma_address, ttm->num_pages); dma_address 1118 drivers/gpu/drm/ttm/ttm_page_alloc.c tt->dma_address[i] = dma_map_page(dev, tt->ttm.pages[i], dma_address 1121 drivers/gpu/drm/ttm/ttm_page_alloc.c if (dma_mapping_error(dev, tt->dma_address[i])) { dma_address 1123 drivers/gpu/drm/ttm/ttm_page_alloc.c dma_unmap_page(dev, tt->dma_address[i], dma_address 1125 drivers/gpu/drm/ttm/ttm_page_alloc.c tt->dma_address[i] = 0; dma_address 1132 drivers/gpu/drm/ttm/ttm_page_alloc.c tt->dma_address[i + 1] = tt->dma_address[i] + PAGE_SIZE; dma_address 1148 drivers/gpu/drm/ttm/ttm_page_alloc.c if (!tt->dma_address[i] || !tt->ttm.pages[i]) { dma_address 1160 drivers/gpu/drm/ttm/ttm_page_alloc.c dma_unmap_page(dev, tt->dma_address[i], num_pages * PAGE_SIZE, dma_address 848 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma->dma_address[index] = d_page->dma; dma_address 939 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma->dma_address[j] = ttm_dma->dma_address[j - 1] + dma_address 1074 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma->dma_address[i] = 0; dma_address 98 drivers/gpu/drm/ttm/ttm_tt.c sizeof(*ttm->dma_address), dma_address 102 drivers/gpu/drm/ttm/ttm_tt.c ttm->dma_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages); dma_address 108 drivers/gpu/drm/ttm/ttm_tt.c ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, dma_address 109 drivers/gpu/drm/ttm/ttm_tt.c sizeof(*ttm->dma_address), dma_address 111 drivers/gpu/drm/ttm/ttm_tt.c if (!ttm->dma_address) dma_address 305 drivers/gpu/drm/ttm/ttm_tt.c kvfree(ttm_dma->dma_address); dma_address 307 drivers/gpu/drm/ttm/ttm_tt.c ttm_dma->dma_address = NULL; dma_address 325 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h dma_addr_t (*dma_address)(struct vmw_piter *); dma_address 965 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h return viter->dma_address(viter); dma_address 334 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->dma_address = &__vmw_piter_phys_addr; dma_address 338 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->dma_address = &__vmw_piter_dma_addr; dma_address 344 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c viter->dma_address = &__vmw_piter_sg_addr; dma_address 429 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c vsgt->addrs = vmw_tt->dma_ttm.dma_address; dma_address 1741 drivers/infiniband/ulp/srp/ib_srp.c idb_sg->dma_address = req->indirect_dma_addr; /* hack! */ dma_address 2560 drivers/iommu/amd_iommu.c s->dma_address = npages << PAGE_SHIFT; dma_address 2609 drivers/iommu/amd_iommu.c bus_addr = address + s->dma_address + (j << PAGE_SHIFT); dma_address 2628 drivers/iommu/amd_iommu.c s->dma_address += address + (s->offset & ~PAGE_MASK); dma_address 2633 drivers/iommu/amd_iommu.c domain_flush_np_cache(domain, s->dma_address, s->dma_length); dma_address 2647 drivers/iommu/amd_iommu.c bus_addr = address + s->dma_address + (j << PAGE_SHIFT); dma_address 2252 drivers/iommu/intel-iommu.c sg->dma_address = ((dma_addr_t)iov_pfn << VTD_PAGE_SHIFT) + pgoff; dma_address 3955 drivers/iommu/intel-iommu.c bounce_unmap_page(dev, sg->dma_address, dma_address 3967 drivers/iommu/intel-iommu.c sg->dma_address = bounce_map_page(dev, sg_page(sg), dma_address 3970 drivers/iommu/intel-iommu.c if (sg->dma_address == DMA_MAPPING_ERROR) dma_address 3118 drivers/misc/habanalabs/goya/goya.c sg->dma_address += HOST_PHYS_BASE; dma_address 3131 drivers/misc/habanalabs/goya/goya.c sg->dma_address -= HOST_PHYS_BASE; dma_address 615 drivers/mmc/host/mtk-sd.c dma_addr_t dma_address; dma_address 635 drivers/mmc/host/mtk-sd.c dma_address = sg_dma_address(sg); dma_address 641 drivers/mmc/host/mtk-sd.c bd[j].ptr = lower_32_bits(dma_address); dma_address 644 drivers/mmc/host/mtk-sd.c bd[j].bd_info |= (upper_32_bits(dma_address) & 0xf) dma_address 573 drivers/mmc/host/wmt-sdmmc.c u32 dma_address; dma_address 633 drivers/mmc/host/wmt-sdmmc.c dma_address = priv->dma_desc_device_addr + 16; dma_address 641 drivers/mmc/host/wmt-sdmmc.c dma_address, 0); dma_address 645 drivers/mmc/host/wmt-sdmmc.c dma_address += 16; dma_address 2331 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c pio[4] = transfer->sgl.dma_address; dma_address 2377 drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c pio[4] = transfer->sgl.dma_address; dma_address 9588 drivers/net/ethernet/sun/niu.c static void niu_pci_unmap_page(struct device *dev, u64 dma_address, dma_address 9591 drivers/net/ethernet/sun/niu.c dma_unmap_page(dev, dma_address, size, direction); dma_address 9601 drivers/net/ethernet/sun/niu.c static void niu_pci_unmap_single(struct device *dev, u64 dma_address, dma_address 9605 drivers/net/ethernet/sun/niu.c dma_unmap_single(dev, dma_address, size, direction); dma_address 9972 drivers/net/ethernet/sun/niu.c static void niu_phys_unmap_page(struct device *dev, u64 dma_address, dma_address 9985 drivers/net/ethernet/sun/niu.c static void niu_phys_unmap_single(struct device *dev, u64 dma_address, dma_address 3136 drivers/net/ethernet/sun/niu.h void (*unmap_page)(struct device *dev, u64 dma_address, dma_address 3141 drivers/net/ethernet/sun/niu.h void (*unmap_single)(struct device *dev, u64 dma_address, dma_address 1107 drivers/net/fddi/skfp/skfddi.c dma_addr_t dma_address; dma_address 1172 drivers/net/fddi/skfp/skfddi.c dma_address = pci_map_single(&bp->pdev, skb->data, dma_address 1176 drivers/net/fddi/skfp/skfddi.c txd->txd_os.dma_addr = dma_address; // save dma mapping dma_address 1178 drivers/net/fddi/skfp/skfddi.c hwm_tx_frag(smc, skb->data, dma_address, skb->len, dma_address 1182 drivers/net/fddi/skfp/skfddi.c pci_unmap_single(&bp->pdev, dma_address, dma_address 892 drivers/net/wireless/ti/wl1251/acx.c mem_conf->rx_queue_config.dma_address = 0; dma_address 1123 drivers/net/wireless/ti/wl1251/acx.h __le32 dma_address; dma_address 975 drivers/nvme/host/fc.c s->dma_address = 0L; dma_address 297 drivers/nvme/target/fc.c s->dma_address = 0L; dma_address 835 drivers/pci/p2pdma.c s->dma_address = paddr - p2p_pgmap->bus_offset; dma_address 1960 drivers/scsi/bnx2fc/bnx2fc_hwi.c dma_addr_t dma_address; dma_address 1962 drivers/scsi/bnx2fc/bnx2fc_hwi.c dma_address = le32_to_cpu(*pbl); dma_address 1964 drivers/scsi/bnx2fc/bnx2fc_hwi.c dma_address += ((u64)le32_to_cpu(*pbl)) << 32; dma_address 1969 drivers/scsi/bnx2fc/bnx2fc_hwi.c dma_address); dma_address 387 drivers/scsi/esp_scsi.c s->dma_address = (uintptr_t)sg_virt(s); dma_address 6842 drivers/scsi/ips.c dma_addr_t dma_address; dma_address 6951 drivers/scsi/ips.c &dma_address, GFP_KERNEL); dma_address 6957 drivers/scsi/ips.c ha->adapt->hw_status_start = dma_address; dma_address 6963 drivers/scsi/ips.c sizeof (IPS_LD_INFO), &dma_address, GFP_KERNEL); dma_address 6969 drivers/scsi/ips.c ha->logical_drive_info_dma_addr = dma_address; dma_address 225 drivers/scsi/lpfc/lpfc_bsg.c unsigned char *dma_address, *sg_address; dma_address 247 drivers/scsi/lpfc/lpfc_bsg.c dma_address = mp->virt + dma_offset; dma_address 261 drivers/scsi/lpfc/lpfc_bsg.c memcpy(dma_address, sg_address, transfer_bytes); dma_address 263 drivers/scsi/lpfc/lpfc_bsg.c memcpy(sg_address, dma_address, transfer_bytes); dma_address 1137 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_EQ_PAGE]; dma_address 1210 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address lwpd; dma_address 1211 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address buff_fwlog[LPFC_MAX_FWLOG_PAGE]; dma_address 1261 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_CQ_PAGE]; dma_address 1371 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[1]; dma_address 1421 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_WQ_PAGE_V0]; dma_address 1455 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_WQ_PAGE-1]; dma_address 1579 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_RQ_PAGE]; dma_address 1631 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[1]; dma_address 1697 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_MQ_PAGE]; dma_address 1748 drivers/scsi/lpfc/lpfc_hw4.h struct dma_address page[LPFC_MAX_MQ_PAGE]; dma_address 8168 drivers/scsi/lpfc/lpfc_init.c struct dma_address *dma_address; dma_address 8211 drivers/scsi/lpfc/lpfc_init.c dma_address = &phba->sli4_hba.bmbx.dma_address; dma_address 8214 drivers/scsi/lpfc/lpfc_init.c dma_address->addr_hi = (uint32_t) ((pa_addr << 2) | dma_address 8218 drivers/scsi/lpfc/lpfc_init.c dma_address->addr_lo = (uint32_t) ((pa_addr << 2) | dma_address 1411 drivers/scsi/lpfc/lpfc_nvme.c physaddr = data_sg->dma_address; dma_address 8550 drivers/scsi/lpfc/lpfc_sli.c struct dma_address *dma_address; dma_address 8589 drivers/scsi/lpfc/lpfc_sli.c dma_address = &phba->sli4_hba.bmbx.dma_address; dma_address 8590 drivers/scsi/lpfc/lpfc_sli.c writel(dma_address->addr_hi, phba->sli4_hba.BMBXregaddr); dma_address 8598 drivers/scsi/lpfc/lpfc_sli.c writel(dma_address->addr_lo, phba->sli4_hba.BMBXregaddr); dma_address 15052 drivers/scsi/lpfc/lpfc_sli.c sizeof(struct dma_address)); dma_address 15500 drivers/scsi/lpfc/lpfc_sli.c struct dma_address *page; dma_address 16073 drivers/scsi/lpfc/lpfc_sli.c sizeof(struct dma_address)); dma_address 404 drivers/scsi/lpfc/lpfc_sli4.h struct dma_address dma_address; dma_address 2134 drivers/scsi/ufs/ufshcd.c cpu_to_le32(lower_32_bits(sg->dma_address)); dma_address 2136 drivers/scsi/ufs/ufshcd.c cpu_to_le32(upper_32_bits(sg->dma_address)); dma_address 313 drivers/spi/spi-pic32-sqi.c bd->bd_addr = sg->dma_address; dma_address 159 drivers/staging/android/ion/ion.c new_sg->dma_address = 0; dma_address 49 drivers/staging/gasket/gasket.h u64 dma_address; dma_address 196 drivers/staging/gasket/gasket_ioctl.c ibuf.dma_address); dma_address 206 drivers/staging/gasket/gasket_ioctl.c ibuf.dma_address, dma_address 210 drivers/staging/gasket/gasket_ioctl.c &ibuf.dma_address, dma_address 1227 drivers/staging/gasket/gasket_page_table.c dma_addr_t dma_address, ulong vma) dma_address 1253 drivers/staging/gasket/gasket_page_table.c dma_addr_t *dma_address, u64 index) dma_address 1288 drivers/staging/gasket/gasket_page_table.c *dma_address = driver_desc->coherent_buffer_description.base; dma_address 1315 drivers/staging/gasket/gasket_page_table.c dma_addr_t dma_address, u64 index) dma_address 1324 drivers/staging/gasket/gasket_page_table.c if (driver_desc->coherent_buffer_description.base != dma_address) dma_address 225 drivers/staging/gasket/gasket_page_table.h dma_addr_t *dma_address, uint64_t index); dma_address 228 drivers/staging/gasket/gasket_page_table.h dma_addr_t dma_address, uint64_t index); dma_address 247 drivers/staging/gasket/gasket_page_table.h dma_addr_t dma_address, ulong vma); dma_address 1565 drivers/staging/octeon-usb/octeon-hcd.c u64 dma_address = transaction->buffer + dma_address 1569 drivers/staging/octeon-usb/octeon-hcd.c dma_address = transaction->buffer + dma_address 1577 drivers/staging/octeon-usb/octeon-hcd.c cvmx_write64_uint64(reg + channel * 8, dma_address); dma_address 903 drivers/usb/cdns3/gadget.c ? trb_dma : request->sg[sg_iter].dma_address); dma_address 93 drivers/usb/serial/io_ti.c __u16 dma_address; dma_address 546 drivers/usb/serial/io_ti.c status = read_ram(port->port->serial->dev, port->dma_address, dma_address 2599 drivers/usb/serial/io_ti.c edge_port->dma_address = UMPD_OEDB1_ADDRESS; dma_address 2603 drivers/usb/serial/io_ti.c edge_port->dma_address = UMPD_OEDB2_ADDRESS; dma_address 2614 drivers/usb/serial/io_ti.c edge_port->dma_address); dma_address 475 drivers/xen/swiotlb-xen.c xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg), dma_address 490 drivers/xen/swiotlb-xen.c sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg), dma_address 492 drivers/xen/swiotlb-xen.c if (sg->dma_address == DMA_MAPPING_ERROR) dma_address 512 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_single_for_cpu(dev, sg->dma_address, dma_address 525 drivers/xen/swiotlb-xen.c xen_swiotlb_sync_single_for_device(dev, sg->dma_address, dma_address 134 include/drm/ttm/ttm_tt.h dma_addr_t *dma_address; dma_address 58 include/linux/pci-dma-compat.h pci_unmap_page(struct pci_dev *hwdev, dma_addr_t dma_address, dma_address 61 include/linux/pci-dma-compat.h dma_unmap_page(&hwdev->dev, dma_address, size, (enum dma_data_direction)direction); dma_address 186 include/linux/platform_data/dma-ste-dma40.h sg.dma_address = addr; dma_address 15 include/linux/scatterlist.h dma_addr_t dma_address; dma_address 34 include/linux/scatterlist.h #define sg_dma_address(sg) ((sg)->dma_address) dma_address 909 include/trace/events/rpcrdma.h __entry->addr = mr->mr_sg->dma_address; dma_address 939 include/trace/events/rpcrdma.h __entry->addr = mr->mr_sg->dma_address; dma_address 320 kernel/dma/direct.c dma_direct_unmap_page(dev, sg->dma_address, sg_dma_len(sg), dir, dma_address 359 kernel/dma/direct.c sg->dma_address = dma_direct_map_page(dev, sg_page(sg), dma_address 361 kernel/dma/direct.c if (sg->dma_address == DMA_MAPPING_ERROR) dma_address 460 net/smc/smc_ib.c if (!buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address) dma_address 467 net/smc/smc_ib.c buf_slot->sgt[SMC_SINGLE_LINK].sgl->dma_address = 0; dma_address 10 tools/virtio/linux/scatterlist.h dma_addr_t dma_address;