cpu_addr 264 arch/alpha/kernel/pci_iommu.c pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size, cpu_addr 276 arch/alpha/kernel/pci_iommu.c paddr = __pa(cpu_addr); cpu_addr 285 arch/alpha/kernel/pci_iommu.c cpu_addr, size, ret, __builtin_return_address(0)); cpu_addr 296 arch/alpha/kernel/pci_iommu.c cpu_addr, size, ret, __builtin_return_address(0)); cpu_addr 330 arch/alpha/kernel/pci_iommu.c ret += (unsigned long)cpu_addr & ~PAGE_MASK; cpu_addr 333 arch/alpha/kernel/pci_iommu.c cpu_addr, size, npages, ret, __builtin_return_address(0)); cpu_addr 452 arch/alpha/kernel/pci_iommu.c void *cpu_addr; cpu_addr 458 arch/alpha/kernel/pci_iommu.c cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order); cpu_addr 459 arch/alpha/kernel/pci_iommu.c if (! cpu_addr) { cpu_addr 467 arch/alpha/kernel/pci_iommu.c memset(cpu_addr, 0, size); cpu_addr 469 arch/alpha/kernel/pci_iommu.c *dma_addrp = pci_map_single_1(pdev, cpu_addr, size, 0); cpu_addr 471 arch/alpha/kernel/pci_iommu.c free_pages((unsigned long)cpu_addr, order); cpu_addr 481 arch/alpha/kernel/pci_iommu.c size, cpu_addr, *dma_addrp, __builtin_return_address(0)); cpu_addr 483 arch/alpha/kernel/pci_iommu.c return cpu_addr; cpu_addr 493 arch/alpha/kernel/pci_iommu.c void *cpu_addr, dma_addr_t dma_addr, cpu_addr 498 arch/alpha/kernel/pci_iommu.c free_pages((unsigned long)cpu_addr, get_order(size)); cpu_addr 122 arch/arm/include/asm/dma-mapping.h extern void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 139 arch/arm/include/asm/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 195 arch/arm/include/asm/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 268 arch/arm/mach-omap2/common.h extern void omap_auxcoreboot_addr(u32 cpu_addr); cpu_addr 55 arch/arm/mm/dma-mapping-nommu.c void *cpu_addr, dma_addr_t dma_addr, cpu_addr 58 arch/arm/mm/dma-mapping-nommu.c int ret = dma_release_from_global_coherent(get_order(size), cpu_addr); cpu_addr 64 arch/arm/mm/dma-mapping-nommu.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 69 arch/arm/mm/dma-mapping-nommu.c if (dma_mmap_from_global_coherent(vma, cpu_addr, size, &ret)) cpu_addr 71 arch/arm/mm/dma-mapping-nommu.c if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) cpu_addr 57 arch/arm/mm/dma-mapping.c void *cpu_addr; cpu_addr 203 arch/arm/mm/dma-mapping.c static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 206 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 581 arch/arm/mm/dma-mapping.c void *cpu_addr, size_t size, bool want_vaddr) cpu_addr 585 arch/arm/mm/dma-mapping.c dma_common_free_remap(cpu_addr, size); cpu_addr 641 arch/arm/mm/dma-mapping.c __free_from_contiguous(args->dev, args->page, args->cpu_addr, cpu_addr 658 arch/arm/mm/dma-mapping.c __free_from_pool(args->cpu_addr, args->size); cpu_addr 677 arch/arm/mm/dma-mapping.c dma_common_free_remap(args->cpu_addr, args->size); cpu_addr 788 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 797 arch/arm/mm/dma-mapping.c if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) cpu_addr 814 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 817 arch/arm/mm/dma-mapping.c return __arm_dma_mmap(dev, vma, cpu_addr, dma_addr, size, attrs); cpu_addr 821 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 825 arch/arm/mm/dma-mapping.c return __arm_dma_mmap(dev, vma, cpu_addr, dma_addr, size, attrs); cpu_addr 831 arch/arm/mm/dma-mapping.c static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 840 arch/arm/mm/dma-mapping.c .cpu_addr = cpu_addr, cpu_addr 845 arch/arm/mm/dma-mapping.c buf = arm_dma_buffer_find(cpu_addr); cpu_addr 846 arch/arm/mm/dma-mapping.c if (WARN(!buf, "Freeing invalid buffer %p\n", cpu_addr)) cpu_addr 853 arch/arm/mm/dma-mapping.c void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 856 arch/arm/mm/dma-mapping.c __arm_dma_free(dev, size, cpu_addr, handle, attrs, false); cpu_addr 859 arch/arm/mm/dma-mapping.c static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 862 arch/arm/mm/dma-mapping.c __arm_dma_free(dev, size, cpu_addr, handle, attrs, true); cpu_addr 866 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t handle, size_t size, cpu_addr 1415 arch/arm/mm/dma-mapping.c static struct page **__iommu_get_pages(void *cpu_addr, unsigned long attrs) cpu_addr 1417 arch/arm/mm/dma-mapping.c if (__in_atomic_pool(cpu_addr, PAGE_SIZE)) cpu_addr 1418 arch/arm/mm/dma-mapping.c return __atomic_get_pages(cpu_addr); cpu_addr 1421 arch/arm/mm/dma-mapping.c return cpu_addr; cpu_addr 1423 arch/arm/mm/dma-mapping.c return dma_common_find_pages(cpu_addr); cpu_addr 1451 arch/arm/mm/dma-mapping.c static void __iommu_free_atomic(struct device *dev, void *cpu_addr, cpu_addr 1456 arch/arm/mm/dma-mapping.c __dma_free_buffer(virt_to_page(cpu_addr), size); cpu_addr 1458 arch/arm/mm/dma-mapping.c __free_from_pool(cpu_addr, size); cpu_addr 1523 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 1526 arch/arm/mm/dma-mapping.c struct page **pages = __iommu_get_pages(cpu_addr, attrs); cpu_addr 1543 arch/arm/mm/dma-mapping.c struct vm_area_struct *vma, void *cpu_addr, cpu_addr 1548 arch/arm/mm/dma-mapping.c return __arm_iommu_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, attrs); cpu_addr 1552 arch/arm/mm/dma-mapping.c struct vm_area_struct *vma, void *cpu_addr, cpu_addr 1555 arch/arm/mm/dma-mapping.c return __arm_iommu_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, attrs); cpu_addr 1562 arch/arm/mm/dma-mapping.c void __arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, cpu_addr 1568 arch/arm/mm/dma-mapping.c if (coherent_flag == COHERENT || __in_atomic_pool(cpu_addr, size)) { cpu_addr 1569 arch/arm/mm/dma-mapping.c __iommu_free_atomic(dev, cpu_addr, handle, size, coherent_flag); cpu_addr 1573 arch/arm/mm/dma-mapping.c pages = __iommu_get_pages(cpu_addr, attrs); cpu_addr 1575 arch/arm/mm/dma-mapping.c WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); cpu_addr 1580 arch/arm/mm/dma-mapping.c dma_common_free_remap(cpu_addr, size); cpu_addr 1587 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t handle, unsigned long attrs) cpu_addr 1589 arch/arm/mm/dma-mapping.c __arm_iommu_free_attrs(dev, size, cpu_addr, handle, attrs, NORMAL); cpu_addr 1593 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t handle, unsigned long attrs) cpu_addr 1595 arch/arm/mm/dma-mapping.c __arm_iommu_free_attrs(dev, size, cpu_addr, handle, attrs, COHERENT); cpu_addr 1599 arch/arm/mm/dma-mapping.c void *cpu_addr, dma_addr_t dma_addr, cpu_addr 1603 arch/arm/mm/dma-mapping.c struct page **pages = __iommu_get_pages(cpu_addr, attrs); cpu_addr 2349 arch/arm/mm/dma-mapping.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, cpu_addr 2363 arch/arm/mm/dma-mapping.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 2366 arch/arm/mm/dma-mapping.c __arm_dma_free(dev, size, cpu_addr, dma_handle, attrs, false); cpu_addr 17 arch/ia64/kernel/dma-mapping.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 20 arch/ia64/kernel/dma-mapping.c dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs); cpu_addr 23 arch/ia64/kernel/dma-mapping.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, cpu_addr 26 arch/ia64/kernel/dma-mapping.c return page_to_pfn(virt_to_page(cpu_addr)); cpu_addr 439 arch/microblaze/pci/pci-common.c range.cpu_addr, range.size); cpu_addr 446 arch/microblaze/pci/pci-common.c if (range.cpu_addr == OF_BAD_ADDR || range.size == 0) cpu_addr 454 arch/microblaze/pci/pci-common.c range.cpu_addr, range.cpu_addr + range.size - 1, cpu_addr 467 arch/microblaze/pci/pci-common.c hose->io_base_virt = ioremap(range.cpu_addr, cpu_addr 478 arch/microblaze/pci/pci-common.c hose->io_base_phys = range.cpu_addr - range.pci_addr; cpu_addr 482 arch/microblaze/pci/pci-common.c range.cpu_addr = range.pci_addr; cpu_addr 487 arch/microblaze/pci/pci-common.c range.cpu_addr, range.cpu_addr + range.size - 1, cpu_addr 499 arch/microblaze/pci/pci-common.c isa_mb = range.cpu_addr; cpu_addr 502 arch/microblaze/pci/pci-common.c isa_mem_base = range.cpu_addr; cpu_addr 503 arch/microblaze/pci/pci-common.c hose->isa_mem_phys = range.cpu_addr; cpu_addr 514 arch/microblaze/pci/pci-common.c hose->pci_mem_offset = range.cpu_addr - cpu_addr 517 arch/microblaze/pci/pci-common.c hose->pci_mem_offset != range.cpu_addr - cpu_addr 530 arch/microblaze/pci/pci-common.c res->start = range.cpu_addr; cpu_addr 531 arch/microblaze/pci/pci-common.c res->end = range.cpu_addr + range.size - 1; cpu_addr 62 arch/mips/mm/dma-noncoherent.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, cpu_addr 65 arch/mips/mm/dma-noncoherent.c return page_to_pfn(virt_to_page(cached_kernel_address(cpu_addr))); cpu_addr 155 arch/mips/pci/pci-legacy.c range.cpu_addr, cpu_addr 156 arch/mips/pci/pci-legacy.c range.cpu_addr + range.size - 1); cpu_addr 158 arch/mips/pci/pci-legacy.c (unsigned long)ioremap(range.cpu_addr, cpu_addr 164 arch/mips/pci/pci-legacy.c range.cpu_addr, cpu_addr 165 arch/mips/pci/pci-legacy.c range.cpu_addr + range.size - 1); cpu_addr 690 arch/powerpc/kernel/pci-common.c if (range.cpu_addr == OF_BAD_ADDR || range.size == 0) cpu_addr 699 arch/powerpc/kernel/pci-common.c range.cpu_addr, range.cpu_addr + range.size - 1, cpu_addr 714 arch/powerpc/kernel/pci-common.c hose->io_base_virt = ioremap(range.cpu_addr, cpu_addr 726 arch/powerpc/kernel/pci-common.c hose->io_base_phys = range.cpu_addr - range.pci_addr; cpu_addr 730 arch/powerpc/kernel/pci-common.c range.cpu_addr = range.pci_addr; cpu_addr 735 arch/powerpc/kernel/pci-common.c range.cpu_addr, range.cpu_addr + range.size - 1, cpu_addr 749 arch/powerpc/kernel/pci-common.c isa_mem_base = range.cpu_addr; cpu_addr 750 arch/powerpc/kernel/pci-common.c hose->isa_mem_phys = range.cpu_addr; cpu_addr 755 arch/powerpc/kernel/pci-common.c hose->mem_offset[memno] = range.cpu_addr - cpu_addr 763 arch/powerpc/kernel/pci-common.c res->start = range.cpu_addr; cpu_addr 764 arch/powerpc/kernel/pci-common.c res->end = range.cpu_addr + range.size - 1; cpu_addr 116 arch/powerpc/platforms/4xx/pci.c u64 cpu_addr = of_translate_dma_address(hose->dn, ranges + 3); cpu_addr 119 arch/powerpc/platforms/4xx/pci.c if (cpu_addr == OF_BAD_ADDR || size == 0) cpu_addr 129 arch/powerpc/platforms/4xx/pci.c if (cpu_addr != 0 || pci_addr > 0xffffffff) { cpu_addr 133 arch/powerpc/platforms/4xx/pci.c pci_addr, pci_addr + size - 1, cpu_addr); cpu_addr 794 arch/powerpc/platforms/cell/iommu.c u64 cpu_addr, size, best_size, dev_addr = OF_BAD_ADDR; cpu_addr 834 arch/powerpc/platforms/cell/iommu.c cpu_addr = of_translate_dma_address(np, ranges + i + naddr); cpu_addr 837 arch/powerpc/platforms/cell/iommu.c if (cpu_addr == 0 && size > best_size) { cpu_addr 132 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_cpu_info *)hdr)->cpu_addr; cpu_addr 134 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_x_cpu_info *)hdr)->cpu_addr; cpu_addr 192 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_phys_cpu *)hdr)->cpu_addr; cpu_addr 194 arch/s390/hypfs/hypfs_diag.c return ((struct diag204_x_phys_cpu *)hdr)->cpu_addr; cpu_addr 231 arch/s390/include/asm/ccwdev.h void *cpu_addr, size_t size); cpu_addr 363 arch/s390/include/asm/cio.h extern void cio_dma_free(void *cpu_addr, size_t size); cpu_addr 368 arch/s390/include/asm/cio.h void cio_gp_dma_free(struct gen_pool *gp_dma, void *cpu_addr, size_t size); cpu_addr 158 arch/s390/include/asm/diag.h __u16 cpu_addr; cpu_addr 168 arch/s390/include/asm/diag.h __u16 cpu_addr; cpu_addr 204 arch/s390/include/asm/diag.h __u16 cpu_addr; cpu_addr 213 arch/s390/include/asm/diag.h __u16 cpu_addr; cpu_addr 738 arch/s390/kvm/interrupt.c int cpu_addr; cpu_addr 741 arch/s390/kvm/interrupt.c cpu_addr = find_first_bit(li->sigp_emerg_pending, KVM_MAX_VCPUS); cpu_addr 742 arch/s390/kvm/interrupt.c clear_bit(cpu_addr, li->sigp_emerg_pending); cpu_addr 750 arch/s390/kvm/interrupt.c cpu_addr, 0); cpu_addr 754 arch/s390/kvm/interrupt.c rc |= put_guest_lc(vcpu, cpu_addr, (u16 *)__LC_EXT_CPU_ADDR); cpu_addr 283 arch/s390/kvm/sigp.c u16 cpu_addr, u32 parameter, u64 *status_reg) cpu_addr 286 arch/s390/kvm/sigp.c struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); cpu_addr 360 arch/s390/kvm/sigp.c u16 cpu_addr) cpu_addr 404 arch/s390/kvm/sigp.c order_code, cpu_addr); cpu_addr 414 arch/s390/kvm/sigp.c u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; cpu_addr 423 arch/s390/kvm/sigp.c if (handle_sigp_order_in_user_space(vcpu, order_code, cpu_addr)) cpu_addr 431 arch/s390/kvm/sigp.c trace_kvm_s390_handle_sigp(vcpu, order_code, cpu_addr, parameter); cpu_addr 439 arch/s390/kvm/sigp.c rc = handle_sigp_dst(vcpu, order_code, cpu_addr, cpu_addr 463 arch/s390/kvm/sigp.c u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; cpu_addr 467 arch/s390/kvm/sigp.c trace_kvm_s390_handle_sigp_pei(vcpu, order_code, cpu_addr); cpu_addr 470 arch/s390/kvm/sigp.c dest_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); cpu_addr 220 arch/s390/kvm/trace.h TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr, \ cpu_addr 222 arch/s390/kvm/trace.h TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr, parameter), cpu_addr 227 arch/s390/kvm/trace.h __field(__u16, cpu_addr) cpu_addr 234 arch/s390/kvm/trace.h __entry->cpu_addr = cpu_addr; cpu_addr 242 arch/s390/kvm/trace.h __entry->cpu_addr, __entry->parameter) cpu_addr 246 arch/s390/kvm/trace.h TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr), cpu_addr 247 arch/s390/kvm/trace.h TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr), cpu_addr 252 arch/s390/kvm/trace.h __field(__u16, cpu_addr) cpu_addr 258 arch/s390/kvm/trace.h __entry->cpu_addr = cpu_addr; cpu_addr 265 arch/s390/kvm/trace.h __entry->cpu_addr) cpu_addr 97 arch/sparc/include/asm/dma.h bool sparc_dma_free_resource(void *cpu_addr, size_t size); cpu_addr 267 arch/sparc/kernel/ioport.c bool sparc_dma_free_resource(void *cpu_addr, size_t size) cpu_addr 269 arch/sparc/kernel/ioport.c unsigned long addr = (unsigned long)cpu_addr; cpu_addr 274 arch/sparc/kernel/ioport.c printk("%s: cannot free %p\n", __func__, cpu_addr); cpu_addr 279 arch/sparc/kernel/ioport.c printk("%s: unaligned va %p\n", __func__, cpu_addr); cpu_addr 356 arch/sparc/kernel/ioport.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 359 arch/sparc/kernel/ioport.c if (!sparc_dma_free_resource(cpu_addr, PAGE_ALIGN(size))) cpu_addr 363 arch/sparc/kernel/ioport.c srmmu_unmapiorange((unsigned long)cpu_addr, size); cpu_addr 270 arch/sparc/mm/io-unit.c static void iounit_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 391 arch/sparc/mm/iommu.c static void sbus_iommu_free(struct device *dev, size_t len, void *cpu_addr, cpu_addr 396 arch/sparc/mm/iommu.c struct page *page = virt_to_page(cpu_addr); cpu_addr 400 arch/sparc/mm/iommu.c if (!sparc_dma_free_resource(cpu_addr, len)) cpu_addr 18 arch/x86/include/asm/xen/page-coherent.h void *cpu_addr, dma_addr_t dma_handle, cpu_addr 21 arch/x86/include/asm/xen/page-coherent.h free_pages((unsigned long) cpu_addr, get_order(size)); cpu_addr 743 drivers/acpi/pci_root.c resource_size_t cpu_addr = res->start; cpu_addr 744 drivers/acpi/pci_root.c resource_size_t pci_addr = cpu_addr - entry->offset; cpu_addr 748 drivers/acpi/pci_root.c if (pci_register_io_range(fwnode, cpu_addr, length)) cpu_addr 751 drivers/acpi/pci_root.c port = pci_address_to_pio(cpu_addr); cpu_addr 759 drivers/acpi/pci_root.c if (pci_remap_iospace(res, cpu_addr) < 0) cpu_addr 762 drivers/acpi/pci_root.c pr_info("Remapped I/O %pa to %pR\n", &cpu_addr, res); cpu_addr 123 drivers/crypto/img-hash.c void __iomem *cpu_addr; cpu_addr 203 drivers/crypto/img-hash.c writel_relaxed(buffer[count], hdev->cpu_addr); cpu_addr 971 drivers/crypto/img-hash.c hdev->cpu_addr = devm_ioremap_resource(dev, hash_res); cpu_addr 972 drivers/crypto/img-hash.c if (IS_ERR(hdev->cpu_addr)) { cpu_addr 974 drivers/crypto/img-hash.c err = PTR_ERR(hdev->cpu_addr); cpu_addr 185 drivers/dma/st_fdma.h readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ cpu_addr 190 drivers/dma/st_fdma.h writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ cpu_addr 197 drivers/dma/st_fdma.h writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ cpu_addr 207 drivers/dma/st_fdma.h readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ cpu_addr 212 drivers/dma/st_fdma.h writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ cpu_addr 1002 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c preempt_seq = le32_to_cpu(*(drv->cpu_addr + 2)); cpu_addr 102 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c if (drv->cpu_addr) cpu_addr 103 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c *drv->cpu_addr = cpu_to_le32(seq); cpu_addr 119 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c if (drv->cpu_addr) cpu_addr 120 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c seq = le32_to_cpu(*drv->cpu_addr); cpu_addr 398 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; cpu_addr 403 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; cpu_addr 415 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.gpu_addr, ring->fence_drv.cpu_addr); cpu_addr 443 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.cpu_addr = NULL; cpu_addr 730 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c le32_to_cpu(*(ring->fence_drv.cpu_addr + 2))); cpu_addr 733 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c le32_to_cpu(*(ring->fence_drv.cpu_addr + 4))); cpu_addr 736 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c le32_to_cpu(*(ring->fence_drv.cpu_addr + 6))); cpu_addr 234 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c u64 *gpu_addr, void **cpu_addr) cpu_addr 249 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bp.flags = cpu_addr ? AMDGPU_GEM_CREATE_CPU_ACCESS_REQUIRED cpu_addr 286 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (cpu_addr) { cpu_addr 287 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = amdgpu_bo_kmap(*bo_ptr, cpu_addr); cpu_addr 329 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c u64 *gpu_addr, void **cpu_addr) cpu_addr 334 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c gpu_addr, cpu_addr); cpu_addr 362 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_bo **bo_ptr, void **cpu_addr) cpu_addr 391 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (cpu_addr) { cpu_addr 392 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = amdgpu_bo_kmap(*bo_ptr, cpu_addr); cpu_addr 416 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c void **cpu_addr) cpu_addr 422 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (cpu_addr) cpu_addr 433 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (cpu_addr) cpu_addr 434 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c *cpu_addr = NULL; cpu_addr 235 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h u64 *gpu_addr, void **cpu_addr); cpu_addr 239 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h u64 *gpu_addr, void **cpu_addr); cpu_addr 242 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct amdgpu_bo **bo_ptr, void **cpu_addr); cpu_addr 244 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h void **cpu_addr); cpu_addr 73 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h volatile uint32_t *cpu_addr; cpu_addr 260 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c &adev->uvd.inst[j].gpu_addr, &adev->uvd.inst[j].cpu_addr); cpu_addr 309 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c (void **)&adev->uvd.inst[j].cpu_addr); cpu_addr 369 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ptr = adev->uvd.inst[j].cpu_addr; cpu_addr 393 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ptr = adev->uvd.inst[i].cpu_addr; cpu_addr 406 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c memcpy_toio(adev->uvd.inst[i].cpu_addr, adev->uvd.fw->data + offset, cpu_addr 42 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h void *cpu_addr; cpu_addr 183 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c &adev->vce.gpu_addr, &adev->vce.cpu_addr); cpu_addr 217 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c (void **)&adev->vce.cpu_addr); cpu_addr 285 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c void *cpu_addr; cpu_addr 299 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_bo_kmap(adev->vce.vcpu_bo, &cpu_addr); cpu_addr 308 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c memcpy_toio(cpu_addr, adev->vce.fw->data + offset, cpu_addr 38 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h void *cpu_addr; cpu_addr 172 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c &adev->vcn.inst[i].gpu_addr, &adev->vcn.inst[i].cpu_addr); cpu_addr 209 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c (void **)&adev->vcn.inst[j].cpu_addr); cpu_addr 239 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c ptr = adev->vcn.inst[i].cpu_addr; cpu_addr 263 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c ptr = adev->vcn.inst[i].cpu_addr; cpu_addr 276 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c memcpy_toio(adev->vcn.inst[i].cpu_addr, adev->vcn.fw->data + offset, cpu_addr 166 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h void *cpu_addr; cpu_addr 280 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c (void *)&adev->virt.mm_table.cpu_addr); cpu_addr 286 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c memset((void *)adev->virt.mm_table.cpu_addr, 0, PAGE_SIZE); cpu_addr 289 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c adev->virt.mm_table.cpu_addr); cpu_addr 305 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c (void *)&adev->virt.mm_table.cpu_addr); cpu_addr 35 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h uint32_t *cpu_addr; cpu_addr 713 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr; cpu_addr 773 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c uint32_t *init_table = adev->virt.mm_table.cpu_addr; cpu_addr 157 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr; cpu_addr 213 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c uint32_t *init_table = adev->virt.mm_table.cpu_addr; cpu_addr 568 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c void *ptr = adev->vce.cpu_addr; cpu_addr 590 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c void *ptr = adev->vce.cpu_addr; cpu_addr 500 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c &compressor->gpu_addr, &compressor->cpu_addr); cpu_addr 85 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.h void *cpu_addr; cpu_addr 448 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c memcpy(table->cpu_addr, table_data, table->size); cpu_addr 469 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c memcpy(table_data, table->cpu_addr, table->size); cpu_addr 939 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c &tables[i].cpu_addr); cpu_addr 951 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c &tables[i].cpu_addr); cpu_addr 972 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c &tables[i].cpu_addr); cpu_addr 1203 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c &memory_pool->cpu_addr); cpu_addr 1223 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c &memory_pool->cpu_addr); cpu_addr 182 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h void *cpu_addr; cpu_addr 684 drivers/gpu/drm/amd/powerplay/smu_v11_0.c if (memory_pool->size == 0 || memory_pool->cpu_addr == NULL) cpu_addr 687 drivers/gpu/drm/amd/powerplay/smu_v11_0.c address = (uintptr_t)memory_pool->cpu_addr; cpu_addr 782 drivers/gpu/drm/amd/powerplay/smu_v11_0.c if (!table->cpu_addr) cpu_addr 785 drivers/gpu/drm/amd/powerplay/smu_v11_0.c ret = smu_update_table(smu, SMU_TABLE_WATERMARKS, 0, table->cpu_addr, cpu_addr 1342 drivers/gpu/drm/amd/powerplay/smu_v11_0.c void *table = watermarks->cpu_addr; cpu_addr 316 drivers/gpu/drm/amd/powerplay/smu_v12_0.c if (!table->cpu_addr) cpu_addr 363 drivers/gpu/drm/radeon/radeon.h volatile uint32_t *cpu_addr; cpu_addr 1671 drivers/gpu/drm/radeon/radeon.h void *cpu_addr; cpu_addr 72 drivers/gpu/drm/radeon/radeon_fence.c if (drv->cpu_addr) { cpu_addr 73 drivers/gpu/drm/radeon/radeon_fence.c *drv->cpu_addr = cpu_to_le32(seq); cpu_addr 95 drivers/gpu/drm/radeon/radeon_fence.c if (drv->cpu_addr) { cpu_addr 96 drivers/gpu/drm/radeon/radeon_fence.c seq = le32_to_cpu(*drv->cpu_addr); cpu_addr 843 drivers/gpu/drm/radeon/radeon_fence.c rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; cpu_addr 850 drivers/gpu/drm/radeon/radeon_fence.c rdev->fence_drv[ring].cpu_addr = rdev->uvd.cpu_addr + index; cpu_addr 863 drivers/gpu/drm/radeon/radeon_fence.c rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; cpu_addr 869 drivers/gpu/drm/radeon/radeon_fence.c ring, rdev->fence_drv[ring].gpu_addr, rdev->fence_drv[ring].cpu_addr); cpu_addr 888 drivers/gpu/drm/radeon/radeon_fence.c rdev->fence_drv[ring].cpu_addr = NULL; cpu_addr 211 drivers/gpu/drm/radeon/radeon_uvd.c r = radeon_bo_kmap(rdev->uvd.vcpu_bo, &rdev->uvd.cpu_addr); cpu_addr 289 drivers/gpu/drm/radeon/radeon_uvd.c memcpy(rdev->uvd.cpu_addr, rdev->uvd_fw->data, rdev->uvd_fw->size); cpu_addr 294 drivers/gpu/drm/radeon/radeon_uvd.c ptr = rdev->uvd.cpu_addr; cpu_addr 784 drivers/gpu/drm/radeon/radeon_uvd.c uint32_t *msg = rdev->uvd.cpu_addr + offs; cpu_addr 820 drivers/gpu/drm/radeon/radeon_uvd.c uint32_t *msg = rdev->uvd.cpu_addr + offs; cpu_addr 223 drivers/gpu/drm/radeon/radeon_vce.c void *cpu_addr; cpu_addr 235 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_bo_kmap(rdev->vce.vcpu_bo, &cpu_addr); cpu_addr 242 drivers/gpu/drm/radeon/radeon_vce.c memset(cpu_addr, 0, radeon_bo_size(rdev->vce.vcpu_bo)); cpu_addr 244 drivers/gpu/drm/radeon/radeon_vce.c r = vce_v1_0_load_fw(rdev, cpu_addr); cpu_addr 246 drivers/gpu/drm/radeon/radeon_vce.c memcpy(cpu_addr, rdev->vce_fw->data, rdev->vce_fw->size); cpu_addr 145 drivers/gpu/drm/radeon/uvd_v1_0.c WREG32(UVD_FW_START, *((uint32_t*)rdev->uvd.cpu_addr)); cpu_addr 93 drivers/infiniband/hw/efa/efa.h void *cpu_addr; cpu_addr 895 drivers/infiniband/hw/efa/efa_verbs.c cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr); cpu_addr 907 drivers/infiniband/hw/efa/efa_verbs.c virt_to_phys(cq->cpu_addr), cpu_addr 988 drivers/infiniband/hw/efa/efa_verbs.c cq->cpu_addr = efa_zalloc_mapped(dev, &cq->dma_addr, cq->size, cpu_addr 990 drivers/infiniband/hw/efa/efa_verbs.c if (!cq->cpu_addr) { cpu_addr 1029 drivers/infiniband/hw/efa/efa_verbs.c cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr); cpu_addr 1039 drivers/infiniband/hw/efa/efa_verbs.c free_pages_exact(cq->cpu_addr, cq->size); cpu_addr 1186 drivers/infiniband/hw/hns/hns_roce_hem.c struct roce_hem_item *hem, void *cpu_addr, cpu_addr 1189 drivers/infiniband/hw/hns/hns_roce_hem.c hem->addr = cpu_addr; cpu_addr 922 drivers/iommu/dma-iommu.c static void __iommu_dma_free(struct device *dev, size_t size, void *cpu_addr) cpu_addr 930 drivers/iommu/dma-iommu.c dma_free_from_pool(cpu_addr, alloc_size)) cpu_addr 933 drivers/iommu/dma-iommu.c if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) { cpu_addr 938 drivers/iommu/dma-iommu.c pages = dma_common_find_pages(cpu_addr); cpu_addr 940 drivers/iommu/dma-iommu.c page = vmalloc_to_page(cpu_addr); cpu_addr 941 drivers/iommu/dma-iommu.c dma_common_free_remap(cpu_addr, alloc_size); cpu_addr 944 drivers/iommu/dma-iommu.c page = virt_to_page(cpu_addr); cpu_addr 953 drivers/iommu/dma-iommu.c static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 957 drivers/iommu/dma-iommu.c __iommu_dma_free(dev, size, cpu_addr); cpu_addr 967 drivers/iommu/dma-iommu.c void *cpu_addr; cpu_addr 978 drivers/iommu/dma-iommu.c cpu_addr = dma_common_contiguous_remap(page, alloc_size, cpu_addr 980 drivers/iommu/dma-iommu.c if (!cpu_addr) cpu_addr 986 drivers/iommu/dma-iommu.c cpu_addr = page_address(page); cpu_addr 990 drivers/iommu/dma-iommu.c memset(cpu_addr, 0, alloc_size); cpu_addr 991 drivers/iommu/dma-iommu.c return cpu_addr; cpu_addr 1003 drivers/iommu/dma-iommu.c void *cpu_addr; cpu_addr 1013 drivers/iommu/dma-iommu.c cpu_addr = dma_alloc_from_pool(PAGE_ALIGN(size), &page, gfp); cpu_addr 1015 drivers/iommu/dma-iommu.c cpu_addr = iommu_dma_alloc_pages(dev, size, &page, gfp, attrs); cpu_addr 1016 drivers/iommu/dma-iommu.c if (!cpu_addr) cpu_addr 1021 drivers/iommu/dma-iommu.c __iommu_dma_free(dev, size, cpu_addr); cpu_addr 1025 drivers/iommu/dma-iommu.c return cpu_addr; cpu_addr 1029 drivers/iommu/dma-iommu.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 1038 drivers/iommu/dma-iommu.c if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) cpu_addr 1044 drivers/iommu/dma-iommu.c if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) { cpu_addr 1045 drivers/iommu/dma-iommu.c struct page **pages = dma_common_find_pages(cpu_addr); cpu_addr 1049 drivers/iommu/dma-iommu.c pfn = vmalloc_to_pfn(cpu_addr); cpu_addr 1051 drivers/iommu/dma-iommu.c pfn = page_to_pfn(virt_to_page(cpu_addr)); cpu_addr 1060 drivers/iommu/dma-iommu.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 1066 drivers/iommu/dma-iommu.c if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) { cpu_addr 1067 drivers/iommu/dma-iommu.c struct page **pages = dma_common_find_pages(cpu_addr); cpu_addr 1075 drivers/iommu/dma-iommu.c page = vmalloc_to_page(cpu_addr); cpu_addr 1077 drivers/iommu/dma-iommu.c page = virt_to_page(cpu_addr); cpu_addr 649 drivers/iommu/qcom_iommu.c void *cpu_addr; cpu_addr 669 drivers/iommu/qcom_iommu.c cpu_addr = dma_alloc_attrs(dev, psize, &paddr, GFP_KERNEL, attrs); cpu_addr 670 drivers/iommu/qcom_iommu.c if (!cpu_addr) { cpu_addr 686 drivers/iommu/qcom_iommu.c dma_free_attrs(dev, psize, cpu_addr, paddr, attrs); cpu_addr 415 drivers/media/common/saa7146/saa7146_core.c dev->d_rps0.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, cpu_addr 417 drivers/media/common/saa7146/saa7146_core.c if (!dev->d_rps0.cpu_addr) cpu_addr 420 drivers/media/common/saa7146/saa7146_core.c dev->d_rps1.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, cpu_addr 422 drivers/media/common/saa7146/saa7146_core.c if (!dev->d_rps1.cpu_addr) cpu_addr 425 drivers/media/common/saa7146/saa7146_core.c dev->d_i2c.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, cpu_addr 427 drivers/media/common/saa7146/saa7146_core.c if (!dev->d_i2c.cpu_addr) cpu_addr 474 drivers/media/common/saa7146/saa7146_core.c pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_i2c.cpu_addr, cpu_addr 477 drivers/media/common/saa7146/saa7146_core.c pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps1.cpu_addr, cpu_addr 480 drivers/media/common/saa7146/saa7146_core.c pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps0.cpu_addr, cpu_addr 503 drivers/media/common/saa7146/saa7146_core.c { dev->d_i2c.cpu_addr, dev->d_i2c.dma_handle }, cpu_addr 504 drivers/media/common/saa7146/saa7146_core.c { dev->d_rps1.cpu_addr, dev->d_rps1.dma_handle }, cpu_addr 505 drivers/media/common/saa7146/saa7146_core.c { dev->d_rps0.cpu_addr, dev->d_rps0.dma_handle }, cpu_addr 521 drivers/media/common/saa7146/saa7146_fops.c vv->d_clipping.cpu_addr = cpu_addr 524 drivers/media/common/saa7146/saa7146_fops.c if( NULL == vv->d_clipping.cpu_addr ) { cpu_addr 581 drivers/media/common/saa7146/saa7146_fops.c pci_free_consistent(dev->pci, SAA7146_CLIPPING_MEM, vv->d_clipping.cpu_addr, vv->d_clipping.dma_handle); cpu_addr 342 drivers/media/common/saa7146/saa7146_hlp.c __le32 *clipping = vv->d_clipping.cpu_addr; cpu_addr 294 drivers/media/common/saa7146/saa7146_i2c.c __le32 *buffer = dev->d_i2c.cpu_addr; cpu_addr 314 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[0], (finishedData / 188)); cpu_addr 337 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[1], (finishedData / 188)); cpu_addr 353 drivers/media/pci/smipcie/smipcie-main.c if (port->cpu_addr[0]) { cpu_addr 355 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[0], port->dma_addr[0]); cpu_addr 356 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[0] = NULL; cpu_addr 358 drivers/media/pci/smipcie/smipcie-main.c if (port->cpu_addr[1]) { cpu_addr 360 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[1], port->dma_addr[1]); cpu_addr 361 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[1] = NULL; cpu_addr 401 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[0] = pci_alloc_consistent(port->dev->pci_dev, cpu_addr 404 drivers/media/pci/smipcie/smipcie-main.c if (!port->cpu_addr[0]) { cpu_addr 413 drivers/media/pci/smipcie/smipcie-main.c port->cpu_addr[1] = pci_alloc_consistent(port->dev->pci_dev, cpu_addr 416 drivers/media/pci/smipcie/smipcie-main.c if (!port->cpu_addr[1]) { cpu_addr 257 drivers/media/pci/smipcie/smipcie.h u8 *cpu_addr[2]; cpu_addr 2815 drivers/misc/habanalabs/goya/goya.c void *cpu_addr, dma_addr_t dma_handle) cpu_addr 2820 drivers/misc/habanalabs/goya/goya.c dma_free_coherent(&hdev->pdev->dev, size, cpu_addr, fixed_dma_handle); cpu_addr 526 drivers/misc/habanalabs/habanalabs.h void *cpu_addr, dma_addr_t dma_handle); cpu_addr 87 drivers/net/caif/caif_spi.c static inline void dma_free(struct cfspi *cfspi, void *cpu_addr, cpu_addr 90 drivers/net/caif/caif_spi.c kfree(cpu_addr); cpu_addr 101 drivers/net/caif/caif_spi.c static inline void dma_free(struct cfspi *cfspi, void *cpu_addr, cpu_addr 104 drivers/net/caif/caif_spi.c dma_free_coherent(&cfspi->pdev->dev, SPI_DMA_BUF_LEN, cpu_addr, handle); cpu_addr 538 drivers/net/ethernet/cavium/thunder/nicvf_main.c u64 dma_addr, cpu_addr; cpu_addr 545 drivers/net/ethernet/cavium/thunder/nicvf_main.c cpu_addr = nicvf_iova_to_phys(nic, dma_addr); cpu_addr 546 drivers/net/ethernet/cavium/thunder/nicvf_main.c if (!cpu_addr) cpu_addr 548 drivers/net/ethernet/cavium/thunder/nicvf_main.c cpu_addr = (u64)phys_to_virt(cpu_addr); cpu_addr 549 drivers/net/ethernet/cavium/thunder/nicvf_main.c page = virt_to_page((void *)cpu_addr); cpu_addr 552 drivers/net/ethernet/cavium/thunder/nicvf_main.c xdp.data = (void *)cpu_addr; cpu_addr 9576 drivers/net/ethernet/sun/niu.c void *cpu_addr, u64 handle) cpu_addr 9578 drivers/net/ethernet/sun/niu.c dma_free_coherent(dev, size, cpu_addr, handle); cpu_addr 9594 drivers/net/ethernet/sun/niu.c static u64 niu_pci_map_single(struct device *dev, void *cpu_addr, cpu_addr 9598 drivers/net/ethernet/sun/niu.c return dma_map_single(dev, cpu_addr, size, direction); cpu_addr 9958 drivers/net/ethernet/sun/niu.c void *cpu_addr, u64 handle) cpu_addr 9962 drivers/net/ethernet/sun/niu.c free_pages((unsigned long) cpu_addr, order); cpu_addr 9978 drivers/net/ethernet/sun/niu.c static u64 niu_phys_map_single(struct device *dev, void *cpu_addr, cpu_addr 9982 drivers/net/ethernet/sun/niu.c return __pa(cpu_addr); cpu_addr 3132 drivers/net/ethernet/sun/niu.h void *cpu_addr, u64 handle); cpu_addr 3138 drivers/net/ethernet/sun/niu.h u64 (*map_single)(struct device *dev, void *cpu_addr, cpu_addr 177 drivers/net/wireless/ath/wcn36xx/dxe.c wcn_ch->cpu_addr = dma_alloc_coherent(dev, size, &wcn_ch->dma_addr, cpu_addr 179 drivers/net/wireless/ath/wcn36xx/dxe.c if (!wcn_ch->cpu_addr) cpu_addr 182 drivers/net/wireless/ath/wcn36xx/dxe.c cur_dxe = (struct wcn36xx_dxe_desc *)wcn_ch->cpu_addr; cpu_addr 232 drivers/net/wireless/ath/wcn36xx/dxe.c dma_free_coherent(dev, size,wcn_ch->cpu_addr, wcn_ch->dma_addr); cpu_addr 620 drivers/net/wireless/ath/wcn36xx/dxe.c void *cpu_addr; cpu_addr 629 drivers/net/wireless/ath/wcn36xx/dxe.c cpu_addr = dma_alloc_coherent(wcn->dev, s, cpu_addr 632 drivers/net/wireless/ath/wcn36xx/dxe.c if (!cpu_addr) cpu_addr 635 drivers/net/wireless/ath/wcn36xx/dxe.c wcn->mgmt_mem_pool.virt_addr = cpu_addr; cpu_addr 644 drivers/net/wireless/ath/wcn36xx/dxe.c cpu_addr = dma_alloc_coherent(wcn->dev, s, cpu_addr 647 drivers/net/wireless/ath/wcn36xx/dxe.c if (!cpu_addr) cpu_addr 650 drivers/net/wireless/ath/wcn36xx/dxe.c wcn->data_mem_pool.virt_addr = cpu_addr; cpu_addr 432 drivers/net/wireless/ath/wcn36xx/dxe.h void *cpu_addr; cpu_addr 210 drivers/net/wireless/intel/iwlwifi/pcie/trans.c void *cpu_addr = NULL; cpu_addr 217 drivers/net/wireless/intel/iwlwifi/pcie/trans.c cpu_addr = dma_alloc_coherent(trans->dev, size, &phys, cpu_addr 219 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (!cpu_addr) cpu_addr 228 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (WARN_ON_ONCE(!cpu_addr)) cpu_addr 237 drivers/net/wireless/intel/iwlwifi/pcie/trans.c trans->dbg.fw_mon[trans->dbg.num_blocks].block = cpu_addr; cpu_addr 2835 drivers/net/wireless/intel/iwlwifi/pcie/trans.c void *cpu_addr = (void *)trans->dbg.fw_mon[0].block, *curr_buf; cpu_addr 2866 drivers/net/wireless/intel/iwlwifi/pcie/trans.c curr_buf = cpu_addr + data->prev_wr_ptr; cpu_addr 2875 drivers/net/wireless/intel/iwlwifi/pcie/trans.c curr_buf = cpu_addr + data->prev_wr_ptr; cpu_addr 2884 drivers/net/wireless/intel/iwlwifi/pcie/trans.c cpu_addr, &size, cpu_addr 2901 drivers/net/wireless/intel/iwlwifi/pcie/trans.c cpu_addr, &size, cpu_addr 282 drivers/of/address.c range->cpu_addr = of_translate_address(parser->node, cpu_addr 291 drivers/of/address.c u64 pci_addr, cpu_addr, size; cpu_addr 295 drivers/of/address.c cpu_addr = of_translate_address(parser->node, cpu_addr 302 drivers/of/address.c cpu_addr != range->cpu_addr + range->size) cpu_addr 338 drivers/of/address.c err = pci_register_io_range(&np->fwnode, range->cpu_addr, cpu_addr 342 drivers/of/address.c port = pci_address_to_pio(range->cpu_addr); cpu_addr 350 drivers/of/address.c upper_32_bits(range->cpu_addr)) { cpu_addr 355 drivers/of/address.c res->start = range->cpu_addr; cpu_addr 879 drivers/parisc/ccio-dma.c ccio_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 883 drivers/parisc/ccio-dma.c free_pages((unsigned long)cpu_addr, get_order(size)); cpu_addr 69 drivers/pci/controller/dwc/pcie-designware-ep.c dma_addr_t cpu_addr, cpu_addr 82 drivers/pci/controller/dwc/pcie-designware-ep.c ret = dw_pcie_prog_inbound_atu(pci, free_win, bar, cpu_addr, cpu_addr 533 drivers/pci/controller/dwc/pcie-designware-host.c u64 cpu_addr; cpu_addr 542 drivers/pci/controller/dwc/pcie-designware-host.c cpu_addr = pp->cfg0_base; cpu_addr 547 drivers/pci/controller/dwc/pcie-designware-host.c cpu_addr = pp->cfg1_base; cpu_addr 553 drivers/pci/controller/dwc/pcie-designware-host.c type, cpu_addr, cpu_addr 242 drivers/pci/controller/dwc/pcie-designware.c int type, u64 cpu_addr, cpu_addr 248 drivers/pci/controller/dwc/pcie-designware.c lower_32_bits(cpu_addr)); cpu_addr 250 drivers/pci/controller/dwc/pcie-designware.c upper_32_bits(cpu_addr)); cpu_addr 252 drivers/pci/controller/dwc/pcie-designware.c lower_32_bits(cpu_addr + size - 1)); cpu_addr 278 drivers/pci/controller/dwc/pcie-designware.c u64 cpu_addr, u64 pci_addr, u32 size) cpu_addr 283 drivers/pci/controller/dwc/pcie-designware.c cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr); cpu_addr 286 drivers/pci/controller/dwc/pcie-designware.c dw_pcie_prog_outbound_atu_unroll(pci, index, type, cpu_addr, cpu_addr 294 drivers/pci/controller/dwc/pcie-designware.c lower_32_bits(cpu_addr)); cpu_addr 296 drivers/pci/controller/dwc/pcie-designware.c upper_32_bits(cpu_addr)); cpu_addr 298 drivers/pci/controller/dwc/pcie-designware.c lower_32_bits(cpu_addr + size - 1)); cpu_addr 336 drivers/pci/controller/dwc/pcie-designware.c int bar, u64 cpu_addr, cpu_addr 343 drivers/pci/controller/dwc/pcie-designware.c lower_32_bits(cpu_addr)); cpu_addr 345 drivers/pci/controller/dwc/pcie-designware.c upper_32_bits(cpu_addr)); cpu_addr 381 drivers/pci/controller/dwc/pcie-designware.c u64 cpu_addr, enum dw_pcie_as_type as_type) cpu_addr 388 drivers/pci/controller/dwc/pcie-designware.c cpu_addr, as_type); cpu_addr 392 drivers/pci/controller/dwc/pcie-designware.c dw_pcie_writel_dbi(pci, PCIE_ATU_LOWER_TARGET, lower_32_bits(cpu_addr)); cpu_addr 393 drivers/pci/controller/dwc/pcie-designware.c dw_pcie_writel_dbi(pci, PCIE_ATU_UPPER_TARGET, upper_32_bits(cpu_addr)); cpu_addr 230 drivers/pci/controller/dwc/pcie-designware.h u64 (*cpu_addr_fixup)(struct dw_pcie *pcie, u64 cpu_addr); cpu_addr 278 drivers/pci/controller/dwc/pcie-designware.h int type, u64 cpu_addr, u64 pci_addr, cpu_addr 281 drivers/pci/controller/dwc/pcie-designware.h u64 cpu_addr, enum dw_pcie_as_type as_type); cpu_addr 312 drivers/pci/controller/pci-rcar-gen2.c pci->window_addr = (unsigned long)range.cpu_addr; cpu_addr 2193 drivers/pci/controller/pci-tegra.c pcie->io.start = range.cpu_addr; cpu_addr 2194 drivers/pci/controller/pci-tegra.c pcie->io.end = range.cpu_addr + range.size - 1; cpu_addr 620 drivers/pci/controller/pci-v3-semi.c u64 cpu_end = range->cpu_addr + range->size - 1; cpu_addr 631 drivers/pci/controller/pci-v3-semi.c if (range->cpu_addr & ~V3_PCI_MAP_M_MAP_ADR) { cpu_addr 635 drivers/pci/controller/pci-v3-semi.c val = ((u32)range->cpu_addr) & V3_PCI_MAP_M_MAP_ADR; cpu_addr 685 drivers/pci/controller/pci-v3-semi.c range->cpu_addr, cpu_end, cpu_addr 369 drivers/pci/controller/pci-xgene.c u64 cpu_addr, u64 pci_addr) cpu_addr 391 drivers/pci/controller/pci-xgene.c xgene_pcie_writel(port, offset, lower_32_bits(cpu_addr)); cpu_addr 392 drivers/pci/controller/pci-xgene.c xgene_pcie_writel(port, offset + 0x04, upper_32_bits(cpu_addr)); cpu_addr 494 drivers/pci/controller/pci-xgene.c u64 cpu_addr = range->cpu_addr; cpu_addr 511 drivers/pci/controller/pci-xgene.c bar_low = pcie_bar_low_val((u32)cpu_addr, flags); cpu_addr 517 drivers/pci/controller/pci-xgene.c writel(upper_32_bits(cpu_addr), bar_addr + 0x4); cpu_addr 527 drivers/pci/controller/pci-xgene.c xgene_pcie_writel(port, IBAR3L + 0x4, upper_32_bits(cpu_addr)); cpu_addr 552 drivers/pci/controller/pci-xgene.c u64 end = range.cpu_addr + range.size - 1; cpu_addr 555 drivers/pci/controller/pci-xgene.c range.flags, range.cpu_addr, end, range.pci_addr); cpu_addr 147 drivers/pci/controller/pcie-cadence-host.c u64 cpu_addr; cpu_addr 160 drivers/pci/controller/pcie-cadence-host.c cpu_addr = cfg_res->start - mem_res->start; cpu_addr 162 drivers/pci/controller/pcie-cadence-host.c (lower_32_bits(cpu_addr) & GENMASK(31, 8)); cpu_addr 163 drivers/pci/controller/pcie-cadence-host.c addr1 = upper_32_bits(cpu_addr); cpu_addr 186 drivers/pci/controller/pcie-cadence-host.c range.cpu_addr, cpu_addr 12 drivers/pci/controller/pcie-cadence.c u64 cpu_addr, u64 pci_addr, size_t size) cpu_addr 76 drivers/pci/controller/pcie-cadence.c cpu_addr -= pcie->mem_res->start; cpu_addr 78 drivers/pci/controller/pcie-cadence.c (lower_32_bits(cpu_addr) & GENMASK(31, 8)); cpu_addr 79 drivers/pci/controller/pcie-cadence.c addr1 = upper_32_bits(cpu_addr); cpu_addr 86 drivers/pci/controller/pcie-cadence.c u32 r, u64 cpu_addr) cpu_addr 103 drivers/pci/controller/pcie-cadence.c cpu_addr -= pcie->mem_res->start; cpu_addr 105 drivers/pci/controller/pcie-cadence.c (lower_32_bits(cpu_addr) & GENMASK(31, 8)); cpu_addr 106 drivers/pci/controller/pcie-cadence.c addr1 = upper_32_bits(cpu_addr); cpu_addr 311 drivers/pci/controller/pcie-cadence.h u64 cpu_addr, u64 pci_addr, size_t size); cpu_addr 314 drivers/pci/controller/pcie-cadence.h u32 r, u64 cpu_addr); cpu_addr 1132 drivers/pci/controller/pcie-iproc.c u64 axi_addr = range->cpu_addr, pci_addr = range->pci_addr; cpu_addr 1198 drivers/pci/controller/pcie-iproc.c if (tmp->res->start < range->cpu_addr) cpu_addr 1202 drivers/pci/controller/pcie-iproc.c res->start = range->cpu_addr; cpu_addr 1209 drivers/pci/controller/pcie-iproc.c entry->offset = res->start - range->cpu_addr; cpu_addr 1283 drivers/pci/controller/pcie-iproc.c range.pci_addr = range.cpu_addr = msi_addr & ~(range.size - 1); cpu_addr 469 drivers/pci/controller/pcie-mobiveil.c u64 cpu_addr, u64 pci_addr, u32 type, u64 size) cpu_addr 489 drivers/pci/controller/pcie-mobiveil.c mobiveil_csr_writel(pcie, lower_32_bits(cpu_addr), cpu_addr 491 drivers/pci/controller/pcie-mobiveil.c mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), cpu_addr 506 drivers/pci/controller/pcie-mobiveil.c u64 cpu_addr, u64 pci_addr, u32 type, u64 size) cpu_addr 535 drivers/pci/controller/pcie-mobiveil.c lower_32_bits(cpu_addr) & (~AXI_WINDOW_ALIGN_MASK), cpu_addr 537 drivers/pci/controller/pcie-mobiveil.c mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), cpu_addr 1026 drivers/pci/controller/pcie-rcar.c u64 cpu_addr = range->cpu_addr; cpu_addr 1027 drivers/pci/controller/pcie-rcar.c u64 cpu_end = range->cpu_addr + range->size; cpu_addr 1041 drivers/pci/controller/pcie-rcar.c if (cpu_addr > 0) { cpu_addr 1042 drivers/pci/controller/pcie-rcar.c unsigned long nr_zeros = __ffs64(cpu_addr); cpu_addr 1055 drivers/pci/controller/pcie-rcar.c while (cpu_addr < cpu_end) { cpu_addr 1062 drivers/pci/controller/pcie-rcar.c rcar_pci_write_reg(pcie, lower_32_bits(cpu_addr), PCIELAR(idx)); cpu_addr 1068 drivers/pci/controller/pcie-rcar.c rcar_pci_write_reg(pcie, upper_32_bits(cpu_addr), cpu_addr 1073 drivers/pci/controller/pcie-rcar.c cpu_addr += size; cpu_addr 1099 drivers/pci/controller/pcie-rcar.c u64 end = range.cpu_addr + range.size - 1; cpu_addr 1102 drivers/pci/controller/pcie-rcar.c range.flags, range.cpu_addr, end, range.pci_addr); cpu_addr 70 drivers/pci/controller/pcie-rockchip-ep.c u32 r, u32 type, u64 cpu_addr, cpu_addr 82 drivers/pci/controller/pcie-rockchip-ep.c cpu_addr -= rockchip->mem_res->start; cpu_addr 85 drivers/pci/controller/pcie-rockchip-ep.c (lower_32_bits(cpu_addr) & PCIE_CORE_OB_REGION_ADDR0_LO_ADDR); cpu_addr 86 drivers/pci/controller/pcie-rockchip-ep.c addr1 = upper_32_bits(is_nor_msg ? cpu_addr : pci_addr); cpu_addr 112 drivers/pci/controller/pcie-rockchip-ep.c (lower_32_bits(cpu_addr) & cpu_addr 114 drivers/pci/controller/pcie-rockchip-ep.c addr1 = upper_32_bits(cpu_addr); cpu_addr 324 drivers/pci/controller/vmd.c void *cpu_addr, dma_addr_t addr, size_t size, cpu_addr 327 drivers/pci/controller/vmd.c return dma_mmap_attrs(to_vmd_dev(dev), vma, cpu_addr, addr, size, cpu_addr 332 drivers/pci/controller/vmd.c void *cpu_addr, dma_addr_t addr, size_t size, cpu_addr 335 drivers/pci/controller/vmd.c return dma_get_sgtable_attrs(to_vmd_dev(dev), sgt, cpu_addr, addr, size, cpu_addr 307 drivers/pci/of.c range_type, range.cpu_addr, cpu_addr 308 drivers/pci/of.c range.cpu_addr + range.size - 1, range.pci_addr); cpu_addr 314 drivers/pci/of.c if (range.cpu_addr == OF_BAD_ADDR || range.size == 0) cpu_addr 337 drivers/pci/of.c *io_base = range.cpu_addr; cpu_addr 48 drivers/remoteproc/da8xx_remoteproc.c void __iomem *cpu_addr; cpu_addr 215 drivers/remoteproc/da8xx_remoteproc.c drproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res); cpu_addr 216 drivers/remoteproc/da8xx_remoteproc.c if (IS_ERR(drproc->mem[i].cpu_addr)) { cpu_addr 219 drivers/remoteproc/da8xx_remoteproc.c return PTR_ERR(drproc->mem[i].cpu_addr); cpu_addr 228 drivers/remoteproc/da8xx_remoteproc.c drproc->mem[i].size, drproc->mem[i].cpu_addr, cpu_addr 55 drivers/remoteproc/imx_rproc.c void __iomem *cpu_addr; cpu_addr 233 drivers/remoteproc/imx_rproc.c va = (__force void *)(priv->mem[i].cpu_addr + offset); cpu_addr 267 drivers/remoteproc/imx_rproc.c priv->mem[b].cpu_addr = devm_ioremap(&pdev->dev, cpu_addr 269 drivers/remoteproc/imx_rproc.c if (!priv->mem[b].cpu_addr) { cpu_addr 298 drivers/remoteproc/imx_rproc.c priv->mem[b].cpu_addr = devm_ioremap_resource(&pdev->dev, &res); cpu_addr 299 drivers/remoteproc/imx_rproc.c if (IS_ERR(priv->mem[b].cpu_addr)) { cpu_addr 301 drivers/remoteproc/imx_rproc.c err = PTR_ERR(priv->mem[b].cpu_addr); cpu_addr 35 drivers/remoteproc/keystone_remoteproc.c void __iomem *cpu_addr; cpu_addr 271 drivers/remoteproc/keystone_remoteproc.c va = ksproc->mem[i].cpu_addr + offset; cpu_addr 279 drivers/remoteproc/keystone_remoteproc.c va = ksproc->mem[i].cpu_addr + offset; cpu_addr 313 drivers/remoteproc/keystone_remoteproc.c ksproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res); cpu_addr 314 drivers/remoteproc/keystone_remoteproc.c if (IS_ERR(ksproc->mem[i].cpu_addr)) { cpu_addr 317 drivers/remoteproc/keystone_remoteproc.c return PTR_ERR(ksproc->mem[i].cpu_addr); cpu_addr 325 drivers/remoteproc/keystone_remoteproc.c memset((__force void *)ksproc->mem[i].cpu_addr, 0, cpu_addr 144 drivers/remoteproc/remoteproc_core.c phys_addr_t rproc_va_to_pa(void *cpu_addr) cpu_addr 151 drivers/remoteproc/remoteproc_core.c if (is_vmalloc_addr(cpu_addr)) { cpu_addr 152 drivers/remoteproc/remoteproc_core.c return page_to_phys(vmalloc_to_page(cpu_addr)) + cpu_addr 153 drivers/remoteproc/remoteproc_core.c offset_in_page(cpu_addr); cpu_addr 156 drivers/remoteproc/remoteproc_core.c WARN_ON(!virt_addr_valid(cpu_addr)); cpu_addr 157 drivers/remoteproc/remoteproc_core.c return virt_to_phys(cpu_addr); cpu_addr 54 drivers/remoteproc/remoteproc_internal.h phys_addr_t rproc_va_to_pa(void *cpu_addr); cpu_addr 144 drivers/remoteproc/st_slim_rproc.c fw_rev = readl(slim_rproc->mem[ST_SLIM_DMEM].cpu_addr + cpu_addr 189 drivers/remoteproc/st_slim_rproc.c va = (__force void *)slim_rproc->mem[i].cpu_addr; cpu_addr 253 drivers/remoteproc/st_slim_rproc.c slim_rproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res); cpu_addr 254 drivers/remoteproc/st_slim_rproc.c if (IS_ERR(slim_rproc->mem[i].cpu_addr)) { cpu_addr 256 drivers/remoteproc/st_slim_rproc.c err = PTR_ERR(slim_rproc->mem[i].cpu_addr); cpu_addr 45 drivers/remoteproc/stm32_rproc.c void __iomem *cpu_addr; cpu_addr 35 drivers/remoteproc/wkup_m3_rproc.c void __iomem *cpu_addr; cpu_addr 98 drivers/remoteproc/wkup_m3_rproc.c va = (__force void *)(wkupm3->mem[i].cpu_addr + offset); cpu_addr 171 drivers/remoteproc/wkup_m3_rproc.c wkupm3->mem[i].cpu_addr = devm_ioremap_resource(dev, res); cpu_addr 172 drivers/remoteproc/wkup_m3_rproc.c if (IS_ERR(wkupm3->mem[i].cpu_addr)) { cpu_addr 175 drivers/remoteproc/wkup_m3_rproc.c ret = PTR_ERR(wkupm3->mem[i].cpu_addr); cpu_addr 199 drivers/rpmsg/virtio_rpmsg_bus.c rpmsg_sg_init(struct scatterlist *sg, void *cpu_addr, unsigned int len) cpu_addr 201 drivers/rpmsg/virtio_rpmsg_bus.c if (is_vmalloc_addr(cpu_addr)) { cpu_addr 203 drivers/rpmsg/virtio_rpmsg_bus.c sg_set_page(sg, vmalloc_to_page(cpu_addr), len, cpu_addr 204 drivers/rpmsg/virtio_rpmsg_bus.c offset_in_page(cpu_addr)); cpu_addr 206 drivers/rpmsg/virtio_rpmsg_bus.c WARN_ON(!virt_addr_valid(cpu_addr)); cpu_addr 207 drivers/rpmsg/virtio_rpmsg_bus.c sg_init_one(sg, cpu_addr, len); cpu_addr 935 drivers/rpmsg/virtio_rpmsg_bus.c void *cpu_addr = vrp->rbufs + i * vrp->buf_size; cpu_addr 937 drivers/rpmsg/virtio_rpmsg_bus.c rpmsg_sg_init(&sg, cpu_addr, vrp->buf_size); cpu_addr 939 drivers/rpmsg/virtio_rpmsg_bus.c err = virtqueue_add_inbuf(vrp->rvq, &sg, 1, cpu_addr, cpu_addr 854 drivers/s390/cio/css.c css->global_pgid.pgid_high.cpu_addr = stap(); cpu_addr 1102 drivers/s390/cio/css.c void *cpu_addr; cpu_addr 1110 drivers/s390/cio/css.c cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr, cpu_addr 1112 drivers/s390/cio/css.c if (!cpu_addr) cpu_addr 1114 drivers/s390/cio/css.c gen_pool_add_virt(gp_dma, (unsigned long) cpu_addr, cpu_addr 1170 drivers/s390/cio/css.c void cio_gp_dma_free(struct gen_pool *gp_dma, void *cpu_addr, size_t size) cpu_addr 1172 drivers/s390/cio/css.c if (!cpu_addr) cpu_addr 1174 drivers/s390/cio/css.c memset(cpu_addr, 0, size); cpu_addr 1175 drivers/s390/cio/css.c gen_pool_free(gp_dma, (unsigned long) cpu_addr, size); cpu_addr 1190 drivers/s390/cio/css.c void cio_dma_free(void *cpu_addr, size_t size) cpu_addr 1192 drivers/s390/cio/css.c cio_gp_dma_free(cio_dma_pool, cpu_addr, size); cpu_addr 55 drivers/s390/cio/css.h __u32 cpu_addr : 16; /* CPU address */ cpu_addr 724 drivers/s390/cio/device_ops.c void ccw_device_dma_free(struct ccw_device *cdev, void *cpu_addr, size_t size) cpu_addr 726 drivers/s390/cio/device_ops.c cio_gp_dma_free(cdev->private->dma_pool, cpu_addr, size); cpu_addr 220 drivers/s390/net/ism_drv.c dmb->cpu_addr, dmb->dma_addr); cpu_addr 242 drivers/s390/net/ism_drv.c dmb->cpu_addr = dma_alloc_coherent(&ism->pdev->dev, dmb->dmb_len, cpu_addr 245 drivers/s390/net/ism_drv.c if (!dmb->cpu_addr) cpu_addr 248 drivers/s390/net/ism_drv.c return dmb->cpu_addr ? 0 : -ENOMEM; cpu_addr 520 drivers/scsi/3w-9xxx.c unsigned long *cpu_addr; cpu_addr 523 drivers/scsi/3w-9xxx.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, cpu_addr 525 drivers/scsi/3w-9xxx.c if (!cpu_addr) { cpu_addr 530 drivers/scsi/3w-9xxx.c if ((unsigned long)cpu_addr % (TW_ALIGNMENT_9000)) { cpu_addr 533 drivers/scsi/3w-9xxx.c cpu_addr, dma_handle); cpu_addr 537 drivers/scsi/3w-9xxx.c memset(cpu_addr, 0, size*TW_Q_LENGTH); cpu_addr 543 drivers/scsi/3w-9xxx.c tw_dev->command_packet_virt[i] = (TW_Command_Full *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 547 drivers/scsi/3w-9xxx.c tw_dev->generic_buffer_virt[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 641 drivers/scsi/3w-9xxx.c unsigned long *cpu_addr, data_buffer_length_adjusted = 0, flags = 0; cpu_addr 679 drivers/scsi/3w-9xxx.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, &dma_handle, GFP_KERNEL); cpu_addr 680 drivers/scsi/3w-9xxx.c if (!cpu_addr) { cpu_addr 685 drivers/scsi/3w-9xxx.c tw_ioctl = (TW_Ioctl_Buf_Apache *)cpu_addr; cpu_addr 874 drivers/scsi/3w-9xxx.c dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, cpu_addr, dma_handle); cpu_addr 646 drivers/scsi/3w-sas.c unsigned long *cpu_addr; cpu_addr 649 drivers/scsi/3w-sas.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, cpu_addr 652 drivers/scsi/3w-sas.c if (!cpu_addr) { cpu_addr 661 drivers/scsi/3w-sas.c tw_dev->command_packet_virt[i] = (TW_Command_Full *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 665 drivers/scsi/3w-sas.c tw_dev->generic_buffer_virt[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 669 drivers/scsi/3w-sas.c tw_dev->sense_buffer_virt[i] = (TW_Command_Apache_Header *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 719 drivers/scsi/3w-sas.c unsigned long *cpu_addr, data_buffer_length_adjusted = 0, flags = 0; cpu_addr 752 drivers/scsi/3w-sas.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, &dma_handle, GFP_KERNEL); cpu_addr 753 drivers/scsi/3w-sas.c if (!cpu_addr) { cpu_addr 758 drivers/scsi/3w-sas.c tw_ioctl = (TW_Ioctl_Buf_Apache *)cpu_addr; cpu_addr 823 drivers/scsi/3w-sas.c dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, cpu_addr, dma_handle); cpu_addr 833 drivers/scsi/3w-xxxx.c unsigned long *cpu_addr = NULL; cpu_addr 837 drivers/scsi/3w-xxxx.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, cpu_addr 839 drivers/scsi/3w-xxxx.c if (cpu_addr == NULL) { cpu_addr 844 drivers/scsi/3w-xxxx.c if ((unsigned long)cpu_addr % (tw_dev->tw_pci_dev->device == TW_DEVICE_ID ? TW_ALIGNMENT_6000 : TW_ALIGNMENT_7000)) { cpu_addr 847 drivers/scsi/3w-xxxx.c cpu_addr, dma_handle); cpu_addr 851 drivers/scsi/3w-xxxx.c memset(cpu_addr, 0, size*TW_Q_LENGTH); cpu_addr 857 drivers/scsi/3w-xxxx.c tw_dev->command_packet_virtual_address[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 861 drivers/scsi/3w-xxxx.c tw_dev->alignment_virtual_address[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size)); cpu_addr 882 drivers/scsi/3w-xxxx.c unsigned long *cpu_addr; cpu_addr 913 drivers/scsi/3w-xxxx.c cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_New_Ioctl) - 1, &dma_handle, GFP_KERNEL); cpu_addr 914 drivers/scsi/3w-xxxx.c if (cpu_addr == NULL) { cpu_addr 919 drivers/scsi/3w-xxxx.c tw_ioctl = (TW_New_Ioctl *)cpu_addr; cpu_addr 1023 drivers/scsi/3w-xxxx.c dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_New_Ioctl) - 1, cpu_addr, dma_handle); cpu_addr 299 drivers/staging/mt7621-pci/pci-mt7621.c ioremap(range.cpu_addr, range.size); cpu_addr 481 drivers/tty/serial/msm_serial.c void *cpu_addr; cpu_addr 485 drivers/tty/serial/msm_serial.c cpu_addr = &xmit->buf[xmit->tail]; cpu_addr 487 drivers/tty/serial/msm_serial.c dma->phys = dma_map_single(port->dev, cpu_addr, count, dma->dir); cpu_addr 30 drivers/usb/host/xhci-dbgcap.c void *cpu_addr, dma_addr_t dma_handle) cpu_addr 32 drivers/usb/host/xhci-dbgcap.c if (cpu_addr) cpu_addr 34 drivers/usb/host/xhci-dbgcap.c size, cpu_addr, dma_handle); cpu_addr 343 drivers/virtio/virtio_ring.c void *cpu_addr, size_t size, cpu_addr 347 drivers/virtio/virtio_ring.c return (dma_addr_t)virt_to_phys(cpu_addr); cpu_addr 350 drivers/virtio/virtio_ring.c cpu_addr, size, direction); cpu_addr 71 include/linux/dma-direct.h void dma_direct_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 75 include/linux/dma-direct.h void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr, cpu_addr 163 include/linux/dma-mapping.h void *cpu_addr, size_t size, int *ret); cpu_addr 167 include/linux/dma-mapping.h int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr, cpu_addr 187 include/linux/dma-mapping.h void *cpu_addr, size_t size, cpu_addr 447 include/linux/dma-mapping.h void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, cpu_addr 456 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 459 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 524 include/linux/dma-mapping.h static void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, cpu_addr 542 include/linux/dma-mapping.h struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, cpu_addr 548 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 625 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 628 include/linux/dma-mapping.h struct page **dma_common_find_pages(void *cpu_addr); cpu_addr 634 include/linux/dma-mapping.h void dma_common_free_remap(void *cpu_addr, size_t size); cpu_addr 641 include/linux/dma-mapping.h dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr, cpu_addr 653 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_handle) cpu_addr 655 include/linux/dma-mapping.h return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0); cpu_addr 793 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr) cpu_addr 795 include/linux/dma-mapping.h return dma_free_attrs(dev, size, cpu_addr, dma_addr, cpu_addr 801 include/linux/dma-mapping.h void *cpu_addr, dma_addr_t dma_addr, cpu_addr 804 include/linux/dma-mapping.h return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, cpu_addr 42 include/linux/dma-noncoherent.h void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, cpu_addr 44 include/linux/dma-noncoherent.h long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, cpu_addr 20 include/linux/of_address.h u64 cpu_addr; cpu_addr 27 include/linux/remoteproc/st_slim_rproc.h void __iomem *cpu_addr; cpu_addr 114 include/media/drv-intf/saa7146.h __le32 *cpu_addr; cpu_addr 15 include/media/drv-intf/saa7146_vv.h dev->d_rps0.cpu_addr[ count++ ] = cpu_to_le32(x); \ cpu_addr 19 include/media/drv-intf/saa7146_vv.h dev->d_rps1.cpu_addr[ count++ ] = cpu_to_le32(x); \ cpu_addr 32 include/net/smc.h void *cpu_addr; cpu_addr 3945 include/rdma/ib_verbs.h void *cpu_addr, size_t size, cpu_addr 3948 include/rdma/ib_verbs.h return dma_map_single(dev->dma_device, cpu_addr, size, direction); cpu_addr 4105 include/rdma/ib_verbs.h size_t size, void *cpu_addr, cpu_addr 4108 include/rdma/ib_verbs.h dma_free_coherent(dev->dma_device, size, cpu_addr, dma_handle); cpu_addr 15 include/xen/arm/page-coherent.h void *cpu_addr, dma_addr_t dma_handle, unsigned long attrs) cpu_addr 17 include/xen/arm/page-coherent.h dma_direct_free(hwdev, size, cpu_addr, dma_handle, attrs); cpu_addr 184 kernel/dma/direct.c void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr, cpu_addr 192 kernel/dma/direct.c __dma_direct_free_pages(dev, size, cpu_addr); cpu_addr 197 kernel/dma/direct.c set_memory_encrypted((unsigned long)cpu_addr, 1 << page_order); cpu_addr 201 kernel/dma/direct.c cpu_addr = cached_kernel_address(cpu_addr); cpu_addr 202 kernel/dma/direct.c __dma_direct_free_pages(dev, size, virt_to_page(cpu_addr)); cpu_addr 215 kernel/dma/direct.c void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs) cpu_addr 219 kernel/dma/direct.c arch_dma_free(dev, size, cpu_addr, dma_addr, attrs); cpu_addr 221 kernel/dma/direct.c dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs); cpu_addr 8 kernel/dma/dummy.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 112 kernel/dma/mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 125 kernel/dma/mapping.c pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr); cpu_addr 130 kernel/dma/mapping.c page = virt_to_page(cpu_addr); cpu_addr 151 kernel/dma/mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 157 kernel/dma/mapping.c return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, cpu_addr 161 kernel/dma/mapping.c return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); cpu_addr 190 kernel/dma/mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 202 kernel/dma/mapping.c if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) cpu_addr 213 kernel/dma/mapping.c pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr); cpu_addr 217 kernel/dma/mapping.c pfn = page_to_pfn(virt_to_page(cpu_addr)); cpu_addr 262 kernel/dma/mapping.c void *cpu_addr, dma_addr_t dma_addr, size_t size, cpu_addr 268 kernel/dma/mapping.c return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size, cpu_addr 272 kernel/dma/mapping.c return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs); cpu_addr 301 kernel/dma/mapping.c void *cpu_addr; cpu_addr 305 kernel/dma/mapping.c if (dma_alloc_from_dev_coherent(dev, size, dma_handle, &cpu_addr)) cpu_addr 306 kernel/dma/mapping.c return cpu_addr; cpu_addr 312 kernel/dma/mapping.c cpu_addr = dma_direct_alloc(dev, size, dma_handle, flag, attrs); cpu_addr 314 kernel/dma/mapping.c cpu_addr = ops->alloc(dev, size, dma_handle, flag, attrs); cpu_addr 318 kernel/dma/mapping.c debug_dma_alloc_coherent(dev, size, *dma_handle, cpu_addr); cpu_addr 319 kernel/dma/mapping.c return cpu_addr; cpu_addr 323 kernel/dma/mapping.c void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, cpu_addr 328 kernel/dma/mapping.c if (dma_release_from_dev_coherent(dev, get_order(size), cpu_addr)) cpu_addr 339 kernel/dma/mapping.c if (!cpu_addr) cpu_addr 342 kernel/dma/mapping.c debug_dma_free_coherent(dev, size, cpu_addr, dma_handle); cpu_addr 344 kernel/dma/mapping.c dma_direct_free(dev, size, cpu_addr, dma_handle, attrs); cpu_addr 346 kernel/dma/mapping.c ops->free(dev, size, cpu_addr, dma_handle, attrs); cpu_addr 14 kernel/dma/remap.c struct page **dma_common_find_pages(void *cpu_addr) cpu_addr 16 kernel/dma/remap.c struct vm_struct *area = find_vm_area(cpu_addr); cpu_addr 88 kernel/dma/remap.c void dma_common_free_remap(void *cpu_addr, size_t size) cpu_addr 90 kernel/dma/remap.c struct vm_struct *area = find_vm_area(cpu_addr); cpu_addr 93 kernel/dma/remap.c WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); cpu_addr 97 kernel/dma/remap.c unmap_kernel_range((unsigned long)cpu_addr, PAGE_ALIGN(size)); cpu_addr 98 kernel/dma/remap.c vunmap(cpu_addr); cpu_addr 263 kernel/dma/remap.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, cpu_addr 23 kernel/dma/virt.c void *cpu_addr, dma_addr_t dma_addr, cpu_addr 26 kernel/dma/virt.c free_pages((unsigned long)cpu_addr, get_order(size)); cpu_addr 231 net/smc/smc_cdc.c base = (char *)conn->rmb_desc->cpu_addr + conn->rx_off; cpu_addr 334 net/smc/smc_cdc.c data_cdc = (struct smcd_cdc_msg *)conn->rmb_desc->cpu_addr; cpu_addr 393 net/smc/smc_core.c kfree(buf_desc->cpu_addr); cpu_addr 760 net/smc/smc_core.c buf_desc->cpu_addr = (void *)page_address(buf_desc->pages); cpu_addr 771 net/smc/smc_core.c buf_desc->cpu_addr, bufsize); cpu_addr 819 net/smc/smc_core.c buf_desc->pages = virt_to_page(buf_desc->cpu_addr); cpu_addr 823 net/smc/smc_core.c buf_desc->cpu_addr = kzalloc(bufsize, GFP_KERNEL | cpu_addr 826 net/smc/smc_core.c if (!buf_desc->cpu_addr) { cpu_addr 869 net/smc/smc_core.c memset(buf_desc->cpu_addr, 0, bufsize); cpu_addr 149 net/smc/smc_core.h void *cpu_addr; /* virtual address of buffer */ cpu_addr 153 net/smc/smc_ism.c dmb.cpu_addr = dmb_desc->cpu_addr; cpu_addr 174 net/smc/smc_ism.c dmb_desc->cpu_addr = dmb.cpu_addr; cpu_addr 164 net/smc/smc_rx.c partial.offset = src - (char *)smc->conn.rmb_desc->cpu_addr; cpu_addr 311 net/smc/smc_rx.c rcvbuf_base = conn->rx_off + conn->rmb_desc->cpu_addr; cpu_addr 181 net/smc/smc_tx.c sndbuf_base = conn->sndbuf_desc->cpu_addr; cpu_addr 370 net/smc/smc_tx.c void *data = conn->sndbuf_desc->cpu_addr + src_off;