cpu_addr          264 arch/alpha/kernel/pci_iommu.c pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
cpu_addr          276 arch/alpha/kernel/pci_iommu.c 	paddr = __pa(cpu_addr);
cpu_addr          285 arch/alpha/kernel/pci_iommu.c 		      cpu_addr, size, ret, __builtin_return_address(0));
cpu_addr          296 arch/alpha/kernel/pci_iommu.c 		      cpu_addr, size, ret, __builtin_return_address(0));
cpu_addr          330 arch/alpha/kernel/pci_iommu.c 	ret += (unsigned long)cpu_addr & ~PAGE_MASK;
cpu_addr          333 arch/alpha/kernel/pci_iommu.c 	      cpu_addr, size, npages, ret, __builtin_return_address(0));
cpu_addr          452 arch/alpha/kernel/pci_iommu.c 	void *cpu_addr;
cpu_addr          458 arch/alpha/kernel/pci_iommu.c 	cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order);
cpu_addr          459 arch/alpha/kernel/pci_iommu.c 	if (! cpu_addr) {
cpu_addr          467 arch/alpha/kernel/pci_iommu.c 	memset(cpu_addr, 0, size);
cpu_addr          469 arch/alpha/kernel/pci_iommu.c 	*dma_addrp = pci_map_single_1(pdev, cpu_addr, size, 0);
cpu_addr          471 arch/alpha/kernel/pci_iommu.c 		free_pages((unsigned long)cpu_addr, order);
cpu_addr          481 arch/alpha/kernel/pci_iommu.c 	      size, cpu_addr, *dma_addrp, __builtin_return_address(0));
cpu_addr          483 arch/alpha/kernel/pci_iommu.c 	return cpu_addr;
cpu_addr          493 arch/alpha/kernel/pci_iommu.c 				    void *cpu_addr, dma_addr_t dma_addr,
cpu_addr          498 arch/alpha/kernel/pci_iommu.c 	free_pages((unsigned long)cpu_addr, get_order(size));
cpu_addr          122 arch/arm/include/asm/dma-mapping.h extern void arm_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          139 arch/arm/include/asm/dma-mapping.h 			void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          195 arch/arm/include/asm/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          268 arch/arm/mach-omap2/common.h extern void omap_auxcoreboot_addr(u32 cpu_addr);
cpu_addr           55 arch/arm/mm/dma-mapping-nommu.c 			       void *cpu_addr, dma_addr_t dma_addr,
cpu_addr           58 arch/arm/mm/dma-mapping-nommu.c 	int ret = dma_release_from_global_coherent(get_order(size), cpu_addr);
cpu_addr           64 arch/arm/mm/dma-mapping-nommu.c 			      void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr           69 arch/arm/mm/dma-mapping-nommu.c 	if (dma_mmap_from_global_coherent(vma, cpu_addr, size, &ret))
cpu_addr           71 arch/arm/mm/dma-mapping-nommu.c 	if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
cpu_addr           57 arch/arm/mm/dma-mapping.c 	void *cpu_addr;
cpu_addr          203 arch/arm/mm/dma-mapping.c static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          206 arch/arm/mm/dma-mapping.c 		 void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          581 arch/arm/mm/dma-mapping.c 				   void *cpu_addr, size_t size, bool want_vaddr)
cpu_addr          585 arch/arm/mm/dma-mapping.c 			dma_common_free_remap(cpu_addr, size);
cpu_addr          641 arch/arm/mm/dma-mapping.c 	__free_from_contiguous(args->dev, args->page, args->cpu_addr,
cpu_addr          658 arch/arm/mm/dma-mapping.c 	__free_from_pool(args->cpu_addr, args->size);
cpu_addr          677 arch/arm/mm/dma-mapping.c 		dma_common_free_remap(args->cpu_addr, args->size);
cpu_addr          788 arch/arm/mm/dma-mapping.c 		 void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          797 arch/arm/mm/dma-mapping.c 	if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
cpu_addr          814 arch/arm/mm/dma-mapping.c 		 void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          817 arch/arm/mm/dma-mapping.c 	return __arm_dma_mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
cpu_addr          821 arch/arm/mm/dma-mapping.c 		 void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          825 arch/arm/mm/dma-mapping.c 	return __arm_dma_mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
cpu_addr          831 arch/arm/mm/dma-mapping.c static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          840 arch/arm/mm/dma-mapping.c 		.cpu_addr = cpu_addr,
cpu_addr          845 arch/arm/mm/dma-mapping.c 	buf = arm_dma_buffer_find(cpu_addr);
cpu_addr          846 arch/arm/mm/dma-mapping.c 	if (WARN(!buf, "Freeing invalid buffer %p\n", cpu_addr))
cpu_addr          853 arch/arm/mm/dma-mapping.c void arm_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          856 arch/arm/mm/dma-mapping.c 	__arm_dma_free(dev, size, cpu_addr, handle, attrs, false);
cpu_addr          859 arch/arm/mm/dma-mapping.c static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          862 arch/arm/mm/dma-mapping.c 	__arm_dma_free(dev, size, cpu_addr, handle, attrs, true);
cpu_addr          866 arch/arm/mm/dma-mapping.c 		 void *cpu_addr, dma_addr_t handle, size_t size,
cpu_addr         1415 arch/arm/mm/dma-mapping.c static struct page **__iommu_get_pages(void *cpu_addr, unsigned long attrs)
cpu_addr         1417 arch/arm/mm/dma-mapping.c 	if (__in_atomic_pool(cpu_addr, PAGE_SIZE))
cpu_addr         1418 arch/arm/mm/dma-mapping.c 		return __atomic_get_pages(cpu_addr);
cpu_addr         1421 arch/arm/mm/dma-mapping.c 		return cpu_addr;
cpu_addr         1423 arch/arm/mm/dma-mapping.c 	return dma_common_find_pages(cpu_addr);
cpu_addr         1451 arch/arm/mm/dma-mapping.c static void __iommu_free_atomic(struct device *dev, void *cpu_addr,
cpu_addr         1456 arch/arm/mm/dma-mapping.c 		__dma_free_buffer(virt_to_page(cpu_addr), size);
cpu_addr         1458 arch/arm/mm/dma-mapping.c 		__free_from_pool(cpu_addr, size);
cpu_addr         1523 arch/arm/mm/dma-mapping.c 		    void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr         1526 arch/arm/mm/dma-mapping.c 	struct page **pages = __iommu_get_pages(cpu_addr, attrs);
cpu_addr         1543 arch/arm/mm/dma-mapping.c 		struct vm_area_struct *vma, void *cpu_addr,
cpu_addr         1548 arch/arm/mm/dma-mapping.c 	return __arm_iommu_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, attrs);
cpu_addr         1552 arch/arm/mm/dma-mapping.c 		struct vm_area_struct *vma, void *cpu_addr,
cpu_addr         1555 arch/arm/mm/dma-mapping.c 	return __arm_iommu_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, attrs);
cpu_addr         1562 arch/arm/mm/dma-mapping.c void __arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr,
cpu_addr         1568 arch/arm/mm/dma-mapping.c 	if (coherent_flag == COHERENT || __in_atomic_pool(cpu_addr, size)) {
cpu_addr         1569 arch/arm/mm/dma-mapping.c 		__iommu_free_atomic(dev, cpu_addr, handle, size, coherent_flag);
cpu_addr         1573 arch/arm/mm/dma-mapping.c 	pages = __iommu_get_pages(cpu_addr, attrs);
cpu_addr         1575 arch/arm/mm/dma-mapping.c 		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr);
cpu_addr         1580 arch/arm/mm/dma-mapping.c 		dma_common_free_remap(cpu_addr, size);
cpu_addr         1587 arch/arm/mm/dma-mapping.c 		    void *cpu_addr, dma_addr_t handle, unsigned long attrs)
cpu_addr         1589 arch/arm/mm/dma-mapping.c 	__arm_iommu_free_attrs(dev, size, cpu_addr, handle, attrs, NORMAL);
cpu_addr         1593 arch/arm/mm/dma-mapping.c 		    void *cpu_addr, dma_addr_t handle, unsigned long attrs)
cpu_addr         1595 arch/arm/mm/dma-mapping.c 	__arm_iommu_free_attrs(dev, size, cpu_addr, handle, attrs, COHERENT);
cpu_addr         1599 arch/arm/mm/dma-mapping.c 				 void *cpu_addr, dma_addr_t dma_addr,
cpu_addr         1603 arch/arm/mm/dma-mapping.c 	struct page **pages = __iommu_get_pages(cpu_addr, attrs);
cpu_addr         2349 arch/arm/mm/dma-mapping.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
cpu_addr         2363 arch/arm/mm/dma-mapping.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr         2366 arch/arm/mm/dma-mapping.c 	__arm_dma_free(dev, size, cpu_addr, dma_handle, attrs, false);
cpu_addr           17 arch/ia64/kernel/dma-mapping.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr           20 arch/ia64/kernel/dma-mapping.c 	dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs);
cpu_addr           23 arch/ia64/kernel/dma-mapping.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
cpu_addr           26 arch/ia64/kernel/dma-mapping.c 	return page_to_pfn(virt_to_page(cpu_addr));
cpu_addr          439 arch/microblaze/pci/pci-common.c 					range.cpu_addr, range.size);
cpu_addr          446 arch/microblaze/pci/pci-common.c 		if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
cpu_addr          454 arch/microblaze/pci/pci-common.c 				range.cpu_addr, range.cpu_addr + range.size - 1,
cpu_addr          467 arch/microblaze/pci/pci-common.c 			hose->io_base_virt = ioremap(range.cpu_addr,
cpu_addr          478 arch/microblaze/pci/pci-common.c 			hose->io_base_phys = range.cpu_addr - range.pci_addr;
cpu_addr          482 arch/microblaze/pci/pci-common.c 			range.cpu_addr = range.pci_addr;
cpu_addr          487 arch/microblaze/pci/pci-common.c 				range.cpu_addr, range.cpu_addr + range.size - 1,
cpu_addr          499 arch/microblaze/pci/pci-common.c 				isa_mb = range.cpu_addr;
cpu_addr          502 arch/microblaze/pci/pci-common.c 					isa_mem_base = range.cpu_addr;
cpu_addr          503 arch/microblaze/pci/pci-common.c 				hose->isa_mem_phys = range.cpu_addr;
cpu_addr          514 arch/microblaze/pci/pci-common.c 				hose->pci_mem_offset = range.cpu_addr -
cpu_addr          517 arch/microblaze/pci/pci-common.c 				 hose->pci_mem_offset != range.cpu_addr -
cpu_addr          530 arch/microblaze/pci/pci-common.c 			res->start = range.cpu_addr;
cpu_addr          531 arch/microblaze/pci/pci-common.c 			res->end = range.cpu_addr + range.size - 1;
cpu_addr           62 arch/mips/mm/dma-noncoherent.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
cpu_addr           65 arch/mips/mm/dma-noncoherent.c 	return page_to_pfn(virt_to_page(cached_kernel_address(cpu_addr)));
cpu_addr          155 arch/mips/pci/pci-legacy.c 				range.cpu_addr,
cpu_addr          156 arch/mips/pci/pci-legacy.c 				range.cpu_addr + range.size - 1);
cpu_addr          158 arch/mips/pci/pci-legacy.c 				(unsigned long)ioremap(range.cpu_addr,
cpu_addr          164 arch/mips/pci/pci-legacy.c 				range.cpu_addr,
cpu_addr          165 arch/mips/pci/pci-legacy.c 				range.cpu_addr + range.size - 1);
cpu_addr          690 arch/powerpc/kernel/pci-common.c 		if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
cpu_addr          699 arch/powerpc/kernel/pci-common.c 			       range.cpu_addr, range.cpu_addr + range.size - 1,
cpu_addr          714 arch/powerpc/kernel/pci-common.c 			hose->io_base_virt = ioremap(range.cpu_addr,
cpu_addr          726 arch/powerpc/kernel/pci-common.c 			hose->io_base_phys = range.cpu_addr - range.pci_addr;
cpu_addr          730 arch/powerpc/kernel/pci-common.c 			range.cpu_addr = range.pci_addr;
cpu_addr          735 arch/powerpc/kernel/pci-common.c 			       range.cpu_addr, range.cpu_addr + range.size - 1,
cpu_addr          749 arch/powerpc/kernel/pci-common.c 					isa_mem_base = range.cpu_addr;
cpu_addr          750 arch/powerpc/kernel/pci-common.c 				hose->isa_mem_phys = range.cpu_addr;
cpu_addr          755 arch/powerpc/kernel/pci-common.c 			hose->mem_offset[memno] = range.cpu_addr -
cpu_addr          763 arch/powerpc/kernel/pci-common.c 			res->start = range.cpu_addr;
cpu_addr          764 arch/powerpc/kernel/pci-common.c 			res->end = range.cpu_addr + range.size - 1;
cpu_addr          116 arch/powerpc/platforms/4xx/pci.c 		u64 cpu_addr = of_translate_dma_address(hose->dn, ranges + 3);
cpu_addr          119 arch/powerpc/platforms/4xx/pci.c 		if (cpu_addr == OF_BAD_ADDR || size == 0)
cpu_addr          129 arch/powerpc/platforms/4xx/pci.c 		if (cpu_addr != 0 || pci_addr > 0xffffffff) {
cpu_addr          133 arch/powerpc/platforms/4xx/pci.c 			       pci_addr, pci_addr + size - 1, cpu_addr);
cpu_addr          794 arch/powerpc/platforms/cell/iommu.c 	u64 cpu_addr, size, best_size, dev_addr = OF_BAD_ADDR;
cpu_addr          834 arch/powerpc/platforms/cell/iommu.c 		cpu_addr = of_translate_dma_address(np, ranges + i + naddr);
cpu_addr          837 arch/powerpc/platforms/cell/iommu.c 		if (cpu_addr == 0 && size > best_size) {
cpu_addr          132 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_cpu_info *)hdr)->cpu_addr;
cpu_addr          134 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_x_cpu_info *)hdr)->cpu_addr;
cpu_addr          192 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_phys_cpu *)hdr)->cpu_addr;
cpu_addr          194 arch/s390/hypfs/hypfs_diag.c 		return ((struct diag204_x_phys_cpu *)hdr)->cpu_addr;
cpu_addr          231 arch/s390/include/asm/ccwdev.h 				void *cpu_addr, size_t size);
cpu_addr          363 arch/s390/include/asm/cio.h extern void cio_dma_free(void *cpu_addr, size_t size);
cpu_addr          368 arch/s390/include/asm/cio.h void cio_gp_dma_free(struct gen_pool *gp_dma, void *cpu_addr, size_t size);
cpu_addr          158 arch/s390/include/asm/diag.h 	__u16 cpu_addr;
cpu_addr          168 arch/s390/include/asm/diag.h 	__u16 cpu_addr;
cpu_addr          204 arch/s390/include/asm/diag.h 	__u16 cpu_addr;
cpu_addr          213 arch/s390/include/asm/diag.h 	__u16 cpu_addr;
cpu_addr          738 arch/s390/kvm/interrupt.c 	int cpu_addr;
cpu_addr          741 arch/s390/kvm/interrupt.c 	cpu_addr = find_first_bit(li->sigp_emerg_pending, KVM_MAX_VCPUS);
cpu_addr          742 arch/s390/kvm/interrupt.c 	clear_bit(cpu_addr, li->sigp_emerg_pending);
cpu_addr          750 arch/s390/kvm/interrupt.c 					 cpu_addr, 0);
cpu_addr          754 arch/s390/kvm/interrupt.c 	rc |= put_guest_lc(vcpu, cpu_addr, (u16 *)__LC_EXT_CPU_ADDR);
cpu_addr          283 arch/s390/kvm/sigp.c 			   u16 cpu_addr, u32 parameter, u64 *status_reg)
cpu_addr          286 arch/s390/kvm/sigp.c 	struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr);
cpu_addr          360 arch/s390/kvm/sigp.c 					   u16 cpu_addr)
cpu_addr          404 arch/s390/kvm/sigp.c 		   order_code, cpu_addr);
cpu_addr          414 arch/s390/kvm/sigp.c 	u16 cpu_addr = vcpu->run->s.regs.gprs[r3];
cpu_addr          423 arch/s390/kvm/sigp.c 	if (handle_sigp_order_in_user_space(vcpu, order_code, cpu_addr))
cpu_addr          431 arch/s390/kvm/sigp.c 	trace_kvm_s390_handle_sigp(vcpu, order_code, cpu_addr, parameter);
cpu_addr          439 arch/s390/kvm/sigp.c 		rc = handle_sigp_dst(vcpu, order_code, cpu_addr,
cpu_addr          463 arch/s390/kvm/sigp.c 	u16 cpu_addr = vcpu->run->s.regs.gprs[r3];
cpu_addr          467 arch/s390/kvm/sigp.c 	trace_kvm_s390_handle_sigp_pei(vcpu, order_code, cpu_addr);
cpu_addr          470 arch/s390/kvm/sigp.c 		dest_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr);
cpu_addr          220 arch/s390/kvm/trace.h 	    TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr, \
cpu_addr          222 arch/s390/kvm/trace.h 	    TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr, parameter),
cpu_addr          227 arch/s390/kvm/trace.h 		    __field(__u16, cpu_addr)
cpu_addr          234 arch/s390/kvm/trace.h 		    __entry->cpu_addr = cpu_addr;
cpu_addr          242 arch/s390/kvm/trace.h 			   __entry->cpu_addr, __entry->parameter)
cpu_addr          246 arch/s390/kvm/trace.h 	    TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr),
cpu_addr          247 arch/s390/kvm/trace.h 	    TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr),
cpu_addr          252 arch/s390/kvm/trace.h 		    __field(__u16, cpu_addr)
cpu_addr          258 arch/s390/kvm/trace.h 		    __entry->cpu_addr = cpu_addr;
cpu_addr          265 arch/s390/kvm/trace.h 			   __entry->cpu_addr)
cpu_addr           97 arch/sparc/include/asm/dma.h bool sparc_dma_free_resource(void *cpu_addr, size_t size);
cpu_addr          267 arch/sparc/kernel/ioport.c bool sparc_dma_free_resource(void *cpu_addr, size_t size)
cpu_addr          269 arch/sparc/kernel/ioport.c 	unsigned long addr = (unsigned long)cpu_addr;
cpu_addr          274 arch/sparc/kernel/ioport.c 		printk("%s: cannot free %p\n", __func__, cpu_addr);
cpu_addr          279 arch/sparc/kernel/ioport.c 		printk("%s: unaligned va %p\n", __func__, cpu_addr);
cpu_addr          356 arch/sparc/kernel/ioport.c void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          359 arch/sparc/kernel/ioport.c 	if (!sparc_dma_free_resource(cpu_addr, PAGE_ALIGN(size)))
cpu_addr          363 arch/sparc/kernel/ioport.c 	srmmu_unmapiorange((unsigned long)cpu_addr, size);
cpu_addr          270 arch/sparc/mm/io-unit.c static void iounit_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          391 arch/sparc/mm/iommu.c static void sbus_iommu_free(struct device *dev, size_t len, void *cpu_addr,
cpu_addr          396 arch/sparc/mm/iommu.c 	struct page *page = virt_to_page(cpu_addr);
cpu_addr          400 arch/sparc/mm/iommu.c 	if (!sparc_dma_free_resource(cpu_addr, len))
cpu_addr           18 arch/x86/include/asm/xen/page-coherent.h 		void *cpu_addr, dma_addr_t dma_handle,
cpu_addr           21 arch/x86/include/asm/xen/page-coherent.h 	free_pages((unsigned long) cpu_addr, get_order(size));
cpu_addr          743 drivers/acpi/pci_root.c 	resource_size_t cpu_addr = res->start;
cpu_addr          744 drivers/acpi/pci_root.c 	resource_size_t pci_addr = cpu_addr - entry->offset;
cpu_addr          748 drivers/acpi/pci_root.c 	if (pci_register_io_range(fwnode, cpu_addr, length))
cpu_addr          751 drivers/acpi/pci_root.c 	port = pci_address_to_pio(cpu_addr);
cpu_addr          759 drivers/acpi/pci_root.c 	if (pci_remap_iospace(res, cpu_addr) < 0)
cpu_addr          762 drivers/acpi/pci_root.c 	pr_info("Remapped I/O %pa to %pR\n", &cpu_addr, res);
cpu_addr          123 drivers/crypto/img-hash.c 	void __iomem		*cpu_addr;
cpu_addr          203 drivers/crypto/img-hash.c 		writel_relaxed(buffer[count], hdev->cpu_addr);
cpu_addr          971 drivers/crypto/img-hash.c 	hdev->cpu_addr = devm_ioremap_resource(dev, hash_res);
cpu_addr          972 drivers/crypto/img-hash.c 	if (IS_ERR(hdev->cpu_addr)) {
cpu_addr          974 drivers/crypto/img-hash.c 		err = PTR_ERR(hdev->cpu_addr);
cpu_addr          185 drivers/dma/st_fdma.h 	readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
cpu_addr          190 drivers/dma/st_fdma.h 	writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
cpu_addr          197 drivers/dma/st_fdma.h 	writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
cpu_addr          207 drivers/dma/st_fdma.h 	readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
cpu_addr          212 drivers/dma/st_fdma.h 	writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
cpu_addr         1002 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	preempt_seq = le32_to_cpu(*(drv->cpu_addr + 2));
cpu_addr          102 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	if (drv->cpu_addr)
cpu_addr          103 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		*drv->cpu_addr = cpu_to_le32(seq);
cpu_addr          119 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	if (drv->cpu_addr)
cpu_addr          120 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		seq = le32_to_cpu(*drv->cpu_addr);
cpu_addr          398 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs];
cpu_addr          403 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index;
cpu_addr          415 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		      ring->fence_drv.gpu_addr, ring->fence_drv.cpu_addr);
cpu_addr          443 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	ring->fence_drv.cpu_addr = NULL;
cpu_addr          730 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 			   le32_to_cpu(*(ring->fence_drv.cpu_addr + 2)));
cpu_addr          733 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 			   le32_to_cpu(*(ring->fence_drv.cpu_addr + 4)));
cpu_addr          736 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 			   le32_to_cpu(*(ring->fence_drv.cpu_addr + 6)));
cpu_addr          234 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			      u64 *gpu_addr, void **cpu_addr)
cpu_addr          249 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	bp.flags = cpu_addr ? AMDGPU_GEM_CREATE_CPU_ACCESS_REQUIRED
cpu_addr          286 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (cpu_addr) {
cpu_addr          287 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		r = amdgpu_bo_kmap(*bo_ptr, cpu_addr);
cpu_addr          329 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			    u64 *gpu_addr, void **cpu_addr)
cpu_addr          334 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 				      gpu_addr, cpu_addr);
cpu_addr          362 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			       struct amdgpu_bo **bo_ptr, void **cpu_addr)
cpu_addr          391 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (cpu_addr) {
cpu_addr          392 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		r = amdgpu_bo_kmap(*bo_ptr, cpu_addr);
cpu_addr          416 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			   void **cpu_addr)
cpu_addr          422 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		if (cpu_addr)
cpu_addr          433 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (cpu_addr)
cpu_addr          434 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		*cpu_addr = NULL;
cpu_addr          235 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 			      u64 *gpu_addr, void **cpu_addr);
cpu_addr          239 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 			    u64 *gpu_addr, void **cpu_addr);
cpu_addr          242 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 			       struct amdgpu_bo **bo_ptr, void **cpu_addr);
cpu_addr          244 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 			   void **cpu_addr);
cpu_addr           73 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h 	volatile uint32_t		*cpu_addr;
cpu_addr          260 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 					    &adev->uvd.inst[j].gpu_addr, &adev->uvd.inst[j].cpu_addr);
cpu_addr          309 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 				      (void **)&adev->uvd.inst[j].cpu_addr);
cpu_addr          369 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		ptr = adev->uvd.inst[j].cpu_addr;
cpu_addr          393 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		ptr = adev->uvd.inst[i].cpu_addr;
cpu_addr          406 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 				memcpy_toio(adev->uvd.inst[i].cpu_addr, adev->uvd.fw->data + offset,
cpu_addr           42 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h 	void			*cpu_addr;
cpu_addr          183 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 				    &adev->vce.gpu_addr, &adev->vce.cpu_addr);
cpu_addr          217 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 		(void **)&adev->vce.cpu_addr);
cpu_addr          285 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 	void *cpu_addr;
cpu_addr          299 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 	r = amdgpu_bo_kmap(adev->vce.vcpu_bo, &cpu_addr);
cpu_addr          308 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 	memcpy_toio(cpu_addr, adev->vce.fw->data + offset,
cpu_addr           38 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.h 	void			*cpu_addr;
cpu_addr          172 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 						&adev->vcn.inst[i].gpu_addr, &adev->vcn.inst[i].cpu_addr);
cpu_addr          209 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 					  (void **)&adev->vcn.inst[j].cpu_addr);
cpu_addr          239 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 		ptr = adev->vcn.inst[i].cpu_addr;
cpu_addr          263 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 		ptr = adev->vcn.inst[i].cpu_addr;
cpu_addr          276 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 				memcpy_toio(adev->vcn.inst[i].cpu_addr, adev->vcn.fw->data + offset,
cpu_addr          166 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h 	void			*cpu_addr;
cpu_addr          280 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c 				    (void *)&adev->virt.mm_table.cpu_addr);
cpu_addr          286 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c 	memset((void *)adev->virt.mm_table.cpu_addr, 0, PAGE_SIZE);
cpu_addr          289 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c 		 adev->virt.mm_table.cpu_addr);
cpu_addr          305 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c 			      (void *)&adev->virt.mm_table.cpu_addr);
cpu_addr           35 drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h 	uint32_t		*cpu_addr;
cpu_addr          713 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 	struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr;
cpu_addr          773 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c 	uint32_t *init_table = adev->virt.mm_table.cpu_addr;
cpu_addr          157 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 	struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr;
cpu_addr          213 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 	uint32_t *init_table = adev->virt.mm_table.cpu_addr;
cpu_addr          568 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 		void *ptr = adev->vce.cpu_addr;
cpu_addr          590 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c 		void *ptr = adev->vce.cpu_addr;
cpu_addr          500 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 			    &compressor->gpu_addr, &compressor->cpu_addr);
cpu_addr           85 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.h 	void *cpu_addr;
cpu_addr          448 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 		memcpy(table->cpu_addr, table_data, table->size);
cpu_addr          469 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 		memcpy(table_data, table->cpu_addr, table->size);
cpu_addr          939 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 					      &tables[i].cpu_addr);
cpu_addr          951 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 				      &tables[i].cpu_addr);
cpu_addr          972 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 				      &tables[i].cpu_addr);
cpu_addr         1203 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 					      &memory_pool->cpu_addr);
cpu_addr         1223 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c 			      &memory_pool->cpu_addr);
cpu_addr          182 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h 	void *cpu_addr;
cpu_addr          684 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	if (memory_pool->size == 0 || memory_pool->cpu_addr == NULL)
cpu_addr          687 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	address = (uintptr_t)memory_pool->cpu_addr;
cpu_addr          782 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	if (!table->cpu_addr)
cpu_addr          785 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	ret = smu_update_table(smu, SMU_TABLE_WATERMARKS, 0, table->cpu_addr,
cpu_addr         1342 drivers/gpu/drm/amd/powerplay/smu_v11_0.c 	void *table = watermarks->cpu_addr;
cpu_addr          316 drivers/gpu/drm/amd/powerplay/smu_v12_0.c 	if (!table->cpu_addr)
cpu_addr          363 drivers/gpu/drm/radeon/radeon.h 	volatile uint32_t		*cpu_addr;
cpu_addr         1671 drivers/gpu/drm/radeon/radeon.h 	void			*cpu_addr;
cpu_addr           72 drivers/gpu/drm/radeon/radeon_fence.c 		if (drv->cpu_addr) {
cpu_addr           73 drivers/gpu/drm/radeon/radeon_fence.c 			*drv->cpu_addr = cpu_to_le32(seq);
cpu_addr           95 drivers/gpu/drm/radeon/radeon_fence.c 		if (drv->cpu_addr) {
cpu_addr           96 drivers/gpu/drm/radeon/radeon_fence.c 			seq = le32_to_cpu(*drv->cpu_addr);
cpu_addr          843 drivers/gpu/drm/radeon/radeon_fence.c 			rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4];
cpu_addr          850 drivers/gpu/drm/radeon/radeon_fence.c 			rdev->fence_drv[ring].cpu_addr = rdev->uvd.cpu_addr + index;
cpu_addr          863 drivers/gpu/drm/radeon/radeon_fence.c 		rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4];
cpu_addr          869 drivers/gpu/drm/radeon/radeon_fence.c 		 ring, rdev->fence_drv[ring].gpu_addr, rdev->fence_drv[ring].cpu_addr);
cpu_addr          888 drivers/gpu/drm/radeon/radeon_fence.c 	rdev->fence_drv[ring].cpu_addr = NULL;
cpu_addr          211 drivers/gpu/drm/radeon/radeon_uvd.c 	r = radeon_bo_kmap(rdev->uvd.vcpu_bo, &rdev->uvd.cpu_addr);
cpu_addr          289 drivers/gpu/drm/radeon/radeon_uvd.c 	memcpy(rdev->uvd.cpu_addr, rdev->uvd_fw->data, rdev->uvd_fw->size);
cpu_addr          294 drivers/gpu/drm/radeon/radeon_uvd.c 	ptr = rdev->uvd.cpu_addr;
cpu_addr          784 drivers/gpu/drm/radeon/radeon_uvd.c 	uint32_t *msg = rdev->uvd.cpu_addr + offs;
cpu_addr          820 drivers/gpu/drm/radeon/radeon_uvd.c 	uint32_t *msg = rdev->uvd.cpu_addr + offs;
cpu_addr          223 drivers/gpu/drm/radeon/radeon_vce.c 	void *cpu_addr;
cpu_addr          235 drivers/gpu/drm/radeon/radeon_vce.c 	r = radeon_bo_kmap(rdev->vce.vcpu_bo, &cpu_addr);
cpu_addr          242 drivers/gpu/drm/radeon/radeon_vce.c 	memset(cpu_addr, 0, radeon_bo_size(rdev->vce.vcpu_bo));
cpu_addr          244 drivers/gpu/drm/radeon/radeon_vce.c 		r = vce_v1_0_load_fw(rdev, cpu_addr);
cpu_addr          246 drivers/gpu/drm/radeon/radeon_vce.c 		memcpy(cpu_addr, rdev->vce_fw->data, rdev->vce_fw->size);
cpu_addr          145 drivers/gpu/drm/radeon/uvd_v1_0.c 	WREG32(UVD_FW_START, *((uint32_t*)rdev->uvd.cpu_addr));
cpu_addr           93 drivers/infiniband/hw/efa/efa.h 	void *cpu_addr;
cpu_addr          895 drivers/infiniband/hw/efa/efa_verbs.c 		  cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr);
cpu_addr          907 drivers/infiniband/hw/efa/efa_verbs.c 					     virt_to_phys(cq->cpu_addr),
cpu_addr          988 drivers/infiniband/hw/efa/efa_verbs.c 	cq->cpu_addr = efa_zalloc_mapped(dev, &cq->dma_addr, cq->size,
cpu_addr          990 drivers/infiniband/hw/efa/efa_verbs.c 	if (!cq->cpu_addr) {
cpu_addr         1029 drivers/infiniband/hw/efa/efa_verbs.c 		  cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr);
cpu_addr         1039 drivers/infiniband/hw/efa/efa_verbs.c 		free_pages_exact(cq->cpu_addr, cq->size);
cpu_addr         1186 drivers/infiniband/hw/hns/hns_roce_hem.c 			       struct roce_hem_item *hem, void *cpu_addr,
cpu_addr         1189 drivers/infiniband/hw/hns/hns_roce_hem.c 	hem->addr = cpu_addr;
cpu_addr          922 drivers/iommu/dma-iommu.c static void __iommu_dma_free(struct device *dev, size_t size, void *cpu_addr)
cpu_addr          930 drivers/iommu/dma-iommu.c 	    dma_free_from_pool(cpu_addr, alloc_size))
cpu_addr          933 drivers/iommu/dma-iommu.c 	if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
cpu_addr          938 drivers/iommu/dma-iommu.c 		pages = dma_common_find_pages(cpu_addr);
cpu_addr          940 drivers/iommu/dma-iommu.c 			page = vmalloc_to_page(cpu_addr);
cpu_addr          941 drivers/iommu/dma-iommu.c 		dma_common_free_remap(cpu_addr, alloc_size);
cpu_addr          944 drivers/iommu/dma-iommu.c 		page = virt_to_page(cpu_addr);
cpu_addr          953 drivers/iommu/dma-iommu.c static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          957 drivers/iommu/dma-iommu.c 	__iommu_dma_free(dev, size, cpu_addr);
cpu_addr          967 drivers/iommu/dma-iommu.c 	void *cpu_addr;
cpu_addr          978 drivers/iommu/dma-iommu.c 		cpu_addr = dma_common_contiguous_remap(page, alloc_size,
cpu_addr          980 drivers/iommu/dma-iommu.c 		if (!cpu_addr)
cpu_addr          986 drivers/iommu/dma-iommu.c 		cpu_addr = page_address(page);
cpu_addr          990 drivers/iommu/dma-iommu.c 	memset(cpu_addr, 0, alloc_size);
cpu_addr          991 drivers/iommu/dma-iommu.c 	return cpu_addr;
cpu_addr         1003 drivers/iommu/dma-iommu.c 	void *cpu_addr;
cpu_addr         1013 drivers/iommu/dma-iommu.c 		cpu_addr = dma_alloc_from_pool(PAGE_ALIGN(size), &page, gfp);
cpu_addr         1015 drivers/iommu/dma-iommu.c 		cpu_addr = iommu_dma_alloc_pages(dev, size, &page, gfp, attrs);
cpu_addr         1016 drivers/iommu/dma-iommu.c 	if (!cpu_addr)
cpu_addr         1021 drivers/iommu/dma-iommu.c 		__iommu_dma_free(dev, size, cpu_addr);
cpu_addr         1025 drivers/iommu/dma-iommu.c 	return cpu_addr;
cpu_addr         1029 drivers/iommu/dma-iommu.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr         1038 drivers/iommu/dma-iommu.c 	if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
cpu_addr         1044 drivers/iommu/dma-iommu.c 	if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
cpu_addr         1045 drivers/iommu/dma-iommu.c 		struct page **pages = dma_common_find_pages(cpu_addr);
cpu_addr         1049 drivers/iommu/dma-iommu.c 		pfn = vmalloc_to_pfn(cpu_addr);
cpu_addr         1051 drivers/iommu/dma-iommu.c 		pfn = page_to_pfn(virt_to_page(cpu_addr));
cpu_addr         1060 drivers/iommu/dma-iommu.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr         1066 drivers/iommu/dma-iommu.c 	if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
cpu_addr         1067 drivers/iommu/dma-iommu.c 		struct page **pages = dma_common_find_pages(cpu_addr);
cpu_addr         1075 drivers/iommu/dma-iommu.c 		page = vmalloc_to_page(cpu_addr);
cpu_addr         1077 drivers/iommu/dma-iommu.c 		page = virt_to_page(cpu_addr);
cpu_addr          649 drivers/iommu/qcom_iommu.c 	void *cpu_addr;
cpu_addr          669 drivers/iommu/qcom_iommu.c 	cpu_addr = dma_alloc_attrs(dev, psize, &paddr, GFP_KERNEL, attrs);
cpu_addr          670 drivers/iommu/qcom_iommu.c 	if (!cpu_addr) {
cpu_addr          686 drivers/iommu/qcom_iommu.c 	dma_free_attrs(dev, psize, cpu_addr, paddr, attrs);
cpu_addr          415 drivers/media/common/saa7146/saa7146_core.c 	dev->d_rps0.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM,
cpu_addr          417 drivers/media/common/saa7146/saa7146_core.c 	if (!dev->d_rps0.cpu_addr)
cpu_addr          420 drivers/media/common/saa7146/saa7146_core.c 	dev->d_rps1.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM,
cpu_addr          422 drivers/media/common/saa7146/saa7146_core.c 	if (!dev->d_rps1.cpu_addr)
cpu_addr          425 drivers/media/common/saa7146/saa7146_core.c 	dev->d_i2c.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM,
cpu_addr          427 drivers/media/common/saa7146/saa7146_core.c 	if (!dev->d_i2c.cpu_addr)
cpu_addr          474 drivers/media/common/saa7146/saa7146_core.c 	pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_i2c.cpu_addr,
cpu_addr          477 drivers/media/common/saa7146/saa7146_core.c 	pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps1.cpu_addr,
cpu_addr          480 drivers/media/common/saa7146/saa7146_core.c 	pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps0.cpu_addr,
cpu_addr          503 drivers/media/common/saa7146/saa7146_core.c 		{ dev->d_i2c.cpu_addr, dev->d_i2c.dma_handle },
cpu_addr          504 drivers/media/common/saa7146/saa7146_core.c 		{ dev->d_rps1.cpu_addr, dev->d_rps1.dma_handle },
cpu_addr          505 drivers/media/common/saa7146/saa7146_core.c 		{ dev->d_rps0.cpu_addr, dev->d_rps0.dma_handle },
cpu_addr          521 drivers/media/common/saa7146/saa7146_fops.c 	vv->d_clipping.cpu_addr =
cpu_addr          524 drivers/media/common/saa7146/saa7146_fops.c 	if( NULL == vv->d_clipping.cpu_addr ) {
cpu_addr          581 drivers/media/common/saa7146/saa7146_fops.c 	pci_free_consistent(dev->pci, SAA7146_CLIPPING_MEM, vv->d_clipping.cpu_addr, vv->d_clipping.dma_handle);
cpu_addr          342 drivers/media/common/saa7146/saa7146_hlp.c 	__le32 *clipping = vv->d_clipping.cpu_addr;
cpu_addr          294 drivers/media/common/saa7146/saa7146_i2c.c 	__le32 *buffer = dev->d_i2c.cpu_addr;
cpu_addr          314 drivers/media/pci/smipcie/smipcie-main.c 			port->cpu_addr[0], (finishedData / 188));
cpu_addr          337 drivers/media/pci/smipcie/smipcie-main.c 			port->cpu_addr[1], (finishedData / 188));
cpu_addr          353 drivers/media/pci/smipcie/smipcie-main.c 	if (port->cpu_addr[0]) {
cpu_addr          355 drivers/media/pci/smipcie/smipcie-main.c 				    port->cpu_addr[0], port->dma_addr[0]);
cpu_addr          356 drivers/media/pci/smipcie/smipcie-main.c 		port->cpu_addr[0] = NULL;
cpu_addr          358 drivers/media/pci/smipcie/smipcie-main.c 	if (port->cpu_addr[1]) {
cpu_addr          360 drivers/media/pci/smipcie/smipcie-main.c 				    port->cpu_addr[1], port->dma_addr[1]);
cpu_addr          361 drivers/media/pci/smipcie/smipcie-main.c 		port->cpu_addr[1] = NULL;
cpu_addr          401 drivers/media/pci/smipcie/smipcie-main.c 		port->cpu_addr[0] = pci_alloc_consistent(port->dev->pci_dev,
cpu_addr          404 drivers/media/pci/smipcie/smipcie-main.c 		if (!port->cpu_addr[0]) {
cpu_addr          413 drivers/media/pci/smipcie/smipcie-main.c 		port->cpu_addr[1] = pci_alloc_consistent(port->dev->pci_dev,
cpu_addr          416 drivers/media/pci/smipcie/smipcie-main.c 		if (!port->cpu_addr[1]) {
cpu_addr          257 drivers/media/pci/smipcie/smipcie.h 	u8 *cpu_addr[2];
cpu_addr         2815 drivers/misc/habanalabs/goya/goya.c 					void *cpu_addr, dma_addr_t dma_handle)
cpu_addr         2820 drivers/misc/habanalabs/goya/goya.c 	dma_free_coherent(&hdev->pdev->dev, size, cpu_addr, fixed_dma_handle);
cpu_addr          526 drivers/misc/habanalabs/habanalabs.h 					void *cpu_addr, dma_addr_t dma_handle);
cpu_addr           87 drivers/net/caif/caif_spi.c static inline void dma_free(struct cfspi *cfspi, void *cpu_addr,
cpu_addr           90 drivers/net/caif/caif_spi.c 	kfree(cpu_addr);
cpu_addr          101 drivers/net/caif/caif_spi.c static inline void dma_free(struct cfspi *cfspi, void *cpu_addr,
cpu_addr          104 drivers/net/caif/caif_spi.c 	dma_free_coherent(&cfspi->pdev->dev, SPI_DMA_BUF_LEN, cpu_addr, handle);
cpu_addr          538 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	u64 dma_addr, cpu_addr;
cpu_addr          545 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	cpu_addr = nicvf_iova_to_phys(nic, dma_addr);
cpu_addr          546 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	if (!cpu_addr)
cpu_addr          548 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	cpu_addr = (u64)phys_to_virt(cpu_addr);
cpu_addr          549 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	page = virt_to_page((void *)cpu_addr);
cpu_addr          552 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	xdp.data = (void *)cpu_addr;
cpu_addr         9576 drivers/net/ethernet/sun/niu.c 				  void *cpu_addr, u64 handle)
cpu_addr         9578 drivers/net/ethernet/sun/niu.c 	dma_free_coherent(dev, size, cpu_addr, handle);
cpu_addr         9594 drivers/net/ethernet/sun/niu.c static u64 niu_pci_map_single(struct device *dev, void *cpu_addr,
cpu_addr         9598 drivers/net/ethernet/sun/niu.c 	return dma_map_single(dev, cpu_addr, size, direction);
cpu_addr         9958 drivers/net/ethernet/sun/niu.c 				   void *cpu_addr, u64 handle)
cpu_addr         9962 drivers/net/ethernet/sun/niu.c 	free_pages((unsigned long) cpu_addr, order);
cpu_addr         9978 drivers/net/ethernet/sun/niu.c static u64 niu_phys_map_single(struct device *dev, void *cpu_addr,
cpu_addr         9982 drivers/net/ethernet/sun/niu.c 	return __pa(cpu_addr);
cpu_addr         3132 drivers/net/ethernet/sun/niu.h 			      void *cpu_addr, u64 handle);
cpu_addr         3138 drivers/net/ethernet/sun/niu.h 	u64 (*map_single)(struct device *dev, void *cpu_addr,
cpu_addr          177 drivers/net/wireless/ath/wcn36xx/dxe.c 	wcn_ch->cpu_addr = dma_alloc_coherent(dev, size, &wcn_ch->dma_addr,
cpu_addr          179 drivers/net/wireless/ath/wcn36xx/dxe.c 	if (!wcn_ch->cpu_addr)
cpu_addr          182 drivers/net/wireless/ath/wcn36xx/dxe.c 	cur_dxe = (struct wcn36xx_dxe_desc *)wcn_ch->cpu_addr;
cpu_addr          232 drivers/net/wireless/ath/wcn36xx/dxe.c 	dma_free_coherent(dev, size,wcn_ch->cpu_addr, wcn_ch->dma_addr);
cpu_addr          620 drivers/net/wireless/ath/wcn36xx/dxe.c 	void *cpu_addr;
cpu_addr          629 drivers/net/wireless/ath/wcn36xx/dxe.c 	cpu_addr = dma_alloc_coherent(wcn->dev, s,
cpu_addr          632 drivers/net/wireless/ath/wcn36xx/dxe.c 	if (!cpu_addr)
cpu_addr          635 drivers/net/wireless/ath/wcn36xx/dxe.c 	wcn->mgmt_mem_pool.virt_addr = cpu_addr;
cpu_addr          644 drivers/net/wireless/ath/wcn36xx/dxe.c 	cpu_addr = dma_alloc_coherent(wcn->dev, s,
cpu_addr          647 drivers/net/wireless/ath/wcn36xx/dxe.c 	if (!cpu_addr)
cpu_addr          650 drivers/net/wireless/ath/wcn36xx/dxe.c 	wcn->data_mem_pool.virt_addr = cpu_addr;
cpu_addr          432 drivers/net/wireless/ath/wcn36xx/dxe.h 	void				*cpu_addr;
cpu_addr          210 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	void *cpu_addr = NULL;
cpu_addr          217 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		cpu_addr = dma_alloc_coherent(trans->dev, size, &phys,
cpu_addr          219 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		if (!cpu_addr)
cpu_addr          228 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	if (WARN_ON_ONCE(!cpu_addr))
cpu_addr          237 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	trans->dbg.fw_mon[trans->dbg.num_blocks].block = cpu_addr;
cpu_addr         2835 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	void *cpu_addr = (void *)trans->dbg.fw_mon[0].block, *curr_buf;
cpu_addr         2866 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		curr_buf = cpu_addr + data->prev_wr_ptr;
cpu_addr         2875 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		curr_buf = cpu_addr + data->prev_wr_ptr;
cpu_addr         2884 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 						       cpu_addr, &size,
cpu_addr         2901 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 					       cpu_addr, &size,
cpu_addr          282 drivers/of/address.c 	range->cpu_addr = of_translate_address(parser->node,
cpu_addr          291 drivers/of/address.c 		u64 pci_addr, cpu_addr, size;
cpu_addr          295 drivers/of/address.c 		cpu_addr = of_translate_address(parser->node,
cpu_addr          302 drivers/of/address.c 		    cpu_addr != range->cpu_addr + range->size)
cpu_addr          338 drivers/of/address.c 		err = pci_register_io_range(&np->fwnode, range->cpu_addr,
cpu_addr          342 drivers/of/address.c 		port = pci_address_to_pio(range->cpu_addr);
cpu_addr          350 drivers/of/address.c 		    upper_32_bits(range->cpu_addr)) {
cpu_addr          355 drivers/of/address.c 		res->start = range->cpu_addr;
cpu_addr          879 drivers/parisc/ccio-dma.c ccio_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          883 drivers/parisc/ccio-dma.c 	free_pages((unsigned long)cpu_addr, get_order(size));
cpu_addr           69 drivers/pci/controller/dwc/pcie-designware-ep.c 				  dma_addr_t cpu_addr,
cpu_addr           82 drivers/pci/controller/dwc/pcie-designware-ep.c 	ret = dw_pcie_prog_inbound_atu(pci, free_win, bar, cpu_addr,
cpu_addr          533 drivers/pci/controller/dwc/pcie-designware-host.c 	u64 cpu_addr;
cpu_addr          542 drivers/pci/controller/dwc/pcie-designware-host.c 		cpu_addr = pp->cfg0_base;
cpu_addr          547 drivers/pci/controller/dwc/pcie-designware-host.c 		cpu_addr = pp->cfg1_base;
cpu_addr          553 drivers/pci/controller/dwc/pcie-designware-host.c 				  type, cpu_addr,
cpu_addr          242 drivers/pci/controller/dwc/pcie-designware.c 					     int type, u64 cpu_addr,
cpu_addr          248 drivers/pci/controller/dwc/pcie-designware.c 				 lower_32_bits(cpu_addr));
cpu_addr          250 drivers/pci/controller/dwc/pcie-designware.c 				 upper_32_bits(cpu_addr));
cpu_addr          252 drivers/pci/controller/dwc/pcie-designware.c 				 lower_32_bits(cpu_addr + size - 1));
cpu_addr          278 drivers/pci/controller/dwc/pcie-designware.c 			       u64 cpu_addr, u64 pci_addr, u32 size)
cpu_addr          283 drivers/pci/controller/dwc/pcie-designware.c 		cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr);
cpu_addr          286 drivers/pci/controller/dwc/pcie-designware.c 		dw_pcie_prog_outbound_atu_unroll(pci, index, type, cpu_addr,
cpu_addr          294 drivers/pci/controller/dwc/pcie-designware.c 			   lower_32_bits(cpu_addr));
cpu_addr          296 drivers/pci/controller/dwc/pcie-designware.c 			   upper_32_bits(cpu_addr));
cpu_addr          298 drivers/pci/controller/dwc/pcie-designware.c 			   lower_32_bits(cpu_addr + size - 1));
cpu_addr          336 drivers/pci/controller/dwc/pcie-designware.c 					   int bar, u64 cpu_addr,
cpu_addr          343 drivers/pci/controller/dwc/pcie-designware.c 				 lower_32_bits(cpu_addr));
cpu_addr          345 drivers/pci/controller/dwc/pcie-designware.c 				 upper_32_bits(cpu_addr));
cpu_addr          381 drivers/pci/controller/dwc/pcie-designware.c 			     u64 cpu_addr, enum dw_pcie_as_type as_type)
cpu_addr          388 drivers/pci/controller/dwc/pcie-designware.c 						       cpu_addr, as_type);
cpu_addr          392 drivers/pci/controller/dwc/pcie-designware.c 	dw_pcie_writel_dbi(pci, PCIE_ATU_LOWER_TARGET, lower_32_bits(cpu_addr));
cpu_addr          393 drivers/pci/controller/dwc/pcie-designware.c 	dw_pcie_writel_dbi(pci, PCIE_ATU_UPPER_TARGET, upper_32_bits(cpu_addr));
cpu_addr          230 drivers/pci/controller/dwc/pcie-designware.h 	u64	(*cpu_addr_fixup)(struct dw_pcie *pcie, u64 cpu_addr);
cpu_addr          278 drivers/pci/controller/dwc/pcie-designware.h 			       int type, u64 cpu_addr, u64 pci_addr,
cpu_addr          281 drivers/pci/controller/dwc/pcie-designware.h 			     u64 cpu_addr, enum dw_pcie_as_type as_type);
cpu_addr          312 drivers/pci/controller/pci-rcar-gen2.c 		pci->window_addr = (unsigned long)range.cpu_addr;
cpu_addr         2193 drivers/pci/controller/pci-tegra.c 			pcie->io.start = range.cpu_addr;
cpu_addr         2194 drivers/pci/controller/pci-tegra.c 			pcie->io.end = range.cpu_addr + range.size - 1;
cpu_addr          620 drivers/pci/controller/pci-v3-semi.c 	u64 cpu_end = range->cpu_addr + range->size - 1;
cpu_addr          631 drivers/pci/controller/pci-v3-semi.c 	if (range->cpu_addr & ~V3_PCI_MAP_M_MAP_ADR) {
cpu_addr          635 drivers/pci/controller/pci-v3-semi.c 	val = ((u32)range->cpu_addr) & V3_PCI_MAP_M_MAP_ADR;
cpu_addr          685 drivers/pci/controller/pci-v3-semi.c 		range->cpu_addr, cpu_end,
cpu_addr          369 drivers/pci/controller/pci-xgene.c 				    u64 cpu_addr, u64 pci_addr)
cpu_addr          391 drivers/pci/controller/pci-xgene.c 	xgene_pcie_writel(port, offset, lower_32_bits(cpu_addr));
cpu_addr          392 drivers/pci/controller/pci-xgene.c 	xgene_pcie_writel(port, offset + 0x04, upper_32_bits(cpu_addr));
cpu_addr          494 drivers/pci/controller/pci-xgene.c 	u64 cpu_addr = range->cpu_addr;
cpu_addr          511 drivers/pci/controller/pci-xgene.c 	bar_low = pcie_bar_low_val((u32)cpu_addr, flags);
cpu_addr          517 drivers/pci/controller/pci-xgene.c 		writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
cpu_addr          527 drivers/pci/controller/pci-xgene.c 		xgene_pcie_writel(port, IBAR3L + 0x4, upper_32_bits(cpu_addr));
cpu_addr          552 drivers/pci/controller/pci-xgene.c 		u64 end = range.cpu_addr + range.size - 1;
cpu_addr          555 drivers/pci/controller/pci-xgene.c 			range.flags, range.cpu_addr, end, range.pci_addr);
cpu_addr          147 drivers/pci/controller/pcie-cadence-host.c 	u64 cpu_addr;
cpu_addr          160 drivers/pci/controller/pcie-cadence-host.c 	cpu_addr = cfg_res->start - mem_res->start;
cpu_addr          162 drivers/pci/controller/pcie-cadence-host.c 		(lower_32_bits(cpu_addr) & GENMASK(31, 8));
cpu_addr          163 drivers/pci/controller/pcie-cadence-host.c 	addr1 = upper_32_bits(cpu_addr);
cpu_addr          186 drivers/pci/controller/pcie-cadence-host.c 					      range.cpu_addr,
cpu_addr           12 drivers/pci/controller/pcie-cadence.c 				   u64 cpu_addr, u64 pci_addr, size_t size)
cpu_addr           76 drivers/pci/controller/pcie-cadence.c 	cpu_addr -= pcie->mem_res->start;
cpu_addr           78 drivers/pci/controller/pcie-cadence.c 		(lower_32_bits(cpu_addr) & GENMASK(31, 8));
cpu_addr           79 drivers/pci/controller/pcie-cadence.c 	addr1 = upper_32_bits(cpu_addr);
cpu_addr           86 drivers/pci/controller/pcie-cadence.c 						  u32 r, u64 cpu_addr)
cpu_addr          103 drivers/pci/controller/pcie-cadence.c 	cpu_addr -= pcie->mem_res->start;
cpu_addr          105 drivers/pci/controller/pcie-cadence.c 		(lower_32_bits(cpu_addr) & GENMASK(31, 8));
cpu_addr          106 drivers/pci/controller/pcie-cadence.c 	addr1 = upper_32_bits(cpu_addr);
cpu_addr          311 drivers/pci/controller/pcie-cadence.h 				   u64 cpu_addr, u64 pci_addr, size_t size);
cpu_addr          314 drivers/pci/controller/pcie-cadence.h 						  u32 r, u64 cpu_addr);
cpu_addr         1132 drivers/pci/controller/pcie-iproc.c 	u64 axi_addr = range->cpu_addr, pci_addr = range->pci_addr;
cpu_addr         1198 drivers/pci/controller/pcie-iproc.c 		if (tmp->res->start < range->cpu_addr)
cpu_addr         1202 drivers/pci/controller/pcie-iproc.c 	res->start = range->cpu_addr;
cpu_addr         1209 drivers/pci/controller/pcie-iproc.c 	entry->offset = res->start - range->cpu_addr;
cpu_addr         1283 drivers/pci/controller/pcie-iproc.c 	range.pci_addr = range.cpu_addr = msi_addr & ~(range.size - 1);
cpu_addr          469 drivers/pci/controller/pcie-mobiveil.c 			       u64 cpu_addr, u64 pci_addr, u32 type, u64 size)
cpu_addr          489 drivers/pci/controller/pcie-mobiveil.c 	mobiveil_csr_writel(pcie, lower_32_bits(cpu_addr),
cpu_addr          491 drivers/pci/controller/pcie-mobiveil.c 	mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr),
cpu_addr          506 drivers/pci/controller/pcie-mobiveil.c 			       u64 cpu_addr, u64 pci_addr, u32 type, u64 size)
cpu_addr          535 drivers/pci/controller/pcie-mobiveil.c 			    lower_32_bits(cpu_addr) & (~AXI_WINDOW_ALIGN_MASK),
cpu_addr          537 drivers/pci/controller/pcie-mobiveil.c 	mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr),
cpu_addr         1026 drivers/pci/controller/pcie-rcar.c 	u64 cpu_addr = range->cpu_addr;
cpu_addr         1027 drivers/pci/controller/pcie-rcar.c 	u64 cpu_end = range->cpu_addr + range->size;
cpu_addr         1041 drivers/pci/controller/pcie-rcar.c 	if (cpu_addr > 0) {
cpu_addr         1042 drivers/pci/controller/pcie-rcar.c 		unsigned long nr_zeros = __ffs64(cpu_addr);
cpu_addr         1055 drivers/pci/controller/pcie-rcar.c 	while (cpu_addr < cpu_end) {
cpu_addr         1062 drivers/pci/controller/pcie-rcar.c 		rcar_pci_write_reg(pcie, lower_32_bits(cpu_addr), PCIELAR(idx));
cpu_addr         1068 drivers/pci/controller/pcie-rcar.c 		rcar_pci_write_reg(pcie, upper_32_bits(cpu_addr),
cpu_addr         1073 drivers/pci/controller/pcie-rcar.c 		cpu_addr += size;
cpu_addr         1099 drivers/pci/controller/pcie-rcar.c 		u64 end = range.cpu_addr + range.size - 1;
cpu_addr         1102 drivers/pci/controller/pcie-rcar.c 			range.flags, range.cpu_addr, end, range.pci_addr);
cpu_addr           70 drivers/pci/controller/pcie-rockchip-ep.c 					 u32 r, u32 type, u64 cpu_addr,
cpu_addr           82 drivers/pci/controller/pcie-rockchip-ep.c 	cpu_addr -= rockchip->mem_res->start;
cpu_addr           85 drivers/pci/controller/pcie-rockchip-ep.c 		(lower_32_bits(cpu_addr) & PCIE_CORE_OB_REGION_ADDR0_LO_ADDR);
cpu_addr           86 drivers/pci/controller/pcie-rockchip-ep.c 	addr1 = upper_32_bits(is_nor_msg ? cpu_addr : pci_addr);
cpu_addr          112 drivers/pci/controller/pcie-rockchip-ep.c 		    (lower_32_bits(cpu_addr) &
cpu_addr          114 drivers/pci/controller/pcie-rockchip-ep.c 		addr1 = upper_32_bits(cpu_addr);
cpu_addr          324 drivers/pci/controller/vmd.c 		    void *cpu_addr, dma_addr_t addr, size_t size,
cpu_addr          327 drivers/pci/controller/vmd.c 	return dma_mmap_attrs(to_vmd_dev(dev), vma, cpu_addr, addr, size,
cpu_addr          332 drivers/pci/controller/vmd.c 			   void *cpu_addr, dma_addr_t addr, size_t size,
cpu_addr          335 drivers/pci/controller/vmd.c 	return dma_get_sgtable_attrs(to_vmd_dev(dev), sgt, cpu_addr, addr, size,
cpu_addr          307 drivers/pci/of.c 			 range_type, range.cpu_addr,
cpu_addr          308 drivers/pci/of.c 			 range.cpu_addr + range.size - 1, range.pci_addr);
cpu_addr          314 drivers/pci/of.c 		if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
cpu_addr          337 drivers/pci/of.c 			*io_base = range.cpu_addr;
cpu_addr           48 drivers/remoteproc/da8xx_remoteproc.c 	void __iomem *cpu_addr;
cpu_addr          215 drivers/remoteproc/da8xx_remoteproc.c 		drproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
cpu_addr          216 drivers/remoteproc/da8xx_remoteproc.c 		if (IS_ERR(drproc->mem[i].cpu_addr)) {
cpu_addr          219 drivers/remoteproc/da8xx_remoteproc.c 			return PTR_ERR(drproc->mem[i].cpu_addr);
cpu_addr          228 drivers/remoteproc/da8xx_remoteproc.c 			drproc->mem[i].size, drproc->mem[i].cpu_addr,
cpu_addr           55 drivers/remoteproc/imx_rproc.c 	void __iomem *cpu_addr;
cpu_addr          233 drivers/remoteproc/imx_rproc.c 			va = (__force void *)(priv->mem[i].cpu_addr + offset);
cpu_addr          267 drivers/remoteproc/imx_rproc.c 		priv->mem[b].cpu_addr = devm_ioremap(&pdev->dev,
cpu_addr          269 drivers/remoteproc/imx_rproc.c 		if (!priv->mem[b].cpu_addr) {
cpu_addr          298 drivers/remoteproc/imx_rproc.c 		priv->mem[b].cpu_addr = devm_ioremap_resource(&pdev->dev, &res);
cpu_addr          299 drivers/remoteproc/imx_rproc.c 		if (IS_ERR(priv->mem[b].cpu_addr)) {
cpu_addr          301 drivers/remoteproc/imx_rproc.c 			err = PTR_ERR(priv->mem[b].cpu_addr);
cpu_addr           35 drivers/remoteproc/keystone_remoteproc.c 	void __iomem *cpu_addr;
cpu_addr          271 drivers/remoteproc/keystone_remoteproc.c 				va = ksproc->mem[i].cpu_addr + offset;
cpu_addr          279 drivers/remoteproc/keystone_remoteproc.c 				va = ksproc->mem[i].cpu_addr + offset;
cpu_addr          313 drivers/remoteproc/keystone_remoteproc.c 		ksproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
cpu_addr          314 drivers/remoteproc/keystone_remoteproc.c 		if (IS_ERR(ksproc->mem[i].cpu_addr)) {
cpu_addr          317 drivers/remoteproc/keystone_remoteproc.c 			return PTR_ERR(ksproc->mem[i].cpu_addr);
cpu_addr          325 drivers/remoteproc/keystone_remoteproc.c 		memset((__force void *)ksproc->mem[i].cpu_addr, 0,
cpu_addr          144 drivers/remoteproc/remoteproc_core.c phys_addr_t rproc_va_to_pa(void *cpu_addr)
cpu_addr          151 drivers/remoteproc/remoteproc_core.c 	if (is_vmalloc_addr(cpu_addr)) {
cpu_addr          152 drivers/remoteproc/remoteproc_core.c 		return page_to_phys(vmalloc_to_page(cpu_addr)) +
cpu_addr          153 drivers/remoteproc/remoteproc_core.c 				    offset_in_page(cpu_addr);
cpu_addr          156 drivers/remoteproc/remoteproc_core.c 	WARN_ON(!virt_addr_valid(cpu_addr));
cpu_addr          157 drivers/remoteproc/remoteproc_core.c 	return virt_to_phys(cpu_addr);
cpu_addr           54 drivers/remoteproc/remoteproc_internal.h phys_addr_t rproc_va_to_pa(void *cpu_addr);
cpu_addr          144 drivers/remoteproc/st_slim_rproc.c 	fw_rev = readl(slim_rproc->mem[ST_SLIM_DMEM].cpu_addr +
cpu_addr          189 drivers/remoteproc/st_slim_rproc.c 			va = (__force void *)slim_rproc->mem[i].cpu_addr;
cpu_addr          253 drivers/remoteproc/st_slim_rproc.c 		slim_rproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
cpu_addr          254 drivers/remoteproc/st_slim_rproc.c 		if (IS_ERR(slim_rproc->mem[i].cpu_addr)) {
cpu_addr          256 drivers/remoteproc/st_slim_rproc.c 			err = PTR_ERR(slim_rproc->mem[i].cpu_addr);
cpu_addr           45 drivers/remoteproc/stm32_rproc.c 	void __iomem *cpu_addr;
cpu_addr           35 drivers/remoteproc/wkup_m3_rproc.c 	void __iomem *cpu_addr;
cpu_addr           98 drivers/remoteproc/wkup_m3_rproc.c 			va = (__force void *)(wkupm3->mem[i].cpu_addr + offset);
cpu_addr          171 drivers/remoteproc/wkup_m3_rproc.c 		wkupm3->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
cpu_addr          172 drivers/remoteproc/wkup_m3_rproc.c 		if (IS_ERR(wkupm3->mem[i].cpu_addr)) {
cpu_addr          175 drivers/remoteproc/wkup_m3_rproc.c 			ret = PTR_ERR(wkupm3->mem[i].cpu_addr);
cpu_addr          199 drivers/rpmsg/virtio_rpmsg_bus.c rpmsg_sg_init(struct scatterlist *sg, void *cpu_addr, unsigned int len)
cpu_addr          201 drivers/rpmsg/virtio_rpmsg_bus.c 	if (is_vmalloc_addr(cpu_addr)) {
cpu_addr          203 drivers/rpmsg/virtio_rpmsg_bus.c 		sg_set_page(sg, vmalloc_to_page(cpu_addr), len,
cpu_addr          204 drivers/rpmsg/virtio_rpmsg_bus.c 			    offset_in_page(cpu_addr));
cpu_addr          206 drivers/rpmsg/virtio_rpmsg_bus.c 		WARN_ON(!virt_addr_valid(cpu_addr));
cpu_addr          207 drivers/rpmsg/virtio_rpmsg_bus.c 		sg_init_one(sg, cpu_addr, len);
cpu_addr          935 drivers/rpmsg/virtio_rpmsg_bus.c 		void *cpu_addr = vrp->rbufs + i * vrp->buf_size;
cpu_addr          937 drivers/rpmsg/virtio_rpmsg_bus.c 		rpmsg_sg_init(&sg, cpu_addr, vrp->buf_size);
cpu_addr          939 drivers/rpmsg/virtio_rpmsg_bus.c 		err = virtqueue_add_inbuf(vrp->rvq, &sg, 1, cpu_addr,
cpu_addr          854 drivers/s390/cio/css.c 		css->global_pgid.pgid_high.cpu_addr = stap();
cpu_addr         1102 drivers/s390/cio/css.c 	void *cpu_addr;
cpu_addr         1110 drivers/s390/cio/css.c 		cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr,
cpu_addr         1112 drivers/s390/cio/css.c 		if (!cpu_addr)
cpu_addr         1114 drivers/s390/cio/css.c 		gen_pool_add_virt(gp_dma, (unsigned long) cpu_addr,
cpu_addr         1170 drivers/s390/cio/css.c void cio_gp_dma_free(struct gen_pool *gp_dma, void *cpu_addr, size_t size)
cpu_addr         1172 drivers/s390/cio/css.c 	if (!cpu_addr)
cpu_addr         1174 drivers/s390/cio/css.c 	memset(cpu_addr, 0, size);
cpu_addr         1175 drivers/s390/cio/css.c 	gen_pool_free(gp_dma, (unsigned long) cpu_addr, size);
cpu_addr         1190 drivers/s390/cio/css.c void cio_dma_free(void *cpu_addr, size_t size)
cpu_addr         1192 drivers/s390/cio/css.c 	cio_gp_dma_free(cio_dma_pool, cpu_addr, size);
cpu_addr           55 drivers/s390/cio/css.h 		__u32 cpu_addr	: 16;	/* CPU address */
cpu_addr          724 drivers/s390/cio/device_ops.c void ccw_device_dma_free(struct ccw_device *cdev, void *cpu_addr, size_t size)
cpu_addr          726 drivers/s390/cio/device_ops.c 	cio_gp_dma_free(cdev->private->dma_pool, cpu_addr, size);
cpu_addr          220 drivers/s390/net/ism_drv.c 			  dmb->cpu_addr, dmb->dma_addr);
cpu_addr          242 drivers/s390/net/ism_drv.c 	dmb->cpu_addr = dma_alloc_coherent(&ism->pdev->dev, dmb->dmb_len,
cpu_addr          245 drivers/s390/net/ism_drv.c 	if (!dmb->cpu_addr)
cpu_addr          248 drivers/s390/net/ism_drv.c 	return dmb->cpu_addr ? 0 : -ENOMEM;
cpu_addr          520 drivers/scsi/3w-9xxx.c 	unsigned long *cpu_addr;
cpu_addr          523 drivers/scsi/3w-9xxx.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev,
cpu_addr          525 drivers/scsi/3w-9xxx.c 	if (!cpu_addr) {
cpu_addr          530 drivers/scsi/3w-9xxx.c 	if ((unsigned long)cpu_addr % (TW_ALIGNMENT_9000)) {
cpu_addr          533 drivers/scsi/3w-9xxx.c 				cpu_addr, dma_handle);
cpu_addr          537 drivers/scsi/3w-9xxx.c 	memset(cpu_addr, 0, size*TW_Q_LENGTH);
cpu_addr          543 drivers/scsi/3w-9xxx.c 			tw_dev->command_packet_virt[i] = (TW_Command_Full *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          547 drivers/scsi/3w-9xxx.c 			tw_dev->generic_buffer_virt[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          641 drivers/scsi/3w-9xxx.c 	unsigned long *cpu_addr, data_buffer_length_adjusted = 0, flags = 0;
cpu_addr          679 drivers/scsi/3w-9xxx.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, &dma_handle, GFP_KERNEL);
cpu_addr          680 drivers/scsi/3w-9xxx.c 	if (!cpu_addr) {
cpu_addr          685 drivers/scsi/3w-9xxx.c 	tw_ioctl = (TW_Ioctl_Buf_Apache *)cpu_addr;
cpu_addr          874 drivers/scsi/3w-9xxx.c 	dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, cpu_addr, dma_handle);
cpu_addr          646 drivers/scsi/3w-sas.c 	unsigned long *cpu_addr;
cpu_addr          649 drivers/scsi/3w-sas.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev,
cpu_addr          652 drivers/scsi/3w-sas.c 	if (!cpu_addr) {
cpu_addr          661 drivers/scsi/3w-sas.c 			tw_dev->command_packet_virt[i] = (TW_Command_Full *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          665 drivers/scsi/3w-sas.c 			tw_dev->generic_buffer_virt[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          669 drivers/scsi/3w-sas.c 			tw_dev->sense_buffer_virt[i] = (TW_Command_Apache_Header *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          719 drivers/scsi/3w-sas.c 	unsigned long *cpu_addr, data_buffer_length_adjusted = 0, flags = 0;
cpu_addr          752 drivers/scsi/3w-sas.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, &dma_handle, GFP_KERNEL);
cpu_addr          753 drivers/scsi/3w-sas.c 	if (!cpu_addr) {
cpu_addr          758 drivers/scsi/3w-sas.c 	tw_ioctl = (TW_Ioctl_Buf_Apache *)cpu_addr;
cpu_addr          823 drivers/scsi/3w-sas.c 	dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_Ioctl_Buf_Apache) - 1, cpu_addr, dma_handle);
cpu_addr          833 drivers/scsi/3w-xxxx.c 	unsigned long *cpu_addr = NULL;
cpu_addr          837 drivers/scsi/3w-xxxx.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev,
cpu_addr          839 drivers/scsi/3w-xxxx.c 	if (cpu_addr == NULL) {
cpu_addr          844 drivers/scsi/3w-xxxx.c 	if ((unsigned long)cpu_addr % (tw_dev->tw_pci_dev->device == TW_DEVICE_ID ? TW_ALIGNMENT_6000 : TW_ALIGNMENT_7000)) {
cpu_addr          847 drivers/scsi/3w-xxxx.c 				cpu_addr, dma_handle);
cpu_addr          851 drivers/scsi/3w-xxxx.c 	memset(cpu_addr, 0, size*TW_Q_LENGTH);
cpu_addr          857 drivers/scsi/3w-xxxx.c 			tw_dev->command_packet_virtual_address[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          861 drivers/scsi/3w-xxxx.c 			tw_dev->alignment_virtual_address[i] = (unsigned long *)((unsigned char *)cpu_addr + (i*size));
cpu_addr          882 drivers/scsi/3w-xxxx.c 	unsigned long *cpu_addr;
cpu_addr          913 drivers/scsi/3w-xxxx.c 	cpu_addr = dma_alloc_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_New_Ioctl) - 1, &dma_handle, GFP_KERNEL);
cpu_addr          914 drivers/scsi/3w-xxxx.c 	if (cpu_addr == NULL) {
cpu_addr          919 drivers/scsi/3w-xxxx.c 	tw_ioctl = (TW_New_Ioctl *)cpu_addr;
cpu_addr         1023 drivers/scsi/3w-xxxx.c 	dma_free_coherent(&tw_dev->tw_pci_dev->dev, data_buffer_length_adjusted+sizeof(TW_New_Ioctl) - 1, cpu_addr, dma_handle);
cpu_addr          299 drivers/staging/mt7621-pci/pci-mt7621.c 			ioremap(range.cpu_addr, range.size);
cpu_addr          481 drivers/tty/serial/msm_serial.c 	void *cpu_addr;
cpu_addr          485 drivers/tty/serial/msm_serial.c 	cpu_addr = &xmit->buf[xmit->tail];
cpu_addr          487 drivers/tty/serial/msm_serial.c 	dma->phys = dma_map_single(port->dev, cpu_addr, count, dma->dir);
cpu_addr           30 drivers/usb/host/xhci-dbgcap.c 		      void *cpu_addr, dma_addr_t dma_handle)
cpu_addr           32 drivers/usb/host/xhci-dbgcap.c 	if (cpu_addr)
cpu_addr           34 drivers/usb/host/xhci-dbgcap.c 				  size, cpu_addr, dma_handle);
cpu_addr          343 drivers/virtio/virtio_ring.c 				   void *cpu_addr, size_t size,
cpu_addr          347 drivers/virtio/virtio_ring.c 		return (dma_addr_t)virt_to_phys(cpu_addr);
cpu_addr          350 drivers/virtio/virtio_ring.c 			      cpu_addr, size, direction);
cpu_addr           71 include/linux/dma-direct.h void dma_direct_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr           75 include/linux/dma-direct.h void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          163 include/linux/dma-mapping.h 			    void *cpu_addr, size_t size, int *ret);
cpu_addr          167 include/linux/dma-mapping.h int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr,
cpu_addr          187 include/linux/dma-mapping.h 						void *cpu_addr, size_t size,
cpu_addr          447 include/linux/dma-mapping.h void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          456 include/linux/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          459 include/linux/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          524 include/linux/dma-mapping.h static void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          542 include/linux/dma-mapping.h 		struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr,
cpu_addr          548 include/linux/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          625 include/linux/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          628 include/linux/dma-mapping.h struct page **dma_common_find_pages(void *cpu_addr);
cpu_addr          634 include/linux/dma-mapping.h void dma_common_free_remap(void *cpu_addr, size_t size);
cpu_addr          641 include/linux/dma-mapping.h dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr,
cpu_addr          653 include/linux/dma-mapping.h 		void *cpu_addr, dma_addr_t dma_handle)
cpu_addr          655 include/linux/dma-mapping.h 	return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0);
cpu_addr          793 include/linux/dma-mapping.h 			       void *cpu_addr, dma_addr_t dma_addr)
cpu_addr          795 include/linux/dma-mapping.h 	return dma_free_attrs(dev, size, cpu_addr, dma_addr,
cpu_addr          801 include/linux/dma-mapping.h 			      void *cpu_addr, dma_addr_t dma_addr,
cpu_addr          804 include/linux/dma-mapping.h 	return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size,
cpu_addr           42 include/linux/dma-noncoherent.h void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
cpu_addr           44 include/linux/dma-noncoherent.h long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
cpu_addr           20 include/linux/of_address.h 	u64 cpu_addr;
cpu_addr           27 include/linux/remoteproc/st_slim_rproc.h 	void __iomem *cpu_addr;
cpu_addr          114 include/media/drv-intf/saa7146.h 	__le32		*cpu_addr;
cpu_addr           15 include/media/drv-intf/saa7146_vv.h 	dev->d_rps0.cpu_addr[ count++ ] = cpu_to_le32(x); \
cpu_addr           19 include/media/drv-intf/saa7146_vv.h 	dev->d_rps1.cpu_addr[ count++ ] = cpu_to_le32(x); \
cpu_addr           32 include/net/smc.h 	void *cpu_addr;
cpu_addr         3945 include/rdma/ib_verbs.h 				    void *cpu_addr, size_t size,
cpu_addr         3948 include/rdma/ib_verbs.h 	return dma_map_single(dev->dma_device, cpu_addr, size, direction);
cpu_addr         4105 include/rdma/ib_verbs.h 					size_t size, void *cpu_addr,
cpu_addr         4108 include/rdma/ib_verbs.h 	dma_free_coherent(dev->dma_device, size, cpu_addr, dma_handle);
cpu_addr           15 include/xen/arm/page-coherent.h 		void *cpu_addr, dma_addr_t dma_handle, unsigned long attrs)
cpu_addr           17 include/xen/arm/page-coherent.h 	dma_direct_free(hwdev, size, cpu_addr, dma_handle, attrs);
cpu_addr          184 kernel/dma/direct.c void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          192 kernel/dma/direct.c 		__dma_direct_free_pages(dev, size, cpu_addr);
cpu_addr          197 kernel/dma/direct.c 		set_memory_encrypted((unsigned long)cpu_addr, 1 << page_order);
cpu_addr          201 kernel/dma/direct.c 		cpu_addr = cached_kernel_address(cpu_addr);
cpu_addr          202 kernel/dma/direct.c 	__dma_direct_free_pages(dev, size, virt_to_page(cpu_addr));
cpu_addr          215 kernel/dma/direct.c 		void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs)
cpu_addr          219 kernel/dma/direct.c 		arch_dma_free(dev, size, cpu_addr, dma_addr, attrs);
cpu_addr          221 kernel/dma/direct.c 		dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs);
cpu_addr            8 kernel/dma/dummy.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          112 kernel/dma/mapping.c 		 void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          125 kernel/dma/mapping.c 		pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr);
cpu_addr          130 kernel/dma/mapping.c 		page = virt_to_page(cpu_addr);
cpu_addr          151 kernel/dma/mapping.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          157 kernel/dma/mapping.c 		return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr,
cpu_addr          161 kernel/dma/mapping.c 	return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs);
cpu_addr          190 kernel/dma/mapping.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          202 kernel/dma/mapping.c 	if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
cpu_addr          213 kernel/dma/mapping.c 		pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr);
cpu_addr          217 kernel/dma/mapping.c 		pfn = page_to_pfn(virt_to_page(cpu_addr));
cpu_addr          262 kernel/dma/mapping.c 		void *cpu_addr, dma_addr_t dma_addr, size_t size,
cpu_addr          268 kernel/dma/mapping.c 		return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size,
cpu_addr          272 kernel/dma/mapping.c 	return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
cpu_addr          301 kernel/dma/mapping.c 	void *cpu_addr;
cpu_addr          305 kernel/dma/mapping.c 	if (dma_alloc_from_dev_coherent(dev, size, dma_handle, &cpu_addr))
cpu_addr          306 kernel/dma/mapping.c 		return cpu_addr;
cpu_addr          312 kernel/dma/mapping.c 		cpu_addr = dma_direct_alloc(dev, size, dma_handle, flag, attrs);
cpu_addr          314 kernel/dma/mapping.c 		cpu_addr = ops->alloc(dev, size, dma_handle, flag, attrs);
cpu_addr          318 kernel/dma/mapping.c 	debug_dma_alloc_coherent(dev, size, *dma_handle, cpu_addr);
cpu_addr          319 kernel/dma/mapping.c 	return cpu_addr;
cpu_addr          323 kernel/dma/mapping.c void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr,
cpu_addr          328 kernel/dma/mapping.c 	if (dma_release_from_dev_coherent(dev, get_order(size), cpu_addr))
cpu_addr          339 kernel/dma/mapping.c 	if (!cpu_addr)
cpu_addr          342 kernel/dma/mapping.c 	debug_dma_free_coherent(dev, size, cpu_addr, dma_handle);
cpu_addr          344 kernel/dma/mapping.c 		dma_direct_free(dev, size, cpu_addr, dma_handle, attrs);
cpu_addr          346 kernel/dma/mapping.c 		ops->free(dev, size, cpu_addr, dma_handle, attrs);
cpu_addr           14 kernel/dma/remap.c struct page **dma_common_find_pages(void *cpu_addr)
cpu_addr           16 kernel/dma/remap.c 	struct vm_struct *area = find_vm_area(cpu_addr);
cpu_addr           88 kernel/dma/remap.c void dma_common_free_remap(void *cpu_addr, size_t size)
cpu_addr           90 kernel/dma/remap.c 	struct vm_struct *area = find_vm_area(cpu_addr);
cpu_addr           93 kernel/dma/remap.c 		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr);
cpu_addr           97 kernel/dma/remap.c 	unmap_kernel_range((unsigned long)cpu_addr, PAGE_ALIGN(size));
cpu_addr           98 kernel/dma/remap.c 	vunmap(cpu_addr);
cpu_addr          263 kernel/dma/remap.c long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
cpu_addr           23 kernel/dma/virt.c 			  void *cpu_addr, dma_addr_t dma_addr,
cpu_addr           26 kernel/dma/virt.c 	free_pages((unsigned long)cpu_addr, get_order(size));
cpu_addr          231 net/smc/smc_cdc.c 	base = (char *)conn->rmb_desc->cpu_addr + conn->rx_off;
cpu_addr          334 net/smc/smc_cdc.c 	data_cdc = (struct smcd_cdc_msg *)conn->rmb_desc->cpu_addr;
cpu_addr          393 net/smc/smc_core.c 		kfree(buf_desc->cpu_addr);
cpu_addr          760 net/smc/smc_core.c 	buf_desc->cpu_addr = (void *)page_address(buf_desc->pages);
cpu_addr          771 net/smc/smc_core.c 		   buf_desc->cpu_addr, bufsize);
cpu_addr          819 net/smc/smc_core.c 		buf_desc->pages = virt_to_page(buf_desc->cpu_addr);
cpu_addr          823 net/smc/smc_core.c 		buf_desc->cpu_addr = kzalloc(bufsize, GFP_KERNEL |
cpu_addr          826 net/smc/smc_core.c 		if (!buf_desc->cpu_addr) {
cpu_addr          869 net/smc/smc_core.c 			memset(buf_desc->cpu_addr, 0, bufsize);
cpu_addr          149 net/smc/smc_core.h 	void			*cpu_addr;	/* virtual address of buffer */
cpu_addr          153 net/smc/smc_ism.c 	dmb.cpu_addr = dmb_desc->cpu_addr;
cpu_addr          174 net/smc/smc_ism.c 		dmb_desc->cpu_addr = dmb.cpu_addr;
cpu_addr          164 net/smc/smc_rx.c 	partial.offset = src - (char *)smc->conn.rmb_desc->cpu_addr;
cpu_addr          311 net/smc/smc_rx.c 	rcvbuf_base = conn->rx_off + conn->rmb_desc->cpu_addr;
cpu_addr          181 net/smc/smc_tx.c 		sndbuf_base = conn->sndbuf_desc->cpu_addr;
cpu_addr          370 net/smc/smc_tx.c 			void *data = conn->sndbuf_desc->cpu_addr + src_off;