__pa               59 arch/alpha/include/asm/mmzone.h #define kvaddr_to_nid(kaddr)	pa_to_nid(__pa(kaddr))
__pa               73 arch/alpha/include/asm/mmzone.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa              109 arch/alpha/include/asm/mmzone.h #define virt_addr_valid(kaddr)	pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
__pa               87 arch/alpha/include/asm/page.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa               90 arch/alpha/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              232 arch/alpha/kernel/core_irongate.c 		if (initrd_end && __pa(initrd_end) > pci_mem) {
__pa              236 arch/alpha/kernel/core_irongate.c 			memblock_free(__pa(initrd_start), PAGE_ALIGN(size));
__pa              276 arch/alpha/kernel/pci_iommu.c 	paddr = __pa(cpu_addr);
__pa              514 arch/alpha/kernel/pci_iommu.c #define SG_ENT_PHYS_ADDRESS(SG) __pa(SG_ENT_VIRT_ADDRESS(SG))
__pa              297 arch/alpha/kernel/setup.c 	if (!start || __pa(start) + size > mem_limit) {
__pa               83 arch/arc/include/asm/page.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa              120 arch/arc/mm/init.c 			 __pa(_end) - CONFIG_LINUX_LINK_BASE);
__pa              397 arch/arm/include/asm/cacheflush.h 	outer_clean_range(__pa(_p), __pa(_p + size));
__pa              419 arch/arm/include/asm/cacheflush.h 		outer_flush_range(__pa(_p), __pa(_p + size));
__pa               69 arch/arm/include/asm/kvm_mmu.h #define kvm_mk_pmd(ptep)	__pmd(__pa(ptep) | PMD_TYPE_TABLE)
__pa               70 arch/arm/include/asm/kvm_mmu.h #define kvm_mk_pud(pmdp)	__pud(__pa(pmdp) | PMD_TYPE_TABLE)
__pa               38 arch/arm/include/asm/pgalloc.h 	set_pud(pud, __pud(__pa(pmd) | PMD_TYPE_TABLE));
__pa              134 arch/arm/include/asm/pgalloc.h 	__pmd_populate(pmdp, __pa(ptep), _PAGE_KERNEL_TABLE);
__pa             1002 arch/arm/kernel/setup.c 		unsigned long long lowmem_max = __pa(high_memory - 1) + 1;
__pa               22 arch/arm/mach-highbank/pm.c 	return psci_ops.cpu_suspend(HIGHBANK_SUSPEND_PARAM, __pa(cpu_resume));
__pa               95 arch/arm/mach-integrator/core.c 	memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET);
__pa              330 arch/arm/mach-omap2/control.c 			(u32) __pa(omap3_secure_ram_storage);
__pa               50 arch/arm/mach-omap2/omap-secure.c 	outer_clean_range(__pa(param), __pa(param + 5));
__pa               51 arch/arm/mach-omap2/omap-secure.c 	ret = omap_smc2(idx, flag, __pa(param));
__pa               82 arch/arm/mach-omap2/omap-secure.c 	param[1] = __pa(addr);	/* Physical address for saving */
__pa               87 arch/arm/mach-omap2/omap-secure.c 	ret = save_secure_ram_context(__pa(param));
__pa              125 arch/arm/mach-omap2/omap-secure.c 	outer_clean_range(__pa(param), __pa(param + 5));
__pa              126 arch/arm/mach-omap2/omap-secure.c 	ret = omap_smc3(idx, process, flag, __pa(param));
__pa              295 arch/arm/mach-omap2/omap-smp.c 	if ((addr >= __pa(PAGE_OFFSET)) && (addr <= __pa(__bss_start)))
__pa               37 arch/arm/mach-shmobile/smp-emev2.c 		iowrite32(__pa(shmobile_boot_vector), smu + SMU_GENERAL_REG0);
__pa               40 arch/arm/mach-shmobile/smp-r8a7779.c 	__raw_writel(__pa(shmobile_boot_vector), AVECR);
__pa               43 arch/arm/mach-shmobile/smp-sh73a0.c 	__raw_writel(__pa(shmobile_boot_vector), SBAR);
__pa               54 arch/arm/mach-zynq/common.c 	if (!__pa(PAGE_OFFSET))
__pa               55 arch/arm/mach-zynq/common.c 		memblock_reserve(__pa(PAGE_OFFSET), 0x80000);
__pa               44 arch/arm/mach-zynq/platsmp.c 			if (__pa(PAGE_OFFSET)) {
__pa               68 arch/arm/mach-zynq/platsmp.c 			if (__pa(PAGE_OFFSET))
__pa              290 arch/arm/mm/dma-mapping.c 			outer_flush_range(__pa(ptr), __pa(ptr) + size);
__pa              267 arch/arm/mm/init.c 	memblock_reserve(__pa(KERNEL_START), KERNEL_END - KERNEL_START);
__pa              342 arch/arm/mm/init.c 	pg = PAGE_ALIGN(__pa(start_pg));
__pa              343 arch/arm/mm/init.c 	pgend = __pa(end_pg) & PAGE_MASK;
__pa              155 arch/arm/mm/mmap.c 	if (addr + size > __pa(high_memory - 1) + 1)
__pa              745 arch/arm/mm/mmu.c 		__pmd_populate(pmd, __pa(pte), prot);
__pa             1318 arch/arm/mm/mmu.c 	memblock_reserve(__pa(swapper_pg_dir), SWAPPER_PG_DIR_SIZE);
__pa             1325 arch/arm/mm/mmu.c 	memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET);
__pa             1453 arch/arm/mm/mmu.c 	phys_addr_t kernel_x_start = round_down(__pa(KERNEL_START), SECTION_SIZE);
__pa             1454 arch/arm/mm/mmu.c 	phys_addr_t kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE);
__pa             1544 arch/arm/mm/mmu.c 	lpae_pgtables_remap = (pgtables_remap *)(unsigned long)__pa(lpae_pgtables_remap_asm);
__pa             1545 arch/arm/mm/mmu.c 	pa_pgd = __pa(swapper_pg_dir);
__pa              259 arch/arm/mm/pmsa-v7.c 	num = allocate_region(CONFIG_XIP_PHYS_ADDR, __pa(_exiprom) - CONFIG_XIP_PHYS_ADDR,
__pa              254 arch/arm/mm/pmsa-v8.c 	subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END));
__pa              255 arch/arm/mm/pmsa-v8.c 	subtract_range(io, ARRAY_SIZE(io),  __pa(KERNEL_START), __pa(KERNEL_END));
__pa              259 arch/arm/mm/pmsa-v8.c 	subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
__pa              260 arch/arm/mm/pmsa-v8.c 	subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
__pa              276 arch/arm/mm/pmsa-v8.c 	err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
__pa              279 arch/arm/mm/pmsa-v8.c 	err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END));
__pa              171 arch/arm64/include/asm/kvm_mmu.h 	__pmd(__phys_to_pmd_val(__pa(ptep)) | PMD_TYPE_TABLE)
__pa              173 arch/arm64/include/asm/kvm_mmu.h 	__pud(__phys_to_pud_val(__pa(pmdp)) | PMD_TYPE_TABLE)
__pa              175 arch/arm64/include/asm/kvm_mmu.h 	__pgd(__phys_to_pgd_val(__pa(pudp)) | PUD_TYPE_TABLE)
__pa              399 arch/arm64/include/asm/kvm_mmu.h 	pgd_addr = __phys_to_pgd_val(__pa(hyp_pgd));
__pa              410 arch/arm64/include/asm/kvm_mmu.h 	pgd_addr = __phys_to_pgd_val(__pa(boot_hyp_pgd));
__pa               54 arch/arm64/include/asm/pgalloc.h 	__pud_populate(pudp, __pa(pmdp), PMD_TYPE_TABLE);
__pa               83 arch/arm64/include/asm/pgalloc.h 	__pgd_populate(pgdp, __pa(pudp), PUD_TYPE_TABLE);
__pa              111 arch/arm64/include/asm/pgalloc.h 	__pmd_populate(pmdp, __pa(ptep), PMD_TYPE_TABLE);
__pa              477 arch/arm64/mm/init.c 	pg = (unsigned long)PAGE_ALIGN(__pa(start_pg));
__pa              478 arch/arm64/mm/init.c 	pgend = (unsigned long)__pa(end_pg) & PAGE_MASK;
__pa               37 arch/arm64/mm/kasan_init.c 					      __pa(MAX_DMA_ADDRESS),
__pa               42 arch/arm64/mm/kasan_init.c 		      __pa(MAX_DMA_ADDRESS));
__pa               44 arch/arm64/mm/kasan_init.c 	return __pa(p);
__pa               50 arch/arm64/mm/kasan_init.c 						__pa(MAX_DMA_ADDRESS),
__pa               55 arch/arm64/mm/kasan_init.c 		      __pa(MAX_DMA_ADDRESS));
__pa               57 arch/arm64/mm/kasan_init.c 	return __pa(p);
__pa              371 arch/arm64/mm/mmu.c 	return __pa(ptr);
__pa              764 arch/arm64/mm/mmu.c 			pmd_set_huge(pmdp, __pa(p), __pgprot(PROT_SECT_NORMAL));
__pa              160 arch/arm64/mm/numa.c 			__pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid);
__pa              165 arch/arm64/mm/numa.c 	memblock_free_early(__pa(ptr), size);
__pa              264 arch/arm64/mm/numa.c 	memblock_free(__pa(numa_distance), size);
__pa               54 arch/c6x/mm/init.c 		__pa(PAGE_OFFSET) >> PAGE_SHIFT;
__pa               18 arch/csky/include/asm/mmu_context.h 	setup_pgd(__pa(pgd), false)
__pa               21 arch/csky/include/asm/mmu_context.h 	setup_pgd(__pa(pgd), true)
__pa               37 arch/csky/include/asm/page.h #define virt_to_pfn(kaddr)      (__pa(kaddr) >> PAGE_SHIFT)
__pa               82 arch/csky/include/asm/page.h #define __pa_symbol(x)	__pa(RELOC_HIDE((unsigned long)(x), 0))
__pa               17 arch/csky/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(pte)));
__pa               23 arch/csky/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(page_address(pte))));
__pa              151 arch/csky/include/asm/pgtable.h 	return pmd_val(pmd) == __pa(invalid_pte_table);
__pa              158 arch/csky/include/asm/pgtable.h 	return (pmd_val(pmd) != __pa(invalid_pte_table));
__pa              163 arch/csky/include/asm/pgtable.h 	pmd_val(*p) = (__pa(invalid_pte_table));
__pa               32 arch/csky/kernel/setup.c 	memblock_reserve(__pa(_stext), _end - _stext);
__pa               34 arch/csky/kernel/setup.c 	memblock_reserve(__pa(initrd_start), initrd_end - initrd_start);
__pa              149 arch/csky/mm/highmem.c 					set_pmd(pmd, __pmd(__pa(pte)));
__pa               89 arch/csky/mm/init.c 		p[i] = __pa(invalid_pte_table);
__pa               91 arch/h8300/kernel/setup.c 	memblock_reserve(__pa(_stext), _end - _stext);
__pa               51 arch/hexagon/include/asm/io.h 	return __pa(address);
__pa               93 arch/hexagon/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
__pa              100 arch/hexagon/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              130 arch/hexagon/include/asm/page.h #define virt_to_pfn(kaddr)      (__pa(kaddr) >> PAGE_SHIFT)
__pa               39 arch/hexagon/include/asm/pgalloc.h 	mm->context.ptbase = __pa(pgd);
__pa               82 arch/hexagon/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
__pa               90 arch/hexagon/include/asm/pgalloc.h 	set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
__pa               71 arch/hexagon/mm/init.c 	init_mm.context.ptbase = __pa(init_mm.pgd);
__pa              193 arch/ia64/include/asm/mmu_context.h 	ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd));
__pa               96 arch/ia64/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              126 arch/ia64/include/asm/page.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa               41 arch/ia64/include/asm/pgalloc.h 	pgd_val(*pgd_entry) = __pa(pud);
__pa               59 arch/ia64/include/asm/pgalloc.h 	pud_val(*pud_entry) = __pa(pmd);
__pa               84 arch/ia64/include/asm/pgalloc.h 	pmd_val(*pmd_entry) = __pa(pte);
__pa              287 arch/ia64/include/asm/uaccess.h 		ptr = (void *)__pa(p) + __IA64_UNCACHED_OFFSET;
__pa              156 arch/ia64/include/asm/uv/uv_hub.h 	return __pa(v) | uv_hub_info->gnode_upper;
__pa              469 arch/ia64/kernel/efi.c 		 pte_val(pfn_pte(__pa(pal_vaddr) >> PAGE_SHIFT, PAGE_KERNEL)),
__pa             1304 arch/ia64/kernel/efi.c 			if (__pa(r[i].start) >= start && __pa(r[i].end) < end) {
__pa             1305 arch/ia64/kernel/efi.c 				if (__pa(r[i].start) > start + size)
__pa             1307 arch/ia64/kernel/efi.c 				start = ALIGN(__pa(r[i].end), alignment);
__pa             1309 arch/ia64/kernel/efi.c 				    __pa(r[i+1].start) < start + size)
__pa             1864 arch/ia64/kernel/mca.c 	__this_cpu_write(ia64_mca_data, (__per_cpu_mca[cpu] = __pa(data)));
__pa             1871 arch/ia64/kernel/mca.c 		pte_val(mk_pte_phys(__pa(cpu_data), PAGE_KERNEL)));
__pa             1882 arch/ia64/kernel/mca.c 	__this_cpu_write(ia64_mca_pal_pte, pte_val(mk_pte_phys(__pa(pal_vaddr),
__pa              167 arch/ia64/kernel/setup.c 			call_pernode_memory(__pa(range_start), range_end - range_start, func);
__pa              197 arch/ia64/kernel/setup.c 		call_pernode_memory(__pa(start), end - start, func);
__pa              318 arch/ia64/kernel/setup.c 	boot_param_res.start = __pa(ia64_boot_param);
__pa              393 arch/ia64/kernel/setup.c 		phys_addr_t addr = __pa(region->start);
__pa              447 arch/ia64/kernel/setup.c 	ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase));
__pa             1009 arch/ia64/kernel/setup.c 	ia64_set_kr(IA64_KR_PT_BASE, __pa(ia64_imva(empty_zero_page)));
__pa              447 arch/ia64/kernel/smpboot.c 	ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase));
__pa               75 arch/ia64/mm/contig.c 			ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) -
__pa               90 arch/ia64/mm/contig.c 				       __pa(MAX_DMA_ADDRESS));
__pa               93 arch/ia64/mm/contig.c 		      __func__, size, PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
__pa              365 arch/ia64/mm/discontig.c 		base = __pa(mem_data[node].pernode_addr);
__pa              450 arch/ia64/mm/discontig.c 				     __pa(MAX_DMA_ADDRESS),
__pa              456 arch/ia64/mm/discontig.c 		      __pa(MAX_DMA_ADDRESS));
__pa              474 arch/ia64/mm/discontig.c 		fill_pernode(node, __pa(pernode), pernodesize);
__pa              437 arch/ia64/mm/init.c 	map_start = vmem_map + (__pa(start) >> PAGE_SHIFT);
__pa              438 arch/ia64/mm/init.c 	map_end   = vmem_map + (__pa(end) >> PAGE_SHIFT);
__pa              442 arch/ia64/mm/init.c 	node = paddr_to_nid(__pa(start));
__pa              475 arch/ia64/mm/init.c 			set_pte(pte, pfn_pte(__pa(page) >> PAGE_SHIFT,
__pa              501 arch/ia64/mm/init.c 	map_start = vmem_map + (__pa(start) >> PAGE_SHIFT);
__pa              502 arch/ia64/mm/init.c 	map_end   = vmem_map + (__pa(end) >> PAGE_SHIFT);
__pa              586 arch/ia64/mm/init.c 		memblock_add_node(__pa(start), end - start, nid);
__pa              595 arch/ia64/mm/init.c 	pfn_start = (PAGE_ALIGN(__pa(start))) >> PAGE_SHIFT;
__pa              596 arch/ia64/mm/init.c 	pfn_end = (PAGE_ALIGN(__pa(end - 1))) >> PAGE_SHIFT;
__pa              598 arch/ia64/mm/init.c 	pfn_start = GRANULEROUNDDOWN(__pa(start)) >> PAGE_SHIFT;
__pa              599 arch/ia64/mm/init.c 	pfn_end = GRANULEROUNDUP(__pa(end - 1)) >> PAGE_SHIFT;
__pa              177 arch/ia64/pci/pci.c 	base = __pa(io_space[space_nr].mmio_base);
__pa              240 arch/m68k/include/asm/cacheflush_mm.h 				     : : "a" (__pa(vaddr)));
__pa              122 arch/m68k/include/asm/motorola_pgtable.h 	pgd_val(*pgdp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp);
__pa              124 arch/m68k/include/asm/page_mm.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa               23 arch/m68k/include/asm/page_no.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa               30 arch/m68k/include/asm/sun3_pgalloc.h 	pmd_val(*pmd) = __pa((unsigned long)pte);
__pa               35 arch/m68k/include/asm/sun3_pgalloc.h 	pmd_val(*pmd) = __pa((unsigned long)page_address(page));
__pa               18 arch/m68k/include/asm/sun3_pgtable.h #define VTOP(addr)	__pa(addr)
__pa               22 arch/m68k/include/asm/virtconvert.h 	return __pa(address);
__pa               34 arch/m68k/include/asm/virtconvert.h 	__pa(PAGE_OFFSET + (((page) - pg_data_map[0].node_mem_map) << PAGE_SHIFT))
__pa               71 arch/m68k/mm/sun3mmu.c 		pg_table = (pte_t *) __pa (next_pgtable);
__pa               97 arch/m68k/mm/sun3mmu.c 			    (__pa(PAGE_OFFSET) >> PAGE_SHIFT) + 1, NULL);
__pa              118 arch/m68k/sun3/config.c 	start_page = __pa(memory_start) >> PAGE_SHIFT;
__pa              119 arch/m68k/sun3/config.c 	max_pfn = num_pages = __pa(memory_end) >> PAGE_SHIFT;
__pa              143 arch/m68k/sun3/mmu_emu.c 	for (i=0; i < __pa(bootmem_end) / SUN3_PMEG_SIZE ; ++i)
__pa              124 arch/m68k/sun3x/dvma.c 					 __pa(kaddr), vaddr);
__pa              160 arch/m68k/sun3x/dvma.c 		dvma_entry_set(index, __pa(kaddr));
__pa              148 arch/microblaze/include/asm/page.h # define virt_to_pfn(vaddr)	(phys_to_pfn((__pa(vaddr))))
__pa              153 arch/microblaze/include/asm/page.h #  define virt_to_page(kaddr)	(pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
__pa              473 arch/microblaze/include/asm/pgtable.h #define pmd_page(pmd)	(pfn_to_page(__pa(pmd_val(pmd)) >> PAGE_SHIFT))
__pa              311 arch/microblaze/mm/init.c 	kstart = __pa(CONFIG_KERNEL_START); /* kernel start */
__pa              163 arch/mips/bcm47xx/prom.c 	off = EXTVBASE + __pa(off);
__pa              113 arch/mips/dec/prom/memory.c 		end = __pa(&_text) - 0x00020000;
__pa              116 arch/mips/dec/prom/memory.c 		end = __pa(&_text);
__pa              109 arch/mips/include/asm/io.h 	return __pa(address);
__pa              216 arch/mips/include/asm/page.h #define __pa_symbol(x)	__pa(RELOC_HIDE((unsigned long)(x), 0))
__pa              205 arch/mips/kernel/setup.c 	end = __pa(initrd_end);
__pa              207 arch/mips/kernel/setup.c 	initrd_start = (unsigned long)__va(__pa(initrd_start));
__pa              254 arch/mips/kernel/setup.c 	if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
__pa              261 arch/mips/kernel/setup.c 	memblock_reserve(__pa(initrd_start), size);
__pa               96 arch/mips/loongson64/common/mem.c 	return addr >= __pa(high_memory) ||
__pa              242 arch/mips/mm/cache.c 	return addr >= __pa(high_memory);
__pa               54 arch/mips/mm/dma-noncoherent.c 	return (void *)(__pa(addr) + UNCAC_BASE);
__pa               38 arch/mips/power/cpu.c 	unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
__pa               39 arch/mips/power/cpu.c 	unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end));
__pa               84 arch/nds32/include/asm/memory.h #define virt_to_page(kaddr)	(pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
__pa               55 arch/nds32/include/asm/pgalloc.h 	pmdval = __pa(pte_ptr) | _PAGE_KERNEL_TABLE;
__pa               76 arch/nds32/include/asm/pgtable.h #define MAXMEM			__pa(VMALLOC_START)
__pa              265 arch/nds32/kernel/setup.c 	free_ram_start_pfn = PFN_UP(__pa(&_end));
__pa               65 arch/nds32/mm/init.c 	e = min((u32) memblock_end_of_DRAM(), (u32) __pa(high_memory));
__pa               85 arch/nds32/mm/init.c 		set_pmd(pme, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE));
__pa              120 arch/nds32/mm/init.c 	set_pmd(pmd, __pmd(__pa(fixmap_pmd_p) + _PAGE_KERNEL_TABLE));
__pa              135 arch/nds32/mm/init.c 	set_pmd(pmd, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE));
__pa              173 arch/nds32/mm/init.c 	for (pfn = PFN_UP(__pa(high_memory)); pfn < max_pfn; pfn++) {
__pa              410 arch/nds32/mm/proc.c 		unsigned long p_start = __pa(start);
__pa              411 arch/nds32/mm/proc.c 		unsigned long p_end = __pa(end);
__pa              532 arch/nds32/mm/proc.c 	__nds32__mtsr_isb(__pa(mm->pgd), NDS32_SR_L1_PPTB);
__pa              152 arch/nios2/kernel/setup.c 	memory_start = PAGE_ALIGN((unsigned long)__pa(_end));
__pa               75 arch/openrisc/include/asm/page.h #define virt_to_pfn(kaddr)      (__pa(kaddr) >> PAGE_SHIFT)
__pa               26 arch/openrisc/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
__pa              317 arch/openrisc/include/asm/pgtable.h 	pte_val(pte) = __pa(page) | pgprot_val(pgprot);
__pa               41 arch/openrisc/kernel/dma.c 	for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size)
__pa               98 arch/openrisc/kernel/dma.c 	*dma_handle = __pa(page);
__pa               82 arch/openrisc/kernel/setup.c 	memblock_reserve(__pa(_stext), _end - _stext);
__pa              160 arch/openrisc/kernel/traps.c 	       ((struct task_struct *)(__pa(current)))->comm,
__pa              161 arch/openrisc/kernel/traps.c 	       ((struct task_struct *)(__pa(current)))->pid,
__pa              172 arch/openrisc/kernel/traps.c 		       *((unsigned long *)(__pa(stack))));
__pa              179 arch/openrisc/kernel/traps.c 		addr = *((unsigned long *)__pa(stack));
__pa              195 arch/openrisc/kernel/traps.c 		c = ((unsigned char *)(__pa(regs->pc)))[i];
__pa              107 arch/openrisc/mm/init.c 			set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
__pa              302 arch/parisc/boot/compressed/misc.c 	kernel_len = __pa(SZ_end) - __pa(SZparisc_kernel_start);
__pa                8 arch/parisc/include/asm/io.h #define virt_to_phys(a) ((unsigned long)__pa(a))
__pa               57 arch/parisc/include/asm/mmu_context.h 		mtctl(__pa(next->pgd), 25);
__pa              172 arch/parisc/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              175 arch/parisc/include/asm/page.h #define virt_to_page(kaddr)     pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa               40 arch/parisc/include/asm/pgalloc.h 			+ (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT));
__pa               65 arch/parisc/include/asm/pgalloc.h 		        (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT));
__pa              116 arch/parisc/include/asm/pgalloc.h 			+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
__pa              120 arch/parisc/include/asm/pgalloc.h 			+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
__pa              477 arch/parisc/kernel/cache.c 	flush_dcache_page_asm(__pa(vfrom), vaddr);
__pa              161 arch/parisc/kernel/firmware.c 		__pa(pdc_result), 0);
__pa              242 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result));
__pa              267 arch/parisc/kernel/firmware.c                               __pa(pdc_result), __pa(pdc_result2), len);
__pa              291 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data));
__pa              337 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result));
__pa              348 arch/parisc/kernel/firmware.c 	ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result));
__pa              395 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, 
__pa              396 arch/parisc/kernel/firmware.c 			      index, __pa(pdc_result2), iodc_data_size);
__pa              422 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), 
__pa              423 arch/parisc/kernel/firmware.c 			      __pa(pdc_result2), mod_index);
__pa              449 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_ADDRESS, __pa(pdc_result),
__pa              471 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_INFO, __pa(pdc_result), 0);
__pa              493 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_SYSMODEL, __pa(pdc_result),
__pa              494 arch/parisc/kernel/firmware.c                               OS_ID_HPUX, __pa(name));
__pa              523 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_VERSIONS, __pa(pdc_result), id);
__pa              545 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CPU_ID, __pa(pdc_result), 0);
__pa              567 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CAPABILITIES, __pa(pdc_result), 0);
__pa              596 arch/parisc/kernel/firmware.c 		__pa(orig_prod_num), __pa(current_prod_num), __pa(serial_no));
__pa              615 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_INFO, __pa(pdc_result), 0);
__pa              636 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_RET_SPID, __pa(pdc_result), 0);
__pa              657 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_BLOCK_TLB, PDC_BTLB_INFO, __pa(pdc_result), 0);
__pa              687 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_MEM_MAP, PDC_MEM_MAP_HPA, __pa(pdc_result),
__pa              688 arch/parisc/kernel/firmware.c 				__pa(pdc_result2));
__pa              710 arch/parisc/kernel/firmware.c 			__pa(pdc_result), hpa);
__pa              740 arch/parisc/kernel/firmware.c                __pa(pdc_result), count);
__pa              768 arch/parisc/kernel/firmware.c                __pa(pdc_result), count);
__pa              790 arch/parisc/kernel/firmware.c        retval = mem_pdc_call(PDC_STABLE, PDC_STABLE_RETURN_SIZE, __pa(pdc_result));
__pa              862 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), __pa(hwpath));
__pa              918 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), hpa);
__pa              945 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), hpa, __pa(tbl));
__pa              970 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), hpa, cfg_addr&~3UL, 4UL);
__pa              993 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), hpa,
__pa             1013 arch/parisc/kernel/firmware.c         retval = mem_pdc_call(PDC_TOD, PDC_TOD_READ, __pa(pdc_result), 0);
__pa             1028 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_MEM, PDC_MEM_MEMINFO, __pa(pdc_result), 0);
__pa             1043 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_MEM, PDC_MEM_READ_PDT, __pa(pdc_result),
__pa             1044 arch/parisc/kernel/firmware.c 			__pa(pdt_entries_ptr));
__pa             1091 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_MEM, PDC_MEM_TABLE, __pa(pdc_result), __pa(pdc_result2), entries);
__pa             1149 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_INFO, __pa(pdc_result), 0);
__pa             1176 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_ENABLE, __pa(pdc_result), sw_control);
__pa             1253 arch/parisc/kernel/firmware.c                     PAGE0->mem_cons.spa, __pa(PAGE0->mem_cons.dp.layers),
__pa             1254 arch/parisc/kernel/firmware.c                     __pa(iodc_retbuf), 0, __pa(iodc_dbuf), i, 0);
__pa             1281 arch/parisc/kernel/firmware.c 		    PAGE0->mem_kbd.spa, __pa(PAGE0->mem_kbd.dp.layers), 
__pa             1282 arch/parisc/kernel/firmware.c 		    __pa(iodc_retbuf), 0, __pa(iodc_dbuf), 1, 0);
__pa             1323 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_GET_NUMBER, __pa(pdc_result));
__pa             1349 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_MODULE, __pa(pdc_result), 
__pa             1350 arch/parisc/kernel/firmware.c 			      ploc, mod, view_type, __pa(&result));
__pa             1379 arch/parisc/kernel/firmware.c 			__pa(pdc_result), __pa(&result), *actcnt,
__pa             1404 arch/parisc/kernel/firmware.c 			      __pa(&pdc_result), hpa);
__pa             1426 arch/parisc/kernel/firmware.c 			      __pa(pdc_result), cell_num);
__pa             1447 arch/parisc/kernel/firmware.c 			      __pa(r_addr), cell_num);
__pa             1468 arch/parisc/kernel/firmware.c 	retval = mem_pdc_call(PDC_PAT_PD, PDC_PAT_PD_GET_ADDR_MAP, __pa(pdc_result), 
__pa             1469 arch/parisc/kernel/firmware.c 			      __pa(pdc_result2), count, offset);
__pa             1492 arch/parisc/kernel/firmware.c 				__pa(pdc_result));
__pa             1518 arch/parisc/kernel/firmware.c 					__pa(pdc_result), pci_addr, pci_size);
__pa             1562 arch/parisc/kernel/firmware.c 			__pa(&pdc_result));
__pa             1585 arch/parisc/kernel/firmware.c 			__pa(&pdc_result), cell);
__pa             1609 arch/parisc/kernel/firmware.c 			__pa(&pdc_result), parisc_cell_num,
__pa             1610 arch/parisc/kernel/firmware.c 			__pa(pdt_entries_ptr));
__pa             1641 arch/parisc/kernel/firmware.c 		__pa(&pdc_result), __pa(pdt_entries_ptr),
__pa             1670 arch/parisc/kernel/firmware.c 		__pa(&pdc_result), phys_addr);
__pa               36 arch/parisc/kernel/kexec_file.c 	image->start = __pa(elf_info.ehdr->e_entry);
__pa               39 arch/parisc/kernel/kexec_file.c 		image->segment[i].mem = __pa(image->segment[i].mem);
__pa              412 arch/parisc/kernel/pci-dma.c 	paddr = __pa(paddr);
__pa              191 arch/parisc/kernel/pdt.c 		pdt_status.first_dbe_loc <= __pa((unsigned long)&_end))
__pa              128 arch/parisc/kernel/setup.c 	if (__pa((unsigned long) &_end) >= KERNEL_INITIAL_SIZE)
__pa              839 arch/parisc/kernel/traps.c 	ivap[6] = (u32)__pa(os_hpmc);
__pa              294 arch/parisc/mm/init.c 	memblock_reserve(__pa(KERNEL_BINARY_TEXT_START),
__pa              310 arch/parisc/mm/init.c 		if (__pa(initrd_start) < mem_max) {
__pa              313 arch/parisc/mm/init.c 			if (__pa(initrd_end) > mem_max) {
__pa              314 arch/parisc/mm/init.c 				initrd_reserve = mem_max - __pa(initrd_start);
__pa              319 arch/parisc/mm/init.c 			printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initrd_start) + initrd_reserve, mem_max);
__pa              321 arch/parisc/mm/init.c 			memblock_reserve(__pa(initrd_start), initrd_reserve);
__pa              368 arch/parisc/mm/init.c 	ro_start = __pa((unsigned long)_text);
__pa              369 arch/parisc/mm/init.c 	ro_end   = __pa((unsigned long)&data_start);
__pa              370 arch/parisc/mm/init.c 	kernel_start = __pa((unsigned long)&__init_begin);
__pa              371 arch/parisc/mm/init.c 	kernel_end  = __pa((unsigned long)&_end);
__pa              388 arch/parisc/mm/init.c 		pmd = (pmd_t *)__pa(pg_dir);
__pa              401 arch/parisc/mm/init.c 			pmd = (pmd_t *) __pa(pmd);
__pa              423 arch/parisc/mm/init.c 				pg_table = (pte_t *) __pa(pg_table);
__pa              480 arch/parisc/mm/init.c 	map_pages(start, __pa(start), end-start,
__pa              496 arch/parisc/mm/init.c 	map_pages(init_end, __pa(init_end), kernel_end - init_end,
__pa              506 arch/parisc/mm/init.c 	map_pages(init_begin, __pa(init_begin), init_end - init_begin,
__pa              510 arch/parisc/mm/init.c 	map_pages(init_begin, __pa(init_begin), init_end - init_begin,
__pa              671 arch/parisc/mm/init.c 		map_pages(initrd_start, __pa(initrd_start),
__pa              698 arch/parisc/mm/init.c 	map_pages(linux_gateway_page_addr, __pa(&linux_gateway_page),
__pa               31 arch/powerpc/include/asm/book3s/32/pgalloc.h 	*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
__pa               37 arch/powerpc/include/asm/book3s/32/pgalloc.h 	*pmdp = __pmd(__pa(pte_page) | _PMD_PRESENT);
__pa              309 arch/powerpc/include/asm/book3s/32/pgtable.h 		unsigned long ptephys = __pa(ptep) & PAGE_MASK;
__pa              104 arch/powerpc/include/asm/book3s/64/hugetlb.h 	*hpdp = __hugepd(__pa(new) | HUGEPD_VAL_BITS | (shift_to_mmu_psize(pshift) << 2));
__pa              995 arch/powerpc/include/asm/book3s/64/pgtable.h #define __pgtable_ptr_val(ptr)	__pa(ptr)
__pa              777 arch/powerpc/include/asm/io.h 	return __pa((unsigned long)address);
__pa              821 arch/powerpc/include/asm/io.h         return __pa(address) + PCI_DRAM_OFFSET;
__pa               35 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h 	*hpdp = __hugepd(__pa(new) | _PMD_USER | _PMD_PRESENT |
__pa               23 arch/powerpc/include/asm/nohash/32/pgalloc.h 		*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
__pa               32 arch/powerpc/include/asm/nohash/32/pgalloc.h 		*pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT);
__pa              354 arch/powerpc/include/asm/nohash/32/pgtable.h 	pfn_to_page((__pa(pmd_val(pmd)) >> PAGE_SHIFT))
__pa              131 arch/powerpc/include/asm/page.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa               62 arch/powerpc/kernel/crash_dump.c 	create_trampoline(__pa(system_reset_fwnmi) - PHYSICAL_START);
__pa               63 arch/powerpc/kernel/crash_dump.c 	create_trampoline(__pa(machine_check_fwnmi) - PHYSICAL_START);
__pa             1063 arch/powerpc/kernel/dt_cpu_ftrs.c 	memblock_free(__pa(dt_cpu_features),
__pa              667 arch/powerpc/kernel/fadump.c 		phdr->p_paddr	= __pa(fw_dump.cpu_notes_buf_vaddr);
__pa              170 arch/powerpc/kernel/machine_kexec.c 	if (overlaps_crashkernel(__pa(_stext), _end - _stext)) {
__pa              271 arch/powerpc/kernel/machine_kexec.c 	kernel_end = cpu_to_be_ulong(__pa(_end));
__pa               50 arch/powerpc/kernel/machine_kexec_64.c 		if (image->segment[i].mem < __pa(_end))
__pa              408 arch/powerpc/kernel/machine_kexec_64.c 	htab_base = cpu_to_be64(__pa(htab_address));
__pa              141 arch/powerpc/kernel/machine_kexec_file_64.c 	ret = delete_fdt_mem_rsv(fdt, __pa(initial_boot_params),
__pa               78 arch/powerpc/kernel/paca.c 		uv_share_page(PHYS_PFN(__pa(shared_lppaca)),
__pa              286 arch/powerpc/kernel/paca.c 		memblock_free(__pa(paca_ptrs) + new_ptrs_size,
__pa              295 arch/powerpc/kernel/paca.c 		memblock_free(__pa(paca_ptrs[boot_cpuid]->slb_shadow_ptr),
__pa               37 arch/powerpc/kernel/proc_powerpc.c 			__pa(PDE_DATA(file_inode(file))) >> PAGE_SHIFT,
__pa              120 arch/powerpc/kernel/prom.c 	start = __pa(initial_boot_params);
__pa              626 arch/powerpc/kernel/prom.c 		memblock_reserve(_ALIGN_DOWN(__pa(initrd_start), PAGE_SIZE),
__pa              733 arch/powerpc/kernel/prom.c 	memblock_reserve(PHYSICAL_START, __pa(klimit) - PHYSICAL_START);
__pa             1766 arch/powerpc/kernel/prom_init.c 	os_term_args.args[0] = cpu_to_be32(__pa(str));
__pa             3160 arch/powerpc/kernel/prom_init.c 		prom_initrd_start = is_kernel_addr(r3) ? __pa(r3) : r3;
__pa              379 arch/powerpc/kernel/rtas.c 	err_args.args[0] = cpu_to_be32(__pa(rtas_err_buf));
__pa              386 arch/powerpc/kernel/rtas.c 	enter_rtas(__pa(&rtas.args));
__pa              432 arch/powerpc/kernel/rtas.c 	enter_rtas(__pa(args));
__pa              736 arch/powerpc/kernel/rtas.c 				   __pa(rtas_os_term_buf));
__pa             1123 arch/powerpc/kernel/rtas.c 	enter_rtas(__pa(&rtas.args));
__pa              456 arch/powerpc/kernel/rtas_flash.c 			       (u32) __pa(rtas_data_buf), args_buf->buf_size);
__pa              596 arch/powerpc/kernel/rtas_flash.c 	rtas_block_list = __pa(flist);
__pa              610 arch/powerpc/kernel/rtas_flash.c 			f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data));
__pa              617 arch/powerpc/kernel/rtas_flash.c 			f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next));
__pa              421 arch/powerpc/kernel/rtasd.c 				  __pa(logdata), rtas_error_log_max);
__pa              831 arch/powerpc/kernel/setup-common.c 	memblock_free(__pa(cpu_to_phys_id), nr_cpu_ids * sizeof(u32));
__pa              762 arch/powerpc/kernel/setup_64.c 	return memblock_alloc_try_nid(size, align, __pa(MAX_DMA_ADDRESS),
__pa              770 arch/powerpc/kernel/setup_64.c 	memblock_free(__pa(ptr), size);
__pa               20 arch/powerpc/kernel/suspend.c 	unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
__pa               21 arch/powerpc/kernel/suspend.c 	unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
__pa              120 arch/powerpc/kvm/book3s_64_mmu_hv.c 	kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18);
__pa               41 arch/powerpc/kvm/book3s_64_mmu_radix.c 					  __pa(to), __pa(from), n);
__pa             3613 arch/powerpc/kvm/book3s_hv.c 		trap = plpar_hcall_norets(H_ENTER_NESTED, __pa(&hvregs),
__pa             3614 arch/powerpc/kvm/book3s_hv.c 					  __pa(&vcpu->arch.regs));
__pa             4566 arch/powerpc/kvm/book3s_hv.c 			__pa(kvm->arch.pgtable) | RADIX_PGD_INDEX_SIZE;
__pa              369 arch/powerpc/kvm/book3s_hv_nested.c 	ptcr = __pa(pseries_partition_tb) | (ptb_order - 8);
__pa              429 arch/powerpc/kvm/book3s_hv_nested.c 		__pa(gp->shadow_pgtable) | RADIX_PGD_INDEX_SIZE;
__pa              383 arch/powerpc/mm/book3s32/mmu.c 	_SDR1 = __pa(Hash) | SDR1_LOW_BITS;
__pa               40 arch/powerpc/mm/book3s32/tlb.c 		ptephys = __pa(ptep) & PAGE_MASK;
__pa              801 arch/powerpc/mm/book3s64/hash_utils.c 	rc = htab_bolt_mapping(start, end, __pa(start),
__pa              943 arch/powerpc/mm/book3s64/hash_utils.c 		BUG_ON(htab_bolt_mapping(base, base + size, __pa(base),
__pa              963 arch/powerpc/mm/book3s64/hash_utils.c 					 __pa(tce_alloc_start), prot,
__pa             1078 arch/powerpc/mm/book3s64/hash_utils.c 			set_ptcr_when_no_uv(__pa(partition_tb) |
__pa             1885 arch/powerpc/mm/book3s64/hash_utils.c 	ret = hpte_insert_repeating(hash, vpn, __pa(vaddr), mode,
__pa             1925 arch/powerpc/mm/book3s64/hash_utils.c 		lmi = __pa(vaddr) >> PAGE_SHIFT;
__pa              167 arch/powerpc/mm/book3s64/mmu_context.c 	process_tb[index].prtb0 = cpu_to_be64(rts_field | __pa(mm->pgd) | RADIX_PGD_INDEX_SIZE);
__pa              207 arch/powerpc/mm/book3s64/pgtable.c 	ptcr = __pa(partition_tb) | (PATB_SIZE_SHIFT - 12);
__pa              370 arch/powerpc/mm/book3s64/radix_pgtable.c 	process_tb->prtb0 = cpu_to_be64(rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE);
__pa              395 arch/powerpc/mm/book3s64/radix_pgtable.c 	dw0 = rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE | PATB_HR;
__pa              396 arch/powerpc/mm/book3s64/radix_pgtable.c 	dw1 = __pa(process_tb) | (PRTB_SIZE_SHIFT - 12) | PATB_GR;
__pa              623 arch/powerpc/mm/book3s64/radix_pgtable.c 		set_ptcr_when_no_uv(__pa(partition_tb) |
__pa              710 arch/powerpc/mm/book3s64/radix_pgtable.c 	create_physical_mapping(__pa(params->aligned_start), __pa(params->start), -1);
__pa              711 arch/powerpc/mm/book3s64/radix_pgtable.c 	create_physical_mapping(__pa(params->end), __pa(params->aligned_end), -1);
__pa              875 arch/powerpc/mm/book3s64/radix_pgtable.c 	return create_physical_mapping(__pa(start), __pa(end), nid);
__pa              222 arch/powerpc/mm/init_64.c 		vmemmap_list_populate(__pa(p), start, node);
__pa              227 arch/powerpc/mm/init_64.c 		rc = vmemmap_create_mapping(start, page_size, __pa(p));
__pa               26 arch/powerpc/mm/kasan/kasan_init_32.c 	phys_addr_t pa = __pa(kasan_early_shadow_page);
__pa               99 arch/powerpc/mm/kasan/kasan_init_32.c 		pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
__pa              116 arch/powerpc/mm/kasan/kasan_init_32.c 	phys_addr_t pa = __pa(kasan_early_shadow_page);
__pa              194 arch/powerpc/mm/kasan/kasan_init_32.c 	modify_instruction_site(&patch__hash_page_A0, 0xffff, __pa(early_hash) >> 16);
__pa              195 arch/powerpc/mm/kasan/kasan_init_32.c 	modify_instruction_site(&patch__flush_hash_A0, 0xffff, __pa(early_hash) >> 16);
__pa               36 arch/powerpc/mm/nohash/8xx.c 		return __pa(va);
__pa              121 arch/powerpc/mm/nohash/8xx.c 					    _ALIGN(__pa(_einittext), 8 << 20));
__pa              148 arch/powerpc/mm/nohash/8xx.c 		mmu_patch_cmp_limit(&patch__itlbmiss_linmem_top, __pa(_etext));
__pa              156 arch/powerpc/mm/nohash/8xx.c 				-__pa(((unsigned long)_sinittext) &
__pa              158 arch/powerpc/mm/nohash/8xx.c 	mmu_patch_addis(&patch__dtlbmiss_romem_top, -__pa(_sinittext));
__pa              184 arch/powerpc/mm/nohash/8xx.c 	s16 offset = (s16)(__pa(swapper_pg_dir));
__pa              197 arch/powerpc/mm/nohash/8xx.c 	mtspr(SPRN_M_TWB, __pa(pgd) - offset);
__pa               59 arch/powerpc/mm/nohash/book3e_pgtable.c 				     __pa(MAX_DMA_ADDRESS), NUMA_NO_NODE);
__pa               63 arch/powerpc/mm/nohash/book3e_pgtable.c 		      __func__, size, size, __pa(MAX_DMA_ADDRESS));
__pa              296 arch/powerpc/mm/pgtable.c 	return __pa(pfn_to_kaddr(pfn)) + offset_in_page(va);
__pa              194 arch/powerpc/oprofile/op_model_cell.c 	u64 paddr = __pa(address);
__pa              603 arch/powerpc/perf/imc-pmu.c 				__pa((void *)mem_info->vbase),
__pa             1113 arch/powerpc/perf/imc-pmu.c 		rc = opal_imc_counters_init(OPAL_IMC_COUNTERS_TRACE, __pa((void *)local_mem),
__pa              113 arch/powerpc/platforms/44x/iss4xx.c 	spin_table[1] = __pa(start_secondary_47x);
__pa              184 arch/powerpc/platforms/44x/ppc476.c 	spin_table[1] = __pa(start_secondary_47x);
__pa              253 arch/powerpc/platforms/85xx/smp.c 		__pa(ppc_function_entry(generic_secondary_smp_init)));
__pa              255 arch/powerpc/platforms/85xx/smp.c 	out_be32(&spin_table->addr_l, __pa(__early_start));
__pa              142 arch/powerpc/platforms/cell/iommu.c 			| (__pa(pte) & IOC_IOPT_CacheInvd_IOPTE_Mask)
__pa              189 arch/powerpc/platforms/cell/iommu.c 		io_pte[i] = base_pte | (__pa(uaddr) & CBE_IOPTE_RPN_Mask);
__pa              216 arch/powerpc/platforms/cell/iommu.c 		__pa(window->iommu->pad_page) |
__pa              368 arch/powerpc/platforms/cell/iommu.c 		iommu->stab[i] = reg | (__pa(ptab) + (n_pte_pages << 12) *
__pa              407 arch/powerpc/platforms/cell/iommu.c 	reg = IOC_IOST_Origin_E | __pa(iommu->stab) | IOC_IOST_Origin_HW;
__pa              872 arch/powerpc/platforms/cell/iommu.c 	ptab[offset] = base_pte | (__pa(addr) & CBE_IOPTE_RPN_Mask);
__pa              962 arch/powerpc/platforms/cell/iommu.c 		hbase = __pa(htab_address);
__pa              137 arch/powerpc/platforms/cell/ras.c 	addr = __pa(page_address(area->pages)) + (PAGE_SIZE >> 1);
__pa               71 arch/powerpc/platforms/cell/smp.c 			__pa(ppc_function_entry(generic_secondary_smp_init));
__pa               35 arch/powerpc/platforms/chrp/nvram.c 		       __pa(nvram_buf), 1) != 0) || 1 != done)
__pa               57 arch/powerpc/platforms/chrp/nvram.c 		       __pa(nvram_buf), 1) != 0) || 1 != done)
__pa               92 arch/powerpc/platforms/pasemi/iommu.c 		rpn = __pa(uaddr) >> IOBMAP_PAGE_SHIFT;
__pa              224 arch/powerpc/platforms/pasemi/iommu.c 		regword = IOBMAP_L1E_V | (__pa(iob_l2_base + i*0x2000) >> 12);
__pa              335 arch/powerpc/platforms/powermac/smp.c 	unsigned long start = __pa(__secondary_start_pmac_0) + nr * 8;
__pa              152 arch/powerpc/platforms/powernv/npu-dma.c 			__pa(tbl->it_base),
__pa              259 arch/powerpc/platforms/powernv/opal-dump.c 	addr = __pa(list);
__pa              166 arch/powerpc/platforms/powernv/opal-elog.c 		opal_rc = opal_read_elog(__pa(elog->buffer),
__pa              209 arch/powerpc/platforms/powernv/opal-elog.c 		rc = opal_read_elog(__pa(elog->buffer),
__pa              137 arch/powerpc/platforms/powernv/opal-flash.c 	ret = opal_validate_flash(__pa(buf), &size, &result);
__pa              293 arch/powerpc/platforms/powernv/opal-flash.c 	addr = __pa(list);
__pa              301 arch/powerpc/platforms/powernv/opal-hmi.c 		while (opal_get_msg(__pa(&msg), sizeof(msg)) == OPAL_SUCCESS) {
__pa               36 arch/powerpc/platforms/powernv/opal-nvram.c 	rc = opal_read_nvram(__pa(buf), count, off);
__pa               59 arch/powerpc/platforms/powernv/opal-nvram.c 		rc = opal_write_nvram(__pa(buf), count, off);
__pa               49 arch/powerpc/platforms/powernv/opal-powercap.c 	ret = opal_get_powercap(pcap_attr->handle, token, (u32 *)__pa(&pcap));
__pa               43 arch/powerpc/platforms/powernv/opal-psr.c 					    (u32 *)__pa(&psr));
__pa               61 arch/powerpc/platforms/powernv/opal-xscom.c 	rc = opal_xscom_read(chip, reg, (__be64 *)__pa(&v));
__pa              278 arch/powerpc/platforms/powernv/opal.c 	ret = opal_get_msg(__pa(opal_msg), opal_msg_size);
__pa              840 arch/powerpc/platforms/powernv/opal.c 				       __pa(addr), size);
__pa             1076 arch/powerpc/platforms/powernv/opal.c 			sg->next = cpu_to_be64(__pa(next));
__pa               77 arch/powerpc/platforms/powernv/pci-ioda-tce.c 			tce = __pa(tmp2) | TCE_PCI_READ | TCE_PCI_WRITE;
__pa              101 arch/powerpc/platforms/powernv/pci-ioda-tce.c 	u64 rpn = __pa(uaddr) >> tbl->it_page_shift;
__pa              253 arch/powerpc/platforms/powernv/pci-ioda-tce.c 		addr[i] = cpu_to_be64(__pa(tmp) |
__pa             1828 arch/powerpc/platforms/powernv/pci-ioda.c 					__pa(tces),
__pa             1912 arch/powerpc/platforms/powernv/pci-ioda.c 	start = __pa(((__be64 *)tbl->it_base) + index - tbl->it_offset);
__pa             1913 arch/powerpc/platforms/powernv/pci-ioda.c 	end = __pa(((__be64 *)tbl->it_base) + index - tbl->it_offset +
__pa             2258 arch/powerpc/platforms/powernv/pci-ioda.c 					      __pa(addr) + tce32_segsz * i,
__pa             2319 arch/powerpc/platforms/powernv/pci-ioda.c 			__pa(tbl->it_base),
__pa             3700 arch/powerpc/platforms/powernv/pci-ioda.c 		memblock_free(__pa(phb), sizeof(struct pnv_phb));
__pa               68 arch/powerpc/platforms/powernv/smp.c 			__pa(ppc_function_entry(generic_secondary_smp_init));
__pa              403 arch/powerpc/platforms/powernv/vas-window.c 	val = __pa(winctx->rx_fifo);
__pa              788 arch/powerpc/platforms/ps3/device-init.c 	lpar = ps3_mm_phys_to_lpar(__pa(buf));
__pa              531 arch/powerpc/platforms/ps3/interrupt.c 	lpar_addr = ps3_mm_phys_to_lpar(__pa(virt_addr_bmp));
__pa              757 arch/powerpc/platforms/ps3/interrupt.c 			ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
__pa              760 arch/powerpc/platforms/ps3/interrupt.c 			pd->thread_id, ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
__pa              776 arch/powerpc/platforms/ps3/mm.c 	unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
__pa              831 arch/powerpc/platforms/ps3/mm.c 	unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
__pa             1072 arch/powerpc/platforms/ps3/mm.c 	unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr)
__pa             1120 arch/powerpc/platforms/ps3/mm.c 	lpar_addr = addr ? ps3_mm_phys_to_lpar(__pa(addr)) : 0;
__pa              188 arch/powerpc/platforms/pseries/cmm.c 		if ((rc = plpar_page_set_loaned(__pa(addr)))) {
__pa              232 arch/powerpc/platforms/pseries/cmm.c 		plpar_page_set_active(__pa(addr));
__pa              571 arch/powerpc/platforms/pseries/cmm.c 			plpar_page_set_active(__pa(pa_curr->page[idx]));
__pa              134 arch/powerpc/platforms/pseries/dtl.c 	addr = __pa(dtl->buf);
__pa              932 arch/powerpc/platforms/pseries/hotplug-cpu.c 			 __pa(cede_parameters),
__pa              127 arch/powerpc/platforms/pseries/io_event_irq.c 				    __pa(ioei_rtas_buf),
__pa              106 arch/powerpc/platforms/pseries/iommu.c 		rpn = __pa(uaddr) >> TCE_SHIFT;
__pa              149 arch/powerpc/platforms/pseries/iommu.c 	rpn = __pa(uaddr) >> tceshift;
__pa              220 arch/powerpc/platforms/pseries/iommu.c 	rpn = __pa(uaddr) >> TCE_SHIFT;
__pa              240 arch/powerpc/platforms/pseries/iommu.c 					    (u64)__pa(tcep),
__pa              462 arch/powerpc/platforms/pseries/iommu.c 					    (u64)__pa(tcep),
__pa              129 arch/powerpc/platforms/pseries/lpar.c 		ret = register_dtl(hwcpu, __pa(dtl));
__pa              673 arch/powerpc/platforms/pseries/lpar.c 	addr = __pa(&lppaca_of(cpu));
__pa              688 arch/powerpc/platforms/pseries/lpar.c 		addr = __pa(paca_ptrs[cpu]->slb_shadow_ptr);
__pa             1430 arch/powerpc/platforms/pseries/lpar.c 				__pa(rtas_data_buf),
__pa             1712 arch/powerpc/platforms/pseries/lpar.c 	pseries_lpar_register_process_table(__pa(process_tb),
__pa             1749 arch/powerpc/platforms/pseries/lpar.c 	addr = __pa((unsigned long)page_address(page));
__pa              306 arch/powerpc/platforms/pseries/lparcfg.c 				__pa(rtas_data_buf),
__pa               63 arch/powerpc/platforms/pseries/nvram.c 		if ((rtas_call(nvram_fetch, 3, 2, &done, i, __pa(nvram_buf),
__pa              108 arch/powerpc/platforms/pseries/nvram.c 		if ((rtas_call(nvram_store, 3, 2, &done, i, __pa(nvram_buf),
__pa              207 arch/powerpc/platforms/pseries/pseries_energy.c 	rc = plpar_hcall9(H_BEST_ENERGY, retbuf, flags, 0, __pa(buf_page),
__pa              294 arch/powerpc/platforms/pseries/ras.c 		  RTAS_HOTPLUG_EVENTS, 0, __pa(&ras_log_buf),
__pa              337 arch/powerpc/platforms/pseries/ras.c 			   critical, __pa(&ras_log_buf),
__pa              368 arch/powerpc/platforms/pseries/ras.c 			   __pa(&ras_log_buf),
__pa              519 arch/powerpc/platforms/pseries/rtas-fadump.c 		rtas_fadump_get_config(fadump_conf, (void *)__pa(fdm_active));
__pa               70 arch/powerpc/platforms/pseries/scanlog.c 				   (u32) __pa(rtas_data_buf), (u32) count);
__pa              123 arch/powerpc/platforms/pseries/setup.c 	system_reset_addr  = __pa(system_reset_fwnmi) - PHYSICAL_START;
__pa              124 arch/powerpc/platforms/pseries/setup.c 	machine_check_addr = __pa(machine_check_fwnmi) - PHYSICAL_START;
__pa              862 arch/powerpc/platforms/pseries/setup.c 				__pa(rtas_data_buf),
__pa               93 arch/powerpc/platforms/pseries/smp.c 			__pa(ppc_function_entry(generic_secondary_smp_init));
__pa               42 arch/powerpc/platforms/pseries/svm.c 	uv_unshare_page(PHYS_PFN(__pa(addr)), numpages);
__pa               52 arch/powerpc/platforms/pseries/svm.c 	uv_share_page(PHYS_PFN(__pa(addr)), numpages);
__pa               76 arch/powerpc/platforms/pseries/svm.c 	unsigned long pfn = PHYS_PFN(__pa(addr));
__pa              190 arch/powerpc/sysdev/dart_iommu.c 		rpn = __pa(uaddr) >> DART_PAGE_SHIFT;
__pa              142 arch/powerpc/sysdev/xive/native.c 		qpage_phys = __pa(qpage);
__pa              649 arch/powerpc/sysdev/xive/native.c 		opal_xive_donate_page(chip, __pa(p));
__pa              472 arch/powerpc/sysdev/xive/spapr.c 		qpage_phys = __pa(qpage);
__pa              103 arch/riscv/include/asm/page.h #define virt_to_pfn(vaddr)	(phys_to_pfn(__pa(vaddr)))
__pa               69 arch/riscv/mm/init.c 	if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
__pa               75 arch/riscv/mm/init.c 	memblock_reserve(__pa(initrd_start), size);
__pa               94 arch/riscv/mm/init.c 	phys_addr_t vmlinux_end = __pa(&_end);
__pa               95 arch/riscv/mm/init.c 	phys_addr_t vmlinux_start = __pa(&_start);
__pa              274 arch/riscv/mm/init.c #define get_pgd_next_virt(__pa)	get_pmd_virt(__pa)
__pa              275 arch/riscv/mm/init.c #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot)	\
__pa              276 arch/riscv/mm/init.c 	create_pmd_mapping(__nextp, __va, __pa, __sz, __prot)
__pa              282 arch/riscv/mm/init.c #define get_pgd_next_virt(__pa)	get_pte_virt(__pa)
__pa              283 arch/riscv/mm/init.c #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot)	\
__pa              284 arch/riscv/mm/init.c 	create_pte_mapping(__nextp, __va, __pa, __sz, __prot)
__pa              421 arch/riscv/mm/init.c 			   __pa(fixmap_pgd_next),
__pa              433 arch/riscv/mm/init.c 		if (start <= __pa(PAGE_OFFSET) &&
__pa              434 arch/riscv/mm/init.c 		    __pa(PAGE_OFFSET) < end)
__pa              435 arch/riscv/mm/init.c 			start = __pa(PAGE_OFFSET);
__pa              450 arch/riscv/mm/init.c 	csr_write(CSR_SATP, PFN_DOWN(__pa(swapper_pg_dir)) | SATP_MODE);
__pa               28 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__sdma) = __pa(&_sdma);
__pa               29 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__edma) = __pa(&_edma);
__pa               30 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__stext_dma) = __pa(&_stext_dma);
__pa               31 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__etext_dma) = __pa(&_etext_dma);
__pa               52 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__swsusp_reset_dma) = __pa(_swsusp_reset_dma);
__pa               32 arch/s390/include/asm/idals.h 	return ((__pa(vaddr) + length - 1) >> 31) != 0;
__pa               41 arch/s390/include/asm/idals.h 	return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length +
__pa               54 arch/s390/include/asm/idals.h 	paddr = __pa(vaddr);
__pa               47 arch/s390/include/asm/mmu_context.h 		mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa               52 arch/s390/include/asm/mmu_context.h 		mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa               57 arch/s390/include/asm/mmu_context.h 		mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa               62 arch/s390/include/asm/mmu_context.h 		mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa              164 arch/s390/include/asm/page.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa              177 arch/s390/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              104 arch/s390/include/asm/pgalloc.h 	pgd_val(*pgd) = _REGION1_ENTRY | __pa(p4d);
__pa              109 arch/s390/include/asm/pgalloc.h 	p4d_val(*p4d) = _REGION2_ENTRY | __pa(pud);
__pa              114 arch/s390/include/asm/pgalloc.h 	pud_val(*pud) = _REGION3_ENTRY | __pa(pmd);
__pa              143 arch/s390/include/asm/pgalloc.h 	pmd_val(*pmd) = _SEGMENT_ENTRY + __pa(pte);
__pa             1287 arch/s390/include/asm/pgtable.h #define pfn_pte(pfn,pgprot) mk_pte_phys(__pa((pfn) << PAGE_SHIFT),(pgprot))
__pa             1615 arch/s390/include/asm/pgtable.h #define pfn_pmd(pfn, pgprot)	mk_pmd_phys(__pa((pfn) << PAGE_SHIFT), (pgprot))
__pa              142 arch/s390/kernel/crash_dump.c 		from = __pa(src);
__pa              185 arch/s390/kernel/crash_dump.c 		from = __pa(src);
__pa               58 arch/s390/kernel/early.c 	end_pfn = PFN_UP(__pa(_end));
__pa              832 arch/s390/kernel/setup.c 	unsigned long start_pfn = PFN_UP(__pa(_end));
__pa              586 arch/s390/kernel/smp.c 	pa = __pa(&pcpu->lowcore->floating_pt_save_area);
__pa              592 arch/s390/kernel/smp.c 	pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK);
__pa              154 arch/s390/kernel/suspend.c 	unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin));
__pa              155 arch/s390/kernel/suspend.c 	unsigned long nosave_end_pfn = PFN_DOWN(__pa(&__nosave_end));
__pa              156 arch/s390/kernel/suspend.c 	unsigned long end_rodata_pfn = PFN_DOWN(__pa(__end_rodata)) - 1;
__pa              157 arch/s390/kernel/suspend.c 	unsigned long stext_pfn = PFN_DOWN(__pa(_stext));
__pa             1280 arch/s390/kvm/vsie.c 	struct page *page = pfn_to_page(__pa(vsie_page) >> PAGE_SHIFT);
__pa              158 arch/s390/mm/dump_pagetables.c 	if ((pud_val(*pud) & PAGE_MASK) == __pa(kasan_early_shadow_pmd)) {
__pa              189 arch/s390/mm/dump_pagetables.c 	if ((p4d_val(*p4d) & PAGE_MASK) == __pa(kasan_early_shadow_pud)) {
__pa              219 arch/s390/mm/dump_pagetables.c 	if ((pgd_val(*pgd) & PAGE_MASK) == __pa(kasan_early_shadow_p4d)) {
__pa               80 arch/s390/mm/gmap.c 		_ASCE_USER_BITS | __pa(table);
__pa             1339 arch/s390/mm/gmap.c 	page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT);
__pa             1367 arch/s390/mm/gmap.c 		page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT);
__pa             1396 arch/s390/mm/gmap.c 	page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT);
__pa             1424 arch/s390/mm/gmap.c 		page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT);
__pa             1453 arch/s390/mm/gmap.c 	page = pfn_to_page(__pa(r3t) >> PAGE_SHIFT);
__pa             1481 arch/s390/mm/gmap.c 		page = pfn_to_page(__pa(r3t) >> PAGE_SHIFT);
__pa             1510 arch/s390/mm/gmap.c 	page = pfn_to_page(__pa(r2t) >> PAGE_SHIFT);
__pa             1541 arch/s390/mm/gmap.c 		page = pfn_to_page(__pa(r2t) >> PAGE_SHIFT);
__pa              102 arch/s390/mm/init.c 	init_mm.context.asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits;
__pa              168 arch/s390/mm/kasan_init.c 				pmd_val(*pm_dir) = __pa(page) | sgt_prot;
__pa              187 arch/s390/mm/kasan_init.c 				pte_val(*pt_dir) = __pa(page) | pgt_prot;
__pa              192 arch/s390/mm/kasan_init.c 				pte_val(*pt_dir) = __pa(page) | pgt_prot;
__pa              196 arch/s390/mm/kasan_init.c 				pte_val(*pt_dir) = __pa(page) | pgt_prot_zero;
__pa              209 arch/s390/mm/kasan_init.c 	S390_lowcore.kernel_asce = (__pa(pgd) & PAGE_MASK) | asce_bits;
__pa              249 arch/s390/mm/kasan_init.c 	pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY);
__pa              250 arch/s390/mm/kasan_init.c 	pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY);
__pa              251 arch/s390/mm/kasan_init.c 	p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY);
__pa              256 arch/s390/mm/kasan_init.c 	pte_z = __pte(__pa(kasan_early_shadow_page) | pgt_prot);
__pa              132 arch/s390/mm/pageattr.c 	pmd_val(new) = __pa(pt_dir) | _SEGMENT_ENTRY;
__pa              209 arch/s390/mm/pageattr.c 	pud_val(new) = __pa(pm_dir) | _REGION3_ENTRY;
__pa              112 arch/s390/mm/pgalloc.c 			mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa              120 arch/s390/mm/pgalloc.c 			mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa              147 arch/s390/mm/pgalloc.c 	mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
__pa              253 arch/s390/mm/pgalloc.c 	page = pfn_to_page(__pa(table) >> PAGE_SHIFT);
__pa              256 arch/s390/mm/pgalloc.c 		bit = (__pa(table) & ~PAGE_MASK)/(PTRS_PER_PTE*sizeof(pte_t));
__pa              283 arch/s390/mm/pgalloc.c 	page = pfn_to_page(__pa(table) >> PAGE_SHIFT);
__pa              286 arch/s390/mm/pgalloc.c 		table = (unsigned long *) (__pa(table) | 3);
__pa              290 arch/s390/mm/pgalloc.c 	bit = (__pa(table) & ~PAGE_MASK) / (PTRS_PER_PTE*sizeof(pte_t));
__pa              299 arch/s390/mm/pgalloc.c 	table = (unsigned long *) (__pa(table) | (1U << bit));
__pa              307 arch/s390/mm/pgalloc.c 	struct page *page = pfn_to_page(__pa(table) >> PAGE_SHIFT);
__pa              270 arch/s390/mm/vmem.c 				pmd_val(*pm_dir) = __pa(new_page) | sgt_prot;
__pa              290 arch/s390/mm/vmem.c 			pte_val(*pt_dir) = __pa(new_page) | pgt_prot;
__pa              131 arch/sh/boot/compressed/misc.c 	output_addr = __pa((unsigned long)&_text+PAGE_SIZE);
__pa              101 arch/sh/drivers/pci/pci-sh5.c 	unsigned long memStart = __pa(memory_start);
__pa              102 arch/sh/drivers/pci/pci-sh5.c 	unsigned long memSize = __pa(memory_end) - memStart;
__pa              300 arch/sh/drivers/pci/pci-sh7780.c 	memphys = __pa(memory_start);
__pa              359 arch/sh/drivers/pci/pcie-sh7786.c 	memstart = __pa(memory_start);
__pa              360 arch/sh/drivers/pci/pcie-sh7786.c 	memend   = __pa(memory_end);
__pa              246 arch/sh/include/asm/io.h #define virt_to_phys(address)	(__pa(address))
__pa              179 arch/sh/include/asm/page.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa              183 arch/sh/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              168 arch/sh/kernel/setup.c 	memblock_reserve(__pa(initrd_start), INITRD_SIZE);
__pa               21 arch/sh/kernel/swsusp.c 	unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
__pa               22 arch/sh/kernel/swsusp.c 	unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
__pa               49 arch/sh/mm/cache-sh3.c 			p = __pa(v);
__pa              144 arch/sh/mm/cache-sh7705.c 		__flush_dcache_page(__pa(page_address(page)));
__pa              255 arch/sh/mm/init.c 	start_pfn = PFN_UP(__pa(_end));
__pa              155 arch/sh/mm/mmap.c 	if (addr + count > __pa(high_memory))
__pa               89 arch/sh/mm/pmb.c 	return ppn >= __pa(memory_start) && ppn < __pa(memory_end);
__pa               41 arch/sparc/include/asm/mmu_context_64.h 	__tsb_context_switch(__pa(mm->pgd),
__pa               50 arch/sparc/include/asm/mmu_context_64.h 			     , __pa(&mm->context.tsb_descr[MM_TSB_BASE]),
__pa              127 arch/sparc/include/asm/page_32.h #define virt_to_phys		__pa
__pa              131 arch/sparc/include/asm/page_32.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa              152 arch/sparc/include/asm/page_64.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
__pa              154 arch/sparc/include/asm/page_64.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              156 arch/sparc/include/asm/page_64.h #define virt_to_phys __pa
__pa              400 arch/sparc/include/asm/pgtable_32.h 	(test_bit(__pa((unsigned long)(addr))>>20, sparc_valid_addr_bitmap))
__pa              830 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = __pa((unsigned long) (ptep));
__pa              836 arch/sparc/include/asm/pgtable_64.h 	(pud_val(*(pudp)) = (__pa((unsigned long) (pmdp))))
__pa              888 arch/sparc/include/asm/pgtable_64.h 	(pgd_val(*(pgdp)) = (__pa((unsigned long) (pudp))))
__pa              144 arch/sparc/include/asm/pgtsrmmu.h #define __nocache_pa(VADDR) (((unsigned long)VADDR) - SRMMU_NOCACHE_VADDR + __pa((unsigned long)srmmu_nocache_pool))
__pa              129 arch/sparc/kernel/iommu.c 	iommu->dummy_page_pa = (unsigned long) __pa(iommu->dummy_page);
__pa              234 arch/sparc/kernel/iommu.c 	first_page = __pa(first_page);
__pa              299 arch/sparc/kernel/iommu.c 	base_paddr = __pa(oaddr & IO_PAGE_MASK);
__pa               36 arch/sparc/kernel/iommu_common.h #define SG_ENT_PHYS_ADDRESS(SG)	(__pa(sg_virt((SG))))
__pa              282 arch/sparc/kernel/irq_64.c 	irq = bucket_get_irq(__pa(bucket));
__pa              623 arch/sparc/kernel/irq_64.c 	irq = bucket_get_irq(__pa(bucket));
__pa              626 arch/sparc/kernel/irq_64.c 		bucket_set_irq(__pa(bucket), irq);
__pa              688 arch/sparc/kernel/irq_64.c 	cookie = ~__pa(&ihd->bucket);
__pa              744 arch/sparc/kernel/irq_64.c 	bucket_set_irq(__pa(bucket), irq);
__pa             1030 arch/sparc/kernel/irq_64.c 	*pa_ptr = __pa(p);
__pa             1048 arch/sparc/kernel/irq_64.c 	tb->cpu_mondo_block_pa = __pa(mondo);
__pa             1056 arch/sparc/kernel/irq_64.c 	tb->cpu_list_pa = __pa(page);
__pa             1116 arch/sparc/kernel/irq_64.c 	ivector_table_pa = __pa(ivector_table);
__pa              994 arch/sparc/kernel/ldc.c 	*ra = __pa(q);
__pa             1081 arch/sparc/kernel/ldc.c 	hv_err = sun4v_ldc_set_map_table(lp->id, __pa(table),
__pa             2206 arch/sparc/kernel/ldc.c 	pa = __pa(buf);
__pa             2277 arch/sparc/kernel/ldc.c 	ra = __pa(buf);
__pa              430 arch/sparc/kernel/leon_pci_grpci1.c 	ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN((unsigned long) &_end));
__pa              623 arch/sparc/kernel/leon_pci_grpci2.c 				ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN(
__pa              191 arch/sparc/kernel/mdesc.c 	start = __pa(hp);
__pa              516 arch/sparc/kernel/mdesc.c 	status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len);
__pa             1339 arch/sparc/kernel/mdesc.c 	status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len);
__pa               64 arch/sparc/kernel/pci_fire.c 	upa_writeq(__pa(iommu->page_table) | 0x7UL, iommu->iommu_tsbbase);
__pa              246 arch/sparc/kernel/pci_fire.c 		    __pa(pbm->msi_queues)),
__pa              436 arch/sparc/kernel/pci_msi.c 		       __pa(pbm->msi_queues));
__pa              453 arch/sparc/kernel/pci_psycho.c 		__pa(pbm->stc.strbuf_flushflag);
__pa             1112 arch/sparc/kernel/pci_schizo.c 		__pa(pbm->stc.strbuf_flushflag);
__pa             1205 arch/sparc/kernel/pci_schizo.c 	upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
__pa              105 arch/sparc/kernel/pci_sun4v.c 						  __pa(pglist));
__pa              111 arch/sparc/kernel/pci_sun4v.c 						   npages, prot, __pa(pglist),
__pa              122 arch/sparc/kernel/pci_sun4v.c 						  __pa(pglist),
__pa              129 arch/sparc/kernel/pci_sun4v.c 						   __pa(pglist), ret);
__pa              226 arch/sparc/kernel/pci_sun4v.c 	first_page = __pa(first_page);
__pa              392 arch/sparc/kernel/pci_sun4v.c 	base_paddr = __pa(oaddr & IO_PAGE_MASK);
__pa              767 arch/sparc/kernel/pci_sun4v.c 	iotsb->ra = __pa(table);
__pa             1074 arch/sparc/kernel/pci_sun4v.c 		unsigned long err, base = __pa(pages + (i * q_size));
__pa              429 arch/sparc/kernel/psycho_common.c 	upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
__pa              126 arch/sparc/kernel/ptrace_64.c 		unsigned long start = __pa(kaddr);
__pa              586 arch/sparc/kernel/sbus.c 		__pa(strbuf->strbuf_flushflag);
__pa              625 arch/sparc/kernel/sbus.c 	upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase);
__pa              338 arch/sparc/kernel/smp_64.c 				 __pa(hdesc));
__pa              901 arch/sparc/kernel/smp_64.c 	if (tp->pgd_paddr == __pa(mm->pgd))
__pa              936 arch/sparc/kernel/smp_64.c 		__flush_icache_page(__pa(page_address(page)));
__pa              969 arch/sparc/kernel/smp_64.c 			xcall_deliver(data0, __pa(pg_addr),
__pa             1005 arch/sparc/kernel/smp_64.c 		xcall_deliver(data0, __pa(pg_addr),
__pa             1585 arch/sparc/kernel/smp_64.c 	const unsigned long goal = __pa(MAX_DMA_ADDRESS);
__pa             1595 arch/sparc/kernel/smp_64.c 			 cpu, size, __pa(ptr));
__pa             1600 arch/sparc/kernel/smp_64.c 			 "%016lx\n", cpu, size, node, __pa(ptr));
__pa             1610 arch/sparc/kernel/smp_64.c 	memblock_free(__pa(ptr), size);
__pa              116 arch/sparc/kernel/sysfs.c 		ra = __pa(&per_cpu(mmu_stats, smp_processor_id()));
__pa              173 arch/sparc/mm/init_32.c 	start_pfn  = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end));
__pa              212 arch/sparc/mm/init_64.c 		__flush_icache_page(__pa(page_address(page)));
__pa              274 arch/sparc/mm/init_64.c 		tsb_addr = __pa(tsb_addr);
__pa              759 arch/sparc/mm/init_64.c 		start = __pa(start);
__pa              760 arch/sparc/mm/init_64.c 		end = __pa(end);
__pa             1661 arch/sparc/mm/init_64.c 		unsigned long pa = __pa(addr);
__pa             1801 arch/sparc/mm/init_64.c 		unsigned long this_end, paddr = __pa(vstart);
__pa             2495 arch/sparc/mm/init_64.c 	if (paddr >= __pa(initrd_start) &&
__pa             2496 arch/sparc/mm/init_64.c 	    paddr < __pa(PAGE_ALIGN(initrd_end)))
__pa             2633 arch/sparc/mm/init_64.c 			pmd_val(*pmd) = pte_base | __pa(block);
__pa              133 arch/sparc/mm/io-unit.c 	iopte = MKIOPTE(__pa(vaddr & PAGE_MASK));
__pa              255 arch/sparc/mm/io-unit.c 			sbus_writel(iopte_val(MKIOPTE(__pa(page))), iopte);
__pa              108 arch/sparc/mm/iommu.c 	base = __pa((unsigned long)iommu->page_table) >> 4;
__pa              327 arch/sparc/mm/srmmu.c 	paddr = __pa((unsigned long)srmmu_nocache_pool);
__pa              130 arch/sparc/mm/tsb.c 			base = __pa(base);
__pa              144 arch/sparc/mm/tsb.c 			base = __pa(base);
__pa              163 arch/sparc/mm/tsb.c 			base = __pa(base);
__pa              178 arch/sparc/mm/tsb.c 			base = __pa(base);
__pa              216 arch/sparc/mm/tsb.c 	tsb_paddr = __pa(mm->context.tsb_block[tsb_idx].tsb);
__pa              506 arch/sparc/mm/tsb.c 			old_tsb_base = __pa(old_tsb_base);
__pa              507 arch/sparc/mm/tsb.c 			new_tsb_base = __pa(new_tsb_base);
__pa              114 arch/um/drivers/mmapper_kern.c 	p_buf = __pa(v_buf);
__pa              565 arch/um/drivers/virtio_uml.c 		rc = vhost_user_init_mem_region(__pa(end_iomem), highmem,
__pa              112 arch/um/include/asm/page.h #define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
__pa               16 arch/um/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
__pa               61 arch/um/include/asm/pgtable-3level.h 	set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
__pa              283 arch/um/include/asm/pgtable.h #define __virt_to_page(virt) phys_to_page(__pa(virt))
__pa               49 arch/um/kernel/mem.c 	map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0);
__pa               50 arch/um/kernel/mem.c 	memblock_free(__pa(brk_end), uml_reserved - brk_end);
__pa               75 arch/um/kernel/mem.c 					   (unsigned long) __pa(pte)));
__pa               89 arch/um/kernel/mem.c 	set_pud(pud, __pud(_KERNPG_TABLE + (unsigned long) __pa(pmd_table)));
__pa              143 arch/um/kernel/mem.c 	p = __pa(v);
__pa              107 arch/um/kernel/physmem.c 	os_seek_file(physmem_fd, __pa(__syscall_stub_start));
__pa              111 arch/um/kernel/physmem.c 	memblock_add(__pa(start), len + highmem);
__pa              112 arch/um/kernel/physmem.c 	memblock_reserve(__pa(start), reserve);
__pa              114 arch/um/kernel/physmem.c 	min_low_pfn = PFN_UP(__pa(reserve_end));
__pa              126 arch/um/kernel/physmem.c 	else if (phys < __pa(end_iomem)) {
__pa              139 arch/um/kernel/physmem.c 	else if (phys < __pa(end_iomem) + highmem) {
__pa              212 arch/um/kernel/physmem.c 			region->phys = __pa(region->virt);
__pa               94 arch/unicore32/include/asm/memory.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa               76 arch/unicore32/include/asm/pgalloc.h 	__pmd_populate(pmdp, __pa(pte_ptr) | _PAGE_KERNEL_TABLE);
__pa              156 arch/unicore32/include/asm/tlbflush.h 		: : "r" (__pa(pmd) & ~(L1_CACHE_BYTES - 1)) : "cc");
__pa               55 arch/unicore32/kernel/hibernate.c 		set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_KERNEL_TABLE));
__pa              144 arch/unicore32/kernel/hibernate.c 	unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
__pa              145 arch/unicore32/kernel/hibernate.c 	unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
__pa              134 arch/unicore32/mm/init.c 	memblock_reserve(__pa(_text), _end - _text);
__pa              215 arch/unicore32/mm/init.c 	pg = PAGE_ALIGN(__pa(start_pg));
__pa              216 arch/unicore32/mm/init.c 	pgend = __pa(end_pg) & PAGE_MASK;
__pa              152 arch/unicore32/mm/mmu.c 		__pmd_populate(pmd, __pa(pte) | prot);
__pa              284 arch/unicore32/mm/mmu.c 	lowmem_limit = __pa(vmalloc_min - 1) + 1;
__pa              334 arch/unicore32/mm/mmu.c 	memblock_reserve(__pa(swapper_pg_dir), PTRS_PER_PGD * sizeof(pgd_t));
__pa              117 arch/x86/entry/vdso/vma.c 					__pa(pvti) >> PAGE_SHIFT,
__pa              131 arch/x86/include/asm/io.h 	return __pa(address);
__pa               90 arch/x86/include/asm/mem_encrypt.h #define __sme_pa(x)		(__pa(x) | sme_me_mask)
__pa               41 arch/x86/include/asm/page.h #ifndef __pa
__pa               63 arch/x86/include/asm/page.h #define __boot_pa(x)		__pa(x)
__pa               69 arch/x86/include/asm/page.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa               66 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
__pa               67 arch/x86/include/asm/pgalloc.h 	set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
__pa               73 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
__pa               74 arch/x86/include/asm/pgalloc.h 	set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
__pa              126 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
__pa              127 arch/x86/include/asm/pgalloc.h 	set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd)));
__pa              132 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
__pa              133 arch/x86/include/asm/pgalloc.h 	set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd)));
__pa              140 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
__pa              141 arch/x86/include/asm/pgalloc.h 	set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud)));
__pa              146 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
__pa              147 arch/x86/include/asm/pgalloc.h 	set_p4d_safe(p4d, __p4d(_PAGE_TABLE | __pa(pud)));
__pa              178 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
__pa              179 arch/x86/include/asm/pgalloc.h 	set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d)));
__pa              186 arch/x86/include/asm/pgalloc.h 	paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
__pa              187 arch/x86/include/asm/pgalloc.h 	set_pgd_safe(pgd, __pgd(_PAGE_TABLE | __pa(p4d)));
__pa              492 arch/x86/include/asm/uv/uv_hub.h 	return uv_soc_phys_ram_to_nasid(__pa(v));
__pa              519 arch/x86/include/asm/uv/uv_hub.h 	return uv_soc_phys_ram_to_gpa(__pa(v));
__pa              306 arch/x86/include/asm/xen/page.h #define virt_to_machine(v)	(phys_to_machine(XPADDR(__pa(v))))
__pa              307 arch/x86/include/asm/xen/page.h #define virt_to_pfn(v)          (PFN_DOWN(__pa(v)))
__pa              552 arch/x86/kernel/amd_gart_64.c 		enable_gart_translation(dev, __pa(agp_gatt_table));
__pa              806 arch/x86/kernel/amd_gart_64.c 	gart_unmapped_entry = GPTE_ENCODE(__pa(scratch));
__pa             2692 arch/x86/kernel/apic/io_apic.c 			ioapic_phys = __pa(ioapic_phys);
__pa              155 arch/x86/kernel/check.c 			pr_err("Corrupted low memory at %p (%lx phys) = %08lx\n", addr, __pa(addr), *addr);
__pa             1507 arch/x86/kernel/cpu/resctrl/pseudo_lock.c 	physical = __pa(plr->kmem) >> PAGE_SHIFT;
__pa              170 arch/x86/kernel/espfix_64.c 		pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask));
__pa              171 arch/x86/kernel/espfix_64.c 		paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT);
__pa              182 arch/x86/kernel/espfix_64.c 		pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask));
__pa              183 arch/x86/kernel/espfix_64.c 		paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT);
__pa              194 arch/x86/kernel/espfix_64.c 	pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask));
__pa               74 arch/x86/kernel/head32.c #ifdef __pa
__pa               82 arch/x86/kernel/head32.c 	const unsigned long limit = __pa(_end) +
__pa               85 arch/x86/kernel/head32.c 	pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd);
__pa               88 arch/x86/kernel/head32.c 	pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table);
__pa               92 arch/x86/kernel/head32.c 	ptep = (pte_t *)__pa(__brk_base);
__pa              112 arch/x86/kernel/head32.c 	ptr = (unsigned long *)__pa(&max_pfn_mapped);
__pa              116 arch/x86/kernel/head32.c 	ptr = (unsigned long *)__pa(&_brk_end);
__pa              101 arch/x86/kernel/machine_kexec_32.c 		set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT));
__pa              107 arch/x86/kernel/machine_kexec_32.c 		set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
__pa              123 arch/x86/kernel/machine_kexec_32.c 		(unsigned long)control_page, __pa(control_page));
__pa              129 arch/x86/kernel/machine_kexec_32.c 		__pa(control_page), __pa(control_page));
__pa              213 arch/x86/kernel/machine_kexec_32.c 	page_list[PA_CONTROL_PAGE] = __pa(control_page);
__pa              215 arch/x86/kernel/machine_kexec_32.c 	page_list[PA_PGD] = __pa(image->arch.pgd);
__pa              135 arch/x86/kernel/machine_kexec_64.c 	paddr = __pa(page_address(image->control_code_page)+PAGE_SIZE);
__pa              142 arch/x86/kernel/machine_kexec_64.c 		set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE));
__pa              150 arch/x86/kernel/machine_kexec_64.c 		set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE));
__pa              158 arch/x86/kernel/machine_kexec_64.c 		set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE));
__pa              166 arch/x86/kernel/machine_kexec_64.c 		set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
__pa              409 arch/x86/kernel/machine_kexec_64.c 	  (unsigned long)__pa(page_address(image->control_code_page));
__pa              771 arch/x86/kernel/pci-calgary_64.c 	table_phys = (u64)__pa(tbl->it_base);
__pa              103 arch/x86/kernel/setup_percpu.c 	const unsigned long goal = __pa(MAX_DMA_ADDRESS);
__pa              113 arch/x86/kernel/setup_percpu.c 			 cpu, size, __pa(ptr));
__pa              120 arch/x86/kernel/setup_percpu.c 			 cpu, size, node, __pa(ptr));
__pa              138 arch/x86/kernel/setup_percpu.c 	memblock_free(__pa(ptr), size);
__pa              176 arch/x86/kernel/tce_64.c 	memblock_free(__pa(tbl), size);
__pa              676 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(sptep));
__pa              760 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(sptep));
__pa             1502 arch/x86/kvm/mmu.c 	sp = page_header(__pa(spte));
__pa             1514 arch/x86/kvm/mmu.c 	sp = page_header(__pa(spte));
__pa             1606 arch/x86/kvm/mmu.c 		WARN_ON(page_header(__pa(sptep))->role.level ==
__pa             1619 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *sp = page_header(__pa(sptep));
__pa             2093 arch/x86/kvm/mmu.c 	sp = page_header(__pa(spte));
__pa             2215 arch/x86/kvm/mmu.c 	sp = page_header(__pa(spte));
__pa             2545 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(spte));
__pa             2710 arch/x86/kvm/mmu.c 	spte = __pa(sp->spt) | shadow_present_mask | PT_WRITABLE_MASK |
__pa             3061 arch/x86/kvm/mmu.c 	sp = page_header(__pa(sptep));
__pa             3286 arch/x86/kvm/mmu.c 	sp = page_header(__pa(sptep));
__pa             3561 arch/x86/kvm/mmu.c 		sp = page_header(__pa(iterator.sptep));
__pa             3792 arch/x86/kvm/mmu.c 		vcpu->arch.mmu->root_hpa = __pa(sp->spt);
__pa             3805 arch/x86/kvm/mmu.c 			root = __pa(sp->spt);
__pa             3810 arch/x86/kvm/mmu.c 		vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root);
__pa             3847 arch/x86/kvm/mmu.c 		root = __pa(sp->spt);
__pa             3884 arch/x86/kvm/mmu.c 		root = __pa(sp->spt);
__pa             3890 arch/x86/kvm/mmu.c 	vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root);
__pa             3909 arch/x86/kvm/mmu.c 			lm_root[0] = __pa(vcpu->arch.mmu->pae_root) | pm_mask;
__pa             3914 arch/x86/kvm/mmu.c 		vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->lm_root);
__pa             6023 arch/x86/kvm/mmu.c 		sp = page_header(__pa(sptep));
__pa              100 arch/x86/kvm/mmu_audit.c 	sp = page_header(__pa(sptep));
__pa              135 arch/x86/kvm/mmu_audit.c 	rev_sp = page_header(__pa(sptep));
__pa              168 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *sp = page_header(__pa(sptep));
__pa              586 arch/x86/kvm/paging_tmpl.h 	sp = page_header(__pa(sptep));
__pa              917 arch/x86/kvm/paging_tmpl.h 		sp = page_header(__pa(sptep));
__pa             1603 arch/x86/kvm/svm.c 	control->msrpm_base_pa = __sme_set(__pa(svm->msrpm));
__pa             3521 arch/x86/kvm/svm.c 	svm->vmcb->control.msrpm_base_pa = __sme_set(__pa(svm->nested.msrpm));
__pa             6446 arch/x86/kvm/svm.c 		start->dh_cert_address = __sme_set(__pa(dh_blob));
__pa             6458 arch/x86/kvm/svm.c 		start->session_address = __sme_set(__pa(session_blob));
__pa             2000 arch/x86/kvm/vmx/nested.c 		vmcs_write64(MSR_BITMAP, __pa(vmx->nested.vmcs02.msr_bitmap));
__pa             2023 arch/x86/kvm/vmx/nested.c 	vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val));
__pa             2024 arch/x86/kvm/vmx/nested.c 	vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
__pa             4803 arch/x86/kvm/vmx/nested.c 			     __pa(vmx->vmcs01.shadow_vmcs));
__pa             5778 arch/x86/kvm/vmx/nested.c 		vmcs_write64(VMREAD_BITMAP, __pa(vmx_vmread_bitmap));
__pa             5779 arch/x86/kvm/vmx/nested.c 		vmcs_write64(VMWRITE_BITMAP, __pa(vmx_vmwrite_bitmap));
__pa              236 arch/x86/kvm/vmx/ops.h 	u64 phys_addr = __pa(vmcs);
__pa              243 arch/x86/kvm/vmx/ops.h 	u64 phys_addr = __pa(vmcs);
__pa              554 arch/x86/kvm/vmx/vmx.c 		__pa(*p_hv_pa_pg);
__pa             2226 arch/x86/kvm/vmx/vmx.c 	u64 phys_addr = __pa(per_cpu(vmxarea, cpu));
__pa             4155 arch/x86/kvm/vmx/vmx.c 		vmcs_write64(MSR_BITMAP, __pa(vmx->vmcs01.msr_bitmap));
__pa             4179 arch/x86/kvm/vmx/vmx.c 		vmcs_write64(POSTED_INTR_DESC_ADDR, __pa((&vmx->pi_desc)));
__pa             4203 arch/x86/kvm/vmx/vmx.c 	vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val));
__pa             4205 arch/x86/kvm/vmx/vmx.c 	vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val));
__pa             4331 arch/x86/kvm/vmx/vmx.c 				     __pa(vcpu->arch.apic->regs));
__pa             7519 arch/x86/kvm/vmx/vmx.c 		vcpu_info.pi_desc_addr = __pa(vcpu_to_pi_desc(vcpu));
__pa              385 arch/x86/mm/dump_pagetables.c 	if (__pa(pt) == __pa(kasan_early_shadow_pmd) ||
__pa              387 arch/x86/mm/dump_pagetables.c 			__pa(pt) == __pa(kasan_early_shadow_p4d)) ||
__pa              388 arch/x86/mm/dump_pagetables.c 	    __pa(pt) == __pa(kasan_early_shadow_pud)) {
__pa               55 arch/x86/mm/ident_map.c 		set_pud(pud, __pud(__pa(pmd) | info->kernpg_flag));
__pa               83 arch/x86/mm/ident_map.c 		set_p4d(p4d, __p4d(__pa(pud) | info->kernpg_flag));
__pa              127 arch/x86/mm/ident_map.c 			set_pgd(pgd, __pgd(__pa(p4d) | info->kernpg_flag));
__pa              134 arch/x86/mm/ident_map.c 			set_pgd(pgd, __pgd(__pa(pud) | info->kernpg_flag));
__pa              115 arch/x86/mm/init.c 			ret = __pa(extend_brk(PAGE_SIZE * num, PAGE_SIZE));
__pa              156 arch/x86/mm/init.c 	base = __pa(extend_brk(tables, PAGE_SIZE));
__pa               76 arch/x86/mm/init_32.c 		paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT);
__pa               77 arch/x86/mm/init_32.c 		set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
__pa              101 arch/x86/mm/init_32.c 		paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT);
__pa              102 arch/x86/mm/init_32.c 		set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
__pa              183 arch/x86/mm/init_32.c 		paravirt_alloc_pte(&init_mm, __pa(newpte) >> PAGE_SHIFT);
__pa              184 arch/x86/mm/init_32.c 		set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE));
__pa              188 arch/x86/mm/init_32.c 		paravirt_release_pte(__pa(pte) >> PAGE_SHIFT);
__pa              504 arch/x86/mm/init_32.c 				pfn, pmd, __pa(pmd));
__pa              513 arch/x86/mm/init_32.c 				pfn, pmd, __pa(pmd), pte, __pa(pte));
__pa              516 arch/x86/mm/init_32.c 	paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT);
__pa              376 arch/x86/mm/init_64.c 			set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE |
__pa              382 arch/x86/mm/init_64.c 			set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE |
__pa              388 arch/x86/mm/init_64.c 			set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE |
__pa              689 arch/x86/mm/init_64.c 		paddr = __pa(vaddr);
__pa              692 arch/x86/mm/init_64.c 			paddr_next = __pa(vaddr_next);
__pa              704 arch/x86/mm/init_64.c 			paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
__pa              710 arch/x86/mm/init_64.c 		paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
__pa              743 arch/x86/mm/init_64.c 			paddr_last = phys_p4d_init(p4d, __pa(vaddr),
__pa              744 arch/x86/mm/init_64.c 						   __pa(vaddr_end),
__pa              751 arch/x86/mm/init_64.c 		paddr_last = phys_p4d_init(p4d, __pa(vaddr), __pa(vaddr_end),
__pa             1485 arch/x86/mm/init_64.c 				entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
__pa               30 arch/x86/mm/kasan_init_64.c 			__pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid);
__pa               34 arch/x86/mm/kasan_init_64.c 		      (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS));
__pa               51 arch/x86/mm/kasan_init_64.c 			if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL))
__pa               54 arch/x86/mm/kasan_init_64.c 				memblock_free(__pa(p), PMD_SIZE);
__pa               70 arch/x86/mm/kasan_init_64.c 		entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL);
__pa               88 arch/x86/mm/kasan_init_64.c 			if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL))
__pa               91 arch/x86/mm/kasan_init_64.c 				memblock_free(__pa(p), PUD_SIZE);
__pa              326 arch/x86/mm/kasan_init_64.c 				__pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE));
__pa              367 arch/x86/mm/kasan_init_64.c 			      early_pfn_to_nid(__pa(_stext)));
__pa              388 arch/x86/mm/kasan_init_64.c 		pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot));
__pa              183 arch/x86/mm/kaslr.c 			__p4d(_KERNPG_TABLE | __pa(pud_page_tramp)));
__pa              186 arch/x86/mm/kaslr.c 			__pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)));
__pa              189 arch/x86/mm/kaslr.c 			__pgd(_KERNPG_TABLE | __pa(pud_page_tramp)));
__pa              309 arch/x86/mm/mem_encrypt.c 		kernel_physical_mapping_change(__pa(vaddr & pmask),
__pa              310 arch/x86/mm/mem_encrypt.c 					       __pa((vaddr_end & pmask) + psize),
__pa              117 arch/x86/mm/mem_encrypt_identity.c 		set_pgd(pgd, __pgd(PGD_FLAGS | __pa(p4d)));
__pa              125 arch/x86/mm/mem_encrypt_identity.c 		set_p4d(p4d, __p4d(P4D_FLAGS | __pa(pud)));
__pa              133 arch/x86/mm/mem_encrypt_identity.c 		set_pud(pud, __pud(PUD_FLAGS | __pa(pmd)));
__pa              173 arch/x86/mm/mem_encrypt_identity.c 		set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
__pa              220 arch/x86/mm/mmap.c 	return addr + count - 1 <= __pa(high_memory - 1);
__pa              339 arch/x86/mm/numa.c 		memblock_free(__pa(numa_distance), size);
__pa              521 arch/x86/mm/numa_emulation.c 		memblock_free(__pa(phys_dist), phys_size);
__pa             1006 arch/x86/mm/pageattr.c 		unsigned long pfn = PFN_DOWN(__pa(address));
__pa             1207 arch/x86/mm/pageattr.c 	set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
__pa             1217 arch/x86/mm/pageattr.c 	set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE));
__pa             1411 arch/x86/mm/pageattr.c 		set_pgd(pgd_entry, __pgd(__pa(p4d) | _KERNPG_TABLE));
__pa             1423 arch/x86/mm/pageattr.c 		set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE));
__pa             1475 arch/x86/mm/pageattr.c 		cpa->pfn = __pa(vaddr) >> PAGE_SHIFT;
__pa             1808 arch/x86/mm/pageattr.c 	ret = reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
__pa             1820 arch/x86/mm/pageattr.c 	free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
__pa             1846 arch/x86/mm/pageattr.c 	ret = reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
__pa             1853 arch/x86/mm/pageattr.c 		free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
__pa             1880 arch/x86/mm/pageattr.c 	free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
__pa              850 arch/x86/mm/pat.c 	if (base > __pa(high_memory-1))
__pa              860 arch/x86/mm/pat.c 	id_sz = (__pa(high_memory-1) <= base + size) ?
__pa              861 arch/x86/mm/pat.c 				__pa(high_memory) - base :
__pa               57 arch/x86/mm/pgtable.c 	paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT);
__pa               72 arch/x86/mm/pgtable.c 	paravirt_release_pud(__pa(pud) >> PAGE_SHIFT);
__pa               79 arch/x86/mm/pgtable.c 	paravirt_release_p4d(__pa(p4d) >> PAGE_SHIFT);
__pa              183 arch/x86/mm/pgtable.c 	paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
__pa              187 arch/x86/mm/pgtable.c 	set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT));
__pa              187 arch/x86/mm/pti.c 		set_pgd(pgd, __pgd(_KERNPG_TABLE | __pa(new_p4d_page)));
__pa              216 arch/x86/mm/pti.c 		set_p4d(p4d, __p4d(_KERNPG_TABLE | __pa(new_pud_page)));
__pa              230 arch/x86/mm/pti.c 		set_pud(pud, __pud(_KERNPG_TABLE | __pa(new_pmd_page)));
__pa              266 arch/x86/mm/pti.c 		set_pmd(pmd, __pmd(_KERNPG_TABLE | __pa(new_pte_page)));
__pa              490 arch/x86/mm/tlb.c 	WARN_ON((cr3 & CR3_ADDR_MASK) != __pa(mm->pgd));
__pa              969 arch/x86/platform/efi/efi.c 	pa = __pa(new_memmap);
__pa              400 arch/x86/platform/efi/efi_64.c 	text = __pa(_text);
__pa               41 arch/x86/power/hibernate_32.c 	set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
__pa               66 arch/x86/power/hibernate_32.c 		set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
__pa              141 arch/x86/power/hibernate_32.c 			__pgd(__pa(empty_zero_page) | _PAGE_PRESENT));
__pa              189 arch/x86/power/hibernate_32.c 	temp_pgt = __pa(resume_pg_dir);
__pa               71 arch/x86/power/hibernate_64.c 		__pud(__pa(pmd) | pgprot_val(pgtable_prot)));
__pa               73 arch/x86/power/hibernate_64.c 		p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot));
__pa               74 arch/x86/power/hibernate_64.c 		pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot));
__pa               80 arch/x86/power/hibernate_64.c 		pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot));
__pa              123 arch/x86/power/hibernate_64.c 	temp_pgt = __pa(pgd);
__pa               65 arch/x86/realmode/init.c 	phys_base = __pa(base);
__pa              129 arch/x86/realmode/init.c 		__pa(base);
__pa              160 arch/x86/xen/efi.c 	boot_params->efi_info.efi_systab = (__u32)__pa(efi_systab_xen);
__pa              161 arch/x86/xen/efi.c 	boot_params->efi_info.efi_systab_hi = (__u32)(__pa(efi_systab_xen) >> 32);
__pa              108 arch/x86/xen/enlighten_hvm.c 		pfn = __pa(hypercall_page);
__pa             1342 arch/x86/xen/enlighten_pv.c 		initrd_start = __pa(xen_start_info->mod_start);
__pa             1349 arch/x86/xen/enlighten_pv.c 	boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line);
__pa               36 arch/x86/xen/enlighten_pvh.c 	pfn = __pa(hypercall_page);
__pa               53 arch/x86/xen/mmu_hvm.c 	a.gpa = __pa(mm->pgd);
__pa              807 arch/x86/xen/mmu_pv.c 		xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));
__pa              812 arch/x86/xen/mmu_pv.c 				   PFN_DOWN(__pa(user_pgd)));
__pa              821 arch/x86/xen/mmu_pv.c 	xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd)));
__pa              926 arch/x86/xen/mmu_pv.c 	xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa              934 arch/x86/xen/mmu_pv.c 				   PFN_DOWN(__pa(user_pgd)));
__pa             1002 arch/x86/xen/mmu_pv.c 	if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd))
__pa             1021 arch/x86/xen/mmu_pv.c 			if (per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd))
__pa             1037 arch/x86/xen/mmu_pv.c 		if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd))
__pa             1129 arch/x86/xen/mmu_pv.c 	unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK;
__pa             1250 arch/x86/xen/mmu_pv.c 		memblock_free(__pa(addr), size);
__pa             1274 arch/x86/xen/mmu_pv.c 	xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base));
__pa             1437 arch/x86/xen/mmu_pv.c 			__xen_write_cr3(false, __pa(user_pgd));
__pa             1505 arch/x86/xen/mmu_pv.c 				__pgd(__pa(level3_user_vsyscall) | _PAGE_TABLE);
__pa             1764 arch/x86/xen/mmu_pv.c 	unsigned long pfn = __pa(addr) >> PAGE_SHIFT;
__pa             1800 arch/x86/xen/mmu_pv.c 			pmd[pmdidx] = __pmd(__pa(pte_page) | _PAGE_TABLE);
__pa             1854 arch/x86/xen/mmu_pv.c 	if (*pt_base == PFN_DOWN(__pa(addr))) {
__pa             1859 arch/x86/xen/mmu_pv.c 	if (*pt_end == PFN_DOWN(__pa(addr))) {
__pa             1890 arch/x86/xen/mmu_pv.c 		max_pfn_mapped = PFN_DOWN(__pa(xen_start_info->mfn_list));
__pa             1892 arch/x86/xen/mmu_pv.c 	pt_base = PFN_DOWN(__pa(xen_start_info->pt_base));
__pa             1965 arch/x86/xen/mmu_pv.c 	pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa             1972 arch/x86/xen/mmu_pv.c 	__xen_write_cr3(true, __pa(init_top_pgt));
__pa             1990 arch/x86/xen/mmu_pv.c 	xen_start_info = (struct start_info *)__va(__pa(xen_start_info));
__pa             2179 arch/x86/xen/mmu_pv.c 	unsigned long pfn = PFN_DOWN(__pa(swapper_pg_dir));
__pa             2181 arch/x86/xen/mmu_pv.c 	BUG_ON(read_cr3_pa() != __pa(initial_page_table));
__pa             2182 arch/x86/xen/mmu_pv.c 	BUG_ON(cr3 != __pa(swapper_pg_dir));
__pa             2198 arch/x86/xen/mmu_pv.c 		__pgd(__pa(swapper_kernel_pmd) | _PAGE_PRESENT);
__pa             2206 arch/x86/xen/mmu_pv.c 			  PFN_DOWN(__pa(initial_page_table)));
__pa             2223 arch/x86/xen/mmu_pv.c 	pt_base = min(__pa(xen_start_info->pt_base), __pa(pmd));
__pa             2254 arch/x86/xen/mmu_pv.c 		__pgd(__pa(initial_kernel_pmd) | _PAGE_PRESENT);
__pa             2260 arch/x86/xen/mmu_pv.c 	pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa             2263 arch/x86/xen/mmu_pv.c 			  PFN_DOWN(__pa(initial_page_table)));
__pa             2264 arch/x86/xen/mmu_pv.c 	xen_write_cr3(__pa(initial_page_table));
__pa             2274 arch/x86/xen/mmu_pv.c 	memblock_reserve(__pa(xen_start_info), PAGE_SIZE);
__pa             2317 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
__pa             2327 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
__pa             2800 arch/x86/xen/mmu_pv.c 		return __pa(vmcoreinfo_note);
__pa              346 arch/x86/xen/p2m.c 	paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT);
__pa              348 arch/x86/xen/p2m.c 	paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT);
__pa              351 arch/x86/xen/p2m.c 			pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO));
__pa              353 arch/x86/xen/p2m.c 			pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO));
__pa              390 arch/x86/xen/p2m.c 				pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL));
__pa              400 arch/x86/xen/p2m.c 				pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO));
__pa              410 arch/x86/xen/p2m.c 			set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE));
__pa              459 arch/x86/xen/p2m.c 	if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
__pa              497 arch/x86/xen/p2m.c 		paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT);
__pa              509 arch/x86/xen/p2m.c 				__pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE));
__pa              518 arch/x86/xen/p2m.c 			paravirt_release_pte(__pa(pte_newpg[i]) >> PAGE_SHIFT);
__pa              590 arch/x86/xen/p2m.c 	if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) ||
__pa              591 arch/x86/xen/p2m.c 	    p2m_pfn == PFN_DOWN(__pa(p2m_missing))) {
__pa              599 arch/x86/xen/p2m.c 		if (p2m_pfn == PFN_DOWN(__pa(p2m_missing)))
__pa              610 arch/x86/xen/p2m.c 				pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL));
__pa              674 arch/x86/xen/p2m.c 	if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing)))
__pa              677 arch/x86/xen/p2m.c 	if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity)))
__pa              310 arch/x86/xen/setup.c 	if (pfn >= PFN_UP(__pa(high_memory - 1)))
__pa              712 arch/x86/xen/setup.c 		start = __pa(xen_start_info->mfn_list);
__pa              353 arch/x86/xen/smp_pv.c 	per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir);
__pa              188 arch/xtensa/include/asm/page.h #define virt_to_page(kaddr)	pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa              190 arch/xtensa/include/asm/page.h #define virt_addr_valid(kaddr)	pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa              335 arch/xtensa/kernel/setup.c 	    !mem_reserve(__pa(initrd_start), __pa(initrd_end)))
__pa              341 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(_stext), __pa(_end));
__pa              344 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_WindowVectors_text_start),
__pa              345 arch/xtensa/kernel/setup.c 		    __pa(&_WindowVectors_text_end));
__pa              347 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_DebugInterruptVector_text_start),
__pa              348 arch/xtensa/kernel/setup.c 		    __pa(&_DebugInterruptVector_text_end));
__pa              350 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_KernelExceptionVector_text_start),
__pa              351 arch/xtensa/kernel/setup.c 		    __pa(&_KernelExceptionVector_text_end));
__pa              353 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_UserExceptionVector_text_start),
__pa              354 arch/xtensa/kernel/setup.c 		    __pa(&_UserExceptionVector_text_end));
__pa              356 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_DoubleExceptionVector_text_start),
__pa              357 arch/xtensa/kernel/setup.c 		    __pa(&_DoubleExceptionVector_text_end));
__pa              360 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_Level2InterruptVector_text_start),
__pa              361 arch/xtensa/kernel/setup.c 		    __pa(&_Level2InterruptVector_text_end));
__pa              364 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_Level3InterruptVector_text_start),
__pa              365 arch/xtensa/kernel/setup.c 		    __pa(&_Level3InterruptVector_text_end));
__pa              368 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_Level4InterruptVector_text_start),
__pa              369 arch/xtensa/kernel/setup.c 		    __pa(&_Level4InterruptVector_text_end));
__pa              372 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_Level5InterruptVector_text_start),
__pa              373 arch/xtensa/kernel/setup.c 		    __pa(&_Level5InterruptVector_text_end));
__pa              376 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_Level6InterruptVector_text_start),
__pa              377 arch/xtensa/kernel/setup.c 		    __pa(&_Level6InterruptVector_text_end));
__pa              383 arch/xtensa/kernel/setup.c 	mem_reserve(__pa(&_SecondaryResetVector_text_start),
__pa              384 arch/xtensa/kernel/setup.c 		    __pa(&_SecondaryResetVector_text_end));
__pa              441 arch/xtensa/kernel/setup.c 		unsigned long paddr = __pa(vaddr);
__pa              369 drivers/block/ps3disk.c 				   ps3_mm_phys_to_lpar(__pa(&ata_cmnd)),
__pa              690 drivers/block/ps3vram.c 	xdr_lpar = ps3_mm_phys_to_lpar(__pa(priv->xdr_buf));
__pa              815 drivers/block/ps3vram.c 			      ps3_mm_phys_to_lpar(__pa(priv->xdr_buf)),
__pa              149 drivers/char/hw_random/n2-drv.c 	unsigned long ra = __pa(&np->scratch_control[0]);
__pa              422 drivers/char/hw_random/n2-drv.c 	unsigned long ra = __pa(&np->test_data);
__pa              457 drivers/char/hw_random/n2-drv.c 	unsigned long ra = __pa(&np->test_data);
__pa              467 drivers/char/hw_random/n2-drv.c 	unsigned long post_ctl_ra = __pa(post_control);
__pa              468 drivers/char/hw_random/n2-drv.c 	unsigned long pre_ctl_ra = __pa(pre_control);
__pa              469 drivers/char/hw_random/n2-drv.c 	unsigned long buffer_ra = __pa(buffer);
__pa              621 drivers/char/hw_random/n2-drv.c 		unsigned long ctl_ra = __pa(&up->control[0]);
__pa               54 drivers/char/mem.c 	return addr + count <= __pa(high_memory);
__pa              308 drivers/char/mem.c 	return addr >= __pa(high_memory);
__pa              417 drivers/char/mem.c 	pfn = __pa((u64)vma->vm_pgoff << PAGE_SHIFT) >> PAGE_SHIFT;
__pa              178 drivers/char/powernv-op-panel.c 		oppanel_lines[i].line = cpu_to_be64(__pa(&oppanel_data[i *
__pa               31 drivers/cpuidle/cpuidle-calxeda.c 	return psci_ops.cpu_suspend(CALXEDA_IDLE_PARAM, __pa(cpu_resume));
__pa               85 drivers/crypto/cavium/zip/zip_deflate.c 	zip_cmd->s.inp_ptr_addr.s.addr  = __pa(zip_ops->input);
__pa               91 drivers/crypto/cavium/zip/zip_deflate.c 	zip_cmd->s.out_ptr_addr.s.addr  = __pa(zip_ops->output);
__pa               97 drivers/crypto/cavium/zip/zip_deflate.c 	zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr);
__pa              134 drivers/crypto/cavium/zip/zip_device.c 		ncp.s.addr = __pa(zip_dev->iq[queue].sw_head);
__pa              137 drivers/crypto/cavium/zip/zip_device.c 			*ncb_ptr, __pa(zip_dev->iq[queue].sw_head));
__pa               94 drivers/crypto/cavium/zip/zip_inflate.c 	zip_cmd->s.out_ptr_addr.s.addr  = __pa(zip_ops->output);
__pa              104 drivers/crypto/cavium/zip/zip_inflate.c 	zip_cmd->s.inp_ptr_addr.s.addr  = __pa((u8 *)zip_ops->input);
__pa              108 drivers/crypto/cavium/zip/zip_inflate.c 	zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr);
__pa              185 drivers/crypto/cavium/zip/zip_main.c 		que_sbuf_addr.s.ptr = (__pa(zip->iq[q].sw_head) >>
__pa              569 drivers/crypto/n2_core.c 	ent->src_addr = __pa(walk.data);
__pa              571 drivers/crypto/n2_core.c 	ent->auth_iv_addr = __pa(hash_loc);
__pa              575 drivers/crypto/n2_core.c 	ent->dest_addr = __pa(hash_loc);
__pa              582 drivers/crypto/n2_core.c 		ent->src_addr = __pa(walk.data);
__pa              656 drivers/crypto/n2_core.c 				  __pa(&ctx->hash_key),
__pa              859 drivers/crypto/n2_core.c 	ent->enc_key_addr = __pa(&ctx->key);
__pa             1053 drivers/crypto/n2_core.c 		iv_paddr = __pa(rctx->walk.iv);
__pa             1070 drivers/crypto/n2_core.c 				iv_paddr = __pa(rctx->walk.iv);
__pa             1652 drivers/crypto/n2_core.c 	hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q),
__pa              107 drivers/crypto/nx/nx-842.h 		return __pa(addr);
__pa               94 drivers/crypto/nx/nx.c 		sg_addr = __pa(sg_addr);
__pa              312 drivers/crypto/nx/nx.c 	nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb);
__pa              313 drivers/crypto/nx/nx.c 	nx_ctx->op.in = __pa(nx_ctx->in_sg);
__pa              314 drivers/crypto/nx/nx.c 	nx_ctx->op.out = __pa(nx_ctx->out_sg);
__pa              320 drivers/crypto/nx/nx.c 		nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead);
__pa              321 drivers/crypto/nx/nx.c 		nx_ctx->op_aead.in = __pa(nx_ctx->in_sg);
__pa              322 drivers/crypto/nx/nx.c 		nx_ctx->op_aead.out = __pa(nx_ctx->out_sg);
__pa              133 drivers/dax/super.c 				&& pfn_t_to_pfn(pfn) == PHYS_PFN(__pa(kaddr))
__pa              134 drivers/dax/super.c 				&& pfn_t_to_pfn(end_pfn) == PHYS_PFN(__pa(end_kaddr)))
__pa             4480 drivers/dma/ppc4xx/adma.c 	iowrite32(((u32)__pa(ppc440spe_dma_fifo_buf)), &i2o_reg->ifbal);
__pa             1025 drivers/firmware/efi/efi.c 	rc = efi_mem_reserve_iomem(__pa(rsv), SZ_4K);
__pa             1044 drivers/firmware/efi/efi.c 	efi_memreserve_root->next = __pa(rsv);
__pa              802 drivers/gpu/drm/amd/amdkfd/kfd_events.c 	pfn = __pa(page->kernel_address);
__pa             1143 drivers/gpu/drm/amd/amdkfd/kfd_process.c 			       PFN_DOWN(__pa(qpd->cwsr_kaddr)),
__pa             2009 drivers/gpu/drm/i915/gvt/kvmgt.c 	return PFN_DOWN(__pa(addr));
__pa               35 drivers/hv/channel.c 		paddr = __pa(addr);
__pa               90 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf));
__pa              101 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf));
__pa              126 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(&data->byte));
__pa              134 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(&data->byte));
__pa              146 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(local));
__pa              154 drivers/i2c/busses/i2c-opal.c 		req.buffer_ra = cpu_to_be64(__pa(&data->block[1]));
__pa              459 drivers/infiniband/hw/hfi1/file_ops.c 				PFN_DOWN(__pa(memvirt)),
__pa              588 drivers/infiniband/hw/hfi1/file_ops.c 				      PFN_DOWN(__pa(memvirt)),
__pa              653 drivers/iommu/amd_iommu_v2.c 					__pa(pasid_state->mm->pgd));
__pa              491 drivers/iommu/intel-pasid.c 	pasid_set_flptr(pte, (u64)__pa(pgd));
__pa              358 drivers/iommu/io-pgtable-arm.c 	new = __pa(table) | ARM_LPAE_PTE_TYPE_TABLE;
__pa              573 drivers/macintosh/smu.c 	memblock_free(__pa(smu), sizeof(struct smu_device));
__pa               76 drivers/media/usb/cpia2/cpia2_core.c 	ret = __pa(kva);
__pa              777 drivers/misc/ocxl/config.c 	rc = pnv_ocxl_set_tl_conf(dev, recv_cap, __pa(recv_rate),
__pa               69 drivers/mtd/devices/powernv_flash.c 		rc = opal_flash_read(info->id, offset, __pa(buf), len, token);
__pa               72 drivers/mtd/devices/powernv_flash.c 		rc = opal_flash_write(info->id, offset, __pa(buf), len, token);
__pa             1144 drivers/mtd/mtdchar.c 		if (file->f_flags & O_DSYNC || map->phys >= __pa(high_memory))
__pa               59 drivers/net/ethernet/brocade/bna/bfa_ioc.h #define bfa_alen_set(__alen, __len, __pa)	\
__pa               60 drivers/net/ethernet/brocade/bna/bfa_ioc.h 	__bfa_alen_set(__alen, __len, (u64)__pa)
__pa              130 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 				       __pa(cb_addr),		/* R8 */
__pa              404 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 				 __pa(cb_addr),			/* R8 */
__pa              517 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	cb_logaddr = __pa(cb_addr);
__pa              534 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	u64 cb_logaddr = __pa(cb_addr);
__pa              556 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	u64 cb_logaddr = __pa(cb_addr);
__pa              610 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 				       __pa(rblock),		/* R6 */
__pa              146 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		rpage = __pa(vpage);
__pa              270 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		rpage = __pa(vpage);
__pa              375 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		rpage = __pa(vpage);
__pa              768 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	index = __pa(caddr) >> SECTION_SIZE_BITS;
__pa              800 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	u64 pt_abs = __pa(pt);
__pa              808 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			pt[m] = __pa(pg);
__pa             9952 drivers/net/ethernet/sun/niu.c 	*dma_addr = __pa(page);
__pa             9982 drivers/net/ethernet/sun/niu.c 	return __pa(cpu_addr);
__pa             1686 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		ps3_mm_phys_to_lpar(__pa(&card->irq_status)),
__pa              159 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c 			ps3_mm_phys_to_lpar(__pa(cmd->buffer)) :
__pa              181 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c 		arg1 = ps3_mm_phys_to_lpar(__pa(cmd->buffer));
__pa              266 drivers/net/fjes/fjes_hw.c 	param.req_start = __pa(hw->hw_info.req_buf);
__pa              268 drivers/net/fjes/fjes_hw.c 	param.res_start = __pa(hw->hw_info.res_buf);
__pa              270 drivers/net/fjes/fjes_hw.c 	param.share_start = __pa(hw->hw_info.share->ep_status);
__pa              479 drivers/net/fjes/fjes_main.c 		param.req_start = __pa(hw->hw_info.req_buf);
__pa              481 drivers/net/fjes/fjes_main.c 		param.res_start = __pa(hw->hw_info.res_buf);
__pa              482 drivers/net/fjes/fjes_main.c 		param.share_start = __pa(hw->hw_info.share->ep_status);
__pa              615 drivers/of/fdt.c 	early_init_dt_reserve_memory_arch(__pa(initial_boot_params),
__pa             1100 drivers/of/fdt.c #define MIN_MEMBLOCK_ADDR	__pa(PAGE_OFFSET)
__pa              754 drivers/platform/goldfish/goldfish_pipe.c 			(u64)(unsigned long)__pa(pipe->command_buffer);
__pa              814 drivers/platform/goldfish/goldfish_pipe.c 	const unsigned long paddr = __pa(addr);
__pa             1124 drivers/ps3/ps3-lpm.c 				ps3_mm_phys_to_lpar(__pa(lpm_priv->tb_cache)),
__pa              358 drivers/ps3/ps3-vuart.c 		ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_written);
__pa              390 drivers/ps3/ps3-vuart.c 		ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_read);
__pa              189 drivers/ps3/ps3stor_lib.c 	dev->bounce_lpar = ps3_mm_phys_to_lpar(__pa(dev->bounce_buf));
__pa              290 drivers/s390/block/dasd_eckd.c 		ccw->cda = (__u32)__pa(data);
__pa              400 drivers/s390/block/dasd_eckd.c 		ccw->cda = (__u32)__pa(data);
__pa              546 drivers/s390/block/dasd_eckd.c 		ccw->cda = (__u32) __pa(pfxdata);
__pa              550 drivers/s390/block/dasd_eckd.c 		ccw->cda = (__u32) __pa(pfxdata);
__pa              617 drivers/s390/block/dasd_eckd.c 	ccw->cda = (__u32) __pa(data);
__pa             4138 drivers/s390/block/dasd_eckd.c 				if (__pa(dst) & (IDA_BLOCK_SIZE-1)) {
__pa             4158 drivers/s390/block/dasd_eckd.c 			if (!(__pa(idaw_dst + idaw_len) & (IDA_BLOCK_SIZE-1)))
__pa               94 drivers/s390/block/dasd_fba.c 	ccw->cda = (__u32) __pa(data);
__pa              114 drivers/s390/block/dasd_fba.c 	ccw->cda = (__u32) __pa(data);
__pa              101 drivers/s390/block/xpram.c 		: "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc");
__pa              131 drivers/s390/block/xpram.c 		: "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc");
__pa              162 drivers/s390/char/con3215.c 	ccw->cda = (__u32) __pa(raw->inbuf);
__pa              222 drivers/s390/char/con3215.c 			(__u32) __pa(raw->buffer + ix);
__pa              160 drivers/s390/char/raw3270.c 	rq->ccw.cda = __pa(rq->buffer);
__pa              185 drivers/s390/char/raw3270.c 	rq->ccw.cda = __pa(rq->buffer);
__pa              219 drivers/s390/char/raw3270.c 	rq->ccw.cda = __pa(data);
__pa              229 drivers/s390/char/raw3270.c 	rq->ccw.cda = __pa(ib->data);
__pa              547 drivers/s390/char/raw3270.c 	rp->init_readmod.ccw.cda = (__u32) __pa(rp->init_data);
__pa              566 drivers/s390/char/raw3270.c 	rp->init_readpart.ccw.cda = (__u32) __pa(&rp->init_data);
__pa              605 drivers/s390/char/raw3270.c 	rp->init_reset.ccw.cda = (__u32) __pa(rp->init_data);
__pa              151 drivers/s390/cio/cio.c 	orb->cmd.cpa = (__u32) __pa(cpa);
__pa              334 drivers/s390/cio/device_status.c 	sense_ccw->cda = (__u32) __pa(cdev->private->dma_area->irb.ecw);
__pa               65 drivers/s390/cio/eadm_sch.c 	orb->eadm.aob = (u32)__pa(aob);
__pa              788 drivers/s390/cio/vfio_ccw_cp.c 	orb->cmd.cpa = (__u32) __pa(cpa);
__pa              226 drivers/s390/net/lcs.c 			(__u32) __pa(card->read.iob[cnt].data);
__pa              239 drivers/s390/net/lcs.c 		(__u32) __pa(card->read.ccws);
__pa              281 drivers/s390/net/lcs.c 			(__u32) __pa(card->write.iob[cnt].data);
__pa              286 drivers/s390/net/lcs.c 		(__u32) __pa(card->write.ccws);
__pa              496 drivers/s390/net/qeth_core_main.c 	ccw->cda = (__u32) __pa(data);
__pa              184 drivers/scsi/bfa/bfa_ioc.h #define bfa_alen_set(__alen, __len, __pa)	\
__pa              185 drivers/scsi/bfa/bfa_ioc.h 	__bfa_alen_set(__alen, __len, (u64)__pa)
__pa              125 drivers/scsi/ps3rom.c 	lpar = ps3_mm_phys_to_lpar(__pa(&atapi_cmnd));
__pa               62 drivers/tty/serial/sunhv.c 		unsigned long ra = __pa(xmit->buf + xmit->tail);
__pa              124 drivers/tty/serial/sunhv.c 		unsigned long ra = __pa(con_read_page);
__pa              451 drivers/tty/serial/sunhv.c 		unsigned long ra = __pa(con_write_page);
__pa              560 drivers/tty/serial/sunhv.c 	port->membase = (unsigned char __iomem *) __pa(port);
__pa              102 drivers/usb/early/xhci-dbc.c 		*dma_addr = (dma_addr_t)__pa(virt);
__pa              832 drivers/usb/early/xhci-dbc.c 		xdbc_write64(__pa(xdbc.evt_ring.dequeue), &xdbc.xdbc_reg->erdp);
__pa              494 drivers/vfio/vfio_iommu_spapr_tce.c 	*hpa = __pa((unsigned long) page_address(page));
__pa             1094 drivers/video/fbdev/ps3fb.c 	xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address));
__pa             1142 drivers/video/fbdev/ps3fb.c 	info->fix.smem_start = __pa(fb_start);
__pa             1213 drivers/video/fbdev/ps3fb.c 	u64 xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address));
__pa              711 drivers/video/fbdev/ssd1307fb.c 	info->fix.smem_start = __pa(vmem);
__pa              111 drivers/watchdog/wdrtas.c 			   WDRTAS_SP_SPI, __pa(rtas_data_buf),
__pa              163 drivers/watchdog/wdrtas.c 				   (void *)__pa(wdrtas_logbuffer),
__pa              235 drivers/xen/swiotlb-xen.c 			memblock_free(__pa(xen_io_tlb_start),
__pa              147 fs/proc/kcore.c 	unsigned long pfn = __pa(ent->addr) >> PAGE_SHIFT;
__pa              390 fs/proc/kcore.c 				phdr->p_paddr = __pa(m->addr);
__pa              491 fs/proc/kcore.c 		} else if (!pfn_is_ram(__pa(start) >> PAGE_SHIFT)) {
__pa              576 fs/proc/vmcore.c 		pfn = __pa(elfcorebuf + start) >> PAGE_SHIFT;
__pa              910 include/asm-generic/io.h 	return __pa((unsigned long)address);
__pa               81 include/asm-generic/page.h #define virt_to_pfn(kaddr)	(__pa(kaddr) >> PAGE_SHIFT)
__pa             1179 include/linux/efi.h 		unsigned long paddr = __pa(start + i);
__pa              377 include/linux/kexec.h 	return phys_to_boot_phys(__pa((unsigned long)addr));
__pa              114 include/linux/mm.h #define __pa_symbol(x)  __pa(RELOC_HIDE((unsigned long)(x), 0))
__pa               22 include/linux/psp-sev.h #define __psp_pa(x)	__pa(x)
__pa              153 include/xen/grant_table.h 		map->host_addr = __pa(addr);
__pa              169 include/xen/grant_table.h 		unmap->host_addr = __pa(addr);
__pa              375 kernel/crash_core.c 	return __pa(vmcoreinfo_note);
__pa              208 kernel/dma/swiotlb.c 	io_tlb_start = __pa(tlb);
__pa              395 kernel/dma/swiotlb.c 		memblock_free_late(__pa(io_tlb_orig_addr),
__pa              397 kernel/dma/swiotlb.c 		memblock_free_late(__pa(io_tlb_list),
__pa              191 lib/cpumask.c  	memblock_free_early(__pa(mask), cpumask_size());
__pa              258 mm/cma.c       	highmem_start = __pa(high_memory - 1) + 1;
__pa               86 mm/kasan/init.c 	void *ptr = memblock_alloc_try_nid(size, size, __pa(MAX_DMA_ADDRESS),
__pa               91 mm/kasan/init.c 		      __func__, size, size, node, (u64)__pa(MAX_DMA_ADDRESS));
__pa              383 mm/memblock.c  		addr = __pa(memblock.reserved.regions);
__pa              390 mm/memblock.c  		addr = __pa(memblock.memory.regions);
__pa              448 mm/memblock.c  		addr = new_array ? __pa(new_array) : 0;
__pa              490 mm/memblock.c  		memblock_free(__pa(old_array), old_alloc_size);
__pa             7608 mm/page_alloc.c 			__pa(PAGE_OFFSET) >> PAGE_SHIFT, NULL);
__pa              157 mm/page_ext.c  			table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
__pa             2096 mm/percpu.c    			return __pa(addr);
__pa             2157 mm/percpu.c    	memblock_free_early(__pa(ai), ai->__ai_size);
__pa             2805 mm/percpu.c    		memblock_free_early(__pa(areas), areas_size);
__pa             2927 mm/percpu.c    	memblock_free_early(__pa(pages), pages_size);
__pa             2952 mm/percpu.c    	return  memblock_alloc_from(size, align, __pa(MAX_DMA_ADDRESS));
__pa             2957 mm/percpu.c    	memblock_free_early(__pa(ptr), size);
__pa             3000 mm/percpu.c    	fc = memblock_alloc_from(unit_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
__pa               70 mm/sparse-vmemmap.c 				__pa(MAX_DMA_ADDRESS));
__pa              151 mm/sparse-vmemmap.c 		entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
__pa              371 mm/sparse.c    	goal = __pa(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT);
__pa              398 mm/sparse.c    	usemap_snr = pfn_to_section_nr(__pa(usage) >> PAGE_SHIFT);
__pa              399 mm/sparse.c    	pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT);
__pa              456 mm/sparse.c    	phys_addr_t addr = __pa(MAX_DMA_ADDRESS);
__pa              478 mm/sparse.c    	memblock_free_early(__pa(sparsemap_buf), size);
__pa              483 mm/sparse.c    	phys_addr_t addr = __pa(MAX_DMA_ADDRESS);
__pa              575 sound/parisc/harmony.c 		ss->runtime->dma_addr = __pa(ss->runtime->dma_area);
__pa              733 virt/kvm/arm/mmu.c 		return __pa(kaddr);