__pa 59 arch/alpha/include/asm/mmzone.h #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr)) __pa 73 arch/alpha/include/asm/mmzone.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 109 arch/alpha/include/asm/mmzone.h #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT)) __pa 87 arch/alpha/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 90 arch/alpha/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 232 arch/alpha/kernel/core_irongate.c if (initrd_end && __pa(initrd_end) > pci_mem) { __pa 236 arch/alpha/kernel/core_irongate.c memblock_free(__pa(initrd_start), PAGE_ALIGN(size)); __pa 276 arch/alpha/kernel/pci_iommu.c paddr = __pa(cpu_addr); __pa 514 arch/alpha/kernel/pci_iommu.c #define SG_ENT_PHYS_ADDRESS(SG) __pa(SG_ENT_VIRT_ADDRESS(SG)) __pa 297 arch/alpha/kernel/setup.c if (!start || __pa(start) + size > mem_limit) { __pa 83 arch/arc/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 120 arch/arc/mm/init.c __pa(_end) - CONFIG_LINUX_LINK_BASE); __pa 397 arch/arm/include/asm/cacheflush.h outer_clean_range(__pa(_p), __pa(_p + size)); __pa 419 arch/arm/include/asm/cacheflush.h outer_flush_range(__pa(_p), __pa(_p + size)); __pa 69 arch/arm/include/asm/kvm_mmu.h #define kvm_mk_pmd(ptep) __pmd(__pa(ptep) | PMD_TYPE_TABLE) __pa 70 arch/arm/include/asm/kvm_mmu.h #define kvm_mk_pud(pmdp) __pud(__pa(pmdp) | PMD_TYPE_TABLE) __pa 38 arch/arm/include/asm/pgalloc.h set_pud(pud, __pud(__pa(pmd) | PMD_TYPE_TABLE)); __pa 134 arch/arm/include/asm/pgalloc.h __pmd_populate(pmdp, __pa(ptep), _PAGE_KERNEL_TABLE); __pa 1002 arch/arm/kernel/setup.c unsigned long long lowmem_max = __pa(high_memory - 1) + 1; __pa 22 arch/arm/mach-highbank/pm.c return psci_ops.cpu_suspend(HIGHBANK_SUSPEND_PARAM, __pa(cpu_resume)); __pa 95 arch/arm/mach-integrator/core.c memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET); __pa 330 arch/arm/mach-omap2/control.c (u32) __pa(omap3_secure_ram_storage); __pa 50 arch/arm/mach-omap2/omap-secure.c outer_clean_range(__pa(param), __pa(param + 5)); __pa 51 arch/arm/mach-omap2/omap-secure.c ret = omap_smc2(idx, flag, __pa(param)); __pa 82 arch/arm/mach-omap2/omap-secure.c param[1] = __pa(addr); /* Physical address for saving */ __pa 87 arch/arm/mach-omap2/omap-secure.c ret = save_secure_ram_context(__pa(param)); __pa 125 arch/arm/mach-omap2/omap-secure.c outer_clean_range(__pa(param), __pa(param + 5)); __pa 126 arch/arm/mach-omap2/omap-secure.c ret = omap_smc3(idx, process, flag, __pa(param)); __pa 295 arch/arm/mach-omap2/omap-smp.c if ((addr >= __pa(PAGE_OFFSET)) && (addr <= __pa(__bss_start))) __pa 37 arch/arm/mach-shmobile/smp-emev2.c iowrite32(__pa(shmobile_boot_vector), smu + SMU_GENERAL_REG0); __pa 40 arch/arm/mach-shmobile/smp-r8a7779.c __raw_writel(__pa(shmobile_boot_vector), AVECR); __pa 43 arch/arm/mach-shmobile/smp-sh73a0.c __raw_writel(__pa(shmobile_boot_vector), SBAR); __pa 54 arch/arm/mach-zynq/common.c if (!__pa(PAGE_OFFSET)) __pa 55 arch/arm/mach-zynq/common.c memblock_reserve(__pa(PAGE_OFFSET), 0x80000); __pa 44 arch/arm/mach-zynq/platsmp.c if (__pa(PAGE_OFFSET)) { __pa 68 arch/arm/mach-zynq/platsmp.c if (__pa(PAGE_OFFSET)) __pa 290 arch/arm/mm/dma-mapping.c outer_flush_range(__pa(ptr), __pa(ptr) + size); __pa 267 arch/arm/mm/init.c memblock_reserve(__pa(KERNEL_START), KERNEL_END - KERNEL_START); __pa 342 arch/arm/mm/init.c pg = PAGE_ALIGN(__pa(start_pg)); __pa 343 arch/arm/mm/init.c pgend = __pa(end_pg) & PAGE_MASK; __pa 155 arch/arm/mm/mmap.c if (addr + size > __pa(high_memory - 1) + 1) __pa 745 arch/arm/mm/mmu.c __pmd_populate(pmd, __pa(pte), prot); __pa 1318 arch/arm/mm/mmu.c memblock_reserve(__pa(swapper_pg_dir), SWAPPER_PG_DIR_SIZE); __pa 1325 arch/arm/mm/mmu.c memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET); __pa 1453 arch/arm/mm/mmu.c phys_addr_t kernel_x_start = round_down(__pa(KERNEL_START), SECTION_SIZE); __pa 1454 arch/arm/mm/mmu.c phys_addr_t kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE); __pa 1544 arch/arm/mm/mmu.c lpae_pgtables_remap = (pgtables_remap *)(unsigned long)__pa(lpae_pgtables_remap_asm); __pa 1545 arch/arm/mm/mmu.c pa_pgd = __pa(swapper_pg_dir); __pa 259 arch/arm/mm/pmsa-v7.c num = allocate_region(CONFIG_XIP_PHYS_ADDR, __pa(_exiprom) - CONFIG_XIP_PHYS_ADDR, __pa 254 arch/arm/mm/pmsa-v8.c subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); __pa 255 arch/arm/mm/pmsa-v8.c subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); __pa 259 arch/arm/mm/pmsa-v8.c subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); __pa 260 arch/arm/mm/pmsa-v8.c subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); __pa 276 arch/arm/mm/pmsa-v8.c err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); __pa 279 arch/arm/mm/pmsa-v8.c err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); __pa 171 arch/arm64/include/asm/kvm_mmu.h __pmd(__phys_to_pmd_val(__pa(ptep)) | PMD_TYPE_TABLE) __pa 173 arch/arm64/include/asm/kvm_mmu.h __pud(__phys_to_pud_val(__pa(pmdp)) | PMD_TYPE_TABLE) __pa 175 arch/arm64/include/asm/kvm_mmu.h __pgd(__phys_to_pgd_val(__pa(pudp)) | PUD_TYPE_TABLE) __pa 399 arch/arm64/include/asm/kvm_mmu.h pgd_addr = __phys_to_pgd_val(__pa(hyp_pgd)); __pa 410 arch/arm64/include/asm/kvm_mmu.h pgd_addr = __phys_to_pgd_val(__pa(boot_hyp_pgd)); __pa 54 arch/arm64/include/asm/pgalloc.h __pud_populate(pudp, __pa(pmdp), PMD_TYPE_TABLE); __pa 83 arch/arm64/include/asm/pgalloc.h __pgd_populate(pgdp, __pa(pudp), PUD_TYPE_TABLE); __pa 111 arch/arm64/include/asm/pgalloc.h __pmd_populate(pmdp, __pa(ptep), PMD_TYPE_TABLE); __pa 477 arch/arm64/mm/init.c pg = (unsigned long)PAGE_ALIGN(__pa(start_pg)); __pa 478 arch/arm64/mm/init.c pgend = (unsigned long)__pa(end_pg) & PAGE_MASK; __pa 37 arch/arm64/mm/kasan_init.c __pa(MAX_DMA_ADDRESS), __pa 42 arch/arm64/mm/kasan_init.c __pa(MAX_DMA_ADDRESS)); __pa 44 arch/arm64/mm/kasan_init.c return __pa(p); __pa 50 arch/arm64/mm/kasan_init.c __pa(MAX_DMA_ADDRESS), __pa 55 arch/arm64/mm/kasan_init.c __pa(MAX_DMA_ADDRESS)); __pa 57 arch/arm64/mm/kasan_init.c return __pa(p); __pa 371 arch/arm64/mm/mmu.c return __pa(ptr); __pa 764 arch/arm64/mm/mmu.c pmd_set_huge(pmdp, __pa(p), __pgprot(PROT_SECT_NORMAL)); __pa 160 arch/arm64/mm/numa.c __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); __pa 165 arch/arm64/mm/numa.c memblock_free_early(__pa(ptr), size); __pa 264 arch/arm64/mm/numa.c memblock_free(__pa(numa_distance), size); __pa 54 arch/c6x/mm/init.c __pa(PAGE_OFFSET) >> PAGE_SHIFT; __pa 18 arch/csky/include/asm/mmu_context.h setup_pgd(__pa(pgd), false) __pa 21 arch/csky/include/asm/mmu_context.h setup_pgd(__pa(pgd), true) __pa 37 arch/csky/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 82 arch/csky/include/asm/page.h #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0)) __pa 17 arch/csky/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(pte))); __pa 23 arch/csky/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(page_address(pte)))); __pa 151 arch/csky/include/asm/pgtable.h return pmd_val(pmd) == __pa(invalid_pte_table); __pa 158 arch/csky/include/asm/pgtable.h return (pmd_val(pmd) != __pa(invalid_pte_table)); __pa 163 arch/csky/include/asm/pgtable.h pmd_val(*p) = (__pa(invalid_pte_table)); __pa 32 arch/csky/kernel/setup.c memblock_reserve(__pa(_stext), _end - _stext); __pa 34 arch/csky/kernel/setup.c memblock_reserve(__pa(initrd_start), initrd_end - initrd_start); __pa 149 arch/csky/mm/highmem.c set_pmd(pmd, __pmd(__pa(pte))); __pa 89 arch/csky/mm/init.c p[i] = __pa(invalid_pte_table); __pa 91 arch/h8300/kernel/setup.c memblock_reserve(__pa(_stext), _end - _stext); __pa 51 arch/hexagon/include/asm/io.h return __pa(address); __pa 93 arch/hexagon/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) __pa 100 arch/hexagon/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 130 arch/hexagon/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 39 arch/hexagon/include/asm/pgalloc.h mm->context.ptbase = __pa(pgd); __pa 82 arch/hexagon/include/asm/pgalloc.h set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); __pa 90 arch/hexagon/include/asm/pgalloc.h set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); __pa 71 arch/hexagon/mm/init.c init_mm.context.ptbase = __pa(init_mm.pgd); __pa 193 arch/ia64/include/asm/mmu_context.h ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); __pa 96 arch/ia64/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 126 arch/ia64/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 41 arch/ia64/include/asm/pgalloc.h pgd_val(*pgd_entry) = __pa(pud); __pa 59 arch/ia64/include/asm/pgalloc.h pud_val(*pud_entry) = __pa(pmd); __pa 84 arch/ia64/include/asm/pgalloc.h pmd_val(*pmd_entry) = __pa(pte); __pa 287 arch/ia64/include/asm/uaccess.h ptr = (void *)__pa(p) + __IA64_UNCACHED_OFFSET; __pa 156 arch/ia64/include/asm/uv/uv_hub.h return __pa(v) | uv_hub_info->gnode_upper; __pa 469 arch/ia64/kernel/efi.c pte_val(pfn_pte(__pa(pal_vaddr) >> PAGE_SHIFT, PAGE_KERNEL)), __pa 1304 arch/ia64/kernel/efi.c if (__pa(r[i].start) >= start && __pa(r[i].end) < end) { __pa 1305 arch/ia64/kernel/efi.c if (__pa(r[i].start) > start + size) __pa 1307 arch/ia64/kernel/efi.c start = ALIGN(__pa(r[i].end), alignment); __pa 1309 arch/ia64/kernel/efi.c __pa(r[i+1].start) < start + size) __pa 1864 arch/ia64/kernel/mca.c __this_cpu_write(ia64_mca_data, (__per_cpu_mca[cpu] = __pa(data))); __pa 1871 arch/ia64/kernel/mca.c pte_val(mk_pte_phys(__pa(cpu_data), PAGE_KERNEL))); __pa 1882 arch/ia64/kernel/mca.c __this_cpu_write(ia64_mca_pal_pte, pte_val(mk_pte_phys(__pa(pal_vaddr), __pa 167 arch/ia64/kernel/setup.c call_pernode_memory(__pa(range_start), range_end - range_start, func); __pa 197 arch/ia64/kernel/setup.c call_pernode_memory(__pa(start), end - start, func); __pa 318 arch/ia64/kernel/setup.c boot_param_res.start = __pa(ia64_boot_param); __pa 393 arch/ia64/kernel/setup.c phys_addr_t addr = __pa(region->start); __pa 447 arch/ia64/kernel/setup.c ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase)); __pa 1009 arch/ia64/kernel/setup.c ia64_set_kr(IA64_KR_PT_BASE, __pa(ia64_imva(empty_zero_page))); __pa 447 arch/ia64/kernel/smpboot.c ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase)); __pa 75 arch/ia64/mm/contig.c ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) - __pa 90 arch/ia64/mm/contig.c __pa(MAX_DMA_ADDRESS)); __pa 93 arch/ia64/mm/contig.c __func__, size, PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); __pa 365 arch/ia64/mm/discontig.c base = __pa(mem_data[node].pernode_addr); __pa 450 arch/ia64/mm/discontig.c __pa(MAX_DMA_ADDRESS), __pa 456 arch/ia64/mm/discontig.c __pa(MAX_DMA_ADDRESS)); __pa 474 arch/ia64/mm/discontig.c fill_pernode(node, __pa(pernode), pernodesize); __pa 437 arch/ia64/mm/init.c map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); __pa 438 arch/ia64/mm/init.c map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); __pa 442 arch/ia64/mm/init.c node = paddr_to_nid(__pa(start)); __pa 475 arch/ia64/mm/init.c set_pte(pte, pfn_pte(__pa(page) >> PAGE_SHIFT, __pa 501 arch/ia64/mm/init.c map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); __pa 502 arch/ia64/mm/init.c map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); __pa 586 arch/ia64/mm/init.c memblock_add_node(__pa(start), end - start, nid); __pa 595 arch/ia64/mm/init.c pfn_start = (PAGE_ALIGN(__pa(start))) >> PAGE_SHIFT; __pa 596 arch/ia64/mm/init.c pfn_end = (PAGE_ALIGN(__pa(end - 1))) >> PAGE_SHIFT; __pa 598 arch/ia64/mm/init.c pfn_start = GRANULEROUNDDOWN(__pa(start)) >> PAGE_SHIFT; __pa 599 arch/ia64/mm/init.c pfn_end = GRANULEROUNDUP(__pa(end - 1)) >> PAGE_SHIFT; __pa 177 arch/ia64/pci/pci.c base = __pa(io_space[space_nr].mmio_base); __pa 240 arch/m68k/include/asm/cacheflush_mm.h : : "a" (__pa(vaddr))); __pa 122 arch/m68k/include/asm/motorola_pgtable.h pgd_val(*pgdp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp); __pa 124 arch/m68k/include/asm/page_mm.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 23 arch/m68k/include/asm/page_no.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 30 arch/m68k/include/asm/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)pte); __pa 35 arch/m68k/include/asm/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)page_address(page)); __pa 18 arch/m68k/include/asm/sun3_pgtable.h #define VTOP(addr) __pa(addr) __pa 22 arch/m68k/include/asm/virtconvert.h return __pa(address); __pa 34 arch/m68k/include/asm/virtconvert.h __pa(PAGE_OFFSET + (((page) - pg_data_map[0].node_mem_map) << PAGE_SHIFT)) __pa 71 arch/m68k/mm/sun3mmu.c pg_table = (pte_t *) __pa (next_pgtable); __pa 97 arch/m68k/mm/sun3mmu.c (__pa(PAGE_OFFSET) >> PAGE_SHIFT) + 1, NULL); __pa 118 arch/m68k/sun3/config.c start_page = __pa(memory_start) >> PAGE_SHIFT; __pa 119 arch/m68k/sun3/config.c max_pfn = num_pages = __pa(memory_end) >> PAGE_SHIFT; __pa 143 arch/m68k/sun3/mmu_emu.c for (i=0; i < __pa(bootmem_end) / SUN3_PMEG_SIZE ; ++i) __pa 124 arch/m68k/sun3x/dvma.c __pa(kaddr), vaddr); __pa 160 arch/m68k/sun3x/dvma.c dvma_entry_set(index, __pa(kaddr)); __pa 148 arch/microblaze/include/asm/page.h # define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) __pa 153 arch/microblaze/include/asm/page.h # define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)) __pa 473 arch/microblaze/include/asm/pgtable.h #define pmd_page(pmd) (pfn_to_page(__pa(pmd_val(pmd)) >> PAGE_SHIFT)) __pa 311 arch/microblaze/mm/init.c kstart = __pa(CONFIG_KERNEL_START); /* kernel start */ __pa 163 arch/mips/bcm47xx/prom.c off = EXTVBASE + __pa(off); __pa 113 arch/mips/dec/prom/memory.c end = __pa(&_text) - 0x00020000; __pa 116 arch/mips/dec/prom/memory.c end = __pa(&_text); __pa 109 arch/mips/include/asm/io.h return __pa(address); __pa 216 arch/mips/include/asm/page.h #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0)) __pa 205 arch/mips/kernel/setup.c end = __pa(initrd_end); __pa 207 arch/mips/kernel/setup.c initrd_start = (unsigned long)__va(__pa(initrd_start)); __pa 254 arch/mips/kernel/setup.c if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { __pa 261 arch/mips/kernel/setup.c memblock_reserve(__pa(initrd_start), size); __pa 96 arch/mips/loongson64/common/mem.c return addr >= __pa(high_memory) || __pa 242 arch/mips/mm/cache.c return addr >= __pa(high_memory); __pa 54 arch/mips/mm/dma-noncoherent.c return (void *)(__pa(addr) + UNCAC_BASE); __pa 38 arch/mips/power/cpu.c unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); __pa 39 arch/mips/power/cpu.c unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end)); __pa 84 arch/nds32/include/asm/memory.h #define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)) __pa 55 arch/nds32/include/asm/pgalloc.h pmdval = __pa(pte_ptr) | _PAGE_KERNEL_TABLE; __pa 76 arch/nds32/include/asm/pgtable.h #define MAXMEM __pa(VMALLOC_START) __pa 265 arch/nds32/kernel/setup.c free_ram_start_pfn = PFN_UP(__pa(&_end)); __pa 65 arch/nds32/mm/init.c e = min((u32) memblock_end_of_DRAM(), (u32) __pa(high_memory)); __pa 85 arch/nds32/mm/init.c set_pmd(pme, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE)); __pa 120 arch/nds32/mm/init.c set_pmd(pmd, __pmd(__pa(fixmap_pmd_p) + _PAGE_KERNEL_TABLE)); __pa 135 arch/nds32/mm/init.c set_pmd(pmd, __pmd(__pa(pte) + _PAGE_KERNEL_TABLE)); __pa 173 arch/nds32/mm/init.c for (pfn = PFN_UP(__pa(high_memory)); pfn < max_pfn; pfn++) { __pa 410 arch/nds32/mm/proc.c unsigned long p_start = __pa(start); __pa 411 arch/nds32/mm/proc.c unsigned long p_end = __pa(end); __pa 532 arch/nds32/mm/proc.c __nds32__mtsr_isb(__pa(mm->pgd), NDS32_SR_L1_PPTB); __pa 152 arch/nios2/kernel/setup.c memory_start = PAGE_ALIGN((unsigned long)__pa(_end)); __pa 75 arch/openrisc/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 26 arch/openrisc/include/asm/pgalloc.h set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte))) __pa 317 arch/openrisc/include/asm/pgtable.h pte_val(pte) = __pa(page) | pgprot_val(pgprot); __pa 41 arch/openrisc/kernel/dma.c for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo->dcache_block_size) __pa 98 arch/openrisc/kernel/dma.c *dma_handle = __pa(page); __pa 82 arch/openrisc/kernel/setup.c memblock_reserve(__pa(_stext), _end - _stext); __pa 160 arch/openrisc/kernel/traps.c ((struct task_struct *)(__pa(current)))->comm, __pa 161 arch/openrisc/kernel/traps.c ((struct task_struct *)(__pa(current)))->pid, __pa 172 arch/openrisc/kernel/traps.c *((unsigned long *)(__pa(stack)))); __pa 179 arch/openrisc/kernel/traps.c addr = *((unsigned long *)__pa(stack)); __pa 195 arch/openrisc/kernel/traps.c c = ((unsigned char *)(__pa(regs->pc)))[i]; __pa 107 arch/openrisc/mm/init.c set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte))); __pa 302 arch/parisc/boot/compressed/misc.c kernel_len = __pa(SZ_end) - __pa(SZparisc_kernel_start); __pa 8 arch/parisc/include/asm/io.h #define virt_to_phys(a) ((unsigned long)__pa(a)) __pa 57 arch/parisc/include/asm/mmu_context.h mtctl(__pa(next->pgd), 25); __pa 172 arch/parisc/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 175 arch/parisc/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 40 arch/parisc/include/asm/pgalloc.h + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT)); __pa 65 arch/parisc/include/asm/pgalloc.h (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT)); __pa 116 arch/parisc/include/asm/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); __pa 120 arch/parisc/include/asm/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); __pa 477 arch/parisc/kernel/cache.c flush_dcache_page_asm(__pa(vfrom), vaddr); __pa 161 arch/parisc/kernel/firmware.c __pa(pdc_result), 0); __pa 242 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result)); __pa 267 arch/parisc/kernel/firmware.c __pa(pdc_result), __pa(pdc_result2), len); __pa 291 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data)); __pa 337 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result)); __pa 348 arch/parisc/kernel/firmware.c ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result)); __pa 395 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, __pa 396 arch/parisc/kernel/firmware.c index, __pa(pdc_result2), iodc_data_size); __pa 422 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), __pa 423 arch/parisc/kernel/firmware.c __pa(pdc_result2), mod_index); __pa 449 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_ADDRESS, __pa(pdc_result), __pa 471 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_INFO, __pa(pdc_result), 0); __pa 493 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_SYSMODEL, __pa(pdc_result), __pa 494 arch/parisc/kernel/firmware.c OS_ID_HPUX, __pa(name)); __pa 523 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_VERSIONS, __pa(pdc_result), id); __pa 545 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CPU_ID, __pa(pdc_result), 0); __pa 567 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CAPABILITIES, __pa(pdc_result), 0); __pa 596 arch/parisc/kernel/firmware.c __pa(orig_prod_num), __pa(current_prod_num), __pa(serial_no)); __pa 615 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_INFO, __pa(pdc_result), 0); __pa 636 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_CACHE, PDC_CACHE_RET_SPID, __pa(pdc_result), 0); __pa 657 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_BLOCK_TLB, PDC_BTLB_INFO, __pa(pdc_result), 0); __pa 687 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MEM_MAP, PDC_MEM_MAP_HPA, __pa(pdc_result), __pa 688 arch/parisc/kernel/firmware.c __pa(pdc_result2)); __pa 710 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa); __pa 740 arch/parisc/kernel/firmware.c __pa(pdc_result), count); __pa 768 arch/parisc/kernel/firmware.c __pa(pdc_result), count); __pa 790 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_STABLE, PDC_STABLE_RETURN_SIZE, __pa(pdc_result)); __pa 862 arch/parisc/kernel/firmware.c __pa(pdc_result), __pa(hwpath)); __pa 918 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa); __pa 945 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, __pa(tbl)); __pa 970 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, cfg_addr&~3UL, 4UL); __pa 993 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, __pa 1013 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_TOD, PDC_TOD_READ, __pa(pdc_result), 0); __pa 1028 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MEM, PDC_MEM_MEMINFO, __pa(pdc_result), 0); __pa 1043 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MEM, PDC_MEM_READ_PDT, __pa(pdc_result), __pa 1044 arch/parisc/kernel/firmware.c __pa(pdt_entries_ptr)); __pa 1091 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_MEM, PDC_MEM_TABLE, __pa(pdc_result), __pa(pdc_result2), entries); __pa 1149 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_INFO, __pa(pdc_result), 0); __pa 1176 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_SOFT_POWER, PDC_SOFT_POWER_ENABLE, __pa(pdc_result), sw_control); __pa 1253 arch/parisc/kernel/firmware.c PAGE0->mem_cons.spa, __pa(PAGE0->mem_cons.dp.layers), __pa 1254 arch/parisc/kernel/firmware.c __pa(iodc_retbuf), 0, __pa(iodc_dbuf), i, 0); __pa 1281 arch/parisc/kernel/firmware.c PAGE0->mem_kbd.spa, __pa(PAGE0->mem_kbd.dp.layers), __pa 1282 arch/parisc/kernel/firmware.c __pa(iodc_retbuf), 0, __pa(iodc_dbuf), 1, 0); __pa 1323 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_GET_NUMBER, __pa(pdc_result)); __pa 1349 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_PAT_CELL, PDC_PAT_CELL_MODULE, __pa(pdc_result), __pa 1350 arch/parisc/kernel/firmware.c ploc, mod, view_type, __pa(&result)); __pa 1379 arch/parisc/kernel/firmware.c __pa(pdc_result), __pa(&result), *actcnt, __pa 1404 arch/parisc/kernel/firmware.c __pa(&pdc_result), hpa); __pa 1426 arch/parisc/kernel/firmware.c __pa(pdc_result), cell_num); __pa 1447 arch/parisc/kernel/firmware.c __pa(r_addr), cell_num); __pa 1468 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_PAT_PD, PDC_PAT_PD_GET_ADDR_MAP, __pa(pdc_result), __pa 1469 arch/parisc/kernel/firmware.c __pa(pdc_result2), count, offset); __pa 1492 arch/parisc/kernel/firmware.c __pa(pdc_result)); __pa 1518 arch/parisc/kernel/firmware.c __pa(pdc_result), pci_addr, pci_size); __pa 1562 arch/parisc/kernel/firmware.c __pa(&pdc_result)); __pa 1585 arch/parisc/kernel/firmware.c __pa(&pdc_result), cell); __pa 1609 arch/parisc/kernel/firmware.c __pa(&pdc_result), parisc_cell_num, __pa 1610 arch/parisc/kernel/firmware.c __pa(pdt_entries_ptr)); __pa 1641 arch/parisc/kernel/firmware.c __pa(&pdc_result), __pa(pdt_entries_ptr), __pa 1670 arch/parisc/kernel/firmware.c __pa(&pdc_result), phys_addr); __pa 36 arch/parisc/kernel/kexec_file.c image->start = __pa(elf_info.ehdr->e_entry); __pa 39 arch/parisc/kernel/kexec_file.c image->segment[i].mem = __pa(image->segment[i].mem); __pa 412 arch/parisc/kernel/pci-dma.c paddr = __pa(paddr); __pa 191 arch/parisc/kernel/pdt.c pdt_status.first_dbe_loc <= __pa((unsigned long)&_end)) __pa 128 arch/parisc/kernel/setup.c if (__pa((unsigned long) &_end) >= KERNEL_INITIAL_SIZE) __pa 839 arch/parisc/kernel/traps.c ivap[6] = (u32)__pa(os_hpmc); __pa 294 arch/parisc/mm/init.c memblock_reserve(__pa(KERNEL_BINARY_TEXT_START), __pa 310 arch/parisc/mm/init.c if (__pa(initrd_start) < mem_max) { __pa 313 arch/parisc/mm/init.c if (__pa(initrd_end) > mem_max) { __pa 314 arch/parisc/mm/init.c initrd_reserve = mem_max - __pa(initrd_start); __pa 319 arch/parisc/mm/init.c printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initrd_start) + initrd_reserve, mem_max); __pa 321 arch/parisc/mm/init.c memblock_reserve(__pa(initrd_start), initrd_reserve); __pa 368 arch/parisc/mm/init.c ro_start = __pa((unsigned long)_text); __pa 369 arch/parisc/mm/init.c ro_end = __pa((unsigned long)&data_start); __pa 370 arch/parisc/mm/init.c kernel_start = __pa((unsigned long)&__init_begin); __pa 371 arch/parisc/mm/init.c kernel_end = __pa((unsigned long)&_end); __pa 388 arch/parisc/mm/init.c pmd = (pmd_t *)__pa(pg_dir); __pa 401 arch/parisc/mm/init.c pmd = (pmd_t *) __pa(pmd); __pa 423 arch/parisc/mm/init.c pg_table = (pte_t *) __pa(pg_table); __pa 480 arch/parisc/mm/init.c map_pages(start, __pa(start), end-start, __pa 496 arch/parisc/mm/init.c map_pages(init_end, __pa(init_end), kernel_end - init_end, __pa 506 arch/parisc/mm/init.c map_pages(init_begin, __pa(init_begin), init_end - init_begin, __pa 510 arch/parisc/mm/init.c map_pages(init_begin, __pa(init_begin), init_end - init_begin, __pa 671 arch/parisc/mm/init.c map_pages(initrd_start, __pa(initrd_start), __pa 698 arch/parisc/mm/init.c map_pages(linux_gateway_page_addr, __pa(&linux_gateway_page), __pa 31 arch/powerpc/include/asm/book3s/32/pgalloc.h *pmdp = __pmd(__pa(pte) | _PMD_PRESENT); __pa 37 arch/powerpc/include/asm/book3s/32/pgalloc.h *pmdp = __pmd(__pa(pte_page) | _PMD_PRESENT); __pa 309 arch/powerpc/include/asm/book3s/32/pgtable.h unsigned long ptephys = __pa(ptep) & PAGE_MASK; __pa 104 arch/powerpc/include/asm/book3s/64/hugetlb.h *hpdp = __hugepd(__pa(new) | HUGEPD_VAL_BITS | (shift_to_mmu_psize(pshift) << 2)); __pa 995 arch/powerpc/include/asm/book3s/64/pgtable.h #define __pgtable_ptr_val(ptr) __pa(ptr) __pa 777 arch/powerpc/include/asm/io.h return __pa((unsigned long)address); __pa 821 arch/powerpc/include/asm/io.h return __pa(address) + PCI_DRAM_OFFSET; __pa 35 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h *hpdp = __hugepd(__pa(new) | _PMD_USER | _PMD_PRESENT | __pa 23 arch/powerpc/include/asm/nohash/32/pgalloc.h *pmdp = __pmd(__pa(pte) | _PMD_PRESENT); __pa 32 arch/powerpc/include/asm/nohash/32/pgalloc.h *pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT); __pa 354 arch/powerpc/include/asm/nohash/32/pgtable.h pfn_to_page((__pa(pmd_val(pmd)) >> PAGE_SHIFT)) __pa 131 arch/powerpc/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 62 arch/powerpc/kernel/crash_dump.c create_trampoline(__pa(system_reset_fwnmi) - PHYSICAL_START); __pa 63 arch/powerpc/kernel/crash_dump.c create_trampoline(__pa(machine_check_fwnmi) - PHYSICAL_START); __pa 1063 arch/powerpc/kernel/dt_cpu_ftrs.c memblock_free(__pa(dt_cpu_features), __pa 667 arch/powerpc/kernel/fadump.c phdr->p_paddr = __pa(fw_dump.cpu_notes_buf_vaddr); __pa 170 arch/powerpc/kernel/machine_kexec.c if (overlaps_crashkernel(__pa(_stext), _end - _stext)) { __pa 271 arch/powerpc/kernel/machine_kexec.c kernel_end = cpu_to_be_ulong(__pa(_end)); __pa 50 arch/powerpc/kernel/machine_kexec_64.c if (image->segment[i].mem < __pa(_end)) __pa 408 arch/powerpc/kernel/machine_kexec_64.c htab_base = cpu_to_be64(__pa(htab_address)); __pa 141 arch/powerpc/kernel/machine_kexec_file_64.c ret = delete_fdt_mem_rsv(fdt, __pa(initial_boot_params), __pa 78 arch/powerpc/kernel/paca.c uv_share_page(PHYS_PFN(__pa(shared_lppaca)), __pa 286 arch/powerpc/kernel/paca.c memblock_free(__pa(paca_ptrs) + new_ptrs_size, __pa 295 arch/powerpc/kernel/paca.c memblock_free(__pa(paca_ptrs[boot_cpuid]->slb_shadow_ptr), __pa 37 arch/powerpc/kernel/proc_powerpc.c __pa(PDE_DATA(file_inode(file))) >> PAGE_SHIFT, __pa 120 arch/powerpc/kernel/prom.c start = __pa(initial_boot_params); __pa 626 arch/powerpc/kernel/prom.c memblock_reserve(_ALIGN_DOWN(__pa(initrd_start), PAGE_SIZE), __pa 733 arch/powerpc/kernel/prom.c memblock_reserve(PHYSICAL_START, __pa(klimit) - PHYSICAL_START); __pa 1766 arch/powerpc/kernel/prom_init.c os_term_args.args[0] = cpu_to_be32(__pa(str)); __pa 3160 arch/powerpc/kernel/prom_init.c prom_initrd_start = is_kernel_addr(r3) ? __pa(r3) : r3; __pa 379 arch/powerpc/kernel/rtas.c err_args.args[0] = cpu_to_be32(__pa(rtas_err_buf)); __pa 386 arch/powerpc/kernel/rtas.c enter_rtas(__pa(&rtas.args)); __pa 432 arch/powerpc/kernel/rtas.c enter_rtas(__pa(args)); __pa 736 arch/powerpc/kernel/rtas.c __pa(rtas_os_term_buf)); __pa 1123 arch/powerpc/kernel/rtas.c enter_rtas(__pa(&rtas.args)); __pa 456 arch/powerpc/kernel/rtas_flash.c (u32) __pa(rtas_data_buf), args_buf->buf_size); __pa 596 arch/powerpc/kernel/rtas_flash.c rtas_block_list = __pa(flist); __pa 610 arch/powerpc/kernel/rtas_flash.c f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data)); __pa 617 arch/powerpc/kernel/rtas_flash.c f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next)); __pa 421 arch/powerpc/kernel/rtasd.c __pa(logdata), rtas_error_log_max); __pa 831 arch/powerpc/kernel/setup-common.c memblock_free(__pa(cpu_to_phys_id), nr_cpu_ids * sizeof(u32)); __pa 762 arch/powerpc/kernel/setup_64.c return memblock_alloc_try_nid(size, align, __pa(MAX_DMA_ADDRESS), __pa 770 arch/powerpc/kernel/setup_64.c memblock_free(__pa(ptr), size); __pa 20 arch/powerpc/kernel/suspend.c unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; __pa 21 arch/powerpc/kernel/suspend.c unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; __pa 120 arch/powerpc/kvm/book3s_64_mmu_hv.c kvm->arch.sdr1 = __pa(info->virt) | (info->order - 18); __pa 41 arch/powerpc/kvm/book3s_64_mmu_radix.c __pa(to), __pa(from), n); __pa 3613 arch/powerpc/kvm/book3s_hv.c trap = plpar_hcall_norets(H_ENTER_NESTED, __pa(&hvregs), __pa 3614 arch/powerpc/kvm/book3s_hv.c __pa(&vcpu->arch.regs)); __pa 4566 arch/powerpc/kvm/book3s_hv.c __pa(kvm->arch.pgtable) | RADIX_PGD_INDEX_SIZE; __pa 369 arch/powerpc/kvm/book3s_hv_nested.c ptcr = __pa(pseries_partition_tb) | (ptb_order - 8); __pa 429 arch/powerpc/kvm/book3s_hv_nested.c __pa(gp->shadow_pgtable) | RADIX_PGD_INDEX_SIZE; __pa 383 arch/powerpc/mm/book3s32/mmu.c _SDR1 = __pa(Hash) | SDR1_LOW_BITS; __pa 40 arch/powerpc/mm/book3s32/tlb.c ptephys = __pa(ptep) & PAGE_MASK; __pa 801 arch/powerpc/mm/book3s64/hash_utils.c rc = htab_bolt_mapping(start, end, __pa(start), __pa 943 arch/powerpc/mm/book3s64/hash_utils.c BUG_ON(htab_bolt_mapping(base, base + size, __pa(base), __pa 963 arch/powerpc/mm/book3s64/hash_utils.c __pa(tce_alloc_start), prot, __pa 1078 arch/powerpc/mm/book3s64/hash_utils.c set_ptcr_when_no_uv(__pa(partition_tb) | __pa 1885 arch/powerpc/mm/book3s64/hash_utils.c ret = hpte_insert_repeating(hash, vpn, __pa(vaddr), mode, __pa 1925 arch/powerpc/mm/book3s64/hash_utils.c lmi = __pa(vaddr) >> PAGE_SHIFT; __pa 167 arch/powerpc/mm/book3s64/mmu_context.c process_tb[index].prtb0 = cpu_to_be64(rts_field | __pa(mm->pgd) | RADIX_PGD_INDEX_SIZE); __pa 207 arch/powerpc/mm/book3s64/pgtable.c ptcr = __pa(partition_tb) | (PATB_SIZE_SHIFT - 12); __pa 370 arch/powerpc/mm/book3s64/radix_pgtable.c process_tb->prtb0 = cpu_to_be64(rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE); __pa 395 arch/powerpc/mm/book3s64/radix_pgtable.c dw0 = rts_field | __pa(init_mm.pgd) | RADIX_PGD_INDEX_SIZE | PATB_HR; __pa 396 arch/powerpc/mm/book3s64/radix_pgtable.c dw1 = __pa(process_tb) | (PRTB_SIZE_SHIFT - 12) | PATB_GR; __pa 623 arch/powerpc/mm/book3s64/radix_pgtable.c set_ptcr_when_no_uv(__pa(partition_tb) | __pa 710 arch/powerpc/mm/book3s64/radix_pgtable.c create_physical_mapping(__pa(params->aligned_start), __pa(params->start), -1); __pa 711 arch/powerpc/mm/book3s64/radix_pgtable.c create_physical_mapping(__pa(params->end), __pa(params->aligned_end), -1); __pa 875 arch/powerpc/mm/book3s64/radix_pgtable.c return create_physical_mapping(__pa(start), __pa(end), nid); __pa 222 arch/powerpc/mm/init_64.c vmemmap_list_populate(__pa(p), start, node); __pa 227 arch/powerpc/mm/init_64.c rc = vmemmap_create_mapping(start, page_size, __pa(p)); __pa 26 arch/powerpc/mm/kasan/kasan_init_32.c phys_addr_t pa = __pa(kasan_early_shadow_page); __pa 99 arch/powerpc/mm/kasan/kasan_init_32.c pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL); __pa 116 arch/powerpc/mm/kasan/kasan_init_32.c phys_addr_t pa = __pa(kasan_early_shadow_page); __pa 194 arch/powerpc/mm/kasan/kasan_init_32.c modify_instruction_site(&patch__hash_page_A0, 0xffff, __pa(early_hash) >> 16); __pa 195 arch/powerpc/mm/kasan/kasan_init_32.c modify_instruction_site(&patch__flush_hash_A0, 0xffff, __pa(early_hash) >> 16); __pa 36 arch/powerpc/mm/nohash/8xx.c return __pa(va); __pa 121 arch/powerpc/mm/nohash/8xx.c _ALIGN(__pa(_einittext), 8 << 20)); __pa 148 arch/powerpc/mm/nohash/8xx.c mmu_patch_cmp_limit(&patch__itlbmiss_linmem_top, __pa(_etext)); __pa 156 arch/powerpc/mm/nohash/8xx.c -__pa(((unsigned long)_sinittext) & __pa 158 arch/powerpc/mm/nohash/8xx.c mmu_patch_addis(&patch__dtlbmiss_romem_top, -__pa(_sinittext)); __pa 184 arch/powerpc/mm/nohash/8xx.c s16 offset = (s16)(__pa(swapper_pg_dir)); __pa 197 arch/powerpc/mm/nohash/8xx.c mtspr(SPRN_M_TWB, __pa(pgd) - offset); __pa 59 arch/powerpc/mm/nohash/book3e_pgtable.c __pa(MAX_DMA_ADDRESS), NUMA_NO_NODE); __pa 63 arch/powerpc/mm/nohash/book3e_pgtable.c __func__, size, size, __pa(MAX_DMA_ADDRESS)); __pa 296 arch/powerpc/mm/pgtable.c return __pa(pfn_to_kaddr(pfn)) + offset_in_page(va); __pa 194 arch/powerpc/oprofile/op_model_cell.c u64 paddr = __pa(address); __pa 603 arch/powerpc/perf/imc-pmu.c __pa((void *)mem_info->vbase), __pa 1113 arch/powerpc/perf/imc-pmu.c rc = opal_imc_counters_init(OPAL_IMC_COUNTERS_TRACE, __pa((void *)local_mem), __pa 113 arch/powerpc/platforms/44x/iss4xx.c spin_table[1] = __pa(start_secondary_47x); __pa 184 arch/powerpc/platforms/44x/ppc476.c spin_table[1] = __pa(start_secondary_47x); __pa 253 arch/powerpc/platforms/85xx/smp.c __pa(ppc_function_entry(generic_secondary_smp_init))); __pa 255 arch/powerpc/platforms/85xx/smp.c out_be32(&spin_table->addr_l, __pa(__early_start)); __pa 142 arch/powerpc/platforms/cell/iommu.c | (__pa(pte) & IOC_IOPT_CacheInvd_IOPTE_Mask) __pa 189 arch/powerpc/platforms/cell/iommu.c io_pte[i] = base_pte | (__pa(uaddr) & CBE_IOPTE_RPN_Mask); __pa 216 arch/powerpc/platforms/cell/iommu.c __pa(window->iommu->pad_page) | __pa 368 arch/powerpc/platforms/cell/iommu.c iommu->stab[i] = reg | (__pa(ptab) + (n_pte_pages << 12) * __pa 407 arch/powerpc/platforms/cell/iommu.c reg = IOC_IOST_Origin_E | __pa(iommu->stab) | IOC_IOST_Origin_HW; __pa 872 arch/powerpc/platforms/cell/iommu.c ptab[offset] = base_pte | (__pa(addr) & CBE_IOPTE_RPN_Mask); __pa 962 arch/powerpc/platforms/cell/iommu.c hbase = __pa(htab_address); __pa 137 arch/powerpc/platforms/cell/ras.c addr = __pa(page_address(area->pages)) + (PAGE_SIZE >> 1); __pa 71 arch/powerpc/platforms/cell/smp.c __pa(ppc_function_entry(generic_secondary_smp_init)); __pa 35 arch/powerpc/platforms/chrp/nvram.c __pa(nvram_buf), 1) != 0) || 1 != done) __pa 57 arch/powerpc/platforms/chrp/nvram.c __pa(nvram_buf), 1) != 0) || 1 != done) __pa 92 arch/powerpc/platforms/pasemi/iommu.c rpn = __pa(uaddr) >> IOBMAP_PAGE_SHIFT; __pa 224 arch/powerpc/platforms/pasemi/iommu.c regword = IOBMAP_L1E_V | (__pa(iob_l2_base + i*0x2000) >> 12); __pa 335 arch/powerpc/platforms/powermac/smp.c unsigned long start = __pa(__secondary_start_pmac_0) + nr * 8; __pa 152 arch/powerpc/platforms/powernv/npu-dma.c __pa(tbl->it_base), __pa 259 arch/powerpc/platforms/powernv/opal-dump.c addr = __pa(list); __pa 166 arch/powerpc/platforms/powernv/opal-elog.c opal_rc = opal_read_elog(__pa(elog->buffer), __pa 209 arch/powerpc/platforms/powernv/opal-elog.c rc = opal_read_elog(__pa(elog->buffer), __pa 137 arch/powerpc/platforms/powernv/opal-flash.c ret = opal_validate_flash(__pa(buf), &size, &result); __pa 293 arch/powerpc/platforms/powernv/opal-flash.c addr = __pa(list); __pa 301 arch/powerpc/platforms/powernv/opal-hmi.c while (opal_get_msg(__pa(&msg), sizeof(msg)) == OPAL_SUCCESS) { __pa 36 arch/powerpc/platforms/powernv/opal-nvram.c rc = opal_read_nvram(__pa(buf), count, off); __pa 59 arch/powerpc/platforms/powernv/opal-nvram.c rc = opal_write_nvram(__pa(buf), count, off); __pa 49 arch/powerpc/platforms/powernv/opal-powercap.c ret = opal_get_powercap(pcap_attr->handle, token, (u32 *)__pa(&pcap)); __pa 43 arch/powerpc/platforms/powernv/opal-psr.c (u32 *)__pa(&psr)); __pa 61 arch/powerpc/platforms/powernv/opal-xscom.c rc = opal_xscom_read(chip, reg, (__be64 *)__pa(&v)); __pa 278 arch/powerpc/platforms/powernv/opal.c ret = opal_get_msg(__pa(opal_msg), opal_msg_size); __pa 840 arch/powerpc/platforms/powernv/opal.c __pa(addr), size); __pa 1076 arch/powerpc/platforms/powernv/opal.c sg->next = cpu_to_be64(__pa(next)); __pa 77 arch/powerpc/platforms/powernv/pci-ioda-tce.c tce = __pa(tmp2) | TCE_PCI_READ | TCE_PCI_WRITE; __pa 101 arch/powerpc/platforms/powernv/pci-ioda-tce.c u64 rpn = __pa(uaddr) >> tbl->it_page_shift; __pa 253 arch/powerpc/platforms/powernv/pci-ioda-tce.c addr[i] = cpu_to_be64(__pa(tmp) | __pa 1828 arch/powerpc/platforms/powernv/pci-ioda.c __pa(tces), __pa 1912 arch/powerpc/platforms/powernv/pci-ioda.c start = __pa(((__be64 *)tbl->it_base) + index - tbl->it_offset); __pa 1913 arch/powerpc/platforms/powernv/pci-ioda.c end = __pa(((__be64 *)tbl->it_base) + index - tbl->it_offset + __pa 2258 arch/powerpc/platforms/powernv/pci-ioda.c __pa(addr) + tce32_segsz * i, __pa 2319 arch/powerpc/platforms/powernv/pci-ioda.c __pa(tbl->it_base), __pa 3700 arch/powerpc/platforms/powernv/pci-ioda.c memblock_free(__pa(phb), sizeof(struct pnv_phb)); __pa 68 arch/powerpc/platforms/powernv/smp.c __pa(ppc_function_entry(generic_secondary_smp_init)); __pa 403 arch/powerpc/platforms/powernv/vas-window.c val = __pa(winctx->rx_fifo); __pa 788 arch/powerpc/platforms/ps3/device-init.c lpar = ps3_mm_phys_to_lpar(__pa(buf)); __pa 531 arch/powerpc/platforms/ps3/interrupt.c lpar_addr = ps3_mm_phys_to_lpar(__pa(virt_addr_bmp)); __pa 757 arch/powerpc/platforms/ps3/interrupt.c ps3_mm_phys_to_lpar(__pa(&pd->bmp))); __pa 760 arch/powerpc/platforms/ps3/interrupt.c pd->thread_id, ps3_mm_phys_to_lpar(__pa(&pd->bmp))); __pa 776 arch/powerpc/platforms/ps3/mm.c unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr) __pa 831 arch/powerpc/platforms/ps3/mm.c unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr) __pa 1072 arch/powerpc/platforms/ps3/mm.c unsigned long phys_addr = is_kernel_addr(virt_addr) ? __pa(virt_addr) __pa 1120 arch/powerpc/platforms/ps3/mm.c lpar_addr = addr ? ps3_mm_phys_to_lpar(__pa(addr)) : 0; __pa 188 arch/powerpc/platforms/pseries/cmm.c if ((rc = plpar_page_set_loaned(__pa(addr)))) { __pa 232 arch/powerpc/platforms/pseries/cmm.c plpar_page_set_active(__pa(addr)); __pa 571 arch/powerpc/platforms/pseries/cmm.c plpar_page_set_active(__pa(pa_curr->page[idx])); __pa 134 arch/powerpc/platforms/pseries/dtl.c addr = __pa(dtl->buf); __pa 932 arch/powerpc/platforms/pseries/hotplug-cpu.c __pa(cede_parameters), __pa 127 arch/powerpc/platforms/pseries/io_event_irq.c __pa(ioei_rtas_buf), __pa 106 arch/powerpc/platforms/pseries/iommu.c rpn = __pa(uaddr) >> TCE_SHIFT; __pa 149 arch/powerpc/platforms/pseries/iommu.c rpn = __pa(uaddr) >> tceshift; __pa 220 arch/powerpc/platforms/pseries/iommu.c rpn = __pa(uaddr) >> TCE_SHIFT; __pa 240 arch/powerpc/platforms/pseries/iommu.c (u64)__pa(tcep), __pa 462 arch/powerpc/platforms/pseries/iommu.c (u64)__pa(tcep), __pa 129 arch/powerpc/platforms/pseries/lpar.c ret = register_dtl(hwcpu, __pa(dtl)); __pa 673 arch/powerpc/platforms/pseries/lpar.c addr = __pa(&lppaca_of(cpu)); __pa 688 arch/powerpc/platforms/pseries/lpar.c addr = __pa(paca_ptrs[cpu]->slb_shadow_ptr); __pa 1430 arch/powerpc/platforms/pseries/lpar.c __pa(rtas_data_buf), __pa 1712 arch/powerpc/platforms/pseries/lpar.c pseries_lpar_register_process_table(__pa(process_tb), __pa 1749 arch/powerpc/platforms/pseries/lpar.c addr = __pa((unsigned long)page_address(page)); __pa 306 arch/powerpc/platforms/pseries/lparcfg.c __pa(rtas_data_buf), __pa 63 arch/powerpc/platforms/pseries/nvram.c if ((rtas_call(nvram_fetch, 3, 2, &done, i, __pa(nvram_buf), __pa 108 arch/powerpc/platforms/pseries/nvram.c if ((rtas_call(nvram_store, 3, 2, &done, i, __pa(nvram_buf), __pa 207 arch/powerpc/platforms/pseries/pseries_energy.c rc = plpar_hcall9(H_BEST_ENERGY, retbuf, flags, 0, __pa(buf_page), __pa 294 arch/powerpc/platforms/pseries/ras.c RTAS_HOTPLUG_EVENTS, 0, __pa(&ras_log_buf), __pa 337 arch/powerpc/platforms/pseries/ras.c critical, __pa(&ras_log_buf), __pa 368 arch/powerpc/platforms/pseries/ras.c __pa(&ras_log_buf), __pa 519 arch/powerpc/platforms/pseries/rtas-fadump.c rtas_fadump_get_config(fadump_conf, (void *)__pa(fdm_active)); __pa 70 arch/powerpc/platforms/pseries/scanlog.c (u32) __pa(rtas_data_buf), (u32) count); __pa 123 arch/powerpc/platforms/pseries/setup.c system_reset_addr = __pa(system_reset_fwnmi) - PHYSICAL_START; __pa 124 arch/powerpc/platforms/pseries/setup.c machine_check_addr = __pa(machine_check_fwnmi) - PHYSICAL_START; __pa 862 arch/powerpc/platforms/pseries/setup.c __pa(rtas_data_buf), __pa 93 arch/powerpc/platforms/pseries/smp.c __pa(ppc_function_entry(generic_secondary_smp_init)); __pa 42 arch/powerpc/platforms/pseries/svm.c uv_unshare_page(PHYS_PFN(__pa(addr)), numpages); __pa 52 arch/powerpc/platforms/pseries/svm.c uv_share_page(PHYS_PFN(__pa(addr)), numpages); __pa 76 arch/powerpc/platforms/pseries/svm.c unsigned long pfn = PHYS_PFN(__pa(addr)); __pa 190 arch/powerpc/sysdev/dart_iommu.c rpn = __pa(uaddr) >> DART_PAGE_SHIFT; __pa 142 arch/powerpc/sysdev/xive/native.c qpage_phys = __pa(qpage); __pa 649 arch/powerpc/sysdev/xive/native.c opal_xive_donate_page(chip, __pa(p)); __pa 472 arch/powerpc/sysdev/xive/spapr.c qpage_phys = __pa(qpage); __pa 103 arch/riscv/include/asm/page.h #define virt_to_pfn(vaddr) (phys_to_pfn(__pa(vaddr))) __pa 69 arch/riscv/mm/init.c if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { __pa 75 arch/riscv/mm/init.c memblock_reserve(__pa(initrd_start), size); __pa 94 arch/riscv/mm/init.c phys_addr_t vmlinux_end = __pa(&_end); __pa 95 arch/riscv/mm/init.c phys_addr_t vmlinux_start = __pa(&_start); __pa 274 arch/riscv/mm/init.c #define get_pgd_next_virt(__pa) get_pmd_virt(__pa) __pa 275 arch/riscv/mm/init.c #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ __pa 276 arch/riscv/mm/init.c create_pmd_mapping(__nextp, __va, __pa, __sz, __prot) __pa 282 arch/riscv/mm/init.c #define get_pgd_next_virt(__pa) get_pte_virt(__pa) __pa 283 arch/riscv/mm/init.c #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ __pa 284 arch/riscv/mm/init.c create_pte_mapping(__nextp, __va, __pa, __sz, __prot) __pa 421 arch/riscv/mm/init.c __pa(fixmap_pgd_next), __pa 433 arch/riscv/mm/init.c if (start <= __pa(PAGE_OFFSET) && __pa 434 arch/riscv/mm/init.c __pa(PAGE_OFFSET) < end) __pa 435 arch/riscv/mm/init.c start = __pa(PAGE_OFFSET); __pa 450 arch/riscv/mm/init.c csr_write(CSR_SATP, PFN_DOWN(__pa(swapper_pg_dir)) | SATP_MODE); __pa 28 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__sdma) = __pa(&_sdma); __pa 29 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__edma) = __pa(&_edma); __pa 30 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__stext_dma) = __pa(&_stext_dma); __pa 31 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__etext_dma) = __pa(&_etext_dma); __pa 52 arch/s390/boot/startup.c unsigned long __bootdata_preserved(__swsusp_reset_dma) = __pa(_swsusp_reset_dma); __pa 32 arch/s390/include/asm/idals.h return ((__pa(vaddr) + length - 1) >> 31) != 0; __pa 41 arch/s390/include/asm/idals.h return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + __pa 54 arch/s390/include/asm/idals.h paddr = __pa(vaddr); __pa 47 arch/s390/include/asm/mmu_context.h mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 52 arch/s390/include/asm/mmu_context.h mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 57 arch/s390/include/asm/mmu_context.h mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 62 arch/s390/include/asm/mmu_context.h mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 164 arch/s390/include/asm/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 177 arch/s390/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 104 arch/s390/include/asm/pgalloc.h pgd_val(*pgd) = _REGION1_ENTRY | __pa(p4d); __pa 109 arch/s390/include/asm/pgalloc.h p4d_val(*p4d) = _REGION2_ENTRY | __pa(pud); __pa 114 arch/s390/include/asm/pgalloc.h pud_val(*pud) = _REGION3_ENTRY | __pa(pmd); __pa 143 arch/s390/include/asm/pgalloc.h pmd_val(*pmd) = _SEGMENT_ENTRY + __pa(pte); __pa 1287 arch/s390/include/asm/pgtable.h #define pfn_pte(pfn,pgprot) mk_pte_phys(__pa((pfn) << PAGE_SHIFT),(pgprot)) __pa 1615 arch/s390/include/asm/pgtable.h #define pfn_pmd(pfn, pgprot) mk_pmd_phys(__pa((pfn) << PAGE_SHIFT), (pgprot)) __pa 142 arch/s390/kernel/crash_dump.c from = __pa(src); __pa 185 arch/s390/kernel/crash_dump.c from = __pa(src); __pa 58 arch/s390/kernel/early.c end_pfn = PFN_UP(__pa(_end)); __pa 832 arch/s390/kernel/setup.c unsigned long start_pfn = PFN_UP(__pa(_end)); __pa 586 arch/s390/kernel/smp.c pa = __pa(&pcpu->lowcore->floating_pt_save_area); __pa 592 arch/s390/kernel/smp.c pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK); __pa 154 arch/s390/kernel/suspend.c unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); __pa 155 arch/s390/kernel/suspend.c unsigned long nosave_end_pfn = PFN_DOWN(__pa(&__nosave_end)); __pa 156 arch/s390/kernel/suspend.c unsigned long end_rodata_pfn = PFN_DOWN(__pa(__end_rodata)) - 1; __pa 157 arch/s390/kernel/suspend.c unsigned long stext_pfn = PFN_DOWN(__pa(_stext)); __pa 1280 arch/s390/kvm/vsie.c struct page *page = pfn_to_page(__pa(vsie_page) >> PAGE_SHIFT); __pa 158 arch/s390/mm/dump_pagetables.c if ((pud_val(*pud) & PAGE_MASK) == __pa(kasan_early_shadow_pmd)) { __pa 189 arch/s390/mm/dump_pagetables.c if ((p4d_val(*p4d) & PAGE_MASK) == __pa(kasan_early_shadow_pud)) { __pa 219 arch/s390/mm/dump_pagetables.c if ((pgd_val(*pgd) & PAGE_MASK) == __pa(kasan_early_shadow_p4d)) { __pa 80 arch/s390/mm/gmap.c _ASCE_USER_BITS | __pa(table); __pa 1339 arch/s390/mm/gmap.c page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); __pa 1367 arch/s390/mm/gmap.c page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); __pa 1396 arch/s390/mm/gmap.c page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT); __pa 1424 arch/s390/mm/gmap.c page = pfn_to_page(__pa(sgt) >> PAGE_SHIFT); __pa 1453 arch/s390/mm/gmap.c page = pfn_to_page(__pa(r3t) >> PAGE_SHIFT); __pa 1481 arch/s390/mm/gmap.c page = pfn_to_page(__pa(r3t) >> PAGE_SHIFT); __pa 1510 arch/s390/mm/gmap.c page = pfn_to_page(__pa(r2t) >> PAGE_SHIFT); __pa 1541 arch/s390/mm/gmap.c page = pfn_to_page(__pa(r2t) >> PAGE_SHIFT); __pa 102 arch/s390/mm/init.c init_mm.context.asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits; __pa 168 arch/s390/mm/kasan_init.c pmd_val(*pm_dir) = __pa(page) | sgt_prot; __pa 187 arch/s390/mm/kasan_init.c pte_val(*pt_dir) = __pa(page) | pgt_prot; __pa 192 arch/s390/mm/kasan_init.c pte_val(*pt_dir) = __pa(page) | pgt_prot; __pa 196 arch/s390/mm/kasan_init.c pte_val(*pt_dir) = __pa(page) | pgt_prot_zero; __pa 209 arch/s390/mm/kasan_init.c S390_lowcore.kernel_asce = (__pa(pgd) & PAGE_MASK) | asce_bits; __pa 249 arch/s390/mm/kasan_init.c pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY); __pa 250 arch/s390/mm/kasan_init.c pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY); __pa 251 arch/s390/mm/kasan_init.c p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY); __pa 256 arch/s390/mm/kasan_init.c pte_z = __pte(__pa(kasan_early_shadow_page) | pgt_prot); __pa 132 arch/s390/mm/pageattr.c pmd_val(new) = __pa(pt_dir) | _SEGMENT_ENTRY; __pa 209 arch/s390/mm/pageattr.c pud_val(new) = __pa(pm_dir) | _REGION3_ENTRY; __pa 112 arch/s390/mm/pgalloc.c mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 120 arch/s390/mm/pgalloc.c mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 147 arch/s390/mm/pgalloc.c mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH | __pa 253 arch/s390/mm/pgalloc.c page = pfn_to_page(__pa(table) >> PAGE_SHIFT); __pa 256 arch/s390/mm/pgalloc.c bit = (__pa(table) & ~PAGE_MASK)/(PTRS_PER_PTE*sizeof(pte_t)); __pa 283 arch/s390/mm/pgalloc.c page = pfn_to_page(__pa(table) >> PAGE_SHIFT); __pa 286 arch/s390/mm/pgalloc.c table = (unsigned long *) (__pa(table) | 3); __pa 290 arch/s390/mm/pgalloc.c bit = (__pa(table) & ~PAGE_MASK) / (PTRS_PER_PTE*sizeof(pte_t)); __pa 299 arch/s390/mm/pgalloc.c table = (unsigned long *) (__pa(table) | (1U << bit)); __pa 307 arch/s390/mm/pgalloc.c struct page *page = pfn_to_page(__pa(table) >> PAGE_SHIFT); __pa 270 arch/s390/mm/vmem.c pmd_val(*pm_dir) = __pa(new_page) | sgt_prot; __pa 290 arch/s390/mm/vmem.c pte_val(*pt_dir) = __pa(new_page) | pgt_prot; __pa 131 arch/sh/boot/compressed/misc.c output_addr = __pa((unsigned long)&_text+PAGE_SIZE); __pa 101 arch/sh/drivers/pci/pci-sh5.c unsigned long memStart = __pa(memory_start); __pa 102 arch/sh/drivers/pci/pci-sh5.c unsigned long memSize = __pa(memory_end) - memStart; __pa 300 arch/sh/drivers/pci/pci-sh7780.c memphys = __pa(memory_start); __pa 359 arch/sh/drivers/pci/pcie-sh7786.c memstart = __pa(memory_start); __pa 360 arch/sh/drivers/pci/pcie-sh7786.c memend = __pa(memory_end); __pa 246 arch/sh/include/asm/io.h #define virt_to_phys(address) (__pa(address)) __pa 179 arch/sh/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 183 arch/sh/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 168 arch/sh/kernel/setup.c memblock_reserve(__pa(initrd_start), INITRD_SIZE); __pa 21 arch/sh/kernel/swsusp.c unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; __pa 22 arch/sh/kernel/swsusp.c unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; __pa 49 arch/sh/mm/cache-sh3.c p = __pa(v); __pa 144 arch/sh/mm/cache-sh7705.c __flush_dcache_page(__pa(page_address(page))); __pa 255 arch/sh/mm/init.c start_pfn = PFN_UP(__pa(_end)); __pa 155 arch/sh/mm/mmap.c if (addr + count > __pa(high_memory)) __pa 89 arch/sh/mm/pmb.c return ppn >= __pa(memory_start) && ppn < __pa(memory_end); __pa 41 arch/sparc/include/asm/mmu_context_64.h __tsb_context_switch(__pa(mm->pgd), __pa 50 arch/sparc/include/asm/mmu_context_64.h , __pa(&mm->context.tsb_descr[MM_TSB_BASE]), __pa 127 arch/sparc/include/asm/page_32.h #define virt_to_phys __pa __pa 131 arch/sparc/include/asm/page_32.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 152 arch/sparc/include/asm/page_64.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT) __pa 154 arch/sparc/include/asm/page_64.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 156 arch/sparc/include/asm/page_64.h #define virt_to_phys __pa __pa 400 arch/sparc/include/asm/pgtable_32.h (test_bit(__pa((unsigned long)(addr))>>20, sparc_valid_addr_bitmap)) __pa 830 arch/sparc/include/asm/pgtable_64.h unsigned long val = __pa((unsigned long) (ptep)); __pa 836 arch/sparc/include/asm/pgtable_64.h (pud_val(*(pudp)) = (__pa((unsigned long) (pmdp)))) __pa 888 arch/sparc/include/asm/pgtable_64.h (pgd_val(*(pgdp)) = (__pa((unsigned long) (pudp)))) __pa 144 arch/sparc/include/asm/pgtsrmmu.h #define __nocache_pa(VADDR) (((unsigned long)VADDR) - SRMMU_NOCACHE_VADDR + __pa((unsigned long)srmmu_nocache_pool)) __pa 129 arch/sparc/kernel/iommu.c iommu->dummy_page_pa = (unsigned long) __pa(iommu->dummy_page); __pa 234 arch/sparc/kernel/iommu.c first_page = __pa(first_page); __pa 299 arch/sparc/kernel/iommu.c base_paddr = __pa(oaddr & IO_PAGE_MASK); __pa 36 arch/sparc/kernel/iommu_common.h #define SG_ENT_PHYS_ADDRESS(SG) (__pa(sg_virt((SG)))) __pa 282 arch/sparc/kernel/irq_64.c irq = bucket_get_irq(__pa(bucket)); __pa 623 arch/sparc/kernel/irq_64.c irq = bucket_get_irq(__pa(bucket)); __pa 626 arch/sparc/kernel/irq_64.c bucket_set_irq(__pa(bucket), irq); __pa 688 arch/sparc/kernel/irq_64.c cookie = ~__pa(&ihd->bucket); __pa 744 arch/sparc/kernel/irq_64.c bucket_set_irq(__pa(bucket), irq); __pa 1030 arch/sparc/kernel/irq_64.c *pa_ptr = __pa(p); __pa 1048 arch/sparc/kernel/irq_64.c tb->cpu_mondo_block_pa = __pa(mondo); __pa 1056 arch/sparc/kernel/irq_64.c tb->cpu_list_pa = __pa(page); __pa 1116 arch/sparc/kernel/irq_64.c ivector_table_pa = __pa(ivector_table); __pa 994 arch/sparc/kernel/ldc.c *ra = __pa(q); __pa 1081 arch/sparc/kernel/ldc.c hv_err = sun4v_ldc_set_map_table(lp->id, __pa(table), __pa 2206 arch/sparc/kernel/ldc.c pa = __pa(buf); __pa 2277 arch/sparc/kernel/ldc.c ra = __pa(buf); __pa 430 arch/sparc/kernel/leon_pci_grpci1.c ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN((unsigned long) &_end)); __pa 623 arch/sparc/kernel/leon_pci_grpci2.c ahbadr = 0xf0000000 & (u32)__pa(PAGE_ALIGN( __pa 191 arch/sparc/kernel/mdesc.c start = __pa(hp); __pa 516 arch/sparc/kernel/mdesc.c status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len); __pa 1339 arch/sparc/kernel/mdesc.c status = sun4v_mach_desc(__pa(&hp->mdesc), len, &real_len); __pa 64 arch/sparc/kernel/pci_fire.c upa_writeq(__pa(iommu->page_table) | 0x7UL, iommu->iommu_tsbbase); __pa 246 arch/sparc/kernel/pci_fire.c __pa(pbm->msi_queues)), __pa 436 arch/sparc/kernel/pci_msi.c __pa(pbm->msi_queues)); __pa 453 arch/sparc/kernel/pci_psycho.c __pa(pbm->stc.strbuf_flushflag); __pa 1112 arch/sparc/kernel/pci_schizo.c __pa(pbm->stc.strbuf_flushflag); __pa 1205 arch/sparc/kernel/pci_schizo.c upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); __pa 105 arch/sparc/kernel/pci_sun4v.c __pa(pglist)); __pa 111 arch/sparc/kernel/pci_sun4v.c npages, prot, __pa(pglist), __pa 122 arch/sparc/kernel/pci_sun4v.c __pa(pglist), __pa 129 arch/sparc/kernel/pci_sun4v.c __pa(pglist), ret); __pa 226 arch/sparc/kernel/pci_sun4v.c first_page = __pa(first_page); __pa 392 arch/sparc/kernel/pci_sun4v.c base_paddr = __pa(oaddr & IO_PAGE_MASK); __pa 767 arch/sparc/kernel/pci_sun4v.c iotsb->ra = __pa(table); __pa 1074 arch/sparc/kernel/pci_sun4v.c unsigned long err, base = __pa(pages + (i * q_size)); __pa 429 arch/sparc/kernel/psycho_common.c upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); __pa 126 arch/sparc/kernel/ptrace_64.c unsigned long start = __pa(kaddr); __pa 586 arch/sparc/kernel/sbus.c __pa(strbuf->strbuf_flushflag); __pa 625 arch/sparc/kernel/sbus.c upa_writeq(__pa(iommu->page_table), iommu->iommu_tsbbase); __pa 338 arch/sparc/kernel/smp_64.c __pa(hdesc)); __pa 901 arch/sparc/kernel/smp_64.c if (tp->pgd_paddr == __pa(mm->pgd)) __pa 936 arch/sparc/kernel/smp_64.c __flush_icache_page(__pa(page_address(page))); __pa 969 arch/sparc/kernel/smp_64.c xcall_deliver(data0, __pa(pg_addr), __pa 1005 arch/sparc/kernel/smp_64.c xcall_deliver(data0, __pa(pg_addr), __pa 1585 arch/sparc/kernel/smp_64.c const unsigned long goal = __pa(MAX_DMA_ADDRESS); __pa 1595 arch/sparc/kernel/smp_64.c cpu, size, __pa(ptr)); __pa 1600 arch/sparc/kernel/smp_64.c "%016lx\n", cpu, size, node, __pa(ptr)); __pa 1610 arch/sparc/kernel/smp_64.c memblock_free(__pa(ptr), size); __pa 116 arch/sparc/kernel/sysfs.c ra = __pa(&per_cpu(mmu_stats, smp_processor_id())); __pa 173 arch/sparc/mm/init_32.c start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end)); __pa 212 arch/sparc/mm/init_64.c __flush_icache_page(__pa(page_address(page))); __pa 274 arch/sparc/mm/init_64.c tsb_addr = __pa(tsb_addr); __pa 759 arch/sparc/mm/init_64.c start = __pa(start); __pa 760 arch/sparc/mm/init_64.c end = __pa(end); __pa 1661 arch/sparc/mm/init_64.c unsigned long pa = __pa(addr); __pa 1801 arch/sparc/mm/init_64.c unsigned long this_end, paddr = __pa(vstart); __pa 2495 arch/sparc/mm/init_64.c if (paddr >= __pa(initrd_start) && __pa 2496 arch/sparc/mm/init_64.c paddr < __pa(PAGE_ALIGN(initrd_end))) __pa 2633 arch/sparc/mm/init_64.c pmd_val(*pmd) = pte_base | __pa(block); __pa 133 arch/sparc/mm/io-unit.c iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); __pa 255 arch/sparc/mm/io-unit.c sbus_writel(iopte_val(MKIOPTE(__pa(page))), iopte); __pa 108 arch/sparc/mm/iommu.c base = __pa((unsigned long)iommu->page_table) >> 4; __pa 327 arch/sparc/mm/srmmu.c paddr = __pa((unsigned long)srmmu_nocache_pool); __pa 130 arch/sparc/mm/tsb.c base = __pa(base); __pa 144 arch/sparc/mm/tsb.c base = __pa(base); __pa 163 arch/sparc/mm/tsb.c base = __pa(base); __pa 178 arch/sparc/mm/tsb.c base = __pa(base); __pa 216 arch/sparc/mm/tsb.c tsb_paddr = __pa(mm->context.tsb_block[tsb_idx].tsb); __pa 506 arch/sparc/mm/tsb.c old_tsb_base = __pa(old_tsb_base); __pa 507 arch/sparc/mm/tsb.c new_tsb_base = __pa(new_tsb_base); __pa 114 arch/um/drivers/mmapper_kern.c p_buf = __pa(v_buf); __pa 565 arch/um/drivers/virtio_uml.c rc = vhost_user_init_mem_region(__pa(end_iomem), highmem, __pa 112 arch/um/include/asm/page.h #define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v))) __pa 16 arch/um/include/asm/pgalloc.h set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte))) __pa 61 arch/um/include/asm/pgtable-3level.h set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd))) __pa 283 arch/um/include/asm/pgtable.h #define __virt_to_page(virt) phys_to_page(__pa(virt)) __pa 49 arch/um/kernel/mem.c map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0); __pa 50 arch/um/kernel/mem.c memblock_free(__pa(brk_end), uml_reserved - brk_end); __pa 75 arch/um/kernel/mem.c (unsigned long) __pa(pte))); __pa 89 arch/um/kernel/mem.c set_pud(pud, __pud(_KERNPG_TABLE + (unsigned long) __pa(pmd_table))); __pa 143 arch/um/kernel/mem.c p = __pa(v); __pa 107 arch/um/kernel/physmem.c os_seek_file(physmem_fd, __pa(__syscall_stub_start)); __pa 111 arch/um/kernel/physmem.c memblock_add(__pa(start), len + highmem); __pa 112 arch/um/kernel/physmem.c memblock_reserve(__pa(start), reserve); __pa 114 arch/um/kernel/physmem.c min_low_pfn = PFN_UP(__pa(reserve_end)); __pa 126 arch/um/kernel/physmem.c else if (phys < __pa(end_iomem)) { __pa 139 arch/um/kernel/physmem.c else if (phys < __pa(end_iomem) + highmem) { __pa 212 arch/um/kernel/physmem.c region->phys = __pa(region->virt); __pa 94 arch/unicore32/include/asm/memory.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 76 arch/unicore32/include/asm/pgalloc.h __pmd_populate(pmdp, __pa(pte_ptr) | _PAGE_KERNEL_TABLE); __pa 156 arch/unicore32/include/asm/tlbflush.h : : "r" (__pa(pmd) & ~(L1_CACHE_BYTES - 1)) : "cc"); __pa 55 arch/unicore32/kernel/hibernate.c set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_KERNEL_TABLE)); __pa 144 arch/unicore32/kernel/hibernate.c unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; __pa 145 arch/unicore32/kernel/hibernate.c unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; __pa 134 arch/unicore32/mm/init.c memblock_reserve(__pa(_text), _end - _text); __pa 215 arch/unicore32/mm/init.c pg = PAGE_ALIGN(__pa(start_pg)); __pa 216 arch/unicore32/mm/init.c pgend = __pa(end_pg) & PAGE_MASK; __pa 152 arch/unicore32/mm/mmu.c __pmd_populate(pmd, __pa(pte) | prot); __pa 284 arch/unicore32/mm/mmu.c lowmem_limit = __pa(vmalloc_min - 1) + 1; __pa 334 arch/unicore32/mm/mmu.c memblock_reserve(__pa(swapper_pg_dir), PTRS_PER_PGD * sizeof(pgd_t)); __pa 117 arch/x86/entry/vdso/vma.c __pa(pvti) >> PAGE_SHIFT, __pa 131 arch/x86/include/asm/io.h return __pa(address); __pa 90 arch/x86/include/asm/mem_encrypt.h #define __sme_pa(x) (__pa(x) | sme_me_mask) __pa 41 arch/x86/include/asm/page.h #ifndef __pa __pa 63 arch/x86/include/asm/page.h #define __boot_pa(x) __pa(x) __pa 69 arch/x86/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 66 arch/x86/include/asm/pgalloc.h paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); __pa 67 arch/x86/include/asm/pgalloc.h set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); __pa 73 arch/x86/include/asm/pgalloc.h paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); __pa 74 arch/x86/include/asm/pgalloc.h set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); __pa 126 arch/x86/include/asm/pgalloc.h paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); __pa 127 arch/x86/include/asm/pgalloc.h set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); __pa 132 arch/x86/include/asm/pgalloc.h paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); __pa 133 arch/x86/include/asm/pgalloc.h set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); __pa 140 arch/x86/include/asm/pgalloc.h paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); __pa 141 arch/x86/include/asm/pgalloc.h set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); __pa 146 arch/x86/include/asm/pgalloc.h paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); __pa 147 arch/x86/include/asm/pgalloc.h set_p4d_safe(p4d, __p4d(_PAGE_TABLE | __pa(pud))); __pa 178 arch/x86/include/asm/pgalloc.h paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); __pa 179 arch/x86/include/asm/pgalloc.h set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); __pa 186 arch/x86/include/asm/pgalloc.h paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); __pa 187 arch/x86/include/asm/pgalloc.h set_pgd_safe(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); __pa 492 arch/x86/include/asm/uv/uv_hub.h return uv_soc_phys_ram_to_nasid(__pa(v)); __pa 519 arch/x86/include/asm/uv/uv_hub.h return uv_soc_phys_ram_to_gpa(__pa(v)); __pa 306 arch/x86/include/asm/xen/page.h #define virt_to_machine(v) (phys_to_machine(XPADDR(__pa(v)))) __pa 307 arch/x86/include/asm/xen/page.h #define virt_to_pfn(v) (PFN_DOWN(__pa(v))) __pa 552 arch/x86/kernel/amd_gart_64.c enable_gart_translation(dev, __pa(agp_gatt_table)); __pa 806 arch/x86/kernel/amd_gart_64.c gart_unmapped_entry = GPTE_ENCODE(__pa(scratch)); __pa 2692 arch/x86/kernel/apic/io_apic.c ioapic_phys = __pa(ioapic_phys); __pa 155 arch/x86/kernel/check.c pr_err("Corrupted low memory at %p (%lx phys) = %08lx\n", addr, __pa(addr), *addr); __pa 1507 arch/x86/kernel/cpu/resctrl/pseudo_lock.c physical = __pa(plr->kmem) >> PAGE_SHIFT; __pa 170 arch/x86/kernel/espfix_64.c pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); __pa 171 arch/x86/kernel/espfix_64.c paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); __pa 182 arch/x86/kernel/espfix_64.c pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); __pa 183 arch/x86/kernel/espfix_64.c paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); __pa 194 arch/x86/kernel/espfix_64.c pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); __pa 74 arch/x86/kernel/head32.c #ifdef __pa __pa 82 arch/x86/kernel/head32.c const unsigned long limit = __pa(_end) + __pa 85 arch/x86/kernel/head32.c pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd); __pa 88 arch/x86/kernel/head32.c pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table); __pa 92 arch/x86/kernel/head32.c ptep = (pte_t *)__pa(__brk_base); __pa 112 arch/x86/kernel/head32.c ptr = (unsigned long *)__pa(&max_pfn_mapped); __pa 116 arch/x86/kernel/head32.c ptr = (unsigned long *)__pa(&_brk_end); __pa 101 arch/x86/kernel/machine_kexec_32.c set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); __pa 107 arch/x86/kernel/machine_kexec_32.c set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); __pa 123 arch/x86/kernel/machine_kexec_32.c (unsigned long)control_page, __pa(control_page)); __pa 129 arch/x86/kernel/machine_kexec_32.c __pa(control_page), __pa(control_page)); __pa 213 arch/x86/kernel/machine_kexec_32.c page_list[PA_CONTROL_PAGE] = __pa(control_page); __pa 215 arch/x86/kernel/machine_kexec_32.c page_list[PA_PGD] = __pa(image->arch.pgd); __pa 135 arch/x86/kernel/machine_kexec_64.c paddr = __pa(page_address(image->control_code_page)+PAGE_SIZE); __pa 142 arch/x86/kernel/machine_kexec_64.c set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE)); __pa 150 arch/x86/kernel/machine_kexec_64.c set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE)); __pa 158 arch/x86/kernel/machine_kexec_64.c set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE)); __pa 166 arch/x86/kernel/machine_kexec_64.c set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); __pa 409 arch/x86/kernel/machine_kexec_64.c (unsigned long)__pa(page_address(image->control_code_page)); __pa 771 arch/x86/kernel/pci-calgary_64.c table_phys = (u64)__pa(tbl->it_base); __pa 103 arch/x86/kernel/setup_percpu.c const unsigned long goal = __pa(MAX_DMA_ADDRESS); __pa 113 arch/x86/kernel/setup_percpu.c cpu, size, __pa(ptr)); __pa 120 arch/x86/kernel/setup_percpu.c cpu, size, node, __pa(ptr)); __pa 138 arch/x86/kernel/setup_percpu.c memblock_free(__pa(ptr), size); __pa 176 arch/x86/kernel/tce_64.c memblock_free(__pa(tbl), size); __pa 676 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); __pa 760 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); __pa 1502 arch/x86/kvm/mmu.c sp = page_header(__pa(spte)); __pa 1514 arch/x86/kvm/mmu.c sp = page_header(__pa(spte)); __pa 1606 arch/x86/kvm/mmu.c WARN_ON(page_header(__pa(sptep))->role.level == __pa 1619 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); __pa 2093 arch/x86/kvm/mmu.c sp = page_header(__pa(spte)); __pa 2215 arch/x86/kvm/mmu.c sp = page_header(__pa(spte)); __pa 2545 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(spte)); __pa 2710 arch/x86/kvm/mmu.c spte = __pa(sp->spt) | shadow_present_mask | PT_WRITABLE_MASK | __pa 3061 arch/x86/kvm/mmu.c sp = page_header(__pa(sptep)); __pa 3286 arch/x86/kvm/mmu.c sp = page_header(__pa(sptep)); __pa 3561 arch/x86/kvm/mmu.c sp = page_header(__pa(iterator.sptep)); __pa 3792 arch/x86/kvm/mmu.c vcpu->arch.mmu->root_hpa = __pa(sp->spt); __pa 3805 arch/x86/kvm/mmu.c root = __pa(sp->spt); __pa 3810 arch/x86/kvm/mmu.c vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root); __pa 3847 arch/x86/kvm/mmu.c root = __pa(sp->spt); __pa 3884 arch/x86/kvm/mmu.c root = __pa(sp->spt); __pa 3890 arch/x86/kvm/mmu.c vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root); __pa 3909 arch/x86/kvm/mmu.c lm_root[0] = __pa(vcpu->arch.mmu->pae_root) | pm_mask; __pa 3914 arch/x86/kvm/mmu.c vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->lm_root); __pa 6023 arch/x86/kvm/mmu.c sp = page_header(__pa(sptep)); __pa 100 arch/x86/kvm/mmu_audit.c sp = page_header(__pa(sptep)); __pa 135 arch/x86/kvm/mmu_audit.c rev_sp = page_header(__pa(sptep)); __pa 168 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); __pa 586 arch/x86/kvm/paging_tmpl.h sp = page_header(__pa(sptep)); __pa 917 arch/x86/kvm/paging_tmpl.h sp = page_header(__pa(sptep)); __pa 1603 arch/x86/kvm/svm.c control->msrpm_base_pa = __sme_set(__pa(svm->msrpm)); __pa 3521 arch/x86/kvm/svm.c svm->vmcb->control.msrpm_base_pa = __sme_set(__pa(svm->nested.msrpm)); __pa 6446 arch/x86/kvm/svm.c start->dh_cert_address = __sme_set(__pa(dh_blob)); __pa 6458 arch/x86/kvm/svm.c start->session_address = __sme_set(__pa(session_blob)); __pa 2000 arch/x86/kvm/vmx/nested.c vmcs_write64(MSR_BITMAP, __pa(vmx->nested.vmcs02.msr_bitmap)); __pa 2023 arch/x86/kvm/vmx/nested.c vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val)); __pa 2024 arch/x86/kvm/vmx/nested.c vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val)); __pa 4803 arch/x86/kvm/vmx/nested.c __pa(vmx->vmcs01.shadow_vmcs)); __pa 5778 arch/x86/kvm/vmx/nested.c vmcs_write64(VMREAD_BITMAP, __pa(vmx_vmread_bitmap)); __pa 5779 arch/x86/kvm/vmx/nested.c vmcs_write64(VMWRITE_BITMAP, __pa(vmx_vmwrite_bitmap)); __pa 236 arch/x86/kvm/vmx/ops.h u64 phys_addr = __pa(vmcs); __pa 243 arch/x86/kvm/vmx/ops.h u64 phys_addr = __pa(vmcs); __pa 554 arch/x86/kvm/vmx/vmx.c __pa(*p_hv_pa_pg); __pa 2226 arch/x86/kvm/vmx/vmx.c u64 phys_addr = __pa(per_cpu(vmxarea, cpu)); __pa 4155 arch/x86/kvm/vmx/vmx.c vmcs_write64(MSR_BITMAP, __pa(vmx->vmcs01.msr_bitmap)); __pa 4179 arch/x86/kvm/vmx/vmx.c vmcs_write64(POSTED_INTR_DESC_ADDR, __pa((&vmx->pi_desc))); __pa 4203 arch/x86/kvm/vmx/vmx.c vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host.val)); __pa 4205 arch/x86/kvm/vmx/vmx.c vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest.val)); __pa 4331 arch/x86/kvm/vmx/vmx.c __pa(vcpu->arch.apic->regs)); __pa 7519 arch/x86/kvm/vmx/vmx.c vcpu_info.pi_desc_addr = __pa(vcpu_to_pi_desc(vcpu)); __pa 385 arch/x86/mm/dump_pagetables.c if (__pa(pt) == __pa(kasan_early_shadow_pmd) || __pa 387 arch/x86/mm/dump_pagetables.c __pa(pt) == __pa(kasan_early_shadow_p4d)) || __pa 388 arch/x86/mm/dump_pagetables.c __pa(pt) == __pa(kasan_early_shadow_pud)) { __pa 55 arch/x86/mm/ident_map.c set_pud(pud, __pud(__pa(pmd) | info->kernpg_flag)); __pa 83 arch/x86/mm/ident_map.c set_p4d(p4d, __p4d(__pa(pud) | info->kernpg_flag)); __pa 127 arch/x86/mm/ident_map.c set_pgd(pgd, __pgd(__pa(p4d) | info->kernpg_flag)); __pa 134 arch/x86/mm/ident_map.c set_pgd(pgd, __pgd(__pa(pud) | info->kernpg_flag)); __pa 115 arch/x86/mm/init.c ret = __pa(extend_brk(PAGE_SIZE * num, PAGE_SIZE)); __pa 156 arch/x86/mm/init.c base = __pa(extend_brk(tables, PAGE_SIZE)); __pa 76 arch/x86/mm/init_32.c paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT); __pa 77 arch/x86/mm/init_32.c set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); __pa 101 arch/x86/mm/init_32.c paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); __pa 102 arch/x86/mm/init_32.c set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); __pa 183 arch/x86/mm/init_32.c paravirt_alloc_pte(&init_mm, __pa(newpte) >> PAGE_SHIFT); __pa 184 arch/x86/mm/init_32.c set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE)); __pa 188 arch/x86/mm/init_32.c paravirt_release_pte(__pa(pte) >> PAGE_SHIFT); __pa 504 arch/x86/mm/init_32.c pfn, pmd, __pa(pmd)); __pa 513 arch/x86/mm/init_32.c pfn, pmd, __pa(pmd), pte, __pa(pte)); __pa 516 arch/x86/mm/init_32.c paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT); __pa 376 arch/x86/mm/init_64.c set_pgd(pgd, __pgd(__pa(p4d) | _KERNPG_TABLE | __pa 382 arch/x86/mm/init_64.c set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE | __pa 388 arch/x86/mm/init_64.c set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE | __pa 689 arch/x86/mm/init_64.c paddr = __pa(vaddr); __pa 692 arch/x86/mm/init_64.c paddr_next = __pa(vaddr_next); __pa 704 arch/x86/mm/init_64.c paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), __pa 710 arch/x86/mm/init_64.c paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end), __pa 743 arch/x86/mm/init_64.c paddr_last = phys_p4d_init(p4d, __pa(vaddr), __pa 744 arch/x86/mm/init_64.c __pa(vaddr_end), __pa 751 arch/x86/mm/init_64.c paddr_last = phys_p4d_init(p4d, __pa(vaddr), __pa(vaddr_end), __pa 1485 arch/x86/mm/init_64.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT, __pa 30 arch/x86/mm/kasan_init_64.c __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); __pa 34 arch/x86/mm/kasan_init_64.c (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS)); __pa 51 arch/x86/mm/kasan_init_64.c if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) __pa 54 arch/x86/mm/kasan_init_64.c memblock_free(__pa(p), PMD_SIZE); __pa 70 arch/x86/mm/kasan_init_64.c entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); __pa 88 arch/x86/mm/kasan_init_64.c if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL)) __pa 91 arch/x86/mm/kasan_init_64.c memblock_free(__pa(p), PUD_SIZE); __pa 326 arch/x86/mm/kasan_init_64.c __pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE)); __pa 367 arch/x86/mm/kasan_init_64.c early_pfn_to_nid(__pa(_stext))); __pa 388 arch/x86/mm/kasan_init_64.c pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); __pa 183 arch/x86/mm/kaslr.c __p4d(_KERNPG_TABLE | __pa(pud_page_tramp))); __pa 186 arch/x86/mm/kaslr.c __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp))); __pa 189 arch/x86/mm/kaslr.c __pgd(_KERNPG_TABLE | __pa(pud_page_tramp))); __pa 309 arch/x86/mm/mem_encrypt.c kernel_physical_mapping_change(__pa(vaddr & pmask), __pa 310 arch/x86/mm/mem_encrypt.c __pa((vaddr_end & pmask) + psize), __pa 117 arch/x86/mm/mem_encrypt_identity.c set_pgd(pgd, __pgd(PGD_FLAGS | __pa(p4d))); __pa 125 arch/x86/mm/mem_encrypt_identity.c set_p4d(p4d, __p4d(P4D_FLAGS | __pa(pud))); __pa 133 arch/x86/mm/mem_encrypt_identity.c set_pud(pud, __pud(PUD_FLAGS | __pa(pmd))); __pa 173 arch/x86/mm/mem_encrypt_identity.c set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte))); __pa 220 arch/x86/mm/mmap.c return addr + count - 1 <= __pa(high_memory - 1); __pa 339 arch/x86/mm/numa.c memblock_free(__pa(numa_distance), size); __pa 521 arch/x86/mm/numa_emulation.c memblock_free(__pa(phys_dist), phys_size); __pa 1006 arch/x86/mm/pageattr.c unsigned long pfn = PFN_DOWN(__pa(address)); __pa 1207 arch/x86/mm/pageattr.c set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); __pa 1217 arch/x86/mm/pageattr.c set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE)); __pa 1411 arch/x86/mm/pageattr.c set_pgd(pgd_entry, __pgd(__pa(p4d) | _KERNPG_TABLE)); __pa 1423 arch/x86/mm/pageattr.c set_p4d(p4d, __p4d(__pa(pud) | _KERNPG_TABLE)); __pa 1475 arch/x86/mm/pageattr.c cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; __pa 1808 arch/x86/mm/pageattr.c ret = reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE, __pa 1820 arch/x86/mm/pageattr.c free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE); __pa 1846 arch/x86/mm/pageattr.c ret = reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE, __pa 1853 arch/x86/mm/pageattr.c free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE); __pa 1880 arch/x86/mm/pageattr.c free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE); __pa 850 arch/x86/mm/pat.c if (base > __pa(high_memory-1)) __pa 860 arch/x86/mm/pat.c id_sz = (__pa(high_memory-1) <= base + size) ? __pa 861 arch/x86/mm/pat.c __pa(high_memory) - base : __pa 57 arch/x86/mm/pgtable.c paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT); __pa 72 arch/x86/mm/pgtable.c paravirt_release_pud(__pa(pud) >> PAGE_SHIFT); __pa 79 arch/x86/mm/pgtable.c paravirt_release_p4d(__pa(p4d) >> PAGE_SHIFT); __pa 183 arch/x86/mm/pgtable.c paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); __pa 187 arch/x86/mm/pgtable.c set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT)); __pa 187 arch/x86/mm/pti.c set_pgd(pgd, __pgd(_KERNPG_TABLE | __pa(new_p4d_page))); __pa 216 arch/x86/mm/pti.c set_p4d(p4d, __p4d(_KERNPG_TABLE | __pa(new_pud_page))); __pa 230 arch/x86/mm/pti.c set_pud(pud, __pud(_KERNPG_TABLE | __pa(new_pmd_page))); __pa 266 arch/x86/mm/pti.c set_pmd(pmd, __pmd(_KERNPG_TABLE | __pa(new_pte_page))); __pa 490 arch/x86/mm/tlb.c WARN_ON((cr3 & CR3_ADDR_MASK) != __pa(mm->pgd)); __pa 969 arch/x86/platform/efi/efi.c pa = __pa(new_memmap); __pa 400 arch/x86/platform/efi/efi_64.c text = __pa(_text); __pa 41 arch/x86/power/hibernate_32.c set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); __pa 66 arch/x86/power/hibernate_32.c set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); __pa 141 arch/x86/power/hibernate_32.c __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); __pa 189 arch/x86/power/hibernate_32.c temp_pgt = __pa(resume_pg_dir); __pa 71 arch/x86/power/hibernate_64.c __pud(__pa(pmd) | pgprot_val(pgtable_prot))); __pa 73 arch/x86/power/hibernate_64.c p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot)); __pa 74 arch/x86/power/hibernate_64.c pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); __pa 80 arch/x86/power/hibernate_64.c pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); __pa 123 arch/x86/power/hibernate_64.c temp_pgt = __pa(pgd); __pa 65 arch/x86/realmode/init.c phys_base = __pa(base); __pa 129 arch/x86/realmode/init.c __pa(base); __pa 160 arch/x86/xen/efi.c boot_params->efi_info.efi_systab = (__u32)__pa(efi_systab_xen); __pa 161 arch/x86/xen/efi.c boot_params->efi_info.efi_systab_hi = (__u32)(__pa(efi_systab_xen) >> 32); __pa 108 arch/x86/xen/enlighten_hvm.c pfn = __pa(hypercall_page); __pa 1342 arch/x86/xen/enlighten_pv.c initrd_start = __pa(xen_start_info->mod_start); __pa 1349 arch/x86/xen/enlighten_pv.c boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line); __pa 36 arch/x86/xen/enlighten_pvh.c pfn = __pa(hypercall_page); __pa 53 arch/x86/xen/mmu_hvm.c a.gpa = __pa(mm->pgd); __pa 807 arch/x86/xen/mmu_pv.c xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd))); __pa 812 arch/x86/xen/mmu_pv.c PFN_DOWN(__pa(user_pgd))); __pa 821 arch/x86/xen/mmu_pv.c xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd))); __pa 926 arch/x86/xen/mmu_pv.c xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); __pa 934 arch/x86/xen/mmu_pv.c PFN_DOWN(__pa(user_pgd))); __pa 1002 arch/x86/xen/mmu_pv.c if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd)) __pa 1021 arch/x86/xen/mmu_pv.c if (per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd)) __pa 1037 arch/x86/xen/mmu_pv.c if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd)) __pa 1129 arch/x86/xen/mmu_pv.c unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK; __pa 1250 arch/x86/xen/mmu_pv.c memblock_free(__pa(addr), size); __pa 1274 arch/x86/xen/mmu_pv.c xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base)); __pa 1437 arch/x86/xen/mmu_pv.c __xen_write_cr3(false, __pa(user_pgd)); __pa 1505 arch/x86/xen/mmu_pv.c __pgd(__pa(level3_user_vsyscall) | _PAGE_TABLE); __pa 1764 arch/x86/xen/mmu_pv.c unsigned long pfn = __pa(addr) >> PAGE_SHIFT; __pa 1800 arch/x86/xen/mmu_pv.c pmd[pmdidx] = __pmd(__pa(pte_page) | _PAGE_TABLE); __pa 1854 arch/x86/xen/mmu_pv.c if (*pt_base == PFN_DOWN(__pa(addr))) { __pa 1859 arch/x86/xen/mmu_pv.c if (*pt_end == PFN_DOWN(__pa(addr))) { __pa 1890 arch/x86/xen/mmu_pv.c max_pfn_mapped = PFN_DOWN(__pa(xen_start_info->mfn_list)); __pa 1892 arch/x86/xen/mmu_pv.c pt_base = PFN_DOWN(__pa(xen_start_info->pt_base)); __pa 1965 arch/x86/xen/mmu_pv.c pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); __pa 1972 arch/x86/xen/mmu_pv.c __xen_write_cr3(true, __pa(init_top_pgt)); __pa 1990 arch/x86/xen/mmu_pv.c xen_start_info = (struct start_info *)__va(__pa(xen_start_info)); __pa 2179 arch/x86/xen/mmu_pv.c unsigned long pfn = PFN_DOWN(__pa(swapper_pg_dir)); __pa 2181 arch/x86/xen/mmu_pv.c BUG_ON(read_cr3_pa() != __pa(initial_page_table)); __pa 2182 arch/x86/xen/mmu_pv.c BUG_ON(cr3 != __pa(swapper_pg_dir)); __pa 2198 arch/x86/xen/mmu_pv.c __pgd(__pa(swapper_kernel_pmd) | _PAGE_PRESENT); __pa 2206 arch/x86/xen/mmu_pv.c PFN_DOWN(__pa(initial_page_table))); __pa 2223 arch/x86/xen/mmu_pv.c pt_base = min(__pa(xen_start_info->pt_base), __pa(pmd)); __pa 2254 arch/x86/xen/mmu_pv.c __pgd(__pa(initial_kernel_pmd) | _PAGE_PRESENT); __pa 2260 arch/x86/xen/mmu_pv.c pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); __pa 2263 arch/x86/xen/mmu_pv.c PFN_DOWN(__pa(initial_page_table))); __pa 2264 arch/x86/xen/mmu_pv.c xen_write_cr3(__pa(initial_page_table)); __pa 2274 arch/x86/xen/mmu_pv.c memblock_reserve(__pa(xen_start_info), PAGE_SIZE); __pa 2317 arch/x86/xen/mmu_pv.c pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL); __pa 2327 arch/x86/xen/mmu_pv.c pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL); __pa 2800 arch/x86/xen/mmu_pv.c return __pa(vmcoreinfo_note); __pa 346 arch/x86/xen/p2m.c paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); __pa 348 arch/x86/xen/p2m.c paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); __pa 351 arch/x86/xen/p2m.c pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); __pa 353 arch/x86/xen/p2m.c pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); __pa 390 arch/x86/xen/p2m.c pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); __pa 400 arch/x86/xen/p2m.c pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); __pa 410 arch/x86/xen/p2m.c set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); __pa 459 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) __pa 497 arch/x86/xen/p2m.c paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); __pa 509 arch/x86/xen/p2m.c __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); __pa 518 arch/x86/xen/p2m.c paravirt_release_pte(__pa(pte_newpg[i]) >> PAGE_SHIFT); __pa 590 arch/x86/xen/p2m.c if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) || __pa 591 arch/x86/xen/p2m.c p2m_pfn == PFN_DOWN(__pa(p2m_missing))) { __pa 599 arch/x86/xen/p2m.c if (p2m_pfn == PFN_DOWN(__pa(p2m_missing))) __pa 610 arch/x86/xen/p2m.c pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL)); __pa 674 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing))) __pa 677 arch/x86/xen/p2m.c if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) __pa 310 arch/x86/xen/setup.c if (pfn >= PFN_UP(__pa(high_memory - 1))) __pa 712 arch/x86/xen/setup.c start = __pa(xen_start_info->mfn_list); __pa 353 arch/x86/xen/smp_pv.c per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir); __pa 188 arch/xtensa/include/asm/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) __pa 190 arch/xtensa/include/asm/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) __pa 335 arch/xtensa/kernel/setup.c !mem_reserve(__pa(initrd_start), __pa(initrd_end))) __pa 341 arch/xtensa/kernel/setup.c mem_reserve(__pa(_stext), __pa(_end)); __pa 344 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_WindowVectors_text_start), __pa 345 arch/xtensa/kernel/setup.c __pa(&_WindowVectors_text_end)); __pa 347 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_DebugInterruptVector_text_start), __pa 348 arch/xtensa/kernel/setup.c __pa(&_DebugInterruptVector_text_end)); __pa 350 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_KernelExceptionVector_text_start), __pa 351 arch/xtensa/kernel/setup.c __pa(&_KernelExceptionVector_text_end)); __pa 353 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_UserExceptionVector_text_start), __pa 354 arch/xtensa/kernel/setup.c __pa(&_UserExceptionVector_text_end)); __pa 356 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_DoubleExceptionVector_text_start), __pa 357 arch/xtensa/kernel/setup.c __pa(&_DoubleExceptionVector_text_end)); __pa 360 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_Level2InterruptVector_text_start), __pa 361 arch/xtensa/kernel/setup.c __pa(&_Level2InterruptVector_text_end)); __pa 364 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_Level3InterruptVector_text_start), __pa 365 arch/xtensa/kernel/setup.c __pa(&_Level3InterruptVector_text_end)); __pa 368 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_Level4InterruptVector_text_start), __pa 369 arch/xtensa/kernel/setup.c __pa(&_Level4InterruptVector_text_end)); __pa 372 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_Level5InterruptVector_text_start), __pa 373 arch/xtensa/kernel/setup.c __pa(&_Level5InterruptVector_text_end)); __pa 376 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_Level6InterruptVector_text_start), __pa 377 arch/xtensa/kernel/setup.c __pa(&_Level6InterruptVector_text_end)); __pa 383 arch/xtensa/kernel/setup.c mem_reserve(__pa(&_SecondaryResetVector_text_start), __pa 384 arch/xtensa/kernel/setup.c __pa(&_SecondaryResetVector_text_end)); __pa 441 arch/xtensa/kernel/setup.c unsigned long paddr = __pa(vaddr); __pa 369 drivers/block/ps3disk.c ps3_mm_phys_to_lpar(__pa(&ata_cmnd)), __pa 690 drivers/block/ps3vram.c xdr_lpar = ps3_mm_phys_to_lpar(__pa(priv->xdr_buf)); __pa 815 drivers/block/ps3vram.c ps3_mm_phys_to_lpar(__pa(priv->xdr_buf)), __pa 149 drivers/char/hw_random/n2-drv.c unsigned long ra = __pa(&np->scratch_control[0]); __pa 422 drivers/char/hw_random/n2-drv.c unsigned long ra = __pa(&np->test_data); __pa 457 drivers/char/hw_random/n2-drv.c unsigned long ra = __pa(&np->test_data); __pa 467 drivers/char/hw_random/n2-drv.c unsigned long post_ctl_ra = __pa(post_control); __pa 468 drivers/char/hw_random/n2-drv.c unsigned long pre_ctl_ra = __pa(pre_control); __pa 469 drivers/char/hw_random/n2-drv.c unsigned long buffer_ra = __pa(buffer); __pa 621 drivers/char/hw_random/n2-drv.c unsigned long ctl_ra = __pa(&up->control[0]); __pa 54 drivers/char/mem.c return addr + count <= __pa(high_memory); __pa 308 drivers/char/mem.c return addr >= __pa(high_memory); __pa 417 drivers/char/mem.c pfn = __pa((u64)vma->vm_pgoff << PAGE_SHIFT) >> PAGE_SHIFT; __pa 178 drivers/char/powernv-op-panel.c oppanel_lines[i].line = cpu_to_be64(__pa(&oppanel_data[i * __pa 31 drivers/cpuidle/cpuidle-calxeda.c return psci_ops.cpu_suspend(CALXEDA_IDLE_PARAM, __pa(cpu_resume)); __pa 85 drivers/crypto/cavium/zip/zip_deflate.c zip_cmd->s.inp_ptr_addr.s.addr = __pa(zip_ops->input); __pa 91 drivers/crypto/cavium/zip/zip_deflate.c zip_cmd->s.out_ptr_addr.s.addr = __pa(zip_ops->output); __pa 97 drivers/crypto/cavium/zip/zip_deflate.c zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr); __pa 134 drivers/crypto/cavium/zip/zip_device.c ncp.s.addr = __pa(zip_dev->iq[queue].sw_head); __pa 137 drivers/crypto/cavium/zip/zip_device.c *ncb_ptr, __pa(zip_dev->iq[queue].sw_head)); __pa 94 drivers/crypto/cavium/zip/zip_inflate.c zip_cmd->s.out_ptr_addr.s.addr = __pa(zip_ops->output); __pa 104 drivers/crypto/cavium/zip/zip_inflate.c zip_cmd->s.inp_ptr_addr.s.addr = __pa((u8 *)zip_ops->input); __pa 108 drivers/crypto/cavium/zip/zip_inflate.c zip_cmd->s.res_ptr_addr.s.addr = __pa(result_ptr); __pa 185 drivers/crypto/cavium/zip/zip_main.c que_sbuf_addr.s.ptr = (__pa(zip->iq[q].sw_head) >> __pa 569 drivers/crypto/n2_core.c ent->src_addr = __pa(walk.data); __pa 571 drivers/crypto/n2_core.c ent->auth_iv_addr = __pa(hash_loc); __pa 575 drivers/crypto/n2_core.c ent->dest_addr = __pa(hash_loc); __pa 582 drivers/crypto/n2_core.c ent->src_addr = __pa(walk.data); __pa 656 drivers/crypto/n2_core.c __pa(&ctx->hash_key), __pa 859 drivers/crypto/n2_core.c ent->enc_key_addr = __pa(&ctx->key); __pa 1053 drivers/crypto/n2_core.c iv_paddr = __pa(rctx->walk.iv); __pa 1070 drivers/crypto/n2_core.c iv_paddr = __pa(rctx->walk.iv); __pa 1652 drivers/crypto/n2_core.c hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q), __pa 107 drivers/crypto/nx/nx-842.h return __pa(addr); __pa 94 drivers/crypto/nx/nx.c sg_addr = __pa(sg_addr); __pa 312 drivers/crypto/nx/nx.c nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb); __pa 313 drivers/crypto/nx/nx.c nx_ctx->op.in = __pa(nx_ctx->in_sg); __pa 314 drivers/crypto/nx/nx.c nx_ctx->op.out = __pa(nx_ctx->out_sg); __pa 320 drivers/crypto/nx/nx.c nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead); __pa 321 drivers/crypto/nx/nx.c nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); __pa 322 drivers/crypto/nx/nx.c nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); __pa 133 drivers/dax/super.c && pfn_t_to_pfn(pfn) == PHYS_PFN(__pa(kaddr)) __pa 134 drivers/dax/super.c && pfn_t_to_pfn(end_pfn) == PHYS_PFN(__pa(end_kaddr))) __pa 4480 drivers/dma/ppc4xx/adma.c iowrite32(((u32)__pa(ppc440spe_dma_fifo_buf)), &i2o_reg->ifbal); __pa 1025 drivers/firmware/efi/efi.c rc = efi_mem_reserve_iomem(__pa(rsv), SZ_4K); __pa 1044 drivers/firmware/efi/efi.c efi_memreserve_root->next = __pa(rsv); __pa 802 drivers/gpu/drm/amd/amdkfd/kfd_events.c pfn = __pa(page->kernel_address); __pa 1143 drivers/gpu/drm/amd/amdkfd/kfd_process.c PFN_DOWN(__pa(qpd->cwsr_kaddr)), __pa 2009 drivers/gpu/drm/i915/gvt/kvmgt.c return PFN_DOWN(__pa(addr)); __pa 35 drivers/hv/channel.c paddr = __pa(addr); __pa 90 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf)); __pa 101 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf)); __pa 126 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(&data->byte)); __pa 134 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(&data->byte)); __pa 146 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(local)); __pa 154 drivers/i2c/busses/i2c-opal.c req.buffer_ra = cpu_to_be64(__pa(&data->block[1])); __pa 459 drivers/infiniband/hw/hfi1/file_ops.c PFN_DOWN(__pa(memvirt)), __pa 588 drivers/infiniband/hw/hfi1/file_ops.c PFN_DOWN(__pa(memvirt)), __pa 653 drivers/iommu/amd_iommu_v2.c __pa(pasid_state->mm->pgd)); __pa 491 drivers/iommu/intel-pasid.c pasid_set_flptr(pte, (u64)__pa(pgd)); __pa 358 drivers/iommu/io-pgtable-arm.c new = __pa(table) | ARM_LPAE_PTE_TYPE_TABLE; __pa 573 drivers/macintosh/smu.c memblock_free(__pa(smu), sizeof(struct smu_device)); __pa 76 drivers/media/usb/cpia2/cpia2_core.c ret = __pa(kva); __pa 777 drivers/misc/ocxl/config.c rc = pnv_ocxl_set_tl_conf(dev, recv_cap, __pa(recv_rate), __pa 69 drivers/mtd/devices/powernv_flash.c rc = opal_flash_read(info->id, offset, __pa(buf), len, token); __pa 72 drivers/mtd/devices/powernv_flash.c rc = opal_flash_write(info->id, offset, __pa(buf), len, token); __pa 1144 drivers/mtd/mtdchar.c if (file->f_flags & O_DSYNC || map->phys >= __pa(high_memory)) __pa 59 drivers/net/ethernet/brocade/bna/bfa_ioc.h #define bfa_alen_set(__alen, __len, __pa) \ __pa 60 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_alen_set(__alen, __len, (u64)__pa) __pa 130 drivers/net/ethernet/ibm/ehea/ehea_phyp.c __pa(cb_addr), /* R8 */ __pa 404 drivers/net/ethernet/ibm/ehea/ehea_phyp.c __pa(cb_addr), /* R8 */ __pa 517 drivers/net/ethernet/ibm/ehea/ehea_phyp.c cb_logaddr = __pa(cb_addr); __pa 534 drivers/net/ethernet/ibm/ehea/ehea_phyp.c u64 cb_logaddr = __pa(cb_addr); __pa 556 drivers/net/ethernet/ibm/ehea/ehea_phyp.c u64 cb_logaddr = __pa(cb_addr); __pa 610 drivers/net/ethernet/ibm/ehea/ehea_phyp.c __pa(rblock), /* R6 */ __pa 146 drivers/net/ethernet/ibm/ehea/ehea_qmr.c rpage = __pa(vpage); __pa 270 drivers/net/ethernet/ibm/ehea/ehea_qmr.c rpage = __pa(vpage); __pa 375 drivers/net/ethernet/ibm/ehea/ehea_qmr.c rpage = __pa(vpage); __pa 768 drivers/net/ethernet/ibm/ehea/ehea_qmr.c index = __pa(caddr) >> SECTION_SIZE_BITS; __pa 800 drivers/net/ethernet/ibm/ehea/ehea_qmr.c u64 pt_abs = __pa(pt); __pa 808 drivers/net/ethernet/ibm/ehea/ehea_qmr.c pt[m] = __pa(pg); __pa 9952 drivers/net/ethernet/sun/niu.c *dma_addr = __pa(page); __pa 9982 drivers/net/ethernet/sun/niu.c return __pa(cpu_addr); __pa 1686 drivers/net/ethernet/toshiba/ps3_gelic_net.c ps3_mm_phys_to_lpar(__pa(&card->irq_status)), __pa 159 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c ps3_mm_phys_to_lpar(__pa(cmd->buffer)) : __pa 181 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c arg1 = ps3_mm_phys_to_lpar(__pa(cmd->buffer)); __pa 266 drivers/net/fjes/fjes_hw.c param.req_start = __pa(hw->hw_info.req_buf); __pa 268 drivers/net/fjes/fjes_hw.c param.res_start = __pa(hw->hw_info.res_buf); __pa 270 drivers/net/fjes/fjes_hw.c param.share_start = __pa(hw->hw_info.share->ep_status); __pa 479 drivers/net/fjes/fjes_main.c param.req_start = __pa(hw->hw_info.req_buf); __pa 481 drivers/net/fjes/fjes_main.c param.res_start = __pa(hw->hw_info.res_buf); __pa 482 drivers/net/fjes/fjes_main.c param.share_start = __pa(hw->hw_info.share->ep_status); __pa 615 drivers/of/fdt.c early_init_dt_reserve_memory_arch(__pa(initial_boot_params), __pa 1100 drivers/of/fdt.c #define MIN_MEMBLOCK_ADDR __pa(PAGE_OFFSET) __pa 754 drivers/platform/goldfish/goldfish_pipe.c (u64)(unsigned long)__pa(pipe->command_buffer); __pa 814 drivers/platform/goldfish/goldfish_pipe.c const unsigned long paddr = __pa(addr); __pa 1124 drivers/ps3/ps3-lpm.c ps3_mm_phys_to_lpar(__pa(lpm_priv->tb_cache)), __pa 358 drivers/ps3/ps3-vuart.c ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_written); __pa 390 drivers/ps3/ps3-vuart.c ps3_mm_phys_to_lpar(__pa(buf)), bytes, bytes_read); __pa 189 drivers/ps3/ps3stor_lib.c dev->bounce_lpar = ps3_mm_phys_to_lpar(__pa(dev->bounce_buf)); __pa 290 drivers/s390/block/dasd_eckd.c ccw->cda = (__u32)__pa(data); __pa 400 drivers/s390/block/dasd_eckd.c ccw->cda = (__u32)__pa(data); __pa 546 drivers/s390/block/dasd_eckd.c ccw->cda = (__u32) __pa(pfxdata); __pa 550 drivers/s390/block/dasd_eckd.c ccw->cda = (__u32) __pa(pfxdata); __pa 617 drivers/s390/block/dasd_eckd.c ccw->cda = (__u32) __pa(data); __pa 4138 drivers/s390/block/dasd_eckd.c if (__pa(dst) & (IDA_BLOCK_SIZE-1)) { __pa 4158 drivers/s390/block/dasd_eckd.c if (!(__pa(idaw_dst + idaw_len) & (IDA_BLOCK_SIZE-1))) __pa 94 drivers/s390/block/dasd_fba.c ccw->cda = (__u32) __pa(data); __pa 114 drivers/s390/block/dasd_fba.c ccw->cda = (__u32) __pa(data); __pa 101 drivers/s390/block/xpram.c : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); __pa 131 drivers/s390/block/xpram.c : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); __pa 162 drivers/s390/char/con3215.c ccw->cda = (__u32) __pa(raw->inbuf); __pa 222 drivers/s390/char/con3215.c (__u32) __pa(raw->buffer + ix); __pa 160 drivers/s390/char/raw3270.c rq->ccw.cda = __pa(rq->buffer); __pa 185 drivers/s390/char/raw3270.c rq->ccw.cda = __pa(rq->buffer); __pa 219 drivers/s390/char/raw3270.c rq->ccw.cda = __pa(data); __pa 229 drivers/s390/char/raw3270.c rq->ccw.cda = __pa(ib->data); __pa 547 drivers/s390/char/raw3270.c rp->init_readmod.ccw.cda = (__u32) __pa(rp->init_data); __pa 566 drivers/s390/char/raw3270.c rp->init_readpart.ccw.cda = (__u32) __pa(&rp->init_data); __pa 605 drivers/s390/char/raw3270.c rp->init_reset.ccw.cda = (__u32) __pa(rp->init_data); __pa 151 drivers/s390/cio/cio.c orb->cmd.cpa = (__u32) __pa(cpa); __pa 334 drivers/s390/cio/device_status.c sense_ccw->cda = (__u32) __pa(cdev->private->dma_area->irb.ecw); __pa 65 drivers/s390/cio/eadm_sch.c orb->eadm.aob = (u32)__pa(aob); __pa 788 drivers/s390/cio/vfio_ccw_cp.c orb->cmd.cpa = (__u32) __pa(cpa); __pa 226 drivers/s390/net/lcs.c (__u32) __pa(card->read.iob[cnt].data); __pa 239 drivers/s390/net/lcs.c (__u32) __pa(card->read.ccws); __pa 281 drivers/s390/net/lcs.c (__u32) __pa(card->write.iob[cnt].data); __pa 286 drivers/s390/net/lcs.c (__u32) __pa(card->write.ccws); __pa 496 drivers/s390/net/qeth_core_main.c ccw->cda = (__u32) __pa(data); __pa 184 drivers/scsi/bfa/bfa_ioc.h #define bfa_alen_set(__alen, __len, __pa) \ __pa 185 drivers/scsi/bfa/bfa_ioc.h __bfa_alen_set(__alen, __len, (u64)__pa) __pa 125 drivers/scsi/ps3rom.c lpar = ps3_mm_phys_to_lpar(__pa(&atapi_cmnd)); __pa 62 drivers/tty/serial/sunhv.c unsigned long ra = __pa(xmit->buf + xmit->tail); __pa 124 drivers/tty/serial/sunhv.c unsigned long ra = __pa(con_read_page); __pa 451 drivers/tty/serial/sunhv.c unsigned long ra = __pa(con_write_page); __pa 560 drivers/tty/serial/sunhv.c port->membase = (unsigned char __iomem *) __pa(port); __pa 102 drivers/usb/early/xhci-dbc.c *dma_addr = (dma_addr_t)__pa(virt); __pa 832 drivers/usb/early/xhci-dbc.c xdbc_write64(__pa(xdbc.evt_ring.dequeue), &xdbc.xdbc_reg->erdp); __pa 494 drivers/vfio/vfio_iommu_spapr_tce.c *hpa = __pa((unsigned long) page_address(page)); __pa 1094 drivers/video/fbdev/ps3fb.c xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address)); __pa 1142 drivers/video/fbdev/ps3fb.c info->fix.smem_start = __pa(fb_start); __pa 1213 drivers/video/fbdev/ps3fb.c u64 xdr_lpar = ps3_mm_phys_to_lpar(__pa(ps3fb_videomemory.address)); __pa 711 drivers/video/fbdev/ssd1307fb.c info->fix.smem_start = __pa(vmem); __pa 111 drivers/watchdog/wdrtas.c WDRTAS_SP_SPI, __pa(rtas_data_buf), __pa 163 drivers/watchdog/wdrtas.c (void *)__pa(wdrtas_logbuffer), __pa 235 drivers/xen/swiotlb-xen.c memblock_free(__pa(xen_io_tlb_start), __pa 147 fs/proc/kcore.c unsigned long pfn = __pa(ent->addr) >> PAGE_SHIFT; __pa 390 fs/proc/kcore.c phdr->p_paddr = __pa(m->addr); __pa 491 fs/proc/kcore.c } else if (!pfn_is_ram(__pa(start) >> PAGE_SHIFT)) { __pa 576 fs/proc/vmcore.c pfn = __pa(elfcorebuf + start) >> PAGE_SHIFT; __pa 910 include/asm-generic/io.h return __pa((unsigned long)address); __pa 81 include/asm-generic/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) __pa 1179 include/linux/efi.h unsigned long paddr = __pa(start + i); __pa 377 include/linux/kexec.h return phys_to_boot_phys(__pa((unsigned long)addr)); __pa 114 include/linux/mm.h #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0)) __pa 22 include/linux/psp-sev.h #define __psp_pa(x) __pa(x) __pa 153 include/xen/grant_table.h map->host_addr = __pa(addr); __pa 169 include/xen/grant_table.h unmap->host_addr = __pa(addr); __pa 375 kernel/crash_core.c return __pa(vmcoreinfo_note); __pa 208 kernel/dma/swiotlb.c io_tlb_start = __pa(tlb); __pa 395 kernel/dma/swiotlb.c memblock_free_late(__pa(io_tlb_orig_addr), __pa 397 kernel/dma/swiotlb.c memblock_free_late(__pa(io_tlb_list), __pa 191 lib/cpumask.c memblock_free_early(__pa(mask), cpumask_size()); __pa 258 mm/cma.c highmem_start = __pa(high_memory - 1) + 1; __pa 86 mm/kasan/init.c void *ptr = memblock_alloc_try_nid(size, size, __pa(MAX_DMA_ADDRESS), __pa 91 mm/kasan/init.c __func__, size, size, node, (u64)__pa(MAX_DMA_ADDRESS)); __pa 383 mm/memblock.c addr = __pa(memblock.reserved.regions); __pa 390 mm/memblock.c addr = __pa(memblock.memory.regions); __pa 448 mm/memblock.c addr = new_array ? __pa(new_array) : 0; __pa 490 mm/memblock.c memblock_free(__pa(old_array), old_alloc_size); __pa 7608 mm/page_alloc.c __pa(PAGE_OFFSET) >> PAGE_SHIFT, NULL); __pa 157 mm/page_ext.c table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS), __pa 2096 mm/percpu.c return __pa(addr); __pa 2157 mm/percpu.c memblock_free_early(__pa(ai), ai->__ai_size); __pa 2805 mm/percpu.c memblock_free_early(__pa(areas), areas_size); __pa 2927 mm/percpu.c memblock_free_early(__pa(pages), pages_size); __pa 2952 mm/percpu.c return memblock_alloc_from(size, align, __pa(MAX_DMA_ADDRESS)); __pa 2957 mm/percpu.c memblock_free_early(__pa(ptr), size); __pa 3000 mm/percpu.c fc = memblock_alloc_from(unit_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); __pa 70 mm/sparse-vmemmap.c __pa(MAX_DMA_ADDRESS)); __pa 151 mm/sparse-vmemmap.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL); __pa 371 mm/sparse.c goal = __pa(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); __pa 398 mm/sparse.c usemap_snr = pfn_to_section_nr(__pa(usage) >> PAGE_SHIFT); __pa 399 mm/sparse.c pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); __pa 456 mm/sparse.c phys_addr_t addr = __pa(MAX_DMA_ADDRESS); __pa 478 mm/sparse.c memblock_free_early(__pa(sparsemap_buf), size); __pa 483 mm/sparse.c phys_addr_t addr = __pa(MAX_DMA_ADDRESS); __pa 575 sound/parisc/harmony.c ss->runtime->dma_addr = __pa(ss->runtime->dma_area); __pa 733 virt/kvm/arm/mmu.c return __pa(kaddr);