Home
last modified time | relevance | path

Searched refs:__pa (Results 1 – 200 of 407) sorted by relevance

123

/linux-4.1.27/arch/parisc/kernel/
Dfirmware.c158 __pa(pdc_result), 0); in set_firmware_width_unlocked()
242 __pa(pdc_result), __pa(pdc_result2), len); in pdc_chassis_info()
266 retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data)); in pdc_pat_chassis_send_log()
299 retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result)); in pdc_chassis_warn()
310 ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result)); in pdc_coproc_cfg_unlocked()
357 retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, in pdc_iodc_read()
358 index, __pa(pdc_result2), iodc_data_size); in pdc_iodc_read()
384 retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), in pdc_system_map_find_mods()
385 __pa(pdc_result2), mod_index); in pdc_system_map_find_mods()
411 retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_ADDRESS, __pa(pdc_result), in pdc_system_map_find_addrs()
[all …]
/linux-4.1.27/arch/xtensa/kernel/
Dsetup.c463 initrd_is_mapped = mem_reserve(__pa(initrd_start), in setup_arch()
464 __pa(initrd_end), 0) == 0; in setup_arch()
471 mem_reserve(__pa(&_stext),__pa(&_end), 1); in setup_arch()
473 mem_reserve(__pa(&_WindowVectors_text_start), in setup_arch()
474 __pa(&_WindowVectors_text_end), 0); in setup_arch()
476 mem_reserve(__pa(&_DebugInterruptVector_literal_start), in setup_arch()
477 __pa(&_DebugInterruptVector_text_end), 0); in setup_arch()
479 mem_reserve(__pa(&_KernelExceptionVector_literal_start), in setup_arch()
480 __pa(&_KernelExceptionVector_text_end), 0); in setup_arch()
482 mem_reserve(__pa(&_UserExceptionVector_literal_start), in setup_arch()
[all …]
/linux-4.1.27/arch/ia64/hp/sim/boot/
Dfw-emu.c284 efi_systab->fw_vendor = __pa("H\0e\0w\0l\0e\0t\0t\0-\0P\0a\0c\0k\0a\0r\0d\0\0"); in sys_fw_init()
286 efi_systab->runtime = (void *) __pa(efi_runtime); in sys_fw_init()
288 efi_systab->tables = __pa(efi_tables); in sys_fw_init()
293 efi_runtime->get_time = (void *)__pa(&fw_efi_get_time); in sys_fw_init()
294 efi_runtime->set_time = (void *)__pa(&efi_unimplemented); in sys_fw_init()
295 efi_runtime->get_wakeup_time = (void *)__pa(&efi_unimplemented); in sys_fw_init()
296 efi_runtime->set_wakeup_time = (void *)__pa(&efi_unimplemented); in sys_fw_init()
297 efi_runtime->set_virtual_address_map = (void *)__pa(&efi_unimplemented); in sys_fw_init()
298 efi_runtime->get_variable = (void *)__pa(&efi_unimplemented); in sys_fw_init()
299 efi_runtime->get_next_variable = (void *)__pa(&efi_unimplemented); in sys_fw_init()
[all …]
Dbootloader.c152 req.addr = __pa(elf_phdr->p_paddr); in start_bootloader()
155 memset((char *)__pa(elf_phdr->p_paddr) + elf_phdr->p_filesz, 0, in start_bootloader()
/linux-4.1.27/arch/score/kernel/
Dsetup.c49 start_pfn = PFN_UP(__pa(&_end)); in bootmem_init()
73 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { in bootmem_init()
79 reserve_bootmem(__pa(initrd_start), size, BOOTMEM_DEFAULT); in bootmem_init()
95 code_resource.start = __pa(&_text); in resource_init()
96 code_resource.end = __pa(&_etext) - 1; in resource_init()
97 data_resource.start = __pa(&_etext); in resource_init()
98 data_resource.end = __pa(&_edata) - 1; in resource_init()
/linux-4.1.27/arch/arm/mach-omap2/
Domap-secure.c54 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher()
55 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher()
108 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher()
109 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
/linux-4.1.27/arch/x86/xen/
Dp2m.c322 paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
324 paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
327 pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
329 pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
366 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list()
376 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
386 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list()
433 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine()
471 paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); in alloc_p2m_pmd()
481 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd()
[all …]
Dmmu.c820 xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_pin()
825 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_pin()
834 xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_pin()
934 xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_unpin()
942 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_unpin()
1015 if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd)) in drop_other_mm_ref()
1035 && per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd)) in xen_drop_mm_ref()
1049 if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd)) in xen_drop_mm_ref()
1142 memblock_free(__pa(xen_start_info->mfn_list), size); in xen_pagetable_p2m_free()
1156 xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base)); in xen_pagetable_p2m_free()
[all …]
Defi.c39 boot_params.efi_info.efi_systab = (__u32)__pa(efi_systab_xen); in xen_efi_init()
40 boot_params.efi_info.efi_systab_hi = (__u32)(__pa(efi_systab_xen) >> 32); in xen_efi_init()
/linux-4.1.27/arch/parisc/mm/
Dinit.c268 bootmap_start_pfn = PAGE_ALIGN(__pa((unsigned long) &_end)) >> PAGE_SHIFT; in setup_bootmem()
328 reserve_bootmem_node(NODE_DATA(0), __pa(KERNEL_BINARY_TEXT_START), in setup_bootmem()
350 if (__pa(initrd_start) < mem_max) { in setup_bootmem()
353 if (__pa(initrd_end) > mem_max) { in setup_bootmem()
354 initrd_reserve = mem_max - __pa(initrd_start); in setup_bootmem()
359 …printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initr… in setup_bootmem()
361 reserve_bootmem_node(NODE_DATA(0), __pa(initrd_start), in setup_bootmem()
415 ro_start = __pa((unsigned long)_text); in map_pages()
416 ro_end = __pa((unsigned long)&data_start); in map_pages()
417 fv_addr = __pa((unsigned long)&fault_vector_20) & PAGE_MASK; in map_pages()
[all …]
/linux-4.1.27/arch/x86/kernel/
Dmachine_kexec_32.c111 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one()
116 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one()
132 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables()
138 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables()
222 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec()
224 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
Despfix_64.c174 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
175 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap()
184 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
185 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap()
192 pte = __pte(__pa(stack_page) | (__PAGE_KERNEL_RO & ptemask)); in init_espfix_ap()
Dsetup_percpu.c101 const unsigned long goal = __pa(MAX_DMA_ADDRESS); in pcpu_alloc_bootmem()
111 cpu, size, __pa(ptr)); in pcpu_alloc_bootmem()
116 cpu, size, node, __pa(ptr)); in pcpu_alloc_bootmem()
134 free_bootmem(__pa(ptr), size); in pcpu_fc_free()
Dmachine_kexec_64.c51 paddr = __pa(page_address(image->control_code_page)+PAGE_SIZE); in init_transition_pgtable()
58 set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE)); in init_transition_pgtable()
66 set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE)); in init_transition_pgtable()
74 set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); in init_transition_pgtable()
289 (unsigned long)__pa(page_address(image->control_code_page)); in machine_kexec()
/linux-4.1.27/arch/x86/include/asm/
Dpgalloc.h65 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel()
66 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel()
114 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
115 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate()
122 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in pgd_populate()
123 set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(pud))); in pgd_populate()
Dpage.h40 #define __pa(x) __phys_addr((unsigned long)(x)) macro
57 #define __boot_pa(x) __pa(x)
63 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/arch/um/kernel/
Dphysmem.c84 unsigned long pfn = PFN_UP(__pa(reserve_end)); in setup_physmem()
113 os_seek_file(physmem_fd, __pa(&__syscall_stub_start)); in setup_physmem()
118 free_bootmem(__pa(reserve_end) + bootmap_size, in setup_physmem()
130 else if (phys < __pa(end_iomem)) { in phys_mapping()
143 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping()
215 region->phys = __pa(region->virt); in setup_iomem()
Dmem.c50 map_memory(brk_end, __pa(brk_end), uml_reserved - brk_end, 1, 1, 0); in mem_init()
51 free_bootmem(__pa(brk_end), uml_reserved - brk_end); in mem_init()
71 (unsigned long) __pa(pte))); in one_page_table_init()
81 set_pud(pud, __pud(_KERNPG_TABLE + (unsigned long) __pa(pmd_table))); in one_md_table_init()
131 p = __pa(v); in fixaddr_user_init()
/linux-4.1.27/arch/hexagon/include/asm/
Dpage.h99 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET) macro
106 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
113 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
143 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h52 mm->context.ptbase = __pa(pgd); in pgd_alloc()
129 set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
137 set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
/linux-4.1.27/arch/arm64/mm/
Dmmu.c100 __pmd_populate(pmd, __pa(pte), PMD_TYPE_TABLE); in alloc_init_pte()
164 phys_addr_t table = __pa(pte_offset_map(&old_pmd, 0)); in alloc_init_pmd()
225 phys_addr_t table = __pa(pmd_offset(&old_pud, 0)); in alloc_init_pud()
310 unsigned long kernel_x_start = round_down(__pa(_stext), SECTION_SIZE); in __map_memblock()
311 unsigned long kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE); in __map_memblock()
400 unsigned long aligned_start = round_down(__pa(_stext), in fixup_executable()
404 __pa(_stext) - aligned_start, in fixup_executable()
409 unsigned long aligned_end = round_up(__pa(__init_end), in fixup_executable()
411 create_mapping(__pa(__init_end), (unsigned long)__init_end, in fixup_executable()
412 aligned_end - __pa(__init_end), in fixup_executable()
[all …]
Dinit.c167 memblock_reserve(__pa(_text), _end - _text); in arm64_memblock_init()
224 pg = (unsigned long)PAGE_ALIGN(__pa(start_pg)); in free_memmap()
225 pgend = (unsigned long)__pa(end_pg) & PAGE_MASK; in free_memmap()
Dmmap.c105 if (addr + size > __pa(high_memory - 1) + 1) in valid_phys_addr_range()
/linux-4.1.27/arch/ia64/hp/sim/
Dsimscsi.c134 req.addr = __pa(sg_virt(sl)); in simscsi_sg_readwrite()
140 ia64_ssc(stat.fd, 1, __pa(&req), offset, mode); in simscsi_sg_readwrite()
141 ia64_ssc(__pa(&stat), 0, 0, 0, SSC_WAIT_COMPLETION); in simscsi_sg_readwrite()
181 req.addr = __pa(&buf); in simscsi_get_disk_size()
183 ia64_ssc(fd, 1, __pa(&req), ((sectors | bit) - 1)*512, SSC_READ); in simscsi_get_disk_size()
185 ia64_ssc(__pa(&stat), 0, 0, 0, SSC_WAIT_COMPLETION); in simscsi_get_disk_size()
229 desc[target_id] = ia64_ssc(__pa(fname), SSC_READ_ACCESS|SSC_WRITE_ACCESS, in simscsi_queuecommand_lck()
Dsimeth.c126 return ia64_ssc(__pa(name), __pa(ether), 0,0, SSC_NETDEV_PROBE); in netdev_probe()
149 return ia64_ssc(fd, __pa(buf), len, 0, SSC_NETDEV_SEND); in netdev_send()
155 return ia64_ssc(fd, __pa(buf), len, 0, SSC_NETDEV_RECV); in netdev_read()
/linux-4.1.27/arch/s390/kernel/
Dsuspend.c139 unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); in pfn_is_nosave()
140 unsigned long nosave_end_pfn = PFN_DOWN(__pa(&__nosave_end)); in pfn_is_nosave()
141 unsigned long eshared_pfn = PFN_DOWN(__pa(&_eshared)) - 1; in pfn_is_nosave()
142 unsigned long stext_pfn = PFN_DOWN(__pa(&_stext)); in pfn_is_nosave()
Dearly.c113 stext_pfn = PFN_DOWN(__pa(&_stext)); in create_kernel_nss()
114 eshared_pfn = PFN_DOWN(__pa(&_eshared)); in create_kernel_nss()
115 end_pfn = PFN_UP(__pa(&_end)); in create_kernel_nss()
125 sinitrd_pfn = PFN_DOWN(__pa(INITRD_START)); in create_kernel_nss()
126 einitrd_pfn = PFN_UP(__pa(INITRD_START + INITRD_SIZE)); in create_kernel_nss()
203 end_pfn = PFN_UP(__pa(&_end)); in init_kernel_storage_key()
/linux-4.1.27/arch/parisc/include/asm/
Dpgalloc.h37 + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT)); in pgd_alloc()
61 (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT)); in pgd_populate()
113 + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); in pmd_populate_kernel()
117 + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)); in pmd_populate_kernel()
Dpage.h140 #define __pa(x) ((unsigned long)(x)-PAGE_OFFSET) macro
154 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
157 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
Dmmu_context.h56 mtctl(__pa(next->pgd), 25); in switch_mm()
/linux-4.1.27/arch/sparc/include/asm/
Dpage_64.h135 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) macro
140 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
142 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
144 #define virt_to_phys __pa
Dpage_32.h123 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + phys_base) macro
126 #define virt_to_phys __pa
130 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
Dmmu_context_64.h37 __tsb_context_switch(__pa(mm->pgd), in tsb_context_switch()
46 , __pa(&mm->context.tsb_descr[0])); in tsb_context_switch()
/linux-4.1.27/arch/x86/mm/
Dinit_32.c77 paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT); in one_md_table_init()
78 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in one_md_table_init()
100 paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); in one_page_table_init()
101 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in one_page_table_init()
182 paravirt_alloc_pte(&init_mm, __pa(newpte) >> PAGE_SHIFT); in page_table_kmap_check()
183 set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE)); in page_table_kmap_check()
187 paravirt_release_pte(__pa(pte) >> PAGE_SHIFT); in page_table_kmap_check()
484 pfn, pmd, __pa(pmd)); in native_pagetable_init()
493 pfn, pmd, __pa(pmd), pte, __pa(pte)); in native_pagetable_init()
496 paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT); in native_pagetable_init()
[all …]
Dpageattr.c653 if (pfn_range_is_mapped(PFN_DOWN(__pa(address)), in __split_large_page()
654 PFN_DOWN(__pa(address)) + 1)) in __split_large_page()
864 set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE)); in alloc_pte_page()
874 set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE)); in alloc_pmd_page()
1073 set_pgd(pgd_entry, __pgd(__pa(pud) | _KERNPG_TABLE)); in populate_pgd()
1112 cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; in __cpa_process_fault()
1484 ret = reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE, in set_memory_uc()
1496 free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE); in set_memory_uc()
1512 ret = reserve_memtype(__pa(addr[i]), __pa(addr[i]) + PAGE_SIZE, in _set_memory_array()
1535 free_memtype(__pa(addr[j]), __pa(addr[j]) + PAGE_SIZE); in _set_memory_array()
[all …]
Dpgtable.c66 paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT); in ___pmd_free_tlb()
81 paravirt_release_pud(__pa(pud) >> PAGE_SHIFT); in ___pud_free_tlb()
173 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
177 set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT)); in pud_populate()
Dinit_64.c92 set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE)); in ident_pud_init()
127 set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE)); in kernel_ident_mapping_init()
340 set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE | in __init_extra_mapping()
346 set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE | in __init_extra_mapping()
619 last_map_addr = phys_pud_init(pud, __pa(start), in kernel_physical_mapping_init()
620 __pa(end), page_size_mask); in kernel_physical_mapping_init()
625 last_map_addr = phys_pud_init(pud, __pa(start), __pa(end), in kernel_physical_mapping_init()
1264 entry = pfn_pte(__pa(p) >> PAGE_SHIFT, in vmemmap_populate_hugepages()
Dpat.c667 (pfn << PAGE_SHIFT) >= __pa(high_memory)) { in phys_mem_access_prot_allowed()
686 if (base > __pa(high_memory-1)) in kernel_map_sync_memtype()
696 id_sz = (__pa(high_memory-1) <= base + size) ? in kernel_map_sync_memtype()
697 __pa(high_memory) - base : in kernel_map_sync_memtype()
/linux-4.1.27/drivers/i2c/busses/
Di2c-opal.c104 req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf)); in i2c_opal_master_xfer()
115 req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf)); in i2c_opal_master_xfer()
142 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
150 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
162 req.buffer_ra = cpu_to_be64(__pa(local)); in i2c_opal_smbus_xfer()
170 req.buffer_ra = cpu_to_be64(__pa(&data->block[1])); in i2c_opal_smbus_xfer()
/linux-4.1.27/arch/powerpc/kernel/
Dsuspend.c20 unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave()
21 unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
Dcrash_dump.c64 create_trampoline(__pa(system_reset_fwnmi) - PHYSICAL_START); in setup_kdump_trampoline()
65 create_trampoline(__pa(machine_check_fwnmi) - PHYSICAL_START); in setup_kdump_trampoline()
Dmachine_kexec_64.c51 if (image->segment[i].mem < __pa(_end)) in default_machine_kexec_prepare()
65 low = __pa(htab_address); in default_machine_kexec_prepare()
405 htab_base = cpu_to_be64(__pa(htab_address)); in export_htab_values()
Dpaca.c85 memblock_free(__pa(extra_lppacas) + new_size, lppaca_size - new_size); in free_lppacas()
242 memblock_free(__pa(paca) + new_size, paca_size - new_size); in free_unused_pacas()
Dmachine_kexec.c169 if (overlaps_crashkernel(__pa(_stext), _end - _stext)) { in reserve_crashkernel()
273 kernel_end = cpu_to_be_ulong(__pa(_end)); in kexec_setup()
Drtas.c109 enter_rtas(__pa(args)); in call_rtas_display_status()
386 err_args.args[0] = cpu_to_be32(__pa(rtas_err_buf)); in __fetch_rtas_last_error()
393 enter_rtas(__pa(&rtas.args)); in __fetch_rtas_last_error()
448 enter_rtas(__pa(rtas_args)); in rtas_call()
719 __pa(rtas_os_term_buf)); in rtas_os_term()
1095 enter_rtas(__pa(&rtas.args)); in ppc_rtas()
Dproc_powerpc.c50 __pa(PDE_DATA(file_inode(file))) >> PAGE_SHIFT, in page_map_mmap()
Drtas_flash.c460 (u32) __pa(rtas_data_buf), args_buf->buf_size); in validate_flash()
600 rtas_block_list = __pa(flist); in rtas_flash_firmware()
614 f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data)); in rtas_flash_firmware()
621 f->next = (struct flash_block_list *)cpu_to_be64(__pa(f->next)); in rtas_flash_firmware()
Dprom.c120 start = __pa(initial_boot_params); in move_device_tree()
618 memblock_reserve(_ALIGN_DOWN(__pa(initrd_start), PAGE_SIZE), in early_reserve_mem()
690 memblock_reserve(PHYSICAL_START, __pa(klimit) - PHYSICAL_START); in early_init_devtree()
/linux-4.1.27/arch/ia64/mm/
Dinit.c447 map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); in create_mem_map_page_table()
448 map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); in create_mem_map_page_table()
452 node = paddr_to_nid(__pa(start)); in create_mem_map_page_table()
469 set_pte(pte, pfn_pte(__pa(alloc_bootmem_pages_node(NODE_DATA(node), PAGE_SIZE)) >> PAGE_SHIFT, in create_mem_map_page_table()
489 map_start = vmem_map + (__pa(start) >> PAGE_SHIFT); in virtual_memmap_init()
490 map_end = vmem_map + (__pa(end) >> PAGE_SHIFT); in virtual_memmap_init()
573 memblock_add_node(__pa(start), end - start, nid); in register_active_ranges()
582 pfn_start = (PAGE_ALIGN(__pa(start))) >> PAGE_SHIFT; in find_max_min_low_pfn()
583 pfn_end = (PAGE_ALIGN(__pa(end - 1))) >> PAGE_SHIFT; in find_max_min_low_pfn()
585 pfn_start = GRANULEROUNDDOWN(__pa(start)) >> PAGE_SHIFT; in find_max_min_low_pfn()
[all …]
Dcontig.c73 bootmap_start = __pa(range_start); in find_bootmap_location()
123 ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) - in per_cpu_init()
136 PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); in alloc_per_cpu_data()
Ddiscontig.c404 base = __pa(bdp->node_bootmem_map); in reserve_pernode_space()
409 base = __pa(mem_data[node].pernode_addr); in reserve_pernode_space()
495 PERCPU_PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); in memory_less_node_alloc()
513 fill_pernode(node, __pa(pernode), pernodesize); in memory_less_nodes()
672 if (start <= __pa(MAX_DMA_ADDRESS)) in count_node_pages()
674 (min(end, __pa(MAX_DMA_ADDRESS)) - start) >>PAGE_SHIFT; in count_node_pages()
/linux-4.1.27/arch/powerpc/include/asm/
Dpage.h130 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
132 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
209 #define __pa(x) ((unsigned long)(x) - VIRT_PHYS_OFFSET) macro
217 #define __pa(x) ((unsigned long)(x) & 0x0fffffffffffffffUL) macro
221 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START) macro
Dpgtable-ppc32.h257 unsigned long ptephys = __pa(ptep) & PAGE_MASK; in __ptep_test_and_clear_young()
314 pfn_to_page((__pa(pmd_val(pmd)) >> PAGE_SHIFT))
Dpgalloc-32.h25 (pmd_val(*(pmd)) = __pa(pte) | _PMD_PRESENT)
/linux-4.1.27/arch/powerpc/platforms/powernv/
Dpci-p5ioc2.c116 pr_devel(" TCE AT : 0x%016lx\n", __pa(tce_mem)); in pnv_pci_init_p5ioc2_phb()
119 rc = opal_pci_set_phb_tce_memory(phb_id, __pa(tce_mem), tce_size); in pnv_pci_init_p5ioc2_phb()
219 __pa(tce_mem), __pa(tce_mem) + P5IOC2_TCE_MEMORY - 1); in pnv_pci_init_p5ioc2_hub()
220 rc = opal_pci_set_hub_tce_memory(hub_id, __pa(tce_mem), in pnv_pci_init_p5ioc2_hub()
Dopal-nvram.c39 rc = opal_read_nvram(__pa(buf), count, off); in opal_nvram_read()
58 rc = opal_write_nvram(__pa(buf), count, off); in opal_nvram_write()
Dopal-xscom.c105 rc = opal_xscom_read(m->chip, reg, (__be64 *)__pa(&v)); in opal_scom_read()
Dopal-elog.c169 opal_rc = opal_read_elog(__pa(elog->buffer), in raw_attr_read()
212 rc = opal_read_elog(__pa(elog->buffer), in create_elog_obj()
/linux-4.1.27/arch/m32r/include/asm/
Dpage.h69 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) macro
78 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
79 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h9 set_pmd(pmd, __pmd(_PAGE_TABLE + __pa(pte)))
/linux-4.1.27/arch/arc/include/asm/
Dpage.h91 #define __pa(vaddr) ((unsigned long)vaddr) macro
95 (mem_map + ((__pa(kaddr) - CONFIG_LINUX_LINK_BASE) >> PAGE_SHIFT))
97 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/arch/cris/include/arch-v10/arch/
Dpage.h21 #define __pa(x) ((unsigned long)(x) & 0xdfffffff) macro
25 #define __pa(x) ((unsigned long)(x) & 0x7fffffff) macro
/linux-4.1.27/arch/sh/kernel/
Dswsusp.c24 unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave()
25 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
/linux-4.1.27/arch/avr32/include/asm/
Dpage.h76 #define __pa(x) PHYSADDR(x) macro
93 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
94 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/arch/unicore32/kernel/
Dhibernate.c58 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_KERNEL_TABLE)); in resume_one_page_table_init()
147 unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave()
148 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
/linux-4.1.27/arch/m68k/include/asm/
Dvirtconvert.h20 return __pa(address); in virt_to_phys()
31 __pa(PAGE_OFFSET + (((page) - pg_data_map[0].node_mem_map) << PAGE_SHIFT))
Dpage_mm.h83 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) macro
97 #define __pa(x) ___pa((unsigned long)(x)) macro
126 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
Dpage_no.h22 #define __pa(vaddr) ((unsigned long)(vaddr)) macro
25 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
Dsun3_pgalloc.h68 pmd_val(*pmd) = __pa((unsigned long)pte); in pmd_populate_kernel()
73 pmd_val(*pmd) = __pa((unsigned long)page_address(page)); in pmd_populate()
/linux-4.1.27/arch/alpha/include/asm/
Dpage.h83 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) macro
86 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
89 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dmmzone.h58 #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr))
72 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
108 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
/linux-4.1.27/arch/frv/include/asm/
Dpage.h45 #define __pa(vaddr) virt_to_phys((void *) (unsigned long) (vaddr)) macro
62 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
63 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h23 #define pmd_populate_kernel(mm, pmd, pte) __set_pmd(pmd, __pa(pte) | _PAGE_TABLE)
/linux-4.1.27/arch/x86/lguest/
Dboot.c370 lazy_hcall2(LHCALL_LOAD_TLS, __pa(&t->tls_array), cpu); in lguest_load_tls()
676 lazy_hcall4(LHCALL_SET_PTE, __pa(mm->pgd), addr, in lguest_pte_update()
679 lazy_hcall3(LHCALL_SET_PTE, __pa(mm->pgd), addr, ptep->pte_low); in lguest_pte_update()
704 lazy_hcall2(LHCALL_SET_PGD, __pa(pudp) & 0xFFFFFFE0, in lguest_set_pud()
705 (__pa(pudp) & 0x1F) / sizeof(pud_t)); in lguest_set_pud()
711 lazy_hcall2(LHCALL_SET_PMD, __pa(pmdp) & PAGE_MASK, in lguest_set_pmd()
712 (__pa(pmdp) & (PAGE_SIZE - 1)) / sizeof(pmd_t)); in lguest_set_pmd()
720 lazy_hcall2(LHCALL_SET_PGD, __pa(pmdp) & PAGE_MASK, in lguest_set_pmd()
721 (__pa(pmdp) & (PAGE_SIZE - 1)) / sizeof(pmd_t)); in lguest_set_pmd()
1173 hcall(LHCALL_SHUTDOWN, __pa("Power down"), in lguest_power_off()
[all …]
/linux-4.1.27/arch/mn10300/include/asm/
Dpage.h105 #define __pa(x) ((unsigned long)(x)) macro
117 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
118 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h23 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE))
/linux-4.1.27/arch/arm64/include/asm/
Dpgalloc.h46 set_pud(pud, __pud(__pa(pmd) | PMD_TYPE_TABLE)); in pud_populate()
66 set_pgd(pgd, __pgd(__pa(pud) | PUD_TYPE_TABLE)); in pgd_populate()
126 __pmd_populate(pmdp, __pa(ptep), PMD_TYPE_TABLE); in pmd_populate_kernel()
Dmemory.h154 #define __pa(x) __virt_to_phys((unsigned long)(x)) macro
165 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
166 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dkvm_mmu.h292 merged_hyp_pgd[0] = __pgd(__pa(hyp_pgd) | PMD_TYPE_TABLE);
302 merged_hyp_pgd[idmap_idx] = __pgd(__pa(boot_hyp_pgd) | PMD_TYPE_TABLE);
/linux-4.1.27/arch/mips/power/
Dcpu.c39 unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); in pfn_is_nosave()
40 unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end)); in pfn_is_nosave()
/linux-4.1.27/arch/ia64/include/asm/
Dpgalloc.h39 pgd_val(*pgd_entry) = __pa(pud); in pgd_populate()
57 pud_val(*pud_entry) = __pa(pmd); in pud_populate()
82 pmd_val(*pmd_entry) = __pa(pte); in pmd_populate_kernel()
Dpage.h59 # define __pa(x) ((x) - PAGE_OFFSET) macro
95 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
124 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
142 #define __pa(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = 0; _v.l;}) macro
/linux-4.1.27/arch/x86/power/
Dhibernate_32.c43 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init()
66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init()
141 __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); in resume_init_first_level_page_table()
/linux-4.1.27/arch/arm/mach-zynq/
Dcommon.c61 if (!__pa(PAGE_OFFSET)) in zynq_memory_init()
62 memblock_reserve(__pa(PAGE_OFFSET), __pa(swapper_pg_dir)); in zynq_memory_init()
Dplatsmp.c52 if (__pa(PAGE_OFFSET)) { in zynq_cpun_start()
76 if (__pa(PAGE_OFFSET)) in zynq_cpun_start()
/linux-4.1.27/arch/metag/include/asm/
Dpage.h103 #define __pa(x) ((unsigned long)(((unsigned long)(x)) - meta_memoffset)) macro
108 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
110 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h8 set_pmd(pmd, __pmd(_PAGE_TABLE | __pa(pte)))
Dmmu_context.h68 metag_out32(__pa(pgd) & MMCU_TBLPHYS1_ADDR_BITS, phys1); in load_pgd()
/linux-4.1.27/mm/
Dsparse-vmemmap.c69 __pa(MAX_DMA_ADDRESS)); in vmemmap_alloc_block()
109 entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL); in vmemmap_pte_populate()
206 PMD_SIZE, __pa(MAX_DMA_ADDRESS)); in sparse_mem_maps_populate_node()
230 memblock_free_early(__pa(vmemmap_buf), in sparse_mem_maps_populate_node()
Dsparse.c280 goal = __pa(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section()
302 usemap_snr = pfn_to_section_nr(__pa(usemap) >> PAGE_SHIFT); in check_usemap_section_nr()
303 pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr()
383 PAGE_SIZE, __pa(MAX_DMA_ADDRESS), in sparse_mem_map_populate()
409 PAGE_SIZE, __pa(MAX_DMA_ADDRESS), in sparse_mem_maps_populate_node()
591 memblock_free_early(__pa(map_map), size2); in sparse_init()
593 memblock_free_early(__pa(usemap_map), size); in sparse_init()
/linux-4.1.27/arch/sparc/mm/
Dtsb.c82 base = __pa(base); in flush_tsb_user()
90 base = __pa(base); in flush_tsb_user()
106 base = __pa(base); in flush_tsb_user_page()
114 base = __pa(base); in flush_tsb_user_page()
151 tsb_paddr = __pa(mm->context.tsb_block[tsb_idx].tsb); in setup_tsb_params()
440 old_tsb_base = __pa(old_tsb_base); in tsb_grow()
441 new_tsb_base = __pa(new_tsb_base); in tsb_grow()
Dio-unit.c132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area()
231 sbus_writel(MKIOPTE(__pa(page)), iopte); in iounit_map_dma_area()
/linux-4.1.27/arch/s390/include/asm/
Dpage.h144 #define __pa(x) (unsigned long)(x) macro
146 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
148 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h93 pgd_val(*pgd) = _REGION2_ENTRY | __pa(pud); in pgd_populate()
98 pud_val(*pud) = _REGION3_ENTRY | __pa(pmd); in pud_populate()
127 pmd_val(*pmd) = _SEGMENT_ENTRY + __pa(pte); in pmd_populate()
Didals.h31 return ((__pa(vaddr) + length - 1) >> 31) != 0; in idal_is_needed()
40 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + in idal_nr_words()
53 paddr = __pa(vaddr); in idal_create_words()
Dmmu_context.h45 S390_lowcore.user_asce = mm->context.asce_bits | __pa(mm->pgd); in set_user_asce()
74 S390_lowcore.user_asce = next->context.asce_bits | __pa(next->pgd); in switch_mm()
/linux-4.1.27/arch/m68k/sun3/
Dconfig.c120 start_page = __pa(memory_start) >> PAGE_SHIFT; in sun3_bootmem_alloc()
121 num_pages = __pa(memory_end) >> PAGE_SHIFT; in sun3_bootmem_alloc()
130 free_bootmem(__pa(availmem), memory_end - (availmem)); in sun3_bootmem_alloc()
/linux-4.1.27/arch/xtensa/include/asm/
Dpage.h172 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) macro
181 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
183 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/arch/openrisc/kernel/
Dtraps.c203 ((struct task_struct *)(__pa(current)))->comm, in nommu_dump_state()
204 ((struct task_struct *)(__pa(current)))->pid, in nommu_dump_state()
215 *((unsigned long *)(__pa(stack)))); in nommu_dump_state()
222 addr = *((unsigned long *)__pa(stack)); in nommu_dump_state()
238 c = ((unsigned char *)(__pa(regs->pc)))[i]; in nommu_dump_state()
Ddma.c46 for (cl = __pa(addr); cl < __pa(next); cl += cpuinfo.dcache_block_size) in page_set_nocache()
100 *dma_handle = __pa(page); in or1k_dma_alloc()
/linux-4.1.27/arch/mips/dec/prom/
Dmemory.c112 end = __pa(&_text) - 0x00020000; in prom_free_prom_memory()
115 end = __pa(&_text); in prom_free_prom_memory()
/linux-4.1.27/arch/mips/include/asm/
Dpage.h166 #define __pa(x) \ macro
172 #define __pa(x) \ macro
192 #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0))
/linux-4.1.27/arch/arm/include/asm/
Dcacheflush.h406 outer_clean_range(__pa(_p), __pa(_p + size)); in __sync_cache_range_w()
428 outer_flush_range(__pa(_p), __pa(_p + size)); in __sync_cache_range_r()
Dpgalloc.h43 set_pud(pud, __pud(__pa(pmd) | PMD_TYPE_TABLE)); in pud_populate()
154 __pmd_populate(pmdp, __pa(ptep), _PAGE_KERNEL_TABLE); in pmd_populate_kernel()
Dmemory.h171 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
292 #define __pa(x) __virt_to_phys((unsigned long)(x)) macro
/linux-4.1.27/arch/m68k/mm/
Dsun3mmu.c67 pg_table = (pte_t *) __pa (next_pgtable); in paging_init()
93 (__pa(PAGE_OFFSET) >> PAGE_SHIFT) + 1, NULL); in paging_init()
/linux-4.1.27/arch/powerpc/platforms/chrp/
Dnvram.c39 __pa(nvram_buf), 1) != 0) || 1 != done) in chrp_nvram_read()
61 __pa(nvram_buf), 1) != 0) || 1 != done) in chrp_nvram_write()
/linux-4.1.27/arch/unicore32/include/asm/
Dmemory.h94 #define __pa(x) __virt_to_phys((unsigned long)(x)) macro
109 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
Dtlbflush.h159 : : "r" (__pa(pmd) & ~(L1_CACHE_BYTES - 1)) : "cc"); in clean_pmd_entry()
Dpgalloc.h101 __pmd_populate(pmdp, __pa(pte_ptr) | _PAGE_KERNEL_TABLE); in pmd_populate_kernel()
/linux-4.1.27/drivers/char/hw_random/
Dn2-drv.c148 unsigned long ra = __pa(&np->scratch_control[0]); in n2rng_control_settle_v2()
390 unsigned long ra = __pa(&np->test_data); in n2rng_data_read()
424 unsigned long ra = __pa(&np->test_data); in n2rng_guest_check()
434 unsigned long post_ctl_ra = __pa(post_control); in n2rng_entropy_diag_read()
435 unsigned long pre_ctl_ra = __pa(pre_control); in n2rng_entropy_diag_read()
436 unsigned long buffer_ra = __pa(buffer); in n2rng_entropy_diag_read()
558 unsigned long ctl_ra = __pa(&up->control[0]); in n2rng_control_configure_units()
/linux-4.1.27/arch/sh/include/asm/
Dpage.h149 #define __pa(x) ___pa((unsigned long)x) macro
177 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
181 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/arch/microblaze/include/asm/
Dpage.h147 # define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr))))
152 # define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
177 # define __pa(x) __virt_to_phys((unsigned long)(x)) macro
/linux-4.1.27/arch/m32r/kernel/
Dsetup.c140 start_pfn = PFN_UP( __pa(_end) ); in setup_memory()
141 max_low_pfn = PFN_DOWN( __pa(memory_end) ); in setup_memory()
160 curr_pfn = PFN_UP(__pa(memory_start)); in setup_memory()
165 last_pfn = PFN_DOWN(__pa(memory_end)); in setup_memory()
/linux-4.1.27/arch/s390/mm/
Dvmem.c100 pud_val(*pu_dir) = __pa(address) | in vmem_add_mem()
117 pmd_val(*pm_dir) = __pa(address) | in vmem_add_mem()
133 pte_val(*pt_dir) = __pa(address) | in vmem_add_mem()
233 pmd_val(*pm_dir) = __pa(new_page) | in vmemmap_populate()
255 __pa(new_page) | pgprot_val(PAGE_KERNEL); in vmemmap_populate()
Dpgtable.c188 _ASCE_USER_BITS | __pa(table); in gmap_alloc()
826 page = pfn_to_page(__pa(table) >> PAGE_SHIFT); in page_table_free_pgste()
1043 page = pfn_to_page(__pa(table) >> PAGE_SHIFT); in page_table_free()
1047 bit = 1 << ((__pa(table) & ~PAGE_MASK)/(PTRS_PER_PTE*sizeof(pte_t))); in page_table_free()
1069 page = pfn_to_page(__pa(table) >> PAGE_SHIFT); in __page_table_free_rcu()
1085 page = pfn_to_page(__pa(table) >> PAGE_SHIFT); in page_table_free_rcu()
1088 table = (unsigned long *) (__pa(table) | FRAG_MASK); in page_table_free_rcu()
1092 bit = 1 << ((__pa(table) & ~PAGE_MASK) / (PTRS_PER_PTE*sizeof(pte_t))); in page_table_free_rcu()
1100 table = (unsigned long *) (__pa(table) | (bit << 4)); in page_table_free_rcu()
/linux-4.1.27/arch/openrisc/include/asm/
Dpage.h80 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) macro
82 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
Dpgalloc.h31 set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
/linux-4.1.27/arch/arc/mm/
Dinit.c102 __pa(_end) - CONFIG_LINUX_LINK_BASE); in setup_arch_memory()
107 memblock_reserve(__pa(initrd_start), initrd_end - initrd_start); in setup_arch_memory()
/linux-4.1.27/arch/sparc/kernel/
Dirq_64.c282 irq = bucket_get_irq(__pa(bucket)); in sysino_exists()
618 irq = bucket_get_irq(__pa(bucket)); in build_irq()
621 bucket_set_irq(__pa(bucket), irq); in build_irq()
683 cookie = ~__pa(&ihd->bucket); in cookie_assign()
739 bucket_set_irq(__pa(bucket), irq); in sysino_set_bucket()
1025 *pa_ptr = __pa(p); in alloc_one_queue()
1041 tb->cpu_mondo_block_pa = __pa(page); in init_cpu_send_mondo_info()
1042 tb->cpu_list_pa = __pa(page + 64); in init_cpu_send_mondo_info()
1102 ivector_table_pa = __pa(ivector_table); in irq_ivector_init()
Dsmp_64.c329 __pa(hdesc)); in ldom_startcpu_cpuid()
847 if (tp->pgd_paddr == __pa(mm->pgd)) in tsb_sync()
882 __flush_icache_page(__pa(page_address(page))); in __local_flush_dcache_page()
915 xcall_deliver(data0, __pa(pg_addr), in smp_flush_dcache_page_impl()
951 xcall_deliver(data0, __pa(pg_addr), in flush_dcache_page_all()
1466 const unsigned long goal = __pa(MAX_DMA_ADDRESS); in pcpu_alloc_bootmem()
1476 cpu, size, __pa(ptr)); in pcpu_alloc_bootmem()
1481 "%016lx\n", cpu, size, node, __pa(ptr)); in pcpu_alloc_bootmem()
1491 free_bootmem(__pa(ptr), size); in pcpu_free_bootmem()
Diommu_common.h36 #define SG_ENT_PHYS_ADDRESS(SG) (__pa(sg_virt((SG))))
/linux-4.1.27/arch/mn10300/kernel/
Dsetup.c126 kstart_pfn = PFN_UP(__pa(&_text)); in setup_arch()
127 free_pfn = PFN_UP(__pa(&_end)); in setup_arch()
128 end_pfn = PFN_DOWN(__pa(memory_end)); in setup_arch()
/linux-4.1.27/include/asm-generic/
Dpage.h81 #define __pa(x) ((unsigned long) (x)) macro
83 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
/linux-4.1.27/drivers/crypto/nx/
Dnx.c107 sg_addr = __pa(sg_addr); in nx_build_sg_list()
325 nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb); in nx_ctx_init()
326 nx_ctx->op.in = __pa(nx_ctx->in_sg); in nx_ctx_init()
327 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
333 nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead); in nx_ctx_init()
334 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); in nx_ctx_init()
335 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
/linux-4.1.27/arch/x86/realmode/
Dinit.c47 phys_base = __pa(base); in setup_real_mode()
107 __pa(base); in set_real_mode_permissions()
/linux-4.1.27/arch/score/include/asm/
Dpage.h61 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) macro
67 #define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr))))
/linux-4.1.27/arch/ia64/sn/pci/
Dpci_dma.c111 phys_addr = __pa(cpuaddr); in sn_dma_alloc_coherent()
192 phys_addr = __pa(cpu_addr); in sn_dma_map_page()
383 port, size, __pa(val)); in sn_pci_legacy_read()
429 port, size, __pa(&val)); in sn_pci_legacy_write()
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
Dehea_phyp.c144 __pa(cb_addr), /* R8 */ in ehea_h_query_ehea_qp()
418 __pa(cb_addr), /* R8 */ in ehea_h_modify_ehea_qp()
531 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea()
548 u64 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea_port()
570 u64 cb_logaddr = __pa(cb_addr); in ehea_h_modify_ehea_port()
624 __pa(rblock), /* R6 */ in ehea_h_error_data()
Dehea_qmr.c165 rpage = __pa(vpage); in ehea_create_cq()
290 rpage = __pa(vpage); in ehea_create_eq()
395 rpage = __pa(vpage); in ehea_qp_alloc_register()
788 index = __pa(caddr) >> SECTION_SIZE_BITS; in ehea_map_vaddr()
820 u64 pt_abs = __pa(pt); in ehea_reg_mr_section()
828 pt[m] = __pa(pg); in ehea_reg_mr_section()
/linux-4.1.27/arch/avr32/kernel/
Dsetup.c83 kernel_code.start = __pa(init_mm.start_code); in resource_init()
567 kernel_code.start = __pa(__init_begin); in setup_arch()
568 kernel_code.end = __pa(init_mm.end_code - 1); in setup_arch()
569 kernel_data.start = __pa(init_mm.end_code); in setup_arch()
570 kernel_data.end = __pa(init_mm.brk - 1); in setup_arch()
/linux-4.1.27/arch/metag/mm/
Dmmu-meta2.c180 phys_addr = __pa(addr); in mmu_init()
182 second_level_phys = __pa(pte); in mmu_init()
Dinit.c58 set_pmd(pmd, __pmd(_PAGE_TABLE | __pa(pte))); in insert_gateway_page()
62 set_pte(pte, pfn_pte(__pa(gateway_page) >> PAGE_SHIFT, PAGE_READONLY)); in insert_gateway_page()
239 start_pfn = PFN_UP(__pa(heap_start)); in init_and_reserve_mem()
/linux-4.1.27/arch/sh/drivers/pci/
Dpci-sh5.c104 unsigned long memStart = __pa(memory_start); in sh5pci_init()
105 unsigned long memSize = __pa(memory_end) - memStart; in sh5pci_init()
/linux-4.1.27/arch/m68k/sun3x/
Ddvma.c131 __pa(kaddr), vaddr); in dvma_map_cpu()
167 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu()
/linux-4.1.27/arch/um/include/asm/
Dpage.h109 #define __pa(virt) to_phys((void *) (unsigned long) (virt)) macro
116 #define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
Dpgalloc.h14 set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
/linux-4.1.27/arch/x86/kvm/
Dmmu_audit.c103 sp = page_header(__pa(sptep)); in audit_mappings()
136 rev_sp = page_header(__pa(sptep)); in inspect_spte_has_rmap()
167 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in audit_spte_after_sync()
/linux-4.1.27/arch/ia64/sn/kernel/
Dio_init.c104 status = sal_get_hubdev_info(nasid, (u64) __pa(hubdev)); in sn_fixup_ionodes()
170 (u64) __pa(pcidev_info), in sn_io_slot_fixup()
171 (u64) __pa(sn_irq_info)); in sn_io_slot_fixup()
/linux-4.1.27/arch/powerpc/sysdev/
Ddart_iommu.c181 rpn = __pa(uaddr) >> DART_PAGE_SHIFT; in dart_build()
258 dart_vbase = ioremap(__pa(dart_tablebase), dart_tablesize); in dart_init()
436 tbasepfn = __pa(dart_tablebase) >> PAGE_SHIFT; in iommu_init_late_dart()
/linux-4.1.27/arch/cris/include/arch-v32/arch/
Dpage.h14 #define __pa(x) ((unsigned long)(x) & 0x7fffffff) macro
/linux-4.1.27/include/xen/
Dgrant_table.h142 map->host_addr = __pa(addr); in gnttab_set_map_op()
158 unmap->host_addr = __pa(addr); in gnttab_set_unmap_op()
/linux-4.1.27/arch/mips/kernel/
Dsetup.c236 end = __pa(initrd_end); in init_initrd()
238 initrd_start = (unsigned long)__va(__pa(initrd_start)); in init_initrd()
256 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { in finalize_initrd()
261 reserve_bootmem(__pa(initrd_start), size, BOOTMEM_DEFAULT); in finalize_initrd()
374 mapstart = max(mapstart, (unsigned long)PFN_UP(__pa(initrd_end))); in bootmem_init()
Dprom.c48 return __alloc_bootmem(size, align, __pa(MAX_DMA_ADDRESS)); in early_init_dt_alloc_memory_arch()
/linux-4.1.27/arch/arm/mach-highbank/
Dpm.c31 return psci_ops.cpu_suspend(ps, __pa(cpu_resume)); in highbank_suspend_finish()
/linux-4.1.27/arch/powerpc/platforms/pasemi/
Diommu.c104 rpn = __pa(uaddr) >> IOBMAP_PAGE_SHIFT; in iobmap_build()
216 regword = IOBMAP_L1E_V | (__pa(iob_l2_base + i*0x2000) >> 12); in iob_init()
/linux-4.1.27/arch/powerpc/platforms/pseries/
Dnvram.c71 if ((rtas_call(nvram_fetch, 3, 2, &done, i, __pa(nvram_buf), in pSeries_nvram_read()
116 if ((rtas_call(nvram_store, 3, 2, &done, i, __pa(nvram_buf), in pSeries_nvram_write()
Dsetup.c106 system_reset_addr = __pa(system_reset_fwnmi) - PHYSICAL_START; in fwnmi_init()
107 machine_check_addr = __pa(machine_check_fwnmi) - PHYSICAL_START; in fwnmi_init()
331 ret = register_dtl(hard_smp_processor_id(), __pa(dtl)); in alloc_dispatch_logs()
627 __pa(rtas_data_buf), in pSeries_cmo_feature_init()
Dras.c206 critical, __pa(&ras_log_buf), in ras_epow_interrupt()
237 __pa(&ras_log_buf), in ras_error_interrupt()
Dio_event_irq.c131 __pa(ioei_rtas_buf), in ioei_interrupt()
Dlpar.c84 addr = __pa(&lppaca_of(cpu)); in vpa_init()
96 addr = __pa(paca[cpu].slb_shadow_ptr); in vpa_init()
117 ret = register_dtl(hwcpu, __pa(dtl)); in vpa_init()
639 addr = __pa((unsigned long)page_address(page)); in pSeries_set_page_state()
Dhotplug-cpu.c105 enter_rtas(__pa(&args)); in rtas_stop_self()
377 __pa(cede_parameters), in parse_cede_parameters()
Diommu.c60 start = __pa(startp); in tce_invalidate_pSeries_sw()
61 end = __pa(endp); in tce_invalidate_pSeries_sw()
101 rpn = __pa(uaddr) >> TCE_SHIFT; in tce_build_pSeries()
150 rpn = __pa(uaddr) >> TCE_SHIFT; in tce_build_pSeriesLP()
219 rpn = __pa(uaddr) >> TCE_SHIFT; in tce_buildmulti_pSeriesLP()
239 (u64)__pa(tcep), in tce_buildmulti_pSeriesLP()
445 (u64)__pa(tcep), in tce_setrange_multi_pSeriesLP()
Dcmm.c168 if ((rc = plpar_page_set_loaned(__pa(addr)))) { in cmm_alloc_pages()
212 plpar_page_set_active(__pa(addr)); in cmm_free_pages()
546 plpar_page_set_active(__pa(pa_curr->page[idx])); in cmm_mem_going_offline()
/linux-4.1.27/drivers/tty/serial/
Dsunhv.c62 unsigned long ra = __pa(xmit->buf + xmit->tail); in transmit_chars_write()
124 unsigned long ra = __pa(con_read_page); in receive_chars_read()
439 unsigned long ra = __pa(con_write_page); in sunhv_console_write_paged()
554 port->membase = (unsigned char __iomem *) __pa(port); in hv_probe()
/linux-4.1.27/arch/unicore32/mm/
Dinit.c253 memblock_reserve(__pa(_text), _end - _text); in uc32_memblock_init()
332 pg = PAGE_ALIGN(__pa(start_pg)); in free_memmap()
333 pgend = __pa(end_pg) & PAGE_MASK; in free_memmap()
Dmmu.c157 __pmd_populate(pmd, __pa(pte) | prot); in early_pte_alloc()
289 lowmem_limit = __pa(vmalloc_min - 1) + 1; in sanity_check_meminfo()
339 memblock_reserve(__pa(swapper_pg_dir), PTRS_PER_PGD * sizeof(pgd_t)); in uc32_mm_memblock_reserve()
/linux-4.1.27/arch/arm/mach-shmobile/
Dsmp-emev2.c43 iowrite32(__pa(shmobile_boot_vector), smu + SMU_GENERAL_REG0); in emev2_smp_prepare_cpus()
Dsmp-sh73a0.c60 __raw_writel(__pa(shmobile_boot_vector), SBAR); in sh73a0_smp_prepare_cpus()
Dsmp-r8a7779.c103 __raw_writel(__pa(shmobile_boot_vector), AVECR); in r8a7779_smp_prepare_cpus()
/linux-4.1.27/arch/arm/mach-keystone/
Dplatsmp.c46 cpu_set_ttbr(1, __pa(pgd0) + TTBR1_OFFSET); in keystone_smp_secondary_initmem()
/linux-4.1.27/arch/metag/kernel/
Dsetup.c360 min_low_pfn = PFN_UP(__pa(text_start)); in setup_arch()
361 max_low_pfn = PFN_DOWN(__pa(heap_end)); in setup_arch()
371 start_pfn = PFN_UP(__pa(heap_start)); in setup_arch()
/linux-4.1.27/drivers/cpuidle/
Dcpuidle-calxeda.c37 return psci_ops.cpu_suspend(ps, __pa(cpu_resume)); in calxeda_idle_finish()
/linux-4.1.27/arch/arm/mm/
Dinit.c233 memblock_reserve(__pa(_sdata), _end - _sdata); in arm_memblock_init()
235 memblock_reserve(__pa(_stext), _end - _stext); in arm_memblock_init()
347 pg = PAGE_ALIGN(__pa(start_pg)); in free_memmap()
348 pgend = __pa(end_pg) & PAGE_MASK; in free_memmap()
Dmmu.c689 __pmd_populate(pmd, __pa(pte), prot); in early_pte_alloc()
1073 phys_addr_t vmalloc_limit = __pa(vmalloc_min - 1) + 1; in sanity_check_meminfo()
1203 memblock_reserve(__pa(swapper_pg_dir), SWAPPER_PG_DIR_SIZE); in arm_mm_memblock_reserve()
1210 memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET); in arm_mm_memblock_reserve()
1332 phys_addr_t kernel_x_start = round_down(__pa(_stext), SECTION_SIZE); in map_lowmem()
1333 phys_addr_t kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE); in map_lowmem()
1464 __pud(__pa(pmd0) | PMD_TYPE_TABLE | L_PGD_SWAPPER)); in early_paging_init()
1472 phys = __pa(map_start); in early_paging_init()
1493 cpu_set_ttbr(1, __pa(pgd0) + TTBR1_OFFSET); in early_paging_init()
/linux-4.1.27/drivers/staging/goldfish/
DREADME9 - Use dma coherent memory not kmalloc/__pa for the memory (this is just
/linux-4.1.27/arch/c6x/mm/
Dinit.c54 __pa(PAGE_OFFSET) >> PAGE_SHIFT; in paging_init()
/linux-4.1.27/arch/x86/include/asm/xen/
Dpage.h215 #define virt_to_machine(v) (phys_to_machine(XPADDR(__pa(v))))
216 #define virt_to_pfn(v) (PFN_DOWN(__pa(v)))
/linux-4.1.27/arch/arm/mach-integrator/
Dcore.c98 memblock_reserve(PHYS_OFFSET, __pa(swapper_pg_dir) - PHYS_OFFSET); in integrator_reserve()
/linux-4.1.27/arch/arm64/kernel/
Dsmp_spin_table.c93 writeq_relaxed(__pa(secondary_holding_pen), release_addr); in smp_spin_table_cpu_prepare()
/linux-4.1.27/arch/arm/kernel/
Dpsci_smp.c54 __pa(secondary_startup)); in psci_boot_secondary()
/linux-4.1.27/arch/x86/platform/efi/
Defi_32.c69 gdt_descr.address = __pa(get_cpu_gdt_table(0)); in efi_call_phys_prolog()
/linux-4.1.27/arch/powerpc/platforms/cell/
Diommu.c155 | (__pa(pte) & IOC_IOPT_CacheInvd_IOPTE_Mask) in invalidate_tce_cache()
202 io_pte[i] = base_pte | (__pa(uaddr) & CBE_IOPTE_RPN_Mask); in tce_build_cell()
229 __pa(window->iommu->pad_page) | in tce_free_cell()
381 iommu->stab[i] = reg | (__pa(ptab) + (n_pte_pages << 12) * in cell_iommu_alloc_ptab()
420 reg = IOC_IOST_Origin_E | __pa(iommu->stab) | IOC_IOST_Origin_HW; in cell_iommu_enable_hardware()
997 ptab[offset] = base_pte | (__pa(addr) & CBE_IOPTE_RPN_Mask); in insert_16M_pte()
1087 hbase = __pa(htab_address); in cell_iommu_fixed_mapping_init()
Dsmp.c75 __pa(ppc_function_entry(generic_secondary_smp_init)); in smp_startup_cpu()
/linux-4.1.27/arch/arm/xen/
Denlighten.c96 info.mfn = __pa(vcpup) >> PAGE_SHIFT; in xen_percpu_init()
207 xatp.gpfn = __pa(shared_info_page) >> PAGE_SHIFT; in xen_guest_init()
/linux-4.1.27/arch/ia64/kernel/
Dsetup.c170 call_pernode_memory(__pa(range_start), range_end - range_start, func); in filter_rsvd_memory()
200 call_pernode_memory(__pa(start), end - start, func); in filter_memory()
321 boot_param_res.start = __pa(ia64_boot_param); in setup_crashkernel()
444 ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase)); in io_port_init()
989 ia64_set_kr(IA64_KR_PT_BASE, __pa(ia64_imva(empty_zero_page))); in cpu_init()
/linux-4.1.27/arch/sh/mm/
Dcache-sh3.c50 p = __pa(v); in sh3__flush_wback_region()
/linux-4.1.27/arch/mn10300/mm/
Dinit.c65 set_pgd(swapper_pg_dir + loop, __pgd(__pa(ppte) | _PAGE_TABLE)); in paging_init()
/linux-4.1.27/arch/tile/include/asm/
Dmmu_context.h40 int rc = hv_install_context(__pa(pgdir), prot, asid, in __install_page_table()
/linux-4.1.27/drivers/watchdog/
Dwdrtas.c124 WDRTAS_SP_SPI, __pa(rtas_data_buf), in wdrtas_get_interval()
176 (void *)__pa(wdrtas_logbuffer), in wdrtas_timer_keepalive()
/linux-4.1.27/arch/powerpc/mm/
Dhash_utils_64.c619 return htab_bolt_mapping(start, end, __pa(start), in create_section_mapping()
780 __pa(base), prot, in htab_initialize()
786 __pa(dart_table_end), in htab_initialize()
793 BUG_ON(htab_bolt_mapping(base, base + size, __pa(base), in htab_initialize()
813 __pa(tce_alloc_start), prot, in htab_initialize()
1487 ret = hpte_insert_repeating(hash, vpn, __pa(vaddr), mode, in kernel_map_linear_page()
1526 lmi = __pa(vaddr) >> PAGE_SHIFT; in __kernel_map_pages()
Dinit_64.c318 vmemmap_list_populate(__pa(p), start, node); in vmemmap_populate()
323 vmemmap_create_mapping(start, page_size, __pa(p)); in vmemmap_populate()
/linux-4.1.27/arch/cris/include/asm/
Dpage.h60 #define page_to_phys(page) __pa((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
/linux-4.1.27/arch/s390/net/
Dbpf_jit_comp.c1064 func_addr = __pa(sk_load_byte_pos); in bpf_jit_insn()
1066 func_addr = __pa(sk_load_byte); in bpf_jit_insn()
1071 func_addr = __pa(sk_load_half_pos); in bpf_jit_insn()
1073 func_addr = __pa(sk_load_half); in bpf_jit_insn()
1078 func_addr = __pa(sk_load_word_pos); in bpf_jit_insn()
1080 func_addr = __pa(sk_load_word); in bpf_jit_insn()
/linux-4.1.27/arch/mips/bcm47xx/
Dprom.c163 off = EXTVBASE + __pa(off); in bcm47xx_prom_highmem_init()
/linux-4.1.27/arch/sh/boot/compressed/
Dmisc.c123 output_addr = __pa((unsigned long)&_text+PAGE_SIZE); in decompress_kernel()
/linux-4.1.27/drivers/char/
Dmspec.c119 nid = nasid_to_cnodeid(get_node_number(__pa(addr))); in mspec_zero_block()
379 phys = __pa(scratch_page[nid]); in mspec_init()
/linux-4.1.27/arch/alpha/kernel/
Dcore_irongate.c232 if (initrd_end && __pa(initrd_end) > pci_mem) { in albacore_init_arch()
236 free_bootmem_node(NODE_DATA(0), __pa(initrd_start), in albacore_init_arch()
/linux-4.1.27/drivers/s390/char/
Draw3270.c157 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_alloc()
182 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_reset()
216 rq->ccw.cda = __pa(data); in raw3270_request_set_data()
226 rq->ccw.cda = __pa(ib->data); in raw3270_request_set_idal()
593 rp->init_readmod.ccw.cda = (__u32) __pa(rp->init_data); in raw3270_read_modified()
612 rp->init_readpart.ccw.cda = (__u32) __pa(&rp->init_data); in raw3270_writesf_readpart()
651 rp->init_reset.ccw.cda = (__u32) __pa(rp->init_data); in __raw3270_reset_device()
/linux-4.1.27/arch/nios2/include/asm/
Dpage.h81 # define __pa(x) \ macro
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dhcp_if.c379 u64 r_cb = __pa(query_port_response_block); in hipz_h_query_port()
421 u64 r_cb = __pa(query_hca_rblock); in hipz_h_query_hca()
560 __pa(mqpcb), /* r7 */ in hipz_h_modify_qp()
578 __pa(qqpcb), /* r6 */ in hipz_h_query_qp()
927 u64 r_cb = __pa(rblock); in hipz_h_error_data()
/linux-4.1.27/arch/powerpc/platforms/44x/
Diss4xx.c117 spin_table[1] = __pa(start_secondary_47x); in smp_iss4xx_kick_cpu()
/linux-4.1.27/lib/
Dcpumask.c136 memblock_free_early(__pa(mask), cpumask_size()); in free_bootmem_cpumask_var()
Dswiotlb.c170 io_tlb_start = __pa(tlb); in swiotlb_init_with_tbl()
182 io_tlb_overflow_buffer = __pa(v_overflow_buffer); in swiotlb_init_with_tbl()
370 memblock_free_late(__pa(io_tlb_orig_addr), in swiotlb_free()
372 memblock_free_late(__pa(io_tlb_list), in swiotlb_free()
/linux-4.1.27/drivers/s390/block/
Dxpram.c100 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_in()
130 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_out()
/linux-4.1.27/arch/um/drivers/
Dmmapper_kern.c113 p_buf = __pa(v_buf); in mmapper_init()
/linux-4.1.27/arch/arm/include/asm/xen/
Dpage.h70 #define virt_to_machine(v) (phys_to_machine(XPADDR(__pa(v))))
/linux-4.1.27/drivers/misc/sgi-xp/
Dxp_sn2.c83 return __pa(addr); in xp_pa_sn2()
/linux-4.1.27/arch/nios2/kernel/
Dsetup.c146 memory_start = PAGE_ALIGN((unsigned long)__pa(_end)); in setup_arch()
/linux-4.1.27/arch/powerpc/platforms/85xx/
Dsmp.c297 out_be32(&spin_table->addr_l, __pa(__early_start)); in smp_85xx_kick_cpu()
315 __pa(ppc_function_entry(generic_secondary_smp_init))); in smp_85xx_kick_cpu()

123