Lines Matching refs:vaddr
127 unsigned long arbitrary_virt_to_mfn(void *vaddr) in arbitrary_virt_to_mfn() argument
129 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); in arbitrary_virt_to_mfn()
134 xmaddr_t arbitrary_virt_to_machine(void *vaddr) in arbitrary_virt_to_machine() argument
136 unsigned long address = (unsigned long)vaddr; in arbitrary_virt_to_machine()
145 if (virt_addr_valid(vaddr)) in arbitrary_virt_to_machine()
146 return virt_to_machine(vaddr); in arbitrary_virt_to_machine()
157 void make_lowmem_page_readonly(void *vaddr) in make_lowmem_page_readonly() argument
160 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readonly()
173 void make_lowmem_page_readwrite(void *vaddr) in make_lowmem_page_readwrite() argument
176 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readwrite()
289 void set_pte_mfn(unsigned long vaddr, unsigned long mfn, pgprot_t flags) in set_pte_mfn() argument
291 set_pte_vaddr(vaddr, mfn_pte(mfn, flags)); in set_pte_mfn()
1108 static void __init xen_cleanhighmap(unsigned long vaddr, in xen_cleanhighmap() argument
1112 pmd_t *pmd = level2_kernel_pgt + pmd_index(vaddr); in xen_cleanhighmap()
1116 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PAGE_SIZE)); in xen_cleanhighmap()
1117 pmd++, vaddr += PMD_SIZE) { in xen_cleanhighmap()
1120 if (vaddr < (unsigned long) _text || vaddr > kernel_end) in xen_cleanhighmap()
1133 void *vaddr = __va(paddr); in xen_free_ro_pages() local
1134 void *vaddr_end = vaddr + size; in xen_free_ro_pages()
1136 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) in xen_free_ro_pages()
1137 make_lowmem_page_readwrite(vaddr); in xen_free_ro_pages()
1156 static void __init xen_cleanmfnmap(unsigned long vaddr) in xen_cleanmfnmap() argument
1158 unsigned long va = vaddr & PMD_MASK; in xen_cleanmfnmap()
1168 unpin = (vaddr == 2 * PGDIR_SIZE); in xen_cleanmfnmap()
2030 unsigned long *vaddr; in xen_read_phys_ulong() local
2033 vaddr = early_memremap_ro(addr, sizeof(val)); in xen_read_phys_ulong()
2034 val = *vaddr; in xen_read_phys_ulong()
2035 early_memunmap(vaddr, sizeof(val)); in xen_read_phys_ulong()
2043 static phys_addr_t __init xen_early_virt_to_phys(unsigned long vaddr) in xen_early_virt_to_phys() argument
2052 pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) * in xen_early_virt_to_phys()
2058 pud = native_make_pud(xen_read_phys_ulong(pa + pud_index(vaddr) * in xen_early_virt_to_phys()
2064 return pa + (vaddr & ~PUD_MASK); in xen_early_virt_to_phys()
2066 pmd = native_make_pmd(xen_read_phys_ulong(pa + pmd_index(vaddr) * in xen_early_virt_to_phys()
2072 return pa + (vaddr & ~PMD_MASK); in xen_early_virt_to_phys()
2074 pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) * in xen_early_virt_to_phys()
2080 return pa | (vaddr & ~PAGE_MASK); in xen_early_virt_to_phys()
2382 unsigned long vaddr = __fix_to_virt(idx); in xen_set_fixmap() local
2383 set_pte_vaddr_pud(level3_user_vsyscall, vaddr, pte); in xen_set_fixmap()
2511 static void xen_zap_pfn_range(unsigned long vaddr, unsigned int order, in xen_zap_pfn_range() argument
2519 for (i = 0; i < (1UL<<order); i++, vaddr += PAGE_SIZE) { in xen_zap_pfn_range()
2523 in_frames[i] = virt_to_mfn(vaddr); in xen_zap_pfn_range()
2525 MULTI_update_va_mapping(mcs.mc, vaddr, VOID_PTE, 0); in xen_zap_pfn_range()
2526 __set_phys_to_machine(virt_to_pfn(vaddr), INVALID_P2M_ENTRY); in xen_zap_pfn_range()
2529 out_frames[i] = virt_to_pfn(vaddr); in xen_zap_pfn_range()
2539 static void xen_remap_exchanged_ptes(unsigned long vaddr, int order, in xen_remap_exchanged_ptes() argument
2549 for (i = 0; i < limit; i++, vaddr += PAGE_SIZE) { in xen_remap_exchanged_ptes()
2568 MULTI_update_va_mapping(mcs.mc, vaddr, in xen_remap_exchanged_ptes()
2571 set_phys_to_machine(virt_to_pfn(vaddr), mfn); in xen_remap_exchanged_ptes()