/linux-4.4.14/arch/nios2/mm/ |
H A D | init.c | 106 unsigned long vpage; alloc_kuser_page() local 108 vpage = get_zeroed_page(GFP_ATOMIC); alloc_kuser_page() 109 if (!vpage) alloc_kuser_page() 113 memcpy((void *)vpage, __kuser_helper_start, kuser_sz); alloc_kuser_page() 115 flush_icache_range(vpage, vpage + KUSER_SIZE); alloc_kuser_page() 116 kuser_page[0] = virt_to_page(vpage); alloc_kuser_page()
|
/linux-4.4.14/arch/arm64/kernel/ |
H A D | vdso.c | 65 unsigned long vpage; alloc_vectors_page() local 67 vpage = get_zeroed_page(GFP_ATOMIC); alloc_vectors_page() 69 if (!vpage) alloc_vectors_page() 73 memcpy((void *)vpage + 0x1000 - kuser_sz, __kuser_helper_start, alloc_vectors_page() 77 memcpy((void *)vpage + AARCH32_KERN_SIGRET_CODE_OFFSET, alloc_vectors_page() 80 flush_icache_range(vpage, vpage + PAGE_SIZE); alloc_vectors_page() 81 vectors_page[0] = virt_to_page(vpage); alloc_vectors_page()
|
/linux-4.4.14/arch/powerpc/kvm/ |
H A D | book3s_mmu_hpte.c | 48 static inline u64 kvmppc_mmu_hash_vpte(u64 vpage) kvmppc_mmu_hash_vpte() argument 50 return hash_64(vpage & 0xfffffffffULL, HPTEG_HASH_BITS_VPTE); kvmppc_mmu_hash_vpte() 53 static inline u64 kvmppc_mmu_hash_vpte_long(u64 vpage) kvmppc_mmu_hash_vpte_long() argument 55 return hash_64((vpage & 0xffffff000ULL) >> 12, kvmppc_mmu_hash_vpte_long() 60 static inline u64 kvmppc_mmu_hash_vpte_64k(u64 vpage) kvmppc_mmu_hash_vpte_64k() argument 62 return hash_64((vpage & 0xffffffff0ULL) >> 4, kvmppc_mmu_hash_vpte_64k() 86 index = kvmppc_mmu_hash_vpte(pte->pte.vpage); kvmppc_mmu_hpte_cache_map() 90 index = kvmppc_mmu_hash_vpte_long(pte->pte.vpage); kvmppc_mmu_hpte_cache_map() 96 index = kvmppc_mmu_hash_vpte_64k(pte->pte.vpage); kvmppc_mmu_hpte_cache_map() 236 if ((pte->pte.vpage & vp_mask) == guest_vp) kvmppc_mmu_pte_vflush_short() 258 if ((pte->pte.vpage & vp_mask) == guest_vp) kvmppc_mmu_pte_vflush_64k() 280 if ((pte->pte.vpage & vp_mask) == guest_vp) kvmppc_mmu_pte_vflush_long()
|
H A D | trace_pr.h | 43 __field( unsigned long long, vpage ) 53 __entry->vpage = orig_pte->vpage; 59 __entry->hpteg, __entry->va, __entry->vpage, __entry->hpaddr) 72 __field( u64, vpage ) 81 __entry->vpage = pte->pte.vpage; 90 __entry->vpage, __entry->raddr, __entry->flags) 101 __field( u64, vpage ) 110 __entry->vpage = pte->pte.vpage; 119 __entry->vpage, __entry->raddr, __entry->flags)
|
H A D | book3s_32_mmu.c | 99 return pte.vpage; kvmppc_mmu_book3s_32_ea_to_vp() 177 pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid; kvmppc_mmu_book3s_32_xlate_bat() 216 pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data); kvmppc_mmu_book3s_32_xlate_pte() 319 pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data); kvmppc_mmu_book3s_32_xlate()
|
H A D | book3s_64_mmu.c | 231 gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data); kvmppc_mmu_book3s_64_xlate() 310 gpte->vpage = kvmppc_mmu_book3s_64_ea_to_vp(vcpu, eaddr, data); kvmppc_mmu_book3s_64_xlate() 339 eaddr, avpn, gpte->vpage, gpte->raddr); kvmppc_mmu_book3s_64_xlate()
|
H A D | book3s_32_mmu_host.c | 256 orig_pte->vpage, hpaddr); kvmppc_mmu_map_page() 272 kvmppc_mmu_pte_vflush(vcpu, pte->vpage, 0xfffffffffULL); kvmppc_mmu_unmap_page()
|
H A D | book3s_pr.c | 557 pte.vpage = eaddr >> 12; kvmppc_handle_pagefault() 563 pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 575 pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 577 pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 578 pte.vpage |= vsid; kvmppc_handle_pagefault()
|
H A D | book3s_64_mmu_host.c | 225 kvmppc_mmu_pte_vflush(vcpu, pte->vpage, mask); kvmppc_mmu_unmap_page()
|
H A D | book3s.c | 408 pte->vpage = VSID_REAL | eaddr >> 12; kvmppc_xlate()
|
H A D | booke.c | 1946 pte->vpage = eaddr >> PAGE_SHIFT; kvmppc_xlate() 1975 pte->vpage = eaddr >> PAGE_SHIFT; kvmppc_xlate()
|
H A D | book3s_64_mmu_hv.c | 347 gpte->vpage = ((v & HPTE_V_AVPN) << 4) | ((eaddr >> 12) & 0xfff); kvmppc_mmu_book3s_64_hv_xlate()
|
/linux-4.4.14/drivers/staging/rdma/ehca/ |
H A D | ehca_eq.c | 61 void *vpage; ehca_create_eq() local 100 vpage = ipz_qpageit_get_inc(&eq->ipz_queue); ehca_create_eq() 101 if (!vpage) ehca_create_eq() 104 rpage = __pa(vpage); ehca_create_eq() 112 vpage = ipz_qpageit_get_inc(&eq->ipz_queue); ehca_create_eq() 113 if (h_ret != H_SUCCESS || vpage) ehca_create_eq()
|
H A D | ehca_cq.c | 130 void *vpage; ehca_create_cq() local 209 vpage = ipz_qpageit_get_inc(&my_cq->ipz_queue); ehca_create_cq() 210 if (!vpage) { ehca_create_cq() 216 rpage = __pa(vpage); ehca_create_cq() 238 vpage = ipz_qpageit_get_inc(&my_cq->ipz_queue); ehca_create_cq() 239 if ((h_ret != H_SUCCESS) || vpage) { ehca_create_cq()
|
H A D | ehca_qp.c | 289 void *vpage; init_qp_queue() local 317 vpage = ipz_qpageit_get_inc(queue); init_qp_queue() 318 if (!vpage) { init_qp_queue() 320 "failed p_vpage= %p", vpage); init_qp_queue() 324 rpage = __pa(vpage); init_qp_queue() 338 vpage = ipz_qpageit_get_inc(&my_qp->ipz_rqueue); init_qp_queue() 339 if (vpage) { init_qp_queue() 341 "should not succeed vpage=%p", vpage); init_qp_queue()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 130 void *vpage; ehea_create_cq() local 159 vpage = hw_qpageit_get_inc(&cq->hw_queue); ehea_create_cq() 160 if (!vpage) { ehea_create_cq() 165 rpage = __pa(vpage); ehea_create_cq() 176 vpage = hw_qpageit_get_inc(&cq->hw_queue); ehea_create_cq() 178 if ((hret != H_SUCCESS) || (vpage)) { ehea_create_cq() 255 void *vpage; ehea_create_eq() local 283 vpage = hw_qpageit_get_inc(&eq->hw_queue); ehea_create_eq() 284 if (!vpage) { ehea_create_eq() 290 rpage = __pa(vpage); ehea_create_eq() 298 vpage = hw_qpageit_get_inc(&eq->hw_queue); ehea_create_eq() 299 if ((hret != H_SUCCESS) || (vpage)) ehea_create_eq() 383 void *vpage; ehea_qp_alloc_register() local 390 vpage = hw_qpageit_get_inc(hw_queue); ehea_qp_alloc_register() 391 if (!vpage) { ehea_qp_alloc_register() 395 rpage = __pa(vpage); ehea_qp_alloc_register()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | kvm_host.h | 340 u64 vpage; member in struct:kvmppc_pte
|