Lines Matching refs:FNAME
36 #define FNAME(name) paging##64_##name macro
56 #define FNAME(name) paging##32_##name macro
71 #define FNAME(name) ept_##name macro
87 #define gpte_to_gfn_lvl FNAME(gpte_to_gfn_lvl)
114 static inline void FNAME(protect_clean_gpte)(unsigned *access, unsigned gpte) in FNAME() function
131 static inline int FNAME(is_present_gpte)(unsigned long pte) in FNAME() function
140 static int FNAME(cmpxchg_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in FNAME() function
163 static bool FNAME(prefetch_invalid_gpte)(struct kvm_vcpu *vcpu, in FNAME() function
170 if (!FNAME(is_present_gpte)(gpte)) in FNAME()
184 static inline unsigned FNAME(gpte_access)(struct kvm_vcpu *vcpu, u64 gpte) in FNAME() function
199 static int FNAME(update_accessed_dirty_bits)(struct kvm_vcpu *vcpu, in FNAME() function
247 ret = FNAME(cmpxchg_gpte)(vcpu, mmu, ptep_user, index, orig_pte, pte); in FNAME()
260 static int FNAME(walk_addr_generic)(struct guest_walker *walker, in FNAME() function
287 if (!FNAME(is_present_gpte)(pte)) in FNAME()
345 if (unlikely(!FNAME(is_present_gpte)(pte))) in FNAME()
354 pte_access = pt_access & FNAME(gpte_access)(vcpu, pte); in FNAME()
377 FNAME(protect_clean_gpte)(&pte_access, pte); in FNAME()
388 ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault); in FNAME()
436 static int FNAME(walk_addr)(struct guest_walker *walker, in FNAME() function
439 return FNAME(walk_addr_generic)(walker, vcpu, &vcpu->arch.mmu, addr, in FNAME()
444 static int FNAME(walk_addr_nested)(struct guest_walker *walker, in FNAME() function
448 return FNAME(walk_addr_generic)(walker, vcpu, &vcpu->arch.nested_mmu, in FNAME()
454 FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME() function
461 if (FNAME(prefetch_invalid_gpte)(vcpu, sp, spte, gpte)) in FNAME()
467 pte_access = sp->role.access & FNAME(gpte_access)(vcpu, gpte); in FNAME()
468 FNAME(protect_clean_gpte)(&pte_access, gpte); in FNAME()
484 static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in FNAME() function
489 FNAME(prefetch_gpte)(vcpu, sp, spte, gpte, false); in FNAME()
492 static bool FNAME(gpte_changed)(struct kvm_vcpu *vcpu, in FNAME() function
515 static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw, in FNAME() function
541 if (!FNAME(prefetch_gpte)(vcpu, sp, spte, gptep[i], true)) in FNAME()
551 static int FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr, in FNAME() function
572 if (FNAME(gpte_changed)(vcpu, gw, top_level)) in FNAME()
597 if (FNAME(gpte_changed)(vcpu, gw, it.level - 1)) in FNAME()
627 FNAME(pte_prefetch)(vcpu, gw, it.sptep); in FNAME()
656 FNAME(is_self_change_mapping)(struct kvm_vcpu *vcpu, in FNAME() function
692 static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, u32 error_code, in FNAME() function
727 r = FNAME(walk_addr)(&walker, vcpu, addr, error_code); in FNAME()
742 is_self_change_mapping = FNAME(is_self_change_mapping)(vcpu, in FNAME()
793 r = FNAME(fetch)(vcpu, addr, &walker, write_fault, in FNAME()
807 static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) in FNAME() function
819 static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva) in FNAME() function
852 pte_gpa = FNAME(get_level1_sp_gpa)(sp); in FNAME()
865 FNAME(update_pte)(vcpu, sp, sptep, &gpte); in FNAME()
874 static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, gva_t vaddr, u32 access, in FNAME() function
881 r = FNAME(walk_addr)(&walker, vcpu, vaddr, access); in FNAME()
893 static gpa_t FNAME(gva_to_gpa_nested)(struct kvm_vcpu *vcpu, gva_t vaddr, in FNAME() function
901 r = FNAME(walk_addr_nested)(&walker, vcpu, vaddr, access); in FNAME()
926 static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in FNAME() function
935 first_pte_gpa = FNAME(get_level1_sp_gpa)(sp); in FNAME()
952 if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) { in FNAME()
959 pte_access &= FNAME(gpte_access)(vcpu, gpte); in FNAME()
960 FNAME(protect_clean_gpte)(&pte_access, gpte); in FNAME()
987 #undef FNAME