Lines Matching refs:gfn_t

249 static gfn_t get_mmio_spte_gfn(u64 spte)  in get_mmio_spte_gfn()
261 static bool set_mmio_spte(struct kvm *kvm, u64 *sptep, gfn_t gfn, in set_mmio_spte()
333 static gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta()
754 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
762 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
774 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
784 static void account_shadowed(struct kvm *kvm, gfn_t gfn) in account_shadowed()
799 static void unaccount_shadowed(struct kvm *kvm, gfn_t gfn) in unaccount_shadowed()
816 gfn_t gfn, in has_wrprotected_page()
831 static int host_mapping_level(struct kvm *kvm, gfn_t gfn) in host_mapping_level()
850 gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
863 static bool mapping_level_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t large_gfn) in mapping_level_dirty_bitmap()
868 static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn) in mapping_level()
1010 static unsigned long *__gfn_to_rmap(gfn_t gfn, int level, in __gfn_to_rmap()
1022 static unsigned long *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, int level) in gfn_to_rmap()
1038 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
1052 gfn_t gfn; in rmap_remove()
1261 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked()
1286 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_clear_dirty_pt_masked()
1313 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1341 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_unmap_rmapp()
1361 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_set_pte_rmapp()
1411 gfn_t gfn, in kvm_handle_hva_range()
1424 gfn_t gfn_start, gfn_end; in kvm_handle_hva_range()
1442 gfn_t gfn = gfn_start; in kvm_handle_hva_range()
1467 gfn_t gfn, int level, in kvm_handle_hva()
1489 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_age_rmapp()
1513 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_test_age_rmapp()
1543 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1625 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
1872 static void kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_sync_pages()
2011 gfn_t gfn, in kvm_mmu_get_page()
2344 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2459 u8 kvm_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_get_guest_memory_type()
2480 static void kvm_unsync_pages(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_unsync_pages()
2492 static int mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, in mmu_need_write_protect()
2515 gfn_t gfn, pfn_t pfn, bool speculative, in set_spte()
2595 int level, gfn_t gfn, pfn_t pfn, bool speculative, in mmu_set_spte()
2655 static pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in pte_prefetch_gfn_to_pfn()
2674 gfn_t gfn; in direct_pte_prefetch_many()
2736 int map_writable, int level, gfn_t gfn, pfn_t pfn, in __direct_map()
2742 gfn_t pseudo_gfn; in __direct_map()
2786 static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, pfn_t pfn) in kvm_handle_bad_page()
2806 gfn_t *gfnp, pfn_t *pfnp, int *levelp) in transparent_hugepage_adjust()
2809 gfn_t gfn = *gfnp; in transparent_hugepage_adjust()
2846 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn()
2890 gfn_t gfn; in fast_pf_fix_direct_spte()
3000 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn,
3005 gfn_t gfn, bool prefault) in nonpaging_map()
3106 static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn) in mmu_check_root()
3158 gfn_t root_gfn; in mmu_alloc_shadow_roots()
3350 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault_common()
3385 gfn_t gfn; in nonpaging_page_fault()
3409 static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn) in kvm_arch_setup_async_pf()
3430 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, in try_async_pf()
3462 gfn_t gfn = gpa >> PAGE_SHIFT; in tdp_page_fault()
3548 static bool sync_mmio_spte(struct kvm *kvm, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
4167 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_pte_write()
4381 gfn_t last_gfn; in kvm_mmu_slot_remove_write_access()
4501 gfn_t last_gfn; in kvm_mmu_slot_leaf_clear_dirty()
4540 gfn_t last_gfn; in kvm_mmu_slot_largepage_remove_write_access()
4578 gfn_t last_gfn; in kvm_mmu_slot_set_dirty()