Lines Matching refs:kvm_mmu_page
371 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in count_spte_clear()
455 struct kvm_mmu_page *sp = page_header(__pa(sptep)); in __get_spte_lockless()
754 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
762 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
784 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
802 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
1047 static unsigned long *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, struct kvm_mmu_page *sp) in gfn_to_rmap()
1067 struct kvm_mmu_page *sp; in rmap_add()
1078 struct kvm_mmu_page *sp; in rmap_remove()
1617 struct kvm_mmu_page *sp; in rmap_recycle()
1685 static void kvm_mmu_free_page(struct kvm_mmu_page *sp) in kvm_mmu_free_page()
1702 struct kvm_mmu_page *sp, u64 *parent_pte) in mmu_page_add_parent_pte()
1710 static void mmu_page_remove_parent_pte(struct kvm_mmu_page *sp, in mmu_page_remove_parent_pte()
1716 static void drop_parent_pte(struct kvm_mmu_page *sp, in drop_parent_pte()
1723 static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, in kvm_mmu_alloc_page()
1726 struct kvm_mmu_page *sp; in kvm_mmu_alloc_page()
1747 static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp) in kvm_mmu_mark_parents_unsync()
1754 struct kvm_mmu_page *sp; in mark_unsync()
1767 struct kvm_mmu_page *sp) in nonpaging_sync_page()
1777 struct kvm_mmu_page *sp, u64 *spte, in nonpaging_update_pte()
1787 struct kvm_mmu_page *sp;
1793 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add()
1809 static int __mmu_unsync_walk(struct kvm_mmu_page *sp, in __mmu_unsync_walk()
1815 struct kvm_mmu_page *child; in __mmu_unsync_walk()
1853 static int mmu_unsync_walk(struct kvm_mmu_page *sp, in mmu_unsync_walk()
1863 static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unlink_unsync_page()
1871 static int kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
1896 static int __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in __kvm_sync_page()
1917 struct kvm_mmu_page *sp) in kvm_sync_page_transient()
1936 static int kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in kvm_sync_page()
1945 struct kvm_mmu_page *s; in kvm_sync_pages()
1969 struct kvm_mmu_page *parent[PT64_ROOT_LEVEL-1];
1986 struct kvm_mmu_page *sp = pvec->page[n].sp; in mmu_pages_next()
2002 struct kvm_mmu_page *sp; in mmu_pages_clear_parents()
2019 static void kvm_mmu_pages_init(struct kvm_mmu_page *parent, in kvm_mmu_pages_init()
2028 struct kvm_mmu_page *parent) in mmu_sync_children()
2031 struct kvm_mmu_page *sp; in mmu_sync_children()
2056 static void init_shadow_page_table(struct kvm_mmu_page *sp) in init_shadow_page_table()
2064 static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp) in __clear_sp_write_flooding_count()
2071 struct kvm_mmu_page *sp = page_header(__pa(spte)); in clear_sp_write_flooding_count()
2076 static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in is_obsolete_sp()
2081 static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, in kvm_mmu_get_page()
2091 struct kvm_mmu_page *sp; in kvm_mmu_get_page()
2201 static void link_shadow_page(u64 *sptep, struct kvm_mmu_page *sp, bool accessed) in link_shadow_page()
2221 struct kvm_mmu_page *child; in validate_direct_spte()
2239 static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, in mmu_page_zap_pte()
2243 struct kvm_mmu_page *child; in mmu_page_zap_pte()
2265 struct kvm_mmu_page *sp) in kvm_mmu_page_unlink_children()
2273 static void kvm_mmu_put_page(struct kvm_mmu_page *sp, u64 *parent_pte) in kvm_mmu_put_page()
2278 static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_mmu_unlink_parents()
2288 struct kvm_mmu_page *parent, in mmu_zap_unsync_children()
2300 struct kvm_mmu_page *sp; in mmu_zap_unsync_children()
2313 static int kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, in kvm_mmu_prepare_zap_page()
2352 struct kvm_mmu_page *sp, *nsp; in kvm_mmu_commit_zap_page()
2378 struct kvm_mmu_page *sp; in prepare_zap_oldest_mmu_page()
2384 struct kvm_mmu_page, link); in prepare_zap_oldest_mmu_page()
2417 struct kvm_mmu_page *sp; in kvm_mmu_unprotect_page()
2437 static void __kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in __kvm_unsync_page()
2448 struct kvm_mmu_page *s; in kvm_unsync_pages()
2461 struct kvm_mmu_page *s; in mmu_need_write_protect()
2585 struct kvm_mmu_page *child; in mmu_set_spte()
2642 struct kvm_mmu_page *sp, in direct_pte_prefetch_many()
2669 struct kvm_mmu_page *sp, u64 *sptep) in __direct_pte_prefetch()
2693 struct kvm_mmu_page *sp; in direct_pte_prefetch()
2716 struct kvm_mmu_page *sp; in __direct_map()
2863 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in fast_pf_fix_direct_spte()
2903 struct kvm_mmu_page *sp; in fast_page_fault()
3038 struct kvm_mmu_page *sp; in mmu_free_roots()
3094 struct kvm_mmu_page *sp; in mmu_alloc_direct_roots()
3130 struct kvm_mmu_page *sp; in mmu_alloc_shadow_roots()
3235 struct kvm_mmu_page *sp; in mmu_sync_roots()
4136 struct kvm_mmu_page *sp, u64 *spte, in mmu_pte_write_new_pte()
4213 static bool detect_write_flooding(struct kvm_mmu_page *sp) in detect_write_flooding()
4229 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, in detect_write_misaligned()
4253 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) in get_written_sptes()
4288 struct kvm_mmu_page *sp; in kvm_mmu_pte_write()
4638 struct kvm_mmu_page *sp; in kvm_mmu_zap_collapsible_spte()
4734 struct kvm_mmu_page *sp, *node; in kvm_zap_obsolete_pages()
4919 sizeof(struct kvm_mmu_page), in kvm_mmu_module_init()