kvm_mmu_page 392 arch/x86/include/asm/kvm_host.h struct kvm_mmu_page *sp); kvm_mmu_page 394 arch/x86/include/asm/kvm_host.h void (*update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 1470 arch/x86/include/asm/kvm_host.h static inline struct kvm_mmu_page *page_header(hpa_t shadow_page) kvm_mmu_page 1474 arch/x86/include/asm/kvm_host.h return (struct kvm_mmu_page *)page_private(page); kvm_mmu_page 359 arch/x86/kvm/mmu.c static inline bool sp_ad_disabled(struct kvm_mmu_page *sp) kvm_mmu_page 676 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); kvm_mmu_page 760 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); kvm_mmu_page 1155 arch/x86/kvm/mmu.c static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) kvm_mmu_page 1163 arch/x86/kvm/mmu.c static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) kvm_mmu_page 1214 arch/x86/kvm/mmu.c static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 1233 arch/x86/kvm/mmu.c static void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 1244 arch/x86/kvm/mmu.c static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 1261 arch/x86/kvm/mmu.c static void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 1479 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp) kvm_mmu_page 1499 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 1510 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 1619 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); kvm_mmu_page 2091 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2140 arch/x86/kvm/mmu.c static void kvm_mmu_free_page(struct kvm_mmu_page *sp) kvm_mmu_page 2157 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, u64 *parent_pte) kvm_mmu_page 2165 arch/x86/kvm/mmu.c static void mmu_page_remove_parent_pte(struct kvm_mmu_page *sp, kvm_mmu_page 2171 arch/x86/kvm/mmu.c static void drop_parent_pte(struct kvm_mmu_page *sp, kvm_mmu_page 2178 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct) kvm_mmu_page 2180 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2200 arch/x86/kvm/mmu.c static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp) kvm_mmu_page 2212 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2225 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp) kvm_mmu_page 2235 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, u64 *spte, kvm_mmu_page 2245 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2251 arch/x86/kvm/mmu.c static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, kvm_mmu_page 2267 arch/x86/kvm/mmu.c static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx) kvm_mmu_page 2274 arch/x86/kvm/mmu.c static int __mmu_unsync_walk(struct kvm_mmu_page *sp, kvm_mmu_page 2280 arch/x86/kvm/mmu.c struct kvm_mmu_page *child; kvm_mmu_page 2315 arch/x86/kvm/mmu.c static int mmu_unsync_walk(struct kvm_mmu_page *sp, kvm_mmu_page 2326 arch/x86/kvm/mmu.c static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 2334 arch/x86/kvm/mmu.c static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, kvm_mmu_page 2350 arch/x86/kvm/mmu.c static inline bool is_ept_sp(struct kvm_mmu_page *sp) kvm_mmu_page 2356 arch/x86/kvm/mmu.c static bool __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 2400 arch/x86/kvm/mmu.c static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 2406 arch/x86/kvm/mmu.c static bool kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 2417 arch/x86/kvm/mmu.c struct kvm_mmu_page *s; kvm_mmu_page 2432 arch/x86/kvm/mmu.c struct kvm_mmu_page *parent[PT64_ROOT_MAX_LEVEL]; kvm_mmu_page 2448 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = pvec->page[n].sp; kvm_mmu_page 2465 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2488 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2504 arch/x86/kvm/mmu.c struct kvm_mmu_page *parent) kvm_mmu_page 2507 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2538 arch/x86/kvm/mmu.c static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp) kvm_mmu_page 2545 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(spte)); kvm_mmu_page 2550 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, kvm_mmu_page 2559 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2704 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp) kvm_mmu_page 2730 arch/x86/kvm/mmu.c struct kvm_mmu_page *child; kvm_mmu_page 2748 arch/x86/kvm/mmu.c static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, kvm_mmu_page 2752 arch/x86/kvm/mmu.c struct kvm_mmu_page *child; kvm_mmu_page 2774 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp) kvm_mmu_page 2782 arch/x86/kvm/mmu.c static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 2792 arch/x86/kvm/mmu.c struct kvm_mmu_page *parent, kvm_mmu_page 2803 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2816 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, kvm_mmu_page 2860 arch/x86/kvm/mmu.c static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, kvm_mmu_page 2872 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *nsp; kvm_mmu_page 2897 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2903 arch/x86/kvm/mmu.c struct kvm_mmu_page, link); kvm_mmu_page 2934 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 2954 arch/x86/kvm/mmu.c static void kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) kvm_mmu_page 2966 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3056 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3173 arch/x86/kvm/mmu.c struct kvm_mmu_page *child; kvm_mmu_page 3232 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, kvm_mmu_page 3260 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, u64 *sptep) kvm_mmu_page 3284 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3330 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3485 arch/x86/kvm/mmu.c fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 3540 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3703 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3779 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3820 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 3934 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 5304 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, u64 *spte, kvm_mmu_page 5359 arch/x86/kvm/mmu.c static bool detect_write_flooding(struct kvm_mmu_page *sp) kvm_mmu_page 5376 arch/x86/kvm/mmu.c static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, kvm_mmu_page 5400 arch/x86/kvm/mmu.c static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte) kvm_mmu_page 5436 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 5831 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node; kvm_mmu_page 6019 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 6123 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node; kvm_mmu_page 6339 arch/x86/kvm/mmu.c sizeof(struct kvm_mmu_page), kvm_mmu_page 6426 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp; kvm_mmu_page 6443 arch/x86/kvm/mmu.c struct kvm_mmu_page, kvm_mmu_page 34 arch/x86/kvm/mmu_audit.c static void __mmu_spte_walk(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 46 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *child; kvm_mmu_page 57 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *sp; kvm_mmu_page 83 arch/x86/kvm/mmu_audit.c typedef void (*sp_handler) (struct kvm *kvm, struct kvm_mmu_page *sp); kvm_mmu_page 87 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *sp; kvm_mmu_page 95 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *sp; kvm_mmu_page 130 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *rev_sp; kvm_mmu_page 168 arch/x86/kvm/mmu_audit.c struct kvm_mmu_page *sp = page_header(__pa(sptep)); kvm_mmu_page 175 arch/x86/kvm/mmu_audit.c static void check_mappings_rmap(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 190 arch/x86/kvm/mmu_audit.c static void audit_write_protection(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 213 arch/x86/kvm/mmu_audit.c static void audit_sp(struct kvm *kvm, struct kvm_mmu_page *sp) kvm_mmu_page 152 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_mmu_page *sp, bool created), kvm_mmu_page 171 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_mmu_page *sp), kvm_mmu_page 186 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_mmu_page *sp), kvm_mmu_page 192 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_mmu_page *sp), kvm_mmu_page 198 arch/x86/kvm/mmutrace.h TP_PROTO(struct kvm_mmu_page *sp), kvm_mmu_page 175 arch/x86/kvm/paging_tmpl.h struct kvm_mmu_page *sp, u64 *spte, kvm_mmu_page 516 arch/x86/kvm/paging_tmpl.h FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 547 arch/x86/kvm/paging_tmpl.h static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, kvm_mmu_page 581 arch/x86/kvm/paging_tmpl.h struct kvm_mmu_page *sp; kvm_mmu_page 620 arch/x86/kvm/paging_tmpl.h struct kvm_mmu_page *sp = NULL; kvm_mmu_page 880 arch/x86/kvm/paging_tmpl.h static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp) kvm_mmu_page 895 arch/x86/kvm/paging_tmpl.h struct kvm_mmu_page *sp; kvm_mmu_page 1007 arch/x86/kvm/paging_tmpl.h static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)