kvm_mmu_page      392 arch/x86/include/asm/kvm_host.h 			 struct kvm_mmu_page *sp);
kvm_mmu_page      394 arch/x86/include/asm/kvm_host.h 	void (*update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page     1470 arch/x86/include/asm/kvm_host.h static inline struct kvm_mmu_page *page_header(hpa_t shadow_page)
kvm_mmu_page     1474 arch/x86/include/asm/kvm_host.h 	return (struct kvm_mmu_page *)page_private(page);
kvm_mmu_page      359 arch/x86/kvm/mmu.c static inline bool sp_ad_disabled(struct kvm_mmu_page *sp)
kvm_mmu_page      676 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(sptep));
kvm_mmu_page      760 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(sptep));
kvm_mmu_page     1155 arch/x86/kvm/mmu.c static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index)
kvm_mmu_page     1163 arch/x86/kvm/mmu.c static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn)
kvm_mmu_page     1214 arch/x86/kvm/mmu.c static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     1233 arch/x86/kvm/mmu.c static void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     1244 arch/x86/kvm/mmu.c static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     1261 arch/x86/kvm/mmu.c static void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     1479 arch/x86/kvm/mmu.c 					 struct kvm_mmu_page *sp)
kvm_mmu_page     1499 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     1510 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     1619 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *sp = page_header(__pa(sptep));
kvm_mmu_page     2091 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2140 arch/x86/kvm/mmu.c static void kvm_mmu_free_page(struct kvm_mmu_page *sp)
kvm_mmu_page     2157 arch/x86/kvm/mmu.c 				    struct kvm_mmu_page *sp, u64 *parent_pte)
kvm_mmu_page     2165 arch/x86/kvm/mmu.c static void mmu_page_remove_parent_pte(struct kvm_mmu_page *sp,
kvm_mmu_page     2171 arch/x86/kvm/mmu.c static void drop_parent_pte(struct kvm_mmu_page *sp,
kvm_mmu_page     2178 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, int direct)
kvm_mmu_page     2180 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2200 arch/x86/kvm/mmu.c static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp)
kvm_mmu_page     2212 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2225 arch/x86/kvm/mmu.c 			       struct kvm_mmu_page *sp)
kvm_mmu_page     2235 arch/x86/kvm/mmu.c 				 struct kvm_mmu_page *sp, u64 *spte,
kvm_mmu_page     2245 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *sp;
kvm_mmu_page     2251 arch/x86/kvm/mmu.c static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp,
kvm_mmu_page     2267 arch/x86/kvm/mmu.c static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx)
kvm_mmu_page     2274 arch/x86/kvm/mmu.c static int __mmu_unsync_walk(struct kvm_mmu_page *sp,
kvm_mmu_page     2280 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *child;
kvm_mmu_page     2315 arch/x86/kvm/mmu.c static int mmu_unsync_walk(struct kvm_mmu_page *sp,
kvm_mmu_page     2326 arch/x86/kvm/mmu.c static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     2334 arch/x86/kvm/mmu.c static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
kvm_mmu_page     2350 arch/x86/kvm/mmu.c static inline bool is_ept_sp(struct kvm_mmu_page *sp)
kvm_mmu_page     2356 arch/x86/kvm/mmu.c static bool __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page     2400 arch/x86/kvm/mmu.c static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     2406 arch/x86/kvm/mmu.c static bool kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page     2417 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *s;
kvm_mmu_page     2432 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *parent[PT64_ROOT_MAX_LEVEL];
kvm_mmu_page     2448 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *sp = pvec->page[n].sp;
kvm_mmu_page     2465 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2488 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2504 arch/x86/kvm/mmu.c 			      struct kvm_mmu_page *parent)
kvm_mmu_page     2507 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2538 arch/x86/kvm/mmu.c static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp)
kvm_mmu_page     2545 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp =  page_header(__pa(spte));
kvm_mmu_page     2550 arch/x86/kvm/mmu.c static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu,
kvm_mmu_page     2559 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2704 arch/x86/kvm/mmu.c 			     struct kvm_mmu_page *sp)
kvm_mmu_page     2730 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *child;
kvm_mmu_page     2748 arch/x86/kvm/mmu.c static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
kvm_mmu_page     2752 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *child;
kvm_mmu_page     2774 arch/x86/kvm/mmu.c 					 struct kvm_mmu_page *sp)
kvm_mmu_page     2782 arch/x86/kvm/mmu.c static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page     2792 arch/x86/kvm/mmu.c 				   struct kvm_mmu_page *parent,
kvm_mmu_page     2803 arch/x86/kvm/mmu.c 		struct kvm_mmu_page *sp;
kvm_mmu_page     2816 arch/x86/kvm/mmu.c 				       struct kvm_mmu_page *sp,
kvm_mmu_page     2860 arch/x86/kvm/mmu.c static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
kvm_mmu_page     2872 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *nsp;
kvm_mmu_page     2897 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2903 arch/x86/kvm/mmu.c 			     struct kvm_mmu_page, link);
kvm_mmu_page     2934 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     2954 arch/x86/kvm/mmu.c static void kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
kvm_mmu_page     2966 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3056 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3173 arch/x86/kvm/mmu.c 			struct kvm_mmu_page *child;
kvm_mmu_page     3232 arch/x86/kvm/mmu.c 				    struct kvm_mmu_page *sp,
kvm_mmu_page     3260 arch/x86/kvm/mmu.c 				  struct kvm_mmu_page *sp, u64 *sptep)
kvm_mmu_page     3284 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3330 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3485 arch/x86/kvm/mmu.c fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page     3540 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3703 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3779 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3820 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     3934 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     5304 arch/x86/kvm/mmu.c 				  struct kvm_mmu_page *sp, u64 *spte,
kvm_mmu_page     5359 arch/x86/kvm/mmu.c static bool detect_write_flooding(struct kvm_mmu_page *sp)
kvm_mmu_page     5376 arch/x86/kvm/mmu.c static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa,
kvm_mmu_page     5400 arch/x86/kvm/mmu.c static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte)
kvm_mmu_page     5436 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     5831 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *node;
kvm_mmu_page     6019 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     6123 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *node;
kvm_mmu_page     6339 arch/x86/kvm/mmu.c 						  sizeof(struct kvm_mmu_page),
kvm_mmu_page     6426 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp;
kvm_mmu_page     6443 arch/x86/kvm/mmu.c 				      struct kvm_mmu_page,
kvm_mmu_page       34 arch/x86/kvm/mmu_audit.c static void __mmu_spte_walk(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page       46 arch/x86/kvm/mmu_audit.c 			struct kvm_mmu_page *child;
kvm_mmu_page       57 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *sp;
kvm_mmu_page       83 arch/x86/kvm/mmu_audit.c typedef void (*sp_handler) (struct kvm *kvm, struct kvm_mmu_page *sp);
kvm_mmu_page       87 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *sp;
kvm_mmu_page       95 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *sp;
kvm_mmu_page      130 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *rev_sp;
kvm_mmu_page      168 arch/x86/kvm/mmu_audit.c 	struct kvm_mmu_page *sp = page_header(__pa(sptep));
kvm_mmu_page      175 arch/x86/kvm/mmu_audit.c static void check_mappings_rmap(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page      190 arch/x86/kvm/mmu_audit.c static void audit_write_protection(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page      213 arch/x86/kvm/mmu_audit.c static void audit_sp(struct kvm *kvm, struct kvm_mmu_page *sp)
kvm_mmu_page      152 arch/x86/kvm/mmutrace.h 	TP_PROTO(struct kvm_mmu_page *sp, bool created),
kvm_mmu_page      171 arch/x86/kvm/mmutrace.h 	TP_PROTO(struct kvm_mmu_page *sp),
kvm_mmu_page      186 arch/x86/kvm/mmutrace.h 	TP_PROTO(struct kvm_mmu_page *sp),
kvm_mmu_page      192 arch/x86/kvm/mmutrace.h 	TP_PROTO(struct kvm_mmu_page *sp),
kvm_mmu_page      198 arch/x86/kvm/mmutrace.h 	TP_PROTO(struct kvm_mmu_page *sp),
kvm_mmu_page      175 arch/x86/kvm/paging_tmpl.h 				  struct kvm_mmu_page *sp, u64 *spte,
kvm_mmu_page      516 arch/x86/kvm/paging_tmpl.h FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page      547 arch/x86/kvm/paging_tmpl.h static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
kvm_mmu_page      581 arch/x86/kvm/paging_tmpl.h 	struct kvm_mmu_page *sp;
kvm_mmu_page      620 arch/x86/kvm/paging_tmpl.h 	struct kvm_mmu_page *sp = NULL;
kvm_mmu_page      880 arch/x86/kvm/paging_tmpl.h static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp)
kvm_mmu_page      895 arch/x86/kvm/paging_tmpl.h 	struct kvm_mmu_page *sp;
kvm_mmu_page     1007 arch/x86/kvm/paging_tmpl.h static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)