to_kvm_vmx        470 arch/x86/kvm/vmx/vmx.c 			to_kvm_vmx(kvm)->ept_pointers_match
to_kvm_vmx        476 arch/x86/kvm/vmx/vmx.c 	to_kvm_vmx(kvm)->ept_pointers_match = EPT_POINTERS_MATCH;
to_kvm_vmx        511 arch/x86/kvm/vmx/vmx.c 	spin_lock(&to_kvm_vmx(kvm)->ept_pointer_lock);
to_kvm_vmx        513 arch/x86/kvm/vmx/vmx.c 	if (to_kvm_vmx(kvm)->ept_pointers_match == EPT_POINTERS_CHECK)
to_kvm_vmx        516 arch/x86/kvm/vmx/vmx.c 	if (to_kvm_vmx(kvm)->ept_pointers_match != EPT_POINTERS_MATCH) {
to_kvm_vmx        528 arch/x86/kvm/vmx/vmx.c 	spin_unlock(&to_kvm_vmx(kvm)->ept_pointer_lock);
to_kvm_vmx       2720 arch/x86/kvm/vmx/vmx.c 	struct kvm_vmx *kvm_vmx = to_kvm_vmx(vcpu->kvm);
to_kvm_vmx       2983 arch/x86/kvm/vmx/vmx.c 			spin_lock(&to_kvm_vmx(kvm)->ept_pointer_lock);
to_kvm_vmx       2985 arch/x86/kvm/vmx/vmx.c 			to_kvm_vmx(kvm)->ept_pointers_match
to_kvm_vmx       2987 arch/x86/kvm/vmx/vmx.c 			spin_unlock(&to_kvm_vmx(kvm)->ept_pointer_lock);
to_kvm_vmx       2996 arch/x86/kvm/vmx/vmx.c 			guest_cr3 = to_kvm_vmx(kvm)->ept_identity_map_addr;
to_kvm_vmx       3433 arch/x86/kvm/vmx/vmx.c 	fn = to_kvm_vmx(kvm)->tss_addr >> PAGE_SHIFT;
to_kvm_vmx       3459 arch/x86/kvm/vmx/vmx.c 	struct kvm_vmx *kvm_vmx = to_kvm_vmx(kvm);
to_kvm_vmx       4500 arch/x86/kvm/vmx/vmx.c 	to_kvm_vmx(kvm)->tss_addr = addr;
to_kvm_vmx       4506 arch/x86/kvm/vmx/vmx.c 	to_kvm_vmx(kvm)->ept_identity_map_addr = ident_addr;
to_kvm_vmx       6655 arch/x86/kvm/vmx/vmx.c 	vfree(to_kvm_vmx(kvm));
to_kvm_vmx       6815 arch/x86/kvm/vmx/vmx.c 	spin_lock_init(&to_kvm_vmx(kvm)->ept_pointer_lock);