to_kvm_vmx 470 arch/x86/kvm/vmx/vmx.c to_kvm_vmx(kvm)->ept_pointers_match to_kvm_vmx 476 arch/x86/kvm/vmx/vmx.c to_kvm_vmx(kvm)->ept_pointers_match = EPT_POINTERS_MATCH; to_kvm_vmx 511 arch/x86/kvm/vmx/vmx.c spin_lock(&to_kvm_vmx(kvm)->ept_pointer_lock); to_kvm_vmx 513 arch/x86/kvm/vmx/vmx.c if (to_kvm_vmx(kvm)->ept_pointers_match == EPT_POINTERS_CHECK) to_kvm_vmx 516 arch/x86/kvm/vmx/vmx.c if (to_kvm_vmx(kvm)->ept_pointers_match != EPT_POINTERS_MATCH) { to_kvm_vmx 528 arch/x86/kvm/vmx/vmx.c spin_unlock(&to_kvm_vmx(kvm)->ept_pointer_lock); to_kvm_vmx 2720 arch/x86/kvm/vmx/vmx.c struct kvm_vmx *kvm_vmx = to_kvm_vmx(vcpu->kvm); to_kvm_vmx 2983 arch/x86/kvm/vmx/vmx.c spin_lock(&to_kvm_vmx(kvm)->ept_pointer_lock); to_kvm_vmx 2985 arch/x86/kvm/vmx/vmx.c to_kvm_vmx(kvm)->ept_pointers_match to_kvm_vmx 2987 arch/x86/kvm/vmx/vmx.c spin_unlock(&to_kvm_vmx(kvm)->ept_pointer_lock); to_kvm_vmx 2996 arch/x86/kvm/vmx/vmx.c guest_cr3 = to_kvm_vmx(kvm)->ept_identity_map_addr; to_kvm_vmx 3433 arch/x86/kvm/vmx/vmx.c fn = to_kvm_vmx(kvm)->tss_addr >> PAGE_SHIFT; to_kvm_vmx 3459 arch/x86/kvm/vmx/vmx.c struct kvm_vmx *kvm_vmx = to_kvm_vmx(kvm); to_kvm_vmx 4500 arch/x86/kvm/vmx/vmx.c to_kvm_vmx(kvm)->tss_addr = addr; to_kvm_vmx 4506 arch/x86/kvm/vmx/vmx.c to_kvm_vmx(kvm)->ept_identity_map_addr = ident_addr; to_kvm_vmx 6655 arch/x86/kvm/vmx/vmx.c vfree(to_kvm_vmx(kvm)); to_kvm_vmx 6815 arch/x86/kvm/vmx/vmx.c spin_lock_init(&to_kvm_vmx(kvm)->ept_pointer_lock);