EFER_LME 4260 arch/x86/kvm/emulate.c if ((new_val & X86_CR0_PG) && (efer & EFER_LME) && EFER_LME 743 arch/x86/kvm/svm.c efer &= ~EFER_LME; EFER_LME 2604 arch/x86/kvm/svm.c if (vcpu->arch.efer & EFER_LME) { EFER_LME 2607 arch/x86/kvm/svm.c svm->vmcb->save.efer |= EFER_LMA | EFER_LME; EFER_LME 2612 arch/x86/kvm/svm.c svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME); EFER_LME 1967 arch/x86/kvm/vmx/nested.c return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); EFER_LME 1969 arch/x86/kvm/vmx/nested.c return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); EFER_LME 2727 arch/x86/kvm/vmx/nested.c CC(ia32e != !!(vmcs12->host_ia32_efer & EFER_LME))) EFER_LME 2808 arch/x86/kvm/vmx/nested.c ia32e != !!(vmcs12->guest_ia32_efer & EFER_LME)))) EFER_LME 3805 arch/x86/kvm/vmx/nested.c vcpu->arch.efer |= (EFER_LMA | EFER_LME); EFER_LME 3807 arch/x86/kvm/vmx/nested.c vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); EFER_LME 958 arch/x86/kvm/vmx/vmx.c ignore_bits |= EFER_LMA | EFER_LME; EFER_LME 972 arch/x86/kvm/vmx/vmx.c guest_efer &= ~EFER_LME; EFER_LME 2780 arch/x86/kvm/vmx/vmx.c msr->data = efer & ~EFER_LME; EFER_LME 2927 arch/x86/kvm/vmx/vmx.c if (vcpu->arch.efer & EFER_LME) { EFER_LME 90 arch/x86/kvm/x86.c u64 __read_mostly efer_reserved_bits = ~((u64)(EFER_SCE | EFER_LME | EFER_LMA)); EFER_LME 783 arch/x86/kvm/x86.c if ((vcpu->arch.efer & EFER_LME)) { EFER_LME 1439 arch/x86/kvm/x86.c if (efer & (EFER_LME | EFER_LMA) && EFER_LME 1471 arch/x86/kvm/x86.c (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME)) EFER_LME 8843 arch/x86/kvm/x86.c if ((sregs->efer & EFER_LME) && (sregs->cr0 & X86_CR0_PG)) { EFER_LME 627 tools/testing/selftests/kvm/lib/x86_64/processor.c sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX);