EFER_LME         4260 arch/x86/kvm/emulate.c 		if ((new_val & X86_CR0_PG) && (efer & EFER_LME) &&
EFER_LME          743 arch/x86/kvm/svm.c 			efer &= ~EFER_LME;
EFER_LME         2604 arch/x86/kvm/svm.c 	if (vcpu->arch.efer & EFER_LME) {
EFER_LME         2607 arch/x86/kvm/svm.c 			svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
EFER_LME         2612 arch/x86/kvm/svm.c 			svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
EFER_LME         1967 arch/x86/kvm/vmx/nested.c 		return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME);
EFER_LME         1969 arch/x86/kvm/vmx/nested.c 		return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME);
EFER_LME         2727 arch/x86/kvm/vmx/nested.c 		    CC(ia32e != !!(vmcs12->host_ia32_efer & EFER_LME)))
EFER_LME         2808 arch/x86/kvm/vmx/nested.c 		     ia32e != !!(vmcs12->guest_ia32_efer & EFER_LME))))
EFER_LME         3805 arch/x86/kvm/vmx/nested.c 		vcpu->arch.efer |= (EFER_LMA | EFER_LME);
EFER_LME         3807 arch/x86/kvm/vmx/nested.c 		vcpu->arch.efer &= ~(EFER_LMA | EFER_LME);
EFER_LME          958 arch/x86/kvm/vmx/vmx.c 	ignore_bits |= EFER_LMA | EFER_LME;
EFER_LME          972 arch/x86/kvm/vmx/vmx.c 			guest_efer &= ~EFER_LME;
EFER_LME         2780 arch/x86/kvm/vmx/vmx.c 		msr->data = efer & ~EFER_LME;
EFER_LME         2927 arch/x86/kvm/vmx/vmx.c 	if (vcpu->arch.efer & EFER_LME) {
EFER_LME           90 arch/x86/kvm/x86.c u64 __read_mostly efer_reserved_bits = ~((u64)(EFER_SCE | EFER_LME | EFER_LMA));
EFER_LME          783 arch/x86/kvm/x86.c 		if ((vcpu->arch.efer & EFER_LME)) {
EFER_LME         1439 arch/x86/kvm/x86.c 	if (efer & (EFER_LME | EFER_LMA) &&
EFER_LME         1471 arch/x86/kvm/x86.c 		    (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME))
EFER_LME         8843 arch/x86/kvm/x86.c 	if ((sregs->efer & EFER_LME) && (sregs->cr0 & X86_CR0_PG)) {
EFER_LME          627 tools/testing/selftests/kvm/lib/x86_64/processor.c 		sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX);