Lines Matching refs:efer
1795 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer()
1829 (enable_ept && ((vmx->vcpu.arch.efer ^ host_efer) & EFER_NX))) { in update_transition_efer()
2379 if ((index >= 0) && (vmx->vcpu.arch.efer & EFER_SCE)) in setup_msrs()
3554 static void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument
3567 vcpu->arch.efer = efer; in vmx_set_efer()
3568 if (efer & EFER_LMA) { in vmx_set_efer()
3570 msr->data = efer; in vmx_set_efer()
3574 msr->data = efer & ~EFER_LME; in vmx_set_efer()
3595 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode()
3601 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode()
3726 if (vcpu->arch.efer & EFER_LME) { in vmx_set_cr0()
7939 u64 efer = vmcs_readl(GUEST_IA32_EFER); in dump_vmcs() local
7953 (cr4 & X86_CR4_PAE) && !(efer & EFER_LMA)) in dump_vmcs()
7980 efer, vmcs_readl(GUEST_IA32_PAT)); in dump_vmcs()
9749 vcpu->arch.efer = vmcs12->guest_ia32_efer; in prepare_vmcs02()
9751 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in prepare_vmcs02()
9753 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in prepare_vmcs02()
9755 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02()
10232 vmcs12->guest_ia32_efer = vcpu->arch.efer; in prepare_vmcs12()
10292 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state()
10294 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state()
10296 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state()
10297 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()