Lines Matching refs:efer
746 u64 efer = 0; in assign_eip_far() local
748 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in assign_eip_far()
749 if (efer & EFER_LMA) in assign_eip_far()
1493 u64 efer = 0; in get_descriptor_ptr() local
1495 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr()
1496 if (!(efer & EFER_LMA)) in get_descriptor_ptr()
1633 u64 efer = 0; in __load_segment_descriptor() local
1635 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor()
1636 if (efer & EFER_LMA) in __load_segment_descriptor()
2478 unsigned long cr0, cr4, efer; in em_rsm() local
2517 efer = 0; in em_rsm()
2518 ctxt->ops->set_msr(ctxt, MSR_EFER, efer); in em_rsm()
2627 u64 efer = 0; in em_syscall() local
2637 ops->get_msr(ctxt, MSR_EFER, &efer); in em_syscall()
2640 if (!(efer & EFER_SCE)) in em_syscall()
2648 if (efer & EFER_LMA) { in em_syscall()
2656 if (efer & EFER_LMA) { in em_syscall()
2686 u64 efer = 0; in em_sysenter() local
2688 ops->get_msr(ctxt, MSR_EFER, &efer); in em_sysenter()
2697 if ((ctxt->mode != X86EMUL_MODE_PROT64) && (efer & EFER_LMA) in em_sysenter()
2714 if (efer & EFER_LMA) { in em_sysenter()
2723 ctxt->_eip = (efer & EFER_LMA) ? msr_data : (u32)msr_data; in em_sysenter()
2726 *reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data : in em_sysenter()
3871 u64 efer = 0; in check_cr_write() local
3895 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
3897 if ((new_val & X86_CR0_PG) && (efer & EFER_LME) && in check_cr_write()
3906 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
3907 if (efer & EFER_LMA) in check_cr_write()
3916 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_cr_write()
3918 if ((efer & EFER_LMA) && !(new_val & X86_CR4_PAE)) in check_cr_write()
3976 u64 efer; in check_svme() local
3978 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in check_svme()
3980 if (!(efer & EFER_SVME)) in check_svme()