emul_to_vcpu 5422 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5466 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5478 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5519 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5751 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5837 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5939 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5963 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 5977 arch/x86/kvm/x86.c kvm_mmu_invlpg(emul_to_vcpu(ctxt), address); emul_to_vcpu 6009 arch/x86/kvm/x86.c kvm_emulate_wbinvd_noskip(emul_to_vcpu(ctxt)); emul_to_vcpu 6015 arch/x86/kvm/x86.c return kvm_get_dr(emul_to_vcpu(ctxt), dr, dest); emul_to_vcpu 6022 arch/x86/kvm/x86.c return __kvm_set_dr(emul_to_vcpu(ctxt), dr, value); emul_to_vcpu 6032 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 6061 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 6090 arch/x86/kvm/x86.c return kvm_x86_ops->get_cpl(emul_to_vcpu(ctxt)); emul_to_vcpu 6095 arch/x86/kvm/x86.c kvm_x86_ops->get_gdt(emul_to_vcpu(ctxt), dt); emul_to_vcpu 6100 arch/x86/kvm/x86.c kvm_x86_ops->get_idt(emul_to_vcpu(ctxt), dt); emul_to_vcpu 6105 arch/x86/kvm/x86.c kvm_x86_ops->set_gdt(emul_to_vcpu(ctxt), dt); emul_to_vcpu 6110 arch/x86/kvm/x86.c kvm_x86_ops->set_idt(emul_to_vcpu(ctxt), dt); emul_to_vcpu 6116 arch/x86/kvm/x86.c return get_segment_base(emul_to_vcpu(ctxt), seg); emul_to_vcpu 6125 arch/x86/kvm/x86.c kvm_get_segment(emul_to_vcpu(ctxt), &var, seg); emul_to_vcpu 6159 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 6188 arch/x86/kvm/x86.c return kvm_get_msr(emul_to_vcpu(ctxt), msr_index, pdata); emul_to_vcpu 6194 arch/x86/kvm/x86.c return kvm_set_msr(emul_to_vcpu(ctxt), msr_index, data); emul_to_vcpu 6199 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 6206 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 6214 arch/x86/kvm/x86.c return kvm_pmu_is_valid_msr_idx(emul_to_vcpu(ctxt), pmc); emul_to_vcpu 6220 arch/x86/kvm/x86.c return kvm_pmu_rdpmc(emul_to_vcpu(ctxt), pmc, pdata); emul_to_vcpu 6225 arch/x86/kvm/x86.c emul_to_vcpu(ctxt)->arch.halt_request = 1; emul_to_vcpu 6232 arch/x86/kvm/x86.c return kvm_x86_ops->check_intercept(emul_to_vcpu(ctxt), info, stage); emul_to_vcpu 6238 arch/x86/kvm/x86.c return kvm_cpuid(emul_to_vcpu(ctxt), eax, ebx, ecx, edx, check_limit); emul_to_vcpu 6243 arch/x86/kvm/x86.c return kvm_register_read(emul_to_vcpu(ctxt), reg); emul_to_vcpu 6248 arch/x86/kvm/x86.c kvm_register_write(emul_to_vcpu(ctxt), reg, val); emul_to_vcpu 6253 arch/x86/kvm/x86.c kvm_x86_ops->set_nmi_mask(emul_to_vcpu(ctxt), masked); emul_to_vcpu 6258 arch/x86/kvm/x86.c return emul_to_vcpu(ctxt)->arch.hflags; emul_to_vcpu 6263 arch/x86/kvm/x86.c emul_to_vcpu(ctxt)->arch.hflags = emul_flags; emul_to_vcpu 6269 arch/x86/kvm/x86.c return kvm_x86_ops->pre_leave_smm(emul_to_vcpu(ctxt), smstate); emul_to_vcpu 6274 arch/x86/kvm/x86.c kvm_smm_changed(emul_to_vcpu(ctxt)); emul_to_vcpu 6279 arch/x86/kvm/x86.c return __kvm_set_xcr(emul_to_vcpu(ctxt), index, xcr); emul_to_vcpu 6511 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); emul_to_vcpu 7517 arch/x86/kvm/x86.c struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);