emul_to_vcpu     5422 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5466 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5478 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5519 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5751 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5837 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5939 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5963 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     5977 arch/x86/kvm/x86.c 	kvm_mmu_invlpg(emul_to_vcpu(ctxt), address);
emul_to_vcpu     6009 arch/x86/kvm/x86.c 	kvm_emulate_wbinvd_noskip(emul_to_vcpu(ctxt));
emul_to_vcpu     6015 arch/x86/kvm/x86.c 	return kvm_get_dr(emul_to_vcpu(ctxt), dr, dest);
emul_to_vcpu     6022 arch/x86/kvm/x86.c 	return __kvm_set_dr(emul_to_vcpu(ctxt), dr, value);
emul_to_vcpu     6032 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     6061 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     6090 arch/x86/kvm/x86.c 	return kvm_x86_ops->get_cpl(emul_to_vcpu(ctxt));
emul_to_vcpu     6095 arch/x86/kvm/x86.c 	kvm_x86_ops->get_gdt(emul_to_vcpu(ctxt), dt);
emul_to_vcpu     6100 arch/x86/kvm/x86.c 	kvm_x86_ops->get_idt(emul_to_vcpu(ctxt), dt);
emul_to_vcpu     6105 arch/x86/kvm/x86.c 	kvm_x86_ops->set_gdt(emul_to_vcpu(ctxt), dt);
emul_to_vcpu     6110 arch/x86/kvm/x86.c 	kvm_x86_ops->set_idt(emul_to_vcpu(ctxt), dt);
emul_to_vcpu     6116 arch/x86/kvm/x86.c 	return get_segment_base(emul_to_vcpu(ctxt), seg);
emul_to_vcpu     6125 arch/x86/kvm/x86.c 	kvm_get_segment(emul_to_vcpu(ctxt), &var, seg);
emul_to_vcpu     6159 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     6188 arch/x86/kvm/x86.c 	return kvm_get_msr(emul_to_vcpu(ctxt), msr_index, pdata);
emul_to_vcpu     6194 arch/x86/kvm/x86.c 	return kvm_set_msr(emul_to_vcpu(ctxt), msr_index, data);
emul_to_vcpu     6199 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     6206 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     6214 arch/x86/kvm/x86.c 	return kvm_pmu_is_valid_msr_idx(emul_to_vcpu(ctxt), pmc);
emul_to_vcpu     6220 arch/x86/kvm/x86.c 	return kvm_pmu_rdpmc(emul_to_vcpu(ctxt), pmc, pdata);
emul_to_vcpu     6225 arch/x86/kvm/x86.c 	emul_to_vcpu(ctxt)->arch.halt_request = 1;
emul_to_vcpu     6232 arch/x86/kvm/x86.c 	return kvm_x86_ops->check_intercept(emul_to_vcpu(ctxt), info, stage);
emul_to_vcpu     6238 arch/x86/kvm/x86.c 	return kvm_cpuid(emul_to_vcpu(ctxt), eax, ebx, ecx, edx, check_limit);
emul_to_vcpu     6243 arch/x86/kvm/x86.c 	return kvm_register_read(emul_to_vcpu(ctxt), reg);
emul_to_vcpu     6248 arch/x86/kvm/x86.c 	kvm_register_write(emul_to_vcpu(ctxt), reg, val);
emul_to_vcpu     6253 arch/x86/kvm/x86.c 	kvm_x86_ops->set_nmi_mask(emul_to_vcpu(ctxt), masked);
emul_to_vcpu     6258 arch/x86/kvm/x86.c 	return emul_to_vcpu(ctxt)->arch.hflags;
emul_to_vcpu     6263 arch/x86/kvm/x86.c 	emul_to_vcpu(ctxt)->arch.hflags = emul_flags;
emul_to_vcpu     6269 arch/x86/kvm/x86.c 	return kvm_x86_ops->pre_leave_smm(emul_to_vcpu(ctxt), smstate);
emul_to_vcpu     6274 arch/x86/kvm/x86.c 	kvm_smm_changed(emul_to_vcpu(ctxt));
emul_to_vcpu     6279 arch/x86/kvm/x86.c 	return __kvm_set_xcr(emul_to_vcpu(ctxt), index, xcr);
emul_to_vcpu     6511 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
emul_to_vcpu     7517 arch/x86/kvm/x86.c 	struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);