init_event       1028 arch/x86/include/asm/kvm_host.h 	void (*vcpu_reset)(struct kvm_vcpu *vcpu, bool init_event);
init_event       1564 arch/x86/include/asm/kvm_host.h void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
init_event       2155 arch/x86/kvm/lapic.c void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event)
init_event       2166 arch/x86/kvm/lapic.c 	if (!init_event) {
init_event         74 arch/x86/kvm/lapic.h void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event);
init_event       2138 arch/x86/kvm/svm.c static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
init_event       2148 arch/x86/kvm/svm.c 	if (!init_event) {
init_event       2159 arch/x86/kvm/svm.c 	if (kvm_vcpu_apicv_active(vcpu) && !init_event)
init_event       4254 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
init_event       4270 arch/x86/kvm/vmx/vmx.c 	if (!init_event) {
init_event       4301 arch/x86/kvm/vmx/vmx.c 	if (!init_event) {
init_event       4327 arch/x86/kvm/vmx/vmx.c 	if (cpu_has_vmx_tpr_shadow() && !init_event) {
init_event       4349 arch/x86/kvm/vmx/vmx.c 	if (init_event)
init_event       9212 arch/x86/kvm/x86.c void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
init_event       9214 arch/x86/kvm/x86.c 	kvm_lapic_reset(vcpu, init_event);
init_event       9253 arch/x86/kvm/x86.c 		if (init_event)
init_event       9263 arch/x86/kvm/x86.c 		if (init_event)
init_event       9267 arch/x86/kvm/x86.c 	if (!init_event) {
init_event       9282 arch/x86/kvm/x86.c 	kvm_x86_ops->vcpu_reset(vcpu, init_event);