init_event 1028 arch/x86/include/asm/kvm_host.h void (*vcpu_reset)(struct kvm_vcpu *vcpu, bool init_event); init_event 1564 arch/x86/include/asm/kvm_host.h void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event); init_event 2155 arch/x86/kvm/lapic.c void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event) init_event 2166 arch/x86/kvm/lapic.c if (!init_event) { init_event 74 arch/x86/kvm/lapic.h void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event); init_event 2138 arch/x86/kvm/svm.c static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) init_event 2148 arch/x86/kvm/svm.c if (!init_event) { init_event 2159 arch/x86/kvm/svm.c if (kvm_vcpu_apicv_active(vcpu) && !init_event) init_event 4254 arch/x86/kvm/vmx/vmx.c static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) init_event 4270 arch/x86/kvm/vmx/vmx.c if (!init_event) { init_event 4301 arch/x86/kvm/vmx/vmx.c if (!init_event) { init_event 4327 arch/x86/kvm/vmx/vmx.c if (cpu_has_vmx_tpr_shadow() && !init_event) { init_event 4349 arch/x86/kvm/vmx/vmx.c if (init_event) init_event 9212 arch/x86/kvm/x86.c void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) init_event 9214 arch/x86/kvm/x86.c kvm_lapic_reset(vcpu, init_event); init_event 9253 arch/x86/kvm/x86.c if (init_event) init_event 9263 arch/x86/kvm/x86.c if (init_event) init_event 9267 arch/x86/kvm/x86.c if (!init_event) { init_event 9282 arch/x86/kvm/x86.c kvm_x86_ops->vcpu_reset(vcpu, init_event);