nmi_pending 694 arch/x86/include/asm/kvm_host.h unsigned nmi_pending; /* NMI queued after currently running handler */ nmi_pending 3495 arch/x86/kvm/vmx/nested.c if (vcpu->arch.nmi_pending && nested_exit_on_nmi(vcpu)) { nmi_pending 3505 arch/x86/kvm/vmx/nested.c vcpu->arch.nmi_pending = 0; nmi_pending 5916 arch/x86/kvm/vmx/vmx.c vcpu->arch.nmi_pending) { nmi_pending 3804 arch/x86/kvm/x86.c events->nmi.pending = vcpu->arch.nmi_pending != 0; nmi_pending 3877 arch/x86/kvm/x86.c vcpu->arch.nmi_pending = events->nmi.pending; nmi_pending 7657 arch/x86/kvm/x86.c } else if (vcpu->arch.nmi_pending && kvm_x86_ops->nmi_allowed(vcpu)) { nmi_pending 7658 arch/x86/kvm/x86.c --vcpu->arch.nmi_pending; nmi_pending 7696 arch/x86/kvm/x86.c vcpu->arch.nmi_pending += atomic_xchg(&vcpu->arch.nmi_queued, 0); nmi_pending 7697 arch/x86/kvm/x86.c vcpu->arch.nmi_pending = min(vcpu->arch.nmi_pending, limit); nmi_pending 8164 arch/x86/kvm/x86.c if (vcpu->arch.nmi_pending) nmi_pending 9221 arch/x86/kvm/x86.c vcpu->arch.nmi_pending = 0; nmi_pending 9967 arch/x86/kvm/x86.c (vcpu->arch.nmi_pending &&