nmi_pending       694 arch/x86/include/asm/kvm_host.h 	unsigned nmi_pending; /* NMI queued after currently running handler */
nmi_pending      3495 arch/x86/kvm/vmx/nested.c 	if (vcpu->arch.nmi_pending && nested_exit_on_nmi(vcpu)) {
nmi_pending      3505 arch/x86/kvm/vmx/nested.c 		vcpu->arch.nmi_pending = 0;
nmi_pending      5916 arch/x86/kvm/vmx/vmx.c 			   vcpu->arch.nmi_pending) {
nmi_pending      3804 arch/x86/kvm/x86.c 	events->nmi.pending = vcpu->arch.nmi_pending != 0;
nmi_pending      3877 arch/x86/kvm/x86.c 		vcpu->arch.nmi_pending = events->nmi.pending;
nmi_pending      7657 arch/x86/kvm/x86.c 	} else if (vcpu->arch.nmi_pending && kvm_x86_ops->nmi_allowed(vcpu)) {
nmi_pending      7658 arch/x86/kvm/x86.c 		--vcpu->arch.nmi_pending;
nmi_pending      7696 arch/x86/kvm/x86.c 	vcpu->arch.nmi_pending += atomic_xchg(&vcpu->arch.nmi_queued, 0);
nmi_pending      7697 arch/x86/kvm/x86.c 	vcpu->arch.nmi_pending = min(vcpu->arch.nmi_pending, limit);
nmi_pending      8164 arch/x86/kvm/x86.c 			if (vcpu->arch.nmi_pending)
nmi_pending      9221 arch/x86/kvm/x86.c 	vcpu->arch.nmi_pending = 0;
nmi_pending      9967 arch/x86/kvm/x86.c 	    (vcpu->arch.nmi_pending &&