is_guest_mode 81 arch/x86/kvm/irq.c if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) is_guest_mode 4205 arch/x86/kvm/mmu.c if (is_guest_mode(vcpu) && !kvm_is_visible_gfn(vcpu->kvm, gfn)) { is_guest_mode 4993 arch/x86/kvm/mmu.c role.base.guest_mode = is_guest_mode(vcpu); is_guest_mode 5590 arch/x86/kvm/mmu.c if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu)) is_guest_mode 508 arch/x86/kvm/svm.c if (!is_guest_mode(&svm->vcpu)) is_guest_mode 523 arch/x86/kvm/svm.c if (is_guest_mode(&svm->vcpu)) is_guest_mode 1048 arch/x86/kvm/svm.c msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm: is_guest_mode 1494 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) is_guest_mode 1505 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) { is_guest_mode 3070 arch/x86/kvm/svm.c if (!is_guest_mode(&svm->vcpu)) is_guest_mode 3099 arch/x86/kvm/svm.c if (!is_guest_mode(&svm->vcpu)) is_guest_mode 3138 arch/x86/kvm/svm.c if (!is_guest_mode(&svm->vcpu)) is_guest_mode 4012 arch/x86/kvm/svm.c if (!is_guest_mode(&svm->vcpu) || is_guest_mode 4356 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) is_guest_mode 4996 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) { is_guest_mode 5134 arch/x86/kvm/svm.c return is_guest_mode(vcpu) && (vcpu->arch.hflags & HF_VINTR_MASK); is_guest_mode 5361 arch/x86/kvm/svm.c pi.is_guest_mode = true; is_guest_mode 5372 arch/x86/kvm/svm.c if (!ret && pi.is_guest_mode) is_guest_mode 5383 arch/x86/kvm/svm.c pi.is_guest_mode = false; is_guest_mode 5464 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) is_guest_mode 6263 arch/x86/kvm/svm.c if (is_guest_mode(&svm->vcpu) && is_guest_mode 6279 arch/x86/kvm/svm.c if (is_guest_mode(vcpu)) { is_guest_mode 426 arch/x86/kvm/vmx/nested.c WARN_ON(!is_guest_mode(vcpu)); is_guest_mode 4608 arch/x86/kvm/vmx/nested.c struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) is_guest_mode 4621 arch/x86/kvm/vmx/nested.c (is_guest_mode(vcpu) && is_guest_mode 4633 arch/x86/kvm/vmx/nested.c if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) is_guest_mode 4703 arch/x86/kvm/vmx/nested.c struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) is_guest_mode 4715 arch/x86/kvm/vmx/nested.c (is_guest_mode(vcpu) && is_guest_mode 4754 arch/x86/kvm/vmx/nested.c if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) is_guest_mode 4776 arch/x86/kvm/vmx/nested.c if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) { is_guest_mode 5092 arch/x86/kvm/vmx/nested.c if (!is_guest_mode(vcpu)) { is_guest_mode 5539 arch/x86/kvm/vmx/nested.c if (is_guest_mode(vcpu) && is_guest_mode 5551 arch/x86/kvm/vmx/nested.c if (is_guest_mode(vcpu)) { is_guest_mode 5575 arch/x86/kvm/vmx/nested.c if (is_guest_mode(vcpu)) { is_guest_mode 5611 arch/x86/kvm/vmx/nested.c if (is_guest_mode(vcpu)) { is_guest_mode 58 arch/x86/kvm/vmx/nested.h return is_guest_mode(vcpu) || vmx->nested.current_vmptr != -1ull || is_guest_mode 779 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) is_guest_mode 1683 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && is_guest_mode 1701 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && is_guest_mode 1917 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) is_guest_mode 1922 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) is_guest_mode 1927 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) is_guest_mode 1932 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls & is_guest_mode 2020 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && is_guest_mode 2949 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && nested_cpu_has_ept(get_vmcs12(vcpu))) is_guest_mode 2963 arch/x86/kvm/vmx/vmx.c (!is_guest_mode(vcpu) || nested_ept_ad_enabled(vcpu))) is_guest_mode 2991 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) is_guest_mode 3026 arch/x86/kvm/vmx/vmx.c } else if (!is_guest_mode(vcpu) || is_guest_mode 3747 arch/x86/kvm/vmx/vmx.c if (WARN_ON_ONCE(!is_guest_mode(vcpu)) || is_guest_mode 3804 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && is_guest_mode 3919 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(&vmx->vcpu)) is_guest_mode 4481 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && nested_exit_on_intr(vcpu)) is_guest_mode 4747 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) { is_guest_mode 4780 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) { is_guest_mode 5200 arch/x86/kvm/vmx/vmx.c if (!is_guest_mode(vcpu) && is_guest_mode 5866 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && nested_vmx_exit_reflected(vcpu, exit_reason)) is_guest_mode 6020 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && is_guest_mode 6045 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) { is_guest_mode 6079 arch/x86/kvm/vmx/vmx.c if (!is_guest_mode(vcpu)) { is_guest_mode 6129 arch/x86/kvm/vmx/vmx.c if (!is_guest_mode(vcpu)) is_guest_mode 6158 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu) && max_irr_updated) { is_guest_mode 7281 arch/x86/kvm/vmx/vmx.c if (is_guest_mode(vcpu)) { is_guest_mode 7565 arch/x86/kvm/vmx/vmx.c vmx->nested.smm.guest_mode = is_guest_mode(vcpu); is_guest_mode 520 arch/x86/kvm/x86.c !is_guest_mode(vcpu)) is_guest_mode 595 arch/x86/kvm/x86.c is_guest_mode(vcpu) && fault->async_page_fault; is_guest_mode 6424 arch/x86/kvm/x86.c if (!is_guest_mode(vcpu) && kvm_x86_ops->get_cpl(vcpu) == 0) { is_guest_mode 6444 arch/x86/kvm/x86.c if (WARN_ON_ONCE(is_guest_mode(vcpu))) is_guest_mode 6535 arch/x86/kvm/x86.c if (WARN_ON_ONCE(is_guest_mode(vcpu))) is_guest_mode 7607 arch/x86/kvm/x86.c if (is_guest_mode(vcpu) && kvm_x86_ops->check_nested_events) { is_guest_mode 7669 arch/x86/kvm/x86.c if (is_guest_mode(vcpu) && kvm_x86_ops->check_nested_events) { is_guest_mode 7963 arch/x86/kvm/x86.c if (is_guest_mode(vcpu)) is_guest_mode 8376 arch/x86/kvm/x86.c if (is_guest_mode(vcpu) && kvm_x86_ops->check_nested_events) is_guest_mode 9947 arch/x86/kvm/x86.c return (is_guest_mode(vcpu) && is_guest_mode 10159 arch/x86/kvm/x86.c if (!vcpu->arch.apf.delivery_as_pf_vmexit && is_guest_mode(vcpu)) is_guest_mode 4482 drivers/iommu/amd_iommu.c pi_data->is_guest_mode = false; is_guest_mode 4490 drivers/iommu/amd_iommu.c if (pi_data->is_guest_mode) { is_guest_mode 22 include/linux/amd-iommu.h bool is_guest_mode;