apf 744 arch/x86/include/asm/kvm_host.h } apf; apf 4186 arch/x86/kvm/mmu.c arch.token = (vcpu->arch.apf.id++ << 12) | vcpu->vcpu_id; apf 4242 arch/x86/kvm/mmu.c switch (vcpu->arch.apf.host_apf_reason) { apf 4252 arch/x86/kvm/mmu.c vcpu->arch.apf.host_apf_reason = 0; apf 4258 arch/x86/kvm/mmu.c vcpu->arch.apf.host_apf_reason = 0; apf 3086 arch/x86/kvm/svm.c svm->vmcb->control.exit_info_2 = svm->vcpu.arch.apf.nested_apf_token; apf 3241 arch/x86/kvm/svm.c if (!npt_enabled || svm->vcpu.arch.apf.host_apf_reason) apf 5854 arch/x86/kvm/svm.c svm->vcpu.arch.apf.host_apf_reason = kvm_read_and_reset_pf_reason(); apf 396 arch/x86/kvm/vmx/nested.c *exit_qual = vcpu->arch.apf.nested_apf_token; apf 5365 arch/x86/kvm/vmx/nested.c return !vmx->vcpu.arch.apf.host_apf_reason && enable_ept; apf 4649 arch/x86/kvm/vmx/vmx.c WARN_ON_ONCE(!vcpu->arch.apf.host_apf_reason && enable_ept); apf 6204 arch/x86/kvm/vmx/vmx.c vmx->vcpu.arch.apf.host_apf_reason = kvm_read_and_reset_pf_reason(); apf 235 arch/x86/kvm/x86.c vcpu->arch.apf.gfns[i] = ~0; apf 597 arch/x86/kvm/x86.c vcpu->arch.apf.nested_apf_token = fault->address; apf 2617 arch/x86/kvm/x86.c vcpu->arch.apf.msr_val = data; apf 2625 arch/x86/kvm/x86.c if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, apf 2629 arch/x86/kvm/x86.c vcpu->arch.apf.send_user_only = !(data & KVM_ASYNC_PF_SEND_ALWAYS); apf 2630 arch/x86/kvm/x86.c vcpu->arch.apf.delivery_as_pf_vmexit = data & KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT; apf 3098 arch/x86/kvm/x86.c msr_info->data = vcpu->arch.apf.msr_val; apf 8077 arch/x86/kvm/x86.c vcpu->arch.apf.halted = true; apf 8363 arch/x86/kvm/x86.c vcpu->arch.apf.halted = false; apf 8380 arch/x86/kvm/x86.c !vcpu->arch.apf.halted); apf 9237 arch/x86/kvm/x86.c vcpu->arch.apf.msr_val = 0; apf 9244 arch/x86/kvm/x86.c vcpu->arch.apf.halted = false; apf 10096 arch/x86/kvm/x86.c while (vcpu->arch.apf.gfns[key] != ~0) apf 10099 arch/x86/kvm/x86.c vcpu->arch.apf.gfns[key] = gfn; apf 10108 arch/x86/kvm/x86.c (vcpu->arch.apf.gfns[key] != gfn && apf 10109 arch/x86/kvm/x86.c vcpu->arch.apf.gfns[key] != ~0); i++) apf 10117 arch/x86/kvm/x86.c return vcpu->arch.apf.gfns[kvm_async_pf_gfn_slot(vcpu, gfn)] == gfn; apf 10126 arch/x86/kvm/x86.c vcpu->arch.apf.gfns[i] = ~0; apf 10129 arch/x86/kvm/x86.c if (vcpu->arch.apf.gfns[j] == ~0) apf 10131 arch/x86/kvm/x86.c k = kvm_async_pf_hash_fn(vcpu->arch.apf.gfns[j]); apf 10138 arch/x86/kvm/x86.c vcpu->arch.apf.gfns[i] = vcpu->arch.apf.gfns[j]; apf 10146 arch/x86/kvm/x86.c return kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.apf.data, &val, apf 10153 arch/x86/kvm/x86.c return kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.apf.data, val, apf 10159 arch/x86/kvm/x86.c if (!vcpu->arch.apf.delivery_as_pf_vmexit && is_guest_mode(vcpu)) apf 10162 arch/x86/kvm/x86.c if (!(vcpu->arch.apf.msr_val & KVM_ASYNC_PF_ENABLED) || apf 10163 arch/x86/kvm/x86.c (vcpu->arch.apf.send_user_only && apf 10229 arch/x86/kvm/x86.c if (vcpu->arch.apf.msr_val & KVM_ASYNC_PF_ENABLED && apf 10252 arch/x86/kvm/x86.c vcpu->arch.apf.halted = false; apf 10258 arch/x86/kvm/x86.c if (!(vcpu->arch.apf.msr_val & KVM_ASYNC_PF_ENABLED)) apf 62 virt/kvm/async_pf.c struct kvm_async_pf *apf = apf 64 virt/kvm/async_pf.c struct mm_struct *mm = apf->mm; apf 65 virt/kvm/async_pf.c struct kvm_vcpu *vcpu = apf->vcpu; apf 66 virt/kvm/async_pf.c unsigned long addr = apf->addr; apf 67 virt/kvm/async_pf.c gpa_t cr2_or_gpa = apf->cr2_or_gpa; apf 83 virt/kvm/async_pf.c kvm_async_page_present_sync(vcpu, apf); apf 86 virt/kvm/async_pf.c list_add_tail(&apf->link, &vcpu->async_pf.done); apf 87 virt/kvm/async_pf.c apf->vcpu = NULL;