kvm_read_cr3 4454 arch/x86/kvm/mmu.c return kvm_read_cr3(vcpu); kvm_read_cr3 95 arch/x86/kvm/mmu.h return kvm_get_pcid(vcpu, kvm_read_cr3(vcpu)); kvm_read_cr3 2416 arch/x86/kvm/svm.c load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu)); kvm_read_cr3 3389 arch/x86/kvm/svm.c nested_vmcb->save.cr3 = kvm_read_cr3(&svm->vcpu); kvm_read_cr3 3713 arch/x86/kvm/svm.c hsave->save.cr3 = kvm_read_cr3(&svm->vcpu); kvm_read_cr3 4082 arch/x86/kvm/svm.c val = kvm_read_cr3(&svm->vcpu); kvm_read_cr3 5889 arch/x86/kvm/svm.c svm->vmcb->save.cr3 = kvm_read_cr3(vcpu); kvm_read_cr3 993 arch/x86/kvm/vmx/nested.c if (cr3 != kvm_read_cr3(vcpu) || (!nested_ept && pdptrs_changed(vcpu))) { kvm_read_cr3 2994 arch/x86/kvm/vmx/vmx.c guest_cr3 = kvm_read_cr3(vcpu); kvm_read_cr3 4855 arch/x86/kvm/vmx/vmx.c val = kvm_read_cr3(vcpu); kvm_read_cr3 748 arch/x86/kvm/x86.c gfn = (kvm_read_cr3(vcpu) & 0xffffffe0ul) >> PAGE_SHIFT; kvm_read_cr3 749 arch/x86/kvm/x86.c offset = (kvm_read_cr3(vcpu) & 0xffffffe0ul) & (PAGE_SIZE - 1); kvm_read_cr3 794 arch/x86/kvm/x86.c kvm_read_cr3(vcpu))) kvm_read_cr3 986 arch/x86/kvm/x86.c kvm_read_cr3(vcpu))) kvm_read_cr3 994 arch/x86/kvm/x86.c if ((kvm_read_cr3(vcpu) & X86_CR3_PCID_MASK) || !is_long_mode(vcpu)) kvm_read_cr3 1024 arch/x86/kvm/x86.c if (cr3 == kvm_read_cr3(vcpu) && !pdptrs_changed(vcpu)) { kvm_read_cr3 6043 arch/x86/kvm/x86.c value = kvm_read_cr3(vcpu); kvm_read_cr3 7759 arch/x86/kvm/x86.c put_smstate(u32, buf, 0x7ff8, kvm_read_cr3(vcpu)); kvm_read_cr3 7821 arch/x86/kvm/x86.c put_smstate(u64, buf, 0x7f50, kvm_read_cr3(vcpu)); kvm_read_cr3 8745 arch/x86/kvm/x86.c sregs->cr3 = kvm_read_cr3(vcpu); kvm_read_cr3 8889 arch/x86/kvm/x86.c mmu_reset_needed |= kvm_read_cr3(vcpu) != sregs->cr3; kvm_read_cr3 8911 arch/x86/kvm/x86.c load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu));