kvm_read_cr3     4454 arch/x86/kvm/mmu.c 	return kvm_read_cr3(vcpu);
kvm_read_cr3       95 arch/x86/kvm/mmu.h 	return kvm_get_pcid(vcpu, kvm_read_cr3(vcpu));
kvm_read_cr3     2416 arch/x86/kvm/svm.c 		load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu));
kvm_read_cr3     3389 arch/x86/kvm/svm.c 	nested_vmcb->save.cr3    = kvm_read_cr3(&svm->vcpu);
kvm_read_cr3     3713 arch/x86/kvm/svm.c 		hsave->save.cr3    = kvm_read_cr3(&svm->vcpu);
kvm_read_cr3     4082 arch/x86/kvm/svm.c 			val = kvm_read_cr3(&svm->vcpu);
kvm_read_cr3     5889 arch/x86/kvm/svm.c 	svm->vmcb->save.cr3 = kvm_read_cr3(vcpu);
kvm_read_cr3      993 arch/x86/kvm/vmx/nested.c 	if (cr3 != kvm_read_cr3(vcpu) || (!nested_ept && pdptrs_changed(vcpu))) {
kvm_read_cr3     2994 arch/x86/kvm/vmx/vmx.c 			guest_cr3 = kvm_read_cr3(vcpu);
kvm_read_cr3     4855 arch/x86/kvm/vmx/vmx.c 			val = kvm_read_cr3(vcpu);
kvm_read_cr3      748 arch/x86/kvm/x86.c 	gfn = (kvm_read_cr3(vcpu) & 0xffffffe0ul) >> PAGE_SHIFT;
kvm_read_cr3      749 arch/x86/kvm/x86.c 	offset = (kvm_read_cr3(vcpu) & 0xffffffe0ul) & (PAGE_SIZE - 1);
kvm_read_cr3      794 arch/x86/kvm/x86.c 						 kvm_read_cr3(vcpu)))
kvm_read_cr3      986 arch/x86/kvm/x86.c 				   kvm_read_cr3(vcpu)))
kvm_read_cr3      994 arch/x86/kvm/x86.c 		if ((kvm_read_cr3(vcpu) & X86_CR3_PCID_MASK) || !is_long_mode(vcpu))
kvm_read_cr3     1024 arch/x86/kvm/x86.c 	if (cr3 == kvm_read_cr3(vcpu) && !pdptrs_changed(vcpu)) {
kvm_read_cr3     6043 arch/x86/kvm/x86.c 		value = kvm_read_cr3(vcpu);
kvm_read_cr3     7759 arch/x86/kvm/x86.c 	put_smstate(u32, buf, 0x7ff8, kvm_read_cr3(vcpu));
kvm_read_cr3     7821 arch/x86/kvm/x86.c 	put_smstate(u64, buf, 0x7f50, kvm_read_cr3(vcpu));
kvm_read_cr3     8745 arch/x86/kvm/x86.c 	sregs->cr3 = kvm_read_cr3(vcpu);
kvm_read_cr3     8889 arch/x86/kvm/x86.c 	mmu_reset_needed |= kvm_read_cr3(vcpu) != sregs->cr3;
kvm_read_cr3     8911 arch/x86/kvm/x86.c 		load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu));