accessed_dirty   5113 arch/x86/kvm/mmu.c kvm_calc_shadow_ept_root_page_role(struct kvm_vcpu *vcpu, bool accessed_dirty,
accessed_dirty   5124 arch/x86/kvm/mmu.c 	role.base.ad_disabled = !accessed_dirty;
accessed_dirty   5142 arch/x86/kvm/mmu.c 			     bool accessed_dirty, gpa_t new_eptp)
accessed_dirty   5146 arch/x86/kvm/mmu.c 		kvm_calc_shadow_ept_root_page_role(vcpu, accessed_dirty,
accessed_dirty   5158 arch/x86/kvm/mmu.c 	context->ept_ad = accessed_dirty;
accessed_dirty     62 arch/x86/kvm/mmu.h 			     bool accessed_dirty, gpa_t new_eptp);
accessed_dirty    305 arch/x86/kvm/paging_tmpl.h 	unsigned index, accessed_dirty, pte_pkey;
accessed_dirty    412 arch/x86/kvm/paging_tmpl.h 	accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0;
accessed_dirty    441 arch/x86/kvm/paging_tmpl.h 		accessed_dirty &= pte >>
accessed_dirty    444 arch/x86/kvm/paging_tmpl.h 	if (unlikely(!accessed_dirty)) {
accessed_dirty   5041 arch/x86/kvm/vmx/nested.c 	bool accessed_dirty;
accessed_dirty   5056 arch/x86/kvm/vmx/nested.c 	accessed_dirty = !!(address & VMX_EPTP_AD_ENABLE_BIT);
accessed_dirty   5067 arch/x86/kvm/vmx/nested.c 		mmu->ept_ad = accessed_dirty;
accessed_dirty   5068 arch/x86/kvm/vmx/nested.c 		mmu->mmu_role.base.ad_disabled = !accessed_dirty;