accessed_dirty 5113 arch/x86/kvm/mmu.c kvm_calc_shadow_ept_root_page_role(struct kvm_vcpu *vcpu, bool accessed_dirty, accessed_dirty 5124 arch/x86/kvm/mmu.c role.base.ad_disabled = !accessed_dirty; accessed_dirty 5142 arch/x86/kvm/mmu.c bool accessed_dirty, gpa_t new_eptp) accessed_dirty 5146 arch/x86/kvm/mmu.c kvm_calc_shadow_ept_root_page_role(vcpu, accessed_dirty, accessed_dirty 5158 arch/x86/kvm/mmu.c context->ept_ad = accessed_dirty; accessed_dirty 62 arch/x86/kvm/mmu.h bool accessed_dirty, gpa_t new_eptp); accessed_dirty 305 arch/x86/kvm/paging_tmpl.h unsigned index, accessed_dirty, pte_pkey; accessed_dirty 412 arch/x86/kvm/paging_tmpl.h accessed_dirty = have_ad ? pte_access & PT_GUEST_ACCESSED_MASK : 0; accessed_dirty 441 arch/x86/kvm/paging_tmpl.h accessed_dirty &= pte >> accessed_dirty 444 arch/x86/kvm/paging_tmpl.h if (unlikely(!accessed_dirty)) { accessed_dirty 5041 arch/x86/kvm/vmx/nested.c bool accessed_dirty; accessed_dirty 5056 arch/x86/kvm/vmx/nested.c accessed_dirty = !!(address & VMX_EPTP_AD_ENABLE_BIT); accessed_dirty 5067 arch/x86/kvm/vmx/nested.c mmu->ept_ad = accessed_dirty; accessed_dirty 5068 arch/x86/kvm/vmx/nested.c mmu->mmu_role.base.ad_disabled = !accessed_dirty;