Lines Matching refs:old_spte
507 static bool spte_is_bit_cleared(u64 old_spte, u64 new_spte, u64 bit_mask) in spte_is_bit_cleared() argument
509 return (old_spte & bit_mask) && !(new_spte & bit_mask); in spte_is_bit_cleared()
512 static bool spte_is_bit_changed(u64 old_spte, u64 new_spte, u64 bit_mask) in spte_is_bit_changed() argument
514 return (old_spte & bit_mask) != (new_spte & bit_mask); in spte_is_bit_changed()
540 u64 old_spte = *sptep; in mmu_spte_update() local
545 if (!is_shadow_present_pte(old_spte)) { in mmu_spte_update()
550 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_update()
553 old_spte = __update_clear_spte_slow(sptep, new_spte); in mmu_spte_update()
560 if (spte_is_locklessly_modifiable(old_spte) && in mmu_spte_update()
571 if (spte_is_bit_changed(old_spte, new_spte, in mmu_spte_update()
575 if (spte_is_bit_cleared(old_spte, new_spte, shadow_accessed_mask)) in mmu_spte_update()
576 kvm_set_pfn_accessed(spte_to_pfn(old_spte)); in mmu_spte_update()
577 if (spte_is_bit_cleared(old_spte, new_spte, shadow_dirty_mask)) in mmu_spte_update()
578 kvm_set_pfn_dirty(spte_to_pfn(old_spte)); in mmu_spte_update()
591 u64 old_spte = *sptep; in mmu_spte_clear_track_bits() local
593 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_clear_track_bits()
596 old_spte = __update_clear_spte_slow(sptep, 0ull); in mmu_spte_clear_track_bits()
598 if (!is_rmap_spte(old_spte)) in mmu_spte_clear_track_bits()
601 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits()
610 if (!shadow_accessed_mask || old_spte & shadow_accessed_mask) in mmu_spte_clear_track_bits()
612 if (!shadow_dirty_mask || (old_spte & shadow_dirty_mask)) in mmu_spte_clear_track_bits()