Lines Matching defs:spte
161 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte) \ argument
215 static unsigned int get_mmio_spte_generation(u64 spte) in get_mmio_spte_generation()
244 static bool is_mmio_spte(u64 spte) in is_mmio_spte()
249 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn()
255 static unsigned get_mmio_spte_access(u64 spte) in get_mmio_spte_access()
272 static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte) in check_mmio_spte()
341 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
346 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
351 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
366 u64 spte; member
369 static void count_spte_clear(u64 *sptep, u64 spte) in count_spte_clear()
381 static void __set_spte(u64 *sptep, u64 spte) in __set_spte()
400 static void __update_clear_spte_fast(u64 *sptep, u64 spte) in __update_clear_spte_fast()
419 static u64 __update_clear_spte_slow(u64 *sptep, u64 spte) in __update_clear_spte_slow()
456 union split_spte spte, *orig = (union split_spte *)sptep; in __get_spte_lockless() local
477 static bool spte_is_locklessly_modifiable(u64 spte) in spte_is_locklessly_modifiable()
483 static bool spte_has_volatile_bits(u64 spte) in spte_has_volatile_bits()
923 static int pte_list_add(struct kvm_vcpu *vcpu, u64 *spte, in pte_list_add()
979 static void pte_list_remove(u64 *spte, unsigned long *pte_list) in pte_list_remove()
1065 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
1076 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1195 u64 spte = *sptep; in spte_write_protect() local
1225 u64 spte = *sptep; in spte_clear_dirty() local
1248 u64 spte = *sptep; in spte_set_dirty() local
1614 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1752 static void mark_unsync(u64 *spte) in mark_unsync()
1777 struct kvm_mmu_page *sp, u64 *spte, in nonpaging_update_pte()
2069 static void clear_sp_write_flooding_count(u64 *spte) in clear_sp_write_flooding_count()
2185 u64 spte) in __shadow_walk_next()
2203 u64 spte; in link_shadow_page() local
2240 u64 *spte) in mmu_page_zap_pte()
2492 u64 spte; in set_spte() local
2671 u64 *spte, *start = NULL; in __direct_pte_prefetch() local
2864 u64 *sptep, u64 spte) in fast_pf_fix_direct_spte()
2905 u64 spte = 0ull; in fast_page_fault() local
3303 static bool is_shadow_zero_bits_set(struct kvm_mmu *mmu, u64 spte, int level) in is_shadow_zero_bits_set()
3321 u64 sptes[PT64_ROOT_LEVEL], spte = 0ull; in walk_shadow_page_get_mmio_spte() local
3364 u64 spte; in handle_mmio_page_fault() local
4136 struct kvm_mmu_page *sp, u64 *spte, in mmu_pte_write_new_pte()
4256 u64 *spte; in get_written_sptes() local
4290 u64 entry, gentry, *spte; in kvm_mmu_pte_write() local