Lines Matching refs:kvm_vcpu
179 static void mmu_free_roots(struct kvm_vcpu *vcpu);
226 static unsigned int kvm_current_mmio_generation(struct kvm_vcpu *vcpu) in kvm_current_mmio_generation()
231 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
261 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte()
272 static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte) in check_mmio_spte()
299 static int is_nx(struct kvm_vcpu *vcpu) in is_nx()
632 static void walk_shadow_page_lockless_begin(struct kvm_vcpu *vcpu) in walk_shadow_page_lockless_begin()
647 static void walk_shadow_page_lockless_end(struct kvm_vcpu *vcpu) in walk_shadow_page_lockless_end()
709 static int mmu_topup_memory_caches(struct kvm_vcpu *vcpu) in mmu_topup_memory_caches()
726 static void mmu_free_memory_caches(struct kvm_vcpu *vcpu) in mmu_free_memory_caches()
744 static struct pte_list_desc *mmu_alloc_pte_list_desc(struct kvm_vcpu *vcpu) in mmu_alloc_pte_list_desc()
834 static int has_wrprotected_page(struct kvm_vcpu *vcpu, gfn_t gfn, int level) in has_wrprotected_page()
871 gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
883 static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn, in mapping_level()
923 static int pte_list_add(struct kvm_vcpu *vcpu, u64 *spte, in pte_list_add()
1057 static bool rmap_can_add(struct kvm_vcpu *vcpu) in rmap_can_add()
1065 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
1174 static void drop_large_spte(struct kvm_vcpu *vcpu, u64 *sptep) in drop_large_spte()
1342 static bool rmap_write_protect(struct kvm_vcpu *vcpu, u64 gfn) in rmap_write_protect()
1614 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1701 static void mmu_page_add_parent_pte(struct kvm_vcpu *vcpu, in mmu_page_add_parent_pte()
1723 static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, in kvm_mmu_alloc_page()
1766 static int nonpaging_sync_page(struct kvm_vcpu *vcpu, in nonpaging_sync_page()
1772 static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva) in nonpaging_invlpg()
1776 static void nonpaging_update_pte(struct kvm_vcpu *vcpu, in nonpaging_update_pte()
1896 static int __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in __kvm_sync_page()
1916 static int kvm_sync_page_transient(struct kvm_vcpu *vcpu, in kvm_sync_page_transient()
1932 static void kvm_mmu_audit(struct kvm_vcpu *vcpu, int point) { } in kvm_mmu_audit()
1936 static int kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in kvm_sync_page()
1943 static void kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_sync_pages()
2027 static void mmu_sync_children(struct kvm_vcpu *vcpu, in mmu_sync_children()
2081 static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, in kvm_mmu_get_page()
2153 struct kvm_vcpu *vcpu, u64 addr) in shadow_walk_init()
2217 static void validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep, in validate_direct_spte()
2437 static void __kvm_unsync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) in __kvm_unsync_page()
2446 static void kvm_unsync_pages(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_unsync_pages()
2458 static int mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, in mmu_need_write_protect()
2487 static int set_spte(struct kvm_vcpu *vcpu, u64 *sptep, in set_spte()
2567 static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, in mmu_set_spte()
2629 static pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in pte_prefetch_gfn_to_pfn()
2641 static int direct_pte_prefetch_many(struct kvm_vcpu *vcpu, in direct_pte_prefetch_many()
2668 static void __direct_pte_prefetch(struct kvm_vcpu *vcpu, in __direct_pte_prefetch()
2691 static void direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep) in direct_pte_prefetch()
2711 static int __direct_map(struct kvm_vcpu *vcpu, gpa_t v, int write, in __direct_map()
2762 static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, pfn_t pfn) in kvm_handle_bad_page()
2781 static void transparent_hugepage_adjust(struct kvm_vcpu *vcpu, in transparent_hugepage_adjust()
2822 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn()
2863 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, in fast_pf_fix_direct_spte()
2899 static bool fast_page_fault(struct kvm_vcpu *vcpu, gva_t gva, int level, in fast_page_fault()
2976 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn,
2978 static void make_mmu_pages_available(struct kvm_vcpu *vcpu);
2980 static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, u32 error_code, in nonpaging_map()
3035 static void mmu_free_roots(struct kvm_vcpu *vcpu) in mmu_free_roots()
3080 static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn) in mmu_check_root()
3092 static int mmu_alloc_direct_roots(struct kvm_vcpu *vcpu) in mmu_alloc_direct_roots()
3128 static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu) in mmu_alloc_shadow_roots()
3224 static int mmu_alloc_roots(struct kvm_vcpu *vcpu) in mmu_alloc_roots()
3232 static void mmu_sync_roots(struct kvm_vcpu *vcpu) in mmu_sync_roots()
3264 void kvm_mmu_sync_roots(struct kvm_vcpu *vcpu) in kvm_mmu_sync_roots()
3272 static gpa_t nonpaging_gva_to_gpa(struct kvm_vcpu *vcpu, gva_t vaddr, in nonpaging_gva_to_gpa()
3280 static gpa_t nonpaging_gva_to_gpa_nested(struct kvm_vcpu *vcpu, gva_t vaddr, in nonpaging_gva_to_gpa_nested()
3308 static bool quickly_check_mmio_pf(struct kvm_vcpu *vcpu, u64 addr, bool direct) in quickly_check_mmio_pf()
3318 walk_shadow_page_get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep) in walk_shadow_page_get_mmio_spte()
3362 int handle_mmio_page_fault(struct kvm_vcpu *vcpu, u64 addr, bool direct) in handle_mmio_page_fault()
3397 static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva, in nonpaging_page_fault()
3424 static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn) in kvm_arch_setup_async_pf()
3436 static bool can_do_async_pf(struct kvm_vcpu *vcpu) in can_do_async_pf()
3445 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, in try_async_pf()
3472 check_hugepage_cache_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, int level) in check_hugepage_cache_consistency()
3481 static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, in tdp_page_fault()
3546 static void nonpaging_init_context(struct kvm_vcpu *vcpu, in nonpaging_init_context()
3561 void kvm_mmu_new_cr3(struct kvm_vcpu *vcpu) in kvm_mmu_new_cr3()
3566 static unsigned long get_cr3(struct kvm_vcpu *vcpu) in get_cr3()
3571 static void inject_page_fault(struct kvm_vcpu *vcpu, in inject_page_fault()
3577 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
3617 __reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, in __reset_rsvds_bits_mask()
3699 static void reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, in reset_rsvds_bits_mask()
3742 static void reset_rsvds_bits_mask_ept(struct kvm_vcpu *vcpu, in reset_rsvds_bits_mask_ept()
3755 reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) in reset_shadow_zero_bits_mask()
3782 reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, in reset_tdp_shadow_zero_bits_mask()
3802 reset_ept_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, in reset_ept_shadow_zero_bits_mask()
3809 static void update_permission_bitmask(struct kvm_vcpu *vcpu, in update_permission_bitmask()
3874 static void update_last_pte_bitmap(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu) in update_last_pte_bitmap()
3892 static void paging64_init_context_common(struct kvm_vcpu *vcpu, in paging64_init_context_common()
3914 static void paging64_init_context(struct kvm_vcpu *vcpu, in paging64_init_context()
3920 static void paging32_init_context(struct kvm_vcpu *vcpu, in paging32_init_context()
3940 static void paging32E_init_context(struct kvm_vcpu *vcpu, in paging32E_init_context()
3946 static void init_kvm_tdp_mmu(struct kvm_vcpu *vcpu) in init_kvm_tdp_mmu()
3990 void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu) in kvm_init_shadow_mmu()
4019 void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly) in kvm_init_shadow_ept_mmu()
4043 static void init_kvm_softmmu(struct kvm_vcpu *vcpu) in init_kvm_softmmu()
4054 static void init_kvm_nested_mmu(struct kvm_vcpu *vcpu) in init_kvm_nested_mmu()
4093 static void init_kvm_mmu(struct kvm_vcpu *vcpu) in init_kvm_mmu()
4103 void kvm_mmu_reset_context(struct kvm_vcpu *vcpu) in kvm_mmu_reset_context()
4110 int kvm_mmu_load(struct kvm_vcpu *vcpu) in kvm_mmu_load()
4128 void kvm_mmu_unload(struct kvm_vcpu *vcpu) in kvm_mmu_unload()
4135 static void mmu_pte_write_new_pte(struct kvm_vcpu *vcpu, in mmu_pte_write_new_pte()
4161 static void mmu_pte_write_flush_tlb(struct kvm_vcpu *vcpu, bool zap_page, in mmu_pte_write_flush_tlb()
4173 static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, in mmu_pte_write_fetch_gpte()
4284 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, in kvm_mmu_pte_write()
4358 int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva) in kvm_mmu_unprotect_page_virt()
4374 static void make_mmu_pages_available(struct kvm_vcpu *vcpu) in make_mmu_pages_available()
4390 static bool is_mmio_page_fault(struct kvm_vcpu *vcpu, gva_t addr) in is_mmio_page_fault()
4398 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t cr2, u32 error_code, in kvm_mmu_page_fault()
4434 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva) in kvm_mmu_invlpg()
4454 static void free_mmu_pages(struct kvm_vcpu *vcpu) in free_mmu_pages()
4461 static int alloc_mmu_pages(struct kvm_vcpu *vcpu) in alloc_mmu_pages()
4482 int kvm_mmu_create(struct kvm_vcpu *vcpu) in kvm_mmu_create()
4492 void kvm_mmu_setup(struct kvm_vcpu *vcpu) in kvm_mmu_setup()
4961 void kvm_mmu_destroy(struct kvm_vcpu *vcpu) in kvm_mmu_destroy()