Lines Matching refs:invalid_list
1872 struct list_head *invalid_list);
1874 struct list_head *invalid_list);
1897 struct list_head *invalid_list, bool clear_unsync) in __kvm_sync_page() argument
1900 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, invalid_list); in __kvm_sync_page()
1908 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, invalid_list); in __kvm_sync_page()
1919 LIST_HEAD(invalid_list); in kvm_sync_page_transient()
1922 ret = __kvm_sync_page(vcpu, sp, &invalid_list, false); in kvm_sync_page_transient()
1924 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_sync_page_transient()
1937 struct list_head *invalid_list) in kvm_sync_page() argument
1939 return __kvm_sync_page(vcpu, sp, invalid_list, true); in kvm_sync_page()
1946 LIST_HEAD(invalid_list); in kvm_sync_pages()
1957 kvm_mmu_prepare_zap_page(vcpu->kvm, s, &invalid_list); in kvm_sync_pages()
1963 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_sync_pages()
2034 LIST_HEAD(invalid_list); in mmu_sync_children()
2047 kvm_sync_page(vcpu, sp, &invalid_list); in mmu_sync_children()
2050 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_sync_children()
2289 struct list_head *invalid_list) in mmu_zap_unsync_children() argument
2303 kvm_mmu_prepare_zap_page(kvm, sp, invalid_list); in mmu_zap_unsync_children()
2314 struct list_head *invalid_list) in kvm_mmu_prepare_zap_page() argument
2320 ret = mmu_zap_unsync_children(kvm, sp, invalid_list); in kvm_mmu_prepare_zap_page()
2332 list_move(&sp->link, invalid_list); in kvm_mmu_prepare_zap_page()
2350 struct list_head *invalid_list) in kvm_mmu_commit_zap_page() argument
2354 if (list_empty(invalid_list)) in kvm_mmu_commit_zap_page()
2369 list_for_each_entry_safe(sp, nsp, invalid_list, link) { in kvm_mmu_commit_zap_page()
2376 struct list_head *invalid_list) in prepare_zap_oldest_mmu_page() argument
2385 kvm_mmu_prepare_zap_page(kvm, sp, invalid_list); in prepare_zap_oldest_mmu_page()
2396 LIST_HEAD(invalid_list); in kvm_mmu_change_mmu_pages()
2403 if (!prepare_zap_oldest_mmu_page(kvm, &invalid_list)) in kvm_mmu_change_mmu_pages()
2406 kvm_mmu_commit_zap_page(kvm, &invalid_list); in kvm_mmu_change_mmu_pages()
2418 LIST_HEAD(invalid_list); in kvm_mmu_unprotect_page()
2428 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); in kvm_mmu_unprotect_page()
2430 kvm_mmu_commit_zap_page(kvm, &invalid_list); in kvm_mmu_unprotect_page()
3039 LIST_HEAD(invalid_list); in mmu_free_roots()
3053 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, &invalid_list); in mmu_free_roots()
3054 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_free_roots()
3071 &invalid_list); in mmu_free_roots()
3075 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_free_roots()
4289 LIST_HEAD(invalid_list); in kvm_mmu_pte_write()
4330 &invalid_list); in kvm_mmu_pte_write()
4353 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_mmu_pte_write()
4376 LIST_HEAD(invalid_list); in make_mmu_pages_available()
4382 if (!prepare_zap_oldest_mmu_page(vcpu->kvm, &invalid_list)) in make_mmu_pages_available()
4387 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in make_mmu_pages_available()
4840 LIST_HEAD(invalid_list); in mmu_shrink_scan()
4869 if (prepare_zap_oldest_mmu_page(kvm, &invalid_list)) in mmu_shrink_scan()
4871 kvm_mmu_commit_zap_page(kvm, &invalid_list); in mmu_shrink_scan()