Lines Matching refs:invalid_list
1801 struct list_head *invalid_list);
1803 struct list_head *invalid_list);
1826 struct list_head *invalid_list, bool clear_unsync) in __kvm_sync_page() argument
1829 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, invalid_list); in __kvm_sync_page()
1837 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, invalid_list); in __kvm_sync_page()
1848 LIST_HEAD(invalid_list); in kvm_sync_page_transient()
1851 ret = __kvm_sync_page(vcpu, sp, &invalid_list, false); in kvm_sync_page_transient()
1853 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_sync_page_transient()
1866 struct list_head *invalid_list) in kvm_sync_page() argument
1868 return __kvm_sync_page(vcpu, sp, invalid_list, true); in kvm_sync_page()
1875 LIST_HEAD(invalid_list); in kvm_sync_pages()
1886 kvm_mmu_prepare_zap_page(vcpu->kvm, s, &invalid_list); in kvm_sync_pages()
1892 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_sync_pages()
1963 LIST_HEAD(invalid_list); in mmu_sync_children()
1976 kvm_sync_page(vcpu, sp, &invalid_list); in mmu_sync_children()
1979 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_sync_children()
2218 struct list_head *invalid_list) in mmu_zap_unsync_children() argument
2232 kvm_mmu_prepare_zap_page(kvm, sp, invalid_list); in mmu_zap_unsync_children()
2243 struct list_head *invalid_list) in kvm_mmu_prepare_zap_page() argument
2249 ret = mmu_zap_unsync_children(kvm, sp, invalid_list); in kvm_mmu_prepare_zap_page()
2261 list_move(&sp->link, invalid_list); in kvm_mmu_prepare_zap_page()
2279 struct list_head *invalid_list) in kvm_mmu_commit_zap_page() argument
2283 if (list_empty(invalid_list)) in kvm_mmu_commit_zap_page()
2298 list_for_each_entry_safe(sp, nsp, invalid_list, link) { in kvm_mmu_commit_zap_page()
2305 struct list_head *invalid_list) in prepare_zap_oldest_mmu_page() argument
2314 kvm_mmu_prepare_zap_page(kvm, sp, invalid_list); in prepare_zap_oldest_mmu_page()
2325 LIST_HEAD(invalid_list); in kvm_mmu_change_mmu_pages()
2332 if (!prepare_zap_oldest_mmu_page(kvm, &invalid_list)) in kvm_mmu_change_mmu_pages()
2335 kvm_mmu_commit_zap_page(kvm, &invalid_list); in kvm_mmu_change_mmu_pages()
2347 LIST_HEAD(invalid_list); in kvm_mmu_unprotect_page()
2357 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); in kvm_mmu_unprotect_page()
2359 kvm_mmu_commit_zap_page(kvm, &invalid_list); in kvm_mmu_unprotect_page()
3065 LIST_HEAD(invalid_list); in mmu_free_roots()
3079 kvm_mmu_prepare_zap_page(vcpu->kvm, sp, &invalid_list); in mmu_free_roots()
3080 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_free_roots()
3097 &invalid_list); in mmu_free_roots()
3101 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in mmu_free_roots()
4169 LIST_HEAD(invalid_list); in kvm_mmu_pte_write()
4209 &invalid_list); in kvm_mmu_pte_write()
4232 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in kvm_mmu_pte_write()
4255 LIST_HEAD(invalid_list); in make_mmu_pages_available()
4261 if (!prepare_zap_oldest_mmu_page(vcpu->kvm, &invalid_list)) in make_mmu_pages_available()
4266 kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); in make_mmu_pages_available()
4722 LIST_HEAD(invalid_list); in mmu_shrink_scan()
4751 if (prepare_zap_oldest_mmu_page(kvm, &invalid_list)) in mmu_shrink_scan()
4753 kvm_mmu_commit_zap_page(kvm, &invalid_list); in mmu_shrink_scan()