Lines Matching refs:memslot
107 struct kvm_memory_slot *memslot, gfn_t gfn);
544 static void kvm_destroy_dirty_bitmap(struct kvm_memory_slot *memslot) in kvm_destroy_dirty_bitmap() argument
546 if (!memslot->dirty_bitmap) in kvm_destroy_dirty_bitmap()
549 kvfree(memslot->dirty_bitmap); in kvm_destroy_dirty_bitmap()
550 memslot->dirty_bitmap = NULL; in kvm_destroy_dirty_bitmap()
570 struct kvm_memory_slot *memslot; in kvm_free_physmem() local
572 kvm_for_each_memslot(memslot, slots) in kvm_free_physmem()
573 kvm_free_physmem_slot(kvm, memslot, NULL); in kvm_free_physmem()
647 static int kvm_create_dirty_bitmap(struct kvm_memory_slot *memslot) in kvm_create_dirty_bitmap() argument
649 unsigned long dirty_bytes = 2 * kvm_dirty_bitmap_bytes(memslot); in kvm_create_dirty_bitmap()
651 memslot->dirty_bitmap = kvm_kvzalloc(dirty_bytes); in kvm_create_dirty_bitmap()
652 if (!memslot->dirty_bitmap) in kvm_create_dirty_bitmap()
964 struct kvm_memory_slot *memslot; in kvm_get_dirty_log() local
973 memslot = id_to_memslot(kvm->memslots, log->slot); in kvm_get_dirty_log()
975 if (!memslot->dirty_bitmap) in kvm_get_dirty_log()
978 n = kvm_dirty_bitmap_bytes(memslot); in kvm_get_dirty_log()
981 any = memslot->dirty_bitmap[i]; in kvm_get_dirty_log()
984 if (copy_to_user(log->dirty_bitmap, memslot->dirty_bitmap, n)) in kvm_get_dirty_log()
1022 struct kvm_memory_slot *memslot; in kvm_get_dirty_log_protect() local
1032 memslot = id_to_memslot(kvm->memslots, log->slot); in kvm_get_dirty_log_protect()
1034 dirty_bitmap = memslot->dirty_bitmap; in kvm_get_dirty_log_protect()
1039 n = kvm_dirty_bitmap_bytes(memslot); in kvm_get_dirty_log_protect()
1060 kvm_arch_mmu_enable_log_dirty_pt_masked(kvm, memslot, in kvm_get_dirty_log_protect()
1097 struct kvm_memory_slot *memslot = gfn_to_memslot(kvm, gfn); in kvm_is_visible_gfn() local
1099 if (!memslot || memslot->id >= KVM_USER_MEM_SLOTS || in kvm_is_visible_gfn()
1100 memslot->flags & KVM_MEMSLOT_INVALID) in kvm_is_visible_gfn()
1641 ghc->memslot = gfn_to_memslot(kvm, start_gfn); in kvm_gfn_to_hva_cache_init()
1642 ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, NULL); in kvm_gfn_to_hva_cache_init()
1651 ghc->memslot = gfn_to_memslot(kvm, start_gfn); in kvm_gfn_to_hva_cache_init()
1652 ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, in kvm_gfn_to_hva_cache_init()
1659 ghc->memslot = NULL; in kvm_gfn_to_hva_cache_init()
1676 if (unlikely(!ghc->memslot)) in kvm_write_guest_cached()
1685 mark_page_dirty_in_slot(kvm, ghc->memslot, ghc->gpa >> PAGE_SHIFT); in kvm_write_guest_cached()
1702 if (unlikely(!ghc->memslot)) in kvm_read_guest_cached()
1744 struct kvm_memory_slot *memslot, in mark_page_dirty_in_slot() argument
1747 if (memslot && memslot->dirty_bitmap) { in mark_page_dirty_in_slot()
1748 unsigned long rel_gfn = gfn - memslot->base_gfn; in mark_page_dirty_in_slot()
1750 set_bit_le(rel_gfn, memslot->dirty_bitmap); in mark_page_dirty_in_slot()
1756 struct kvm_memory_slot *memslot; in mark_page_dirty() local
1758 memslot = gfn_to_memslot(kvm, gfn); in mark_page_dirty()
1759 mark_page_dirty_in_slot(kvm, memslot, gfn); in mark_page_dirty()