Lines Matching refs:memslot

1506 	struct kvm_memory_slot *memslot;  in kvm_handle_hva_range()  local
1513 kvm_for_each_memslot(memslot, slots) { in kvm_handle_hva_range()
1517 hva_start = max(start, memslot->userspace_addr); in kvm_handle_hva_range()
1518 hva_end = min(end, memslot->userspace_addr + in kvm_handle_hva_range()
1519 (memslot->npages << PAGE_SHIFT)); in kvm_handle_hva_range()
1526 gfn_start = hva_to_gfn_memslot(hva_start, memslot); in kvm_handle_hva_range()
1527 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); in kvm_handle_hva_range()
1529 for_each_slot_rmap_range(memslot, PT_PAGE_TABLE_LEVEL, in kvm_handle_hva_range()
1533 ret |= handler(kvm, iterator.rmap, memslot, in kvm_handle_hva_range()
4504 slot_handle_level_range(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level_range() argument
4511 for_each_slot_rmap_range(memslot, start_level, end_level, start_gfn, in slot_handle_level_range()
4534 slot_handle_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level() argument
4538 return slot_handle_level_range(kvm, memslot, fn, start_level, in slot_handle_level()
4539 end_level, memslot->base_gfn, in slot_handle_level()
4540 memslot->base_gfn + memslot->npages - 1, in slot_handle_level()
4545 slot_handle_all_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_all_level() argument
4548 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL, in slot_handle_all_level()
4553 slot_handle_large_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_large_level() argument
4556 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL + 1, in slot_handle_large_level()
4561 slot_handle_leaf(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_leaf() argument
4564 return slot_handle_level(kvm, memslot, fn, PT_PAGE_TABLE_LEVEL, in slot_handle_leaf()
4571 struct kvm_memory_slot *memslot; in kvm_zap_gfn_range() local
4577 kvm_for_each_memslot(memslot, slots) { in kvm_zap_gfn_range()
4580 start = max(gfn_start, memslot->base_gfn); in kvm_zap_gfn_range()
4581 end = min(gfn_end, memslot->base_gfn + memslot->npages); in kvm_zap_gfn_range()
4585 slot_handle_level_range(kvm, memslot, kvm_zap_rmapp, in kvm_zap_gfn_range()
4600 struct kvm_memory_slot *memslot) in kvm_mmu_slot_remove_write_access() argument
4605 flush = slot_handle_all_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_remove_write_access()
4665 const struct kvm_memory_slot *memslot) in kvm_mmu_zap_collapsible_sptes() argument
4669 slot_handle_leaf(kvm, (struct kvm_memory_slot *)memslot, in kvm_mmu_zap_collapsible_sptes()
4675 struct kvm_memory_slot *memslot) in kvm_mmu_slot_leaf_clear_dirty() argument
4680 flush = slot_handle_leaf(kvm, memslot, __rmap_clear_dirty, false); in kvm_mmu_slot_leaf_clear_dirty()
4697 struct kvm_memory_slot *memslot) in kvm_mmu_slot_largepage_remove_write_access() argument
4702 flush = slot_handle_large_level(kvm, memslot, slot_rmap_write_protect, in kvm_mmu_slot_largepage_remove_write_access()
4715 struct kvm_memory_slot *memslot) in kvm_mmu_slot_set_dirty() argument
4720 flush = slot_handle_all_level(kvm, memslot, __rmap_set_dirty, false); in kvm_mmu_slot_set_dirty()
4944 struct kvm_memory_slot *memslot; in kvm_mmu_calculate_mmu_pages() local
4950 kvm_for_each_memslot(memslot, slots) in kvm_mmu_calculate_mmu_pages()
4951 nr_pages += memslot->npages; in kvm_mmu_calculate_mmu_pages()