Lines Matching refs:memslot

53 static bool memslot_is_logging(struct kvm_memory_slot *memslot)  in memslot_is_logging()  argument
55 return memslot->dirty_bitmap && !(memslot->flags & KVM_MEM_READONLY); in memslot_is_logging()
355 struct kvm_memory_slot *memslot) in stage2_flush_memslot() argument
357 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
358 phys_addr_t end = addr + PAGE_SIZE * memslot->npages; in stage2_flush_memslot()
379 struct kvm_memory_slot *memslot; in stage2_flush_vm() local
386 kvm_for_each_memslot(memslot, slots) in stage2_flush_vm()
387 stage2_flush_memslot(kvm, memslot); in stage2_flush_vm()
745 struct kvm_memory_slot *memslot) in stage2_unmap_memslot() argument
747 hva_t hva = memslot->userspace_addr; in stage2_unmap_memslot()
748 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
749 phys_addr_t size = PAGE_SIZE * memslot->npages; in stage2_unmap_memslot()
778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot()
795 struct kvm_memory_slot *memslot; in stage2_unmap_vm() local
802 kvm_for_each_memslot(memslot, slots) in stage2_unmap_vm()
803 stage2_unmap_memslot(kvm, memslot); in stage2_unmap_vm()
1162 struct kvm_memory_slot *memslot = id_to_memslot(kvm->memslots, slot); in kvm_mmu_wp_memory_region() local
1163 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1164 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1215 struct kvm_memory_slot *memslot, unsigned long hva, in user_mem_abort() argument
1228 bool logging_active = memslot_is_logging(memslot); in user_mem_abort()
1259 if ((memslot->userspace_addr & ~PMD_MASK) != in user_mem_abort()
1260 ((memslot->base_gfn << PAGE_SHIFT) & ~PMD_MASK)) in user_mem_abort()
1314 fault_ipa_uncached = memslot->flags & KVM_MEMSLOT_INCOHERENT; in user_mem_abort()
1400 struct kvm_memory_slot *memslot; in kvm_handle_guest_abort() local
1426 memslot = gfn_to_memslot(vcpu->kvm, gfn); in kvm_handle_guest_abort()
1427 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable); in kvm_handle_guest_abort()
1457 ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status); in kvm_handle_guest_abort()
1473 struct kvm_memory_slot *memslot; in handle_hva_to_gpa() local
1479 kvm_for_each_memslot(memslot, slots) { in handle_hva_to_gpa()
1483 hva_start = max(start, memslot->userspace_addr); in handle_hva_to_gpa()
1484 hva_end = min(end, memslot->userspace_addr + in handle_hva_to_gpa()
1485 (memslot->npages << PAGE_SHIFT)); in handle_hva_to_gpa()
1493 gfn = hva_to_gfn_memslot(hva_start, memslot); in handle_hva_to_gpa()
1494 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); in handle_hva_to_gpa()
1739 struct kvm_memory_slot *memslot, in kvm_arch_prepare_memory_region() argument
1756 if (memslot->base_gfn + memslot->npages >= in kvm_arch_prepare_memory_region()
1803 if (memslot->flags & KVM_MEM_LOG_DIRTY_PAGES) in kvm_arch_prepare_memory_region()
1822 stage2_flush_memslot(kvm, memslot); in kvm_arch_prepare_memory_region()