Home
last modified time | relevance | path

Searched refs:base_gfn (Results 1 – 16 of 16) sorted by relevance

/linux-4.4.14/arch/x86/kvm/
Diommu.c44 gfn_t base_gfn, unsigned long npages);
85 gfn = slot->base_gfn; in kvm_iommu_map_pages()
147 kvm_iommu_put_pages(kvm, slot->base_gfn, gfn - slot->base_gfn); in kvm_iommu_map_pages()
274 gfn_t base_gfn, unsigned long npages) in kvm_iommu_put_pages() argument
282 end_gfn = base_gfn + npages; in kvm_iommu_put_pages()
283 gfn = base_gfn; in kvm_iommu_put_pages()
318 kvm_iommu_put_pages(kvm, slot->base_gfn, slot->npages); in kvm_iommu_unmap_pages()
Dmmu.c780 idx = gfn_to_index(gfn, slot->base_gfn, level); in lpage_info_slot()
1040 idx = gfn_to_index(gfn, slot->base_gfn, level); in __gfn_to_rmap()
1286 rmapp = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_write_protect_pt_masked()
1311 rmapp = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_clear_dirty_pt_masked()
4539 end_level, memslot->base_gfn, in slot_handle_level()
4540 memslot->base_gfn + memslot->npages - 1, in slot_handle_level()
4580 start = max(gfn_start, memslot->base_gfn); in kvm_zap_gfn_range()
4581 end = min(gfn_end, memslot->base_gfn + memslot->npages); in kvm_zap_gfn_range()
Dx86.c7812 lpages = gfn_to_index(slot->base_gfn + npages - 1, in kvm_arch_create_memslot()
7813 slot->base_gfn, level) + 1; in kvm_arch_create_memslot()
7827 if (slot->base_gfn & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot()
7829 if ((slot->base_gfn + npages) & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot()
7837 if ((slot->base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE(level) - 1) || in kvm_arch_create_memslot()
/linux-4.4.14/arch/powerpc/kvm/
Dtrace_hv.h286 __field(u64, base_gfn)
298 __entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
306 __entry->base_gfn, __entry->slot_flags)
Dbook3s_64_mmu_hv.c498 if (gfn_base < memslot->base_gfn) in kvmppc_book3s_hv_page_fault()
594 rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; in kvmppc_book3s_hv_page_fault()
696 gfn_t gfn_offset = gfn - memslot->base_gfn; in kvm_handle_hva_range()
798 gfn = memslot->base_gfn; in kvmppc_core_flush_memslot_hv()
1019 if (gfn < memslot->base_gfn || in harvest_vpa_dirty()
1020 gfn >= memslot->base_gfn + memslot->npages) in harvest_vpa_dirty()
1025 __set_bit_le(gfn - memslot->base_gfn, map); in harvest_vpa_dirty()
1112 rmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvmppc_unpin_guest_page()
Dbook3s_hv_rm_mmu.c128 rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]); in revmap_for_hpte()
214 slot_fn = gfn - memslot->base_gfn; in kvmppc_do_h_enter()
De500_mmu_host.c383 slot_start = pfn - (gfn - slot->base_gfn); in kvmppc_e500_shadow_map()
Dbook3s_pr.c1552 ga = memslot->base_gfn << PAGE_SHIFT; in kvm_vm_ioctl_get_dirty_log_pr()
/linux-4.4.14/include/linux/
Dkvm_host.h299 gfn_t base_gfn; member
924 if (gfn >= memslots[slot].base_gfn && in search_memslots()
925 gfn < memslots[slot].base_gfn + memslots[slot].npages) in search_memslots()
931 if (gfn >= memslots[slot].base_gfn) in search_memslots()
937 if (gfn >= memslots[start].base_gfn && in search_memslots()
938 gfn < memslots[start].base_gfn + memslots[start].npages) { in search_memslots()
955 return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; in __gfn_to_hva_memslot()
968 return slot->base_gfn + gfn_offset; in hva_to_gfn_memslot()
/linux-4.4.14/arch/arm/kvm/
Dmmu.c357 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
748 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
1164 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1165 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1188 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_mmu_write_protect_pt_masked() local
1189 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1190 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1261 ((memslot->base_gfn << PAGE_SHIFT) & ~PMD_MASK)) in user_mem_abort()
1758 if (memslot->base_gfn + memslot->npages >= in kvm_arch_prepare_memory_region()
1860 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
/linux-4.4.14/virt/kvm/
Dkvm_main.c738 new->base_gfn <= mslots[i + 1].base_gfn) { in update_memslots()
757 new->base_gfn >= mslots[i - 1].base_gfn) { in update_memslots()
822 gfn_t base_gfn; in __kvm_set_memory_region() local
856 base_gfn = mem->guest_phys_addr >> PAGE_SHIFT; in __kvm_set_memory_region()
865 new.base_gfn = base_gfn; in __kvm_set_memory_region()
878 if (base_gfn != old.base_gfn) in __kvm_set_memory_region()
892 new.base_gfn = 0; in __kvm_set_memory_region()
903 if (!((base_gfn + npages <= slot->base_gfn) || in __kvm_set_memory_region()
904 (base_gfn >= slot->base_gfn + slot->npages))) in __kvm_set_memory_region()
1219 *nr_pages = slot->npages - (gfn - slot->base_gfn); in __gfn_to_hva_many()
[all …]
/linux-4.4.14/arch/powerpc/include/asm/
Dkvm_book3s_64.h388 return !(memslot->base_gfn & mask) && !(memslot->npages & mask); in slot_is_aligned()
/linux-4.4.14/include/trace/events/
Dkvm.h262 __entry->hva = ((gfn - slot->base_gfn) <<
/linux-4.4.14/arch/x86/include/asm/
Dkvm_host.h79 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() argument
83 (base_gfn >> KVM_HPAGE_GFN_SHIFT(level)); in gfn_to_index()
/linux-4.4.14/arch/s390/kvm/
Dkvm-s390.c275 last_gfn = memslot->base_gfn + memslot->npages; in kvm_s390_sync_dirty_log()
276 for (cur_gfn = memslot->base_gfn; cur_gfn <= last_gfn; cur_gfn++) { in kvm_s390_sync_dirty_log()
2718 old->base_gfn * PAGE_SIZE == mem->guest_phys_addr && in kvm_arch_commit_memory_region()
/linux-4.4.14/arch/mips/kvm/
Dmips.c994 ga = memslot->base_gfn << PAGE_SHIFT; in kvm_vm_ioctl_get_dirty_log()