base_gfn 216 arch/mips/kvm/mips.c kvm_mips_flush_gpa_pt(kvm, slot->base_gfn, base_gfn 217 arch/mips/kvm/mips.c slot->base_gfn + slot->npages - 1); base_gfn 257 arch/mips/kvm/mips.c needs_flush = kvm_mips_mkclean_gpa_pt(kvm, new->base_gfn, base_gfn 258 arch/mips/kvm/mips.c new->base_gfn + new->npages - 1); base_gfn 445 arch/mips/kvm/mmu.c gfn_t base_gfn = slot->base_gfn + gfn_offset; base_gfn 446 arch/mips/kvm/mmu.c gfn_t start = base_gfn + __ffs(mask); base_gfn 447 arch/mips/kvm/mmu.c gfn_t end = base_gfn + __fls(mask); base_gfn 493 arch/powerpc/include/asm/kvm_book3s_64.h return !(memslot->base_gfn & mask) && !(memslot->npages & mask); base_gfn 576 arch/powerpc/kvm/book3s_64_mmu_hv.c if (gfn_base < memslot->base_gfn) base_gfn 687 arch/powerpc/kvm/book3s_64_mmu_hv.c rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; base_gfn 861 arch/powerpc/kvm/book3s_64_mmu_hv.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 907 arch/powerpc/kvm/book3s_64_mmu_hv.c gfn = memslot->base_gfn; base_gfn 936 arch/powerpc/kvm/book3s_64_mmu_hv.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 999 arch/powerpc/kvm/book3s_64_mmu_hv.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 1138 arch/powerpc/kvm/book3s_64_mmu_hv.c if (gfn < memslot->base_gfn || base_gfn 1139 arch/powerpc/kvm/book3s_64_mmu_hv.c gfn >= memslot->base_gfn + memslot->npages) base_gfn 1144 arch/powerpc/kvm/book3s_64_mmu_hv.c __set_bit_le(gfn - memslot->base_gfn, map); base_gfn 1219 arch/powerpc/kvm/book3s_64_mmu_hv.c set_bit_le(gfn - memslot->base_gfn, memslot->dirty_bitmap); base_gfn 1297 arch/powerpc/kvm/book3s_64_mmu_hv.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 998 arch/powerpc/kvm/book3s_64_mmu_radix.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 1026 arch/powerpc/kvm/book3s_64_mmu_radix.c unsigned long gfn = memslot->base_gfn + pagenum; base_gfn 1043 arch/powerpc/kvm/book3s_64_mmu_radix.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 1085 arch/powerpc/kvm/book3s_64_mmu_radix.c gpa = memslot->base_gfn << PAGE_SHIFT; base_gfn 92 arch/powerpc/kvm/book3s_64_vio_hv.c *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 813 arch/powerpc/kvm/book3s_hv.c if ((from + len) >= ((from_memslot->base_gfn + from_memslot->npages) base_gfn 825 arch/powerpc/kvm/book3s_hv.c if ((to + len) >= ((to_memslot->base_gfn + to_memslot->npages) base_gfn 887 arch/powerpc/kvm/book3s_hv_nested.c gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; base_gfn 1416 arch/powerpc/kvm/book3s_hv_nested.c rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; base_gfn 117 arch/powerpc/kvm/book3s_hv_rm_mmu.c gfn -= memslot->base_gfn; base_gfn 155 arch/powerpc/kvm/book3s_hv_rm_mmu.c rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]); base_gfn 246 arch/powerpc/kvm/book3s_hv_rm_mmu.c slot_fn = gfn - memslot->base_gfn; base_gfn 1884 arch/powerpc/kvm/book3s_pr.c ga = memslot->base_gfn << PAGE_SHIFT; base_gfn 381 arch/powerpc/kvm/e500_mmu_host.c slot_start = pfn - (gfn - slot->base_gfn); base_gfn 285 arch/powerpc/kvm/trace_hv.h __field(u64, base_gfn) base_gfn 297 arch/powerpc/kvm/trace_hv.h __entry->base_gfn = memslot ? memslot->base_gfn : -1UL; base_gfn 305 arch/powerpc/kvm/trace_hv.h __entry->base_gfn, __entry->slot_flags) base_gfn 585 arch/s390/kvm/kvm-s390.c cur_gfn = memslot->base_gfn; base_gfn 586 arch/s390/kvm/kvm-s390.c last_gfn = memslot->base_gfn + memslot->npages; base_gfn 1922 arch/s390/kvm/kvm-s390.c if (gfn >= memslots[slot].base_gfn && base_gfn 1923 arch/s390/kvm/kvm-s390.c gfn < memslots[slot].base_gfn + memslots[slot].npages) base_gfn 1929 arch/s390/kvm/kvm-s390.c if (gfn >= memslots[slot].base_gfn) base_gfn 1938 arch/s390/kvm/kvm-s390.c if (gfn >= memslots[start].base_gfn && base_gfn 1939 arch/s390/kvm/kvm-s390.c gfn < memslots[start].base_gfn + memslots[start].npages) { base_gfn 1974 arch/s390/kvm/kvm-s390.c unsigned long ofs = cur_gfn - ms->base_gfn; base_gfn 1976 arch/s390/kvm/kvm-s390.c if (ms->base_gfn + ms->npages <= cur_gfn) { base_gfn 1991 arch/s390/kvm/kvm-s390.c return ms->base_gfn + ofs; base_gfn 2008 arch/s390/kvm/kvm-s390.c mem_end = slots->memslots[0].base_gfn + slots->memslots[0].npages; base_gfn 2015 arch/s390/kvm/kvm-s390.c if (test_and_clear_bit(cur_gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) base_gfn 2033 arch/s390/kvm/kvm-s390.c if (cur_gfn - ms->base_gfn >= ms->npages) { base_gfn 4537 arch/s390/kvm/kvm-s390.c rc = gmap_unmap_segment(kvm->arch.gmap, old->base_gfn * PAGE_SIZE, base_gfn 4541 arch/s390/kvm/kvm-s390.c rc = gmap_unmap_segment(kvm->arch.gmap, old->base_gfn * PAGE_SIZE, base_gfn 1177 arch/s390/kvm/priv.c if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) base_gfn 120 arch/x86/include/asm/kvm_host.h static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) base_gfn 124 arch/x86/include/asm/kvm_host.h (base_gfn >> KVM_HPAGE_GFN_SHIFT(level)); base_gfn 1187 arch/x86/kvm/mmu.c idx = gfn_to_index(gfn, slot->base_gfn, level); base_gfn 1474 arch/x86/kvm/mmu.c idx = gfn_to_index(gfn, slot->base_gfn, level); base_gfn 1758 arch/x86/kvm/mmu.c rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), base_gfn 1784 arch/x86/kvm/mmu.c rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), base_gfn 3333 arch/x86/kvm/mmu.c gfn_t base_gfn = gfn; base_gfn 3346 arch/x86/kvm/mmu.c base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); base_gfn 3352 arch/x86/kvm/mmu.c sp = kvm_mmu_get_page(vcpu, base_gfn, it.addr, base_gfn 3362 arch/x86/kvm/mmu.c write, level, base_gfn, pfn, prefault, base_gfn 5729 arch/x86/kvm/mmu.c end_level, memslot->base_gfn, base_gfn 5730 arch/x86/kvm/mmu.c memslot->base_gfn + memslot->npages - 1, base_gfn 5959 arch/x86/kvm/mmu.c start = max(gfn_start, memslot->base_gfn); base_gfn 5960 arch/x86/kvm/mmu.c end = min(gfn_end, memslot->base_gfn + memslot->npages); base_gfn 6008 arch/x86/kvm/mmu.c kvm_flush_remote_tlbs_with_address(kvm, memslot->base_gfn, base_gfn 6079 arch/x86/kvm/mmu.c kvm_flush_remote_tlbs_with_address(kvm, memslot->base_gfn, base_gfn 6098 arch/x86/kvm/mmu.c kvm_flush_remote_tlbs_with_address(kvm, memslot->base_gfn, base_gfn 6116 arch/x86/kvm/mmu.c kvm_flush_remote_tlbs_with_address(kvm, memslot->base_gfn, base_gfn 68 arch/x86/kvm/page_track.c index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); base_gfn 158 arch/x86/kvm/page_track.c index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); base_gfn 624 arch/x86/kvm/paging_tmpl.h gfn_t gfn, base_gfn; base_gfn 674 arch/x86/kvm/paging_tmpl.h base_gfn = gfn; base_gfn 687 arch/x86/kvm/paging_tmpl.h base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); base_gfn 696 arch/x86/kvm/paging_tmpl.h sp = kvm_mmu_get_page(vcpu, base_gfn, addr, base_gfn 705 arch/x86/kvm/paging_tmpl.h it.level, base_gfn, pfn, prefault, map_writable); base_gfn 9756 arch/x86/kvm/x86.c lpages = gfn_to_index(slot->base_gfn + npages - 1, base_gfn 9757 arch/x86/kvm/x86.c slot->base_gfn, level) + 1; base_gfn 9773 arch/x86/kvm/x86.c if (slot->base_gfn & (KVM_PAGES_PER_HPAGE(level) - 1)) base_gfn 9775 arch/x86/kvm/x86.c if ((slot->base_gfn + npages) & (KVM_PAGES_PER_HPAGE(level) - 1)) base_gfn 9783 arch/x86/kvm/x86.c if ((slot->base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE(level) - 1) || base_gfn 1725 drivers/gpu/drm/i915/gvt/kvmgt.c gfn = slot->base_gfn + i; base_gfn 344 include/linux/kvm_host.h gfn_t base_gfn; base_gfn 1017 include/linux/kvm_host.h if (gfn >= memslots[slot].base_gfn && base_gfn 1018 include/linux/kvm_host.h gfn < memslots[slot].base_gfn + memslots[slot].npages) base_gfn 1024 include/linux/kvm_host.h if (gfn >= memslots[slot].base_gfn) base_gfn 1030 include/linux/kvm_host.h if (start < slots->used_slots && gfn >= memslots[start].base_gfn && base_gfn 1031 include/linux/kvm_host.h gfn < memslots[start].base_gfn + memslots[start].npages) { base_gfn 1048 include/linux/kvm_host.h return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; base_gfn 1061 include/linux/kvm_host.h return slot->base_gfn + gfn_offset; base_gfn 272 include/trace/events/kvm.h __entry->hva = ((gfn - slot->base_gfn) << base_gfn 416 virt/kvm/arm/mmu.c phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; base_gfn 927 virt/kvm/arm/mmu.c phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; base_gfn 1544 virt/kvm/arm/mmu.c phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; base_gfn 1545 virt/kvm/arm/mmu.c phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; base_gfn 1568 virt/kvm/arm/mmu.c phys_addr_t base_gfn = slot->base_gfn + gfn_offset; base_gfn 1569 virt/kvm/arm/mmu.c phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; base_gfn 1570 virt/kvm/arm/mmu.c phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; base_gfn 1622 virt/kvm/arm/mmu.c gpa_start = memslot->base_gfn << PAGE_SHIFT; base_gfn 2294 virt/kvm/arm/mmu.c if (memslot->base_gfn + memslot->npages >= base_gfn 2393 virt/kvm/arm/mmu.c gpa_t gpa = slot->base_gfn << PAGE_SHIFT; base_gfn 902 virt/kvm/kvm_main.c new->base_gfn <= mslots[i + 1].base_gfn) { base_gfn 921 virt/kvm/kvm_main.c new->base_gfn >= mslots[i - 1].base_gfn) { base_gfn 995 virt/kvm/kvm_main.c gfn_t base_gfn; base_gfn 1028 virt/kvm/kvm_main.c base_gfn = mem->guest_phys_addr >> PAGE_SHIFT; base_gfn 1037 virt/kvm/kvm_main.c new.base_gfn = base_gfn; base_gfn 1050 virt/kvm/kvm_main.c if (base_gfn != old.base_gfn) base_gfn 1064 virt/kvm/kvm_main.c new.base_gfn = 0; base_gfn 1074 virt/kvm/kvm_main.c if (!((base_gfn + npages <= slot->base_gfn) || base_gfn 1075 virt/kvm/kvm_main.c (base_gfn >= slot->base_gfn + slot->npages))) base_gfn 1442 virt/kvm/kvm_main.c *nr_pages = slot->npages - (gfn - slot->base_gfn); base_gfn 2362 virt/kvm/kvm_main.c unsigned long rel_gfn = gfn - memslot->base_gfn;