gfn_end           472 arch/mips/kvm/mmu.c 					    gpa_t gfn_end,
gfn_end           486 arch/mips/kvm/mmu.c 		gfn_t gfn, gfn_end;
gfn_end           499 arch/mips/kvm/mmu.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
gfn_end           501 arch/mips/kvm/mmu.c 		ret |= handler(kvm, gfn, gfn_end, memslot, data);
gfn_end           508 arch/mips/kvm/mmu.c static int kvm_unmap_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_end           511 arch/mips/kvm/mmu.c 	kvm_mips_flush_gpa_pt(kvm, gfn, gfn_end);
gfn_end           523 arch/mips/kvm/mmu.c static int kvm_set_spte_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_end           565 arch/mips/kvm/mmu.c static int kvm_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_end           568 arch/mips/kvm/mmu.c 	return kvm_mips_mkold_gpa_pt(kvm, gfn, gfn_end);
gfn_end           571 arch/mips/kvm/mmu.c static int kvm_test_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_end           782 arch/powerpc/kvm/book3s_64_mmu_hv.c 		gfn_t gfn, gfn_end;
gfn_end           794 arch/powerpc/kvm/book3s_64_mmu_hv.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
gfn_end           796 arch/powerpc/kvm/book3s_64_mmu_hv.c 		for (; gfn < gfn_end; ++gfn) {
gfn_end           403 arch/powerpc/kvm/book3s_pr.c 		gfn_t gfn, gfn_end;
gfn_end           415 arch/powerpc/kvm/book3s_pr.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
gfn_end           418 arch/powerpc/kvm/book3s_pr.c 					      gfn_end << PAGE_SHIFT);
gfn_end           406 arch/powerpc/kvm/e500_mmu_host.c 				unsigned long gfn_start, gfn_end;
gfn_end           410 arch/powerpc/kvm/e500_mmu_host.c 				gfn_end = gfn_start + tsize_pages;
gfn_end           414 arch/powerpc/kvm/e500_mmu_host.c 				if (gfn_end + pfn - gfn > end)
gfn_end          2011 arch/x86/kvm/mmu.c 			gfn_t gfn_start, gfn_end;
gfn_end          2023 arch/x86/kvm/mmu.c 			gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
gfn_end          2027 arch/x86/kvm/mmu.c 						 gfn_start, gfn_end - 1,
gfn_end          5947 arch/x86/kvm/mmu.c void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end)
gfn_end          5960 arch/x86/kvm/mmu.c 			end = min(gfn_end, memslot->base_gfn + memslot->npages);
gfn_end           206 arch/x86/kvm/mmu.h void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end);