hva_end           485 arch/mips/kvm/mmu.c 		unsigned long hva_start, hva_end;
hva_end           489 arch/mips/kvm/mmu.c 		hva_end = min(end, memslot->userspace_addr +
hva_end           491 arch/mips/kvm/mmu.c 		if (hva_start >= hva_end)
hva_end           499 arch/mips/kvm/mmu.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
hva_end           781 arch/powerpc/kvm/book3s_64_mmu_hv.c 		unsigned long hva_start, hva_end;
hva_end           785 arch/powerpc/kvm/book3s_64_mmu_hv.c 		hva_end = min(end, memslot->userspace_addr +
hva_end           787 arch/powerpc/kvm/book3s_64_mmu_hv.c 		if (hva_start >= hva_end)
hva_end           794 arch/powerpc/kvm/book3s_64_mmu_hv.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
hva_end           402 arch/powerpc/kvm/book3s_pr.c 		unsigned long hva_start, hva_end;
hva_end           406 arch/powerpc/kvm/book3s_pr.c 		hva_end = min(end, memslot->userspace_addr +
hva_end           408 arch/powerpc/kvm/book3s_pr.c 		if (hva_start >= hva_end)
hva_end           415 arch/powerpc/kvm/book3s_pr.c 		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
hva_end          2010 arch/x86/kvm/mmu.c 			unsigned long hva_start, hva_end;
hva_end          2014 arch/x86/kvm/mmu.c 			hva_end = min(end, memslot->userspace_addr +
hva_end          2016 arch/x86/kvm/mmu.c 			if (hva_start >= hva_end)
hva_end          2023 arch/x86/kvm/mmu.c 			gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
hva_end          2029 virt/kvm/arm/mmu.c 		unsigned long hva_start, hva_end;
hva_end          2033 virt/kvm/arm/mmu.c 		hva_end = min(end, memslot->userspace_addr +
hva_end          2035 virt/kvm/arm/mmu.c 		if (hva_start >= hva_end)
hva_end          2039 virt/kvm/arm/mmu.c 		ret |= handler(kvm, gpa, (u64)(hva_end - hva_start), data);