hva_end 485 arch/mips/kvm/mmu.c unsigned long hva_start, hva_end; hva_end 489 arch/mips/kvm/mmu.c hva_end = min(end, memslot->userspace_addr + hva_end 491 arch/mips/kvm/mmu.c if (hva_start >= hva_end) hva_end 499 arch/mips/kvm/mmu.c gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); hva_end 781 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long hva_start, hva_end; hva_end 785 arch/powerpc/kvm/book3s_64_mmu_hv.c hva_end = min(end, memslot->userspace_addr + hva_end 787 arch/powerpc/kvm/book3s_64_mmu_hv.c if (hva_start >= hva_end) hva_end 794 arch/powerpc/kvm/book3s_64_mmu_hv.c gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); hva_end 402 arch/powerpc/kvm/book3s_pr.c unsigned long hva_start, hva_end; hva_end 406 arch/powerpc/kvm/book3s_pr.c hva_end = min(end, memslot->userspace_addr + hva_end 408 arch/powerpc/kvm/book3s_pr.c if (hva_start >= hva_end) hva_end 415 arch/powerpc/kvm/book3s_pr.c gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); hva_end 2010 arch/x86/kvm/mmu.c unsigned long hva_start, hva_end; hva_end 2014 arch/x86/kvm/mmu.c hva_end = min(end, memslot->userspace_addr + hva_end 2016 arch/x86/kvm/mmu.c if (hva_start >= hva_end) hva_end 2023 arch/x86/kvm/mmu.c gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); hva_end 2029 virt/kvm/arm/mmu.c unsigned long hva_start, hva_end; hva_end 2033 virt/kvm/arm/mmu.c hva_end = min(end, memslot->userspace_addr + hva_end 2035 virt/kvm/arm/mmu.c if (hva_start >= hva_end) hva_end 2039 virt/kvm/arm/mmu.c ret |= handler(kvm, gpa, (u64)(hva_end - hva_start), data);