hpt 546 arch/powerpc/include/asm/kvm_book3s_64.h static inline unsigned long kvmppc_hpt_npte(struct kvm_hpt_info *hpt) hpt 549 arch/powerpc/include/asm/kvm_book3s_64.h return 1UL << (hpt->order - 4); hpt 552 arch/powerpc/include/asm/kvm_book3s_64.h static inline unsigned long kvmppc_hpt_mask(struct kvm_hpt_info *hpt) hpt 555 arch/powerpc/include/asm/kvm_book3s_64.h return (1UL << (hpt->order - 7)) - 1; hpt 284 arch/powerpc/include/asm/kvm_host.h struct kvm_hpt_info hpt; hpt 67 arch/powerpc/kvm/book3s_64_mmu_hv.c struct kvm_hpt_info hpt; hpt 72 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long hpt = 0; hpt 83 arch/powerpc/kvm/book3s_64_mmu_hv.c hpt = (unsigned long)pfn_to_kaddr(page_to_pfn(page)); hpt 84 arch/powerpc/kvm/book3s_64_mmu_hv.c memset((void *)hpt, 0, (1ul << order)); hpt 88 arch/powerpc/kvm/book3s_64_mmu_hv.c if (!hpt) hpt 89 arch/powerpc/kvm/book3s_64_mmu_hv.c hpt = __get_free_pages(GFP_KERNEL|__GFP_ZERO|__GFP_RETRY_MAYFAIL hpt 92 arch/powerpc/kvm/book3s_64_mmu_hv.c if (!hpt) hpt 104 arch/powerpc/kvm/book3s_64_mmu_hv.c free_pages(hpt, order - PAGE_SHIFT); hpt 109 arch/powerpc/kvm/book3s_64_mmu_hv.c info->virt = hpt; hpt 119 arch/powerpc/kvm/book3s_64_mmu_hv.c kvm->arch.hpt = *info; hpt 147 arch/powerpc/kvm/book3s_64_mmu_hv.c if (kvm->arch.hpt.order == order) { hpt 151 arch/powerpc/kvm/book3s_64_mmu_hv.c memset((void *)kvm->arch.hpt.virt, 0, 1ul << order); hpt 160 arch/powerpc/kvm/book3s_64_mmu_hv.c if (kvm->arch.hpt.virt) { hpt 161 arch/powerpc/kvm/book3s_64_mmu_hv.c kvmppc_free_hpt(&kvm->arch.hpt); hpt 224 arch/powerpc/kvm/book3s_64_mmu_hv.c if (npages > kvmppc_hpt_mask(&kvm->arch.hpt) + 1) hpt 225 arch/powerpc/kvm/book3s_64_mmu_hv.c npages = kvmppc_hpt_mask(&kvm->arch.hpt) + 1; hpt 236 arch/powerpc/kvm/book3s_64_mmu_hv.c & kvmppc_hpt_mask(&kvm->arch.hpt); hpt 374 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4)); hpt 378 arch/powerpc/kvm/book3s_64_mmu_hv.c gr = kvm->arch.hpt.rev[index].guest_rpte; hpt 538 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *)(kvm->arch.hpt.virt + (index << 4)); hpt 539 arch/powerpc/kvm/book3s_64_mmu_hv.c rev = &kvm->arch.hpt.rev[index]; hpt 816 arch/powerpc/kvm/book3s_64_mmu_hv.c __be64 *hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); hpt 817 arch/powerpc/kvm/book3s_64_mmu_hv.c struct revmap_entry *rev = kvm->arch.hpt.rev; hpt 875 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); hpt 930 arch/powerpc/kvm/book3s_64_mmu_hv.c struct revmap_entry *rev = kvm->arch.hpt.rev; hpt 950 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); hpt 993 arch/powerpc/kvm/book3s_64_mmu_hv.c struct revmap_entry *rev = kvm->arch.hpt.rev; hpt 1010 arch/powerpc/kvm/book3s_64_mmu_hv.c hp = (unsigned long *)(kvm->arch.hpt.virt + (i << 4)); hpt 1050 arch/powerpc/kvm/book3s_64_mmu_hv.c struct revmap_entry *rev = kvm->arch.hpt.rev; hpt 1067 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *) (kvm->arch.hpt.virt + (i << 4)); hpt 1230 arch/powerpc/kvm/book3s_64_mmu_hv.c rc = kvmppc_allocate_hpt(&resize->hpt, resize->order); hpt 1235 arch/powerpc/kvm/book3s_64_mmu_hv.c resize->hpt.virt); hpt 1244 arch/powerpc/kvm/book3s_64_mmu_hv.c struct kvm_hpt_info *old = &kvm->arch.hpt; hpt 1245 arch/powerpc/kvm/book3s_64_mmu_hv.c struct kvm_hpt_info *new = &resize->hpt; hpt 1401 arch/powerpc/kvm/book3s_64_mmu_hv.c for (i = 0; i < kvmppc_hpt_npte(&kvm->arch.hpt); i++) { hpt 1423 arch/powerpc/kvm/book3s_64_mmu_hv.c hpt_tmp = kvm->arch.hpt; hpt 1424 arch/powerpc/kvm/book3s_64_mmu_hv.c kvmppc_set_hpt(kvm, &resize->hpt); hpt 1425 arch/powerpc/kvm/book3s_64_mmu_hv.c resize->hpt = hpt_tmp; hpt 1446 arch/powerpc/kvm/book3s_64_mmu_hv.c if (resize->hpt.virt) hpt 1447 arch/powerpc/kvm/book3s_64_mmu_hv.c kvmppc_free_hpt(&resize->hpt); hpt 1763 arch/powerpc/kvm/book3s_64_mmu_hv.c hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE)); hpt 1764 arch/powerpc/kvm/book3s_64_mmu_hv.c revp = kvm->arch.hpt.rev + i; hpt 1779 arch/powerpc/kvm/book3s_64_mmu_hv.c while (i < kvmppc_hpt_npte(&kvm->arch.hpt) && hpt 1789 arch/powerpc/kvm/book3s_64_mmu_hv.c while (i < kvmppc_hpt_npte(&kvm->arch.hpt) && hpt 1805 arch/powerpc/kvm/book3s_64_mmu_hv.c while (i < kvmppc_hpt_npte(&kvm->arch.hpt) && hpt 1826 arch/powerpc/kvm/book3s_64_mmu_hv.c if (i >= kvmppc_hpt_npte(&kvm->arch.hpt)) { hpt 1888 arch/powerpc/kvm/book3s_64_mmu_hv.c if (i >= kvmppc_hpt_npte(&kvm->arch.hpt) || hpt 1889 arch/powerpc/kvm/book3s_64_mmu_hv.c i + hdr.n_valid + hdr.n_invalid > kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 1892 arch/powerpc/kvm/book3s_64_mmu_hv.c hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE)); hpt 2090 arch/powerpc/kvm/book3s_64_mmu_hv.c hptp = (__be64 *)(kvm->arch.hpt.virt + (i * HPTE_SIZE)); hpt 2091 arch/powerpc/kvm/book3s_64_mmu_hv.c for (; len != 0 && i < kvmppc_hpt_npte(&kvm->arch.hpt); hpt 2102 arch/powerpc/kvm/book3s_64_mmu_hv.c gr = kvm->arch.hpt.rev[i].guest_rpte; hpt 4588 arch/powerpc/kvm/book3s_hv.c if (!kvm->arch.hpt.virt) { hpt 4698 arch/powerpc/kvm/book3s_hv.c kvmppc_free_hpt(&kvm->arch.hpt); hpt 4952 arch/powerpc/kvm/book3s_hv.c kvmppc_free_hpt(&kvm->arch.hpt); hpt 89 arch/powerpc/kvm/book3s_hv_rm_mmu.c head = &kvm->arch.hpt.rev[i]; hpt 92 arch/powerpc/kvm/book3s_hv_rm_mmu.c tail = &kvm->arch.hpt.rev[head->back]; hpt 179 arch/powerpc/kvm/book3s_hv_rm_mmu.c next = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->forw]); hpt 180 arch/powerpc/kvm/book3s_hv_rm_mmu.c prev = real_vmalloc_addr(&kvm->arch.hpt.rev[rev->back]); hpt 317 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 321 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 352 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 369 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = &kvm->arch.hpt.rev[pte_index]; hpt 509 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 511 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 527 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 597 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) { hpt 603 arch/powerpc/kvm/book3s_hv_rm_mmu.c hp = (__be64 *) (kvm->arch.hpt.virt + (pte_index << 4)); hpt 640 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 697 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 700 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 720 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 767 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 773 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 775 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 808 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 811 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 812 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 855 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) hpt 858 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]); hpt 859 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4)); hpt 1144 arch/powerpc/kvm/book3s_hv_rm_mmu.c hash = (vsid ^ ((eaddr & somask) >> pshift)) & kvmppc_hpt_mask(&kvm->arch.hpt); hpt 1155 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (hash << 7)); hpt 1191 arch/powerpc/kvm/book3s_hv_rm_mmu.c hash = hash ^ kvmppc_hpt_mask(&kvm->arch.hpt); hpt 1240 arch/powerpc/kvm/book3s_hv_rm_mmu.c hpte = (__be64 *)(kvm->arch.hpt.virt + (index << 4)); hpt 1247 arch/powerpc/kvm/book3s_hv_rm_mmu.c rev = real_vmalloc_addr(&kvm->arch.hpt.rev[index]);