Lines Matching refs:pte_index

74 			     unsigned long *rmap, long pte_index, int realmode)  in kvmppc_add_revmap_chain()  argument
89 tail->forw = pte_index; in kvmppc_add_revmap_chain()
90 head->back = pte_index; in kvmppc_add_revmap_chain()
92 rev->forw = rev->back = pte_index; in kvmppc_add_revmap_chain()
94 pte_index | KVMPPC_RMAP_PRESENT; in kvmppc_add_revmap_chain()
101 static void remove_revmap_chain(struct kvm *kvm, long pte_index, in remove_revmap_chain() argument
126 if (head == pte_index) { in remove_revmap_chain()
128 if (head == pte_index) in remove_revmap_chain()
138 long pte_index, unsigned long pteh, unsigned long ptel, in kvmppc_do_h_enter() argument
252 if (pte_index >= kvm->arch.hpt_npte) in kvmppc_do_h_enter()
255 pte_index &= ~7UL; in kvmppc_do_h_enter()
256 hpte = (__be64 *)(kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_do_h_enter()
285 pte_index += i; in kvmppc_do_h_enter()
287 hpte = (__be64 *)(kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_do_h_enter()
304 rev = &kvm->arch.revmap[pte_index]; in kvmppc_do_h_enter()
324 kvmppc_add_revmap_chain(kvm, rev, rmap, pte_index, in kvmppc_do_h_enter()
339 *pte_idx_ret = pte_index; in kvmppc_do_h_enter()
345 long pte_index, unsigned long pteh, unsigned long ptel) in kvmppc_h_enter() argument
347 return kvmppc_do_h_enter(vcpu->kvm, flags, pte_index, pteh, ptel, in kvmppc_h_enter()
400 unsigned long pte_index, unsigned long avpn, in kvmppc_do_h_remove() argument
408 if (pte_index >= kvm->arch.hpt_npte) in kvmppc_do_h_remove()
410 hpte = (__be64 *)(kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_do_h_remove()
421 rev = real_vmalloc_addr(&kvm->arch.revmap[pte_index]); in kvmppc_do_h_remove()
425 rb = compute_tlbie_rb(v, be64_to_cpu(hpte[1]), pte_index); in kvmppc_do_h_remove()
436 remove_revmap_chain(kvm, pte_index, rev, v, in kvmppc_do_h_remove()
450 unsigned long pte_index, unsigned long avpn) in kvmppc_h_remove() argument
452 return kvmppc_do_h_remove(vcpu->kvm, flags, pte_index, avpn, in kvmppc_h_remove()
463 unsigned long flags, req, pte_index, rcbits; in kvmppc_h_bulk_remove() local
474 pte_index = args[j]; in kvmppc_h_bulk_remove()
475 flags = pte_index >> 56; in kvmppc_h_bulk_remove()
476 pte_index &= ((1ul << 56) - 1); in kvmppc_h_bulk_remove()
484 pte_index >= kvm->arch.hpt_npte) { in kvmppc_h_bulk_remove()
486 args[j] = ((0xa0 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
490 hp = (__be64 *) (kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_h_bulk_remove()
517 args[j] = ((0x90 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
521 args[j] = ((0x80 | flags) << 56) + pte_index; in kvmppc_h_bulk_remove()
522 rev = real_vmalloc_addr(&kvm->arch.revmap[pte_index]); in kvmppc_h_bulk_remove()
536 be64_to_cpu(hp[1]), pte_index); in kvmppc_h_bulk_remove()
552 pte_index = args[j] & ((1ul << 56) - 1); in kvmppc_h_bulk_remove()
555 remove_revmap_chain(kvm, pte_index, rev, in kvmppc_h_bulk_remove()
567 unsigned long pte_index, unsigned long avpn, in kvmppc_h_protect() argument
576 if (pte_index >= kvm->arch.hpt_npte) in kvmppc_h_protect()
579 hpte = (__be64 *)(kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_h_protect()
597 rev = real_vmalloc_addr(&kvm->arch.revmap[pte_index]); in kvmppc_h_protect()
617 rb = compute_tlbie_rb(v, r, pte_index); in kvmppc_h_protect()
631 unsigned long pte_index) in kvmppc_h_read() argument
639 if (pte_index >= kvm->arch.hpt_npte) in kvmppc_h_read()
642 pte_index &= ~3; in kvmppc_h_read()
645 rev = real_vmalloc_addr(&kvm->arch.revmap[pte_index]); in kvmppc_h_read()
646 for (i = 0; i < n; ++i, ++pte_index) { in kvmppc_h_read()
647 hpte = (__be64 *)(kvm->arch.hpt_virt + (pte_index << 4)); in kvmppc_h_read()
665 unsigned long pte_index) in kvmppc_invalidate_hpte() argument
671 pte_index); in kvmppc_invalidate_hpte()
677 unsigned long pte_index) in kvmppc_clear_ref_hpte() argument
683 pte_index); in kvmppc_clear_ref_hpte()