Lines Matching refs:arch
41 kvmppc_set_msr(vcpu, vcpu->arch.intr_msr); in kvmppc_mmu_book3s_64_reset_msr()
52 for (i = 0; i < vcpu->arch.slb_nr; i++) { in kvmppc_mmu_book3s_64_find_slbe()
55 if (!vcpu->arch.slb[i].valid) in kvmppc_mmu_book3s_64_find_slbe()
58 if (vcpu->arch.slb[i].tb) in kvmppc_mmu_book3s_64_find_slbe()
61 if (vcpu->arch.slb[i].esid == cmp_esid) in kvmppc_mmu_book3s_64_find_slbe()
62 return &vcpu->arch.slb[i]; in kvmppc_mmu_book3s_64_find_slbe()
67 for (i = 0; i < vcpu->arch.slb_nr; i++) { in kvmppc_mmu_book3s_64_find_slbe()
68 if (vcpu->arch.slb[i].vsid) in kvmppc_mmu_book3s_64_find_slbe()
70 vcpu->arch.slb[i].valid ? 'v' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
71 vcpu->arch.slb[i].large ? 'l' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
72 vcpu->arch.slb[i].tb ? 't' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
73 vcpu->arch.slb[i].esid, in kvmppc_mmu_book3s_64_find_slbe()
74 vcpu->arch.slb[i].vsid); in kvmppc_mmu_book3s_64_find_slbe()
162 if (vcpu->arch.papr_enabled) in kvmppc_mmu_book3s_64_get_pteg()
224 ulong mp_ea = vcpu->arch.magic_page_ea; in kvmppc_mmu_book3s_64_xlate()
232 gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); in kvmppc_mmu_book3s_64_xlate()
260 mutex_lock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
285 (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE)) { in kvmppc_mmu_book3s_64_xlate()
316 if (unlikely(vcpu->arch.disable_kernel_nx) && in kvmppc_mmu_book3s_64_xlate()
363 mutex_unlock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
370 mutex_unlock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
393 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmte()
396 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmte()
410 if (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE) { in kvmppc_mmu_book3s_64_slbmte()
434 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmfee()
437 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmfee()
446 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmfev()
449 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmfev()
482 for (i = 1; i < vcpu->arch.slb_nr; i++) { in kvmppc_mmu_book3s_64_slbia()
483 vcpu->arch.slb[i].valid = false; in kvmppc_mmu_book3s_64_slbia()
484 vcpu->arch.slb[i].orige = 0; in kvmppc_mmu_book3s_64_slbia()
485 vcpu->arch.slb[i].origv = 0; in kvmppc_mmu_book3s_64_slbia()
549 if (vcpu->arch.hflags & BOOK3S_HFLAG_NEW_TLBIE) { in kvmppc_mmu_book3s_64_tlbie()
570 ulong mp_ea = vcpu->arch.magic_page_ea; in segment_contains_magic_page()
583 ulong mp_ea = vcpu->arch.magic_page_ea; in kvmppc_mmu_book3s_64_esid_to_vsid()
658 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_64_init()
674 vcpu->arch.hflags |= BOOK3S_HFLAG_SLB; in kvmppc_mmu_book3s_64_init()