slb_nr 394 arch/powerpc/include/asm/kvm_host.h u64 (*slbmfee)(struct kvm_vcpu *vcpu, u64 slb_nr); slb_nr 395 arch/powerpc/include/asm/kvm_host.h u64 (*slbmfev)(struct kvm_vcpu *vcpu, u64 slb_nr); slb_nr 397 arch/powerpc/include/asm/kvm_host.h void (*slbie)(struct kvm_vcpu *vcpu, u64 slb_nr); slb_nr 520 arch/powerpc/include/asm/kvm_host.h int slb_nr; /* total number of entries in SLB */ slb_nr 549 arch/powerpc/kernel/asm-offsets.c OFFSET(VCPU_SLB_NR, kvm_vcpu, arch.slb_nr); slb_nr 49 arch/powerpc/kvm/book3s_64_mmu.c for (i = 0; i < vcpu->arch.slb_nr; i++) { slb_nr 64 arch/powerpc/kvm/book3s_64_mmu.c for (i = 0; i < vcpu->arch.slb_nr; i++) { slb_nr 381 arch/powerpc/kvm/book3s_64_mmu.c int slb_nr; slb_nr 388 arch/powerpc/kvm/book3s_64_mmu.c slb_nr = rb & 0xfff; slb_nr 390 arch/powerpc/kvm/book3s_64_mmu.c if (slb_nr > vcpu->arch.slb_nr) slb_nr 393 arch/powerpc/kvm/book3s_64_mmu.c slbe = &vcpu->arch.slb[slb_nr]; slb_nr 440 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_slbmfee(struct kvm_vcpu *vcpu, u64 slb_nr) slb_nr 444 arch/powerpc/kvm/book3s_64_mmu.c if (slb_nr > vcpu->arch.slb_nr) slb_nr 447 arch/powerpc/kvm/book3s_64_mmu.c slbe = &vcpu->arch.slb[slb_nr]; slb_nr 452 arch/powerpc/kvm/book3s_64_mmu.c static u64 kvmppc_mmu_book3s_64_slbmfev(struct kvm_vcpu *vcpu, u64 slb_nr) slb_nr 456 arch/powerpc/kvm/book3s_64_mmu.c if (slb_nr > vcpu->arch.slb_nr) slb_nr 459 arch/powerpc/kvm/book3s_64_mmu.c slbe = &vcpu->arch.slb[slb_nr]; slb_nr 492 arch/powerpc/kvm/book3s_64_mmu.c for (i = 1; i < vcpu->arch.slb_nr; i++) { slb_nr 316 arch/powerpc/kvm/book3s_64_mmu_hv.c for (i = 0; i < vcpu->arch.slb_nr; i++) { slb_nr 2161 arch/powerpc/kvm/book3s_64_mmu_hv.c vcpu->arch.slb_nr = 32; /* POWER7/POWER8 */ slb_nr 1562 arch/powerpc/kvm/book3s_hv.c for (i = 0; i < vcpu->arch.slb_nr; i++) { slb_nr 1766 arch/powerpc/kvm/book3s_pr.c vcpu->arch.slb_nr = 64;