ESID_MASK         112 arch/powerpc/kvm/book3s_32_mmu_host.c 	page = (eaddr & ~ESID_MASK) >> 12;
ESID_MASK         168 arch/powerpc/kvm/book3s_32_mmu_host.c 		((eaddr & ~ESID_MASK) >> VPN_SHIFT);
ESID_MASK         420 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->orige = rb & (ESID_MASK | SLB_ESID_V);
ESID_MASK         281 arch/powerpc/kvm/book3s_64_mmu_host.c 		else if ((svcpu->slb[i].esid & ESID_MASK) == esid) {
ESID_MASK         314 arch/powerpc/kvm/book3s_64_mmu_host.c 	u64 slb_esid = (eaddr & ESID_MASK) | SLB_ESID_V;
ESID_MASK         321 arch/powerpc/kvm/book3s_64_mmu_host.c 	slb_index = kvmppc_mmu_next_segment(vcpu, eaddr & ESID_MASK);
ESID_MASK         323 arch/powerpc/kvm/book3s_64_mmu_hv.c 			mask = ESID_MASK;
ESID_MASK          36 arch/powerpc/mm/book3s64/slb.c 	(((ssize) == MMU_SEGSIZE_256M)? ESID_MASK: ESID_MASK_1T)
ESID_MASK         139 arch/powerpc/mm/copro_fault.c 	slb->esid = (ea & (ssize == MMU_SEGSIZE_1T ? ESID_MASK_1T : ESID_MASK)) | SLB_ESID_V;
ESID_MASK         221 arch/powerpc/platforms/cell/spu_base.c 	slb->esid = (ea & ESID_MASK) | SLB_ESID_V;
ESID_MASK         235 arch/powerpc/platforms/cell/spu_base.c 		if (!((slbs[i].esid ^ ea) & ESID_MASK))
ESID_MASK         147 drivers/misc/cxl/main.c 			(u64)ctx->sstp, (u64)ctx->sstp & ESID_MASK, mmu_kernel_ssize, vsid, sstp0, sstp1);