ESID_MASK 112 arch/powerpc/kvm/book3s_32_mmu_host.c page = (eaddr & ~ESID_MASK) >> 12; ESID_MASK 168 arch/powerpc/kvm/book3s_32_mmu_host.c ((eaddr & ~ESID_MASK) >> VPN_SHIFT); ESID_MASK 420 arch/powerpc/kvm/book3s_64_mmu.c slbe->orige = rb & (ESID_MASK | SLB_ESID_V); ESID_MASK 281 arch/powerpc/kvm/book3s_64_mmu_host.c else if ((svcpu->slb[i].esid & ESID_MASK) == esid) { ESID_MASK 314 arch/powerpc/kvm/book3s_64_mmu_host.c u64 slb_esid = (eaddr & ESID_MASK) | SLB_ESID_V; ESID_MASK 321 arch/powerpc/kvm/book3s_64_mmu_host.c slb_index = kvmppc_mmu_next_segment(vcpu, eaddr & ESID_MASK); ESID_MASK 323 arch/powerpc/kvm/book3s_64_mmu_hv.c mask = ESID_MASK; ESID_MASK 36 arch/powerpc/mm/book3s64/slb.c (((ssize) == MMU_SEGSIZE_256M)? ESID_MASK: ESID_MASK_1T) ESID_MASK 139 arch/powerpc/mm/copro_fault.c slb->esid = (ea & (ssize == MMU_SEGSIZE_1T ? ESID_MASK_1T : ESID_MASK)) | SLB_ESID_V; ESID_MASK 221 arch/powerpc/platforms/cell/spu_base.c slb->esid = (ea & ESID_MASK) | SLB_ESID_V; ESID_MASK 235 arch/powerpc/platforms/cell/spu_base.c if (!((slbs[i].esid ^ ea) & ESID_MASK)) ESID_MASK 147 drivers/misc/cxl/main.c (u64)ctx->sstp, (u64)ctx->sstp & ESID_MASK, mmu_kernel_ssize, vsid, sstp0, sstp1);