esid 509 arch/powerpc/include/asm/book3s/64/mmu-hash.h u64 esid; esid 13 arch/powerpc/include/asm/copro.h u64 esid, vsid; esid 102 arch/powerpc/include/asm/kvm_book3s.h u64 esid; esid 160 arch/powerpc/include/asm/kvm_book3s_asm.h u64 esid; esid 406 arch/powerpc/include/asm/kvm_host.h int (*esid_to_vsid)(struct kvm_vcpu *vcpu, ulong esid, u64 *vsid); esid 412 arch/powerpc/include/asm/kvm_host.h u64 esid; esid 144 arch/powerpc/include/asm/lppaca.h __be64 esid; esid 241 arch/powerpc/kernel/asm-offsets.c OFFSET(SLBSHADOW_STACKESID, slb_shadow, save_area[SLB_NUM_BOLTED - 1].esid); esid 72 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, esid 353 arch/powerpc/kvm/book3s_32_mmu.c static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, esid 356 arch/powerpc/kvm/book3s_32_mmu.c ulong ea = esid << SID_SHIFT; esid 358 arch/powerpc/kvm/book3s_32_mmu.c u64 gvsid = esid; esid 372 arch/powerpc/kvm/book3s_32_mmu.c *vsid = VSID_REAL | esid; esid 304 arch/powerpc/kvm/book3s_32_mmu_host.c u32 esid = eaddr >> SID_SHIFT; esid 311 arch/powerpc/kvm/book3s_32_mmu_host.c if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) { esid 313 arch/powerpc/kvm/book3s_32_mmu_host.c svcpu->sr[esid] = SR_INVALID; esid 322 arch/powerpc/kvm/book3s_32_mmu_host.c map->guest_esid = esid; esid 324 arch/powerpc/kvm/book3s_32_mmu_host.c svcpu->sr[esid] = sr; esid 326 arch/powerpc/kvm/book3s_32_mmu_host.c dprintk_sr("MMU: mtsr %d, 0x%x\n", esid, sr); esid 46 arch/powerpc/kvm/book3s_64_mmu.c u64 esid = GET_ESID(eaddr); esid 50 arch/powerpc/kvm/book3s_64_mmu.c u64 cmp_esid = esid; esid 58 arch/powerpc/kvm/book3s_64_mmu.c if (vcpu->arch.slb[i].esid == cmp_esid) esid 63 arch/powerpc/kvm/book3s_64_mmu.c eaddr, esid, esid_1t); esid 70 arch/powerpc/kvm/book3s_64_mmu.c vcpu->arch.slb[i].esid, esid 380 arch/powerpc/kvm/book3s_64_mmu.c u64 esid, esid_1t; esid 386 arch/powerpc/kvm/book3s_64_mmu.c esid = GET_ESID(rb); esid 397 arch/powerpc/kvm/book3s_64_mmu.c slbe->esid = slbe->tb ? esid_1t : esid; esid 424 arch/powerpc/kvm/book3s_64_mmu.c kvmppc_mmu_map_segment(vcpu, esid << SID_SHIFT); esid 476 arch/powerpc/kvm/book3s_64_mmu.c dprintk("KVM MMU: slbie(0x%llx, 0x%llx)\n", ea, slbe->esid); esid 578 arch/powerpc/kvm/book3s_64_mmu.c static int segment_contains_magic_page(struct kvm_vcpu *vcpu, ulong esid) esid 583 arch/powerpc/kvm/book3s_64_mmu.c (mp_ea >> SID_SHIFT) == esid; esid 587 arch/powerpc/kvm/book3s_64_mmu.c static int kvmppc_mmu_book3s_64_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid, esid 590 arch/powerpc/kvm/book3s_64_mmu.c ulong ea = esid << SID_SHIFT; esid 592 arch/powerpc/kvm/book3s_64_mmu.c u64 gvsid = esid; esid 604 arch/powerpc/kvm/book3s_64_mmu.c gvsid |= esid & ((1ul << (SID_SHIFT_1T - SID_SHIFT)) - 1); esid 612 arch/powerpc/kvm/book3s_64_mmu.c gvsid = VSID_REAL | esid; esid 639 arch/powerpc/kvm/book3s_64_mmu.c !segment_contains_magic_page(vcpu, esid)) esid 652 arch/powerpc/kvm/book3s_64_mmu.c unlikely(esid == (mp_ea >> SID_SHIFT)) && esid 654 arch/powerpc/kvm/book3s_64_mmu.c *vsid = VSID_REAL | esid; esid 269 arch/powerpc/kvm/book3s_64_mmu_host.c static int kvmppc_mmu_next_segment(struct kvm_vcpu *vcpu, ulong esid) esid 279 arch/powerpc/kvm/book3s_64_mmu_host.c if (!(svcpu->slb[i].esid & SLB_ESID_V)) esid 281 arch/powerpc/kvm/book3s_64_mmu_host.c else if ((svcpu->slb[i].esid & ESID_MASK) == esid) { esid 313 arch/powerpc/kvm/book3s_64_mmu_host.c u64 esid = eaddr >> SID_SHIFT; esid 323 arch/powerpc/kvm/book3s_64_mmu_host.c if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) { esid 325 arch/powerpc/kvm/book3s_64_mmu_host.c svcpu->slb[slb_index].esid = 0; esid 334 arch/powerpc/kvm/book3s_64_mmu_host.c map->guest_esid = esid; esid 346 arch/powerpc/kvm/book3s_64_mmu_host.c svcpu->slb[slb_index].esid = slb_esid; esid 363 arch/powerpc/kvm/book3s_64_mmu_host.c if ((svcpu->slb[i].esid & SLB_ESID_V) && esid 364 arch/powerpc/kvm/book3s_64_mmu_host.c (svcpu->slb[i].esid & seg_mask) == ea) { esid 366 arch/powerpc/kvm/book3s_64_mmu_host.c svcpu->slb[i].esid = 0; esid 377 arch/powerpc/kvm/book3s_64_mmu_host.c svcpu->slb[0].esid = 0; esid 56 arch/powerpc/kvm/book3s_hv_ras.c unsigned long rb = be64_to_cpu(slb->save_area[i].esid); esid 89 arch/powerpc/mm/book3s64/slb.c WRITE_ONCE(p->save_area[index].esid, 0); esid 91 arch/powerpc/mm/book3s64/slb.c WRITE_ONCE(p->save_area[index].esid, cpu_to_be64(mk_esid_data(ea, ssize, index))); esid 96 arch/powerpc/mm/book3s64/slb.c WRITE_ONCE(get_slb_shadow()->save_area[index].esid, cpu_to_be64(index)); esid 130 arch/powerpc/mm/book3s64/slb.c "r" (be64_to_cpu(p->save_area[index].esid))); esid 179 arch/powerpc/mm/book3s64/slb.c "r" (be64_to_cpu(p->save_area[KSTACK_INDEX].esid)) esid 203 arch/powerpc/mm/book3s64/slb.c slb_ptr->esid = e; esid 222 arch/powerpc/mm/book3s64/slb.c e = slb_ptr->esid; esid 267 arch/powerpc/mm/book3s64/slb.c static bool preload_hit(struct thread_info *ti, unsigned long esid) esid 275 arch/powerpc/mm/book3s64/slb.c if (esid == ti->slb_preload_esid[idx]) esid 284 arch/powerpc/mm/book3s64/slb.c unsigned long esid; esid 292 arch/powerpc/mm/book3s64/slb.c esid = ea >> SID_SHIFT; esid 294 arch/powerpc/mm/book3s64/slb.c if (preload_hit(ti, esid)) esid 298 arch/powerpc/mm/book3s64/slb.c ti->slb_preload_esid[idx] = esid; esid 458 arch/powerpc/mm/book3s64/slb.c be64_to_cpu(p->save_area[KSTACK_INDEX].esid); esid 139 arch/powerpc/mm/copro_fault.c slb->esid = (ea & (ssize == MMU_SEGSIZE_1T ? ESID_MASK_1T : ESID_MASK)) | SLB_ESID_V; esid 140 arch/powerpc/platforms/cell/spu_base.c __func__, slbe, slb->vsid, slb->esid); esid 148 arch/powerpc/platforms/cell/spu_base.c out_be64(&priv2->slb_esid_RW, slb->esid); esid 221 arch/powerpc/platforms/cell/spu_base.c slb->esid = (ea & ESID_MASK) | SLB_ESID_V; esid 235 arch/powerpc/platforms/cell/spu_base.c if (!((slbs[i].esid ^ ea) & ESID_MASK)) esid 2457 arch/powerpc/xmon/xmon.c u64 esid, vsid; esid 2462 arch/powerpc/xmon/xmon.c esid = be64_to_cpu(p->slb_shadow_ptr->save_area[i].esid); esid 2465 arch/powerpc/xmon/xmon.c if (esid || vsid) { esid 2467 arch/powerpc/xmon/xmon.c 22, "slb_shadow", i, esid, vsid); esid 3563 arch/powerpc/xmon/xmon.c unsigned long esid,vsid; esid 3569 arch/powerpc/xmon/xmon.c asm volatile("slbmfee %0,%1" : "=r" (esid) : "r" (i)); esid 3572 arch/powerpc/xmon/xmon.c if (!esid && !vsid) esid 3575 arch/powerpc/xmon/xmon.c printf("%02d %016lx %016lx", i, esid, vsid); esid 3577 arch/powerpc/xmon/xmon.c if (!(esid & SLB_ESID_V)) { esid 3585 arch/powerpc/xmon/xmon.c GET_ESID_1T(esid), esid 3590 arch/powerpc/xmon/xmon.c GET_ESID(esid), esid 25 drivers/misc/cxl/fault.c (sste->esid_data == cpu_to_be64(slb->esid))); esid 41 drivers/misc/cxl/fault.c hash = (slb->esid >> SID_SHIFT_1T) & mask; esid 43 drivers/misc/cxl/fault.c hash = (slb->esid >> SID_SHIFT) & mask; esid 75 drivers/misc/cxl/fault.c sste - ctx->sstp, slb->vsid, slb->esid); esid 76 drivers/misc/cxl/fault.c trace_cxl_ste_write(ctx, sste - ctx->sstp, slb->esid, slb->vsid); esid 79 drivers/misc/cxl/fault.c sste->esid_data = cpu_to_be64(slb->esid); esid 332 drivers/misc/cxl/fault.c if (last_esid == slb.esid) esid 336 drivers/misc/cxl/fault.c last_esid = slb.esid; esid 136 drivers/s390/cio/device_id.c cdev->private->flags.esid = 0; esid 161 drivers/s390/cio/device_id.c cdev->private->flags.esid = 1; esid 440 drivers/s390/cio/device_ops.c if (cdev->private->flags.esid == 0) esid 157 drivers/s390/cio/io_sch.h unsigned int esid:1; /* Ext. SenseID supported by HW */ esid 144 tools/testing/selftests/powerpc/vphn/asm/lppaca.h __be64 esid;