gvsid             358 arch/powerpc/kvm/book3s_32_mmu.c 	u64 gvsid = esid;
gvsid             364 arch/powerpc/kvm/book3s_32_mmu.c 			gvsid = sr_vsid(sr);
gvsid             375 arch/powerpc/kvm/book3s_32_mmu.c 		*vsid = VSID_REAL_IR | gvsid;
gvsid             378 arch/powerpc/kvm/book3s_32_mmu.c 		*vsid = VSID_REAL_DR | gvsid;
gvsid             384 arch/powerpc/kvm/book3s_32_mmu.c 			*vsid = VSID_BAT | gvsid;
gvsid              66 arch/powerpc/kvm/book3s_32_mmu_host.c static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid              68 arch/powerpc/kvm/book3s_32_mmu_host.c 	return (u16)(((gvsid >> (SID_MAP_BITS * 7)) & SID_MAP_MASK) ^
gvsid              69 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 6)) & SID_MAP_MASK) ^
gvsid              70 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 5)) & SID_MAP_MASK) ^
gvsid              71 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 4)) & SID_MAP_MASK) ^
gvsid              72 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 3)) & SID_MAP_MASK) ^
gvsid              73 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 2)) & SID_MAP_MASK) ^
gvsid              74 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 1)) & SID_MAP_MASK) ^
gvsid              75 arch/powerpc/kvm/book3s_32_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 0)) & SID_MAP_MASK));
gvsid              79 arch/powerpc/kvm/book3s_32_mmu_host.c static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid              85 arch/powerpc/kvm/book3s_32_mmu_host.c 		gvsid |= VSID_PR;
gvsid              87 arch/powerpc/kvm/book3s_32_mmu_host.c 	sid_map_mask = kvmppc_sid_hash(vcpu, gvsid);
gvsid              89 arch/powerpc/kvm/book3s_32_mmu_host.c 	if (map->guest_vsid == gvsid) {
gvsid              91 arch/powerpc/kvm/book3s_32_mmu_host.c 			    gvsid, map->host_vsid);
gvsid              96 arch/powerpc/kvm/book3s_32_mmu_host.c 	if (map->guest_vsid == gvsid) {
gvsid              98 arch/powerpc/kvm/book3s_32_mmu_host.c 			    gvsid, map->host_vsid);
gvsid             102 arch/powerpc/kvm/book3s_32_mmu_host.c 	dprintk_sr("SR: Searching 0x%llx -> not found\n", gvsid);
gvsid             263 arch/powerpc/kvm/book3s_32_mmu_host.c static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid             271 arch/powerpc/kvm/book3s_32_mmu_host.c 		gvsid |= VSID_PR;
gvsid             276 arch/powerpc/kvm/book3s_32_mmu_host.c 	sid_map_mask = kvmppc_sid_hash(vcpu, gvsid);
gvsid             296 arch/powerpc/kvm/book3s_32_mmu_host.c 	map->guest_vsid = gvsid;
gvsid             305 arch/powerpc/kvm/book3s_32_mmu_host.c 	u64 gvsid;
gvsid             311 arch/powerpc/kvm/book3s_32_mmu_host.c 	if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) {
gvsid             318 arch/powerpc/kvm/book3s_32_mmu_host.c 	map = find_sid_vsid(vcpu, gvsid);
gvsid             320 arch/powerpc/kvm/book3s_32_mmu_host.c 		map = create_sid_map(vcpu, gvsid);
gvsid             592 arch/powerpc/kvm/book3s_64_mmu.c 	u64 gvsid = esid;
gvsid             600 arch/powerpc/kvm/book3s_64_mmu.c 			gvsid = slb->vsid;
gvsid             603 arch/powerpc/kvm/book3s_64_mmu.c 				gvsid <<= SID_SHIFT_1T - SID_SHIFT;
gvsid             604 arch/powerpc/kvm/book3s_64_mmu.c 				gvsid |= esid & ((1ul << (SID_SHIFT_1T - SID_SHIFT)) - 1);
gvsid             605 arch/powerpc/kvm/book3s_64_mmu.c 				gvsid |= VSID_1T;
gvsid             612 arch/powerpc/kvm/book3s_64_mmu.c 		gvsid = VSID_REAL | esid;
gvsid             615 arch/powerpc/kvm/book3s_64_mmu.c 		gvsid |= VSID_REAL_IR;
gvsid             618 arch/powerpc/kvm/book3s_64_mmu.c 		gvsid |= VSID_REAL_DR;
gvsid             640 arch/powerpc/kvm/book3s_64_mmu.c 		gvsid |= VSID_64K;
gvsid             644 arch/powerpc/kvm/book3s_64_mmu.c 		gvsid |= VSID_PR;
gvsid             646 arch/powerpc/kvm/book3s_64_mmu.c 	*vsid = gvsid;
gvsid              32 arch/powerpc/kvm/book3s_64_mmu_host.c static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid              34 arch/powerpc/kvm/book3s_64_mmu_host.c 	return (u16)(((gvsid >> (SID_MAP_BITS * 7)) & SID_MAP_MASK) ^
gvsid              35 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 6)) & SID_MAP_MASK) ^
gvsid              36 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 5)) & SID_MAP_MASK) ^
gvsid              37 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 4)) & SID_MAP_MASK) ^
gvsid              38 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 3)) & SID_MAP_MASK) ^
gvsid              39 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 2)) & SID_MAP_MASK) ^
gvsid              40 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 1)) & SID_MAP_MASK) ^
gvsid              41 arch/powerpc/kvm/book3s_64_mmu_host.c 		     ((gvsid >> (SID_MAP_BITS * 0)) & SID_MAP_MASK));
gvsid              45 arch/powerpc/kvm/book3s_64_mmu_host.c static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid              51 arch/powerpc/kvm/book3s_64_mmu_host.c 		gvsid |= VSID_PR;
gvsid              53 arch/powerpc/kvm/book3s_64_mmu_host.c 	sid_map_mask = kvmppc_sid_hash(vcpu, gvsid);
gvsid              55 arch/powerpc/kvm/book3s_64_mmu_host.c 	if (map->valid && (map->guest_vsid == gvsid)) {
gvsid              56 arch/powerpc/kvm/book3s_64_mmu_host.c 		trace_kvm_book3s_slb_found(gvsid, map->host_vsid);
gvsid              61 arch/powerpc/kvm/book3s_64_mmu_host.c 	if (map->valid && (map->guest_vsid == gvsid)) {
gvsid              62 arch/powerpc/kvm/book3s_64_mmu_host.c 		trace_kvm_book3s_slb_found(gvsid, map->host_vsid);
gvsid              66 arch/powerpc/kvm/book3s_64_mmu_host.c 	trace_kvm_book3s_slb_fail(sid_map_mask, gvsid);
gvsid             223 arch/powerpc/kvm/book3s_64_mmu_host.c static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid)
gvsid             232 arch/powerpc/kvm/book3s_64_mmu_host.c 		gvsid |= VSID_PR;
gvsid             237 arch/powerpc/kvm/book3s_64_mmu_host.c 	sid_map_mask = kvmppc_sid_hash(vcpu, gvsid);
gvsid             261 arch/powerpc/kvm/book3s_64_mmu_host.c 	map->guest_vsid = gvsid;
gvsid             264 arch/powerpc/kvm/book3s_64_mmu_host.c 	trace_kvm_book3s_slb_map(sid_map_mask, gvsid, map->host_vsid);
gvsid             316 arch/powerpc/kvm/book3s_64_mmu_host.c 	u64 gvsid;
gvsid             323 arch/powerpc/kvm/book3s_64_mmu_host.c 	if (vcpu->arch.mmu.esid_to_vsid(vcpu, esid, &gvsid)) {
gvsid             330 arch/powerpc/kvm/book3s_64_mmu_host.c 	map = find_sid_vsid(vcpu, gvsid);
gvsid             332 arch/powerpc/kvm/book3s_64_mmu_host.c 		map = create_sid_map(vcpu, gvsid);
gvsid             342 arch/powerpc/kvm/book3s_64_mmu_host.c 	if (gvsid & VSID_64K)
gvsid             145 arch/powerpc/kvm/trace_pr.h 	TP_PROTO(unsigned long long gvsid, unsigned long long hvsid),
gvsid             146 arch/powerpc/kvm/trace_pr.h 	TP_ARGS(gvsid, hvsid),
gvsid             149 arch/powerpc/kvm/trace_pr.h 		__field(	unsigned long long,	gvsid		)
gvsid             154 arch/powerpc/kvm/trace_pr.h 		__entry->gvsid		= gvsid;
gvsid             158 arch/powerpc/kvm/trace_pr.h 	TP_printk("%llx -> %llx", __entry->gvsid, __entry->hvsid)
gvsid             162 arch/powerpc/kvm/trace_pr.h 	TP_PROTO(u16 sid_map_mask, unsigned long long gvsid),
gvsid             163 arch/powerpc/kvm/trace_pr.h 	TP_ARGS(sid_map_mask, gvsid),
gvsid             167 arch/powerpc/kvm/trace_pr.h 		__field(	unsigned long long,	gvsid		)
gvsid             172 arch/powerpc/kvm/trace_pr.h 		__entry->gvsid		= gvsid;
gvsid             176 arch/powerpc/kvm/trace_pr.h 		  SID_MAP_MASK - __entry->sid_map_mask, __entry->gvsid)
gvsid             180 arch/powerpc/kvm/trace_pr.h 	TP_PROTO(u16 sid_map_mask, unsigned long long gvsid,
gvsid             182 arch/powerpc/kvm/trace_pr.h 	TP_ARGS(sid_map_mask, gvsid, hvsid),
gvsid             192 arch/powerpc/kvm/trace_pr.h 		__entry->guest_vsid	= gvsid;