kvm_vcpu_get_hsr  164 arch/arm/include/asm/kvm_emulate.h 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  184 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_ISV;
kvm_vcpu_get_hsr  189 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_WNR;
kvm_vcpu_get_hsr  194 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_SSE;
kvm_vcpu_get_hsr  204 arch/arm/include/asm/kvm_emulate.h 	return (kvm_vcpu_get_hsr(vcpu) & HSR_SRT_MASK) >> HSR_SRT_SHIFT;
kvm_vcpu_get_hsr  209 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_DABT_S1PTW;
kvm_vcpu_get_hsr  214 arch/arm/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & HSR_DABT_CM);
kvm_vcpu_get_hsr  220 arch/arm/include/asm/kvm_emulate.h 	switch ((kvm_vcpu_get_hsr(vcpu) >> 22) & 0x3) {
kvm_vcpu_get_hsr  236 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_IL;
kvm_vcpu_get_hsr  241 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) >> HSR_EC_SHIFT;
kvm_vcpu_get_hsr  251 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_FSC;
kvm_vcpu_get_hsr  256 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_FSC_TYPE;
kvm_vcpu_get_hsr  288 arch/arm/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & HSR_HVC_IMM_MASK;
kvm_vcpu_get_hsr  611 arch/arm/kvm/coproc.c 	params.CRn = (kvm_vcpu_get_hsr(vcpu) >> 1) & 0xf;
kvm_vcpu_get_hsr  612 arch/arm/kvm/coproc.c 	params.Rt1 = (kvm_vcpu_get_hsr(vcpu) >> 5) & 0xf;
kvm_vcpu_get_hsr  613 arch/arm/kvm/coproc.c 	params.is_write = ((kvm_vcpu_get_hsr(vcpu) & 1) == 0);
kvm_vcpu_get_hsr  616 arch/arm/kvm/coproc.c 	params.Op1 = (kvm_vcpu_get_hsr(vcpu) >> 16) & 0xf;
kvm_vcpu_get_hsr  618 arch/arm/kvm/coproc.c 	params.Rt2 = (kvm_vcpu_get_hsr(vcpu) >> 10) & 0xf;
kvm_vcpu_get_hsr  676 arch/arm/kvm/coproc.c 	params.CRm = (kvm_vcpu_get_hsr(vcpu) >> 1) & 0xf;
kvm_vcpu_get_hsr  677 arch/arm/kvm/coproc.c 	params.Rt1 = (kvm_vcpu_get_hsr(vcpu) >> 5) & 0xf;
kvm_vcpu_get_hsr  678 arch/arm/kvm/coproc.c 	params.is_write = ((kvm_vcpu_get_hsr(vcpu) & 1) == 0);
kvm_vcpu_get_hsr  681 arch/arm/kvm/coproc.c 	params.CRn = (kvm_vcpu_get_hsr(vcpu) >> 10) & 0xf;
kvm_vcpu_get_hsr  682 arch/arm/kvm/coproc.c 	params.Op1 = (kvm_vcpu_get_hsr(vcpu) >> 14) & 0x7;
kvm_vcpu_get_hsr  683 arch/arm/kvm/coproc.c 	params.Op2 = (kvm_vcpu_get_hsr(vcpu) >> 17) & 0x7;
kvm_vcpu_get_hsr   64 arch/arm/kvm/handle_exit.c 	if (kvm_vcpu_get_hsr(vcpu) & HSR_WFI_IS_WFE) {
kvm_vcpu_get_hsr   82 arch/arm/kvm/handle_exit.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  254 arch/arm64/include/asm/kvm_emulate.h 	u32 esr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  279 arch/arm64/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_xVC_IMM_MASK;
kvm_vcpu_get_hsr  284 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_ISV);
kvm_vcpu_get_hsr  289 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SSE);
kvm_vcpu_get_hsr  294 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SF);
kvm_vcpu_get_hsr  299 arch/arm64/include/asm/kvm_emulate.h 	return (kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT;
kvm_vcpu_get_hsr  304 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_S1PTW);
kvm_vcpu_get_hsr  309 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_WNR) ||
kvm_vcpu_get_hsr  315 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_CM);
kvm_vcpu_get_hsr  320 arch/arm64/include/asm/kvm_emulate.h 	return 1 << ((kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT);
kvm_vcpu_get_hsr  326 arch/arm64/include/asm/kvm_emulate.h 	return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_IL);
kvm_vcpu_get_hsr  331 arch/arm64/include/asm/kvm_emulate.h 	return ESR_ELx_EC(kvm_vcpu_get_hsr(vcpu));
kvm_vcpu_get_hsr  341 arch/arm64/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_FSC;
kvm_vcpu_get_hsr  346 arch/arm64/include/asm/kvm_emulate.h 	return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_FSC_TYPE;
kvm_vcpu_get_hsr  370 arch/arm64/include/asm/kvm_emulate.h 	u32 esr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr   92 arch/arm64/kvm/handle_exit.c 	if (kvm_vcpu_get_hsr(vcpu) & ESR_ELx_WFx_ISS_WFE) {
kvm_vcpu_get_hsr  122 arch/arm64/kvm/handle_exit.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  149 arch/arm64/kvm/handle_exit.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  214 arch/arm64/kvm/handle_exit.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr  255 arch/arm64/kvm/handle_exit.c 		u8 hsr_ec = ESR_ELx_EC(kvm_vcpu_get_hsr(vcpu));
kvm_vcpu_get_hsr  321 arch/arm64/kvm/handle_exit.c 		kvm_handle_guest_serror(vcpu, kvm_vcpu_get_hsr(vcpu));
kvm_vcpu_get_hsr  398 arch/arm64/kvm/hyp/switch.c 	u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_hsr(vcpu));
kvm_vcpu_get_hsr 2125 arch/arm64/kvm/sys_regs.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr 2182 arch/arm64/kvm/sys_regs.c 	u32 hsr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr 2290 arch/arm64/kvm/sys_regs.c 	unsigned long esr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr   54 virt/kvm/arm/hyp/aarch32.c 	if (kvm_vcpu_get_hsr(vcpu) >> 30)
kvm_vcpu_get_hsr  444 virt/kvm/arm/hyp/vgic-v3-sr.c 	u32 esr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr 1010 virt/kvm/arm/hyp/vgic-v3-sr.c 	esr = kvm_vcpu_get_hsr(vcpu);
kvm_vcpu_get_hsr 1929 virt/kvm/arm/mmu.c 		if (!kvm_handle_guest_sea(fault_ipa, kvm_vcpu_get_hsr(vcpu)))
kvm_vcpu_get_hsr 1938 virt/kvm/arm/mmu.c 	trace_kvm_guest_fault(*vcpu_pc(vcpu), kvm_vcpu_get_hsr(vcpu),
kvm_vcpu_get_hsr 1947 virt/kvm/arm/mmu.c 			(unsigned long)kvm_vcpu_get_hsr(vcpu));