kvm_vcpu_get_hsr 164 arch/arm/include/asm/kvm_emulate.h u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 184 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_ISV; kvm_vcpu_get_hsr 189 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_WNR; kvm_vcpu_get_hsr 194 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_SSE; kvm_vcpu_get_hsr 204 arch/arm/include/asm/kvm_emulate.h return (kvm_vcpu_get_hsr(vcpu) & HSR_SRT_MASK) >> HSR_SRT_SHIFT; kvm_vcpu_get_hsr 209 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_DABT_S1PTW; kvm_vcpu_get_hsr 214 arch/arm/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & HSR_DABT_CM); kvm_vcpu_get_hsr 220 arch/arm/include/asm/kvm_emulate.h switch ((kvm_vcpu_get_hsr(vcpu) >> 22) & 0x3) { kvm_vcpu_get_hsr 236 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_IL; kvm_vcpu_get_hsr 241 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) >> HSR_EC_SHIFT; kvm_vcpu_get_hsr 251 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_FSC; kvm_vcpu_get_hsr 256 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_FSC_TYPE; kvm_vcpu_get_hsr 288 arch/arm/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & HSR_HVC_IMM_MASK; kvm_vcpu_get_hsr 611 arch/arm/kvm/coproc.c params.CRn = (kvm_vcpu_get_hsr(vcpu) >> 1) & 0xf; kvm_vcpu_get_hsr 612 arch/arm/kvm/coproc.c params.Rt1 = (kvm_vcpu_get_hsr(vcpu) >> 5) & 0xf; kvm_vcpu_get_hsr 613 arch/arm/kvm/coproc.c params.is_write = ((kvm_vcpu_get_hsr(vcpu) & 1) == 0); kvm_vcpu_get_hsr 616 arch/arm/kvm/coproc.c params.Op1 = (kvm_vcpu_get_hsr(vcpu) >> 16) & 0xf; kvm_vcpu_get_hsr 618 arch/arm/kvm/coproc.c params.Rt2 = (kvm_vcpu_get_hsr(vcpu) >> 10) & 0xf; kvm_vcpu_get_hsr 676 arch/arm/kvm/coproc.c params.CRm = (kvm_vcpu_get_hsr(vcpu) >> 1) & 0xf; kvm_vcpu_get_hsr 677 arch/arm/kvm/coproc.c params.Rt1 = (kvm_vcpu_get_hsr(vcpu) >> 5) & 0xf; kvm_vcpu_get_hsr 678 arch/arm/kvm/coproc.c params.is_write = ((kvm_vcpu_get_hsr(vcpu) & 1) == 0); kvm_vcpu_get_hsr 681 arch/arm/kvm/coproc.c params.CRn = (kvm_vcpu_get_hsr(vcpu) >> 10) & 0xf; kvm_vcpu_get_hsr 682 arch/arm/kvm/coproc.c params.Op1 = (kvm_vcpu_get_hsr(vcpu) >> 14) & 0x7; kvm_vcpu_get_hsr 683 arch/arm/kvm/coproc.c params.Op2 = (kvm_vcpu_get_hsr(vcpu) >> 17) & 0x7; kvm_vcpu_get_hsr 64 arch/arm/kvm/handle_exit.c if (kvm_vcpu_get_hsr(vcpu) & HSR_WFI_IS_WFE) { kvm_vcpu_get_hsr 82 arch/arm/kvm/handle_exit.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 254 arch/arm64/include/asm/kvm_emulate.h u32 esr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 279 arch/arm64/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_xVC_IMM_MASK; kvm_vcpu_get_hsr 284 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_ISV); kvm_vcpu_get_hsr 289 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SSE); kvm_vcpu_get_hsr 294 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SF); kvm_vcpu_get_hsr 299 arch/arm64/include/asm/kvm_emulate.h return (kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT; kvm_vcpu_get_hsr 304 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_S1PTW); kvm_vcpu_get_hsr 309 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_WNR) || kvm_vcpu_get_hsr 315 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_CM); kvm_vcpu_get_hsr 320 arch/arm64/include/asm/kvm_emulate.h return 1 << ((kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT); kvm_vcpu_get_hsr 326 arch/arm64/include/asm/kvm_emulate.h return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_IL); kvm_vcpu_get_hsr 331 arch/arm64/include/asm/kvm_emulate.h return ESR_ELx_EC(kvm_vcpu_get_hsr(vcpu)); kvm_vcpu_get_hsr 341 arch/arm64/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_FSC; kvm_vcpu_get_hsr 346 arch/arm64/include/asm/kvm_emulate.h return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_FSC_TYPE; kvm_vcpu_get_hsr 370 arch/arm64/include/asm/kvm_emulate.h u32 esr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 92 arch/arm64/kvm/handle_exit.c if (kvm_vcpu_get_hsr(vcpu) & ESR_ELx_WFx_ISS_WFE) { kvm_vcpu_get_hsr 122 arch/arm64/kvm/handle_exit.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 149 arch/arm64/kvm/handle_exit.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 214 arch/arm64/kvm/handle_exit.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 255 arch/arm64/kvm/handle_exit.c u8 hsr_ec = ESR_ELx_EC(kvm_vcpu_get_hsr(vcpu)); kvm_vcpu_get_hsr 321 arch/arm64/kvm/handle_exit.c kvm_handle_guest_serror(vcpu, kvm_vcpu_get_hsr(vcpu)); kvm_vcpu_get_hsr 398 arch/arm64/kvm/hyp/switch.c u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_hsr(vcpu)); kvm_vcpu_get_hsr 2125 arch/arm64/kvm/sys_regs.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 2182 arch/arm64/kvm/sys_regs.c u32 hsr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 2290 arch/arm64/kvm/sys_regs.c unsigned long esr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 54 virt/kvm/arm/hyp/aarch32.c if (kvm_vcpu_get_hsr(vcpu) >> 30) kvm_vcpu_get_hsr 444 virt/kvm/arm/hyp/vgic-v3-sr.c u32 esr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 1010 virt/kvm/arm/hyp/vgic-v3-sr.c esr = kvm_vcpu_get_hsr(vcpu); kvm_vcpu_get_hsr 1929 virt/kvm/arm/mmu.c if (!kvm_handle_guest_sea(fault_ipa, kvm_vcpu_get_hsr(vcpu))) kvm_vcpu_get_hsr 1938 virt/kvm/arm/mmu.c trace_kvm_guest_fault(*vcpu_pc(vcpu), kvm_vcpu_get_hsr(vcpu), kvm_vcpu_get_hsr 1947 virt/kvm/arm/mmu.c (unsigned long)kvm_vcpu_get_hsr(vcpu));