kvmppc_get_msr 388 arch/powerpc/include/asm/kvm_book3s.h static inline u64 kvmppc_get_msr(struct kvm_vcpu *vcpu); kvmppc_get_msr 391 arch/powerpc/include/asm/kvm_book3s.h return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); kvmppc_get_msr 1041 arch/powerpc/include/asm/kvm_ppc.h if (!(kvmppc_get_msr(vcpu) & msr_64bit)) kvmppc_get_msr 122 arch/powerpc/kvm/book3s.c if (!(kvmppc_get_msr(vcpu) & MSR_SF)) { kvmppc_get_msr 130 arch/powerpc/kvm/book3s.c crit = crit && !(kvmppc_get_msr(vcpu) & MSR_PR); kvmppc_get_msr 139 arch/powerpc/kvm/book3s.c kvmppc_set_srr1(vcpu, (kvmppc_get_msr(vcpu) & ~0x783f0000ul) | flags); kvmppc_get_msr 303 arch/powerpc/kvm/book3s.c deliver = (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; kvmppc_get_msr 307 arch/powerpc/kvm/book3s.c deliver = (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; kvmppc_get_msr 432 arch/powerpc/kvm/book3s.c if (!(kvmppc_get_msr(vcpu) & MSR_SF)) kvmppc_get_msr 457 arch/powerpc/kvm/book3s.c int relocated = (kvmppc_get_msr(vcpu) & (data ? MSR_DR : MSR_IR)); kvmppc_get_msr 471 arch/powerpc/kvm/book3s.c if ((kvmppc_get_msr(vcpu) & (MSR_IR | MSR_DR)) == MSR_DR && kvmppc_get_msr 546 arch/powerpc/kvm/book3s.c regs->msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 146 arch/powerpc/kvm/book3s_32_mmu.c if (kvmppc_get_msr(vcpu) & MSR_PR) { kvmppc_get_msr 229 arch/powerpc/kvm/book3s_32_mmu.c if ((sr_kp(sre) && (kvmppc_get_msr(vcpu) & MSR_PR)) || kvmppc_get_msr 230 arch/powerpc/kvm/book3s_32_mmu.c (sr_ks(sre) && !(kvmppc_get_msr(vcpu) & MSR_PR))) kvmppc_get_msr 308 arch/powerpc/kvm/book3s_32_mmu.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 359 arch/powerpc/kvm/book3s_32_mmu.c u64 msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 84 arch/powerpc/kvm/book3s_32_mmu_host.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 270 arch/powerpc/kvm/book3s_32_mmu_host.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 30 arch/powerpc/kvm/book3s_64_mmu.c unsigned long cur_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 226 arch/powerpc/kvm/book3s_64_mmu.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 271 arch/powerpc/kvm/book3s_64_mmu.c if ((kvmppc_get_msr(vcpu) & MSR_PR) && slbe->Kp) kvmppc_get_msr 273 arch/powerpc/kvm/book3s_64_mmu.c else if (!(kvmppc_get_msr(vcpu) & MSR_PR) && slbe->Ks) kvmppc_get_msr 316 arch/powerpc/kvm/book3s_64_mmu.c !(kvmppc_get_msr(vcpu) & MSR_PR)) kvmppc_get_msr 498 arch/powerpc/kvm/book3s_64_mmu.c if (kvmppc_get_msr(vcpu) & MSR_IR) { kvmppc_get_msr 582 arch/powerpc/kvm/book3s_64_mmu.c return mp_ea && !(kvmppc_get_msr(vcpu) & MSR_PR) && kvmppc_get_msr 595 arch/powerpc/kvm/book3s_64_mmu.c u64 msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 643 arch/powerpc/kvm/book3s_64_mmu.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 653 arch/powerpc/kvm/book3s_64_mmu.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 50 arch/powerpc/kvm/book3s_64_mmu_host.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 231 arch/powerpc/kvm/book3s_64_mmu_host.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 274 arch/powerpc/kvm/book3s_64_mmu_radix.c if (kvmppc_get_msr(vcpu) & MSR_PR) { kvmppc_get_msr 81 arch/powerpc/kvm/book3s_emulate.c if ((kvmppc_get_msr(vcpu) & MSR_PR) && level > PRIV_PROBLEM) kvmppc_get_msr 128 arch/powerpc/kvm/book3s_emulate.c unsigned long guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 149 arch/powerpc/kvm/book3s_emulate.c if (kvmppc_get_msr(vcpu) & MSR_PR) kvmppc_get_msr 152 arch/powerpc/kvm/book3s_emulate.c if (kvmppc_get_msr(vcpu) & MSR_HV) kvmppc_get_msr 174 arch/powerpc/kvm/book3s_emulate.c unsigned long guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 203 arch/powerpc/kvm/book3s_emulate.c unsigned long guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 251 arch/powerpc/kvm/book3s_emulate.c if ((kvmppc_get_msr(vcpu) & MSR_LE) && kvmppc_get_msr 271 arch/powerpc/kvm/book3s_emulate.c unsigned long cur_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 299 arch/powerpc/kvm/book3s_emulate.c kvmppc_set_gpr(vcpu, rt, kvmppc_get_msr(vcpu)); kvmppc_get_msr 305 arch/powerpc/kvm/book3s_emulate.c ulong new_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 365 arch/powerpc/kvm/book3s_emulate.c if ((kvmppc_get_msr(vcpu) & MSR_PR) || kvmppc_get_msr 464 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_SF)) kvmppc_get_msr 495 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 501 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 513 arch/powerpc/kvm/book3s_emulate.c if (kvmppc_get_msr(vcpu) & MSR_HV) kvmppc_get_msr 527 arch/powerpc/kvm/book3s_emulate.c ulong guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 533 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 553 arch/powerpc/kvm/book3s_emulate.c ulong guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 559 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 587 arch/powerpc/kvm/book3s_emulate.c ulong guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 593 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 789 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 795 arch/powerpc/kvm/book3s_emulate.c if (MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)) && kvmppc_get_msr 796 arch/powerpc/kvm/book3s_emulate.c !((MSR_TM_SUSPENDED(kvmppc_get_msr(vcpu))) && kvmppc_get_msr 849 arch/powerpc/kvm/book3s_emulate.c if (kvmppc_get_msr(vcpu) & MSR_PR) { kvmppc_get_msr 854 arch/powerpc/kvm/book3s_emulate.c if ((kvmppc_get_msr(vcpu) & MSR_PR) || sprn == 0) { kvmppc_get_msr 967 arch/powerpc/kvm/book3s_emulate.c if (!(kvmppc_get_msr(vcpu) & MSR_TM)) { kvmppc_get_msr 1014 arch/powerpc/kvm/book3s_emulate.c if (kvmppc_get_msr(vcpu) & MSR_PR) { kvmppc_get_msr 1019 arch/powerpc/kvm/book3s_emulate.c if ((kvmppc_get_msr(vcpu) & MSR_PR) || sprn == 0 || kvmppc_get_msr 1502 arch/powerpc/kvm/book3s_hv.c vcpu->arch.fault_dsisr = kvmppc_get_msr(vcpu) & kvmppc_get_msr 158 arch/powerpc/kvm/book3s_paired_singles.c u64 msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 656 arch/powerpc/kvm/book3s_paired_singles.c if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { kvmppc_get_msr 68 arch/powerpc/kvm/book3s_pr.c ulong msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 74 arch/powerpc/kvm/book3s_pr.c ulong msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 190 arch/powerpc/kvm/book3s_pr.c ulong guest_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 282 arch/powerpc/kvm/book3s_pr.c old_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 324 arch/powerpc/kvm/book3s_pr.c ulong ext_diff = (kvmppc_get_msr(vcpu) & ~vcpu->arch.guest_owned_ext) & kvmppc_get_msr 342 arch/powerpc/kvm/book3s_pr.c if (!(MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)))) { kvmppc_get_msr 357 arch/powerpc/kvm/book3s_pr.c if (!MSR_TM_ACTIVE(kvmppc_get_msr(vcpu))) { kvmppc_get_msr 359 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_TM) { kvmppc_get_msr 368 arch/powerpc/kvm/book3s_pr.c _kvmppc_restore_tm_pr(vcpu, kvmppc_get_msr(vcpu)); kvmppc_get_msr 371 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_TM) { kvmppc_get_msr 473 arch/powerpc/kvm/book3s_pr.c old_msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 495 arch/powerpc/kvm/book3s_pr.c if ((kvmppc_get_msr(vcpu) & (MSR_PR|MSR_IR|MSR_DR)) != kvmppc_get_msr 527 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_FP) kvmppc_get_msr 531 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_TM) kvmppc_get_msr 656 arch/powerpc/kvm/book3s_pr.c if (!(kvmppc_get_msr(vcpu) & MSR_SF)) kvmppc_get_msr 676 arch/powerpc/kvm/book3s_pr.c bool dr = (kvmppc_get_msr(vcpu) & MSR_DR) ? true : false; kvmppc_get_msr 677 arch/powerpc/kvm/book3s_pr.c bool ir = (kvmppc_get_msr(vcpu) & MSR_IR) ? true : false; kvmppc_get_msr 698 arch/powerpc/kvm/book3s_pr.c switch (kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) { kvmppc_get_msr 711 arch/powerpc/kvm/book3s_pr.c if ((kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) == MSR_DR) kvmppc_get_msr 855 arch/powerpc/kvm/book3s_pr.c if (!(kvmppc_get_msr(vcpu) & msr)) { kvmppc_get_msr 958 arch/powerpc/kvm/book3s_pr.c if (!(kvmppc_get_msr(vcpu) & MSR_PR)) kvmppc_get_msr 983 arch/powerpc/kvm/book3s_pr.c guest_fac_enabled = kvmppc_get_msr(vcpu) & MSR_TM; kvmppc_get_msr 1016 arch/powerpc/kvm/book3s_pr.c if ((fac == FSCR_TM_LG) && !(kvmppc_get_msr(vcpu) & MSR_PR)) kvmppc_get_msr 1041 arch/powerpc/kvm/book3s_pr.c u64 msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 1050 arch/powerpc/kvm/book3s_pr.c u64 msr = kvmppc_get_msr(vcpu); kvmppc_get_msr 1079 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_PR) { kvmppc_get_msr 1279 arch/powerpc/kvm/book3s_pr.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 1311 arch/powerpc/kvm/book3s_pr.c } else if (!(kvmppc_get_msr(vcpu) & MSR_PR) && kvmppc_get_msr 1836 arch/powerpc/kvm/book3s_pr.c if (kvmppc_get_msr(vcpu) & MSR_FP) kvmppc_get_msr 378 arch/powerpc/kvm/book3s_pr_papr.c kvmppc_set_msr_fast(vcpu, kvmppc_get_msr(vcpu) | MSR_EE); kvmppc_get_msr 30 arch/powerpc/kvm/emulate_loadstore.c if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { kvmppc_get_msr 42 arch/powerpc/kvm/emulate_loadstore.c if (!(kvmppc_get_msr(vcpu) & MSR_VSX)) { kvmppc_get_msr 54 arch/powerpc/kvm/emulate_loadstore.c if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) { kvmppc_get_msr 170 arch/powerpc/kvm/powerpc.c if (!(kvmppc_get_msr(vcpu) & MSR_SF)) { kvmppc_get_msr 352 arch/powerpc/kvm/powerpc.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 398 arch/powerpc/kvm/powerpc.c !(kvmppc_get_msr(vcpu) & MSR_PR)) { kvmppc_get_msr 234 arch/powerpc/kvm/trace_pr.h __entry->msr = kvmppc_get_msr(vcpu);