Lines Matching refs:msr
79 printk("pc: %08lx msr: %08llx\n", vcpu->arch.pc, vcpu->arch.shared->msr); in kvmppc_dump_vcpu()
116 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()
141 if (!(current->thread.regs->msr & MSR_FP)) { in kvmppc_load_guest_fp()
145 current->thread.regs->msr |= MSR_FP; in kvmppc_load_guest_fp()
157 if (current->thread.regs->msr & MSR_FP) in kvmppc_save_guest_fp()
169 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_FP; in kvmppc_vcpu_sync_fpu()
182 if (!(current->thread.regs->msr & MSR_VEC)) { in kvmppc_load_guest_altivec()
186 current->thread.regs->msr |= MSR_VEC; in kvmppc_load_guest_altivec()
200 if (current->thread.regs->msr & MSR_VEC) in kvmppc_save_guest_altivec()
212 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_DE; in kvmppc_vcpu_sync_debug()
222 vcpu->arch.shared->msr |= MSR_DE; in kvmppc_vcpu_sync_debug()
225 vcpu->arch.shared->msr &= ~MSR_DE; in kvmppc_vcpu_sync_debug()
236 u32 old_msr = vcpu->arch.shared->msr; in kvmppc_set_msr()
242 vcpu->arch.shared->msr = new_msr; in kvmppc_set_msr()
390 ulong new_msr = vcpu->arch.shared->msr; in kvmppc_booke_irqprio_deliver()
393 if (!(vcpu->arch.shared->msr & MSR_SF)) { in kvmppc_booke_irqprio_deliver()
401 crit = crit && !(vcpu->arch.shared->msr & MSR_PR); in kvmppc_booke_irqprio_deliver()
441 allowed = vcpu->arch.shared->msr & MSR_CE; in kvmppc_booke_irqprio_deliver()
447 allowed = vcpu->arch.shared->msr & MSR_ME; in kvmppc_booke_irqprio_deliver()
457 allowed = vcpu->arch.shared->msr & MSR_EE; in kvmppc_booke_irqprio_deliver()
463 allowed = vcpu->arch.shared->msr & MSR_DE; in kvmppc_booke_irqprio_deliver()
478 vcpu->arch.shared->msr); in kvmppc_booke_irqprio_deliver()
482 vcpu->arch.shared->msr); in kvmppc_booke_irqprio_deliver()
486 vcpu->arch.shared->msr); in kvmppc_booke_irqprio_deliver()
490 vcpu->arch.shared->msr); in kvmppc_booke_irqprio_deliver()
685 if (vcpu->arch.shared->msr & MSR_WE) { in kvmppc_core_prepare_to_enter()
844 if (dbsr && (vcpu->arch.shared->msr & MSR_DE) && in kvmppc_handle_debug()
849 if ((dbsr & DBSR_TIE) && !(vcpu->arch.shared->msr & MSR_DE)) in kvmppc_handle_debug()
881 ulong r1, ip, msr, lr; in kvmppc_fill_pt_regs() local
885 asm("mfmsr %0" : "=r"(msr)); in kvmppc_fill_pt_regs()
891 regs->msr = msr; in kvmppc_fill_pt_regs()
1094 if (vcpu->arch.shared->msr & (MSR_PR | MSR_GS)) { in kvmppc_handle_exit()
1120 if (vcpu->arch.shared->msr & MSR_SPE) in kvmppc_handle_exit()
1198 if (!(vcpu->arch.shared->msr & MSR_PR)) { in kvmppc_handle_exit()
1212 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_handle_exit()
1233 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_handle_exit()
1385 vcpu->arch.shared->msr = 0; in kvm_arch_vcpu_setup()
1430 regs->msr = vcpu->arch.shared->msr; in kvm_arch_vcpu_ioctl_get_regs()
1458 kvmppc_set_msr(vcpu, regs->msr); in kvm_arch_vcpu_ioctl_set_regs()
1941 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_xlate()