MSR_SPE 88 arch/powerpc/include/asm/reg.h #ifndef MSR_SPE MSR_SPE 72 arch/powerpc/include/asm/switch_to.h msr_check_and_clear(MSR_SPE); MSR_SPE 422 arch/powerpc/kernel/process.c msr_check_and_set(MSR_SPE); MSR_SPE 424 arch/powerpc/kernel/process.c msr_check_and_clear(MSR_SPE); MSR_SPE 432 arch/powerpc/kernel/process.c msr_check_and_set(MSR_SPE); MSR_SPE 434 arch/powerpc/kernel/process.c if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) { MSR_SPE 445 arch/powerpc/kernel/process.c if (tsk->thread.regs->msr & MSR_SPE) { MSR_SPE 472 arch/powerpc/kernel/process.c msr_all_available |= MSR_SPE; MSR_SPE 506 arch/powerpc/kernel/process.c if (usermsr & MSR_SPE) MSR_SPE 577 arch/powerpc/kernel/process.c if (usermsr & MSR_SPE) MSR_SPE 590 arch/powerpc/kernel/process.c if (tsk->thread.regs->msr & MSR_SPE) MSR_SPE 451 arch/powerpc/kernel/signal_32.c msr |= MSR_SPE; MSR_SPE 599 arch/powerpc/kernel/signal_32.c msr |= MSR_SPE; MSR_SPE 706 arch/powerpc/kernel/signal_32.c regs->msr &= ~MSR_SPE; MSR_SPE 707 arch/powerpc/kernel/signal_32.c if (msr & MSR_SPE) { MSR_SPE 817 arch/powerpc/kernel/signal_32.c regs->msr &= ~MSR_SPE; MSR_SPE 818 arch/powerpc/kernel/signal_32.c if (msr & MSR_SPE) { MSR_SPE 2149 arch/powerpc/kernel/traps.c if (regs->msr & MSR_SPE) MSR_SPE 94 arch/powerpc/kvm/booke.c vcpu->arch.shadow_msr &= ~MSR_SPE; MSR_SPE 104 arch/powerpc/kvm/booke.c vcpu->arch.shadow_msr |= MSR_SPE; MSR_SPE 110 arch/powerpc/kvm/booke.c if (vcpu->arch.shared->msr & MSR_SPE) { MSR_SPE 111 arch/powerpc/kvm/booke.c if (!(vcpu->arch.shadow_msr & MSR_SPE)) MSR_SPE 113 arch/powerpc/kvm/booke.c } else if (vcpu->arch.shadow_msr & MSR_SPE) { MSR_SPE 1129 arch/powerpc/kvm/booke.c if (vcpu->arch.shared->msr & MSR_SPE) MSR_SPE 310 arch/powerpc/kvm/e500.c if (vcpu->arch.shadow_msr & MSR_SPE)