MSR_VEC            49 arch/powerpc/include/asm/switch_to.h 	msr_check_and_clear(MSR_VEC);
MSR_VEC            61 arch/powerpc/include/asm/switch_to.h 	msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
MSR_VEC           262 arch/powerpc/kernel/process.c 	msr &= ~MSR_VEC;
MSR_VEC           274 arch/powerpc/kernel/process.c 	msr_check_and_set(MSR_VEC);
MSR_VEC           276 arch/powerpc/kernel/process.c 	msr_check_and_clear(MSR_VEC);
MSR_VEC           286 arch/powerpc/kernel/process.c 	cpumsr = msr_check_and_set(MSR_VEC);
MSR_VEC           288 arch/powerpc/kernel/process.c 	if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) {
MSR_VEC           313 arch/powerpc/kernel/process.c 		if (tsk->thread.regs->msr & MSR_VEC) {
MSR_VEC           347 arch/powerpc/kernel/process.c 	WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC)));
MSR_VEC           352 arch/powerpc/kernel/process.c 	if (msr & MSR_VEC)
MSR_VEC           360 arch/powerpc/kernel/process.c 	msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
MSR_VEC           362 arch/powerpc/kernel/process.c 	msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
MSR_VEC           371 arch/powerpc/kernel/process.c 	cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
MSR_VEC           374 arch/powerpc/kernel/process.c 	    (current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) {
MSR_VEC           395 arch/powerpc/kernel/process.c 		if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) {
MSR_VEC           464 arch/powerpc/kernel/process.c 		msr_all_available |= MSR_VEC;
MSR_VEC           495 arch/powerpc/kernel/process.c 	WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
MSR_VEC           502 arch/powerpc/kernel/process.c 	if (usermsr & MSR_VEC)
MSR_VEC           542 arch/powerpc/kernel/process.c 	if ((!(msr & MSR_VEC)) && restore_altivec(current))
MSR_VEC           543 arch/powerpc/kernel/process.c 		msr |= MSR_VEC;
MSR_VEC           545 arch/powerpc/kernel/process.c 	if ((msr & (MSR_FP | MSR_VEC)) == (MSR_FP | MSR_VEC) &&
MSR_VEC           569 arch/powerpc/kernel/process.c 	WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
MSR_VEC           574 arch/powerpc/kernel/process.c 	if (usermsr & MSR_VEC)
MSR_VEC           868 arch/powerpc/kernel/process.c 	if ((thr->ckpt_regs.msr & MSR_VEC) == 0)
MSR_VEC           978 arch/powerpc/kernel/process.c 	new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC          1032 arch/powerpc/kernel/process.c 	msr_diff &= MSR_FP | MSR_VEC | MSR_VSX;
MSR_VEC          1038 arch/powerpc/kernel/process.c 	if (cpu_has_feature(CPU_FTR_ALTIVEC) && msr_diff & MSR_VEC)
MSR_VEC          1307 arch/powerpc/kernel/process.c 	{MSR_VEC,	"VEC"},
MSR_VEC          1641 arch/powerpc/kernel/process.c 	childregs->msr &= ~(MSR_FP|MSR_VEC|MSR_VSX);
MSR_VEC           405 arch/powerpc/kernel/signal_32.c 		msr |= MSR_VEC;
MSR_VEC           517 arch/powerpc/kernel/signal_32.c 		if (msr & MSR_VEC) {
MSR_VEC           532 arch/powerpc/kernel/signal_32.c 		msr |= MSR_VEC;
MSR_VEC           545 arch/powerpc/kernel/signal_32.c 	if (msr & MSR_VEC) {
MSR_VEC           659 arch/powerpc/kernel/signal_32.c 	regs->msr &= ~MSR_VEC;
MSR_VEC           660 arch/powerpc/kernel/signal_32.c 	if (msr & MSR_VEC) {
MSR_VEC           762 arch/powerpc/kernel/signal_32.c 	regs->msr &= ~MSR_VEC;
MSR_VEC           763 arch/powerpc/kernel/signal_32.c 	if (msr & MSR_VEC) {
MSR_VEC           868 arch/powerpc/kernel/signal_32.c 	msr_check_and_set(msr & (MSR_FP | MSR_VEC));
MSR_VEC           874 arch/powerpc/kernel/signal_32.c 	if (msr & MSR_VEC) {
MSR_VEC           876 arch/powerpc/kernel/signal_32.c 		regs->msr |= MSR_VEC;
MSR_VEC           127 arch/powerpc/kernel/signal_64.c 		msr |= MSR_VEC;
MSR_VEC           223 arch/powerpc/kernel/signal_64.c 	msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC           237 arch/powerpc/kernel/signal_64.c 		if (msr & MSR_VEC)
MSR_VEC           249 arch/powerpc/kernel/signal_64.c 		msr |= MSR_VEC;
MSR_VEC           257 arch/powerpc/kernel/signal_64.c 	if (msr & MSR_VEC)
MSR_VEC           369 arch/powerpc/kernel/signal_64.c 	regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX);
MSR_VEC           378 arch/powerpc/kernel/signal_64.c 	if (v_regs != NULL && (msr & MSR_VEC) != 0) {
MSR_VEC           489 arch/powerpc/kernel/signal_64.c 	regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX);
MSR_VEC           501 arch/powerpc/kernel/signal_64.c 	if (v_regs != NULL && tm_v_regs != NULL && (msr & MSR_VEC) != 0) {
MSR_VEC           582 arch/powerpc/kernel/signal_64.c 	msr_check_and_set(msr & (MSR_FP | MSR_VEC));
MSR_VEC           587 arch/powerpc/kernel/signal_64.c 	if (msr & MSR_VEC) {
MSR_VEC           589 arch/powerpc/kernel/signal_64.c 		regs->msr |= MSR_VEC;
MSR_VEC           968 arch/powerpc/kernel/traps.c 		msr_mask = MSR_VEC;
MSR_VEC          3561 arch/powerpc/kvm/book3s_hv.c 	msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC          3667 arch/powerpc/kvm/book3s_hv.c 	msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC           136 arch/powerpc/kvm/book3s_pr.c 	kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC           325 arch/powerpc/kvm/book3s_pr.c 		(MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC           332 arch/powerpc/kvm/book3s_pr.c 	else if (ext_diff == MSR_VEC)
MSR_VEC           793 arch/powerpc/kvm/book3s_pr.c 		msr |= MSR_FP | MSR_VEC;
MSR_VEC           815 arch/powerpc/kvm/book3s_pr.c 	if (msr & MSR_VEC) {
MSR_VEC           816 arch/powerpc/kvm/book3s_pr.c 		if (current->thread.regs->msr & MSR_VEC)
MSR_VEC           874 arch/powerpc/kvm/book3s_pr.c 		msr = MSR_FP | MSR_VEC | MSR_VSX;
MSR_VEC           895 arch/powerpc/kvm/book3s_pr.c 	if (msr & MSR_VEC) {
MSR_VEC           933 arch/powerpc/kvm/book3s_pr.c 	if (lost_ext & MSR_VEC) {
MSR_VEC          1351 arch/powerpc/kvm/book3s_pr.c 			ext_msr = MSR_VEC;
MSR_VEC          1849 arch/powerpc/kvm/book3s_pr.c 	kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
MSR_VEC           177 arch/powerpc/kvm/booke.c 		if (!(current->thread.regs->msr & MSR_VEC)) {
MSR_VEC           182 arch/powerpc/kvm/booke.c 			current->thread.regs->msr |= MSR_VEC;
MSR_VEC           196 arch/powerpc/kvm/booke.c 		if (current->thread.regs->msr & MSR_VEC)
MSR_VEC            54 arch/powerpc/kvm/emulate_loadstore.c 	if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) {
MSR_VEC           272 arch/powerpc/kvm/emulate_loadstore.c 						MSR_VEC);
MSR_VEC          1195 arch/powerpc/kvm/powerpc.c 			vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC);
MSR_VEC           576 arch/powerpc/lib/sstep.c 	if (regs->msr & MSR_VEC)
MSR_VEC           599 arch/powerpc/lib/sstep.c 	if (regs->msr & MSR_VEC)
MSR_VEC           839 arch/powerpc/lib/sstep.c 		if (regs->msr & MSR_VEC)
MSR_VEC           870 arch/powerpc/lib/sstep.c 		if (regs->msr & MSR_VEC)
MSR_VEC          2970 arch/powerpc/lib/sstep.c 		if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC))
MSR_VEC          2984 arch/powerpc/lib/sstep.c 			msrbit = MSR_VEC;
MSR_VEC          3041 arch/powerpc/lib/sstep.c 		if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC))
MSR_VEC          3055 arch/powerpc/lib/sstep.c 			msrbit = MSR_VEC;
MSR_VEC            78 arch/powerpc/lib/test_emulate_step.c 	regs->msr |= MSR_VEC;