Lines Matching refs:current

415 	flush_fp_to_thread(current);  in save_user_regs()
423 if (current->thread.used_vr) { in save_user_regs()
424 flush_altivec_to_thread(current); in save_user_regs()
425 if (__copy_to_user(&frame->mc_vregs, &current->thread.vr_state, in save_user_regs()
441 current->thread.vrsave = mfspr(SPRN_VRSAVE); in save_user_regs()
442 if (__put_user(current->thread.vrsave, (u32 __user *)&frame->mc_vregs[32])) in save_user_regs()
445 if (copy_fpr_to_user(&frame->mc_fregs, current)) in save_user_regs()
460 if (current->thread.used_vsr && ctx_has_vsx_region) { in save_user_regs()
461 __giveup_vsx(current); in save_user_regs()
462 if (copy_vsx_to_user(&frame->mc_vsregs, current)) in save_user_regs()
469 if (current->thread.used_spe) { in save_user_regs()
470 flush_spe_to_thread(current); in save_user_regs()
471 if (__copy_to_user(&frame->mc_vregs, current->thread.evr, in save_user_regs()
481 if (__put_user(current->thread.spefscr, (u32 __user *)&frame->mc_vregs + ELF_NEVRREG)) in save_user_regs()
529 flush_fp_to_thread(current); in save_tm_user_regs()
532 if (save_general_regs(&current->thread.ckpt_regs, frame) in save_tm_user_regs()
547 if (current->thread.used_vr) { in save_tm_user_regs()
548 flush_altivec_to_thread(current); in save_tm_user_regs()
549 if (__copy_to_user(&frame->mc_vregs, &current->thread.vr_state, in save_tm_user_regs()
554 &current->thread.transact_vr, in save_tm_user_regs()
559 &current->thread.vr_state, in save_tm_user_regs()
576 current->thread.vrsave = mfspr(SPRN_VRSAVE); in save_tm_user_regs()
577 if (__put_user(current->thread.vrsave, in save_tm_user_regs()
581 if (__put_user(current->thread.transact_vrsave, in save_tm_user_regs()
585 if (__put_user(current->thread.vrsave, in save_tm_user_regs()
591 if (copy_fpr_to_user(&frame->mc_fregs, current)) in save_tm_user_regs()
594 if (copy_transact_fpr_to_user(&tm_frame->mc_fregs, current)) in save_tm_user_regs()
597 if (copy_fpr_to_user(&tm_frame->mc_fregs, current)) in save_tm_user_regs()
608 if (current->thread.used_vsr) { in save_tm_user_regs()
609 __giveup_vsx(current); in save_tm_user_regs()
610 if (copy_vsx_to_user(&frame->mc_vsregs, current)) in save_tm_user_regs()
614 current)) in save_tm_user_regs()
617 if (copy_vsx_to_user(&tm_frame->mc_vsregs, current)) in save_tm_user_regs()
628 if (current->thread.used_spe) { in save_tm_user_regs()
629 flush_spe_to_thread(current); in save_tm_user_regs()
630 if (__copy_to_user(&frame->mc_vregs, current->thread.evr, in save_tm_user_regs()
639 if (__put_user(current->thread.spefscr, (u32 __user *)&frame->mc_vregs + ELF_NEVRREG)) in save_tm_user_regs()
707 if (__copy_from_user(&current->thread.vr_state, &sr->mc_vregs, in restore_user_regs()
710 } else if (current->thread.used_vr) in restore_user_regs()
711 memset(&current->thread.vr_state, 0, in restore_user_regs()
715 if (__get_user(current->thread.vrsave, (u32 __user *)&sr->mc_vregs[32])) in restore_user_regs()
718 mtspr(SPRN_VRSAVE, current->thread.vrsave); in restore_user_regs()
720 if (copy_fpr_from_user(current, &sr->mc_fregs)) in restore_user_regs()
734 if (copy_vsx_from_user(current, &sr->mc_vsregs)) in restore_user_regs()
736 } else if (current->thread.used_vsr) in restore_user_regs()
738 current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_user_regs()
752 if (__copy_from_user(current->thread.evr, &sr->mc_vregs, in restore_user_regs()
755 } else if (current->thread.used_spe) in restore_user_regs()
756 memset(current->thread.evr, 0, ELF_NEVRREG * sizeof(u32)); in restore_user_regs()
759 if (__get_user(current->thread.spefscr, (u32 __user *)&sr->mc_vregs + ELF_NEVRREG)) in restore_user_regs()
790 err |= restore_general_regs(&current->thread.ckpt_regs, sr); in restore_tm_user_regs()
792 err |= __get_user(current->thread.tm_tfhar, &sr->mc_gregs[PT_NIP]); in restore_tm_user_regs()
814 if (__copy_from_user(&current->thread.vr_state, &sr->mc_vregs, in restore_tm_user_regs()
816 __copy_from_user(&current->thread.transact_vr, in restore_tm_user_regs()
820 } else if (current->thread.used_vr) { in restore_tm_user_regs()
821 memset(&current->thread.vr_state, 0, in restore_tm_user_regs()
823 memset(&current->thread.transact_vr, 0, in restore_tm_user_regs()
828 if (__get_user(current->thread.vrsave, in restore_tm_user_regs()
830 __get_user(current->thread.transact_vrsave, in restore_tm_user_regs()
834 mtspr(SPRN_VRSAVE, current->thread.vrsave); in restore_tm_user_regs()
839 if (copy_fpr_from_user(current, &sr->mc_fregs) || in restore_tm_user_regs()
840 copy_transact_fpr_from_user(current, &tm_sr->mc_fregs)) in restore_tm_user_regs()
850 if (copy_vsx_from_user(current, &sr->mc_vsregs) || in restore_tm_user_regs()
851 copy_transact_vsx_from_user(current, &tm_sr->mc_vsregs)) in restore_tm_user_regs()
853 } else if (current->thread.used_vsr) in restore_tm_user_regs()
855 current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_tm_user_regs()
856 current->thread.transact_fp.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_tm_user_regs()
866 if (__copy_from_user(current->thread.evr, &sr->mc_vregs, in restore_tm_user_regs()
869 } else if (current->thread.used_spe) in restore_tm_user_regs()
870 memset(current->thread.evr, 0, ELF_NEVRREG * sizeof(u32)); in restore_tm_user_regs()
873 if (__get_user(current->thread.spefscr, (u32 __user *)&sr->mc_vregs in restore_tm_user_regs()
893 current->thread.tm_texasr |= TEXASR_FS; in restore_tm_user_regs()
895 tm_recheckpoint(&current->thread, msr); in restore_tm_user_regs()
899 do_load_up_transact_fpu(&current->thread); in restore_tm_user_regs()
900 regs->msr |= (MSR_FP | current->thread.fpexc_mode); in restore_tm_user_regs()
904 do_load_up_transact_altivec(&current->thread); in restore_tm_user_regs()
1021 if (vdso32_rt_sigtramp && current->mm->context.vdso_base) { in handle_rt_signal32()
1023 tramp = current->mm->context.vdso_base + vdso32_rt_sigtramp; in handle_rt_signal32()
1050 current->thread.fp_state.fpscr = 0; /* turn off all fp exceptions */ in handle_rt_signal32()
1075 current->comm, current->pid, in handle_rt_signal32()
1201 || put_sigset_t(&old_ctx->uc_sigmask, &current->blocked) in sys_swapcontext()
1241 current->restart_block.fn = do_no_restart_syscall; in sys_rt_sigreturn()
1300 current->comm, current->pid, in sys_rt_sigreturn()
1303 force_sig(SIGSEGV, current); in sys_rt_sigreturn()
1318 unsigned long new_dbcr0 = current->thread.debug.dbcr0; in sys_debug_setcontext()
1333 current->thread.debug.dbcr1)) { in sys_debug_setcontext()
1368 current->thread.debug.dbcr0 = new_dbcr0; in sys_debug_setcontext()
1392 current->comm, current->pid, in sys_debug_setcontext()
1395 force_sig(SIGSEGV, current); in sys_debug_setcontext()
1446 if (vdso32_sigtramp && current->mm->context.vdso_base) { in handle_signal32()
1448 tramp = current->mm->context.vdso_base + vdso32_sigtramp; in handle_signal32()
1470 current->thread.fp_state.fpscr = 0; /* turn off all fp exceptions */ in handle_signal32()
1490 current->comm, current->pid, in handle_signal32()
1514 current->restart_block.fn = do_no_restart_syscall; in sys_sigreturn()
1562 current->comm, current->pid, in sys_sigreturn()
1565 force_sig(SIGSEGV, current); in sys_sigreturn()