fpscr              33 arch/arm/include/asm/fpstate.h 	__u32 fpscr;
fpscr              91 arch/arm/include/asm/user.h 	unsigned long fpscr;
fpscr             660 arch/arm/kernel/ptrace.c 	const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
fpscr             678 arch/arm/kernel/ptrace.c 				   &vfp->fpscr,
fpscr             680 arch/arm/kernel/ptrace.c 				   user_fpscr_offset + sizeof(vfp->fpscr));
fpscr             697 arch/arm/kernel/ptrace.c 	const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
fpscr             716 arch/arm/kernel/ptrace.c 				 &new_vfp.fpscr,
fpscr             718 arch/arm/kernel/ptrace.c 				 user_fpscr_offset + sizeof(new_vfp.fpscr));
fpscr            1111 arch/arm/kvm/coproc.c 		return reg_to_user(uaddr, &vcpu->arch.ctxt.vfp.fpscr, id);
fpscr            1155 arch/arm/kvm/coproc.c 		return reg_from_user(&vcpu->arch.ctxt.vfp.fpscr, uaddr, id);
fpscr             245 arch/arm/vfp/vfp.h #define vfp_single_normaliseround(sd,vsd,fpscr,except,func) __vfp_single_normaliseround(sd,vsd,fpscr,except)
fpscr             246 arch/arm/vfp/vfp.h u32 __vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions);
fpscr             248 arch/arm/vfp/vfp.h u32 vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions, const char *func);
fpscr             343 arch/arm/vfp/vfp.h u32 vfp_double_normaliseround(int dd, struct vfp_double *vd, u32 fpscr, u32 exceptions, const char *func);
fpscr             373 arch/arm/vfp/vfp.h 	u32 (* const fn)(int dd, int dn, int dm, u32 fpscr);
fpscr              70 arch/arm/vfp/vfpdouble.c u32 vfp_double_normaliseround(int dd, struct vfp_double *vd, u32 fpscr, u32 exceptions, const char *func)
fpscr             129 arch/arm/vfp/vfpdouble.c 	rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             208 arch/arm/vfp/vfpdouble.c 		  struct vfp_double *vdm, u32 fpscr)
fpscr             218 arch/arm/vfp/vfpdouble.c 	if (fpscr & FPSCR_DEFAULT_NAN)
fpscr             250 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fabs(int dd, int unused, int dm, u32 fpscr)
fpscr             256 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcpy(int dd, int unused, int dm, u32 fpscr)
fpscr             262 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fneg(int dd, int unused, int dm, u32 fpscr)
fpscr             268 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fsqrt(int dd, int unused, int dm, u32 fpscr)
fpscr             279 arch/arm/vfp/vfpdouble.c 			ret = vfp_propagate_nan(vdp, &vdm, NULL, fpscr);
fpscr             349 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, 0, "fsqrt");
fpscr             358 arch/arm/vfp/vfpdouble.c static u32 vfp_compare(int dd, int signal_on_qnan, int dm, u32 fpscr)
fpscr             419 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcmp(int dd, int unused, int dm, u32 fpscr)
fpscr             421 arch/arm/vfp/vfpdouble.c 	return vfp_compare(dd, 0, dm, fpscr);
fpscr             424 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcmpe(int dd, int unused, int dm, u32 fpscr)
fpscr             426 arch/arm/vfp/vfpdouble.c 	return vfp_compare(dd, 1, dm, fpscr);
fpscr             429 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcmpz(int dd, int unused, int dm, u32 fpscr)
fpscr             431 arch/arm/vfp/vfpdouble.c 	return vfp_compare(dd, 0, VFP_REG_ZERO, fpscr);
fpscr             434 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcmpez(int dd, int unused, int dm, u32 fpscr)
fpscr             436 arch/arm/vfp/vfpdouble.c 	return vfp_compare(dd, 1, VFP_REG_ZERO, fpscr);
fpscr             439 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fcvts(int sd, int unused, int dm, u32 fpscr)
fpscr             475 arch/arm/vfp/vfpdouble.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fcvts");
fpscr             482 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fuito(int dd, int unused, int dm, u32 fpscr)
fpscr             491 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdm, fpscr, 0, "fuito");
fpscr             494 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fsito(int dd, int unused, int dm, u32 fpscr)
fpscr             503 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdm, fpscr, 0, "fsito");
fpscr             506 arch/arm/vfp/vfpdouble.c static u32 vfp_double_ftoui(int sd, int unused, int dm, u32 fpscr)
fpscr             510 arch/arm/vfp/vfpdouble.c 	int rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             580 arch/arm/vfp/vfpdouble.c static u32 vfp_double_ftouiz(int sd, int unused, int dm, u32 fpscr)
fpscr             585 arch/arm/vfp/vfpdouble.c static u32 vfp_double_ftosi(int sd, int unused, int dm, u32 fpscr)
fpscr             589 arch/arm/vfp/vfpdouble.c 	int rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             655 arch/arm/vfp/vfpdouble.c static u32 vfp_double_ftosiz(int dd, int unused, int dm, u32 fpscr)
fpscr             684 arch/arm/vfp/vfpdouble.c 			  struct vfp_double *vdm, u32 fpscr)
fpscr             718 arch/arm/vfp/vfpdouble.c 		return vfp_propagate_nan(vdd, vdn, vdm, fpscr);
fpscr             726 arch/arm/vfp/vfpdouble.c 	       struct vfp_double *vdm, u32 fpscr)
fpscr             754 arch/arm/vfp/vfpdouble.c 		return vfp_double_fadd_nonnumber(vdd, vdn, vdm, fpscr);
fpscr             778 arch/arm/vfp/vfpdouble.c 			vdd->sign = (fpscr & FPSCR_RMODE_MASK) ==
fpscr             791 arch/arm/vfp/vfpdouble.c 		    struct vfp_double *vdm, u32 fpscr)
fpscr             815 arch/arm/vfp/vfpdouble.c 			return vfp_propagate_nan(vdd, vdn, vdm, fpscr);
fpscr             851 arch/arm/vfp/vfpdouble.c vfp_double_multiply_accumulate(int dd, int dn, int dm, u32 fpscr, u32 negate, char *func)
fpscr             864 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_double_multiply(&vdp, &vdn, &vdm, fpscr);
fpscr             874 arch/arm/vfp/vfpdouble.c 	exceptions |= vfp_double_add(&vdd, &vdn, &vdp, fpscr);
fpscr             876 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, func);
fpscr             886 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fmac(int dd, int dn, int dm, u32 fpscr)
fpscr             888 arch/arm/vfp/vfpdouble.c 	return vfp_double_multiply_accumulate(dd, dn, dm, fpscr, 0, "fmac");
fpscr             894 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fnmac(int dd, int dn, int dm, u32 fpscr)
fpscr             896 arch/arm/vfp/vfpdouble.c 	return vfp_double_multiply_accumulate(dd, dn, dm, fpscr, NEG_MULTIPLY, "fnmac");
fpscr             902 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fmsc(int dd, int dn, int dm, u32 fpscr)
fpscr             904 arch/arm/vfp/vfpdouble.c 	return vfp_double_multiply_accumulate(dd, dn, dm, fpscr, NEG_SUBTRACT, "fmsc");
fpscr             910 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fnmsc(int dd, int dn, int dm, u32 fpscr)
fpscr             912 arch/arm/vfp/vfpdouble.c 	return vfp_double_multiply_accumulate(dd, dn, dm, fpscr, NEG_SUBTRACT | NEG_MULTIPLY, "fnmsc");
fpscr             918 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fmul(int dd, int dn, int dm, u32 fpscr)
fpscr             931 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_double_multiply(&vdd, &vdn, &vdm, fpscr);
fpscr             932 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fmul");
fpscr             938 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fnmul(int dd, int dn, int dm, u32 fpscr)
fpscr             951 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_double_multiply(&vdd, &vdn, &vdm, fpscr);
fpscr             954 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fnmul");
fpscr             960 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fadd(int dd, int dn, int dm, u32 fpscr)
fpscr             973 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_double_add(&vdd, &vdn, &vdm, fpscr);
fpscr             975 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fadd");
fpscr             981 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fsub(int dd, int dn, int dm, u32 fpscr)
fpscr             999 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_double_add(&vdd, &vdn, &vdm, fpscr);
fpscr            1001 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fsub");
fpscr            1007 arch/arm/vfp/vfpdouble.c static u32 vfp_double_fdiv(int dd, int dn, int dm, u32 fpscr)
fpscr            1083 arch/arm/vfp/vfpdouble.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, 0, "fdiv");
fpscr            1086 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_propagate_nan(&vdd, &vdn, &vdm, fpscr);
fpscr            1092 arch/arm/vfp/vfpdouble.c 	exceptions = vfp_propagate_nan(&vdd, &vdm, &vdn, fpscr);
fpscr            1127 arch/arm/vfp/vfpdouble.c u32 vfp_double_cpdo(u32 inst, u32 fpscr)
fpscr            1137 arch/arm/vfp/vfpdouble.c 	vecstride = (1 + ((fpscr & FPSCR_STRIDE_MASK) == FPSCR_STRIDE_MASK));
fpscr            1165 arch/arm/vfp/vfpdouble.c 		veclen = fpscr & FPSCR_LENGTH_MASK;
fpscr            1187 arch/arm/vfp/vfpdouble.c 		except = fop->fn(dest, dn, dm, fpscr);
fpscr              82 arch/arm/vfp/vfpinstr.h u32 vfp_single_cpdo(u32 inst, u32 fpscr);
fpscr              83 arch/arm/vfp/vfpinstr.h u32 vfp_single_cprt(u32 inst, u32 fpscr, struct pt_regs *regs);
fpscr              85 arch/arm/vfp/vfpinstr.h u32 vfp_double_cpdo(u32 inst, u32 fpscr);
fpscr             112 arch/arm/vfp/vfpmodule.c 	vfp->hard.fpscr = FPSCR_ROUND_NEAREST;
fpscr             243 arch/arm/vfp/vfpmodule.c static void vfp_raise_exceptions(u32 exceptions, u32 inst, u32 fpscr, struct pt_regs *regs)
fpscr             261 arch/arm/vfp/vfpmodule.c 		fpscr &= ~(FPSCR_N|FPSCR_Z|FPSCR_C|FPSCR_V);
fpscr             263 arch/arm/vfp/vfpmodule.c 	fpscr |= exceptions;
fpscr             265 arch/arm/vfp/vfpmodule.c 	fmxr(FPSCR, fpscr);
fpscr             268 arch/arm/vfp/vfpmodule.c 	if (exceptions & stat && fpscr & en)		\
fpscr             287 arch/arm/vfp/vfpmodule.c static u32 vfp_emulate_instruction(u32 inst, u32 fpscr, struct pt_regs *regs)
fpscr             291 arch/arm/vfp/vfpmodule.c 	pr_debug("VFP: emulate: INST=0x%08x SCR=0x%08x\n", inst, fpscr);
fpscr             299 arch/arm/vfp/vfpmodule.c 				exceptions = vfp_single_cpdo(inst, fpscr);
fpscr             301 arch/arm/vfp/vfpmodule.c 				exceptions = vfp_double_cpdo(inst, fpscr);
fpscr             325 arch/arm/vfp/vfpmodule.c 	u32 fpscr, orig_fpscr, fpsid, exceptions;
fpscr             345 arch/arm/vfp/vfpmodule.c 	orig_fpscr = fpscr = fmrx(FPSCR);
fpscr             351 arch/arm/vfp/vfpmodule.c 	    && (fpscr & FPSCR_IXE)) {
fpscr             373 arch/arm/vfp/vfpmodule.c 		 vfp_raise_exceptions(VFP_EXCEPTION_ERROR, trigger, fpscr, regs);
fpscr             387 arch/arm/vfp/vfpmodule.c 		fpscr &= ~FPSCR_LENGTH_MASK;
fpscr             388 arch/arm/vfp/vfpmodule.c 		fpscr |= (len & FPEXC_LENGTH_MASK) << (FPSCR_LENGTH_BIT - FPEXC_LENGTH_BIT);
fpscr             396 arch/arm/vfp/vfpmodule.c 	exceptions = vfp_emulate_instruction(trigger, fpscr, regs);
fpscr             566 arch/arm/vfp/vfpmodule.c 	ufp->fpscr = hwstate->fpscr;
fpscr             582 arch/arm/vfp/vfpmodule.c 	hwstate->fpscr &= ~(FPSCR_LENGTH_MASK | FPSCR_STRIDE_MASK);
fpscr             604 arch/arm/vfp/vfpmodule.c 	hwstate->fpscr = ufp->fpscr;
fpscr              69 arch/arm/vfp/vfpsingle.c #define vfp_single_normaliseround(sd,vsd,fpscr,except,func) __vfp_single_normaliseround(sd,vsd,fpscr,except)
fpscr              70 arch/arm/vfp/vfpsingle.c u32 __vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions)
fpscr              72 arch/arm/vfp/vfpsingle.c u32 vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions, const char *func)
fpscr             134 arch/arm/vfp/vfpsingle.c 	rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             216 arch/arm/vfp/vfpsingle.c 		  struct vfp_single *vsm, u32 fpscr)
fpscr             226 arch/arm/vfp/vfpsingle.c 	if (fpscr & FPSCR_DEFAULT_NAN)
fpscr             259 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fabs(int sd, int unused, s32 m, u32 fpscr)
fpscr             265 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcpy(int sd, int unused, s32 m, u32 fpscr)
fpscr             271 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fneg(int sd, int unused, s32 m, u32 fpscr)
fpscr             316 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fsqrt(int sd, int unused, s32 m, u32 fpscr)
fpscr             327 arch/arm/vfp/vfpsingle.c 			ret = vfp_propagate_nan(vsp, &vsm, NULL, fpscr);
fpscr             394 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, 0, "fsqrt");
fpscr             403 arch/arm/vfp/vfpsingle.c static u32 vfp_compare(int sd, int signal_on_qnan, s32 m, u32 fpscr)
fpscr             462 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcmp(int sd, int unused, s32 m, u32 fpscr)
fpscr             464 arch/arm/vfp/vfpsingle.c 	return vfp_compare(sd, 0, m, fpscr);
fpscr             467 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcmpe(int sd, int unused, s32 m, u32 fpscr)
fpscr             469 arch/arm/vfp/vfpsingle.c 	return vfp_compare(sd, 1, m, fpscr);
fpscr             472 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcmpz(int sd, int unused, s32 m, u32 fpscr)
fpscr             474 arch/arm/vfp/vfpsingle.c 	return vfp_compare(sd, 0, 0, fpscr);
fpscr             477 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcmpez(int sd, int unused, s32 m, u32 fpscr)
fpscr             479 arch/arm/vfp/vfpsingle.c 	return vfp_compare(sd, 1, 0, fpscr);
fpscr             482 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fcvtd(int dd, int unused, s32 m, u32 fpscr)
fpscr             518 arch/arm/vfp/vfpsingle.c 	return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fcvtd");
fpscr             525 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fuito(int sd, int unused, s32 m, u32 fpscr)
fpscr             533 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vs, fpscr, 0, "fuito");
fpscr             536 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fsito(int sd, int unused, s32 m, u32 fpscr)
fpscr             544 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vs, fpscr, 0, "fsito");
fpscr             547 arch/arm/vfp/vfpsingle.c static u32 vfp_single_ftoui(int sd, int unused, s32 m, u32 fpscr)
fpscr             551 arch/arm/vfp/vfpsingle.c 	int rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             622 arch/arm/vfp/vfpsingle.c static u32 vfp_single_ftouiz(int sd, int unused, s32 m, u32 fpscr)
fpscr             627 arch/arm/vfp/vfpsingle.c static u32 vfp_single_ftosi(int sd, int unused, s32 m, u32 fpscr)
fpscr             631 arch/arm/vfp/vfpsingle.c 	int rmode = fpscr & FPSCR_RMODE_MASK;
fpscr             701 arch/arm/vfp/vfpsingle.c static u32 vfp_single_ftosiz(int sd, int unused, s32 m, u32 fpscr)
fpscr             730 arch/arm/vfp/vfpsingle.c 			  struct vfp_single *vsm, u32 fpscr)
fpscr             764 arch/arm/vfp/vfpsingle.c 		return vfp_propagate_nan(vsd, vsn, vsm, fpscr);
fpscr             772 arch/arm/vfp/vfpsingle.c 	       struct vfp_single *vsm, u32 fpscr)
fpscr             799 arch/arm/vfp/vfpsingle.c 		return vfp_single_fadd_nonnumber(vsd, vsn, vsm, fpscr);
fpscr             823 arch/arm/vfp/vfpsingle.c 			vsd->sign = (fpscr & FPSCR_RMODE_MASK) ==
fpscr             835 arch/arm/vfp/vfpsingle.c vfp_single_multiply(struct vfp_single *vsd, struct vfp_single *vsn, struct vfp_single *vsm, u32 fpscr)
fpscr             859 arch/arm/vfp/vfpsingle.c 			return vfp_propagate_nan(vsd, vsn, vsm, fpscr);
fpscr             895 arch/arm/vfp/vfpsingle.c vfp_single_multiply_accumulate(int sd, int sn, s32 m, u32 fpscr, u32 negate, char *func)
fpscr             911 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_single_multiply(&vsp, &vsn, &vsm, fpscr);
fpscr             923 arch/arm/vfp/vfpsingle.c 	exceptions |= vfp_single_add(&vsd, &vsn, &vsp, fpscr);
fpscr             925 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, func);
fpscr             935 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fmac(int sd, int sn, s32 m, u32 fpscr)
fpscr             937 arch/arm/vfp/vfpsingle.c 	return vfp_single_multiply_accumulate(sd, sn, m, fpscr, 0, "fmac");
fpscr             943 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fnmac(int sd, int sn, s32 m, u32 fpscr)
fpscr             945 arch/arm/vfp/vfpsingle.c 	return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_MULTIPLY, "fnmac");
fpscr             951 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fmsc(int sd, int sn, s32 m, u32 fpscr)
fpscr             953 arch/arm/vfp/vfpsingle.c 	return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_SUBTRACT, "fmsc");
fpscr             959 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fnmsc(int sd, int sn, s32 m, u32 fpscr)
fpscr             961 arch/arm/vfp/vfpsingle.c 	return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_SUBTRACT | NEG_MULTIPLY, "fnmsc");
fpscr             967 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fmul(int sd, int sn, s32 m, u32 fpscr)
fpscr             983 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
fpscr             984 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fmul");
fpscr             990 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fnmul(int sd, int sn, s32 m, u32 fpscr)
fpscr            1006 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
fpscr            1008 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fnmul");
fpscr            1014 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fadd(int sd, int sn, s32 m, u32 fpscr)
fpscr            1033 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_single_add(&vsd, &vsn, &vsm, fpscr);
fpscr            1035 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fadd");
fpscr            1041 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fsub(int sd, int sn, s32 m, u32 fpscr)
fpscr            1046 arch/arm/vfp/vfpsingle.c 	return vfp_single_fadd(sd, sn, vfp_single_packed_negate(m), fpscr);
fpscr            1052 arch/arm/vfp/vfpsingle.c static u32 vfp_single_fdiv(int sd, int sn, s32 m, u32 fpscr)
fpscr            1128 arch/arm/vfp/vfpsingle.c 	return vfp_single_normaliseround(sd, &vsd, fpscr, 0, "fdiv");
fpscr            1131 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_propagate_nan(&vsd, &vsn, &vsm, fpscr);
fpscr            1137 arch/arm/vfp/vfpsingle.c 	exceptions = vfp_propagate_nan(&vsd, &vsm, &vsn, fpscr);
fpscr            1172 arch/arm/vfp/vfpsingle.c u32 vfp_single_cpdo(u32 inst, u32 fpscr)
fpscr            1182 arch/arm/vfp/vfpsingle.c 	vecstride = 1 + ((fpscr & FPSCR_STRIDE_MASK) == FPSCR_STRIDE_MASK);
fpscr            1204 arch/arm/vfp/vfpsingle.c 		veclen = fpscr & FPSCR_LENGTH_MASK;
fpscr            1227 arch/arm/vfp/vfpsingle.c 		except = fop->fn(dest, sn, m, fpscr);
fpscr            1362 arch/arm64/kernel/ptrace.c 	compat_ulong_t fpscr;
fpscr            1382 arch/arm64/kernel/ptrace.c 		fpscr = (uregs->fpsr & VFP_FPSCR_STAT_MASK) |
fpscr            1385 arch/arm64/kernel/ptrace.c 		ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf, &fpscr,
fpscr            1398 arch/arm64/kernel/ptrace.c 	compat_ulong_t fpscr;
fpscr            1411 arch/arm64/kernel/ptrace.c 		ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, &fpscr,
fpscr            1414 arch/arm64/kernel/ptrace.c 			uregs->fpsr = fpscr & VFP_FPSCR_STAT_MASK;
fpscr            1415 arch/arm64/kernel/ptrace.c 			uregs->fpcr = fpscr & VFP_FPSCR_CTRL_MASK;
fpscr              28 arch/arm64/kernel/signal32.c 		compat_ulong_t	fpscr;
fpscr             100 arch/arm64/kernel/signal32.c 	compat_ulong_t fpscr, fpexc;
fpscr             128 arch/arm64/kernel/signal32.c 	fpscr = (fpsimd->fpsr & VFP_FPSCR_STAT_MASK) |
fpscr             130 arch/arm64/kernel/signal32.c 	__put_user_error(fpscr, &frame->ufp.fpscr, err);
fpscr             149 arch/arm64/kernel/signal32.c 	compat_ulong_t fpscr;
fpscr             170 arch/arm64/kernel/signal32.c 	__get_user_error(fpscr, &frame->ufp.fpscr, err);
fpscr             171 arch/arm64/kernel/signal32.c 	fpsimd.fpsr = fpscr & VFP_FPSCR_STAT_MASK;
fpscr             172 arch/arm64/kernel/signal32.c 	fpsimd.fpcr = fpscr & VFP_FPSCR_CTRL_MASK;
fpscr              14 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fres(u64 *fpscr, u32 *dst, u32 *src1);
fpscr              15 arch/powerpc/include/asm/kvm_fpu.h extern void fps_frsqrte(u64 *fpscr, u32 *dst, u32 *src1);
fpscr              16 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fsqrts(u64 *fpscr, u32 *dst, u32 *src1);
fpscr              18 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
fpscr              19 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fdivs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
fpscr              20 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fmuls(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
fpscr              21 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
fpscr              23 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
fpscr              25 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
fpscr              27 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fnmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
fpscr              29 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fnmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
fpscr              31 arch/powerpc/include/asm/kvm_fpu.h extern void fps_fsel(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
fpscr              34 arch/powerpc/include/asm/kvm_fpu.h #define FPD_ONE_IN(name) extern void fpd_ ## name(u64 *fpscr, u32 *cr, \
fpscr              36 arch/powerpc/include/asm/kvm_fpu.h #define FPD_TWO_IN(name) extern void fpd_ ## name(u64 *fpscr, u32 *cr, \
fpscr              38 arch/powerpc/include/asm/kvm_fpu.h #define FPD_THREE_IN(name) extern void fpd_ ## name(u64 *fpscr, u32 *cr, \
fpscr              41 arch/powerpc/include/asm/kvm_fpu.h extern void fpd_fcmpu(u64 *fpscr, u32 *cr, u64 *src1, u64 *src2);
fpscr              42 arch/powerpc/include/asm/kvm_fpu.h extern void fpd_fcmpo(u64 *fpscr, u32 *cr, u64 *src1, u64 *src2);
fpscr              96 arch/powerpc/include/asm/processor.h 	u64	fpscr;		/* Floating point status */
fpscr             128 arch/powerpc/include/asm/sfp-machine.h #define __FPU_FPSCR	(current->thread.fp_state.fpscr)
fpscr             113 arch/powerpc/kernel/asm-offsets.c 	OFFSET(FPSTATE_FPSCR, thread_fp_state, fpscr);
fpscr             458 arch/powerpc/kernel/ptrace.c 	buf[32] = target->thread.fp_state.fpscr;
fpscr             461 arch/powerpc/kernel/ptrace.c 	BUILD_BUG_ON(offsetof(struct thread_fp_state, fpscr) !=
fpscr             496 arch/powerpc/kernel/ptrace.c 	buf[32] = target->thread.fp_state.fpscr;
fpscr             505 arch/powerpc/kernel/ptrace.c 	target->thread.fp_state.fpscr = buf[32];
fpscr             508 arch/powerpc/kernel/ptrace.c 	BUILD_BUG_ON(offsetof(struct thread_fp_state, fpscr) !=
fpscr            1020 arch/powerpc/kernel/ptrace.c 	buf[32] = target->thread.ckfp_state.fpscr;
fpscr            1065 arch/powerpc/kernel/ptrace.c 	buf[32] = target->thread.ckfp_state.fpscr;
fpscr            1073 arch/powerpc/kernel/ptrace.c 	target->thread.ckfp_state.fpscr = buf[32];
fpscr            3020 arch/powerpc/kernel/ptrace.c 				tmp = child->thread.fp_state.fpscr;
fpscr            3053 arch/powerpc/kernel/ptrace.c 				child->thread.fp_state.fpscr = data;
fpscr             248 arch/powerpc/kernel/signal_32.c 	buf[i] = task->thread.fp_state.fpscr;
fpscr             262 arch/powerpc/kernel/signal_32.c 	task->thread.fp_state.fpscr = buf[i];
fpscr             302 arch/powerpc/kernel/signal_32.c 	buf[i] = task->thread.ckfp_state.fpscr;
fpscr             316 arch/powerpc/kernel/signal_32.c 	task->thread.ckfp_state.fpscr = buf[i];
fpscr             962 arch/powerpc/kernel/signal_32.c 	tsk->thread.fp_state.fpscr = 0;	/* turn off all fp exceptions */
fpscr            1419 arch/powerpc/kernel/signal_32.c 	tsk->thread.fp_state.fpscr = 0;	/* turn off all fp exceptions */
fpscr             866 arch/powerpc/kernel/signal_64.c 	tsk->thread.fp_state.fpscr = 0;
fpscr            1152 arch/powerpc/kernel/traps.c static inline int __parse_fpscr(unsigned long fpscr)
fpscr            1157 arch/powerpc/kernel/traps.c 	if ((fpscr & FPSCR_VE) && (fpscr & FPSCR_VX))
fpscr            1161 arch/powerpc/kernel/traps.c 	else if ((fpscr & FPSCR_OE) && (fpscr & FPSCR_OX))
fpscr            1165 arch/powerpc/kernel/traps.c 	else if ((fpscr & FPSCR_UE) && (fpscr & FPSCR_UX))
fpscr            1169 arch/powerpc/kernel/traps.c 	else if ((fpscr & FPSCR_ZE) && (fpscr & FPSCR_ZX))
fpscr            1173 arch/powerpc/kernel/traps.c 	else if ((fpscr & FPSCR_XE) && (fpscr & FPSCR_XX))
fpscr            1185 arch/powerpc/kernel/traps.c 	code = __parse_fpscr(current->thread.fp_state.fpscr);
fpscr            1445 arch/powerpc/kernel/traps.c 			code = __parse_fpscr(current->thread.fp_state.fpscr);
fpscr             623 arch/powerpc/kvm/book3s.c 			*val = get_reg_val(id, vcpu->arch.fp.fpscr);
fpscr             711 arch/powerpc/kvm/book3s.c 			vcpu->arch.fp.fpscr = set_reg_val(id, *val);
fpscr            1798 arch/powerpc/kvm/book3s_hv.c 		*val = get_reg_val(id, vcpu->arch.fp_tm.fpscr);
fpscr            2027 arch/powerpc/kvm/book3s_hv.c 		vcpu->arch.fp_tm.fpscr = set_reg_val(id, *val);
fpscr             493 arch/powerpc/kvm/book3s_paired_singles.c 				      void (*func)(u64 *fpscr,
fpscr             513 arch/powerpc/kvm/book3s_paired_singles.c 	func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in1, &ps0_in2, &ps0_in3);
fpscr             530 arch/powerpc/kvm/book3s_paired_singles.c 		func(&vcpu->arch.fp.fpscr, &qpr[reg_out], &ps1_in1, &ps1_in2, &ps1_in3);
fpscr             541 arch/powerpc/kvm/book3s_paired_singles.c 				    void (*func)(u64 *fpscr,
fpscr             562 arch/powerpc/kvm/book3s_paired_singles.c 	func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in1, &ps0_in2);
fpscr             578 arch/powerpc/kvm/book3s_paired_singles.c 	func(&vcpu->arch.fp.fpscr, &ps1_out, &ps1_in1, &ps1_in2);
fpscr             604 arch/powerpc/kvm/book3s_paired_singles.c 	func(&vcpu->arch.fp.fpscr, &ps0_out, &ps0_in);
fpscr             613 arch/powerpc/kvm/book3s_paired_singles.c 	func(&vcpu->arch.fp.fpscr, &qpr[reg_out], &ps1_in);
fpscr            1099 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1103 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1107 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fdivs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1111 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fres(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1115 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_frsqrtes(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1121 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmuls(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c);
fpscr            1125 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1129 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1133 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fnmsubs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1137 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fnmadds(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1152 arch/powerpc/kvm/book3s_paired_singles.c 			*fpr_d = vcpu->arch.fp.fpscr;
fpscr            1157 arch/powerpc/kvm/book3s_paired_singles.c 			vcpu->arch.fp.fpscr = *fpr_b;
fpscr            1165 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fcmpu(&vcpu->arch.fp.fpscr, &tmp_cr, fpr_a, fpr_b);
fpscr            1176 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fcmpo(&vcpu->arch.fp.fpscr, &tmp_cr, fpr_a, fpr_b);
fpscr            1182 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fneg(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1188 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fabs(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1191 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fcpsgn(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1194 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fdiv(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1197 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1200 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_b);
fpscr            1203 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fctiw(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1206 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fctiwz(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1209 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_frsp(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1217 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fsqrt(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_b);
fpscr            1219 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fdiv(&vcpu->arch.fp.fpscr, &cr, fpr_d, (u64*)&one, fpr_d);
fpscr            1225 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmul(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c);
fpscr            1228 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fsel(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1231 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1234 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fmadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1237 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fnmsub(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1240 arch/powerpc/kvm/book3s_paired_singles.c 			fpd_fnmadd(&vcpu->arch.fp.fpscr, &cr, fpr_d, fpr_a, fpr_c, fpr_b);
fpscr            1580 arch/powerpc/kvm/book3s_pr.c 		*val = get_reg_val(id, vcpu->arch.fp_tm.fpscr);
fpscr            1679 arch/powerpc/kvm/book3s_pr.c 		vcpu->arch.fp_tm.fpscr = set_reg_val(id, *val);
fpscr              15 arch/powerpc/math-emu/fctiwz.c 	u32 fpscr;
fpscr              18 arch/powerpc/math-emu/fctiwz.c 	fpscr = __FPU_FPSCR;
fpscr              26 arch/powerpc/math-emu/fctiwz.c 	__FPU_FPSCR = fpscr;
fpscr             170 arch/powerpc/math-emu/math.c 	u32 fpscr;
fpscr             172 arch/powerpc/math-emu/math.c 	fpscr = __FPU_FPSCR;
fpscr             175 arch/powerpc/math-emu/math.c 		fpscr |= FPSCR_FX;
fpscr             177 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_OX;
fpscr             179 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_UX;
fpscr             181 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_ZX;
fpscr             183 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_XX;
fpscr             185 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VX;
fpscr             187 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXSNAN;
fpscr             189 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXISI;
fpscr             191 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXIDI;
fpscr             193 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXZDZ;
fpscr             195 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXIMZ;
fpscr             197 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXVC;
fpscr             199 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXSOFT;
fpscr             201 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXSQRT;
fpscr             203 arch/powerpc/math-emu/math.c 			fpscr |= FPSCR_VXCVI;
fpscr             207 arch/powerpc/math-emu/math.c 	if (fpscr & (FPSCR_VXSNAN | FPSCR_VXISI | FPSCR_VXIDI |
fpscr             210 arch/powerpc/math-emu/math.c 		fpscr |= FPSCR_VX;
fpscr             212 arch/powerpc/math-emu/math.c 	fpscr &= ~(FPSCR_FEX);
fpscr             213 arch/powerpc/math-emu/math.c 	if (((fpscr & FPSCR_VX) && (fpscr & FPSCR_VE)) ||
fpscr             214 arch/powerpc/math-emu/math.c 	    ((fpscr & FPSCR_OX) && (fpscr & FPSCR_OE)) ||
fpscr             215 arch/powerpc/math-emu/math.c 	    ((fpscr & FPSCR_UX) && (fpscr & FPSCR_UE)) ||
fpscr             216 arch/powerpc/math-emu/math.c 	    ((fpscr & FPSCR_ZX) && (fpscr & FPSCR_ZE)) ||
fpscr             217 arch/powerpc/math-emu/math.c 	    ((fpscr & FPSCR_XX) && (fpscr & FPSCR_XE)))
fpscr             218 arch/powerpc/math-emu/math.c 		fpscr |= FPSCR_FEX;
fpscr             220 arch/powerpc/math-emu/math.c 	__FPU_FPSCR = fpscr;
fpscr             222 arch/powerpc/math-emu/math.c 	return (fpscr & FPSCR_FEX) ? 1 : 0;
fpscr              13 arch/powerpc/math-emu/mtfsf.c 	u32 fpscr;
fpscr              30 arch/powerpc/math-emu/mtfsf.c 	fpscr = ((__FPU_FPSCR & ~mask) | (frB[1] & mask)) &
fpscr              33 arch/powerpc/math-emu/mtfsf.c 	if (fpscr & (FPSCR_VXSNAN | FPSCR_VXISI | FPSCR_VXIDI |
fpscr              36 arch/powerpc/math-emu/mtfsf.c 		fpscr |= FPSCR_VX;
fpscr              42 arch/powerpc/math-emu/mtfsf.c 	if (fpscr & (fpscr >> 22) &  0xf8)
fpscr              43 arch/powerpc/math-emu/mtfsf.c 		fpscr |= FPSCR_FEX;
fpscr              45 arch/powerpc/math-emu/mtfsf.c 	__FPU_FPSCR = fpscr;
fpscr              69 arch/sh/include/asm/processor_32.h 	unsigned long fpscr;
fpscr              79 arch/sh/include/asm/processor_32.h 	unsigned long fpscr;
fpscr              67 arch/sh/include/asm/processor_64.h 	unsigned int fpscr;
fpscr              74 arch/sh/include/asm/processor_64.h 	unsigned int fpscr;
fpscr              34 arch/sh/include/asm/user.h 	unsigned int fpscr;
fpscr              40 arch/sh/include/asm/user.h 	unsigned long fpscr;
fpscr              32 arch/sh/kernel/cpu/fpu.c 		fp->fpscr = FPSCR_INIT;
fpscr              36 arch/sh/kernel/cpu/fpu.c 		fp->fpscr = FPSCR_INIT;
fpscr             456 arch/sh/kernel/cpu/sh2a/fpu.c 		if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_FPU_ERROR)) {
fpscr             467 arch/sh/kernel/cpu/sh2a/fpu.c 		int fpscr;
fpscr             475 arch/sh/kernel/cpu/sh2a/fpu.c 		fpscr = tsk->thread.xstate->hardfpu.fpscr;
fpscr             476 arch/sh/kernel/cpu/sh2a/fpu.c 		prec = fpscr & (1 << 19);
fpscr             478 arch/sh/kernel/cpu/sh2a/fpu.c 		if ((fpscr & FPSCR_FPU_ERROR)
fpscr             494 arch/sh/kernel/cpu/sh2a/fpu.c 		} else if ((fpscr & FPSCR_FPU_ERROR)
fpscr             510 arch/sh/kernel/cpu/sh2a/fpu.c 		int fpscr;
fpscr             518 arch/sh/kernel/cpu/sh2a/fpu.c 		fpscr = tsk->thread.xstate->hardfpu.fpscr;
fpscr             519 arch/sh/kernel/cpu/sh2a/fpu.c 		prec = fpscr & (1 << 19);
fpscr             521 arch/sh/kernel/cpu/sh2a/fpu.c 		if ((fpscr & FPSCR_FPU_ERROR)
fpscr             537 arch/sh/kernel/cpu/sh2a/fpu.c 		} else if ((fpscr & FPSCR_FPU_ERROR)
fpscr             563 arch/sh/kernel/cpu/sh2a/fpu.c 		tsk->thread.xstate->hardfpu.fpscr &=
fpscr             230 arch/sh/kernel/cpu/sh4/fpu.c 		if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR))
fpscr             242 arch/sh/kernel/cpu/sh4/fpu.c 		int fpscr;
fpscr             250 arch/sh/kernel/cpu/sh4/fpu.c 		fpscr = tsk->thread.xstate->hardfpu.fpscr;
fpscr             251 arch/sh/kernel/cpu/sh4/fpu.c 		prec = fpscr & FPSCR_DBL_PRECISION;
fpscr             253 arch/sh/kernel/cpu/sh4/fpu.c 		if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             266 arch/sh/kernel/cpu/sh4/fpu.c 		} else if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             280 arch/sh/kernel/cpu/sh4/fpu.c 		int fpscr;
fpscr             288 arch/sh/kernel/cpu/sh4/fpu.c 		fpscr = tsk->thread.xstate->hardfpu.fpscr;
fpscr             289 arch/sh/kernel/cpu/sh4/fpu.c 		prec = fpscr & FPSCR_DBL_PRECISION;
fpscr             291 arch/sh/kernel/cpu/sh4/fpu.c 		if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             307 arch/sh/kernel/cpu/sh4/fpu.c 		} else if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             324 arch/sh/kernel/cpu/sh4/fpu.c 		int fpscr;
fpscr             332 arch/sh/kernel/cpu/sh4/fpu.c 		fpscr = tsk->thread.xstate->hardfpu.fpscr;
fpscr             333 arch/sh/kernel/cpu/sh4/fpu.c 		prec = fpscr & FPSCR_DBL_PRECISION;
fpscr             335 arch/sh/kernel/cpu/sh4/fpu.c 		if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             350 arch/sh/kernel/cpu/sh4/fpu.c 		} else if ((fpscr & FPSCR_CAUSE_ERROR)
fpscr             370 arch/sh/kernel/cpu/sh4/fpu.c 		if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR)
fpscr             397 arch/sh/kernel/cpu/sh4/fpu.c 	int roundingMode = FPSCR_ROUNDING_MODE(tsk->thread.xstate->hardfpu.fpscr);
fpscr             409 arch/sh/kernel/cpu/sh4/fpu.c 		tsk->thread.xstate->hardfpu.fpscr &=
fpscr             411 arch/sh/kernel/cpu/sh4/fpu.c 		tsk->thread.xstate->hardfpu.fpscr |= fpu_exception_flags;
fpscr             414 arch/sh/kernel/cpu/sh4/fpu.c 		tsk->thread.xstate->hardfpu.fpscr |= (fpu_exception_flags >> 10);
fpscr             418 arch/sh/kernel/cpu/sh4/fpu.c 		if ((((tsk->thread.xstate->hardfpu.fpscr & FPSCR_ENABLE_MASK) >> 7) &
fpscr             389 arch/sh/kernel/ptrace_32.c 				if (addr == offsetof(struct user, fpu.fpscr))
fpscr              73 arch/sh/kernel/ptrace_64.c 		if (addr == offsetof(struct user_fpu_struct, fpscr)) {
fpscr              27 arch/sh/math-emu/math.c #define FPSCR		(fregs->fpscr)
fpscr             554 arch/sh/math-emu/math.c 		if ((tsk->thread.xstate->softfpu.fpscr & (1 << 17))) {
fpscr             558 arch/sh/math-emu/math.c 			tsk->thread.xstate->softfpu.fpscr &=
fpscr             581 arch/sh/math-emu/math.c 	fpu->fpscr = FPSCR_INIT;
fpscr              34 tools/testing/selftests/powerpc/ptrace/ptrace.h 	unsigned long fpscr;