Lines Matching refs:fpscr

69 #define vfp_single_normaliseround(sd,vsd,fpscr,except,func) __vfp_single_normaliseround(sd,vsd,fpscr,except)
70 u32 __vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions)
72 u32 vfp_single_normaliseround(int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions, const char *func)
134 rmode = fpscr & FPSCR_RMODE_MASK;
216 struct vfp_single *vsm, u32 fpscr)
226 if (fpscr & FPSCR_DEFAULT_NAN)
259 static u32 vfp_single_fabs(int sd, int unused, s32 m, u32 fpscr)
265 static u32 vfp_single_fcpy(int sd, int unused, s32 m, u32 fpscr)
271 static u32 vfp_single_fneg(int sd, int unused, s32 m, u32 fpscr)
316 static u32 vfp_single_fsqrt(int sd, int unused, s32 m, u32 fpscr)
327 ret = vfp_propagate_nan(vsp, &vsm, NULL, fpscr);
394 return vfp_single_normaliseround(sd, &vsd, fpscr, 0, "fsqrt");
403 static u32 vfp_compare(int sd, int signal_on_qnan, s32 m, u32 fpscr)
462 static u32 vfp_single_fcmp(int sd, int unused, s32 m, u32 fpscr)
464 return vfp_compare(sd, 0, m, fpscr);
467 static u32 vfp_single_fcmpe(int sd, int unused, s32 m, u32 fpscr)
469 return vfp_compare(sd, 1, m, fpscr);
472 static u32 vfp_single_fcmpz(int sd, int unused, s32 m, u32 fpscr)
474 return vfp_compare(sd, 0, 0, fpscr);
477 static u32 vfp_single_fcmpez(int sd, int unused, s32 m, u32 fpscr)
479 return vfp_compare(sd, 1, 0, fpscr);
482 static u32 vfp_single_fcvtd(int dd, int unused, s32 m, u32 fpscr)
518 return vfp_double_normaliseround(dd, &vdd, fpscr, exceptions, "fcvtd");
525 static u32 vfp_single_fuito(int sd, int unused, s32 m, u32 fpscr)
533 return vfp_single_normaliseround(sd, &vs, fpscr, 0, "fuito");
536 static u32 vfp_single_fsito(int sd, int unused, s32 m, u32 fpscr)
544 return vfp_single_normaliseround(sd, &vs, fpscr, 0, "fsito");
547 static u32 vfp_single_ftoui(int sd, int unused, s32 m, u32 fpscr)
551 int rmode = fpscr & FPSCR_RMODE_MASK;
622 static u32 vfp_single_ftouiz(int sd, int unused, s32 m, u32 fpscr)
627 static u32 vfp_single_ftosi(int sd, int unused, s32 m, u32 fpscr)
631 int rmode = fpscr & FPSCR_RMODE_MASK;
701 static u32 vfp_single_ftosiz(int sd, int unused, s32 m, u32 fpscr)
730 struct vfp_single *vsm, u32 fpscr)
764 return vfp_propagate_nan(vsd, vsn, vsm, fpscr);
772 struct vfp_single *vsm, u32 fpscr)
799 return vfp_single_fadd_nonnumber(vsd, vsn, vsm, fpscr);
823 vsd->sign = (fpscr & FPSCR_RMODE_MASK) ==
835 vfp_single_multiply(struct vfp_single *vsd, struct vfp_single *vsn, struct vfp_single *vsm, u32 fpscr)
859 return vfp_propagate_nan(vsd, vsn, vsm, fpscr);
895 vfp_single_multiply_accumulate(int sd, int sn, s32 m, u32 fpscr, u32 negate, char *func)
911 exceptions = vfp_single_multiply(&vsp, &vsn, &vsm, fpscr);
923 exceptions |= vfp_single_add(&vsd, &vsn, &vsp, fpscr);
925 return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, func);
935 static u32 vfp_single_fmac(int sd, int sn, s32 m, u32 fpscr)
937 return vfp_single_multiply_accumulate(sd, sn, m, fpscr, 0, "fmac");
943 static u32 vfp_single_fnmac(int sd, int sn, s32 m, u32 fpscr)
945 return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_MULTIPLY, "fnmac");
951 static u32 vfp_single_fmsc(int sd, int sn, s32 m, u32 fpscr)
953 return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_SUBTRACT, "fmsc");
959 static u32 vfp_single_fnmsc(int sd, int sn, s32 m, u32 fpscr)
961 return vfp_single_multiply_accumulate(sd, sn, m, fpscr, NEG_SUBTRACT | NEG_MULTIPLY, "fnmsc");
967 static u32 vfp_single_fmul(int sd, int sn, s32 m, u32 fpscr)
983 exceptions = vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
984 return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fmul");
990 static u32 vfp_single_fnmul(int sd, int sn, s32 m, u32 fpscr)
1006 exceptions = vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
1008 return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fnmul");
1014 static u32 vfp_single_fadd(int sd, int sn, s32 m, u32 fpscr)
1033 exceptions = vfp_single_add(&vsd, &vsn, &vsm, fpscr);
1035 return vfp_single_normaliseround(sd, &vsd, fpscr, exceptions, "fadd");
1041 static u32 vfp_single_fsub(int sd, int sn, s32 m, u32 fpscr)
1046 return vfp_single_fadd(sd, sn, vfp_single_packed_negate(m), fpscr);
1052 static u32 vfp_single_fdiv(int sd, int sn, s32 m, u32 fpscr)
1128 return vfp_single_normaliseround(sd, &vsd, fpscr, 0, "fdiv");
1131 exceptions = vfp_propagate_nan(&vsd, &vsn, &vsm, fpscr);
1137 exceptions = vfp_propagate_nan(&vsd, &vsm, &vsn, fpscr);
1172 u32 vfp_single_cpdo(u32 inst, u32 fpscr)
1182 vecstride = 1 + ((fpscr & FPSCR_STRIDE_MASK) == FPSCR_STRIDE_MASK);
1204 veclen = fpscr & FPSCR_LENGTH_MASK;
1227 except = fop->fn(dest, sn, m, fpscr);