/netbsd-current/external/lgpl3/gmp/dist/mpn/generic/ |
H A D | gcd_1.c | 44 mp_limb_t ulimb; local 52 ulimb = up[0]; 60 /* Must get common zeros before the mod reduction. If ulimb==0 then 62 if (ulimb != 0) 64 count_trailing_zeros (u_low_zero_bits, ulimb); 68 ulimb = MPN_MOD_OR_MODEXACT_1_ODD (up, size, vlimb); 69 if (ulimb == 0) 72 count_trailing_zeros (c, ulimb); 73 ulimb >>= c; 78 count_trailing_zeros (u_low_zero_bits, ulimb); [all...] |
/netbsd-current/external/lgpl3/gmp/dist/mpf/ |
H A D | cmp_si.c | 40 mp_limb_t ulimb; variable 83 ulimb = up[--usize]; 87 if ((ulimb >> GMP_NAIL_BITS) != 0) 89 ulimb = (ulimb << GMP_NUMB_BITS); 90 if (usize != 0) ulimb |= up[--usize]; 95 if (ulimb != abs_vval) 96 return (ulimb < abs_vval) ? -usign : usign;
|
H A D | cmp_ui.c | 39 mp_limb_t ulimb; variable 65 ulimb = up[--usize]; 69 if ((ulimb >> GMP_NAIL_BITS) != 0) 71 ulimb = (ulimb << GMP_NUMB_BITS); 72 if (usize != 0) ulimb |= up[--usize]; 77 if (ulimb != vval) 78 return (ulimb < vval) ? -1 : 1;
|
H A D | ui_sub.c | 60 mp_limb_t ulimb; 90 ulimb = u; 100 vp = &ulimb; 118 if (ediff == 0 && ulimb == vp[vsize - 1]) 141 up = &ulimb; 205 ASSERT (usize == 1 && vsize >= 1 && ulimb == *up); /* usize is 1>ediff, vsize >= 1 */ 213 if (ulimb > vp[vsize]) 215 tp[vsize] = ulimb - vp[vsize] - 1; 223 tp[vsize] = vp[vsize] - ulimb; 232 negate = ulimb < v [all...] |
/netbsd-current/external/lgpl3/gmp/dist/mpz/ |
H A D | hamdist.c | 69 mp_limb_t ulimb, vlimb; variable 89 ulimb = *up++; 92 if (ulimb != 0) 98 ulimb = vlimb; 104 /* twos complement first non-zero limbs (ulimb is non-zero, but vlimb 106 ulimb = -ulimb; 108 popc_limb (count, (ulimb ^ vlimb) & GMP_NUMB_MASK); 136 complement ulimb. Note -v^(~0^u) == (v-1)^u. */
|
/netbsd-current/external/lgpl3/gmp/dist/mpn/alpha/ev6/nails/ |
H A D | addmul_3.asm | 46 define(`ulimb',`r25') 89 ldq ulimb, 0(up) 91 mulq v0, ulimb, m0a C U1 92 umulh v0, ulimb, m0b C U1 93 mulq v1, ulimb, m1a C U1 94 umulh v1, ulimb, m1b C U1 96 mulq v2, ulimb, m2a C U1 97 umulh v2, ulimb, m2b C U1 102 ldq ulimb, 0(up) C L0 109 mulq v0, ulimb, m0 [all...] |
H A D | addmul_4.asm | 47 define(`ulimb',`r25') 103 ldq ulimb, 0(up) 105 mulq v0, ulimb, m0a C U1 106 umulh v0, ulimb, m0b C U1 107 mulq v1, ulimb, m1a C U1 108 umulh v1, ulimb, m1b C U1 110 mulq v2, ulimb, m2a C U1 111 umulh v2, ulimb, m2b C U1 112 mulq v3, ulimb, m3a C U1 113 umulh v3, ulimb, m3 [all...] |
H A D | addmul_2.asm | 47 define(`ulimb',`r25') 83 ldq ulimb, 0(up) 85 mulq v0, ulimb, m0a C U1 86 umulh v0, ulimb, m0b C U1 87 mulq v1, ulimb, m1a C U1 88 umulh v1, ulimb, m1b C U1 96 ldq ulimb, 0(up) C L1 101 mulq v0, ulimb, m0a C U1 105 umulh v0, ulimb, m0b C U1 111 mulq v1, ulimb, m1 [all...] |
/netbsd-current/external/lgpl3/gmp/dist/tests/mpf/ |
H A D | t-muldiv.c | 40 unsigned long ulimb, vlimb; local 83 ulimb = urandom (); 84 mpf_set_ui (u, ulimb); 134 mpf_ui_div (x, ulimb, v);
|
/netbsd-current/external/lgpl3/gmp/dist/mpn/ia64/ |
H A D | dive_1.asm | 167 .Ln1: xmpy.l f12 = f10, f7 C q = ulimb * inverse 172 xmpy.l f12 = f11, f7 C q = ulimb * inverse 182 xmpy.l f12 = f11, f7 C q = ulimb * inverse 205 xmpy.l f10 = f11, f7 C si = ulimb * inverse 221 xmpy.l f10 = f11, f7 C si = ulimb * inverse 229 xmpy.l f10 = f11, f7 C si = ulimb * inverse
|
H A D | hamdist.asm | 61 {.mmi; ld8 r10 = [up], 8 C load first ulimb M01
|
/netbsd-current/external/lgpl3/gmp/dist/mpn/alpha/ev6/ |
H A D | aorslsh1_n.asm | 113 ADDSUB u0, sl, ps C ulimb + (vlimb << 1)
|
/netbsd-current/external/lgpl3/gmp/dist/mpn/alpha/ |
H A D | aorslsh2_n.asm | 111 ADDSUB u0, sl, ps C ulimb + (vlimb << 1)
|
H A D | aorslsh1_n.asm | 105 ADDSUB u0, sl, ps C ulimb + (vlimb << 1)
|