Lines Matching refs:GMP_LIMB_BITS

181 #ifndef GMP_LIMB_BITS
182 #define GMP_LIMB_BITS (8 * SIZEOF_MP_LIMB_T)
2253 (cout) = __w >> (GMP_LIMB_BITS-1); \
2274 && GMP_LIMB_BITS == 32 && ! defined (NO_ASM) && ! WANT_ASSERT
2463 /* A GMP_LIMB_BITS bit approximation to 1/big_base, represented as a
2576 On most CPUs this is just an arithmetic right shift by GMP_LIMB_BITS-1,
2588 ? (mp_limb_signed_t) (n) >> (GMP_LIMB_BITS - 1) \
2645 limb not larger than (2**(2*GMP_LIMB_BITS))/D - (2**GMP_LIMB_BITS).
2655 _q += (nh); /* Compensate, di is 2**GMP_LIMB_BITS too small */ \
2702 _n2 = ((nh) << (GMP_LIMB_BITS - (lgup))) + ((nl) >> 1 >> (l - 1)); \
2703 _n10 = (nl) << (GMP_LIMB_BITS - (lgup)); \
3068 && HAVE_HOST_CPU_FAMILY_powerpc && GMP_LIMB_BITS == 32
3089 && GMP_LIMB_BITS == 32
3097 && defined (__amd64__) && GMP_LIMB_BITS == 64
3105 && ! defined (NO_ASM) && defined (__ia64) && GMP_LIMB_BITS == 64
3114 && HAVE_HOST_CPU_FAMILY_m68k && GMP_LIMB_BITS == 32
3127 #if GMP_LIMB_BITS == 8
3131 #if GMP_LIMB_BITS == 16
3137 #if GMP_LIMB_BITS == 32
3147 #if GMP_LIMB_BITS == 64
3182 && GMP_LIMB_BITS == 32 && HAVE_LIMB_BIG_ENDIAN \
3207 && GMP_LIMB_BITS == 32 && HAVE_LIMB_BIG_ENDIAN \
3268 #if defined __GNUC__ && defined __sparc_v9__ && GMP_LIMB_BITS == 64
3293 && ! defined (NO_ASM) && defined (__ia64) && GMP_LIMB_BITS == 64
3307 (N is GMP_LIMB_BITS, [] denotes truncation.) */
3309 #if ! defined (popc_limb) && GMP_LIMB_BITS == 8
3320 #if ! defined (popc_limb) && GMP_LIMB_BITS == 16
3332 #if ! defined (popc_limb) && GMP_LIMB_BITS == 32
3345 #if ! defined (popc_limb) && GMP_LIMB_BITS == 64
3382 #define W_TYPE_SIZE GMP_LIMB_BITS
3744 xh, xl and r include nail bits. Must have 0 < count < GMP_LIMB_BITS.
3752 ((xl) >> (GMP_LIMB_BITS - (count))))
3847 #define mpn_dc_set_str_powtab_alloc(n) ((n) + GMP_LIMB_BITS)
3848 #define mpn_dc_set_str_itch(n) ((n) + GMP_LIMB_BITS)
3849 #define mpn_dc_get_str_powtab_alloc(n) ((n) + 2 * GMP_LIMB_BITS)
3850 #define mpn_dc_get_str_itch(n) ((n) + GMP_LIMB_BITS)
3881 GMP_LIMB_BITS then the +2 is unnecessary. This happens always for