Merge branch 'vendor/MDOCML'
[dragonfly.git] / contrib / mpfr / src / mpfr-longlong.h
1 /* longlong.h -- definitions for mixed size 32/64 bit arithmetic.
2
3 Copyright 1991, 1992, 1993, 1994, 1996, 1997, 1999, 2000, 2001, 2002, 2003,
4 2004, 2005, 2007, 2008, 2009, 2010, 2011, 2012, 2013 Free Software Foundation, Inc.
5
6 This file is free software; you can redistribute it and/or modify it under the
7 terms of the GNU Lesser General Public License as published by the Free
8 Software Foundation; either version 3 of the License, or (at your option) any
9 later version.
10
11 This file is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A
13 PARTICULAR PURPOSE.  See the GNU Lesser General Public License for more
14 details.
15
16 You should have received a copy of the GNU Lesser General Public License
17 along with this file.  If not, see http://www.gnu.org/licenses/.  */
18
19 /* You have to define the following before including this file:
20
21    UWtype -- An unsigned type, default type for operations (typically a "word")
22    UHWtype -- An unsigned type, at least half the size of UWtype.
23    UDWtype -- An unsigned type, at least twice as large a UWtype
24    W_TYPE_SIZE -- size in bits of UWtype
25
26    SItype, USItype -- Signed and unsigned 32 bit types.
27    DItype, UDItype -- Signed and unsigned 64 bit types.
28
29    On a 32 bit machine UWtype should typically be USItype;
30    on a 64 bit machine, UWtype should typically be UDItype.
31
32    CAUTION!  Using this file outside of GMP is not safe.  You need to include
33    gmp.h and gmp-impl.h, or certain things might not work as expected.
34 */
35
36 #define __BITS4 (W_TYPE_SIZE / 4)
37 #define __ll_B ((UWtype) 1 << (W_TYPE_SIZE / 2))
38 #define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
39 #define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
40
41 /* This is used to make sure no undesirable sharing between different libraries
42    that use this file takes place.  */
43 #ifndef __MPN
44 #define __MPN(x) __##x
45 #endif
46
47 #ifndef _PROTO
48 #if (__STDC__-0) || defined (__cplusplus)
49 #define _PROTO(x) x
50 #else
51 #define _PROTO(x) ()
52 #endif
53 #endif
54
55 /* Define auxiliary asm macros.
56
57    1) umul_ppmm(high_prod, low_prod, multiplier, multiplicand) multiplies two
58    UWtype integers MULTIPLIER and MULTIPLICAND, and generates a two UWtype
59    word product in HIGH_PROD and LOW_PROD.
60
61    2) __umulsidi3(a,b) multiplies two UWtype integers A and B, and returns a
62    UDWtype product.  This is just a variant of umul_ppmm.
63
64    3) udiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
65    denominator) divides a UDWtype, composed by the UWtype integers
66    HIGH_NUMERATOR and LOW_NUMERATOR, by DENOMINATOR and places the quotient
67    in QUOTIENT and the remainder in REMAINDER.  HIGH_NUMERATOR must be less
68    than DENOMINATOR for correct operation.  If, in addition, the most
69    significant bit of DENOMINATOR must be 1, then the pre-processor symbol
70    UDIV_NEEDS_NORMALIZATION is defined to 1.
71
72    4) sdiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
73    denominator).  Like udiv_qrnnd but the numbers are signed.  The quotient
74    is rounded toward 0.
75
76    5) count_leading_zeros(count, x) counts the number of zero-bits from the
77    msb to the first non-zero bit in the UWtype X.  This is the number of
78    steps X needs to be shifted left to set the msb.  Undefined for X == 0,
79    unless the symbol COUNT_LEADING_ZEROS_0 is defined to some value.
80
81    6) count_trailing_zeros(count, x) like count_leading_zeros, but counts
82    from the least significant end.
83
84    7) add_ssaaaa(high_sum, low_sum, high_addend_1, low_addend_1,
85    high_addend_2, low_addend_2) adds two UWtype integers, composed by
86    HIGH_ADDEND_1 and LOW_ADDEND_1, and HIGH_ADDEND_2 and LOW_ADDEND_2
87    respectively.  The result is placed in HIGH_SUM and LOW_SUM.  Overflow
88    (i.e. carry out) is not stored anywhere, and is lost.
89
90    8) sub_ddmmss(high_difference, low_difference, high_minuend, low_minuend,
91    high_subtrahend, low_subtrahend) subtracts two two-word UWtype integers,
92    composed by HIGH_MINUEND_1 and LOW_MINUEND_1, and HIGH_SUBTRAHEND_2 and
93    LOW_SUBTRAHEND_2 respectively.  The result is placed in HIGH_DIFFERENCE
94    and LOW_DIFFERENCE.  Overflow (i.e. carry out) is not stored anywhere,
95    and is lost.
96
97    If any of these macros are left undefined for a particular CPU,
98    C macros are used.
99
100
101    Notes:
102
103    For add_ssaaaa the two high and two low addends can both commute, but
104    unfortunately gcc only supports one "%" commutative in each asm block.
105    This has always been so but is only documented in recent versions
106    (eg. pre-release 3.3).  Having two or more "%"s can cause an internal
107    compiler error in certain rare circumstances.
108
109    Apparently it was only the last "%" that was ever actually respected, so
110    the code has been updated to leave just that.  Clearly there's a free
111    choice whether high or low should get it, if there's a reason to favour
112    one over the other.  Also obviously when the constraints on the two
113    operands are identical there's no benefit to the reloader in any "%" at
114    all.
115
116    */
117
118 /* The CPUs come in alphabetical order below.
119
120    Please add support for more CPUs here, or improve the current support
121    for the CPUs below!  */
122
123
124 /* count_leading_zeros_gcc_clz is count_leading_zeros implemented with gcc
125    3.4 __builtin_clzl or __builtin_clzll, according to our limb size.
126    Similarly count_trailing_zeros_gcc_ctz using __builtin_ctzl or
127    __builtin_ctzll.
128
129    These builtins are only used when we check what code comes out, on some
130    chips they're merely libgcc calls, where we will instead want an inline
131    in that case (either asm or generic C).
132
133    These builtins are better than an asm block of the same insn, since an
134    asm block doesn't give gcc any information about scheduling or resource
135    usage.  We keep an asm block for use on prior versions of gcc though.
136
137    For reference, __builtin_ffs existed in gcc prior to __builtin_clz, but
138    it's not used (for count_leading_zeros) because it generally gives extra
139    code to ensure the result is 0 when the input is 0, which we don't need
140    or want.  */
141
142 #ifdef _LONG_LONG_LIMB
143 #define count_leading_zeros_gcc_clz(count,x)    \
144   do {                                          \
145     ASSERT ((x) != 0);                          \
146     (count) = __builtin_clzll (x);              \
147   } while (0)
148 #else
149 #define count_leading_zeros_gcc_clz(count,x)    \
150   do {                                          \
151     ASSERT ((x) != 0);                          \
152     (count) = __builtin_clzl (x);               \
153   } while (0)
154 #endif
155
156 #ifdef _LONG_LONG_LIMB
157 #define count_trailing_zeros_gcc_ctz(count,x)   \
158   do {                                          \
159     ASSERT ((x) != 0);                          \
160     (count) = __builtin_ctzll (x);              \
161   } while (0)
162 #else
163 #define count_trailing_zeros_gcc_ctz(count,x)   \
164   do {                                          \
165     ASSERT ((x) != 0);                          \
166     (count) = __builtin_ctzl (x);               \
167   } while (0)
168 #endif
169
170 /* Note: the following FIXME comes from GMP, thus it does make sense to try
171    to resolve it in MPFR. */
172 /* FIXME: The macros using external routines like __MPN(count_leading_zeros)
173    don't need to be under !NO_ASM */
174 #if ! defined (NO_ASM)
175
176 #if defined (__alpha) && W_TYPE_SIZE == 64
177 /* Most alpha-based machines, except Cray systems. */
178 #if defined (__GNUC__)
179 #if __GMP_GNUC_PREREQ (3,3)
180 #define umul_ppmm(ph, pl, m0, m1) \
181   do {                                                                  \
182     UDItype __m0 = (m0), __m1 = (m1);                                   \
183     (ph) = __builtin_alpha_umulh (__m0, __m1);                          \
184     (pl) = __m0 * __m1;                                                 \
185   } while (0)
186 #else
187 #define umul_ppmm(ph, pl, m0, m1) \
188   do {                                                                  \
189     UDItype __m0 = (m0), __m1 = (m1);                                   \
190     __asm__ ("umulh %r1,%2,%0"                                          \
191              : "=r" (ph)                                                \
192              : "%rJ" (m0), "rI" (m1));                                  \
193     (pl) = __m0 * __m1;                                                 \
194   } while (0)
195 #endif
196 #define UMUL_TIME 18
197 #else /* ! __GNUC__ */
198 #include <machine/builtins.h>
199 #define umul_ppmm(ph, pl, m0, m1) \
200   do {                                                                  \
201     UDItype __m0 = (m0), __m1 = (m1);                                   \
202     (ph) = __UMULH (m0, m1);                                            \
203     (pl) = __m0 * __m1;                                                 \
204   } while (0)
205 #endif
206 #ifndef LONGLONG_STANDALONE
207 #define udiv_qrnnd(q, r, n1, n0, d) \
208   do { UWtype __di;                                                     \
209     __di = __MPN(invert_limb) (d);                                      \
210     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);                          \
211   } while (0)
212 #define UDIV_PREINV_ALWAYS  1
213 #define UDIV_NEEDS_NORMALIZATION 1
214 #define UDIV_TIME 220
215 #endif /* LONGLONG_STANDALONE */
216
217 /* clz_tab is required in all configurations, since mpn/alpha/cntlz.asm
218    always goes into libgmp.so, even when not actually used.  */
219 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
220
221 #if defined (__GNUC__) && HAVE_HOST_CPU_alpha_CIX
222 #define count_leading_zeros(COUNT,X) \
223   __asm__("ctlz %1,%0" : "=r"(COUNT) : "r"(X))
224 #define count_trailing_zeros(COUNT,X) \
225   __asm__("cttz %1,%0" : "=r"(COUNT) : "r"(X))
226 #endif /* clz/ctz using cix */
227
228 #if ! defined (count_leading_zeros)                             \
229   && defined (__GNUC__) && ! defined (LONGLONG_STANDALONE)
230 /* ALPHA_CMPBGE_0 gives "cmpbge $31,src,dst", ie. test src bytes == 0.
231    "$31" is written explicitly in the asm, since an "r" constraint won't
232    select reg 31.  There seems no need to worry about "r31" syntax for cray,
233    since gcc itself (pre-release 3.4) emits just $31 in various places.  */
234 #define ALPHA_CMPBGE_0(dst, src)                                        \
235   do { asm ("cmpbge $31, %1, %0" : "=r" (dst) : "r" (src)); } while (0)
236 /* Zero bytes are turned into bits with cmpbge, a __clz_tab lookup counts
237    them, locating the highest non-zero byte.  A second __clz_tab lookup
238    counts the leading zero bits in that byte, giving the result.  */
239 #define count_leading_zeros(count, x)                                   \
240   do {                                                                  \
241     UWtype  __clz__b, __clz__c, __clz__x = (x);                         \
242     ALPHA_CMPBGE_0 (__clz__b,  __clz__x);           /* zero bytes */    \
243     __clz__b = __clz_tab [(__clz__b >> 1) ^ 0x7F];  /* 8 to 1 byte */   \
244     __clz__b = __clz__b * 8 - 7;                    /* 57 to 1 shift */ \
245     __clz__x >>= __clz__b;                                              \
246     __clz__c = __clz_tab [__clz__x];                /* 8 to 1 bit */    \
247     __clz__b = 65 - __clz__b;                                           \
248     (count) = __clz__b - __clz__c;                                      \
249   } while (0)
250 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
251 #endif /* clz using cmpbge */
252
253 #if ! defined (count_leading_zeros) && ! defined (LONGLONG_STANDALONE)
254 #if HAVE_ATTRIBUTE_CONST
255 long __MPN(count_leading_zeros) _PROTO ((UDItype)) __attribute__ ((const));
256 #else
257 long __MPN(count_leading_zeros) _PROTO ((UDItype));
258 #endif
259 #define count_leading_zeros(count, x) \
260   ((count) = __MPN(count_leading_zeros) (x))
261 #endif /* clz using mpn */
262 #endif /* __alpha */
263
264 #if defined (_CRAY) && W_TYPE_SIZE == 64
265 #include <intrinsics.h>
266 #define UDIV_PREINV_ALWAYS  1
267 #define UDIV_NEEDS_NORMALIZATION 1
268 #define UDIV_TIME 220
269 long __MPN(count_leading_zeros) _PROTO ((UDItype));
270 #define count_leading_zeros(count, x) \
271   ((count) = _leadz ((UWtype) (x)))
272 #if defined (_CRAYIEEE)         /* I.e., Cray T90/ieee, T3D, and T3E */
273 #define umul_ppmm(ph, pl, m0, m1) \
274   do {                                                                  \
275     UDItype __m0 = (m0), __m1 = (m1);                                   \
276     (ph) = _int_mult_upper (m0, m1);                                    \
277     (pl) = __m0 * __m1;                                                 \
278   } while (0)
279 #ifndef LONGLONG_STANDALONE
280 #define udiv_qrnnd(q, r, n1, n0, d) \
281   do { UWtype __di;                                                     \
282     __di = __MPN(invert_limb) (d);                                      \
283     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);                          \
284   } while (0)
285 #endif /* LONGLONG_STANDALONE */
286 #endif /* _CRAYIEEE */
287 #endif /* _CRAY */
288
289 #if defined (__ia64) && W_TYPE_SIZE == 64
290 /* This form encourages gcc (pre-release 3.4 at least) to emit predicated
291    "sub r=r,r" and "sub r=r,r,1", giving a 2 cycle latency.  The generic
292    code using "al<bl" arithmetically comes out making an actual 0 or 1 in a
293    register, which takes an extra cycle.  */
294 #define sub_ddmmss(sh, sl, ah, al, bh, bl)      \
295   do {                                          \
296     UWtype __x;                                 \
297     __x = (al) - (bl);                          \
298     if ((al) < (bl))                            \
299       (sh) = (ah) - (bh) - 1;                   \
300     else                                        \
301       (sh) = (ah) - (bh);                       \
302     (sl) = __x;                                 \
303   } while (0)
304 #if defined (__GNUC__) && ! defined (__INTEL_COMPILER)
305 /* Do both product parts in assembly, since that gives better code with
306    all gcc versions.  Some callers will just use the upper part, and in
307    that situation we waste an instruction, but not any cycles.  */
308 #define umul_ppmm(ph, pl, m0, m1) \
309     __asm__ ("xma.hu %0 = %2, %3, f0\n\txma.l %1 = %2, %3, f0"          \
310              : "=&f" (ph), "=f" (pl)                                    \
311              : "f" (m0), "f" (m1))
312 #define UMUL_TIME 14
313 #define count_leading_zeros(count, x) \
314   do {                                                                  \
315     UWtype _x = (x), _y, _a, _c;                                        \
316     __asm__ ("mux1 %0 = %1, @rev" : "=r" (_y) : "r" (_x));              \
317     __asm__ ("czx1.l %0 = %1" : "=r" (_a) : "r" (-_y | _y));            \
318     _c = (_a - 1) << 3;                                                 \
319     _x >>= _c;                                                          \
320     if (_x >= 1 << 4)                                                   \
321       _x >>= 4, _c += 4;                                                \
322     if (_x >= 1 << 2)                                                   \
323       _x >>= 2, _c += 2;                                                \
324     _c += _x >> 1;                                                      \
325     (count) =  W_TYPE_SIZE - 1 - _c;                                    \
326   } while (0)
327 /* similar to what gcc does for __builtin_ffs, but 0 based rather than 1
328    based, and we don't need a special case for x==0 here */
329 #define count_trailing_zeros(count, x)                                  \
330   do {                                                                  \
331     UWtype __ctz_x = (x);                                               \
332     __asm__ ("popcnt %0 = %1"                                           \
333              : "=r" (count)                                             \
334              : "r" ((__ctz_x-1) & ~__ctz_x));                           \
335   } while (0)
336 #endif
337 #if defined (__INTEL_COMPILER)
338 #include <ia64intrin.h>
339 #define umul_ppmm(ph, pl, m0, m1)                                       \
340   do {                                                                  \
341     UWtype _m0 = (m0), _m1 = (m1);                                      \
342     ph = _m64_xmahu (_m0, _m1, 0);                                      \
343     pl = _m0 * _m1;                                                     \
344   } while (0)
345 #endif
346 #ifndef LONGLONG_STANDALONE
347 #define udiv_qrnnd(q, r, n1, n0, d) \
348   do { UWtype __di;                                                     \
349     __di = __MPN(invert_limb) (d);                                      \
350     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);                          \
351   } while (0)
352 #define UDIV_PREINV_ALWAYS  1
353 #define UDIV_NEEDS_NORMALIZATION 1
354 #endif
355 #define UDIV_TIME 220
356 #endif
357
358
359 #if defined (__GNUC__)
360
361 /* We sometimes need to clobber "cc" with gcc2, but that would not be
362    understood by gcc1.  Use cpp to avoid major code duplication.  */
363 #if __GNUC__ < 2
364 #define __CLOBBER_CC
365 #define __AND_CLOBBER_CC
366 #else /* __GNUC__ >= 2 */
367 #define __CLOBBER_CC : "cc"
368 #define __AND_CLOBBER_CC , "cc"
369 #endif /* __GNUC__ < 2 */
370
371 #if (defined (__a29k__) || defined (_AM29K)) && W_TYPE_SIZE == 32
372 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
373   __asm__ ("add %1,%4,%5\n\taddc %0,%2,%3"                              \
374            : "=r" (sh), "=&r" (sl)                                      \
375            : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl))
376 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
377   __asm__ ("sub %1,%4,%5\n\tsubc %0,%2,%3"                              \
378            : "=r" (sh), "=&r" (sl)                                      \
379            : "r" (ah), "rI" (bh), "r" (al), "rI" (bl))
380 #define umul_ppmm(xh, xl, m0, m1) \
381   do {                                                                  \
382     USItype __m0 = (m0), __m1 = (m1);                                   \
383     __asm__ ("multiplu %0,%1,%2"                                        \
384              : "=r" (xl)                                                \
385              : "r" (__m0), "r" (__m1));                                 \
386     __asm__ ("multmu %0,%1,%2"                                          \
387              : "=r" (xh)                                                \
388              : "r" (__m0), "r" (__m1));                                 \
389   } while (0)
390 #define udiv_qrnnd(q, r, n1, n0, d) \
391   __asm__ ("dividu %0,%3,%4"                                            \
392            : "=r" (q), "=q" (r)                                         \
393            : "1" (n1), "r" (n0), "r" (d))
394 #define count_leading_zeros(count, x) \
395     __asm__ ("clz %0,%1"                                                \
396              : "=r" (count)                                             \
397              : "r" (x))
398 #define COUNT_LEADING_ZEROS_0 32
399 #endif /* __a29k__ */
400
401 #if defined (__arc__)
402 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
403   __asm__ ("add.f\t%1, %4, %5\n\tadc\t%0, %2, %3"                       \
404            : "=r" (sh),                                                 \
405              "=&r" (sl)                                                 \
406            : "r"  ((USItype) (ah)),                                     \
407              "rIJ" ((USItype) (bh)),                                    \
408              "%r" ((USItype) (al)),                                     \
409              "rIJ" ((USItype) (bl)))
410 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
411   __asm__ ("sub.f\t%1, %4, %5\n\tsbc\t%0, %2, %3"                       \
412            : "=r" (sh),                                                 \
413              "=&r" (sl)                                                 \
414            : "r" ((USItype) (ah)),                                      \
415              "rIJ" ((USItype) (bh)),                                    \
416              "r" ((USItype) (al)),                                      \
417              "rIJ" ((USItype) (bl)))
418 #endif
419
420 #if defined (__arm__) && W_TYPE_SIZE == 32
421 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
422   __asm__ ("adds\t%1, %4, %5\n\tadc\t%0, %2, %3"                        \
423            : "=r" (sh), "=&r" (sl)                                      \
424            : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC)
425 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
426   do {                                                                  \
427     if (__builtin_constant_p (al))                                      \
428       {                                                                 \
429         if (__builtin_constant_p (ah))                                  \
430           __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2"                \
431                    : "=r" (sh), "=&r" (sl)                              \
432                    : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
433         else                                                            \
434           __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3"                \
435                    : "=r" (sh), "=&r" (sl)                              \
436                    : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
437       }                                                                 \
438     else if (__builtin_constant_p (ah))                                 \
439       {                                                                 \
440         if (__builtin_constant_p (bl))                                  \
441           __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2"                \
442                    : "=r" (sh), "=&r" (sl)                              \
443                    : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
444         else                                                            \
445           __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2"                \
446                    : "=r" (sh), "=&r" (sl)                              \
447                    : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
448       }                                                                 \
449     else if (__builtin_constant_p (bl))                                 \
450       {                                                                 \
451         if (__builtin_constant_p (bh))                                  \
452           __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"                \
453                    : "=r" (sh), "=&r" (sl)                              \
454                    : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
455         else                                                            \
456           __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2"                \
457                    : "=r" (sh), "=&r" (sl)                              \
458                    : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
459       }                                                                 \
460     else /* only bh might be a constant */                              \
461       __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"                    \
462                : "=r" (sh), "=&r" (sl)                                  \
463                : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC);\
464     } while (0)
465 #if 1 || defined (__arm_m__)    /* `M' series has widening multiply support */
466 #define umul_ppmm(xh, xl, a, b) \
467   __asm__ ("umull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
468 #define UMUL_TIME 5
469 #define smul_ppmm(xh, xl, a, b) \
470   __asm__ ("smull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
471 #ifndef LONGLONG_STANDALONE
472 #define udiv_qrnnd(q, r, n1, n0, d) \
473   do { UWtype __di;                                                     \
474     __di = __MPN(invert_limb) (d);                                      \
475     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);                          \
476   } while (0)
477 #define UDIV_PREINV_ALWAYS  1
478 #define UDIV_NEEDS_NORMALIZATION 1
479 #define UDIV_TIME 70
480 #endif /* LONGLONG_STANDALONE */
481 #else
482 #define umul_ppmm(xh, xl, a, b) \
483   __asm__ ("%@ Inlined umul_ppmm\n"                                     \
484 "       mov     %|r0, %2, lsr #16\n"                                    \
485 "       mov     %|r2, %3, lsr #16\n"                                    \
486 "       bic     %|r1, %2, %|r0, lsl #16\n"                              \
487 "       bic     %|r2, %3, %|r2, lsl #16\n"                              \
488 "       mul     %1, %|r1, %|r2\n"                                       \
489 "       mul     %|r2, %|r0, %|r2\n"                                     \
490 "       mul     %|r1, %0, %|r1\n"                                       \
491 "       mul     %0, %|r0, %0\n"                                         \
492 "       adds    %|r1, %|r2, %|r1\n"                                     \
493 "       addcs   %0, %0, #65536\n"                                       \
494 "       adds    %1, %1, %|r1, lsl #16\n"                                \
495 "       adc     %0, %0, %|r1, lsr #16"                                  \
496            : "=&r" (xh), "=r" (xl)                                      \
497            : "r" (a), "r" (b)                                           \
498            : "r0", "r1", "r2")
499 #define UMUL_TIME 20
500 #ifndef LONGLONG_STANDALONE
501 #define udiv_qrnnd(q, r, n1, n0, d) \
502   do { UWtype __r;                                                      \
503     (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d));                    \
504     (r) = __r;                                                          \
505   } while (0)
506 extern UWtype __MPN(udiv_qrnnd) _PROTO ((UWtype *, UWtype, UWtype, UWtype));
507 #define UDIV_TIME 200
508 #endif /* LONGLONG_STANDALONE */
509 #endif
510 #endif /* __arm__ */
511
512 #if defined (__clipper__) && W_TYPE_SIZE == 32
513 #define umul_ppmm(w1, w0, u, v) \
514   ({union {UDItype __ll;                                                \
515            struct {USItype __l, __h;} __i;                              \
516           } __x;                                                        \
517   __asm__ ("mulwux %2,%0"                                               \
518            : "=r" (__x.__ll)                                            \
519            : "%0" ((USItype)(u)), "r" ((USItype)(v)));                  \
520   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
521 #define smul_ppmm(w1, w0, u, v) \
522   ({union {DItype __ll;                                                 \
523            struct {SItype __l, __h;} __i;                               \
524           } __x;                                                        \
525   __asm__ ("mulwx %2,%0"                                                \
526            : "=r" (__x.__ll)                                            \
527            : "%0" ((SItype)(u)), "r" ((SItype)(v)));                    \
528   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
529 #define __umulsidi3(u, v) \
530   ({UDItype __w;                                                        \
531     __asm__ ("mulwux %2,%0"                                             \
532              : "=r" (__w) : "%0" ((USItype)(u)), "r" ((USItype)(v)));   \
533     __w; })
534 #endif /* __clipper__ */
535
536 /* Fujitsu vector computers.  */
537 #if defined (__uxp__) && W_TYPE_SIZE == 32
538 #define umul_ppmm(ph, pl, u, v) \
539   do {                                                                  \
540     union {UDItype __ll;                                                \
541            struct {USItype __h, __l;} __i;                              \
542           } __x;                                                        \
543     __asm__ ("mult.lu %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v));\
544     (ph) = __x.__i.__h;                                                 \
545     (pl) = __x.__i.__l;                                                 \
546   } while (0)
547 #define smul_ppmm(ph, pl, u, v) \
548   do {                                                                  \
549     union {UDItype __ll;                                                \
550            struct {USItype __h, __l;} __i;                              \
551           } __x;                                                        \
552     __asm__ ("mult.l %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v)); \
553     (ph) = __x.__i.__h;                                                 \
554     (pl) = __x.__i.__l;                                                 \
555   } while (0)
556 #endif
557
558 #if defined (__gmicro__) && W_TYPE_SIZE == 32
559 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
560   __asm__ ("add.w %5,%1\n\taddx %3,%0"                                  \
561            : "=g" (sh), "=&g" (sl)                                      \
562            : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),                 \
563              "%1" ((USItype)(al)), "g" ((USItype)(bl)))
564 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
565   __asm__ ("sub.w %5,%1\n\tsubx %3,%0"                                  \
566            : "=g" (sh), "=&g" (sl)                                      \
567            : "0" ((USItype)(ah)), "g" ((USItype)(bh)),                  \
568              "1" ((USItype)(al)), "g" ((USItype)(bl)))
569 #define umul_ppmm(ph, pl, m0, m1) \
570   __asm__ ("mulx %3,%0,%1"                                              \
571            : "=g" (ph), "=r" (pl)                                       \
572            : "%0" ((USItype)(m0)), "g" ((USItype)(m1)))
573 #define udiv_qrnnd(q, r, nh, nl, d) \
574   __asm__ ("divx %4,%0,%1"                                              \
575            : "=g" (q), "=r" (r)                                         \
576            : "1" ((USItype)(nh)), "0" ((USItype)(nl)), "g" ((USItype)(d)))
577 #define count_leading_zeros(count, x) \
578   __asm__ ("bsch/1 %1,%0"                                               \
579            : "=g" (count) : "g" ((USItype)(x)), "0" ((USItype)0))
580 #endif
581
582 #if defined (__hppa) && W_TYPE_SIZE == 32
583 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
584   __asm__ ("add%I5 %5,%r4,%1\n\taddc %r2,%r3,%0"                        \
585            : "=r" (sh), "=&r" (sl)                                      \
586            : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
587 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
588   __asm__ ("sub%I4 %4,%r5,%1\n\tsubb %r2,%r3,%0"                        \
589            : "=r" (sh), "=&r" (sl)                                      \
590            : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
591 #if defined (_PA_RISC1_1)
592 #define umul_ppmm(wh, wl, u, v) \
593   do {                                                                  \
594     union {UDItype __ll;                                                \
595            struct {USItype __h, __l;} __i;                              \
596           } __x;                                                        \
597     __asm__ ("xmpyu %1,%2,%0" : "=*f" (__x.__ll) : "*f" (u), "*f" (v)); \
598     (wh) = __x.__i.__h;                                                 \
599     (wl) = __x.__i.__l;                                                 \
600   } while (0)
601 #define UMUL_TIME 8
602 #define UDIV_TIME 60
603 #else
604 #define UMUL_TIME 40
605 #define UDIV_TIME 80
606 #endif
607 #define count_leading_zeros(count, x) \
608   do {                                                                  \
609     USItype __tmp;                                                      \
610     __asm__ (                                                           \
611        "ldi             1,%0\n"                                         \
612 "       extru,=         %1,15,16,%%r0   ; Bits 31..16 zero?\n"          \
613 "       extru,tr        %1,15,16,%1     ; No.  Shift down, skip add.\n" \
614 "       ldo             16(%0),%0       ; Yes.  Perform add.\n"         \
615 "       extru,=         %1,23,8,%%r0    ; Bits 15..8 zero?\n"           \
616 "       extru,tr        %1,23,8,%1      ; No.  Shift down, skip add.\n" \
617 "       ldo             8(%0),%0        ; Yes.  Perform add.\n"         \
618 "       extru,=         %1,27,4,%%r0    ; Bits 7..4 zero?\n"            \
619 "       extru,tr        %1,27,4,%1      ; No.  Shift down, skip add.\n" \
620 "       ldo             4(%0),%0        ; Yes.  Perform add.\n"         \
621 "       extru,=         %1,29,2,%%r0    ; Bits 3..2 zero?\n"            \
622 "       extru,tr        %1,29,2,%1      ; No.  Shift down, skip add.\n" \
623 "       ldo             2(%0),%0        ; Yes.  Perform add.\n"         \
624 "       extru           %1,30,1,%1      ; Extract bit 1.\n"             \
625 "       sub             %0,%1,%0        ; Subtract it.\n"               \
626         : "=r" (count), "=r" (__tmp) : "1" (x));                        \
627   } while (0)
628 #endif /* hppa */
629
630 /* These macros are for ABI=2.0w.  In ABI=2.0n they can't be used, since GCC
631    (3.2) puts longlong into two adjacent 32-bit registers.  Presumably this
632    is just a case of no direct support for 2.0n but treating it like 1.0. */
633 #if defined (__hppa) && W_TYPE_SIZE == 64 && ! defined (_LONG_LONG_LIMB)
634 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
635   __asm__ ("add%I5 %5,%r4,%1\n\tadd,dc %r2,%r3,%0"                      \
636            : "=r" (sh), "=&r" (sl)                                      \
637            : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
638 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
639   __asm__ ("sub%I4 %4,%r5,%1\n\tsub,db %r2,%r3,%0"                      \
640            : "=r" (sh), "=&r" (sl)                                      \
641            : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
642 #endif /* hppa */
643
644 #if (defined (__i370__) || defined (__s390__) || defined (__mvs__)) && W_TYPE_SIZE == 32
645 #define smul_ppmm(xh, xl, m0, m1) \
646   do {                                                                  \
647     union {DItype __ll;                                                 \
648            struct {USItype __h, __l;} __i;                              \
649           } __x;                                                        \
650     __asm__ ("lr %N0,%1\n\tmr %0,%2"                                    \
651              : "=&r" (__x.__ll)                                         \
652              : "r" (m0), "r" (m1));                                     \
653     (xh) = __x.__i.__h; (xl) = __x.__i.__l;                             \
654   } while (0)
655 #define sdiv_qrnnd(q, r, n1, n0, d) \
656   do {                                                                  \
657     union {DItype __ll;                                                 \
658            struct {USItype __h, __l;} __i;                              \
659           } __x;                                                        \
660     __x.__i.__h = n1; __x.__i.__l = n0;                                 \
661     __asm__ ("dr %0,%2"                                                 \
662              : "=r" (__x.__ll)                                          \
663              : "0" (__x.__ll), "r" (d));                                \
664     (q) = __x.__i.__l; (r) = __x.__i.__h;                               \
665   } while (0)
666 #endif
667
668 #if (defined (__i386__) || defined (__i486__)) && W_TYPE_SIZE == 32
669 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
670   __asm__ ("addl %5,%k1\n\tadcl %3,%k0"                                 \
671            : "=r" (sh), "=&r" (sl)                                      \
672            : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),                 \
673              "%1" ((USItype)(al)), "g" ((USItype)(bl)))
674 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
675   __asm__ ("subl %5,%k1\n\tsbbl %3,%k0"                                 \
676            : "=r" (sh), "=&r" (sl)                                      \
677            : "0" ((USItype)(ah)), "g" ((USItype)(bh)),                  \
678              "1" ((USItype)(al)), "g" ((USItype)(bl)))
679 #define umul_ppmm(w1, w0, u, v) \
680   __asm__ ("mull %3"                                                    \
681            : "=a" (w0), "=d" (w1)                                       \
682            : "%0" ((USItype)(u)), "rm" ((USItype)(v)))
683 #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
684   __asm__ ("divl %4"                 /* stringification in K&R C */     \
685            : "=a" (q), "=d" (r)                                         \
686            : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "rm" ((USItype)(dx)))
687
688 #if HAVE_HOST_CPU_i586 || HAVE_HOST_CPU_pentium || HAVE_HOST_CPU_pentiummmx
689 /* Pentium bsrl takes between 10 and 72 cycles depending where the most
690    significant 1 bit is, hence the use of the following alternatives.  bsfl
691    is slow too, between 18 and 42 depending where the least significant 1
692    bit is, so let the generic count_trailing_zeros below make use of the
693    count_leading_zeros here too.  */
694
695 #if HAVE_HOST_CPU_pentiummmx && ! defined (LONGLONG_STANDALONE)
696 /* The following should be a fixed 14 or 15 cycles, but possibly plus an L1
697    cache miss reading from __clz_tab.  For P55 it's favoured over the float
698    below so as to avoid mixing MMX and x87, since the penalty for switching
699    between the two is about 100 cycles.
700
701    The asm block sets __shift to -3 if the high 24 bits are clear, -2 for
702    16, -1 for 8, or 0 otherwise.  This could be written equivalently as
703    follows, but as of gcc 2.95.2 it results in conditional jumps.
704
705        __shift = -(__n < 0x1000000);
706        __shift -= (__n < 0x10000);
707        __shift -= (__n < 0x100);
708
709    The middle two sbbl and cmpl's pair, and with luck something gcc
710    generates might pair with the first cmpl and the last sbbl.  The "32+1"
711    constant could be folded into __clz_tab[], but it doesn't seem worth
712    making a different table just for that.  */
713
714 #define count_leading_zeros(c,n)                                        \
715   do {                                                                  \
716     USItype  __n = (n);                                                 \
717     USItype  __shift;                                                   \
718     __asm__ ("cmpl  $0x1000000, %1\n"                                   \
719              "sbbl  %0, %0\n"                                           \
720              "cmpl  $0x10000, %1\n"                                     \
721              "sbbl  $0, %0\n"                                           \
722              "cmpl  $0x100, %1\n"                                       \
723              "sbbl  $0, %0\n"                                           \
724              : "=&r" (__shift) : "r"  (__n));                           \
725     __shift = __shift*8 + 24 + 1;                                       \
726     (c) = 32 + 1 - __shift - __clz_tab[__n >> __shift];                 \
727   } while (0)
728 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
729 #define COUNT_LEADING_ZEROS_0   31   /* n==0 indistinguishable from n==1 */
730
731 #else /* ! pentiummmx || LONGLONG_STANDALONE */
732 /* The following should be a fixed 14 cycles or so.  Some scheduling
733    opportunities should be available between the float load/store too.  This
734    sort of code is used in gcc 3 for __builtin_ffs (with "n&-n") and is
735    apparently suggested by the Intel optimizing manual (don't know exactly
736    where).  gcc 2.95 or up will be best for this, so the "double" is
737    correctly aligned on the stack.  */
738 #define count_leading_zeros(c,n)                                        \
739   do {                                                                  \
740     union {                                                             \
741       double    d;                                                      \
742       unsigned  a[2];                                                   \
743     } __u;                                                              \
744     ASSERT ((n) != 0);                                                  \
745     __u.d = (UWtype) (n);                                               \
746     (c) = 0x3FF + 31 - (__u.a[1] >> 20);                                \
747   } while (0)
748 #define COUNT_LEADING_ZEROS_0   (0x3FF + 31)
749 #endif /* pentiummx */
750
751 #else /* ! pentium */
752
753 #if __GMP_GNUC_PREREQ (3,4)  /* using bsrl */
754 #define count_leading_zeros(count,x)  count_leading_zeros_gcc_clz(count,x)
755 #endif /* gcc clz */
756
757 /* On P6, gcc prior to 3.0 generates a partial register stall for
758    __cbtmp^31, due to using "xorb $31" instead of "xorl $31", the former
759    being 1 code byte smaller.  "31-__cbtmp" is a workaround, probably at the
760    cost of one extra instruction.  Do this for "i386" too, since that means
761    generic x86.  */
762 #if ! defined (count_leading_zeros) && __GNUC__ < 3                     \
763   && (HAVE_HOST_CPU_i386                                                \
764       || HAVE_HOST_CPU_i686                                             \
765       || HAVE_HOST_CPU_pentiumpro                                       \
766       || HAVE_HOST_CPU_pentium2                                         \
767       || HAVE_HOST_CPU_pentium3)
768 #define count_leading_zeros(count, x)                                   \
769   do {                                                                  \
770     USItype __cbtmp;                                                    \
771     ASSERT ((x) != 0);                                                  \
772     __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x)));      \
773     (count) = 31 - __cbtmp;                                             \
774   } while (0)
775 #endif /* gcc<3 asm bsrl */
776
777 #ifndef count_leading_zeros
778 #define count_leading_zeros(count, x)                                   \
779   do {                                                                  \
780     USItype __cbtmp;                                                    \
781     ASSERT ((x) != 0);                                                  \
782     __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x)));      \
783     (count) = __cbtmp ^ 31;                                             \
784   } while (0)
785 #endif /* asm bsrl */
786
787 #if __GMP_GNUC_PREREQ (3,4)  /* using bsfl */
788 #define count_trailing_zeros(count,x)  count_trailing_zeros_gcc_ctz(count,x)
789 #endif /* gcc ctz */
790
791 #ifndef count_trailing_zeros
792 #define count_trailing_zeros(count, x)                                  \
793   do {                                                                  \
794     ASSERT ((x) != 0);                                                  \
795     __asm__ ("bsfl %1,%k0" : "=r" (count) : "rm" ((USItype)(x)));       \
796   } while (0)
797 #endif /* asm bsfl */
798
799 #endif /* ! pentium */
800
801 #ifndef UMUL_TIME
802 #define UMUL_TIME 10
803 #endif
804 #ifndef UDIV_TIME
805 #define UDIV_TIME 40
806 #endif
807 #endif /* 80x86 */
808
809 #if defined (__amd64__) && W_TYPE_SIZE == 64
810 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
811   __asm__ ("addq %5,%q1\n\tadcq %3,%q0"                                 \
812            : "=r" (sh), "=&r" (sl)                                      \
813            : "0"  ((UDItype)(ah)), "rme" ((UDItype)(bh)),               \
814              "%1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
815 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
816   __asm__ ("subq %5,%q1\n\tsbbq %3,%q0"                                 \
817            : "=r" (sh), "=&r" (sl)                                      \
818            : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)),                \
819              "1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
820 #define umul_ppmm(w1, w0, u, v) \
821   __asm__ ("mulq %3"                                                    \
822            : "=a" (w0), "=d" (w1)                                       \
823            : "%0" ((UDItype)(u)), "rm" ((UDItype)(v)))
824 #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
825   __asm__ ("divq %4"                 /* stringification in K&R C */     \
826            : "=a" (q), "=d" (r)                                         \
827            : "0" ((UDItype)(n0)), "1" ((UDItype)(n1)), "rm" ((UDItype)(dx)))
828 /* bsrq destination must be a 64-bit register, hence UDItype for __cbtmp. */
829 #define count_leading_zeros(count, x)                                   \
830   do {                                                                  \
831     UDItype __cbtmp;                                                    \
832     ASSERT ((x) != 0);                                                  \
833     __asm__ ("bsrq %1,%0" : "=r" (__cbtmp) : "rm" ((UDItype)(x)));      \
834     (count) = __cbtmp ^ 63;                                             \
835   } while (0)
836 /* bsfq destination must be a 64-bit register, "%q0" forces this in case
837    count is only an int. */
838 #define count_trailing_zeros(count, x)                                  \
839   do {                                                                  \
840     ASSERT ((x) != 0);                                                  \
841     __asm__ ("bsfq %1,%q0" : "=r" (count) : "rm" ((UDItype)(x)));       \
842   } while (0)
843 #endif /* x86_64 */
844
845 #if defined (__i860__) && W_TYPE_SIZE == 32
846 #define rshift_rhlc(r,h,l,c) \
847   __asm__ ("shr %3,r0,r0\;shrd %1,%2,%0"                                \
848            "=r" (r) : "r" (h), "r" (l), "rn" (c))
849 #endif /* i860 */
850
851 #if defined (__i960__) && W_TYPE_SIZE == 32
852 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
853   __asm__ ("cmpo 1,0\;addc %5,%4,%1\;addc %3,%2,%0"                     \
854            : "=r" (sh), "=&r" (sl)                                      \
855            : "dI" (ah), "dI" (bh), "%dI" (al), "dI" (bl))
856 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
857   __asm__ ("cmpo 0,0\;subc %5,%4,%1\;subc %3,%2,%0"                     \
858            : "=r" (sh), "=&r" (sl)                                      \
859            : "dI" (ah), "dI" (bh), "dI" (al), "dI" (bl))
860 #define umul_ppmm(w1, w0, u, v) \
861   ({union {UDItype __ll;                                                \
862            struct {USItype __l, __h;} __i;                              \
863           } __x;                                                        \
864   __asm__ ("emul %2,%1,%0"                                              \
865            : "=d" (__x.__ll) : "%dI" (u), "dI" (v));                    \
866   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
867 #define __umulsidi3(u, v) \
868   ({UDItype __w;                                                        \
869     __asm__ ("emul %2,%1,%0" : "=d" (__w) : "%dI" (u), "dI" (v));       \
870     __w; })
871 #define udiv_qrnnd(q, r, nh, nl, d) \
872   do {                                                                  \
873     union {UDItype __ll;                                                \
874            struct {USItype __l, __h;} __i;                              \
875           } __nn;                                                       \
876     __nn.__i.__h = (nh); __nn.__i.__l = (nl);                           \
877     __asm__ ("ediv %d,%n,%0"                                            \
878            : "=d" (__rq.__ll) : "dI" (__nn.__ll), "dI" (d));            \
879     (r) = __rq.__i.__l; (q) = __rq.__i.__h;                             \
880   } while (0)
881 #define count_leading_zeros(count, x) \
882   do {                                                                  \
883     USItype __cbtmp;                                                    \
884     __asm__ ("scanbit %1,%0" : "=r" (__cbtmp) : "r" (x));               \
885     (count) = __cbtmp ^ 31;                                             \
886   } while (0)
887 #define COUNT_LEADING_ZEROS_0 (-32) /* sic */
888 #if defined (__i960mx)          /* what is the proper symbol to test??? */
889 #define rshift_rhlc(r,h,l,c) \
890   do {                                                                  \
891     union {UDItype __ll;                                                \
892            struct {USItype __l, __h;} __i;                              \
893           } __nn;                                                       \
894     __nn.__i.__h = (h); __nn.__i.__l = (l);                             \
895     __asm__ ("shre %2,%1,%0" : "=d" (r) : "dI" (__nn.__ll), "dI" (c));  \
896   }
897 #endif /* i960mx */
898 #endif /* i960 */
899
900 #if (defined (__mc68000__) || defined (__mc68020__) || defined(mc68020) \
901      || defined (__m68k__) || defined (__mc5200__) || defined (__mc5206e__) \
902      || defined (__mc5307__)) && W_TYPE_SIZE == 32
903 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
904   __asm__ ("add%.l %5,%1\n\taddx%.l %3,%0"                              \
905            : "=d" (sh), "=&d" (sl)                                      \
906            : "0"  ((USItype)(ah)), "d" ((USItype)(bh)),                 \
907              "%1" ((USItype)(al)), "g" ((USItype)(bl)))
908 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
909   __asm__ ("sub%.l %5,%1\n\tsubx%.l %3,%0"                              \
910            : "=d" (sh), "=&d" (sl)                                      \
911            : "0" ((USItype)(ah)), "d" ((USItype)(bh)),                  \
912              "1" ((USItype)(al)), "g" ((USItype)(bl)))
913 /* The '020, '030, '040 and CPU32 have 32x32->64 and 64/32->32q-32r.  */
914 #if defined (__mc68020__) || defined(mc68020) \
915      || defined (__mc68030__) || defined (mc68030) \
916      || defined (__mc68040__) || defined (mc68040) \
917      || defined (__mcpu32__) || defined (mcpu32) \
918      || defined (__NeXT__)
919 #define umul_ppmm(w1, w0, u, v) \
920   __asm__ ("mulu%.l %3,%1:%0"                                           \
921            : "=d" (w0), "=d" (w1)                                       \
922            : "%0" ((USItype)(u)), "dmi" ((USItype)(v)))
923 #define UMUL_TIME 45
924 #define udiv_qrnnd(q, r, n1, n0, d) \
925   __asm__ ("divu%.l %4,%1:%0"                                           \
926            : "=d" (q), "=d" (r)                                         \
927            : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
928 #define UDIV_TIME 90
929 #define sdiv_qrnnd(q, r, n1, n0, d) \
930   __asm__ ("divs%.l %4,%1:%0"                                           \
931            : "=d" (q), "=d" (r)                                         \
932            : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
933 #else /* for other 68k family members use 16x16->32 multiplication */
934 #define umul_ppmm(xh, xl, a, b) \
935   do { USItype __umul_tmp1, __umul_tmp2;                                \
936         __asm__ ("| Inlined umul_ppmm\n"                                \
937 "       move%.l %5,%3\n"                                                \
938 "       move%.l %2,%0\n"                                                \
939 "       move%.w %3,%1\n"                                                \
940 "       swap    %3\n"                                                   \
941 "       swap    %0\n"                                                   \
942 "       mulu%.w %2,%1\n"                                                \
943 "       mulu%.w %3,%0\n"                                                \
944 "       mulu%.w %2,%3\n"                                                \
945 "       swap    %2\n"                                                   \
946 "       mulu%.w %5,%2\n"                                                \
947 "       add%.l  %3,%2\n"                                                \
948 "       jcc     1f\n"                                                   \
949 "       add%.l  %#0x10000,%0\n"                                         \
950 "1:     move%.l %2,%3\n"                                                \
951 "       clr%.w  %2\n"                                                   \
952 "       swap    %2\n"                                                   \
953 "       swap    %3\n"                                                   \
954 "       clr%.w  %3\n"                                                   \
955 "       add%.l  %3,%1\n"                                                \
956 "       addx%.l %2,%0\n"                                                \
957 "       | End inlined umul_ppmm"                                        \
958               : "=&d" (xh), "=&d" (xl),                                 \
959                 "=d" (__umul_tmp1), "=&d" (__umul_tmp2)                 \
960               : "%2" ((USItype)(a)), "d" ((USItype)(b)));               \
961   } while (0)
962 #define UMUL_TIME 100
963 #define UDIV_TIME 400
964 #endif /* not mc68020 */
965 /* The '020, '030, '040 and '060 have bitfield insns.
966    GCC 3.4 defines __mc68020__ when in CPU32 mode, check for __mcpu32__ to
967    exclude bfffo on that chip (bitfield insns not available).  */
968 #if (defined (__mc68020__) || defined (mc68020)    \
969      || defined (__mc68030__) || defined (mc68030) \
970      || defined (__mc68040__) || defined (mc68040) \
971      || defined (__mc68060__) || defined (mc68060) \
972      || defined (__NeXT__))                        \
973   && ! defined (__mcpu32__)
974 #define count_leading_zeros(count, x) \
975   __asm__ ("bfffo %1{%b2:%b2},%0"                                       \
976            : "=d" (count)                                               \
977            : "od" ((USItype) (x)), "n" (0))
978 #define COUNT_LEADING_ZEROS_0 32
979 #endif
980 #endif /* mc68000 */
981
982 #if defined (__m88000__) && W_TYPE_SIZE == 32
983 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
984   __asm__ ("addu.co %1,%r4,%r5\n\taddu.ci %0,%r2,%r3"                   \
985            : "=r" (sh), "=&r" (sl)                                      \
986            : "rJ" (ah), "rJ" (bh), "%rJ" (al), "rJ" (bl))
987 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
988   __asm__ ("subu.co %1,%r4,%r5\n\tsubu.ci %0,%r2,%r3"                   \
989            : "=r" (sh), "=&r" (sl)                                      \
990            : "rJ" (ah), "rJ" (bh), "rJ" (al), "rJ" (bl))
991 #define count_leading_zeros(count, x) \
992   do {                                                                  \
993     USItype __cbtmp;                                                    \
994     __asm__ ("ff1 %0,%1" : "=r" (__cbtmp) : "r" (x));                   \
995     (count) = __cbtmp ^ 31;                                             \
996   } while (0)
997 #define COUNT_LEADING_ZEROS_0 63 /* sic */
998 #if defined (__m88110__)
999 #define umul_ppmm(wh, wl, u, v) \
1000   do {                                                                  \
1001     union {UDItype __ll;                                                \
1002            struct {USItype __h, __l;} __i;                              \
1003           } __x;                                                        \
1004     __asm__ ("mulu.d %0,%1,%2" : "=r" (__x.__ll) : "r" (u), "r" (v));   \
1005     (wh) = __x.__i.__h;                                                 \
1006     (wl) = __x.__i.__l;                                                 \
1007   } while (0)
1008 #define udiv_qrnnd(q, r, n1, n0, d) \
1009   ({union {UDItype __ll;                                                \
1010            struct {USItype __h, __l;} __i;                              \
1011           } __x, __q;                                                   \
1012   __x.__i.__h = (n1); __x.__i.__l = (n0);                               \
1013   __asm__ ("divu.d %0,%1,%2"                                            \
1014            : "=r" (__q.__ll) : "r" (__x.__ll), "r" (d));                \
1015   (r) = (n0) - __q.__l * (d); (q) = __q.__l; })
1016 #define UMUL_TIME 5
1017 #define UDIV_TIME 25
1018 #else
1019 #define UMUL_TIME 17
1020 #define UDIV_TIME 150
1021 #endif /* __m88110__ */
1022 #endif /* __m88000__ */
1023
1024 #if defined (__mips) && W_TYPE_SIZE == 32
1025 #if __GMP_GNUC_PREREQ (4,4)
1026 #define umul_ppmm(w1, w0, u, v) \
1027   do {                                                                  \
1028     UDItype __ll = (UDItype)(u) * (v);                                  \
1029     w1 = __ll >> 32;                                                    \
1030     w0 = __ll;                                                          \
1031   } while (0)
1032 #endif
1033 #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7)
1034 #define umul_ppmm(w1, w0, u, v) \
1035   __asm__ ("multu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v))
1036 #endif
1037 #if !defined (umul_ppmm)
1038 #define umul_ppmm(w1, w0, u, v) \
1039   __asm__ ("multu %2,%3\n\tmflo %0\n\tmfhi %1"                          \
1040            : "=d" (w0), "=d" (w1) : "d" (u), "d" (v))
1041 #endif
1042 #define UMUL_TIME 10
1043 #define UDIV_TIME 100
1044 #endif /* __mips */
1045
1046 #if (defined (__mips) && __mips >= 3) && W_TYPE_SIZE == 64
1047 #if __GMP_GNUC_PREREQ (4,4)
1048 #define umul_ppmm(w1, w0, u, v) \
1049   do {                                                                  \
1050     typedef unsigned int __ll_UTItype __attribute__((mode(TI)));        \
1051     __ll_UTItype __ll = (__ll_UTItype)(u) * (v);                        \
1052     w1 = __ll >> 64;                                                    \
1053     w0 = __ll;                                                          \
1054   } while (0)
1055 #endif
1056 #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7)
1057 #define umul_ppmm(w1, w0, u, v) \
1058   __asm__ ("dmultu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v))
1059 #endif
1060 #if !defined (umul_ppmm)
1061 #define umul_ppmm(w1, w0, u, v) \
1062   __asm__ ("dmultu %2,%3\n\tmflo %0\n\tmfhi %1"                         \
1063            : "=d" (w0), "=d" (w1) : "d" (u), "d" (v))
1064 #endif
1065 #define UMUL_TIME 20
1066 #define UDIV_TIME 140
1067 #endif /* __mips */
1068
1069 #if defined (__mmix__) && W_TYPE_SIZE == 64
1070 #define umul_ppmm(w1, w0, u, v) \
1071   __asm__ ("MULU %0,%2,%3" : "=r" (w0), "=z" (w1) : "r" (u), "r" (v))
1072 #endif
1073
1074 #if defined (__ns32000__) && W_TYPE_SIZE == 32
1075 #define umul_ppmm(w1, w0, u, v) \
1076   ({union {UDItype __ll;                                                \
1077            struct {USItype __l, __h;} __i;                              \
1078           } __x;                                                        \
1079   __asm__ ("meid %2,%0"                                                 \
1080            : "=g" (__x.__ll)                                            \
1081            : "%0" ((USItype)(u)), "g" ((USItype)(v)));                  \
1082   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
1083 #define __umulsidi3(u, v) \
1084   ({UDItype __w;                                                        \
1085     __asm__ ("meid %2,%0"                                               \
1086              : "=g" (__w)                                               \
1087              : "%0" ((USItype)(u)), "g" ((USItype)(v)));                \
1088     __w; })
1089 #define udiv_qrnnd(q, r, n1, n0, d) \
1090   ({union {UDItype __ll;                                                \
1091            struct {USItype __l, __h;} __i;                              \
1092           } __x;                                                        \
1093   __x.__i.__h = (n1); __x.__i.__l = (n0);                               \
1094   __asm__ ("deid %2,%0"                                                 \
1095            : "=g" (__x.__ll)                                            \
1096            : "0" (__x.__ll), "g" ((USItype)(d)));                       \
1097   (r) = __x.__i.__l; (q) = __x.__i.__h; })
1098 #define count_trailing_zeros(count,x) \
1099   do {                                                                  \
1100     __asm__ ("ffsd      %2,%0"                                          \
1101              : "=r" (count)                                             \
1102              : "0" ((USItype) 0), "r" ((USItype) (x)));                 \
1103   } while (0)
1104 #endif /* __ns32000__ */
1105
1106 /* In the past we had a block of various #defines tested
1107        _ARCH_PPC    - AIX
1108        _ARCH_PWR    - AIX
1109        __powerpc__  - gcc
1110        __POWERPC__  - BEOS
1111        __ppc__      - Darwin
1112        PPC          - old gcc, GNU/Linux, SysV
1113    The plain PPC test was not good for vxWorks, since PPC is defined on all
1114    CPUs there (eg. m68k too), as a constant one is expected to compare
1115    CPU_FAMILY against.
1116
1117    At any rate, this was pretty unattractive and a bit fragile.  The use of
1118    HAVE_HOST_CPU_FAMILY is designed to cut through it all and be sure of
1119    getting the desired effect.
1120
1121    ENHANCE-ME: We should test _IBMR2 here when we add assembly support for
1122    the system vendor compilers.  (Is that vendor compilers with inline asm,
1123    or what?)  */
1124
1125 #if (HAVE_HOST_CPU_FAMILY_power || HAVE_HOST_CPU_FAMILY_powerpc)        \
1126   && W_TYPE_SIZE == 32
1127 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1128   do {                                                                  \
1129     if (__builtin_constant_p (bh) && (bh) == 0)                         \
1130       __asm__ ("{a%I4|add%I4c} %1,%3,%4\n\t{aze|addze} %0,%2"           \
1131              : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1132     else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0)         \
1133       __asm__ ("{a%I4|add%I4c} %1,%3,%4\n\t{ame|addme} %0,%2"           \
1134              : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1135     else                                                                \
1136       __asm__ ("{a%I5|add%I5c} %1,%4,%5\n\t{ae|adde} %0,%2,%3"          \
1137              : "=r" (sh), "=&r" (sl)                                    \
1138              : "r" (ah), "r" (bh), "%r" (al), "rI" (bl));               \
1139   } while (0)
1140 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1141   do {                                                                  \
1142     if (__builtin_constant_p (ah) && (ah) == 0)                         \
1143       __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{sfze|subfze} %0,%2"       \
1144                : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\
1145     else if (__builtin_constant_p (ah) && (ah) == ~(USItype) 0)         \
1146       __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{sfme|subfme} %0,%2"       \
1147                : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\
1148     else if (__builtin_constant_p (bh) && (bh) == 0)                    \
1149       __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{ame|addme} %0,%2"         \
1150                : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\
1151     else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0)         \
1152       __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{aze|addze} %0,%2"         \
1153                : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\
1154     else                                                                \
1155       __asm__ ("{sf%I4|subf%I4c} %1,%5,%4\n\t{sfe|subfe} %0,%3,%2"      \
1156                : "=r" (sh), "=&r" (sl)                                  \
1157                : "r" (ah), "r" (bh), "rI" (al), "r" (bl));              \
1158   } while (0)
1159 #define count_leading_zeros(count, x) \
1160   __asm__ ("{cntlz|cntlzw} %0,%1" : "=r" (count) : "r" (x))
1161 #define COUNT_LEADING_ZEROS_0 32
1162 #if HAVE_HOST_CPU_FAMILY_powerpc
1163 #if __GMP_GNUC_PREREQ (4,4)
1164 #define umul_ppmm(w1, w0, u, v) \
1165   do {                                                                  \
1166     UDItype __ll = (UDItype)(u) * (v);                                  \
1167     w1 = __ll >> 32;                                                    \
1168     w0 = __ll;                                                          \
1169   } while (0)
1170 #endif
1171 #if !defined (umul_ppmm)
1172 #define umul_ppmm(ph, pl, m0, m1) \
1173   do {                                                                  \
1174     USItype __m0 = (m0), __m1 = (m1);                                   \
1175     __asm__ ("mulhwu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));      \
1176     (pl) = __m0 * __m1;                                                 \
1177   } while (0)
1178 #endif
1179 #define UMUL_TIME 15
1180 #define smul_ppmm(ph, pl, m0, m1) \
1181   do {                                                                  \
1182     SItype __m0 = (m0), __m1 = (m1);                                    \
1183     __asm__ ("mulhw %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));       \
1184     (pl) = __m0 * __m1;                                                 \
1185   } while (0)
1186 #define SMUL_TIME 14
1187 #define UDIV_TIME 120
1188 #else
1189 #define UMUL_TIME 8
1190 #define smul_ppmm(xh, xl, m0, m1) \
1191   __asm__ ("mul %0,%2,%3" : "=r" (xh), "=q" (xl) : "r" (m0), "r" (m1))
1192 #define SMUL_TIME 4
1193 #define sdiv_qrnnd(q, r, nh, nl, d) \
1194   __asm__ ("div %0,%2,%4" : "=r" (q), "=q" (r) : "r" (nh), "1" (nl), "r" (d))
1195 #define UDIV_TIME 100
1196 #endif
1197 #endif /* 32-bit POWER architecture variants.  */
1198
1199 /* We should test _IBMR2 here when we add assembly support for the system
1200    vendor compilers.  */
1201 #if HAVE_HOST_CPU_FAMILY_powerpc && W_TYPE_SIZE == 64
1202 #if !defined (_LONG_LONG_LIMB)
1203 /* _LONG_LONG_LIMB is ABI=mode32 where adde operates on 32-bit values.  So
1204    use adde etc only when not _LONG_LONG_LIMB.  */
1205 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1206   do {                                                                  \
1207     if (__builtin_constant_p (bh) && (bh) == 0)                         \
1208       __asm__ ("{a%I4|add%I4c} %1,%3,%4\n\t{aze|addze} %0,%2"           \
1209              : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1210     else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)         \
1211       __asm__ ("{a%I4|add%I4c} %1,%3,%4\n\t{ame|addme} %0,%2"           \
1212              : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1213     else                                                                \
1214       __asm__ ("{a%I5|add%I5c} %1,%4,%5\n\t{ae|adde} %0,%2,%3"          \
1215              : "=r" (sh), "=&r" (sl)                                    \
1216              : "r" (ah), "r" (bh), "%r" (al), "rI" (bl));               \
1217   } while (0)
1218 /* We use "*rI" for the constant operand here, since with just "I", gcc barfs.
1219    This might seem strange, but gcc folds away the dead code late.  */
1220 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1221   do {                                                                        \
1222     if (__builtin_constant_p (bl) && bl > -0x8000 && bl <= 0x8000) {          \
1223         if (__builtin_constant_p (ah) && (ah) == 0)                           \
1224           __asm__ ("{ai|addic} %1,%3,%4\n\t{sfze|subfze} %0,%2"               \
1225                    : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "*rI" (-bl)); \
1226         else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0)           \
1227           __asm__ ("{ai|addic} %1,%3,%4\n\t{sfme|subfme} %0,%2"               \
1228                    : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "*rI" (-bl)); \
1229         else if (__builtin_constant_p (bh) && (bh) == 0)                      \
1230           __asm__ ("{ai|addic} %1,%3,%4\n\t{ame|addme} %0,%2"                 \
1231                    : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "*rI" (-bl)); \
1232         else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)           \
1233           __asm__ ("{ai|addic} %1,%3,%4\n\t{aze|addze} %0,%2"                 \
1234                    : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "*rI" (-bl)); \
1235         else                                                                  \
1236           __asm__ ("{ai|addic} %1,%4,%5\n\t{sfe|subfe} %0,%3,%2"              \
1237                    : "=r" (sh), "=&r" (sl)                                    \
1238                    : "r" (ah), "r" (bh), "rI" (al), "*rI" (-bl));             \
1239       } else {                                                                \
1240         if (__builtin_constant_p (ah) && (ah) == 0)                           \
1241           __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{sfze|subfze} %0,%2"         \
1242                    : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));  \
1243         else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0)           \
1244           __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{sfme|subfme} %0,%2"         \
1245                    : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));  \
1246         else if (__builtin_constant_p (bh) && (bh) == 0)                      \
1247           __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{ame|addme} %0,%2"           \
1248                    : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));  \
1249         else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)           \
1250           __asm__ ("{sf%I3|subf%I3c} %1,%4,%3\n\t{aze|addze} %0,%2"           \
1251                    : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));  \
1252         else                                                                  \
1253           __asm__ ("{sf%I4|subf%I4c} %1,%5,%4\n\t{sfe|subfe} %0,%3,%2"        \
1254                    : "=r" (sh), "=&r" (sl)                                    \
1255                    : "r" (ah), "r" (bh), "rI" (al), "r" (bl));                \
1256       }                                                                       \
1257   } while (0)
1258 #endif /* ! _LONG_LONG_LIMB */
1259 #define count_leading_zeros(count, x) \
1260   __asm__ ("cntlzd %0,%1" : "=r" (count) : "r" (x))
1261 #define COUNT_LEADING_ZEROS_0 64
1262 #if __GMP_GNUC_PREREQ (4,4)
1263 #define umul_ppmm(w1, w0, u, v) \
1264   do {                                                                  \
1265     typedef unsigned int __ll_UTItype __attribute__((mode(TI)));        \
1266     __ll_UTItype __ll = (__ll_UTItype)(u) * (v);                        \
1267     w1 = __ll >> 64;                                                    \
1268     w0 = __ll;                                                          \
1269   } while (0)
1270 #endif
1271 #if !defined (umul_ppmm)
1272 #define umul_ppmm(ph, pl, m0, m1) \
1273   do {                                                                  \
1274     UDItype __m0 = (m0), __m1 = (m1);                                   \
1275     __asm__ ("mulhdu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));      \
1276     (pl) = __m0 * __m1;                                                 \
1277   } while (0)
1278 #endif
1279 #define UMUL_TIME 15
1280 #define smul_ppmm(ph, pl, m0, m1) \
1281   do {                                                                  \
1282     DItype __m0 = (m0), __m1 = (m1);                                    \
1283     __asm__ ("mulhd %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));       \
1284     (pl) = __m0 * __m1;                                                 \
1285   } while (0)
1286 #define SMUL_TIME 14  /* ??? */
1287 #define UDIV_TIME 120 /* ??? */
1288 #endif /* 64-bit PowerPC.  */
1289
1290 #if defined (__pyr__) && W_TYPE_SIZE == 32
1291 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1292   __asm__ ("addw %5,%1\n\taddwc %3,%0"                                  \
1293            : "=r" (sh), "=&r" (sl)                                      \
1294            : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),                 \
1295              "%1" ((USItype)(al)), "g" ((USItype)(bl)))
1296 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1297   __asm__ ("subw %5,%1\n\tsubwb %3,%0"                                  \
1298            : "=r" (sh), "=&r" (sl)                                      \
1299            : "0" ((USItype)(ah)), "g" ((USItype)(bh)),                  \
1300              "1" ((USItype)(al)), "g" ((USItype)(bl)))
1301 /* This insn works on Pyramids with AP, XP, or MI CPUs, but not with SP.  */
1302 #define umul_ppmm(w1, w0, u, v) \
1303   ({union {UDItype __ll;                                                \
1304            struct {USItype __h, __l;} __i;                              \
1305           } __x;                                                        \
1306   __asm__ ("movw %1,%R0\n\tuemul %2,%0"                                 \
1307            : "=&r" (__x.__ll)                                           \
1308            : "g" ((USItype) (u)), "g" ((USItype)(v)));                  \
1309   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
1310 #endif /* __pyr__ */
1311
1312 #if defined (__ibm032__) /* RT/ROMP */  && W_TYPE_SIZE == 32
1313 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1314   __asm__ ("a %1,%5\n\tae %0,%3"                                        \
1315            : "=r" (sh), "=&r" (sl)                                      \
1316            : "0"  ((USItype)(ah)), "r" ((USItype)(bh)),                 \
1317              "%1" ((USItype)(al)), "r" ((USItype)(bl)))
1318 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1319   __asm__ ("s %1,%5\n\tse %0,%3"                                        \
1320            : "=r" (sh), "=&r" (sl)                                      \
1321            : "0" ((USItype)(ah)), "r" ((USItype)(bh)),                  \
1322              "1" ((USItype)(al)), "r" ((USItype)(bl)))
1323 #define smul_ppmm(ph, pl, m0, m1) \
1324   __asm__ (                                                             \
1325        "s       r2,r2\n"                                                \
1326 "       mts r10,%2\n"                                                   \
1327 "       m       r2,%3\n"                                                \
1328 "       m       r2,%3\n"                                                \
1329 "       m       r2,%3\n"                                                \
1330 "       m       r2,%3\n"                                                \
1331 "       m       r2,%3\n"                                                \
1332 "       m       r2,%3\n"                                                \
1333 "       m       r2,%3\n"                                                \
1334 "       m       r2,%3\n"                                                \
1335 "       m       r2,%3\n"                                                \
1336 "       m       r2,%3\n"                                                \
1337 "       m       r2,%3\n"                                                \
1338 "       m       r2,%3\n"                                                \
1339 "       m       r2,%3\n"                                                \
1340 "       m       r2,%3\n"                                                \
1341 "       m       r2,%3\n"                                                \
1342 "       m       r2,%3\n"                                                \
1343 "       cas     %0,r2,r0\n"                                             \
1344 "       mfs     r10,%1"                                                 \
1345            : "=r" (ph), "=r" (pl)                                       \
1346            : "%r" ((USItype)(m0)), "r" ((USItype)(m1))                  \
1347            : "r2")
1348 #define UMUL_TIME 20
1349 #define UDIV_TIME 200
1350 #define count_leading_zeros(count, x) \
1351   do {                                                                  \
1352     if ((x) >= 0x10000)                                                 \
1353       __asm__ ("clz     %0,%1"                                          \
1354                : "=r" (count) : "r" ((USItype)(x) >> 16));              \
1355     else                                                                \
1356       {                                                                 \
1357         __asm__ ("clz   %0,%1"                                          \
1358                  : "=r" (count) : "r" ((USItype)(x)));                  \
1359         (count) += 16;                                                  \
1360       }                                                                 \
1361   } while (0)
1362 #endif /* RT/ROMP */
1363
1364 #if defined (__sh2__) && W_TYPE_SIZE == 32
1365 #define umul_ppmm(w1, w0, u, v) \
1366   __asm__ ("dmulu.l %2,%3\n\tsts macl,%1\n\tsts mach,%0"                \
1367            : "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "macl", "mach")
1368 #define UMUL_TIME 5
1369 #endif
1370
1371 #if defined (__sparc__) && W_TYPE_SIZE == 32
1372 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1373   __asm__ ("addcc %r4,%5,%1\n\taddx %r2,%3,%0"                          \
1374            : "=r" (sh), "=&r" (sl)                                      \
1375            : "rJ" (ah), "rI" (bh),"%rJ" (al), "rI" (bl)                 \
1376            __CLOBBER_CC)
1377 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1378   __asm__ ("subcc %r4,%5,%1\n\tsubx %r2,%3,%0"                          \
1379            : "=r" (sh), "=&r" (sl)                                      \
1380            : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl) \
1381            __CLOBBER_CC)
1382 /* Note: the following FIXME comes from GMP, thus it does make sense to try
1383    to resolve it in MPFR. */
1384 /* FIXME: When gcc -mcpu=v9 is used on solaris, gcc/config/sol2-sld-64.h
1385    doesn't define anything to indicate that to us, it only sets __sparcv8. */
1386 #if defined (__sparc_v9__) || defined (__sparcv9)
1387 /* Perhaps we should use floating-point operations here?  */
1388 #if 0
1389 /* Triggers a bug making mpz/tests/t-gcd.c fail.
1390    Perhaps we simply need explicitly zero-extend the inputs?  */
1391 #define umul_ppmm(w1, w0, u, v) \
1392   __asm__ ("mulx %2,%3,%%g1; srl %%g1,0,%1; srlx %%g1,32,%0" :          \
1393            "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "g1")
1394 #else
1395 /* Use v8 umul until above bug is fixed.  */
1396 #define umul_ppmm(w1, w0, u, v) \
1397   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1398 #endif
1399 /* Use a plain v8 divide for v9.  */
1400 #define udiv_qrnnd(q, r, n1, n0, d) \
1401   do {                                                                  \
1402     USItype __q;                                                        \
1403     __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0"                     \
1404              : "=r" (__q) : "r" (n1), "r" (n0), "r" (d));               \
1405     (r) = (n0) - __q * (d);                                             \
1406     (q) = __q;                                                          \
1407   } while (0)
1408 #else
1409 #if defined (__sparc_v8__)   /* gcc normal */                           \
1410   || defined (__sparcv8)     /* gcc solaris */                          \
1411   || HAVE_HOST_CPU_supersparc
1412 /* Don't match immediate range because, 1) it is not often useful,
1413    2) the 'I' flag thinks of the range as a 13 bit signed interval,
1414    while we want to match a 13 bit interval, sign extended to 32 bits,
1415    but INTERPRETED AS UNSIGNED.  */
1416 #define umul_ppmm(w1, w0, u, v) \
1417   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1418 #define UMUL_TIME 5
1419
1420 #if HAVE_HOST_CPU_supersparc
1421 #define UDIV_TIME 60            /* SuperSPARC timing */
1422 #else
1423 /* Don't use this on SuperSPARC because its udiv only handles 53 bit
1424    dividends and will trap to the kernel for the rest. */
1425 #define udiv_qrnnd(q, r, n1, n0, d) \
1426   do {                                                                  \
1427     USItype __q;                                                        \
1428     __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0"                     \
1429              : "=r" (__q) : "r" (n1), "r" (n0), "r" (d));               \
1430     (r) = (n0) - __q * (d);                                             \
1431     (q) = __q;                                                          \
1432   } while (0)
1433 #define UDIV_TIME 25
1434 #endif /* HAVE_HOST_CPU_supersparc */
1435
1436 #else /* ! __sparc_v8__ */
1437 #if defined (__sparclite__)
1438 /* This has hardware multiply but not divide.  It also has two additional
1439    instructions scan (ffs from high bit) and divscc.  */
1440 #define umul_ppmm(w1, w0, u, v) \
1441   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1442 #define UMUL_TIME 5
1443 #define udiv_qrnnd(q, r, n1, n0, d) \
1444   __asm__ ("! Inlined udiv_qrnnd\n"                                     \
1445 "       wr      %%g0,%2,%%y     ! Not a delayed write for sparclite\n"  \
1446 "       tst     %%g0\n"                                                 \
1447 "       divscc  %3,%4,%%g1\n"                                           \
1448 "       divscc  %%g1,%4,%%g1\n"                                         \
1449 "       divscc  %%g1,%4,%%g1\n"                                         \
1450 "       divscc  %%g1,%4,%%g1\n"                                         \
1451 "       divscc  %%g1,%4,%%g1\n"                                         \
1452 "       divscc  %%g1,%4,%%g1\n"                                         \
1453 "       divscc  %%g1,%4,%%g1\n"                                         \
1454 "       divscc  %%g1,%4,%%g1\n"                                         \
1455 "       divscc  %%g1,%4,%%g1\n"                                         \
1456 "       divscc  %%g1,%4,%%g1\n"                                         \
1457 "       divscc  %%g1,%4,%%g1\n"                                         \
1458 "       divscc  %%g1,%4,%%g1\n"                                         \
1459 "       divscc  %%g1,%4,%%g1\n"                                         \
1460 "       divscc  %%g1,%4,%%g1\n"                                         \
1461 "       divscc  %%g1,%4,%%g1\n"                                         \
1462 "       divscc  %%g1,%4,%%g1\n"                                         \
1463 "       divscc  %%g1,%4,%%g1\n"                                         \
1464 "       divscc  %%g1,%4,%%g1\n"                                         \
1465 "       divscc  %%g1,%4,%%g1\n"                                         \
1466 "       divscc  %%g1,%4,%%g1\n"                                         \
1467 "       divscc  %%g1,%4,%%g1\n"                                         \
1468 "       divscc  %%g1,%4,%%g1\n"                                         \
1469 "       divscc  %%g1,%4,%%g1\n"                                         \
1470 "       divscc  %%g1,%4,%%g1\n"                                         \
1471 "       divscc  %%g1,%4,%%g1\n"                                         \
1472 "       divscc  %%g1,%4,%%g1\n"                                         \
1473 "       divscc  %%g1,%4,%%g1\n"                                         \
1474 "       divscc  %%g1,%4,%%g1\n"                                         \
1475 "       divscc  %%g1,%4,%%g1\n"                                         \
1476 "       divscc  %%g1,%4,%%g1\n"                                         \
1477 "       divscc  %%g1,%4,%%g1\n"                                         \
1478 "       divscc  %%g1,%4,%0\n"                                           \
1479 "       rd      %%y,%1\n"                                               \
1480 "       bl,a 1f\n"                                                      \
1481 "       add     %1,%4,%1\n"                                             \
1482 "1:     ! End of inline udiv_qrnnd"                                     \
1483            : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d)          \
1484            : "%g1" __AND_CLOBBER_CC)
1485 #define UDIV_TIME 37
1486 #define count_leading_zeros(count, x) \
1487   __asm__ ("scan %1,1,%0" : "=r" (count) : "r" (x))
1488 /* Early sparclites return 63 for an argument of 0, but they warn that future
1489    implementations might change this.  Therefore, leave COUNT_LEADING_ZEROS_0
1490    undefined.  */
1491 #endif /* __sparclite__ */
1492 #endif /* __sparc_v8__ */
1493 #endif /* __sparc_v9__ */
1494 /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd.  */
1495 #ifndef umul_ppmm
1496 #define umul_ppmm(w1, w0, u, v) \
1497   __asm__ ("! Inlined umul_ppmm\n"                                      \
1498 "       wr      %%g0,%2,%%y     ! SPARC has 0-3 delay insn after a wr\n" \
1499 "       sra     %3,31,%%g2      ! Don't move this insn\n"               \
1500 "       and     %2,%%g2,%%g2    ! Don't move this insn\n"               \
1501 "       andcc   %%g0,0,%%g1     ! Don't move this insn\n"               \
1502 "       mulscc  %%g1,%3,%%g1\n"                                         \
1503 "       mulscc  %%g1,%3,%%g1\n"                                         \
1504 "       mulscc  %%g1,%3,%%g1\n"                                         \
1505 "       mulscc  %%g1,%3,%%g1\n"                                         \
1506 "       mulscc  %%g1,%3,%%g1\n"                                         \
1507 "       mulscc  %%g1,%3,%%g1\n"                                         \
1508 "       mulscc  %%g1,%3,%%g1\n"                                         \
1509 "       mulscc  %%g1,%3,%%g1\n"                                         \
1510 "       mulscc  %%g1,%3,%%g1\n"                                         \
1511 "       mulscc  %%g1,%3,%%g1\n"                                         \
1512 "       mulscc  %%g1,%3,%%g1\n"                                         \
1513 "       mulscc  %%g1,%3,%%g1\n"                                         \
1514 "       mulscc  %%g1,%3,%%g1\n"                                         \
1515 "       mulscc  %%g1,%3,%%g1\n"                                         \
1516 "       mulscc  %%g1,%3,%%g1\n"                                         \
1517 "       mulscc  %%g1,%3,%%g1\n"                                         \
1518 "       mulscc  %%g1,%3,%%g1\n"                                         \
1519 "       mulscc  %%g1,%3,%%g1\n"                                         \
1520 "       mulscc  %%g1,%3,%%g1\n"                                         \
1521 "       mulscc  %%g1,%3,%%g1\n"                                         \
1522 "       mulscc  %%g1,%3,%%g1\n"                                         \
1523 "       mulscc  %%g1,%3,%%g1\n"                                         \
1524 "       mulscc  %%g1,%3,%%g1\n"                                         \
1525 "       mulscc  %%g1,%3,%%g1\n"                                         \
1526 "       mulscc  %%g1,%3,%%g1\n"                                         \
1527 "       mulscc  %%g1,%3,%%g1\n"                                         \
1528 "       mulscc  %%g1,%3,%%g1\n"                                         \
1529 "       mulscc  %%g1,%3,%%g1\n"                                         \
1530 "       mulscc  %%g1,%3,%%g1\n"                                         \
1531 "       mulscc  %%g1,%3,%%g1\n"                                         \
1532 "       mulscc  %%g1,%3,%%g1\n"                                         \
1533 "       mulscc  %%g1,%3,%%g1\n"                                         \
1534 "       mulscc  %%g1,0,%%g1\n"                                          \
1535 "       add     %%g1,%%g2,%0\n"                                         \
1536 "       rd      %%y,%1"                                                 \
1537            : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v)                  \
1538            : "%g1", "%g2" __AND_CLOBBER_CC)
1539 #define UMUL_TIME 39            /* 39 instructions */
1540 #endif
1541 #ifndef udiv_qrnnd
1542 #ifndef LONGLONG_STANDALONE
1543 #define udiv_qrnnd(q, r, n1, n0, d) \
1544   do { UWtype __r;                                                      \
1545     (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d));                    \
1546     (r) = __r;                                                          \
1547   } while (0)
1548 extern UWtype __MPN(udiv_qrnnd) _PROTO ((UWtype *, UWtype, UWtype, UWtype));
1549 #ifndef UDIV_TIME
1550 #define UDIV_TIME 140
1551 #endif
1552 #endif /* LONGLONG_STANDALONE */
1553 #endif /* udiv_qrnnd */
1554 #endif /* __sparc__ */
1555
1556 #if defined (__sparc__) && W_TYPE_SIZE == 64
1557 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1558   __asm__ (                                                             \
1559        "addcc   %r4,%5,%1\n"                                            \
1560       " addccc  %r6,%7,%%g0\n"                                          \
1561       " addc    %r2,%3,%0"                                              \
1562           : "=r" (sh), "=&r" (sl)                                       \
1563           : "rJ" (ah), "rI" (bh), "%rJ" (al), "rI" (bl),                \
1564             "%rJ" ((al) >> 32), "rI" ((bl) >> 32)                       \
1565            __CLOBBER_CC)
1566 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1567   __asm__ (                                                             \
1568        "subcc   %r4,%5,%1\n"                                            \
1569       " subccc  %r6,%7,%%g0\n"                                          \
1570       " subc    %r2,%3,%0"                                              \
1571           : "=r" (sh), "=&r" (sl)                                       \
1572           : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl),         \
1573             "rJ" ((al) >> 32), "rI" ((bl) >> 32)                        \
1574            __CLOBBER_CC)
1575 #endif
1576
1577 #if defined (__vax__) && W_TYPE_SIZE == 32
1578 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1579   __asm__ ("addl2 %5,%1\n\tadwc %3,%0"                                  \
1580            : "=g" (sh), "=&g" (sl)                                      \
1581            : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),                 \
1582              "%1" ((USItype)(al)), "g" ((USItype)(bl)))
1583 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1584   __asm__ ("subl2 %5,%1\n\tsbwc %3,%0"                                  \
1585            : "=g" (sh), "=&g" (sl)                                      \
1586            : "0" ((USItype)(ah)), "g" ((USItype)(bh)),                  \
1587              "1" ((USItype)(al)), "g" ((USItype)(bl)))
1588 #define smul_ppmm(xh, xl, m0, m1) \
1589   do {                                                                  \
1590     union {UDItype __ll;                                                \
1591            struct {USItype __l, __h;} __i;                              \
1592           } __x;                                                        \
1593     USItype __m0 = (m0), __m1 = (m1);                                   \
1594     __asm__ ("emul %1,%2,$0,%0"                                         \
1595              : "=g" (__x.__ll) : "g" (__m0), "g" (__m1));               \
1596     (xh) = __x.__i.__h; (xl) = __x.__i.__l;                             \
1597   } while (0)
1598 #define sdiv_qrnnd(q, r, n1, n0, d) \
1599   do {                                                                  \
1600     union {DItype __ll;                                                 \
1601            struct {SItype __l, __h;} __i;                               \
1602           } __x;                                                        \
1603     __x.__i.__h = n1; __x.__i.__l = n0;                                 \
1604     __asm__ ("ediv %3,%2,%0,%1"                                         \
1605              : "=g" (q), "=g" (r) : "g" (__x.__ll), "g" (d));           \
1606   } while (0)
1607 #if 0
1608 /* Note: the following FIXME comes from GMP, thus it does make sense to try
1609    to resolve it in MPFR. */
1610 /* FIXME: This instruction appears to be unimplemented on some systems (vax
1611    8800 maybe). */
1612 #define count_trailing_zeros(count,x)                                   \
1613   do {                                                                  \
1614     __asm__ ("ffs 0, 31, %1, %0"                                        \
1615              : "=g" (count)                                             \
1616              : "g" ((USItype) (x)));                                    \
1617   } while (0)
1618 #endif
1619 #endif /* __vax__ */
1620
1621 #if defined (__z8000__) && W_TYPE_SIZE == 16
1622 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1623   __asm__ ("add %H1,%H5\n\tadc  %H0,%H3"                                \
1624            : "=r" (sh), "=&r" (sl)                                      \
1625            : "0"  ((unsigned int)(ah)), "r" ((unsigned int)(bh)),       \
1626              "%1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
1627 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1628   __asm__ ("sub %H1,%H5\n\tsbc  %H0,%H3"                                \
1629            : "=r" (sh), "=&r" (sl)                                      \
1630            : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)),        \
1631              "1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
1632 #define umul_ppmm(xh, xl, m0, m1) \
1633   do {                                                                  \
1634     union {long int __ll;                                               \
1635            struct {unsigned int __h, __l;} __i;                         \
1636           } __x;                                                        \
1637     unsigned int __m0 = (m0), __m1 = (m1);                              \
1638     __asm__ ("mult      %S0,%H3"                                        \
1639              : "=r" (__x.__i.__h), "=r" (__x.__i.__l)                   \
1640              : "%1" (m0), "rQR" (m1));                                  \
1641     (xh) = __x.__i.__h; (xl) = __x.__i.__l;                             \
1642     (xh) += ((((signed int) __m0 >> 15) & __m1)                         \
1643              + (((signed int) __m1 >> 15) & __m0));                     \
1644   } while (0)
1645 #endif /* __z8000__ */
1646
1647 #endif /* __GNUC__ */
1648
1649 #endif /* NO_ASM */
1650
1651
1652 #if !defined (umul_ppmm) && defined (__umulsidi3)
1653 #define umul_ppmm(ph, pl, m0, m1) \
1654   {                                                                     \
1655     UDWtype __ll = __umulsidi3 (m0, m1);                                \
1656     ph = (UWtype) (__ll >> W_TYPE_SIZE);                                \
1657     pl = (UWtype) __ll;                                                 \
1658   }
1659 #endif
1660
1661 #if !defined (__umulsidi3)
1662 #define __umulsidi3(u, v) \
1663   ({UWtype __hi, __lo;                                                  \
1664     umul_ppmm (__hi, __lo, u, v);                                       \
1665     ((UDWtype) __hi << W_TYPE_SIZE) | __lo; })
1666 #endif
1667
1668
1669 /* Use mpn_umul_ppmm or mpn_udiv_qrnnd functions, if they exist.  The "_r"
1670    forms have "reversed" arguments, meaning the pointer is last, which
1671    sometimes allows better parameter passing, in particular on 64-bit
1672    hppa. */
1673
1674 #define mpn_umul_ppmm  __MPN(umul_ppmm)
1675 extern UWtype mpn_umul_ppmm _PROTO ((UWtype *, UWtype, UWtype));
1676
1677 #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm  \
1678   && ! defined (LONGLONG_STANDALONE)
1679 #define umul_ppmm(wh, wl, u, v)                                               \
1680   do {                                                                        \
1681     UWtype __umul_ppmm__p0;                                                   \
1682     (wh) = mpn_umul_ppmm (&__umul_ppmm__p0, (UWtype) (u), (UWtype) (v));      \
1683     (wl) = __umul_ppmm__p0;                                                   \
1684   } while (0)
1685 #endif
1686
1687 #define mpn_umul_ppmm_r  __MPN(umul_ppmm_r)
1688 extern UWtype mpn_umul_ppmm_r _PROTO ((UWtype, UWtype, UWtype *));
1689
1690 #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm_r        \
1691   && ! defined (LONGLONG_STANDALONE)
1692 #define umul_ppmm(wh, wl, u, v)                                               \
1693   do {                                                                        \
1694     UWtype __umul_ppmm__p0;                                                   \
1695     (wh) = mpn_umul_ppmm_r ((UWtype) (u), (UWtype) (v), &__umul_ppmm__p0);    \
1696     (wl) = __umul_ppmm__p0;                                                   \
1697   } while (0)
1698 #endif
1699
1700 #define mpn_udiv_qrnnd  __MPN(udiv_qrnnd)
1701 extern UWtype mpn_udiv_qrnnd _PROTO ((UWtype *, UWtype, UWtype, UWtype));
1702
1703 #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd        \
1704   && ! defined (LONGLONG_STANDALONE)
1705 #define udiv_qrnnd(q, r, n1, n0, d)                                     \
1706   do {                                                                  \
1707     UWtype __udiv_qrnnd__r;                                             \
1708     (q) = mpn_udiv_qrnnd (&__udiv_qrnnd__r,                             \
1709                           (UWtype) (n1), (UWtype) (n0), (UWtype) d);    \
1710     (r) = __udiv_qrnnd__r;                                              \
1711   } while (0)
1712 #endif
1713
1714 #define mpn_udiv_qrnnd_r  __MPN(udiv_qrnnd_r)
1715 extern UWtype mpn_udiv_qrnnd_r _PROTO ((UWtype, UWtype, UWtype, UWtype *));
1716
1717 #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd_r      \
1718   && ! defined (LONGLONG_STANDALONE)
1719 #define udiv_qrnnd(q, r, n1, n0, d)                                     \
1720   do {                                                                  \
1721     UWtype __udiv_qrnnd__r;                                             \
1722     (q) = mpn_udiv_qrnnd_r ((UWtype) (n1), (UWtype) (n0), (UWtype) d,   \
1723                             &__udiv_qrnnd__r);                          \
1724     (r) = __udiv_qrnnd__r;                                              \
1725   } while (0)
1726 #endif
1727
1728
1729 /* If this machine has no inline assembler, use C macros.  */
1730
1731 #if !defined (add_ssaaaa)
1732 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1733   do {                                                                  \
1734     UWtype __x;                                                         \
1735     __x = (al) + (bl);                                                  \
1736     (sh) = (ah) + (bh) + (__x < (al));                                  \
1737     (sl) = __x;                                                         \
1738   } while (0)
1739 #endif
1740
1741 #if !defined (sub_ddmmss)
1742 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1743   do {                                                                  \
1744     UWtype __x;                                                         \
1745     __x = (al) - (bl);                                                  \
1746     (sh) = (ah) - (bh) - ((al) < (bl));                                 \
1747     (sl) = __x;                                                         \
1748   } while (0)
1749 #endif
1750
1751 /* If we lack umul_ppmm but have smul_ppmm, define umul_ppmm in terms of
1752    smul_ppmm.  */
1753 #if !defined (umul_ppmm) && defined (smul_ppmm)
1754 #define umul_ppmm(w1, w0, u, v)                                         \
1755   do {                                                                  \
1756     UWtype __w1;                                                        \
1757     UWtype __xm0 = (u), __xm1 = (v);                                    \
1758     smul_ppmm (__w1, w0, __xm0, __xm1);                                 \
1759     (w1) = __w1 + (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1)               \
1760                 + (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0);              \
1761   } while (0)
1762 #endif
1763
1764 /* If we still don't have umul_ppmm, define it using plain C.
1765
1766    For reference, when this code is used for squaring (ie. u and v identical
1767    expressions), gcc recognises __x1 and __x2 are the same and generates 3
1768    multiplies, not 4.  The subsequent additions could be optimized a bit,
1769    but the only place GMP currently uses such a square is mpn_sqr_basecase,
1770    and chips obliged to use this generic C umul will have plenty of worse
1771    performance problems than a couple of extra instructions on the diagonal
1772    of sqr_basecase.  */
1773
1774 #if !defined (umul_ppmm)
1775 #define umul_ppmm(w1, w0, u, v)                                         \
1776   do {                                                                  \
1777     UWtype __x0, __x1, __x2, __x3;                                      \
1778     UHWtype __ul, __vl, __uh, __vh;                                     \
1779     UWtype __u = (u), __v = (v);                                        \
1780                                                                         \
1781     __ul = __ll_lowpart (__u);                                          \
1782     __uh = __ll_highpart (__u);                                         \
1783     __vl = __ll_lowpart (__v);                                          \
1784     __vh = __ll_highpart (__v);                                         \
1785                                                                         \
1786     __x0 = (UWtype) __ul * __vl;                                        \
1787     __x1 = (UWtype) __ul * __vh;                                        \
1788     __x2 = (UWtype) __uh * __vl;                                        \
1789     __x3 = (UWtype) __uh * __vh;                                        \
1790                                                                         \
1791     __x1 += __ll_highpart (__x0);/* this can't give carry */            \
1792     __x1 += __x2;               /* but this indeed can */               \
1793     if (__x1 < __x2)            /* did we get it? */                    \
1794       __x3 += __ll_B;           /* yes, add it in the proper pos. */    \
1795                                                                         \
1796     (w1) = __x3 + __ll_highpart (__x1);                                 \
1797     (w0) = (__x1 << W_TYPE_SIZE/2) + __ll_lowpart (__x0);               \
1798   } while (0)
1799 #endif
1800
1801 /* If we don't have smul_ppmm, define it using umul_ppmm (which surely will
1802    exist in one form or another.  */
1803 #if !defined (smul_ppmm)
1804 #define smul_ppmm(w1, w0, u, v)                                         \
1805   do {                                                                  \
1806     UWtype __w1;                                                        \
1807     UWtype __xm0 = (u), __xm1 = (v);                                    \
1808     umul_ppmm (__w1, w0, __xm0, __xm1);                                 \
1809     (w1) = __w1 - (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1)               \
1810                 - (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0);              \
1811   } while (0)
1812 #endif
1813
1814 /* Define this unconditionally, so it can be used for debugging.  */
1815 #define __udiv_qrnnd_c(q, r, n1, n0, d) \
1816   do {                                                                  \
1817     UWtype __d1, __d0, __q1, __q0, __r1, __r0, __m;                     \
1818                                                                         \
1819     ASSERT ((d) != 0);                                                  \
1820     ASSERT ((n1) < (d));                                                \
1821                                                                         \
1822     __d1 = __ll_highpart (d);                                           \
1823     __d0 = __ll_lowpart (d);                                            \
1824                                                                         \
1825     __q1 = (n1) / __d1;                                                 \
1826     __r1 = (n1) - __q1 * __d1;                                          \
1827     __m = __q1 * __d0;                                                  \
1828     __r1 = __r1 * __ll_B | __ll_highpart (n0);                          \
1829     if (__r1 < __m)                                                     \
1830       {                                                                 \
1831         __q1--, __r1 += (d);                                            \
1832         if (__r1 >= (d)) /* i.e. we didn't get carry when adding to __r1 */\
1833           if (__r1 < __m)                                               \
1834             __q1--, __r1 += (d);                                        \
1835       }                                                                 \
1836     __r1 -= __m;                                                        \
1837                                                                         \
1838     __q0 = __r1 / __d1;                                                 \
1839     __r0 = __r1  - __q0 * __d1;                                         \
1840     __m = __q0 * __d0;                                                  \
1841     __r0 = __r0 * __ll_B | __ll_lowpart (n0);                           \
1842     if (__r0 < __m)                                                     \
1843       {                                                                 \
1844         __q0--, __r0 += (d);                                            \
1845         if (__r0 >= (d))                                                \
1846           if (__r0 < __m)                                               \
1847             __q0--, __r0 += (d);                                        \
1848       }                                                                 \
1849     __r0 -= __m;                                                        \
1850                                                                         \
1851     (q) = __q1 * __ll_B | __q0;                                         \
1852     (r) = __r0;                                                         \
1853   } while (0)
1854
1855 /* If the processor has no udiv_qrnnd but sdiv_qrnnd, go through
1856    __udiv_w_sdiv (defined in libgcc or elsewhere).  */
1857 #if !defined (udiv_qrnnd) && defined (sdiv_qrnnd)
1858 #define udiv_qrnnd(q, r, nh, nl, d) \
1859   do {                                                                  \
1860     UWtype __r;                                                         \
1861     (q) = __MPN(udiv_w_sdiv) (&__r, nh, nl, d);                         \
1862     (r) = __r;                                                          \
1863   } while (0)
1864 #endif
1865
1866 /* If udiv_qrnnd was not defined for this processor, use __udiv_qrnnd_c.  */
1867 #if !defined (udiv_qrnnd)
1868 #define UDIV_NEEDS_NORMALIZATION 1
1869 #define udiv_qrnnd __udiv_qrnnd_c
1870 #endif
1871
1872 #if !defined (count_leading_zeros)
1873 #define count_leading_zeros(count, x) \
1874   do {                                                                  \
1875     UWtype __xr = (x);                                                  \
1876     UWtype __a;                                                         \
1877                                                                         \
1878     if (W_TYPE_SIZE == 32)                                              \
1879       {                                                                 \
1880         __a = __xr < ((UWtype) 1 << 2*__BITS4)                          \
1881           ? (__xr < ((UWtype) 1 << __BITS4) ? 1 : __BITS4 + 1)          \
1882           : (__xr < ((UWtype) 1 << 3*__BITS4) ? 2*__BITS4 + 1           \
1883           : 3*__BITS4 + 1);                                             \
1884       }                                                                 \
1885     else                                                                \
1886       {                                                                 \
1887         for (__a = W_TYPE_SIZE - 8; __a > 0; __a -= 8)                  \
1888           if (((__xr >> __a) & 0xff) != 0)                              \
1889             break;                                                      \
1890         ++__a;                                                          \
1891       }                                                                 \
1892                                                                         \
1893     (count) = W_TYPE_SIZE + 1 - __a - __clz_tab[__xr >> __a];           \
1894   } while (0)
1895 /* This version gives a well-defined value for zero. */
1896 #define COUNT_LEADING_ZEROS_0 (W_TYPE_SIZE - 1)
1897 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
1898 #endif
1899
1900 /* clz_tab needed by mpn/x86/pentium/mod_1.asm in a fat binary */
1901 #if HAVE_HOST_CPU_FAMILY_x86 && WANT_FAT_BINARY
1902 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
1903 #endif
1904
1905 #ifdef COUNT_LEADING_ZEROS_NEED_CLZ_TAB
1906 # ifdef MPFR_HAVE_GMP_IMPL
1907     extern const unsigned char __GMP_DECLSPEC __clz_tab[128];
1908 # else
1909     extern const unsigned char __clz_tab[128];
1910 # endif
1911 #endif
1912
1913 #if !defined (count_trailing_zeros)
1914 /* Define count_trailing_zeros using count_leading_zeros.  The latter might be
1915    defined in asm, but if it is not, the C version above is good enough.  */
1916 #define count_trailing_zeros(count, x) \
1917   do {                                                                  \
1918     UWtype __ctz_x = (x);                                               \
1919     UWtype __ctz_c;                                                     \
1920     ASSERT (__ctz_x != 0);                                              \
1921     count_leading_zeros (__ctz_c, __ctz_x & -__ctz_x);                  \
1922     (count) = W_TYPE_SIZE - 1 - __ctz_c;                                \
1923   } while (0)
1924 #endif
1925
1926 #ifndef UDIV_NEEDS_NORMALIZATION
1927 #define UDIV_NEEDS_NORMALIZATION 0
1928 #endif
1929
1930 /* Whether udiv_qrnnd is actually implemented with udiv_qrnnd_preinv, and
1931    that hence the latter should always be used.  */
1932 #ifndef UDIV_PREINV_ALWAYS
1933 #define UDIV_PREINV_ALWAYS 0
1934 #endif
1935
1936 /* Give defaults for UMUL_TIME and UDIV_TIME.  */
1937 #ifndef UMUL_TIME
1938 #define UMUL_TIME 1
1939 #endif
1940
1941 #ifndef UDIV_TIME
1942 #define UDIV_TIME UMUL_TIME
1943 #endif