[Gmp-commit] /var/hg/gmp: Whitespace cleanup.
mercurial at gmplib.org
mercurial at gmplib.org
Sat Jun 2 19:03:09 CEST 2012
details: /var/hg/gmp/rev/21c17d4c7a2b
changeset: 15044:21c17d4c7a2b
user: Torbjorn Granlund <tege at gmplib.org>
date: Sat Jun 02 19:03:06 2012 +0200
description:
Whitespace cleanup.
diffstat:
AUTHORS | 8 ++--
Makefile.am | 2 +-
gmp-impl.h | 48 +++++++++++++++++-----------------
mpn/powerpc32/invert_limb.asm | 2 +-
mpn/powerpc64/mode64/sqr_basecase.asm | 2 +-
mpn/s390_32/logops_n.asm | 2 +-
mpn/s390_64/logops_n.asm | 2 +-
mpn/x86/atom/aorrlshC_n.asm | 2 +-
mpn/x86/atom/aorslshC_n.asm | 6 ++--
mpn/x86_64/k8/aorrlsh_n.asm | 6 ++--
mpn/x86_64/mod_1_1.asm | 4 +-
mpn/x86_64/mullo_basecase.asm | 4 +-
mpz/jacobi.c | 2 +-
tests/mpz/reuse.c | 6 ++--
tests/refmpn.c | 12 ++++----
tune/speed.h | 32 +++++++++++-----------
16 files changed, 70 insertions(+), 70 deletions(-)
diffs (truncated from 473 to 300 lines):
diff -r 41347195eaa3 -r 21c17d4c7a2b AUTHORS
--- a/AUTHORS Sat Jun 02 07:51:02 2012 +0200
+++ b/AUTHORS Sat Jun 02 19:03:06 2012 +0200
@@ -51,11 +51,11 @@
mullo_n.c, invert.c, invertappr.c.
David Harvey mpn/generic/add_err1_n.c, add_err2_n.c,
- add_err3_n.c, sub_err1_n.c, sub_err2_n.c,
- sub_err3_n.c, mulmid_basecase.c, mulmid_n.c,
- toom42_mulmid.c,
+ add_err3_n.c, sub_err1_n.c, sub_err2_n.c,
+ sub_err3_n.c, mulmid_basecase.c, mulmid_n.c,
+ toom42_mulmid.c,
mpn/x86_64/mul_basecase.asm, aors_err1_n.asm,
- aors_err2_n.asm, aors_err3_n.asm,
+ aors_err2_n.asm, aors_err3_n.asm,
mulmid_basecase.asm,
mpn/x86_64/core2/aors_err1_n.asm.
diff -r 41347195eaa3 -r 21c17d4c7a2b Makefile.am
--- a/Makefile.am Sat Jun 02 07:51:02 2012 +0200
+++ b/Makefile.am Sat Jun 02 19:03:06 2012 +0200
@@ -167,7 +167,7 @@
mpz/gcd_ui$U.lo mpz/gcdext$U.lo mpz/get_d$U.lo mpz/get_d_2exp$U.lo \
mpz/get_si$U.lo mpz/get_str$U.lo mpz/get_ui$U.lo mpz/getlimbn$U.lo \
mpz/hamdist$U.lo \
- mpz/import$U.lo mpz/init$U.lo mpz/init2$U.lo mpz/inits$U.lo \
+ mpz/import$U.lo mpz/init$U.lo mpz/init2$U.lo mpz/inits$U.lo \
mpz/inp_raw$U.lo mpz/inp_str$U.lo mpz/invert$U.lo \
mpz/ior$U.lo mpz/iset$U.lo mpz/iset_d$U.lo mpz/iset_si$U.lo \
mpz/iset_str$U.lo mpz/iset_ui$U.lo mpz/jacobi$U.lo mpz/kronsz$U.lo \
diff -r 41347195eaa3 -r 21c17d4c7a2b gmp-impl.h
--- a/gmp-impl.h Sat Jun 02 07:51:02 2012 +0200
+++ b/gmp-impl.h Sat Jun 02 19:03:06 2012 +0200
@@ -314,9 +314,9 @@
/* Usage: TMP_DECL;
- TMP_MARK;
- ptr = TMP_ALLOC (bytes);
- TMP_FREE;
+ TMP_MARK;
+ ptr = TMP_ALLOC (bytes);
+ TMP_FREE;
Small allocations should use TMP_SALLOC, big allocations should use
TMP_BALLOC. Allocations that might be small or big should use TMP_ALLOC.
@@ -1820,11 +1820,11 @@
std/repe/scasl/cld and cld/repe/scasl (the latter would be for stripping
low zeros).
- std cld
- P5 18 16
- P6 46 38
- K6 36 13
- K7 21 20
+ std cld
+ P5 18 16
+ P6 46 38
+ K6 36 13
+ K7 21 20
*/
#ifndef MPN_NORMALIZE
#define MPN_NORMALIZE(DST, NLIMBS) \
@@ -2166,7 +2166,7 @@
where FFT_FIRST_K+1 should be used, the second FFT_FIRST_K+2,
etc. See mpn_fft_best_k(). */
#ifndef MUL_FFT_TABLE
-#define MUL_FFT_TABLE \
+#define MUL_FFT_TABLE \
{ MUL_TOOM33_THRESHOLD * 4, /* k=5 */ \
MUL_TOOM33_THRESHOLD * 8, /* k=6 */ \
MUL_TOOM33_THRESHOLD * 16, /* k=7 */ \
@@ -2176,7 +2176,7 @@
0 }
#endif
#ifndef SQR_FFT_TABLE
-#define SQR_FFT_TABLE \
+#define SQR_FFT_TABLE \
{ SQR_TOOM3_THRESHOLD * 4, /* k=5 */ \
SQR_TOOM3_THRESHOLD * 8, /* k=6 */ \
SQR_TOOM3_THRESHOLD * 16, /* k=7 */ \
@@ -2941,7 +2941,7 @@
_p += _t1; \
if (_p < _t1) \
{ \
- _v--; \
+ _v--; \
if (UNLIKELY (_p >= (d1))) \
{ \
if (_p > (d1) || _t0 >= (d0)) \
@@ -3221,17 +3221,17 @@
\
if ((a) <= (d)) \
{ \
- /* small a is reasonably likely */ \
- (r) = (d) - (a); \
+ /* small a is reasonably likely */ \
+ (r) = (d) - (a); \
} \
else \
{ \
- unsigned __twos; \
- mp_limb_t __dnorm; \
- count_leading_zeros (__twos, d); \
- __twos -= GMP_NAIL_BITS; \
- __dnorm = (d) << __twos; \
- (r) = ((a) <= __dnorm ? __dnorm : 2*__dnorm) - (a); \
+ unsigned __twos; \
+ mp_limb_t __dnorm; \
+ count_leading_zeros (__twos, d); \
+ __twos -= GMP_NAIL_BITS; \
+ __dnorm = (d) << __twos; \
+ (r) = ((a) <= __dnorm ? __dnorm : 2*__dnorm) - (a); \
} \
\
ASSERT_LIMB (r); \
@@ -3304,8 +3304,8 @@
int __p = 0; \
do \
{ \
- __p ^= 0x96696996L >> (__n & 0x1F); \
- __n >>= 5; \
+ __p ^= 0x96696996L >> (__n & 0x1F); \
+ __n >>= 5; \
} \
while (__n != 0); \
\
@@ -3378,7 +3378,7 @@
#if ! defined (BSWAP_LIMB)
#if GMP_LIMB_BITS == 8
-#define BSWAP_LIMB(dst, src) \
+#define BSWAP_LIMB(dst, src) \
do { (dst) = (src); } while (0)
#endif
#if GMP_LIMB_BITS == 16
@@ -4888,11 +4888,11 @@
#undef FAC_ODD_THRESHOLD
#define FAC_ODD_THRESHOLD fac_odd_threshold
-extern mp_size_t fac_odd_threshold;
+extern mp_size_t fac_odd_threshold;
#undef FAC_DSC_THRESHOLD
#define FAC_DSC_THRESHOLD fac_dsc_threshold
-extern mp_size_t fac_dsc_threshold;
+extern mp_size_t fac_dsc_threshold;
#undef FFT_TABLE_ATTRS
#define FFT_TABLE_ATTRS
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/powerpc32/invert_limb.asm
--- a/mpn/powerpc32/invert_limb.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/powerpc32/invert_limb.asm Sat Jun 02 19:03:06 2012 +0200
@@ -33,7 +33,7 @@
ASM_START()
PROLOGUE(mpn_invert_limb)
rlwinm r6, r3, 11, 22, 30 C extract bits 30..22 to pos 2^1
- srwi r10, r3, 11 C extract bits 31..11
+ srwi r10, r3, 11 C extract bits 31..11
LEA( r9, approx_tab) C N.B. clobbers r0 for ELF and Darwin
lhzx r9, r9, r6 C w2
addi r0, r10, 1
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/powerpc64/mode64/sqr_basecase.asm
--- a/mpn/powerpc64/mode64/sqr_basecase.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/powerpc64/mode64/sqr_basecase.asm Sat Jun 02 19:03:06 2012 +0200
@@ -673,7 +673,7 @@
rldicl. r0, n, 0,62 C r0 = n & 3, set cr0
cmpdi cr6, r0, 2
- addi n, n, 2 C compute count...
+ addi n, n, 2 C compute count...
srdi n, n, 2 C ...for ctr
mtctr n C put loop count into ctr
beq cr0, L(xb0)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/s390_32/logops_n.asm
--- a/mpn/s390_32/logops_n.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/s390_32/logops_n.asm Sat Jun 02 19:03:06 2012 +0200
@@ -84,7 +84,7 @@
ltr %r1, %r1 C < 256 bytes to copy?
je L(1)
-L(tp): LOGOPC 0(256, rp), 0(vp)
+L(tp): LOGOPC 0(256, rp), 0(vp)
la rp, 256(rp)
la vp, 256(vp)
brct %r1, L(tp)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/s390_64/logops_n.asm
--- a/mpn/s390_64/logops_n.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/s390_64/logops_n.asm Sat Jun 02 19:03:06 2012 +0200
@@ -83,7 +83,7 @@
ltgr %r1, %r1 C < 256 bytes to copy?
je L(1)
-L(tp): LOGOPC 0(256, rp), 0(vp)
+L(tp): LOGOPC 0(256, rp), 0(vp)
la rp, 256(rp)
la vp, 256(vp)
brctg %r1, L(tp)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/x86/atom/aorrlshC_n.asm
--- a/mpn/x86/atom/aorrlshC_n.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/x86/atom/aorrlshC_n.asm Sat Jun 02 19:03:06 2012 +0200
@@ -112,7 +112,7 @@
shr $RSH, %ecx
mov 4(vp), %eax
shr %edx
- lea 8(vp), vp
+ lea 8(vp), vp
M4_inst (up), %ebp
lea (%ecx,%eax,M), %edx
mov %ebp, (rp)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/x86/atom/aorslshC_n.asm
--- a/mpn/x86/atom/aorslshC_n.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/x86/atom/aorslshC_n.asm Sat Jun 02 19:03:06 2012 +0200
@@ -120,7 +120,7 @@
shr $RSH, %eax
mov 4(up), %ecx
add %edx, %edx
- lea 8(up), up
+ lea 8(up), up
M4_inst %ebp, (rp)
lea (%eax,%ecx,M), %eax
@@ -208,12 +208,12 @@
mov %edx, (rp)
L(enteven2):
mov 4(%ebx), %edx
- lea 8(%ebx), %ebx
+ lea 8(%ebx), %ebx
M4_inst %eax, %edx
mov %edx, 4(rp)
sbb %edx, %edx
shr $RSH, %ecx
- lea 8(rp), rp
+ lea 8(rp), rp
L(entry2):
mov (up), %eax
decl GPARAM_SIZE
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/x86_64/k8/aorrlsh_n.asm
--- a/mpn/x86_64/k8/aorrlsh_n.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/x86_64/k8/aorrlsh_n.asm Sat Jun 02 19:03:06 2012 +0200
@@ -59,8 +59,8 @@
ABI_SUPPORT(STD64)
ASM_START()
- TEXT
- ALIGN(16)
+ TEXT
+ ALIGN(16)
PROLOGUE(func)
FUNC_ENTRY(4)
IFDOS(` mov 56(%rsp), %r8d ')
@@ -83,7 +83,7 @@
mov $1, R32(%r8)
shl R8(%rcx), %r8
- mul %r8 C initial multiply
+ mul %r8 C initial multiply
and $3, R32(%r9)
jz L(b0)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/x86_64/mod_1_1.asm
--- a/mpn/x86_64/mod_1_1.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/x86_64/mod_1_1.asm Sat Jun 02 19:03:06 2012 +0200
@@ -111,8 +111,8 @@
add %rax, r0
mov r2, %rax
adc %rdx, %rax
- sbb r2, r2
- sub $1, n
+ sbb r2, r2
+ sub $1, n
jnc L(top)
L(reduce_three):
diff -r 41347195eaa3 -r 21c17d4c7a2b mpn/x86_64/mullo_basecase.asm
--- a/mpn/x86_64/mullo_basecase.asm Sat Jun 02 07:51:02 2012 +0200
+++ b/mpn/x86_64/mullo_basecase.asm Sat Jun 02 19:03:06 2012 +0200
@@ -90,7 +90,7 @@
L(2): mov 8(vp_param), %r11
imul %rax, %r11 C u0 x v1
- mul %r8 C u0 x v0
+ mul %r8 C u0 x v0
mov %rax, (rp)
imul 8(up), %r8 C u1 x v0
lea (%r11, %rdx), %rax
@@ -117,7 +117,7 @@
add %rax, %rcx
adc %rdx, %r9
mov %r11, %rax
- imul (up), %rax C u0 x v2 -> r2
+ imul (up), %rax C u0 x v2 -> r2
add %rax, %r9
mov %rcx, 8(rp)
mov %r9, 16(rp)
diff -r 41347195eaa3 -r 21c17d4c7a2b mpz/jacobi.c
--- a/mpz/jacobi.c Sat Jun 02 07:51:02 2012 +0200
+++ b/mpz/jacobi.c Sat Jun 02 19:03:06 2012 +0200
@@ -131,7 +131,7 @@
More information about the gmp-commit
mailing list