diff options
author | jschauma <jschauma> | 2004-01-13 23:40:24 +0000 |
---|---|---|
committer | jschauma <jschauma> | 2004-01-13 23:40:24 +0000 |
commit | e761c26f134aa8a1d803907e40ebb6ae2c7fbdad (patch) | |
tree | 94fadcdf5b10a05d23e9f3bd0451b258c9059b9d /lang/drscheme | |
parent | 451051bb5571dd02e9fb40bc3f9b8ea66ae6f4a6 (diff) | |
download | pkgsrc-e761c26f134aa8a1d803907e40ebb6ae2c7fbdad.tar.gz |
Make this compile with gcc-3.3 by replacing multi-line strings
"foo
string"
in inline asm with
"foo\n"
"string"
(Noted in agc's latest bulk build.)
Diffstat (limited to 'lang/drscheme')
-rw-r--r-- | lang/drscheme/distinfo | 3 | ||||
-rw-r--r-- | lang/drscheme/patches/patch-aq | 346 |
2 files changed, 348 insertions, 1 deletions
diff --git a/lang/drscheme/distinfo b/lang/drscheme/distinfo index afd7824070b..0016d80faf4 100644 --- a/lang/drscheme/distinfo +++ b/lang/drscheme/distinfo @@ -1,4 +1,4 @@ -$NetBSD: distinfo,v 1.8 2003/05/03 20:57:03 jmc Exp $ +$NetBSD: distinfo,v 1.9 2004/01/13 23:40:24 jschauma Exp $ SHA1 (drscheme/202/plt.src.x.tar.gz) = bb06b6e3f8214db71ebb12e95e744f996622ee06 Size (drscheme/202/plt.src.x.tar.gz) = 6133563 bytes @@ -18,3 +18,4 @@ SHA1 (patch-am) = afeb2c469b55f1b9fdadf4e064928b59bf28d794 SHA1 (patch-an) = bca5a79edd48a2670d02496b6da3799c13c0aed6 SHA1 (patch-ao) = 667593d28a695fd7caee210deb8ee06e72019c40 SHA1 (patch-ap) = 2395a864a40af34f295e0ed782979d90d75f1ec4 +SHA1 (patch-aq) = 058affb4c3f840f27ce88aba0baaeacba2eb8282 diff --git a/lang/drscheme/patches/patch-aq b/lang/drscheme/patches/patch-aq new file mode 100644 index 00000000000..cae92ec4a6f --- /dev/null +++ b/lang/drscheme/patches/patch-aq @@ -0,0 +1,346 @@ +$NetBSD: patch-aq,v 1.1 2004/01/13 23:40:25 jschauma Exp $ + +--- ./mzscheme/src/gmp/gmplonglong.h.orig 2001-06-26 12:46:25.000000000 -0400 ++++ ./mzscheme/src/gmp/gmplonglong.h 2004-01-13 18:20:47.000000000 -0500 +@@ -231,19 +231,19 @@ + #define UMUL_TIME 5 + #else + #define umul_ppmm(xh, xl, a, b) \ +- __asm__ ("%@ Inlined umul_ppmm +- mov %|r0, %2, lsr #16 +- mov %|r2, %3, lsr #16 +- bic %|r1, %2, %|r0, lsl #16 +- bic %|r2, %3, %|r2, lsl #16 +- mul %1, %|r1, %|r2 +- mul %|r2, %|r0, %|r2 +- mul %|r1, %0, %|r1 +- mul %0, %|r0, %0 +- adds %|r1, %|r2, %|r1 +- addcs %0, %0, #65536 +- adds %1, %1, %|r1, lsl #16 +- adc %0, %0, %|r1, lsr #16" \ ++ __asm__ ("%@ Inlined umul_ppmm\n" ++" mov %|r0, %2, lsr #16\n" ++" mov %|r2, %3, lsr #16\n" ++" bic %|r1, %2, %|r0, lsl #16\n" ++" bic %|r2, %3, %|r2, lsl #16\n" ++" mul %1, %|r1, %|r2\n" ++" mul %|r2, %|r0, %|r2\n" ++" mul %|r1, %0, %|r1\n" ++" mul %0, %|r0, %0\n" ++" adds %|r1, %|r2, %|r1\n" ++" addcs %0, %0, #65536\n" ++" adds %1, %1, %|r1, lsl #16\n" ++" adc %0, %0, %|r1, lsr #16" \ + : "=&r" (xh), "=r" (xl) \ + : "r" (a), "r" (b) \ + : "r0", "r1", "r2") +@@ -360,22 +360,22 @@ + do { \ + USItype __tmp; \ + __asm__ ( \ +- "ldi 1,%0 +- extru,= %1,15,16,%%r0 ; Bits 31..16 zero? +- extru,tr %1,15,16,%1 ; No. Shift down, skip add. +- ldo 16(%0),%0 ; Yes. Perform add. +- extru,= %1,23,8,%%r0 ; Bits 15..8 zero? +- extru,tr %1,23,8,%1 ; No. Shift down, skip add. +- ldo 8(%0),%0 ; Yes. Perform add. +- extru,= %1,27,4,%%r0 ; Bits 7..4 zero? +- extru,tr %1,27,4,%1 ; No. Shift down, skip add. +- ldo 4(%0),%0 ; Yes. Perform add. +- extru,= %1,29,2,%%r0 ; Bits 3..2 zero? +- extru,tr %1,29,2,%1 ; No. Shift down, skip add. +- ldo 2(%0),%0 ; Yes. Perform add. +- extru %1,30,1,%1 ; Extract bit 1. +- sub %0,%1,%0 ; Subtract it. +- " : "=r" (count), "=r" (__tmp) : "1" (x)); \ ++ "ldi 1,%0\n" ++" extru,= %1,15,16,%%r0 ; Bits 31..16 zero?\n" ++" extru,tr %1,15,16,%1 ; No. Shift down, skip add.\n" ++" ldo 16(%0),%0 ; Yes. Perform add.\n" ++" extru,= %1,23,8,%%r0 ; Bits 15..8 zero?\n" ++" extru,tr %1,23,8,%1 ; No. Shift down, skip add.\n" ++" ldo 8(%0),%0 ; Yes. Perform add.\n" ++" extru,= %1,27,4,%%r0 ; Bits 7..4 zero?\n" ++" extru,tr %1,27,4,%1 ; No. Shift down, skip add.\n" ++" ldo 4(%0),%0 ; Yes. Perform add.\n" ++" extru,= %1,29,2,%%r0 ; Bits 3..2 zero?\n" ++" extru,tr %1,29,2,%1 ; No. Shift down, skip add.\n" ++" ldo 2(%0),%0 ; Yes. Perform add.\n" ++" extru %1,30,1,%1 ; Extract bit 1.\n" ++" sub %0,%1,%0 ; Subtract it.\n" ++ : "=r" (count), "=r" (__tmp) : "1" (x)); \ + } while (0) + #endif /* hppa */ + +@@ -529,28 +529,28 @@ + #else /* for other 68k family members use 16x16->32 multiplication */ + #define umul_ppmm(xh, xl, a, b) \ + do { USItype __umul_tmp1, __umul_tmp2; \ +- __asm__ ("| Inlined umul_ppmm +- move%.l %5,%3 +- move%.l %2,%0 +- move%.w %3,%1 +- swap %3 +- swap %0 +- mulu%.w %2,%1 +- mulu%.w %3,%0 +- mulu%.w %2,%3 +- swap %2 +- mulu%.w %5,%2 +- add%.l %3,%2 +- jcc 1f +- add%.l %#0x10000,%0 +-1: move%.l %2,%3 +- clr%.w %2 +- swap %2 +- swap %3 +- clr%.w %3 +- add%.l %3,%1 +- addx%.l %2,%0 +- | End inlined umul_ppmm" \ ++ __asm__ ("| Inlined umul_ppmm\n" ++" move%.l %5,%3\n" ++" move%.l %2,%0\n" ++" move%.w %3,%1\n" ++" swap %3\n" ++" swap %0\n" ++" mulu%.w %2,%1\n" ++" mulu%.w %3,%0\n" ++" mulu%.w %2,%3\n" ++" swap %2\n" ++" mulu%.w %5,%2\n" ++" add%.l %3,%2\n" ++" jcc 1f\n" ++" add%.l %#0x10000,%0\n" ++"1: move%.l %2,%3\n" ++" clr%.w %2\n" ++" swap %2\n" ++" swap %3\n" ++" clr%.w %3\n" ++" add%.l %3,%1\n" ++" addx%.l %2,%0\n" ++" | End inlined umul_ppmm" \ + : "=&d" ((USItype)(xh)), "=&d" ((USItype)(xl)), \ + "=d" (__umul_tmp1), "=&d" (__umul_tmp2) \ + : "%2" ((USItype)(a)), "d" ((USItype)(b))); \ +@@ -827,26 +827,26 @@ + "1" ((USItype)(al)), "r" ((USItype)(bl))) + #define smul_ppmm(ph, pl, m0, m1) \ + __asm__ ( \ +- "s r2,r2 +- mts r10,%2 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- m r2,%3 +- cas %0,r2,r0 +- mfs r10,%1" \ ++ "s r2,r2\n" ++" mts r10,%2\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" m r2,%3\n" ++" cas %0,r2,r0\n" ++" mfs r10,%1" \ + : "=r" ((USItype)(ph)), "=r" ((USItype)(pl)) \ + : "%r" ((USItype)(m0)), "r" ((USItype)(m1)) \ + : "r2"); \ +@@ -936,45 +936,45 @@ + __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) + #define UMUL_TIME 5 + #define udiv_qrnnd(q, r, n1, n0, d) \ +- __asm__ ("! Inlined udiv_qrnnd +- wr %%g0,%2,%%y ! Not a delayed write for sparclite +- tst %%g0 +- divscc %3,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%%g1 +- divscc %%g1,%4,%0 +- rd %%y,%1 +- bl,a 1f +- add %1,%4,%1 +-1: ! End of inline udiv_qrnnd" \ ++ __asm__ ("! Inlined udiv_qrnnd\n" ++" wr %%g0,%2,%%y ! Not a delayed write for sparclite\n" ++" tst %%g0\n" ++" divscc %3,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%%g1\n" ++" divscc %%g1,%4,%0\n" ++" rd %%y,%1\n" ++" bl,a 1f\n" ++" add %1,%4,%1\n" ++"1: ! End of inline udiv_qrnnd" \ + : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d) + : "%g1" __AND_CLOBBER_CC) + #define UDIV_TIME 37 +@@ -989,46 +989,46 @@ + /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd. */ + #ifndef umul_ppmm + #define umul_ppmm(w1, w0, u, v) \ +- __asm__ ("! Inlined umul_ppmm +- wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr +- sra %3,31,%%g2 ! Don't move this insn +- and %2,%%g2,%%g2 ! Don't move this insn +- andcc %%g0,0,%%g1 ! Don't move this insn +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,%3,%%g1 +- mulscc %%g1,0,%%g1 +- add %%g1,%%g2,%0 +- rd %%y,%1" \ ++ __asm__ ("! Inlined umul_ppmm\n" ++" wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n" ++" sra %3,31,%%g2 ! Don't move this insn\n" ++" and %2,%%g2,%%g2 ! Don't move this insn\n" ++" andcc %%g0,0,%%g1 ! Don't move this insn\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,%3,%%g1\n" ++" mulscc %%g1,0,%%g1\n" ++" add %%g1,%%g2,%0\n" ++" rd %%y,%1" \ + : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v) \ + : "%g1", "%g2" __AND_CLOBBER_CC) + #define UMUL_TIME 39 /* 39 instructions */ |