summaryrefslogtreecommitdiff
path: root/lang/drscheme
diff options
context:
space:
mode:
authorjschauma <jschauma>2004-01-13 23:40:24 +0000
committerjschauma <jschauma>2004-01-13 23:40:24 +0000
commite761c26f134aa8a1d803907e40ebb6ae2c7fbdad (patch)
tree94fadcdf5b10a05d23e9f3bd0451b258c9059b9d /lang/drscheme
parent451051bb5571dd02e9fb40bc3f9b8ea66ae6f4a6 (diff)
downloadpkgsrc-e761c26f134aa8a1d803907e40ebb6ae2c7fbdad.tar.gz
Make this compile with gcc-3.3 by replacing multi-line strings
"foo string" in inline asm with "foo\n" "string" (Noted in agc's latest bulk build.)
Diffstat (limited to 'lang/drscheme')
-rw-r--r--lang/drscheme/distinfo3
-rw-r--r--lang/drscheme/patches/patch-aq346
2 files changed, 348 insertions, 1 deletions
diff --git a/lang/drscheme/distinfo b/lang/drscheme/distinfo
index afd7824070b..0016d80faf4 100644
--- a/lang/drscheme/distinfo
+++ b/lang/drscheme/distinfo
@@ -1,4 +1,4 @@
-$NetBSD: distinfo,v 1.8 2003/05/03 20:57:03 jmc Exp $
+$NetBSD: distinfo,v 1.9 2004/01/13 23:40:24 jschauma Exp $
SHA1 (drscheme/202/plt.src.x.tar.gz) = bb06b6e3f8214db71ebb12e95e744f996622ee06
Size (drscheme/202/plt.src.x.tar.gz) = 6133563 bytes
@@ -18,3 +18,4 @@ SHA1 (patch-am) = afeb2c469b55f1b9fdadf4e064928b59bf28d794
SHA1 (patch-an) = bca5a79edd48a2670d02496b6da3799c13c0aed6
SHA1 (patch-ao) = 667593d28a695fd7caee210deb8ee06e72019c40
SHA1 (patch-ap) = 2395a864a40af34f295e0ed782979d90d75f1ec4
+SHA1 (patch-aq) = 058affb4c3f840f27ce88aba0baaeacba2eb8282
diff --git a/lang/drscheme/patches/patch-aq b/lang/drscheme/patches/patch-aq
new file mode 100644
index 00000000000..cae92ec4a6f
--- /dev/null
+++ b/lang/drscheme/patches/patch-aq
@@ -0,0 +1,346 @@
+$NetBSD: patch-aq,v 1.1 2004/01/13 23:40:25 jschauma Exp $
+
+--- ./mzscheme/src/gmp/gmplonglong.h.orig 2001-06-26 12:46:25.000000000 -0400
++++ ./mzscheme/src/gmp/gmplonglong.h 2004-01-13 18:20:47.000000000 -0500
+@@ -231,19 +231,19 @@
+ #define UMUL_TIME 5
+ #else
+ #define umul_ppmm(xh, xl, a, b) \
+- __asm__ ("%@ Inlined umul_ppmm
+- mov %|r0, %2, lsr #16
+- mov %|r2, %3, lsr #16
+- bic %|r1, %2, %|r0, lsl #16
+- bic %|r2, %3, %|r2, lsl #16
+- mul %1, %|r1, %|r2
+- mul %|r2, %|r0, %|r2
+- mul %|r1, %0, %|r1
+- mul %0, %|r0, %0
+- adds %|r1, %|r2, %|r1
+- addcs %0, %0, #65536
+- adds %1, %1, %|r1, lsl #16
+- adc %0, %0, %|r1, lsr #16" \
++ __asm__ ("%@ Inlined umul_ppmm\n"
++" mov %|r0, %2, lsr #16\n"
++" mov %|r2, %3, lsr #16\n"
++" bic %|r1, %2, %|r0, lsl #16\n"
++" bic %|r2, %3, %|r2, lsl #16\n"
++" mul %1, %|r1, %|r2\n"
++" mul %|r2, %|r0, %|r2\n"
++" mul %|r1, %0, %|r1\n"
++" mul %0, %|r0, %0\n"
++" adds %|r1, %|r2, %|r1\n"
++" addcs %0, %0, #65536\n"
++" adds %1, %1, %|r1, lsl #16\n"
++" adc %0, %0, %|r1, lsr #16" \
+ : "=&r" (xh), "=r" (xl) \
+ : "r" (a), "r" (b) \
+ : "r0", "r1", "r2")
+@@ -360,22 +360,22 @@
+ do { \
+ USItype __tmp; \
+ __asm__ ( \
+- "ldi 1,%0
+- extru,= %1,15,16,%%r0 ; Bits 31..16 zero?
+- extru,tr %1,15,16,%1 ; No. Shift down, skip add.
+- ldo 16(%0),%0 ; Yes. Perform add.
+- extru,= %1,23,8,%%r0 ; Bits 15..8 zero?
+- extru,tr %1,23,8,%1 ; No. Shift down, skip add.
+- ldo 8(%0),%0 ; Yes. Perform add.
+- extru,= %1,27,4,%%r0 ; Bits 7..4 zero?
+- extru,tr %1,27,4,%1 ; No. Shift down, skip add.
+- ldo 4(%0),%0 ; Yes. Perform add.
+- extru,= %1,29,2,%%r0 ; Bits 3..2 zero?
+- extru,tr %1,29,2,%1 ; No. Shift down, skip add.
+- ldo 2(%0),%0 ; Yes. Perform add.
+- extru %1,30,1,%1 ; Extract bit 1.
+- sub %0,%1,%0 ; Subtract it.
+- " : "=r" (count), "=r" (__tmp) : "1" (x)); \
++ "ldi 1,%0\n"
++" extru,= %1,15,16,%%r0 ; Bits 31..16 zero?\n"
++" extru,tr %1,15,16,%1 ; No. Shift down, skip add.\n"
++" ldo 16(%0),%0 ; Yes. Perform add.\n"
++" extru,= %1,23,8,%%r0 ; Bits 15..8 zero?\n"
++" extru,tr %1,23,8,%1 ; No. Shift down, skip add.\n"
++" ldo 8(%0),%0 ; Yes. Perform add.\n"
++" extru,= %1,27,4,%%r0 ; Bits 7..4 zero?\n"
++" extru,tr %1,27,4,%1 ; No. Shift down, skip add.\n"
++" ldo 4(%0),%0 ; Yes. Perform add.\n"
++" extru,= %1,29,2,%%r0 ; Bits 3..2 zero?\n"
++" extru,tr %1,29,2,%1 ; No. Shift down, skip add.\n"
++" ldo 2(%0),%0 ; Yes. Perform add.\n"
++" extru %1,30,1,%1 ; Extract bit 1.\n"
++" sub %0,%1,%0 ; Subtract it.\n"
++ : "=r" (count), "=r" (__tmp) : "1" (x)); \
+ } while (0)
+ #endif /* hppa */
+
+@@ -529,28 +529,28 @@
+ #else /* for other 68k family members use 16x16->32 multiplication */
+ #define umul_ppmm(xh, xl, a, b) \
+ do { USItype __umul_tmp1, __umul_tmp2; \
+- __asm__ ("| Inlined umul_ppmm
+- move%.l %5,%3
+- move%.l %2,%0
+- move%.w %3,%1
+- swap %3
+- swap %0
+- mulu%.w %2,%1
+- mulu%.w %3,%0
+- mulu%.w %2,%3
+- swap %2
+- mulu%.w %5,%2
+- add%.l %3,%2
+- jcc 1f
+- add%.l %#0x10000,%0
+-1: move%.l %2,%3
+- clr%.w %2
+- swap %2
+- swap %3
+- clr%.w %3
+- add%.l %3,%1
+- addx%.l %2,%0
+- | End inlined umul_ppmm" \
++ __asm__ ("| Inlined umul_ppmm\n"
++" move%.l %5,%3\n"
++" move%.l %2,%0\n"
++" move%.w %3,%1\n"
++" swap %3\n"
++" swap %0\n"
++" mulu%.w %2,%1\n"
++" mulu%.w %3,%0\n"
++" mulu%.w %2,%3\n"
++" swap %2\n"
++" mulu%.w %5,%2\n"
++" add%.l %3,%2\n"
++" jcc 1f\n"
++" add%.l %#0x10000,%0\n"
++"1: move%.l %2,%3\n"
++" clr%.w %2\n"
++" swap %2\n"
++" swap %3\n"
++" clr%.w %3\n"
++" add%.l %3,%1\n"
++" addx%.l %2,%0\n"
++" | End inlined umul_ppmm" \
+ : "=&d" ((USItype)(xh)), "=&d" ((USItype)(xl)), \
+ "=d" (__umul_tmp1), "=&d" (__umul_tmp2) \
+ : "%2" ((USItype)(a)), "d" ((USItype)(b))); \
+@@ -827,26 +827,26 @@
+ "1" ((USItype)(al)), "r" ((USItype)(bl)))
+ #define smul_ppmm(ph, pl, m0, m1) \
+ __asm__ ( \
+- "s r2,r2
+- mts r10,%2
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- m r2,%3
+- cas %0,r2,r0
+- mfs r10,%1" \
++ "s r2,r2\n"
++" mts r10,%2\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" m r2,%3\n"
++" cas %0,r2,r0\n"
++" mfs r10,%1" \
+ : "=r" ((USItype)(ph)), "=r" ((USItype)(pl)) \
+ : "%r" ((USItype)(m0)), "r" ((USItype)(m1)) \
+ : "r2"); \
+@@ -936,45 +936,45 @@
+ __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
+ #define UMUL_TIME 5
+ #define udiv_qrnnd(q, r, n1, n0, d) \
+- __asm__ ("! Inlined udiv_qrnnd
+- wr %%g0,%2,%%y ! Not a delayed write for sparclite
+- tst %%g0
+- divscc %3,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%%g1
+- divscc %%g1,%4,%0
+- rd %%y,%1
+- bl,a 1f
+- add %1,%4,%1
+-1: ! End of inline udiv_qrnnd" \
++ __asm__ ("! Inlined udiv_qrnnd\n"
++" wr %%g0,%2,%%y ! Not a delayed write for sparclite\n"
++" tst %%g0\n"
++" divscc %3,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%%g1\n"
++" divscc %%g1,%4,%0\n"
++" rd %%y,%1\n"
++" bl,a 1f\n"
++" add %1,%4,%1\n"
++"1: ! End of inline udiv_qrnnd" \
+ : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d)
+ : "%g1" __AND_CLOBBER_CC)
+ #define UDIV_TIME 37
+@@ -989,46 +989,46 @@
+ /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd. */
+ #ifndef umul_ppmm
+ #define umul_ppmm(w1, w0, u, v) \
+- __asm__ ("! Inlined umul_ppmm
+- wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr
+- sra %3,31,%%g2 ! Don't move this insn
+- and %2,%%g2,%%g2 ! Don't move this insn
+- andcc %%g0,0,%%g1 ! Don't move this insn
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,%3,%%g1
+- mulscc %%g1,0,%%g1
+- add %%g1,%%g2,%0
+- rd %%y,%1" \
++ __asm__ ("! Inlined umul_ppmm\n"
++" wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n"
++" sra %3,31,%%g2 ! Don't move this insn\n"
++" and %2,%%g2,%%g2 ! Don't move this insn\n"
++" andcc %%g0,0,%%g1 ! Don't move this insn\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,%3,%%g1\n"
++" mulscc %%g1,0,%%g1\n"
++" add %%g1,%%g2,%0\n"
++" rd %%y,%1" \
+ : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v) \
+ : "%g1", "%g2" __AND_CLOBBER_CC)
+ #define UMUL_TIME 39 /* 39 instructions */