.explicit
.text
-.asciz "ia64.S, Version 1.0"
-.asciz "IA-64 ISA artwork by Andy Polyakov <appro@fy.chalmers.se>"
+.ident "ia64.S, Version 2.1"
+.ident "IA-64 ISA artwork by Andy Polyakov <appro@fy.chalmers.se>"
//
// ====================================================================
// disclaimed.
// ====================================================================
//
+// Version 2.x is Itanium2 re-tune. Few words about how Itanum2 is
+// different from Itanium to this module viewpoint. Most notably, is it
+// "wider" than Itanium? Can you experience loop scalability as
+// discussed in commentary sections? Not really:-( Itanium2 has 6
+// integer ALU ports, i.e. it's 2 ports wider, but it's not enough to
+// spin twice as fast, as I need 8 IALU ports. Amount of floating point
+// ports is the same, i.e. 2, while I need 4. In other words, to this
+// module Itanium2 remains effectively as "wide" as Itanium. Yet it's
+// essentially different in respect to this module, and a re-tune was
+// required. Well, because some intruction latencies has changed. Most
+// noticeably those intensively used:
+//
+// Itanium Itanium2
+// ldf8 9 6 L2 hit
+// ld8 2 1 L1 hit
+// getf 2 5
+// xma[->getf] 7[+1] 4[+0]
+// add[->st8] 1[+1] 1[+0]
+//
+// What does it mean? You might ratiocinate that the original code
+// should run just faster... Because sum of latencies is smaller...
+// Wrong! Note that getf latency increased. This means that if a loop is
+// scheduled for lower latency (as they were), then it will suffer from
+// stall condition and the code will therefore turn anti-scalable, e.g.
+// original bn_mul_words spun at 5*n or 2.5 times slower than expected
+// on Itanium2! What to do? Reschedule loops for Itanium2? But then
+// Itanium would exhibit anti-scalability. So I've chosen to reschedule
+// for worst latency for every instruction aiming for best *all-round*
+// performance.
// Q. How much faster does it get?
// A. Here is the output from 'openssl speed rsa dsa' for vanilla
// -Drum=nop.m in command line.
//
+#if defined(_HPUX_SOURCE) && !defined(_LP64)
+#define ADDP addp4
+#else
+#define ADDP add
+#endif
+
#if 1
//
// bn_[add|sub]_words routines.
.global bn_add_words#
.proc bn_add_words#
.align 64
-.space 32 // makes the loop body aligned at 64-byte boundary
+.skip 32 // makes the loop body aligned at 64-byte boundary
bn_add_words:
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,4,12,0,16
+{ .mii; alloc r2=ar.pfs,4,12,0,16
cmp4.le p6,p0=r35,r0 };;
-{ .mfb mov r8=r0 // return value
+{ .mfb; mov r8=r0 // return value
(p6) br.ret.spnt.many b0 };;
.save ar.lc,r3
-{ .mib sub r10=r35,r0,1
+{ .mib; sub r10=r35,r0,1
mov r3=ar.lc
brp.loop.imp .L_bn_add_words_ctop,.L_bn_add_words_cend-16
}
.body
-{ .mib mov r14=r32 // rp
+{ .mib; ADDP r14=0,r32 // rp
mov r9=pr };;
-{ .mii mov r15=r33 // ap
+{ .mii; ADDP r15=0,r33 // ap
mov ar.lc=r10
mov ar.ec=6 }
-{ .mib mov r16=r34 // bp
- mov pr.rot=1<<16 }
+{ .mib; ADDP r16=0,r34 // bp
+ mov pr.rot=1<<16 };;
-.L_bn_add_words_ctop: ;;
-{ .mii (p16) ld8 r32=[r16],8 // b=*(bp++)
+.L_bn_add_words_ctop:
+{ .mii; (p16) ld8 r32=[r16],8 // b=*(bp++)
(p18) add r39=r37,r34
(p19) cmp.ltu.unc p56,p0=r40,r38 }
-{ .mfb (p0) nop.m 0x0
+{ .mfb; (p0) nop.m 0x0
(p0) nop.f 0x0
(p0) nop.b 0x0 }
-{ .mii (p16) ld8 r35=[r15],8 // a=*(ap++)
+{ .mii; (p16) ld8 r35=[r15],8 // a=*(ap++)
(p58) cmp.eq.or p57,p0=-1,r41 // (p20)
(p58) add r41=1,r41 } // (p20)
-{ .mfb (p21) st8 [r14]=r42,8 // *(rp++)=r
+{ .mfb; (p21) st8 [r14]=r42,8 // *(rp++)=r
(p0) nop.f 0x0
br.ctop.sptk .L_bn_add_words_ctop };;
.L_bn_add_words_cend:
-{ .mii
+{ .mii;
(p59) add r8=1,r8 // return value
- mov pr=r9,-1
+ mov pr=r9,0x1ffff
mov ar.lc=r3 }
-{ .mbb nop.b 0x0
+{ .mbb; nop.b 0x0
br.ret.sptk.many b0 };;
.endp bn_add_words#
.global bn_sub_words#
.proc bn_sub_words#
.align 64
-.space 32 // makes the loop body aligned at 64-byte boundary
+.skip 32 // makes the loop body aligned at 64-byte boundary
bn_sub_words:
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,4,12,0,16
+{ .mii; alloc r2=ar.pfs,4,12,0,16
cmp4.le p6,p0=r35,r0 };;
-{ .mfb mov r8=r0 // return value
+{ .mfb; mov r8=r0 // return value
(p6) br.ret.spnt.many b0 };;
.save ar.lc,r3
-{ .mib sub r10=r35,r0,1
+{ .mib; sub r10=r35,r0,1
mov r3=ar.lc
brp.loop.imp .L_bn_sub_words_ctop,.L_bn_sub_words_cend-16
}
.body
-{ .mib mov r14=r32 // rp
+{ .mib; ADDP r14=0,r32 // rp
mov r9=pr };;
-{ .mii mov r15=r33 // ap
+{ .mii; ADDP r15=0,r33 // ap
mov ar.lc=r10
mov ar.ec=6 }
-{ .mib mov r16=r34 // bp
- mov pr.rot=1<<16 }
+{ .mib; ADDP r16=0,r34 // bp
+ mov pr.rot=1<<16 };;
-.L_bn_sub_words_ctop: ;;
-{ .mii (p16) ld8 r32=[r16],8 // b=*(bp++)
+.L_bn_sub_words_ctop:
+{ .mii; (p16) ld8 r32=[r16],8 // b=*(bp++)
(p18) sub r39=r37,r34
(p19) cmp.gtu.unc p56,p0=r40,r38 }
-{ .mfb (p0) nop.m 0x0
+{ .mfb; (p0) nop.m 0x0
(p0) nop.f 0x0
(p0) nop.b 0x0 }
-{ .mii (p16) ld8 r35=[r15],8 // a=*(ap++)
+{ .mii; (p16) ld8 r35=[r15],8 // a=*(ap++)
(p58) cmp.eq.or p57,p0=0,r41 // (p20)
(p58) add r41=-1,r41 } // (p20)
-{ .mbb (p21) st8 [r14]=r42,8 // *(rp++)=r
+{ .mbb; (p21) st8 [r14]=r42,8 // *(rp++)=r
(p0) nop.b 0x0
br.ctop.sptk .L_bn_sub_words_ctop };;
.L_bn_sub_words_cend:
-{ .mii
+{ .mii;
(p59) add r8=1,r8 // return value
- mov pr=r9,-1
+ mov pr=r9,0x1ffff
mov ar.lc=r3 }
-{ .mbb nop.b 0x0
+{ .mbb; nop.b 0x0
br.ret.sptk.many b0 };;
.endp bn_sub_words#
#endif
.global bn_mul_words#
.proc bn_mul_words#
.align 64
-.space 32 // makes the loop body aligned at 64-byte boundary
+.skip 32 // makes the loop body aligned at 64-byte boundary
bn_mul_words:
.prologue
.fframe 0
.save ar.pfs,r2
#ifdef XMA_TEMPTATION
-{ .mfi alloc r2=ar.pfs,4,0,0,0 };;
+{ .mfi; alloc r2=ar.pfs,4,0,0,0 };;
#else
-{ .mfi alloc r2=ar.pfs,4,4,0,8 };;
+{ .mfi; alloc r2=ar.pfs,4,12,0,16 };;
#endif
-{ .mib mov r8=r0 // return value
+{ .mib; mov r8=r0 // return value
cmp4.le p6,p0=r34,r0
(p6) br.ret.spnt.many b0 };;
.save ar.lc,r3
-{ .mii sub r10=r34,r0,1
+{ .mii; sub r10=r34,r0,1
mov r3=ar.lc
mov r9=pr };;
.body
-{ .mib setf.sig f8=r35 // w
- mov pr.rot=0x400001<<16
- // ------^----- serves as (p48) at first (p26)
+{ .mib; setf.sig f8=r35 // w
+ mov pr.rot=0x800001<<16
+ // ------^----- serves as (p50) at first (p27)
brp.loop.imp .L_bn_mul_words_ctop,.L_bn_mul_words_cend-16
}
#ifndef XMA_TEMPTATION
-{ .mii mov r14=r32 // rp
- mov r15=r33 // ap
+{ .mmi; ADDP r14=0,r32 // rp
+ ADDP r15=0,r33 // ap
mov ar.lc=r10 }
-{ .mii mov r39=0 // serves as r33 at first (p26)
- mov ar.ec=12 }
+{ .mmi; mov r40=0 // serves as r35 at first (p27)
+ mov ar.ec=13 };;
-// This loop spins in 2*(n+11) ticks. It's scheduled for data in L2
-// cache (i.e. 9 ticks away) as floating point load/store instructions
+// This loop spins in 2*(n+12) ticks. It's scheduled for data in Itanium
+// L2 cache (i.e. 9 ticks away) as floating point load/store instructions
// bypass L1 cache and L2 latency is actually best-case scenario for
-// ldf8. The loop is not scalable and shall run in 2*(n+11) even on
-// "wider" IA-64 implementations. It's a trade-off here. n+22 loop
+// ldf8. The loop is not scalable and shall run in 2*(n+12) even on
+// "wider" IA-64 implementations. It's a trade-off here. n+24 loop
// would give us ~5% in *overall* performance improvement on "wider"
// IA-64, but would hurt Itanium for about same because of longer
// epilogue. As it's a matter of few percents in either case I've
// chosen to trade the scalability for development time (you can see
// this very instruction sequence in bn_mul_add_words loop which in
// turn is scalable).
-.L_bn_mul_words_ctop: ;;
-{ .mfi (p25) getf.sig r36=f49 // low
- (p21) xmpy.lu f45=f37,f8
- (p27) cmp.ltu p52,p48=r39,r38 }
-{ .mfi (p16) ldf8 f32=[r15],8
- (p21) xmpy.hu f38=f37,f8
+.L_bn_mul_words_ctop:
+{ .mfi; (p25) getf.sig r36=f52 // low
+ (p21) xmpy.lu f48=f37,f8
+ (p28) cmp.ltu p54,p50=r41,r39 }
+{ .mfi; (p16) ldf8 f32=[r15],8
+ (p21) xmpy.hu f40=f37,f8
(p0) nop.i 0x0 };;
-{ .mii (p26) getf.sig r32=f43 // high
- .pred.rel "mutex",p48,p52
- (p48) add r38=r37,r33 // (p26)
- (p52) add r38=r37,r33,1 } // (p26)
-{ .mfb (p27) st8 [r14]=r39,8
+{ .mii; (p25) getf.sig r32=f44 // high
+ .pred.rel "mutex",p50,p54
+ (p50) add r40=r38,r35 // (p27)
+ (p54) add r40=r38,r35,1 } // (p27)
+{ .mfb; (p28) st8 [r14]=r41,8
(p0) nop.f 0x0
br.ctop.sptk .L_bn_mul_words_ctop };;
.L_bn_mul_words_cend:
-{ .mii nop.m 0x0
-.pred.rel "mutex",p49,p53
-(p49) add r8=r34,r0
-(p53) add r8=r34,r0,1 }
-{ .mfb nop.m 0x0
+{ .mii; nop.m 0x0
+.pred.rel "mutex",p51,p55
+(p51) add r8=r36,r0
+(p55) add r8=r36,r0,1 }
+{ .mfb; nop.m 0x0
nop.f 0x0
nop.b 0x0 }
setf.sig f37=r0 // serves as carry at (p18) tick
mov ar.lc=r10
- mov ar.ec=5
+ mov ar.ec=5;;
// Most of you examining this code very likely wonder why in the name
// of Intel the following loop is commented out? Indeed, it looks so
// by moving the dependency to one-tick latent interger ALU. Note that
// "distance" between ldf8 and xma is not latency of ldf8, but the
// *difference* between xma and ldf8 latencies.
-.L_bn_mul_words_ctop: ;;
-{ .mfi (p16) ldf8 f32=[r33],8
+.L_bn_mul_words_ctop:
+{ .mfi; (p16) ldf8 f32=[r33],8
(p18) xma.hu f38=f34,f8,f39 }
-{ .mfb (p20) stf8 [r32]=f37,8
+{ .mfb; (p20) stf8 [r32]=f37,8
(p18) xma.lu f35=f34,f8,f39
br.ctop.sptk .L_bn_mul_words_ctop };;
.L_bn_mul_words_cend:
#endif // XMA_TEMPTATION
-{ .mii nop.m 0x0
- mov pr=r9,-1
+{ .mii; nop.m 0x0
+ mov pr=r9,0x1ffff
mov ar.lc=r3 }
-{ .mfb rum 1<<5 // clear um.mfh
+{ .mfb; rum 1<<5 // clear um.mfh
nop.f 0x0
br.ret.sptk.many b0 };;
.endp bn_mul_words#
.global bn_mul_add_words#
.proc bn_mul_add_words#
.align 64
-//.space 0 // makes the loop split at 64-byte boundary
+.skip 48 // makes the loop body aligned at 64-byte boundary
bn_mul_add_words:
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,4,12,0,16
- cmp4.le p6,p0=r34,r0 };;
-{ .mfb mov r8=r0 // return value
-(p6) br.ret.spnt.many b0 };;
-
.save ar.lc,r3
-{ .mii sub r10=r34,r0,1
- mov r3=ar.lc
- mov r9=pr };;
+ .save pr,r9
+{ .mmi; alloc r2=ar.pfs,4,4,0,8
+ cmp4.le p6,p0=r34,r0
+ mov r3=ar.lc };;
+{ .mib; mov r8=r0 // return value
+ sub r10=r34,r0,1
+(p6) br.ret.spnt.many b0 };;
.body
-{ .mib setf.sig f8=r35 // w
- mov pr.rot=0x400001<<16
- // ------^----- serves as (p48) at first (p26)
+{ .mib; setf.sig f8=r35 // w
+ mov r9=pr
brp.loop.imp .L_bn_mul_add_words_ctop,.L_bn_mul_add_words_cend-16
}
-{ .mii mov r14=r32 // rp
- mov r15=r33 // ap
+{ .mmi; ADDP r14=0,r32 // rp
+ ADDP r15=0,r33 // ap
mov ar.lc=r10 }
-{ .mii mov r39=0 // serves as r33 at first (p26)
- mov r18=r32 // rp copy
- mov ar.ec=14 }
-
-// This loop spins in 3*(n+13) ticks on Itanium and should spin in
-// 2*(n+13) on "wider" IA-64 implementations (to be verified with new
-// µ-architecture manuals as they become available). As usual it's
-// possible to compress the epilogue, down to 10 in this case, at the
-// cost of scalability. Compressed (and therefore non-scalable) loop
-// running at 3*(n+10) would buy you ~10% on Itanium but take ~35%
-// from "wider" IA-64 so let it be scalable! Special attention was
-// paid for having the loop body split at 64-byte boundary. ld8 is
-// scheduled for L1 cache as the data is more than likely there.
-// Indeed, bn_mul_words has put it there a moment ago:-)
-.L_bn_mul_add_words_ctop: ;;
-{ .mfi (p25) getf.sig r36=f49 // low
- (p21) xmpy.lu f45=f37,f8
- (p27) cmp.ltu p52,p48=r39,r38 }
-{ .mfi (p16) ldf8 f32=[r15],8
- (p21) xmpy.hu f38=f37,f8
- (p27) add r43=r43,r39 };;
-{ .mii (p26) getf.sig r32=f43 // high
- .pred.rel "mutex",p48,p52
- (p48) add r38=r37,r33 // (p26)
- (p52) add r38=r37,r33,1 } // (p26)
-{ .mfb (p27) cmp.ltu.unc p56,p0=r43,r39
- (p0) nop.f 0x0
- (p0) nop.b 0x0 }
-{ .mii (p26) ld8 r42=[r18],8
- (p58) cmp.eq.or p57,p0=-1,r44
- (p58) add r44=1,r44 }
-{ .mfb (p29) st8 [r14]=r45,8
- (p0) nop.f 0x0
+{ .mii; ADDP r16=0,r32 // rp copy
+ mov pr.rot=0x2001<<16
+ // ------^----- serves as (p40) at first (p27)
+ mov ar.ec=11 };;
+
+// This loop spins in 3*(n+10) ticks on Itanium and in 2*(n+10) on
+// Itanium 2. Yes, unlike previous versions it scales:-) Previous
+// version was peforming *all* additions in IALU and was starving
+// for those even on Itanium 2. In this version one addition is
+// moved to FPU and is folded with multiplication. This is at cost
+// of propogating the result from previous call to this subroutine
+// to L2 cache... In other words negligible even for shorter keys.
+// *Overall* performance improvement [over previous version] varies
+// from 11 to 22 percent depending on key length.
+.L_bn_mul_add_words_ctop:
+.pred.rel "mutex",p40,p42
+{ .mfi; (p23) getf.sig r36=f45 // low
+ (p20) xma.lu f42=f36,f8,f50 // low
+ (p40) add r39=r39,r35 } // (p27)
+{ .mfi; (p16) ldf8 f32=[r15],8 // *(ap++)
+ (p20) xma.hu f36=f36,f8,f50 // high
+ (p42) add r39=r39,r35,1 };; // (p27)
+{ .mmi; (p24) getf.sig r32=f40 // high
+ (p16) ldf8 f46=[r16],8 // *(rp1++)
+ (p40) cmp.ltu p41,p39=r39,r35 } // (p27)
+{ .mib; (p26) st8 [r14]=r39,8 // *(rp2++)
+ (p42) cmp.leu p41,p39=r39,r35 // (p27)
br.ctop.sptk .L_bn_mul_add_words_ctop};;
.L_bn_mul_add_words_cend:
-{ .mii nop.m 0x0
-.pred.rel "mutex",p51,p55
-(p51) add r8=r36,r0
-(p55) add r8=r36,r0,1 }
-{ .mfb nop.m 0x0
- nop.f 0x0
- nop.b 0x0 };;
-{ .mii
-(p59) add r8=1,r8
- mov pr=r9,-1
- mov ar.lc=r3 }
-{ .mfb rum 1<<5 // clear um.mfh
- nop.f 0x0
+{ .mmi; .pred.rel "mutex",p40,p42
+(p40) add r8=r35,r0
+(p42) add r8=r35,r0,1
+ mov pr=r9,0x1ffff }
+{ .mib; rum 1<<5 // clear um.mfh
+ mov ar.lc=r3
br.ret.sptk.many b0 };;
.endp bn_mul_add_words#
#endif
.global bn_sqr_words#
.proc bn_sqr_words#
.align 64
-.space 32 // makes the loop body aligned at 64-byte boundary
+.skip 32 // makes the loop body aligned at 64-byte boundary
bn_sqr_words:
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,3,0,0,0
+{ .mii; alloc r2=ar.pfs,3,0,0,0
sxt4 r34=r34 };;
-{ .mii cmp.le p6,p0=r34,r0
+{ .mii; cmp.le p6,p0=r34,r0
mov r8=r0 } // return value
-{ .mfb nop.f 0x0
+{ .mfb; ADDP r32=0,r32
+ nop.f 0x0
(p6) br.ret.spnt.many b0 };;
.save ar.lc,r3
-{ .mii sub r10=r34,r0,1
+{ .mii; sub r10=r34,r0,1
mov r3=ar.lc
mov r9=pr };;
.body
-{ .mib
+{ .mib; ADDP r33=0,r33
mov pr.rot=1<<16
brp.loop.imp .L_bn_sqr_words_ctop,.L_bn_sqr_words_cend-16
}
-{ .mii add r34=8,r32
+{ .mii; add r34=8,r32
mov ar.lc=r10
- mov ar.ec=18 }
+ mov ar.ec=18 };;
// 2*(n+17) on Itanium, (n+17) on "wider" IA-64 implementations. It's
// possible to compress the epilogue (I'm getting tired to write this
// benchmark program is profiled. I.e. if perfomance gain on Itanium
// will appear larger than loss on "wider" IA-64, then the loop should
// be explicitely split and the epilogue compressed.
-.L_bn_sqr_words_ctop: ;;
-{ .mfi (p16) ldf8 f32=[r33],8
+.L_bn_sqr_words_ctop:
+{ .mfi; (p16) ldf8 f32=[r33],8
(p25) xmpy.lu f42=f41,f41
(p0) nop.i 0x0 }
-{ .mib (p33) stf8 [r32]=f50,16
+{ .mib; (p33) stf8 [r32]=f50,16
(p0) nop.i 0x0
(p0) nop.b 0x0 }
-{ .mfi (p0) nop.m 0x0
+{ .mfi; (p0) nop.m 0x0
(p25) xmpy.hu f52=f41,f41
(p0) nop.i 0x0 }
-{ .mib (p33) stf8 [r34]=f60,16
+{ .mib; (p33) stf8 [r34]=f60,16
(p0) nop.i 0x0
br.ctop.sptk .L_bn_sqr_words_ctop };;
.L_bn_sqr_words_cend:
-{ .mii nop.m 0x0
- mov pr=r9,-1
+{ .mii; nop.m 0x0
+ mov pr=r9,0x1ffff
mov ar.lc=r3 }
-{ .mfb rum 1<<5 // clear um.mfh
+{ .mfb; rum 1<<5 // clear um.mfh
nop.f 0x0
br.ret.sptk.many b0 };;
.endp bn_sqr_words#
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,2,1,0,0
+#if defined(_HPUX_SOURCE) && !defined(_LP64)
+{ .mii; alloc r2=ar.pfs,2,1,0,0
+ addp4 r33=0,r33
+ addp4 r32=0,r32 };;
+{ .mii;
+#else
+{ .mii; alloc r2=ar.pfs,2,1,0,0
+#endif
mov r34=r33
add r14=8,r33 };;
.body
-{ .mii add r17=8,r34
+{ .mii; add r17=8,r34
add r15=16,r33
add r18=16,r34 }
-{ .mfb add r16=24,r33
+{ .mfb; add r16=24,r33
br .L_cheat_entry_point8 };;
+.endp bn_sqr_comba8#
#endif
#if 1
// clause in Itanium µ-architecture manual? Comments are welcomed and
// highly appreciated.
//
+// On Itanium 2 it takes ~190 ticks. This is because of stalls on
+// result from getf.sig. I do nothing about it at this point for
+// reasons depicted below.
+//
// However! It should be noted that even 160 ticks is darn good result
// as it's over 10 (yes, ten, spelled as t-e-n) times faster than the
// C version (compiled with gcc with inline assembler). I really
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,3,0,0,0
+#if defined(_HPUX_SOURCE) && !defined(_LP64)
+{ .mii; alloc r2=ar.pfs,3,0,0,0
+ addp4 r33=0,r33
+ addp4 r34=0,r34 };;
+{ .mii; addp4 r32=0,r32
+#else
+{ .mii; alloc r2=ar.pfs,3,0,0,0
+#endif
add r14=8,r33
add r17=8,r34 }
.body
-{ .mii add r15=16,r33
+{ .mii; add r15=16,r33
add r18=16,r34
- add r16=24,r33 };;
+ add r16=24,r33 }
.L_cheat_entry_point8:
-{ .mmi add r19=24,r34
+{ .mmi; add r19=24,r34
- ldf8 f32=[r33],32 }
+ ldf8 f32=[r33],32 };;
-{ .mmi ldf8 f120=[r34],32
- ldf8 f121=[r17],32 };;
-{ .mmi ldf8 f122=[r18],32
- ldf8 f123=[r19],32 }
-{ .mmi ldf8 f124=[r34]
+{ .mmi; ldf8 f120=[r34],32
+ ldf8 f121=[r17],32 }
+{ .mmi; ldf8 f122=[r18],32
+ ldf8 f123=[r19],32 };;
+{ .mmi; ldf8 f124=[r34]
ldf8 f125=[r17] }
-{ .mmi ldf8 f126=[r18]
+{ .mmi; ldf8 f126=[r18]
ldf8 f127=[r19] }
-{ .mmi ldf8 f33=[r14],32
+{ .mmi; ldf8 f33=[r14],32
ldf8 f34=[r15],32 }
-{ .mmi ldf8 f35=[r16],32
+{ .mmi; ldf8 f35=[r16],32;;
ldf8 f36=[r33] }
-{ .mmi ldf8 f37=[r14]
+{ .mmi; ldf8 f37=[r14]
ldf8 f38=[r15] }
-{ .mfi ldf8 f39=[r16]
+{ .mfi; ldf8 f39=[r16]
// -------\ Entering multiplier's heaven /-------
// ------------\ /------------
// -----------------\ /-----------------
// ----------------------\/----------------------
xma.hu f41=f32,f120,f0 }
-{ .mfi xma.lu f40=f32,f120,f0 };; // (*)
-{ .mfi xma.hu f51=f32,f121,f0 }
-{ .mfi xma.lu f50=f32,f121,f0 };;
-{ .mfi xma.hu f61=f32,f122,f0 }
-{ .mfi xma.lu f60=f32,f122,f0 };;
-{ .mfi xma.hu f71=f32,f123,f0 }
-{ .mfi xma.lu f70=f32,f123,f0 };;
-{ .mfi xma.hu f81=f32,f124,f0 }
-{ .mfi xma.lu f80=f32,f124,f0 };;
-{ .mfi xma.hu f91=f32,f125,f0 }
-{ .mfi xma.lu f90=f32,f125,f0 };;
-{ .mfi xma.hu f101=f32,f126,f0 }
-{ .mfi xma.lu f100=f32,f126,f0 };;
-{ .mfi xma.hu f111=f32,f127,f0 }
-{ .mfi xma.lu f110=f32,f127,f0 };;;;
+{ .mfi; xma.lu f40=f32,f120,f0 };; // (*)
+{ .mfi; xma.hu f51=f32,f121,f0 }
+{ .mfi; xma.lu f50=f32,f121,f0 };;
+{ .mfi; xma.hu f61=f32,f122,f0 }
+{ .mfi; xma.lu f60=f32,f122,f0 };;
+{ .mfi; xma.hu f71=f32,f123,f0 }
+{ .mfi; xma.lu f70=f32,f123,f0 };;
+{ .mfi; xma.hu f81=f32,f124,f0 }
+{ .mfi; xma.lu f80=f32,f124,f0 };;
+{ .mfi; xma.hu f91=f32,f125,f0 }
+{ .mfi; xma.lu f90=f32,f125,f0 };;
+{ .mfi; xma.hu f101=f32,f126,f0 }
+{ .mfi; xma.lu f100=f32,f126,f0 };;
+{ .mfi; xma.hu f111=f32,f127,f0 }
+{ .mfi; xma.lu f110=f32,f127,f0 };;//
// (*) You can argue that splitting at every second bundle would
// prevent "wider" IA-64 implementations from achieving the peak
// performance. Well, not really... The catch is that if you
// implementation will have same latency:-). This stall will hold
// you back and the performance would be as if every second bundle
// were split *anyway*...
-{ .mfi getf.sig r16=f40
+{ .mfi; getf.sig r16=f40
xma.hu f42=f33,f120,f41
add r33=8,r32 }
-{ .mfi xma.lu f41=f33,f120,f41 };;
-{ .mfi getf.sig r24=f50
+{ .mfi; xma.lu f41=f33,f120,f41 };;
+{ .mfi; getf.sig r24=f50
xma.hu f52=f33,f121,f51 }
-{ .mfi xma.lu f51=f33,f121,f51 };;
-{ .mfi st8 [r32]=r16,16
+{ .mfi; xma.lu f51=f33,f121,f51 };;
+{ .mfi; st8 [r32]=r16,16
xma.hu f62=f33,f122,f61 }
-{ .mfi xma.lu f61=f33,f122,f61 };;
-{ .mfi xma.hu f72=f33,f123,f71 }
-{ .mfi xma.lu f71=f33,f123,f71 };;
-{ .mfi xma.hu f82=f33,f124,f81 }
-{ .mfi xma.lu f81=f33,f124,f81 };;
-{ .mfi xma.hu f92=f33,f125,f91 }
-{ .mfi xma.lu f91=f33,f125,f91 };;
-{ .mfi xma.hu f102=f33,f126,f101 }
-{ .mfi xma.lu f101=f33,f126,f101 };;
-{ .mfi xma.hu f112=f33,f127,f111 }
-{ .mfi xma.lu f111=f33,f127,f111 };;;;
+{ .mfi; xma.lu f61=f33,f122,f61 };;
+{ .mfi; xma.hu f72=f33,f123,f71 }
+{ .mfi; xma.lu f71=f33,f123,f71 };;
+{ .mfi; xma.hu f82=f33,f124,f81 }
+{ .mfi; xma.lu f81=f33,f124,f81 };;
+{ .mfi; xma.hu f92=f33,f125,f91 }
+{ .mfi; xma.lu f91=f33,f125,f91 };;
+{ .mfi; xma.hu f102=f33,f126,f101 }
+{ .mfi; xma.lu f101=f33,f126,f101 };;
+{ .mfi; xma.hu f112=f33,f127,f111 }
+{ .mfi; xma.lu f111=f33,f127,f111 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r25=f41
+{ .mfi; getf.sig r25=f41
xma.hu f43=f34,f120,f42 }
-{ .mfi xma.lu f42=f34,f120,f42 };;
-{ .mfi getf.sig r16=f60
+{ .mfi; xma.lu f42=f34,f120,f42 };;
+{ .mfi; getf.sig r16=f60
xma.hu f53=f34,f121,f52 }
-{ .mfi xma.lu f52=f34,f121,f52 };;
-{ .mfi getf.sig r17=f51
+{ .mfi; xma.lu f52=f34,f121,f52 };;
+{ .mfi; getf.sig r17=f51
xma.hu f63=f34,f122,f62
add r25=r25,r24 }
-{ .mfi xma.lu f62=f34,f122,f62
+{ .mfi; xma.lu f62=f34,f122,f62
mov carry1=0 };;
-{ .mfi cmp.ltu p6,p0=r25,r24
+{ .mfi; cmp.ltu p6,p0=r25,r24
xma.hu f73=f34,f123,f72 }
-{ .mfi xma.lu f72=f34,f123,f72 };;
-{ .mfi st8 [r33]=r25,16
+{ .mfi; xma.lu f72=f34,f123,f72 };;
+{ .mfi; st8 [r33]=r25,16
xma.hu f83=f34,f124,f82
(p6) add carry1=1,carry1 }
-{ .mfi xma.lu f82=f34,f124,f82 };;
-{ .mfi xma.hu f93=f34,f125,f92 }
-{ .mfi xma.lu f92=f34,f125,f92 };;
-{ .mfi xma.hu f103=f34,f126,f102 }
-{ .mfi xma.lu f102=f34,f126,f102 };;
-{ .mfi xma.hu f113=f34,f127,f112 }
-{ .mfi xma.lu f112=f34,f127,f112 };;;;
+{ .mfi; xma.lu f82=f34,f124,f82 };;
+{ .mfi; xma.hu f93=f34,f125,f92 }
+{ .mfi; xma.lu f92=f34,f125,f92 };;
+{ .mfi; xma.hu f103=f34,f126,f102 }
+{ .mfi; xma.lu f102=f34,f126,f102 };;
+{ .mfi; xma.hu f113=f34,f127,f112 }
+{ .mfi; xma.lu f112=f34,f127,f112 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r18=f42
+{ .mfi; getf.sig r18=f42
xma.hu f44=f35,f120,f43
add r17=r17,r16 }
-{ .mfi xma.lu f43=f35,f120,f43 };;
-{ .mfi getf.sig r24=f70
+{ .mfi; xma.lu f43=f35,f120,f43 };;
+{ .mfi; getf.sig r24=f70
xma.hu f54=f35,f121,f53 }
-{ .mfi mov carry2=0
+{ .mfi; mov carry2=0
xma.lu f53=f35,f121,f53 };;
-{ .mfi getf.sig r25=f61
+{ .mfi; getf.sig r25=f61
xma.hu f64=f35,f122,f63
cmp.ltu p7,p0=r17,r16 }
-{ .mfi add r18=r18,r17
+{ .mfi; add r18=r18,r17
xma.lu f63=f35,f122,f63 };;
-{ .mfi getf.sig r26=f52
+{ .mfi; getf.sig r26=f52
xma.hu f74=f35,f123,f73
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r18,r17
+{ .mfi; cmp.ltu p7,p0=r18,r17
xma.lu f73=f35,f123,f73
add r18=r18,carry1 };;
-{ .mfi
+{ .mfi;
xma.hu f84=f35,f124,f83
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r18,carry1
+{ .mfi; cmp.ltu p7,p0=r18,carry1
xma.lu f83=f35,f124,f83 };;
-{ .mfi st8 [r32]=r18,16
+{ .mfi; st8 [r32]=r18,16
xma.hu f94=f35,f125,f93
(p7) add carry2=1,carry2 }
-{ .mfi xma.lu f93=f35,f125,f93 };;
-{ .mfi xma.hu f104=f35,f126,f103 }
-{ .mfi xma.lu f103=f35,f126,f103 };;
-{ .mfi xma.hu f114=f35,f127,f113 }
-{ .mfi mov carry1=0
+{ .mfi; xma.lu f93=f35,f125,f93 };;
+{ .mfi; xma.hu f104=f35,f126,f103 }
+{ .mfi; xma.lu f103=f35,f126,f103 };;
+{ .mfi; xma.hu f114=f35,f127,f113 }
+{ .mfi; mov carry1=0
xma.lu f113=f35,f127,f113
- add r25=r25,r24 };;;;
+ add r25=r25,r24 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r27=f43
+{ .mfi; getf.sig r27=f43
xma.hu f45=f36,f120,f44
cmp.ltu p6,p0=r25,r24 }
-{ .mfi xma.lu f44=f36,f120,f44
+{ .mfi; xma.lu f44=f36,f120,f44
add r26=r26,r25 };;
-{ .mfi getf.sig r16=f80
+{ .mfi; getf.sig r16=f80
xma.hu f55=f36,f121,f54
(p6) add carry1=1,carry1 }
-{ .mfi xma.lu f54=f36,f121,f54 };;
-{ .mfi getf.sig r17=f71
+{ .mfi; xma.lu f54=f36,f121,f54 };;
+{ .mfi; getf.sig r17=f71
xma.hu f65=f36,f122,f64
cmp.ltu p6,p0=r26,r25 }
-{ .mfi xma.lu f64=f36,f122,f64
+{ .mfi; xma.lu f64=f36,f122,f64
add r27=r27,r26 };;
-{ .mfi getf.sig r18=f62
+{ .mfi; getf.sig r18=f62
xma.hu f75=f36,f123,f74
(p6) add carry1=1,carry1 }
-{ .mfi cmp.ltu p6,p0=r27,r26
+{ .mfi; cmp.ltu p6,p0=r27,r26
xma.lu f74=f36,f123,f74
add r27=r27,carry2 };;
-{ .mfi getf.sig r19=f53
+{ .mfi; getf.sig r19=f53
xma.hu f85=f36,f124,f84
(p6) add carry1=1,carry1 }
-{ .mfi xma.lu f84=f36,f124,f84
+{ .mfi; xma.lu f84=f36,f124,f84
cmp.ltu p6,p0=r27,carry2 };;
-{ .mfi st8 [r33]=r27,16
+{ .mfi; st8 [r33]=r27,16
xma.hu f95=f36,f125,f94
(p6) add carry1=1,carry1 }
-{ .mfi xma.lu f94=f36,f125,f94 };;
-{ .mfi xma.hu f105=f36,f126,f104 }
-{ .mfi mov carry2=0
+{ .mfi; xma.lu f94=f36,f125,f94 };;
+{ .mfi; xma.hu f105=f36,f126,f104 }
+{ .mfi; mov carry2=0
xma.lu f104=f36,f126,f104
add r17=r17,r16 };;
-{ .mfi xma.hu f115=f36,f127,f114
+{ .mfi; xma.hu f115=f36,f127,f114
cmp.ltu p7,p0=r17,r16 }
-{ .mfi xma.lu f114=f36,f127,f114
- add r18=r18,r17 };;;;
+{ .mfi; xma.lu f114=f36,f127,f114
+ add r18=r18,r17 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r20=f44
+{ .mfi; getf.sig r20=f44
xma.hu f46=f37,f120,f45
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r18,r17
+{ .mfi; cmp.ltu p7,p0=r18,r17
xma.lu f45=f37,f120,f45
add r19=r19,r18 };;
-{ .mfi getf.sig r24=f90
+{ .mfi; getf.sig r24=f90
xma.hu f56=f37,f121,f55 }
-{ .mfi xma.lu f55=f37,f121,f55 };;
-{ .mfi getf.sig r25=f81
+{ .mfi; xma.lu f55=f37,f121,f55 };;
+{ .mfi; getf.sig r25=f81
xma.hu f66=f37,f122,f65
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r19,r18
+{ .mfi; cmp.ltu p7,p0=r19,r18
xma.lu f65=f37,f122,f65
add r20=r20,r19 };;
-{ .mfi getf.sig r26=f72
+{ .mfi; getf.sig r26=f72
xma.hu f76=f37,f123,f75
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r20,r19
+{ .mfi; cmp.ltu p7,p0=r20,r19
xma.lu f75=f37,f123,f75
add r20=r20,carry1 };;
-{ .mfi getf.sig r27=f63
+{ .mfi; getf.sig r27=f63
xma.hu f86=f37,f124,f85
(p7) add carry2=1,carry2 }
-{ .mfi xma.lu f85=f37,f124,f85
+{ .mfi; xma.lu f85=f37,f124,f85
cmp.ltu p7,p0=r20,carry1 };;
-{ .mfi getf.sig r28=f54
+{ .mfi; getf.sig r28=f54
xma.hu f96=f37,f125,f95
(p7) add carry2=1,carry2 }
-{ .mfi st8 [r32]=r20,16
+{ .mfi; st8 [r32]=r20,16
xma.lu f95=f37,f125,f95 };;
-{ .mfi xma.hu f106=f37,f126,f105 }
-{ .mfi mov carry1=0
+{ .mfi; xma.hu f106=f37,f126,f105 }
+{ .mfi; mov carry1=0
xma.lu f105=f37,f126,f105
add r25=r25,r24 };;
-{ .mfi xma.hu f116=f37,f127,f115
+{ .mfi; xma.hu f116=f37,f127,f115
cmp.ltu p6,p0=r25,r24 }
-{ .mfi xma.lu f115=f37,f127,f115
- add r26=r26,r25 };;;;
+{ .mfi; xma.lu f115=f37,f127,f115
+ add r26=r26,r25 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r29=f45
+{ .mfi; getf.sig r29=f45
xma.hu f47=f38,f120,f46
(p6) add carry1=1,carry1 }
-{ .mfi cmp.ltu p6,p0=r26,r25
+{ .mfi; cmp.ltu p6,p0=r26,r25
xma.lu f46=f38,f120,f46
add r27=r27,r26 };;
-{ .mfi getf.sig r16=f100
+{ .mfi; getf.sig r16=f100
xma.hu f57=f38,f121,f56
(p6) add carry1=1,carry1 }
-{ .mfi cmp.ltu p6,p0=r27,r26
+{ .mfi; cmp.ltu p6,p0=r27,r26
xma.lu f56=f38,f121,f56
add r28=r28,r27 };;
-{ .mfi getf.sig r17=f91
+{ .mfi; getf.sig r17=f91
xma.hu f67=f38,f122,f66
(p6) add carry1=1,carry1 }
-{ .mfi cmp.ltu p6,p0=r28,r27
+{ .mfi; cmp.ltu p6,p0=r28,r27
xma.lu f66=f38,f122,f66
add r29=r29,r28 };;
-{ .mfi getf.sig r18=f82
+{ .mfi; getf.sig r18=f82
xma.hu f77=f38,f123,f76
(p6) add carry1=1,carry1 }
-{ .mfi cmp.ltu p6,p0=r29,r28
+{ .mfi; cmp.ltu p6,p0=r29,r28
xma.lu f76=f38,f123,f76
add r29=r29,carry2 };;
-{ .mfi getf.sig r19=f73
+{ .mfi; getf.sig r19=f73
xma.hu f87=f38,f124,f86
(p6) add carry1=1,carry1 }
-{ .mfi xma.lu f86=f38,f124,f86
+{ .mfi; xma.lu f86=f38,f124,f86
cmp.ltu p6,p0=r29,carry2 };;
-{ .mfi getf.sig r20=f64
+{ .mfi; getf.sig r20=f64
xma.hu f97=f38,f125,f96
(p6) add carry1=1,carry1 }
-{ .mfi st8 [r33]=r29,16
+{ .mfi; st8 [r33]=r29,16
xma.lu f96=f38,f125,f96 };;
-{ .mfi getf.sig r21=f55
+{ .mfi; getf.sig r21=f55
xma.hu f107=f38,f126,f106 }
-{ .mfi mov carry2=0
+{ .mfi; mov carry2=0
xma.lu f106=f38,f126,f106
add r17=r17,r16 };;
-{ .mfi xma.hu f117=f38,f127,f116
+{ .mfi; xma.hu f117=f38,f127,f116
cmp.ltu p7,p0=r17,r16 }
-{ .mfi xma.lu f116=f38,f127,f116
- add r18=r18,r17 };;;;
+{ .mfi; xma.lu f116=f38,f127,f116
+ add r18=r18,r17 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r22=f46
+{ .mfi; getf.sig r22=f46
xma.hu f48=f39,f120,f47
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r18,r17
+{ .mfi; cmp.ltu p7,p0=r18,r17
xma.lu f47=f39,f120,f47
add r19=r19,r18 };;
-{ .mfi getf.sig r24=f110
+{ .mfi; getf.sig r24=f110
xma.hu f58=f39,f121,f57
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r19,r18
+{ .mfi; cmp.ltu p7,p0=r19,r18
xma.lu f57=f39,f121,f57
add r20=r20,r19 };;
-{ .mfi getf.sig r25=f101
+{ .mfi; getf.sig r25=f101
xma.hu f68=f39,f122,f67
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r20,r19
+{ .mfi; cmp.ltu p7,p0=r20,r19
xma.lu f67=f39,f122,f67
add r21=r21,r20 };;
-{ .mfi getf.sig r26=f92
+{ .mfi; getf.sig r26=f92
xma.hu f78=f39,f123,f77
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r21,r20
+{ .mfi; cmp.ltu p7,p0=r21,r20
xma.lu f77=f39,f123,f77
add r22=r22,r21 };;
-{ .mfi getf.sig r27=f83
+{ .mfi; getf.sig r27=f83
xma.hu f88=f39,f124,f87
(p7) add carry2=1,carry2 }
-{ .mfi cmp.ltu p7,p0=r22,r21
+{ .mfi; cmp.ltu p7,p0=r22,r21
xma.lu f87=f39,f124,f87
add r22=r22,carry1 };;
-{ .mfi getf.sig r28=f74
+{ .mfi; getf.sig r28=f74
xma.hu f98=f39,f125,f97
(p7) add carry2=1,carry2 }
-{ .mfi xma.lu f97=f39,f125,f97
+{ .mfi; xma.lu f97=f39,f125,f97
cmp.ltu p7,p0=r22,carry1 };;
-{ .mfi getf.sig r29=f65
+{ .mfi; getf.sig r29=f65
xma.hu f108=f39,f126,f107
(p7) add carry2=1,carry2 }
-{ .mfi st8 [r32]=r22,16
+{ .mfi; st8 [r32]=r22,16
xma.lu f107=f39,f126,f107 };;
-{ .mfi getf.sig r30=f56
+{ .mfi; getf.sig r30=f56
xma.hu f118=f39,f127,f117 }
-{ .mfi xma.lu f117=f39,f127,f117 };;;;
+{ .mfi; xma.lu f117=f39,f127,f117 };;//
//-------------------------------------------------//
// Leaving muliplier's heaven... Quite a ride, huh?
-{ .mii getf.sig r31=f47
+{ .mii; getf.sig r31=f47
add r25=r25,r24
mov carry1=0 };;
-{ .mii getf.sig r16=f111
+{ .mii; getf.sig r16=f111
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mfb getf.sig r17=f102 }
-{ .mii
+{ .mfb; getf.sig r17=f102 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r27=r27,r26 };;
-{ .mfb nop.m 0x0 }
-{ .mii
+{ .mfb; nop.m 0x0 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r27,r26
add r28=r28,r27 };;
-{ .mii getf.sig r18=f93
+{ .mii; getf.sig r18=f93
add r17=r17,r16
mov carry3=0 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r28,r27
add r29=r29,r28 };;
-{ .mii getf.sig r19=f84
+{ .mii; getf.sig r19=f84
cmp.ltu p7,p0=r17,r16 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r29,r28
add r30=r30,r29 };;
-{ .mii getf.sig r20=f75
+{ .mii; getf.sig r20=f75
add r18=r18,r17 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r30,r29
add r31=r31,r30 };;
-{ .mfb getf.sig r21=f66 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r21=f66 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r18,r17
add r19=r19,r18 }
-{ .mfb nop.m 0x0 }
-{ .mii
+{ .mfb; nop.m 0x0 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r31,r30
add r31=r31,carry2 };;
-{ .mfb getf.sig r22=f57 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r22=f57 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r19,r18
add r20=r20,r19 }
-{ .mfb nop.m 0x0 }
-{ .mii
+{ .mfb; nop.m 0x0 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r31,carry2 };;
-{ .mfb getf.sig r23=f48 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r23=f48 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r20,r19
add r21=r21,r20 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1 }
-{ .mfb st8 [r33]=r31,16 };;
+{ .mfb; st8 [r33]=r31,16 };;
-{ .mfb getf.sig r24=f112 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r24=f112 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r21,r20
add r22=r22,r21 };;
-{ .mfb getf.sig r25=f103 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r25=f103 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r22,r21
add r23=r23,r22 };;
-{ .mfb getf.sig r26=f94 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r26=f94 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r23,r22
add r23=r23,carry1 };;
-{ .mfb getf.sig r27=f85 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r27=f85 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p8=r23,carry1};;
-{ .mii getf.sig r28=f76
+{ .mii; getf.sig r28=f76
add r25=r25,r24
mov carry1=0 }
-{ .mii st8 [r32]=r23,16
+{ .mii; st8 [r32]=r23,16
(p7) add carry2=1,carry3
(p8) add carry2=0,carry3 };;
-{ .mfb nop.m 0x0 }
-{ .mii getf.sig r29=f67
+{ .mfb; nop.m 0x0 }
+{ .mii; getf.sig r29=f67
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mfb getf.sig r30=f58 }
-{ .mii
+{ .mfb; getf.sig r30=f58 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r27=r27,r26 };;
-{ .mfb getf.sig r16=f113 }
-{ .mii
+{ .mfb; getf.sig r16=f113 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r27,r26
add r28=r28,r27 };;
-{ .mfb getf.sig r17=f104 }
-{ .mii
+{ .mfb; getf.sig r17=f104 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r28,r27
add r29=r29,r28 };;
-{ .mfb getf.sig r18=f95 }
-{ .mii
+{ .mfb; getf.sig r18=f95 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r29,r28
add r30=r30,r29 };;
-{ .mii getf.sig r19=f86
+{ .mii; getf.sig r19=f86
add r17=r17,r16
mov carry3=0 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r30,r29
add r30=r30,carry2 };;
-{ .mii getf.sig r20=f77
+{ .mii; getf.sig r20=f77
cmp.ltu p7,p0=r17,r16
add r18=r18,r17 }
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r30,carry2 };;
-{ .mfb getf.sig r21=f68 }
-{ .mii st8 [r33]=r30,16
+{ .mfb; getf.sig r21=f68 }
+{ .mii; st8 [r33]=r30,16
(p6) add carry1=1,carry1 };;
-{ .mfb getf.sig r24=f114 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r24=f114 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r18,r17
add r19=r19,r18 };;
-{ .mfb getf.sig r25=f105 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r25=f105 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r19,r18
add r20=r20,r19 };;
-{ .mfb getf.sig r26=f96 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r26=f96 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r20,r19
add r21=r21,r20 };;
-{ .mfb getf.sig r27=f87 }
-{ .mii (p7) add carry3=1,carry3
+{ .mfb; getf.sig r27=f87 }
+{ .mii; (p7) add carry3=1,carry3
cmp.ltu p7,p0=r21,r20
add r21=r21,carry1 };;
-{ .mib getf.sig r28=f78
+{ .mib; getf.sig r28=f78
add r25=r25,r24 }
-{ .mib (p7) add carry3=1,carry3
+{ .mib; (p7) add carry3=1,carry3
cmp.ltu p7,p8=r21,carry1};;
-{ .mii st8 [r32]=r21,16
+{ .mii; st8 [r32]=r21,16
(p7) add carry2=1,carry3
(p8) add carry2=0,carry3 }
-{ .mii mov carry1=0
+{ .mii; mov carry1=0
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mfb getf.sig r16=f115 }
-{ .mii
+{ .mfb; getf.sig r16=f115 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r27=r27,r26 };;
-{ .mfb getf.sig r17=f106 }
-{ .mii
+{ .mfb; getf.sig r17=f106 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r27,r26
add r28=r28,r27 };;
-{ .mfb getf.sig r18=f97 }
-{ .mii
+{ .mfb; getf.sig r18=f97 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r28,r27
add r28=r28,carry2 };;
-{ .mib getf.sig r19=f88
+{ .mib; getf.sig r19=f88
add r17=r17,r16 }
-{ .mib
+{ .mib;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r28,carry2 };;
-{ .mii st8 [r33]=r28,16
+{ .mii; st8 [r33]=r28,16
(p6) add carry1=1,carry1 }
-{ .mii mov carry2=0
+{ .mii; mov carry2=0
cmp.ltu p7,p0=r17,r16
add r18=r18,r17 };;
-{ .mfb getf.sig r24=f116 }
-{ .mii (p7) add carry2=1,carry2
+{ .mfb; getf.sig r24=f116 }
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r18,r17
add r19=r19,r18 };;
-{ .mfb getf.sig r25=f107 }
-{ .mii (p7) add carry2=1,carry2
+{ .mfb; getf.sig r25=f107 }
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r19,r18
add r19=r19,carry1 };;
-{ .mfb getf.sig r26=f98 }
-{ .mii (p7) add carry2=1,carry2
+{ .mfb; getf.sig r26=f98 }
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r19,carry1};;
-{ .mii st8 [r32]=r19,16
+{ .mii; st8 [r32]=r19,16
(p7) add carry2=1,carry2 }
-{ .mfb add r25=r25,r24 };;
+{ .mfb; add r25=r25,r24 };;
-{ .mfb getf.sig r16=f117 }
-{ .mii mov carry1=0
+{ .mfb; getf.sig r16=f117 }
+{ .mii; mov carry1=0
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mfb getf.sig r17=f108 }
-{ .mii
+{ .mfb; getf.sig r17=f108 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r26=r26,carry2 };;
-{ .mfb nop.m 0x0 }
-{ .mii
+{ .mfb; nop.m 0x0 }
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,carry2 };;
-{ .mii st8 [r33]=r26,16
+{ .mii; st8 [r33]=r26,16
(p6) add carry1=1,carry1 }
-{ .mfb add r17=r17,r16 };;
-{ .mfb getf.sig r24=f118 }
-{ .mii mov carry2=0
+{ .mfb; add r17=r17,r16 };;
+{ .mfb; getf.sig r24=f118 }
+{ .mii; mov carry2=0
cmp.ltu p7,p0=r17,r16
add r17=r17,carry1 };;
-{ .mii (p7) add carry2=1,carry2
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r17,carry1};;
-{ .mii st8 [r32]=r17
+{ .mii; st8 [r32]=r17
(p7) add carry2=1,carry2 };;
-{ .mfb add r24=r24,carry2 };;
-{ .mib st8 [r33]=r24 }
+{ .mfb; add r24=r24,carry2 };;
+{ .mib; st8 [r33]=r24 }
-{ .mib rum 1<<5 // clear um.mfh
+{ .mib; rum 1<<5 // clear um.mfh
br.ret.sptk.many b0 };;
.endp bn_mul_comba8#
-.endp bn_sqr_comba8#
#undef carry3
#undef carry2
#undef carry1
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,2,1,0,0
+#if defined(_HPUX_SOURCE) && !defined(_LP64)
+{ .mii; alloc r2=ar.pfs,2,1,0,0
+ addp4 r32=0,r32
+ addp4 r33=0,r33 };;
+{ .mii;
+#else
+{ .mii; alloc r2=ar.pfs,2,1,0,0
+#endif
mov r34=r33
add r14=8,r33 };;
.body
-{ .mii add r17=8,r34
+{ .mii; add r17=8,r34
add r15=16,r33
add r18=16,r34 }
-{ .mfb add r16=24,r33
+{ .mfb; add r16=24,r33
br .L_cheat_entry_point4 };;
+.endp bn_sqr_comba4#
#endif
#if 1
//
// void bn_mul_comba4(BN_ULONG *r, BN_ULONG *a, BN_ULONG *b)
//
-#define carry1 r14
+#define carry1 r14
#define carry2 r15
.global bn_mul_comba4#
.proc bn_mul_comba4#
.prologue
.fframe 0
.save ar.pfs,r2
-{ .mii alloc r2=ar.pfs,3,0,0,0
+#if defined(_HPUX_SOURCE) && !defined(_LP64)
+{ .mii; alloc r2=ar.pfs,3,0,0,0
+ addp4 r33=0,r33
+ addp4 r34=0,r34 };;
+{ .mii; addp4 r32=0,r32
+#else
+{ .mii; alloc r2=ar.pfs,3,0,0,0
+#endif
add r14=8,r33
add r17=8,r34 }
.body
-{ .mii add r15=16,r33
+{ .mii; add r15=16,r33
add r18=16,r34
add r16=24,r33 };;
.L_cheat_entry_point4:
-{ .mmi add r19=24,r34
+{ .mmi; add r19=24,r34
ldf8 f32=[r33] }
-{ .mmi ldf8 f120=[r34]
+{ .mmi; ldf8 f120=[r34]
ldf8 f121=[r17] };;
-{ .mmi ldf8 f122=[r18]
+{ .mmi; ldf8 f122=[r18]
ldf8 f123=[r19] }
-{ .mmi ldf8 f33=[r14]
+{ .mmi; ldf8 f33=[r14]
ldf8 f34=[r15] }
-{ .mfi ldf8 f35=[r16]
+{ .mfi; ldf8 f35=[r16]
xma.hu f41=f32,f120,f0 }
-{ .mfi xma.lu f40=f32,f120,f0 };;
-{ .mfi xma.hu f51=f32,f121,f0 }
-{ .mfi xma.lu f50=f32,f121,f0 };;
-{ .mfi xma.hu f61=f32,f122,f0 }
-{ .mfi xma.lu f60=f32,f122,f0 };;
-{ .mfi xma.hu f71=f32,f123,f0 }
-{ .mfi xma.lu f70=f32,f123,f0 };;;;
+{ .mfi; xma.lu f40=f32,f120,f0 };;
+{ .mfi; xma.hu f51=f32,f121,f0 }
+{ .mfi; xma.lu f50=f32,f121,f0 };;
+{ .mfi; xma.hu f61=f32,f122,f0 }
+{ .mfi; xma.lu f60=f32,f122,f0 };;
+{ .mfi; xma.hu f71=f32,f123,f0 }
+{ .mfi; xma.lu f70=f32,f123,f0 };;//
// Major stall takes place here, and 3 more places below. Result from
// first xma is not available for another 3 ticks.
-{ .mfi getf.sig r16=f40
+{ .mfi; getf.sig r16=f40
xma.hu f42=f33,f120,f41
add r33=8,r32 }
-{ .mfi xma.lu f41=f33,f120,f41 };;
-{ .mfi getf.sig r24=f50
+{ .mfi; xma.lu f41=f33,f120,f41 };;
+{ .mfi; getf.sig r24=f50
xma.hu f52=f33,f121,f51 }
-{ .mfi xma.lu f51=f33,f121,f51 };;
-{ .mfi st8 [r32]=r16,16
+{ .mfi; xma.lu f51=f33,f121,f51 };;
+{ .mfi; st8 [r32]=r16,16
xma.hu f62=f33,f122,f61 }
-{ .mfi xma.lu f61=f33,f122,f61 };;
-{ .mfi xma.hu f72=f33,f123,f71 }
-{ .mfi xma.lu f71=f33,f123,f71 };;;;
+{ .mfi; xma.lu f61=f33,f122,f61 };;
+{ .mfi; xma.hu f72=f33,f123,f71 }
+{ .mfi; xma.lu f71=f33,f123,f71 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r25=f41
+{ .mfi; getf.sig r25=f41
xma.hu f43=f34,f120,f42 }
-{ .mfi xma.lu f42=f34,f120,f42 };;
-{ .mfi getf.sig r16=f60
+{ .mfi; xma.lu f42=f34,f120,f42 };;
+{ .mfi; getf.sig r16=f60
xma.hu f53=f34,f121,f52 }
-{ .mfi xma.lu f52=f34,f121,f52 };;
-{ .mfi getf.sig r17=f51
+{ .mfi; xma.lu f52=f34,f121,f52 };;
+{ .mfi; getf.sig r17=f51
xma.hu f63=f34,f122,f62
add r25=r25,r24 }
-{ .mfi mov carry1=0
+{ .mfi; mov carry1=0
xma.lu f62=f34,f122,f62 };;
-{ .mfi st8 [r33]=r25,16
+{ .mfi; st8 [r33]=r25,16
xma.hu f73=f34,f123,f72
cmp.ltu p6,p0=r25,r24 }
-{ .mfi xma.lu f72=f34,f123,f72 };;;;
+{ .mfi; xma.lu f72=f34,f123,f72 };;//
//-------------------------------------------------//
-{ .mfi getf.sig r18=f42
+{ .mfi; getf.sig r18=f42
xma.hu f44=f35,f120,f43
(p6) add carry1=1,carry1 }
-{ .mfi add r17=r17,r16
+{ .mfi; add r17=r17,r16
xma.lu f43=f35,f120,f43
mov carry2=0 };;
-{ .mfi getf.sig r24=f70
+{ .mfi; getf.sig r24=f70
xma.hu f54=f35,f121,f53
cmp.ltu p7,p0=r17,r16 }
-{ .mfi xma.lu f53=f35,f121,f53 };;
-{ .mfi getf.sig r25=f61
+{ .mfi; xma.lu f53=f35,f121,f53 };;
+{ .mfi; getf.sig r25=f61
xma.hu f64=f35,f122,f63
add r18=r18,r17 }
-{ .mfi xma.lu f63=f35,f122,f63
+{ .mfi; xma.lu f63=f35,f122,f63
(p7) add carry2=1,carry2 };;
-{ .mfi getf.sig r26=f52
+{ .mfi; getf.sig r26=f52
xma.hu f74=f35,f123,f73
cmp.ltu p7,p0=r18,r17 }
-{ .mfi xma.lu f73=f35,f123,f73
+{ .mfi; xma.lu f73=f35,f123,f73
add r18=r18,carry1 };;
//-------------------------------------------------//
-{ .mii st8 [r32]=r18,16
+{ .mii; st8 [r32]=r18,16
(p7) add carry2=1,carry2
cmp.ltu p7,p0=r18,carry1 };;
-{ .mfi getf.sig r27=f43 // last major stall
+{ .mfi; getf.sig r27=f43 // last major stall
(p7) add carry2=1,carry2 };;
-{ .mii getf.sig r16=f71
+{ .mii; getf.sig r16=f71
add r25=r25,r24
mov carry1=0 };;
-{ .mii getf.sig r17=f62
+{ .mii; getf.sig r17=f62
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r27=r27,r26 };;
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r27,r26
add r27=r27,carry2 };;
-{ .mii getf.sig r18=f53
+{ .mii; getf.sig r18=f53
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r27,carry2 };;
-{ .mfi st8 [r33]=r27,16
+{ .mfi; st8 [r33]=r27,16
(p6) add carry1=1,carry1 }
-{ .mii getf.sig r19=f44
+{ .mii; getf.sig r19=f44
add r17=r17,r16
mov carry2=0 };;
-{ .mii getf.sig r24=f72
+{ .mii; getf.sig r24=f72
cmp.ltu p7,p0=r17,r16
add r18=r18,r17 };;
-{ .mii (p7) add carry2=1,carry2
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r18,r17
add r19=r19,r18 };;
-{ .mii (p7) add carry2=1,carry2
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r19,r18
add r19=r19,carry1 };;
-{ .mii getf.sig r25=f63
+{ .mii; getf.sig r25=f63
(p7) add carry2=1,carry2
cmp.ltu p7,p0=r19,carry1};;
-{ .mii st8 [r32]=r19,16
+{ .mii; st8 [r32]=r19,16
(p7) add carry2=1,carry2 }
-{ .mii getf.sig r26=f54
+{ .mii; getf.sig r26=f54
add r25=r25,r24
mov carry1=0 };;
-{ .mii getf.sig r16=f73
+{ .mii; getf.sig r16=f73
cmp.ltu p6,p0=r25,r24
add r26=r26,r25 };;
-{ .mii
+{ .mii;
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,r25
add r26=r26,carry2 };;
-{ .mii getf.sig r17=f64
+{ .mii; getf.sig r17=f64
(p6) add carry1=1,carry1
cmp.ltu p6,p0=r26,carry2 };;
-{ .mii st8 [r33]=r26,16
+{ .mii; st8 [r33]=r26,16
(p6) add carry1=1,carry1 }
-{ .mii getf.sig r24=f74
+{ .mii; getf.sig r24=f74
add r17=r17,r16
mov carry2=0 };;
-{ .mii cmp.ltu p7,p0=r17,r16
+{ .mii; cmp.ltu p7,p0=r17,r16
add r17=r17,carry1 };;
-{ .mii (p7) add carry2=1,carry2
+{ .mii; (p7) add carry2=1,carry2
cmp.ltu p7,p0=r17,carry1};;
-{ .mii st8 [r32]=r17,16
+{ .mii; st8 [r32]=r17,16
(p7) add carry2=1,carry2 };;
-{ .mii add r24=r24,carry2 };;
-{ .mii st8 [r33]=r24 }
+{ .mii; add r24=r24,carry2 };;
+{ .mii; st8 [r33]=r24 }
-{ .mib rum 1<<5 // clear um.mfh
+{ .mib; rum 1<<5 // clear um.mfh
br.ret.sptk.many b0 };;
.endp bn_mul_comba4#
-.endp bn_sqr_comba4#
#undef carry2
#undef carry1
#endif
#define DH r22
#define I r21
+#if 0
+// Some preprocessors (most notably HP-UX) appear to be allergic to
+// macros enclosed to parenthesis [as these three were].
#define cont p16
#define break p0 // p20
#define equ p24
+#else
+cont=p16
+break=p0
+equ=p24
+#endif
+
.global abort#
.global bn_div_words#
.proc bn_div_words#
.fframe 0
.save ar.pfs,r2
.save b0,r3
-{ .mii alloc r2=ar.pfs,3,5,0,8
+{ .mii; alloc r2=ar.pfs,3,5,0,8
mov r3=b0
mov r10=pr };;
-{ .mmb cmp.eq p6,p0=r34,r0
+{ .mmb; cmp.eq p6,p0=r34,r0
mov r8=-1
(p6) br.ret.spnt.many b0 };;
.body
-{ .mii mov H=r32 // save h
+{ .mii; mov H=r32 // save h
mov ar.ec=0 // don't rotate at exit
mov pr.rot=0 }
-{ .mii mov L=r33 // save l
- mov r36=r0 }
+{ .mii; mov L=r33 // save l
+ mov r36=r0 };;
-1: ;; // -vv- note signed comparison
-{ .mfi (p0) cmp.lt p16,p0=r0,r34 // d
+.L_divw_shift: // -vv- note signed comparison
+{ .mfi; (p0) cmp.lt p16,p0=r0,r34 // d
(p0) shladd r33=r34,1,r0 }
-{ .mfb (p0) add r35=1,r36
+{ .mfb; (p0) add r35=1,r36
(p0) nop.f 0x0
-(p16) br.wtop.dpnt 1b };;
+(p16) br.wtop.dpnt .L_divw_shift };;
-{ .mii mov D=r34
+{ .mii; mov D=r34
shr.u DH=r34,32
sub r35=64,r36 };;
-{ .mii setf.sig f7=DH
+{ .mii; setf.sig f7=DH
shr.u AT=H,r35
mov I=r36 };;
-{ .mib cmp.ne p6,p0=r0,AT
+{ .mib; cmp.ne p6,p0=r0,AT
shl H=H,r36
(p6) br.call.spnt.clr b0=abort };; // overflow, die...
-{ .mfi fcvt.xuf.s1 f7=f7
+{ .mfi; fcvt.xuf.s1 f7=f7
shr.u AT=L,r35 };;
-{ .mii shl L=L,r36
+{ .mii; shl L=L,r36
or H=H,AT };;
-{ .mii nop.m 0x0
+{ .mii; nop.m 0x0
cmp.leu p6,p0=D,H;;
(p6) sub H=H,D }
-{ .mlx setf.sig f14=D
+{ .mlx; setf.sig f14=D
movl AT=0xffffffff };;
///////////////////////////////////////////////////////////
-{ .mii setf.sig f6=H
+{ .mii; setf.sig f6=H
shr.u HH=H,32;;
cmp.eq p6,p7=HH,DH };;
-{ .mfb
+{ .mfb;
(p6) setf.sig f8=AT
(p7) fcvt.xuf.s1 f6=f6
(p7) br.call.sptk b6=.L_udiv64_32_b6 };;
-{ .mfi getf.sig r33=f8 // q
+{ .mfi; getf.sig r33=f8 // q
xmpy.lu f9=f8,f14 }
-{ .mfi xmpy.hu f10=f8,f14
+{ .mfi; xmpy.hu f10=f8,f14
shrp H=H,L,32 };;
-{ .mmi getf.sig r35=f9 // tl
+{ .mmi; getf.sig r35=f9 // tl
getf.sig r31=f10 };; // th
-2: ;;
-{ .mii (p0) add r32=-1,r33
+.L_divw_1st_iter:
+{ .mii; (p0) add r32=-1,r33
(p0) cmp.eq equ,cont=HH,r31 };;
-{ .mii (p0) cmp.ltu p8,p0=r35,D
+{ .mii; (p0) cmp.ltu p8,p0=r35,D
(p0) sub r34=r35,D
(equ) cmp.leu break,cont=r35,H };;
-{ .mib (cont) cmp.leu cont,break=HH,r31
+{ .mib; (cont) cmp.leu cont,break=HH,r31
(p8) add r31=-1,r31
-(cont) br.wtop.spnt 2b };;
+(cont) br.wtop.spnt .L_divw_1st_iter };;
///////////////////////////////////////////////////////////
-{ .mii sub H=H,r35
+{ .mii; sub H=H,r35
shl r8=r33,32
shl L=L,32 };;
///////////////////////////////////////////////////////////
-{ .mii setf.sig f6=H
+{ .mii; setf.sig f6=H
shr.u HH=H,32;;
cmp.eq p6,p7=HH,DH };;
-{ .mfb
+{ .mfb;
(p6) setf.sig f8=AT
(p7) fcvt.xuf.s1 f6=f6
(p7) br.call.sptk b6=.L_udiv64_32_b6 };;
-{ .mfi getf.sig r33=f8 // q
+{ .mfi; getf.sig r33=f8 // q
xmpy.lu f9=f8,f14 }
-{ .mfi xmpy.hu f10=f8,f14
+{ .mfi; xmpy.hu f10=f8,f14
shrp H=H,L,32 };;
-{ .mmi getf.sig r35=f9 // tl
+{ .mmi; getf.sig r35=f9 // tl
getf.sig r31=f10 };; // th
-2: ;;
-{ .mii (p0) add r32=-1,r33
+.L_divw_2nd_iter:
+{ .mii; (p0) add r32=-1,r33
(p0) cmp.eq equ,cont=HH,r31 };;
-{ .mii (p0) cmp.ltu p8,p0=r35,D
+{ .mii; (p0) cmp.ltu p8,p0=r35,D
(p0) sub r34=r35,D
(equ) cmp.leu break,cont=r35,H };;
-{ .mib (cont) cmp.leu cont,break=HH,r31
+{ .mib; (cont) cmp.leu cont,break=HH,r31
(p8) add r31=-1,r31
-(cont) br.wtop.spnt 2b };;
+(cont) br.wtop.spnt .L_divw_2nd_iter };;
///////////////////////////////////////////////////////////
-{ .mii sub H=H,r35
+{ .mii; sub H=H,r35
or r8=r8,r33
mov ar.pfs=r2 };;
-{ .mii shr.u r9=H,I // remainder if anybody wants it
- mov pr=r10,-1 }
-{ .mfb br.ret.sptk.many b0 };;
+{ .mii; shr.u r9=H,I // remainder if anybody wants it
+ mov pr=r10,0x1ffff }
+{ .mfb; br.ret.sptk.many b0 };;
// Unsigned 64 by 32 (well, by 64 for the moment) bit integer division
// procedure.
//
// inputs: f6 = (double)a, f7 = (double)b
// output: f8 = (int)(a/b)
-// clobbered: f8,f9,f10,f11,PR
-#define PR p15
-// This procedure is essentially Intel code and therefore is
-// copyrighted to Intel Corporation (I suppose...). It's sligtly
-// modified for specific needs.
+// clobbered: f8,f9,f10,f11,pred
+pred=p15
+// One can argue that this snippet is copyrighted to Intel
+// Corporation, as it's essentially identical to one of those
+// found in "Divide, Square Root and Remainder" section at
+// http://www.intel.com/software/products/opensource/libraries/num.htm.
+// Yes, I admit that the referred code was used as template,
+// but after I realized that there hardly is any other instruction
+// sequence which would perform this operation. I mean I figure that
+// any independent attempt to implement high-performance division
+// will result in code virtually identical to the Intel code. It
+// should be noted though that below division kernel is 1 cycle
+// faster than Intel one (note commented splits:-), not to mention
+// original prologue (rather lack of one) and epilogue.
.align 32
-.space 16
+.skip 16
.L_udiv64_32_b6:
- frcpa.s1 f8,PR=f6,f7;; // [0] y0 = 1 / b
-
-(PR) fnma.s1 f9=f7,f8,f1 // [5] e0 = 1 - b * y0
-(PR) fmpy.s1 f10=f6,f8;; // [5] q0 = a * y0
-(PR) fmpy.s1 f11=f9,f9 // [10] e1 = e0 * e0
-(PR) fma.s1 f10=f9,f10,f10;; // [10] q1 = q0 + e0 * q0
-(PR) fma.s1 f8=f9,f8,f8 //;; // [15] y1 = y0 + e0 * y0
-(PR) fma.s1 f9=f11,f10,f10;; // [15] q2 = q1 + e1 * q1
-(PR) fma.s1 f8=f11,f8,f8 //;; // [20] y2 = y1 + e1 * y1
-(PR) fnma.s1 f10=f7,f9,f6;; // [20] r2 = a - b * q2
-(PR) fma.s1 f8=f10,f8,f9;; // [25] q3 = q2 + r2 * y2
+ frcpa.s1 f8,pred=f6,f7;; // [0] y0 = 1 / b
+
+(pred) fnma.s1 f9=f7,f8,f1 // [5] e0 = 1 - b * y0
+(pred) fmpy.s1 f10=f6,f8;; // [5] q0 = a * y0
+(pred) fmpy.s1 f11=f9,f9 // [10] e1 = e0 * e0
+(pred) fma.s1 f10=f9,f10,f10;; // [10] q1 = q0 + e0 * q0
+(pred) fma.s1 f8=f9,f8,f8 //;; // [15] y1 = y0 + e0 * y0
+(pred) fma.s1 f9=f11,f10,f10;; // [15] q2 = q1 + e1 * q1
+(pred) fma.s1 f8=f11,f8,f8 //;; // [20] y2 = y1 + e1 * y1
+(pred) fnma.s1 f10=f7,f9,f6;; // [20] r2 = a - b * q2
+(pred) fma.s1 f8=f10,f8,f9;; // [25] q3 = q2 + r2 * y2
fcvt.fxu.trunc.s1 f8=f8 // [30] q = trunc(q3)
br.ret.sptk.many b6;;