/*********************************************************************/ /* Copyright 2009, 2010 The University of Texas at Austin. */ /* All rights reserved. */ /* */ /* Redistribution and use in source and binary forms, with or */ /* without modification, are permitted provided that the following */ /* conditions are met: */ /* */ /* 1. Redistributions of source code must retain the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer. */ /* */ /* 2. Redistributions in binary form must reproduce the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer in the documentation and/or other materials */ /* provided with the distribution. */ /* */ /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ /* POSSIBILITY OF SUCH DAMAGE. */ /* */ /* The views and conclusions contained in the software and */ /* documentation are those of the authors and should not be */ /* interpreted as representing official policies, either expressed */ /* or implied, of The University of Texas at Austin. */ /*********************************************************************/ #define ASSEMBLER #include "common.h" #include "version.h" #define PREFETCHSIZE 40 #ifndef CONJ #define ADD1 SUB #define ADD2 ADD #else #define ADD1 ADD #define ADD2 SUB #endif PROLOGUE PROFCODE .frame $sp, 16, $26, 0 ldl $19, 0($sp) fmov $f19, $f29 ldq $20, 8($sp) fmov $f20, $f30 mov $21, $18 ldl $21, 16($sp) lda $sp, -64($sp) nop stt $f2, 0($sp) cmpeq $19, 1, $1 stt $f3, 8($sp) cmpeq $21, 1, $2 stt $f4, 16($sp) and $16, 3, $5 stt $f5, 24($sp) stt $f6, 32($sp) stt $f7, 40($sp) stt $f8, 48($sp) #ifndef PROFILE .prologue 0 #else .prologue 1 #endif and $1, $2, $1 ble $16, $End sra $16, 2, $4 beq $1, $Sub ble $4, $Remain subq $4, 1, $4 LD $f0, 0*SIZE($18) LD $f1, 1*SIZE($18) LD $f2, 2*SIZE($18) LD $f3, 3*SIZE($18) LD $f4, 4*SIZE($18) LD $f5, 5*SIZE($18) LD $f6, 6*SIZE($18) LD $f7, 7*SIZE($18) LD $f8, 0*SIZE($20) LD $f28, 1*SIZE($20) LD $f10, 2*SIZE($20) LD $f11, 3*SIZE($20) LD $f12, 4*SIZE($20) LD $f13, 5*SIZE($20) LD $f14, 6*SIZE($20) LD $f15, 7*SIZE($20) addq $18, 8*SIZE, $18 ble $4, $MainLoopEnd .align 4 $MainLoop: ldt $f31, PREFETCHSIZE * SIZE($20) ldl $31, PREFETCHSIZE * SIZE($18) MUL $f29, $f0, $f20 LD $f31, 9*SIZE($18) MUL $f30, $f1, $f21 unop MUL $f30, $f0, $f22 LD $f0, 0*SIZE($18) MUL $f29, $f1, $f23 LD $f1, 1*SIZE($18) MUL $f29, $f2, $f24 unop MUL $f30, $f3, $f25 nop MUL $f30, $f2, $f26 LD $f2, 2*SIZE($18) MUL $f29, $f3, $f27 LD $f3, 3*SIZE($18) ADD1 $f20, $f21, $f16 MUL $f29, $f4, $f20 ADD2 $f22, $f23, $f17 MUL $f30, $f5, $f21 ADD1 $f24, $f25, $f18 unop MUL $f30, $f4, $f22 LD $f4, 4*SIZE($18) ADD2 $f26, $f27, $f19 addq $20, 8*SIZE, $20 MUL $f29, $f5, $f23 LD $f5, 5*SIZE($18) ADD $f16, $f8, $f16 LD $f8, 0*SIZE($20) MUL $f29, $f6, $f24 unop ADD $f17, $f28, $f17 LD $f28, 1*SIZE($20) MUL $f30, $f7, $f25 unop ADD $f18, $f10, $f18 LD $f10, 2*SIZE($20) MUL $f30, $f6, $f26 LD $f6, 6*SIZE($18) ADD $f19, $f11, $f19 LD $f11, 3*SIZE($20) MUL $f29, $f7, $f27 LD $f7, 7*SIZE($18) ST $f16,-8*SIZE($20) ADD1 $f20, $f21, $f16 ST $f17,-7*SIZE($20) ADD2 $f22, $f23, $f17 ST $f18,-6*SIZE($20) ADD1 $f24, $f25, $f18 ST $f19,-5*SIZE($20) ADD2 $f26, $f27, $f19 ADD $f16, $f12, $f16 LD $f12, 4*SIZE($20) ADD $f17, $f13, $f17 LD $f13, 5*SIZE($20) ADD $f18, $f14, $f18 LD $f14, 6*SIZE($20) ADD $f19, $f15, $f19 LD $f15, 7*SIZE($20) ST $f16,-4*SIZE($20) addq $18, 8*SIZE, $18 ST $f17,-3*SIZE($20) subq $4, 1, $4 ST $f18,-2*SIZE($20) nop ST $f19,-1*SIZE($20) bgt $4, $MainLoop .align 4 $MainLoopEnd: MUL $f29, $f0, $f20 MUL $f30, $f1, $f21 MUL $f30, $f0, $f22 MUL $f29, $f1, $f23 MUL $f29, $f2, $f24 MUL $f30, $f3, $f25 MUL $f30, $f2, $f26 MUL $f29, $f3, $f27 ADD1 $f20, $f21, $f16 MUL $f29, $f4, $f20 ADD2 $f22, $f23, $f17 MUL $f30, $f5, $f21 ADD1 $f24, $f25, $f18 MUL $f30, $f4, $f22 ADD2 $f26, $f27, $f19 MUL $f29, $f5, $f23 ADD $f16, $f8, $f16 MUL $f29, $f6, $f24 ADD $f17, $f28, $f17 MUL $f30, $f7, $f25 ADD $f18, $f10, $f18 MUL $f30, $f6, $f26 ADD $f19, $f11, $f19 MUL $f29, $f7, $f27 ST $f16, 0*SIZE($20) ADD1 $f20, $f21, $f16 ST $f17, 1*SIZE($20) ADD2 $f22, $f23, $f17 ST $f18, 2*SIZE($20) ADD1 $f24, $f25, $f18 ST $f19, 3*SIZE($20) ADD2 $f26, $f27, $f19 ADD $f16, $f12, $f16 ADD $f17, $f13, $f17 ADD $f18, $f14, $f18 ADD $f19, $f15, $f19 ST $f16, 4*SIZE($20) ST $f17, 5*SIZE($20) ST $f18, 6*SIZE($20) ST $f19, 7*SIZE($20) unop addq $20, 8*SIZE, $20 unop ble $5, $End .align 4 $Remain: subq $5, 1, $6 ble $5, $End LD $f0, 0*SIZE($18) LD $f1, 1*SIZE($18) LD $f8, 0*SIZE($20) LD $f28, 1*SIZE($20) addq $18, 2*SIZE, $18 ble $6, $RemainLoopEnd .align 4 $RemainLoop: MUL $f29, $f0, $f20 subq $6, 1, $6 MUL $f30, $f1, $f21 addq $20, 2*SIZE, $20 MUL $f30, $f0, $f22 LD $f0, 0*SIZE($18) MUL $f29, $f1, $f23 LD $f1, 1*SIZE($18) ADD1 $f20, $f21, $f16 ADD2 $f22, $f23, $f17 ADD $f16, $f8, $f16 LD $f8, 0*SIZE($20) ADD $f17, $f28, $f17 LD $f28, 1*SIZE($20) ST $f16,-2*SIZE($20) addq $18, 2*SIZE, $18 ST $f17,-1*SIZE($20) bgt $6, $RemainLoop .align 4 $RemainLoopEnd: MUL $f29, $f0, $f20 MUL $f30, $f1, $f21 MUL $f30, $f0, $f22 MUL $f29, $f1, $f23 ADD1 $f20, $f21, $f16 ADD2 $f22, $f23, $f17 ADD $f16, $f8, $f16 ADD $f17, $f28, $f17 ST $f16, 0*SIZE($20) nop ST $f17, 1*SIZE($20) nop .align 4 $End: ldt $f2, 0($sp) ldt $f3, 8($sp) ldt $f4, 16($sp) ldt $f5, 24($sp) ldt $f6, 32($sp) ldt $f7, 40($sp) ldt $f8, 48($sp) lda $sp, 64($sp) ret .align 4 $Sub: SXSUBL $16, SIZE, $22 addq $22, $22, $22 # Complex .align 4 addq $19, $19, $19 # Complex addq $21, $21, $21 # Complex ble $4, $SubRemain LD $f0, 0*SIZE($18) LD $f1, 1*SIZE($18) SXADDQ $19, $18, $18 LD $f2, 0*SIZE($18) LD $f3, 1*SIZE($18) SXADDQ $19, $18, $18 LD $f4, 0*SIZE($18) LD $f5, 1*SIZE($18) SXADDQ $19, $18, $18 LD $f6, 0*SIZE($18) LD $f7, 1*SIZE($18) SXADDQ $19, $18, $18 LD $f8, 0*SIZE($20) LD $f28, 1*SIZE($20) SXADDQ $21, $20, $24 LD $f10, 0*SIZE($24) LD $f11, 1*SIZE($24) SXADDQ $21, $24, $24 LD $f12, 0*SIZE($24) LD $f13, 1*SIZE($24) SXADDQ $21, $24, $24 LD $f14, 0*SIZE($24) LD $f15, 1*SIZE($24) SXADDQ $21, $24, $24 subq $4, 1, $4 ble $4, $SubMainLoopEnd .align 4 $SubMainLoop: MUL $f29, $f0, $f20 unop MUL $f30, $f1, $f21 unop MUL $f30, $f0, $f22 LD $f0, 0*SIZE($18) MUL $f29, $f1, $f23 LD $f1, 1*SIZE($18) MUL $f29, $f2, $f24 SXADDQ $19, $18, $18 MUL $f30, $f3, $f25 unop MUL $f30, $f2, $f26 LD $f2, 0*SIZE($18) MUL $f29, $f3, $f27 LD $f3, 1*SIZE($18) ADD1 $f20, $f21, $f16 SXADDQ $19, $18, $18 MUL $f29, $f4, $f20 unop ADD2 $f22, $f23, $f17 unop MUL $f30, $f5, $f21 unop ADD1 $f24, $f25, $f18 unop MUL $f30, $f4, $f22 LD $f4, 0*SIZE($18) ADD2 $f26, $f27, $f19 unop MUL $f29, $f5, $f23 LD $f5, 1*SIZE($18) ADD $f16, $f8, $f16 LD $f8, 0*SIZE($24) MUL $f29, $f6, $f24 SXADDQ $19, $18, $18 ADD $f17, $f28, $f17 LD $f28, 1*SIZE($24) MUL $f30, $f7, $f25 SXADDQ $21, $24, $24 ADD $f18, $f10, $f18 LD $f10, 0*SIZE($24) MUL $f30, $f6, $f26 LD $f6, 0*SIZE($18) ADD $f19, $f11, $f19 LD $f11, 1*SIZE($24) MUL $f29, $f7, $f27 LD $f7, 1*SIZE($18) ST $f16, 0*SIZE($20) SXADDQ $19, $18, $18 ADD1 $f20, $f21, $f16 unop ST $f17, 1*SIZE($20) SXADDQ $21, $20, $20 ADD2 $f22, $f23, $f17 unop ST $f18, 0*SIZE($20) SXADDQ $21, $24, $24 ADD1 $f24, $f25, $f18 unop ST $f19, 1*SIZE($20) unop ADD2 $f26, $f27, $f19 SXADDQ $21, $20, $20 ADD $f16, $f12, $f16 unop LD $f12, 0*SIZE($24) unop ADD $f17, $f13, $f17 unop LD $f13, 1*SIZE($24) SXADDQ $21, $24, $24 ADD $f18, $f14, $f18 subq $4, 1, $4 LD $f14, 0*SIZE($24) unop ADD $f19, $f15, $f19 unop LD $f15, 1*SIZE($24) SXADDQ $21, $24, $24 ST $f16, 0*SIZE($20) ST $f17, 1*SIZE($20) SXADDQ $21, $20, $20 unop ST $f18, 0*SIZE($20) ST $f19, 1*SIZE($20) SXADDQ $21, $20, $20 bgt $4, $SubMainLoop .align 4 $SubMainLoopEnd: MUL $f29, $f0, $f20 MUL $f30, $f1, $f21 MUL $f30, $f0, $f22 MUL $f29, $f1, $f23 MUL $f29, $f2, $f24 MUL $f30, $f3, $f25 MUL $f30, $f2, $f26 MUL $f29, $f3, $f27 ADD1 $f20, $f21, $f16 MUL $f29, $f4, $f20 ADD2 $f22, $f23, $f17 MUL $f30, $f5, $f21 ADD1 $f24, $f25, $f18 MUL $f30, $f4, $f22 ADD2 $f26, $f27, $f19 MUL $f29, $f5, $f23 ADD $f16, $f8, $f16 MUL $f29, $f6, $f24 ADD $f17, $f28, $f17 MUL $f30, $f7, $f25 ADD $f18, $f10, $f18 MUL $f30, $f6, $f26 ADD $f19, $f11, $f19 MUL $f29, $f7, $f27 ST $f16, 0*SIZE($20) ADD1 $f20, $f21, $f16 ST $f17, 1*SIZE($20) ADD2 $f22, $f23, $f17 SXADDQ $21, $20, $20 nop ST $f18, 0*SIZE($20) ADD1 $f24, $f25, $f18 ST $f19, 1*SIZE($20) ADD2 $f26, $f27, $f19 SXADDQ $21, $20, $20 ADD $f16, $f12, $f16 ADD $f17, $f13, $f17 ADD $f18, $f14, $f18 ADD $f19, $f15, $f19 ST $f16, 0*SIZE($20) ST $f17, 1*SIZE($20) SXADDQ $21, $20, $20 ST $f18, 0*SIZE($20) ST $f19, 1*SIZE($20) SXADDQ $21, $20, $20 ble $5, $SubEnd .align 4 $SubRemain: subq $5, 1, $6 ble $5, $SubEnd LD $f0, 0*SIZE($18) LD $f1, 1*SIZE($18) LD $f8, 0*SIZE($20) LD $f28, 1*SIZE($20) SXADDQ $19, $18, $18 SXADDQ $21, $20, $24 ble $6, $SubRemainLoopEnd .align 4 $SubRemainLoop: MUL $f29, $f0, $f20 MUL $f30, $f1, $f21 MUL $f30, $f0, $f22 LD $f0, 0*SIZE($18) MUL $f29, $f1, $f23 LD $f1, 1*SIZE($18) ADD1 $f20, $f21, $f16 SXADDQ $19, $18, $18 ADD2 $f22, $f23, $f17 nop ADD $f16, $f8, $f16 LD $f8, 0*SIZE($24) ADD $f17, $f28, $f17 LD $f28, 1*SIZE($24) SXADDQ $21, $24, $24 subq $6, 1, $6 ST $f16, 0*SIZE($20) ST $f17, 1*SIZE($20) SXADDQ $21, $20, $20 bgt $6, $SubRemainLoop .align 4 $SubRemainLoopEnd: MUL $f29, $f0, $f20 MUL $f30, $f1, $f21 MUL $f30, $f0, $f22 MUL $f29, $f1, $f23 ADD1 $f20, $f21, $f16 ADD2 $f22, $f23, $f17 ADD $f16, $f8, $f16 ADD $f17, $f28, $f17 ST $f16, 0*SIZE($20) nop ST $f17, 1*SIZE($20) nop .align 4 $SubEnd: ldt $f2, 0($sp) ldt $f3, 8($sp) ldt $f4, 16($sp) ldt $f5, 24($sp) ldt $f6, 32($sp) ldt $f7, 40($sp) ldt $f8, 48($sp) lda $sp, 64($sp) ret EPILOGUE