/*********************************************************************/ /* Copyright 2009, 2010 The University of Texas at Austin. */ /* All rights reserved. */ /* */ /* Redistribution and use in source and binary forms, with or */ /* without modification, are permitted provided that the following */ /* conditions are met: */ /* */ /* 1. Redistributions of source code must retain the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer. */ /* */ /* 2. Redistributions in binary form must reproduce the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer in the documentation and/or other materials */ /* provided with the distribution. */ /* */ /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ /* POSSIBILITY OF SUCH DAMAGE. */ /* */ /* The views and conclusions contained in the software and */ /* documentation are those of the authors and should not be */ /* interpreted as representing official policies, either expressed */ /* or implied, of The University of Texas at Austin. */ /*********************************************************************/ #define ASSEMBLER #include "common.h" #define N %i0 #define X %i1 #define INCX %i2 #define Y %i3 #define INCY %i4 #define I %i5 #define XX %l0 #define YY %l1 #ifdef DOUBLE #define a1 %f4 #define a2 %f6 #define a3 %f8 #define a4 %f10 #define a5 %f12 #define a6 %f14 #define a7 %f16 #define a8 %f18 #define b1 %f20 #define b2 %f22 #define b3 %f24 #define b4 %f26 #define b5 %f28 #define b6 %f30 #define b7 %f32 #define b8 %f34 #define c1 %f36 #define c2 %f38 #define c3 %f40 #define c4 %f42 #define c5 %f44 #define c6 %f46 #define c7 %f48 #define c8 %f50 #define t1 %f52 #define t2 %f54 #define t3 %f56 #define t4 %f58 #else #define a1 %f2 #define a2 %f3 #define a3 %f4 #define a4 %f5 #define a5 %f6 #define a6 %f7 #define a7 %f8 #define a8 %f9 #define b1 %f10 #define b2 %f11 #define b3 %f12 #define b4 %f13 #define b5 %f14 #define b6 %f15 #define b7 %f16 #define b8 %f17 #define c1 %f18 #define c2 %f19 #define c3 %f20 #define c4 %f21 #define c5 %f22 #define c6 %f23 #define c7 %f24 #define c8 %f25 #define t1 %f26 #define t2 %f27 #define t3 %f28 #define t4 %f29 #endif #ifdef DOUBLE #define C %f0 #define S %f2 #else #define C %f0 #define S %f1 #endif PROLOGUE SAVESP #ifndef __64BIT__ #ifdef DOUBLE st %i5, [%sp + STACK_START + 24] LDF [%sp + STACK_START + 24], C LDF [%sp + STACK_START + 32], S #else st %i5, [%sp + STACK_START + 24] LDF [%sp + STACK_START + 24], C LDF [%sp + STACK_START + 28], S #endif #else #ifdef DOUBLE FMOV %f10, C FMOV %f12, S #else FMOV %f11, C FMOV %f13, S #endif #endif cmp N, 0 ble .LL19 nop sll INCX, BASE_SHIFT, INCX sll INCY, BASE_SHIFT, INCY cmp INCX, SIZE bne .LL50 nop cmp INCY, SIZE bne .LL50 nop sra N, 3, I cmp I, 0 ble,pn %icc, .LL15 nop LDF [X + 0 * SIZE], a1 LDF [Y + 0 * SIZE], b1 LDF [X + 1 * SIZE], a2 LDF [Y + 1 * SIZE], b2 LDF [X + 2 * SIZE], a3 LDF [Y + 2 * SIZE], b3 LDF [X + 3 * SIZE], a4 LDF [Y + 3 * SIZE], b4 LDF [X + 4 * SIZE], a5 LDF [Y + 4 * SIZE], b5 LDF [X + 5 * SIZE], a6 LDF [Y + 5 * SIZE], b6 LDF [X + 6 * SIZE], a7 LDF [Y + 6 * SIZE], b7 LDF [X + 7 * SIZE], a8 LDF [Y + 7 * SIZE], b8 FMUL C, a1, c1 FMUL S, b1, c2 FMUL C, b1, c3 LDF [Y + 8 * SIZE], b1 FMUL S, a1, c4 LDF [X + 8 * SIZE], a1 FMUL C, a2, c5 FMUL S, b2, c6 FADD c1, c2, t1 FMUL C, b2, c7 LDF [Y + 9 * SIZE], b2 FMUL S, a2, c8 LDF [X + 9 * SIZE], a2 FSUB c3, c4, t2 addcc I, -1, I ble,pt %icc, .LL12 nop #define PREFETCHSIZE 64 .LL11: FMUL C, a3, c1 nop prefetch [Y + PREFETCHSIZE * SIZE], 1 nop FMUL S, b3, c2 STF t1, [X + 0 * SIZE] FADD c5, c6, t3 nop FMUL C, b3, c3 LDF [Y + 10 * SIZE], b3 nop nop FMUL S, a3, c4 STF t2, [Y + 0 * SIZE] FSUB c7, c8, t4 nop FMUL C, a4, c5 LDF [X + 10 * SIZE], a3 nop nop FMUL S, b4, c6 STF t3, [X + 1 * SIZE] FADD c1, c2, t1 nop FMUL C, b4, c7 LDF [Y + 11 * SIZE], b4 nop nop FMUL S, a4, c8 STF t4, [Y + 1 * SIZE] FSUB c3, c4, t2 nop FMUL C, a5, c1 LDF [X + 11 * SIZE], a4 nop nop FMUL S, b5, c2 STF t1, [X + 2 * SIZE] FADD c5, c6, t3 nop FMUL C, b5, c3 LDF [Y + 12 * SIZE], b5 nop nop FMUL S, a5, c4 STF t2, [Y + 2 * SIZE] FSUB c7, c8, t4 nop FMUL C, a6, c5 LDF [X + 12 * SIZE], a5 nop nop FMUL S, b6, c6 STF t3, [X + 3 * SIZE] FADD c1, c2, t1 nop FMUL C, b6, c7 LDF [Y + 13 * SIZE], b6 nop nop FMUL S, a6, c8 STF t4, [Y + 3 * SIZE] FSUB c3, c4, t2 nop FMUL C, a7, c1 LDF [X + 13 * SIZE], a6 nop nop FMUL S, b7, c2 STF t1, [X + 4 * SIZE] FADD c5, c6, t3 nop FMUL C, b7, c3 LDF [Y + 14 * SIZE], b7 nop nop FMUL S, a7, c4 STF t2, [Y + 4 * SIZE] FSUB c7, c8, t4 nop FMUL C, a8, c5 LDF [X + 14 * SIZE], a7 nop nop FMUL S, b8, c6 STF t3, [X + 5 * SIZE] FADD c1, c2, t1 nop FMUL C, b8, c7 LDF [Y + 15 * SIZE], b8 nop nop FMUL S, a8, c8 STF t4, [Y + 5 * SIZE] FSUB c3, c4, t2 nop FMUL C, a1, c1 LDF [X + 15 * SIZE], a8 addcc I, -1, I nop FMUL S, b1, c2 STF t1, [X + 6 * SIZE] FADD c5, c6, t3 nop FMUL C, b1, c3 LDF [Y + 16 * SIZE], b1 nop nop FMUL S, a1, c4 STF t2, [Y + 6 * SIZE] FSUB c7, c8, t4 nop FMUL C, a2, c5 LDF [X + 16 * SIZE], a1 add Y, 8 * SIZE, Y nop FMUL S, b2, c6 STF t3, [X + 7 * SIZE] FADD c1, c2, t1 nop FMUL C, b2, c7 LDF [Y + 9 * SIZE], b2 add X, 8 * SIZE, X nop FMUL S, a2, c8 STF t4, [Y - 1 * SIZE] FSUB c3, c4, t2 nop bg,pt %icc, .LL11 LDF [X + 9 * SIZE], a2 .LL12: FMUL C, a3, c1 FMUL S, b3, c2 STF t1, [X + 0 * SIZE] FADD c5, c6, t3 FMUL C, b3, c3 FMUL S, a3, c4 STF t2, [Y + 0 * SIZE] FSUB c7, c8, t4 FMUL C, a4, c5 FMUL S, b4, c6 STF t3, [X + 1 * SIZE] FADD c1, c2, t1 FMUL C, b4, c7 FMUL S, a4, c8 STF t4, [Y + 1 * SIZE] FSUB c3, c4, t2 FMUL C, a5, c1 FMUL S, b5, c2 STF t1, [X + 2 * SIZE] FADD c5, c6, t3 FMUL C, b5, c3 FMUL S, a5, c4 STF t2, [Y + 2 * SIZE] FSUB c7, c8, t4 FMUL C, a6, c5 FMUL S, b6, c6 STF t3, [X + 3 * SIZE] FADD c1, c2, t1 FMUL C, b6, c7 FMUL S, a6, c8 STF t4, [Y + 3 * SIZE] FSUB c3, c4, t2 FMUL C, a7, c1 FMUL S, b7, c2 STF t1, [X + 4 * SIZE] FADD c5, c6, t3 FMUL C, b7, c3 FMUL S, a7, c4 STF t2, [Y + 4 * SIZE] FSUB c7, c8, t4 FMUL C, a8, c5 FMUL S, b8, c6 STF t3, [X + 5 * SIZE] FADD c1, c2, t1 FMUL C, b8, c7 FMUL S, a8, c8 STF t4, [Y + 5 * SIZE] FSUB c3, c4, t2 FADD c5, c6, t3 STF t1, [X + 6 * SIZE] FSUB c7, c8, t4 STF t2, [Y + 6 * SIZE] STF t3, [X + 7 * SIZE] STF t4, [Y + 7 * SIZE] add X, 8 * SIZE, X add Y, 8 * SIZE, Y .LL15: andcc N, 7, I nop ble,a,pn %icc, .LL19 nop .LL16: LDF [X + 0 * SIZE], a1 add X, 1 * SIZE, X LDF [Y + 0 * SIZE], b1 add Y, 1 * SIZE, Y FMUL C, a1, c1 FMUL S, b1, c2 FMUL C, b1, c3 FMUL S, a1, c4 FADD c1, c2, c2 addcc I, -1, I FSUB c3, c4, c4 nop STF c2, [X - 1 * SIZE] STF c4, [Y - 1 * SIZE] bg,pt %icc, .LL16 nop .LL19: return %i7 + 8 nop .LL50: mov X, XX mov Y, YY sra N, 3, I cmp I, 0 ble,pn %icc, .LL55 nop .LL51: LDF [X + 0 * SIZE], a1 add X, INCX, X LDF [Y + 0 * SIZE], b1 add Y, INCY, Y LDF [X + 0 * SIZE], a2 add X, INCX, X LDF [Y + 0 * SIZE], b2 add Y, INCY, Y LDF [X + 0 * SIZE], a3 add X, INCX, X LDF [Y + 0 * SIZE], b3 add Y, INCY, Y LDF [X + 0 * SIZE], a4 add X, INCX, X LDF [Y + 0 * SIZE], b4 add Y, INCY, Y LDF [X + 0 * SIZE], a5 add X, INCX, X LDF [Y + 0 * SIZE], b5 add Y, INCY, Y LDF [X + 0 * SIZE], a6 add X, INCX, X LDF [Y + 0 * SIZE], b6 add Y, INCY, Y LDF [X + 0 * SIZE], a7 add X, INCX, X LDF [Y + 0 * SIZE], b7 add Y, INCY, Y LDF [X + 0 * SIZE], a8 add X, INCX, X LDF [Y + 0 * SIZE], b8 add Y, INCY, Y FMUL C, a1, c1 FMUL S, b1, c2 FMUL C, b1, c3 FMUL S, a1, c4 FADD c1, c2, t1 FSUB c3, c4, t2 STF t1, [XX + 0 * SIZE] add XX, INCX, XX STF t2, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a2, c5 FMUL S, b2, c6 FMUL C, b2, c7 FMUL S, a2, c8 FADD c5, c6, t3 FSUB c7, c8, t4 STF t3, [XX + 0 * SIZE] add XX, INCX, XX STF t4, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a3, c1 FMUL S, b3, c2 FMUL C, b3, c3 FMUL S, a3, c4 FADD c1, c2, t1 FSUB c3, c4, t2 STF t1, [XX + 0 * SIZE] add XX, INCX, XX STF t2, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a4, c5 FMUL S, b4, c6 FMUL C, b4, c7 FMUL S, a4, c8 FADD c5, c6, t3 FSUB c7, c8, t4 STF t3, [XX + 0 * SIZE] add XX, INCX, XX STF t4, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a5, c1 FMUL S, b5, c2 FMUL C, b5, c3 FMUL S, a5, c4 FADD c1, c2, t1 FSUB c3, c4, t2 STF t1, [XX + 0 * SIZE] add XX, INCX, XX STF t2, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a6, c5 FMUL S, b6, c6 FMUL C, b6, c7 FMUL S, a6, c8 FADD c5, c6, t3 FSUB c7, c8, t4 STF t3, [XX + 0 * SIZE] add XX, INCX, XX STF t4, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a7, c1 FMUL S, b7, c2 FMUL C, b7, c3 FMUL S, a7, c4 FADD c1, c2, t1 FSUB c3, c4, t2 STF t1, [XX + 0 * SIZE] add XX, INCX, XX STF t2, [YY + 0 * SIZE] add YY, INCY, YY FMUL C, a8, c5 FMUL S, b8, c6 FMUL C, b8, c7 FMUL S, a8, c8 FADD c5, c6, t3 FSUB c7, c8, t4 STF t3, [XX + 0 * SIZE] add XX, INCX, XX STF t4, [YY + 0 * SIZE] add YY, INCY, YY addcc I, -1, I bg,pt %icc, .LL51 nop .LL55: andcc N, 7, I nop ble %icc, .LL59 nop .LL56: LDF [X + 0 * SIZE], a1 LDF [Y + 0 * SIZE], b1 FMUL C, a1, c1 FMUL S, b1, c2 FMUL C, b1, c3 FMUL S, a1, c4 FADD c1, c2, c2 FSUB c3, c4, c4 STF c2, [X + 0 * SIZE] add X, INCX, X STF c4, [Y + 0 * SIZE] addcc I, -1, I bg %icc, .LL56 add Y, INCY, Y .LL59: return %i7 + 8 nop EPILOGUE