/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define N %i0
#define X %i1
#define INCX %i2
#define Y %i3
#define INCY %i4
#define I %i5
#define XX %l0
#define YY %l1
#ifdef DOUBLE
#define a1 %f4
#define a2 %f6
#define a3 %f8
#define a4 %f10
#define a5 %f12
#define a6 %f14
#define a7 %f16
#define a8 %f18
#define b1 %f20
#define b2 %f22
#define b3 %f24
#define b4 %f26
#define b5 %f28
#define b6 %f30
#define b7 %f32
#define b8 %f34
#define c1 %f36
#define c2 %f38
#define c3 %f40
#define c4 %f42
#define c5 %f44
#define c6 %f46
#define c7 %f48
#define c8 %f50
#define t1 %f52
#define t2 %f54
#define t3 %f56
#define t4 %f58
#else
#define a1 %f2
#define a2 %f3
#define a3 %f4
#define a4 %f5
#define a5 %f6
#define a6 %f7
#define a7 %f8
#define a8 %f9
#define b1 %f10
#define b2 %f11
#define b3 %f12
#define b4 %f13
#define b5 %f14
#define b6 %f15
#define b7 %f16
#define b8 %f17
#define c1 %f18
#define c2 %f19
#define c3 %f20
#define c4 %f21
#define c5 %f22
#define c6 %f23
#define c7 %f24
#define c8 %f25
#define t1 %f26
#define t2 %f27
#define t3 %f28
#define t4 %f29
#endif
#ifdef DOUBLE
#define C %f0
#define S %f2
#else
#define C %f0
#define S %f1
#endif
PROLOGUE
SAVESP
#ifndef __64BIT__
#ifdef DOUBLE
st %i5, [%sp + STACK_START + 24]
LDF [%sp + STACK_START + 24], C
LDF [%sp + STACK_START + 32], S
#else
st %i5, [%sp + STACK_START + 24]
LDF [%sp + STACK_START + 24], C
LDF [%sp + STACK_START + 28], S
#endif
#else
#ifdef DOUBLE
FMOV %f10, C
FMOV %f12, S
#else
FMOV %f11, C
FMOV %f13, S
#endif
#endif
cmp N, 0
ble .LL19
nop
sll INCX, BASE_SHIFT, INCX
sll INCY, BASE_SHIFT, INCY
cmp INCX, SIZE
bne .LL50
nop
cmp INCY, SIZE
bne .LL50
nop
sra N, 3, I
cmp I, 0
ble,pn %icc, .LL15
nop
LDF [X + 0 * SIZE], a1
LDF [Y + 0 * SIZE], b1
LDF [X + 1 * SIZE], a2
LDF [Y + 1 * SIZE], b2
LDF [X + 2 * SIZE], a3
LDF [Y + 2 * SIZE], b3
LDF [X + 3 * SIZE], a4
LDF [Y + 3 * SIZE], b4
LDF [X + 4 * SIZE], a5
LDF [Y + 4 * SIZE], b5
LDF [X + 5 * SIZE], a6
LDF [Y + 5 * SIZE], b6
LDF [X + 6 * SIZE], a7
LDF [Y + 6 * SIZE], b7
LDF [X + 7 * SIZE], a8
LDF [Y + 7 * SIZE], b8
FMUL C, a1, c1
FMUL S, b1, c2
FMUL C, b1, c3
LDF [Y + 8 * SIZE], b1
FMUL S, a1, c4
LDF [X + 8 * SIZE], a1
FMUL C, a2, c5
FMUL S, b2, c6
FADD c1, c2, t1
FMUL C, b2, c7
LDF [Y + 9 * SIZE], b2
FMUL S, a2, c8
LDF [X + 9 * SIZE], a2
FSUB c3, c4, t2
addcc I, -1, I
ble,pt %icc, .LL12
nop
#define PREFETCHSIZE 64
.LL11:
FMUL C, a3, c1
nop
prefetch [Y + PREFETCHSIZE * SIZE], 1
nop
FMUL S, b3, c2
STF t1, [X + 0 * SIZE]
FADD c5, c6, t3
nop
FMUL C, b3, c3
LDF [Y + 10 * SIZE], b3
nop
nop
FMUL S, a3, c4
STF t2, [Y + 0 * SIZE]
FSUB c7, c8, t4
nop
FMUL C, a4, c5
LDF [X + 10 * SIZE], a3
nop
nop
FMUL S, b4, c6
STF t3, [X + 1 * SIZE]
FADD c1, c2, t1
nop
FMUL C, b4, c7
LDF [Y + 11 * SIZE], b4
nop
nop
FMUL S, a4, c8
STF t4, [Y + 1 * SIZE]
FSUB c3, c4, t2
nop
FMUL C, a5, c1
LDF [X + 11 * SIZE], a4
nop
nop
FMUL S, b5, c2
STF t1, [X + 2 * SIZE]
FADD c5, c6, t3
nop
FMUL C, b5, c3
LDF [Y + 12 * SIZE], b5
nop
nop
FMUL S, a5, c4
STF t2, [Y + 2 * SIZE]
FSUB c7, c8, t4
nop
FMUL C, a6, c5
LDF [X + 12 * SIZE], a5
nop
nop
FMUL S, b6, c6
STF t3, [X + 3 * SIZE]
FADD c1, c2, t1
nop
FMUL C, b6, c7
LDF [Y + 13 * SIZE], b6
nop
nop
FMUL S, a6, c8
STF t4, [Y + 3 * SIZE]
FSUB c3, c4, t2
nop
FMUL C, a7, c1
LDF [X + 13 * SIZE], a6
nop
nop
FMUL S, b7, c2
STF t1, [X + 4 * SIZE]
FADD c5, c6, t3
nop
FMUL C, b7, c3
LDF [Y + 14 * SIZE], b7
nop
nop
FMUL S, a7, c4
STF t2, [Y + 4 * SIZE]
FSUB c7, c8, t4
nop
FMUL C, a8, c5
LDF [X + 14 * SIZE], a7
nop
nop
FMUL S, b8, c6
STF t3, [X + 5 * SIZE]
FADD c1, c2, t1
nop
FMUL C, b8, c7
LDF [Y + 15 * SIZE], b8
nop
nop
FMUL S, a8, c8
STF t4, [Y + 5 * SIZE]
FSUB c3, c4, t2
nop
FMUL C, a1, c1
LDF [X + 15 * SIZE], a8
addcc I, -1, I
nop
FMUL S, b1, c2
STF t1, [X + 6 * SIZE]
FADD c5, c6, t3
nop
FMUL C, b1, c3
LDF [Y + 16 * SIZE], b1
nop
nop
FMUL S, a1, c4
STF t2, [Y + 6 * SIZE]
FSUB c7, c8, t4
nop
FMUL C, a2, c5
LDF [X + 16 * SIZE], a1
add Y, 8 * SIZE, Y
nop
FMUL S, b2, c6
STF t3, [X + 7 * SIZE]
FADD c1, c2, t1
nop
FMUL C, b2, c7
LDF [Y + 9 * SIZE], b2
add X, 8 * SIZE, X
nop
FMUL S, a2, c8
STF t4, [Y - 1 * SIZE]
FSUB c3, c4, t2
nop
bg,pt %icc, .LL11
LDF [X + 9 * SIZE], a2
.LL12:
FMUL C, a3, c1
FMUL S, b3, c2
STF t1, [X + 0 * SIZE]
FADD c5, c6, t3
FMUL C, b3, c3
FMUL S, a3, c4
STF t2, [Y + 0 * SIZE]
FSUB c7, c8, t4
FMUL C, a4, c5
FMUL S, b4, c6
STF t3, [X + 1 * SIZE]
FADD c1, c2, t1
FMUL C, b4, c7
FMUL S, a4, c8
STF t4, [Y + 1 * SIZE]
FSUB c3, c4, t2
FMUL C, a5, c1
FMUL S, b5, c2
STF t1, [X + 2 * SIZE]
FADD c5, c6, t3
FMUL C, b5, c3
FMUL S, a5, c4
STF t2, [Y + 2 * SIZE]
FSUB c7, c8, t4
FMUL C, a6, c5
FMUL S, b6, c6
STF t3, [X + 3 * SIZE]
FADD c1, c2, t1
FMUL C, b6, c7
FMUL S, a6, c8
STF t4, [Y + 3 * SIZE]
FSUB c3, c4, t2
FMUL C, a7, c1
FMUL S, b7, c2
STF t1, [X + 4 * SIZE]
FADD c5, c6, t3
FMUL C, b7, c3
FMUL S, a7, c4
STF t2, [Y + 4 * SIZE]
FSUB c7, c8, t4
FMUL C, a8, c5
FMUL S, b8, c6
STF t3, [X + 5 * SIZE]
FADD c1, c2, t1
FMUL C, b8, c7
FMUL S, a8, c8
STF t4, [Y + 5 * SIZE]
FSUB c3, c4, t2
FADD c5, c6, t3
STF t1, [X + 6 * SIZE]
FSUB c7, c8, t4
STF t2, [Y + 6 * SIZE]
STF t3, [X + 7 * SIZE]
STF t4, [Y + 7 * SIZE]
add X, 8 * SIZE, X
add Y, 8 * SIZE, Y
.LL15:
andcc N, 7, I
nop
ble,a,pn %icc, .LL19
nop
.LL16:
LDF [X + 0 * SIZE], a1
add X, 1 * SIZE, X
LDF [Y + 0 * SIZE], b1
add Y, 1 * SIZE, Y
FMUL C, a1, c1
FMUL S, b1, c2
FMUL C, b1, c3
FMUL S, a1, c4
FADD c1, c2, c2
addcc I, -1, I
FSUB c3, c4, c4
nop
STF c2, [X - 1 * SIZE]
STF c4, [Y - 1 * SIZE]
bg,pt %icc, .LL16
nop
.LL19:
return %i7 + 8
nop
.LL50:
mov X, XX
mov Y, YY
sra N, 3, I
cmp I, 0
ble,pn %icc, .LL55
nop
.LL51:
LDF [X + 0 * SIZE], a1
add X, INCX, X
LDF [Y + 0 * SIZE], b1
add Y, INCY, Y
LDF [X + 0 * SIZE], a2
add X, INCX, X
LDF [Y + 0 * SIZE], b2
add Y, INCY, Y
LDF [X + 0 * SIZE], a3
add X, INCX, X
LDF [Y + 0 * SIZE], b3
add Y, INCY, Y
LDF [X + 0 * SIZE], a4
add X, INCX, X
LDF [Y + 0 * SIZE], b4
add Y, INCY, Y
LDF [X + 0 * SIZE], a5
add X, INCX, X
LDF [Y + 0 * SIZE], b5
add Y, INCY, Y
LDF [X + 0 * SIZE], a6
add X, INCX, X
LDF [Y + 0 * SIZE], b6
add Y, INCY, Y
LDF [X + 0 * SIZE], a7
add X, INCX, X
LDF [Y + 0 * SIZE], b7
add Y, INCY, Y
LDF [X + 0 * SIZE], a8
add X, INCX, X
LDF [Y + 0 * SIZE], b8
add Y, INCY, Y
FMUL C, a1, c1
FMUL S, b1, c2
FMUL C, b1, c3
FMUL S, a1, c4
FADD c1, c2, t1
FSUB c3, c4, t2
STF t1, [XX + 0 * SIZE]
add XX, INCX, XX
STF t2, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a2, c5
FMUL S, b2, c6
FMUL C, b2, c7
FMUL S, a2, c8
FADD c5, c6, t3
FSUB c7, c8, t4
STF t3, [XX + 0 * SIZE]
add XX, INCX, XX
STF t4, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a3, c1
FMUL S, b3, c2
FMUL C, b3, c3
FMUL S, a3, c4
FADD c1, c2, t1
FSUB c3, c4, t2
STF t1, [XX + 0 * SIZE]
add XX, INCX, XX
STF t2, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a4, c5
FMUL S, b4, c6
FMUL C, b4, c7
FMUL S, a4, c8
FADD c5, c6, t3
FSUB c7, c8, t4
STF t3, [XX + 0 * SIZE]
add XX, INCX, XX
STF t4, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a5, c1
FMUL S, b5, c2
FMUL C, b5, c3
FMUL S, a5, c4
FADD c1, c2, t1
FSUB c3, c4, t2
STF t1, [XX + 0 * SIZE]
add XX, INCX, XX
STF t2, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a6, c5
FMUL S, b6, c6
FMUL C, b6, c7
FMUL S, a6, c8
FADD c5, c6, t3
FSUB c7, c8, t4
STF t3, [XX + 0 * SIZE]
add XX, INCX, XX
STF t4, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a7, c1
FMUL S, b7, c2
FMUL C, b7, c3
FMUL S, a7, c4
FADD c1, c2, t1
FSUB c3, c4, t2
STF t1, [XX + 0 * SIZE]
add XX, INCX, XX
STF t2, [YY + 0 * SIZE]
add YY, INCY, YY
FMUL C, a8, c5
FMUL S, b8, c6
FMUL C, b8, c7
FMUL S, a8, c8
FADD c5, c6, t3
FSUB c7, c8, t4
STF t3, [XX + 0 * SIZE]
add XX, INCX, XX
STF t4, [YY + 0 * SIZE]
add YY, INCY, YY
addcc I, -1, I
bg,pt %icc, .LL51
nop
.LL55:
andcc N, 7, I
nop
ble %icc, .LL59
nop
.LL56:
LDF [X + 0 * SIZE], a1
LDF [Y + 0 * SIZE], b1
FMUL C, a1, c1
FMUL S, b1, c2
FMUL C, b1, c3
FMUL S, a1, c4
FADD c1, c2, c2
FSUB c3, c4, c4
STF c2, [X + 0 * SIZE]
add X, INCX, X
STF c4, [Y + 0 * SIZE]
addcc I, -1, I
bg %icc, .LL56
add Y, INCY, Y
.LL59:
return %i7 + 8
nop
EPILOGUE