/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define M %i0
#define N %i1
#define K %i2
#if defined(DOUBLE) && !defined(__64BIT__)
#define A %i5
#define B %i4
#else
#define A %i4
#define B %i5
#endif
#define C %o4
#define LDC %o5
#define AO %l0
#define BO %l1
#define I %l2
#define J %l3
#define L %l4
#define C1 %o0
#define C2 %o1
#define C3 %o2
#define C4 %o3
#define OFFSET %l5
#define KK %l6
#define TEMP1 %l7
#define TEMP2 %i3
#define AORIG %g1
#ifdef DOUBLE
#define c01 %f0
#define c02 %f2
#define c03 %f4
#define c04 %f6
#define c05 %f8
#define c06 %f10
#define c07 %f12
#define c08 %f14
#define c09 %f16
#define c10 %f18
#define c11 %f20
#define c12 %f22
#define c13 %f24
#define c14 %f26
#define c15 %f28
#define c16 %f30
#define t1 %f32
#define t2 %f34
#define t3 %f36
#define t4 %f38
#define a1 %f40
#define a2 %f42
#define a3 %f44
#define a4 %f46
#define a5 %f58
#define b1 %f48
#define b2 %f50
#define b3 %f52
#define b4 %f54
#define b5 %f56
#define FZERO %f60
#define ALPHA %f62
#else
#define c01 %f0
#define c02 %f1
#define c03 %f2
#define c04 %f3
#define c05 %f4
#define c06 %f5
#define c07 %f6
#define c08 %f7
#define c09 %f8
#define c10 %f9
#define c11 %f10
#define c12 %f11
#define c13 %f12
#define c14 %f13
#define c15 %f14
#define c16 %f15
#define t1 %f16
#define t2 %f17
#define t3 %f18
#define t4 %f19
#define a1 %f20
#define a2 %f21
#define a3 %f22
#define a4 %f23
#define a5 %f31
#define b1 %f24
#define b2 %f25
#define b3 %f26
#define b4 %f27
#define b5 %f28
#define FZERO %f29
#define ALPHA %f30
#endif
#define APREFETCHSIZE 40
#define BPREFETCHSIZE 40
#define APREFETCH_CATEGORY 0
#define BPREFETCH_CATEGORY 0
PROLOGUE
SAVESP
nop
#ifndef __64BIT__
#ifdef DOUBLE
ld [%sp + STACK_START + 28], B
ld [%sp + STACK_START + 32], C
ld [%sp + STACK_START + 36], LDC
ld [%sp + STACK_START + 40], OFFSET
#else
ld [%sp + STACK_START + 28], C
ld [%sp + STACK_START + 32], LDC
ld [%sp + STACK_START + 36], OFFSET
#endif
#else
ldx [%sp+ STACK_START + 56], C
ldx [%sp+ STACK_START + 64], LDC
ldx [%sp+ STACK_START + 72], OFFSET
#endif
FCLR(29)
sll LDC, BASE_SHIFT, LDC
#ifdef LN
smul M, K, TEMP1
sll TEMP1, BASE_SHIFT, TEMP1
add A, TEMP1, A
sll M, BASE_SHIFT, TEMP1
add C, TEMP1, C
#endif
#ifdef RN
neg OFFSET, KK
#endif
#ifdef RT
smul N, K, TEMP1
sll TEMP1, BASE_SHIFT, TEMP1
add B, TEMP1, B
smul N, LDC, TEMP1
add C, TEMP1, C
sub N, OFFSET, KK
#endif
and N, 1, J
cmp J, 0
ble,pn %icc, .LL100
nop
#ifdef RT
sll K, 0 + BASE_SHIFT, TEMP1
sub B, TEMP1, B
sub C, LDC, C
#endif
mov C, C1
#ifdef LN
add M, OFFSET, KK
#endif
#ifdef LT
mov OFFSET, KK
#endif
#if defined(LN) || defined(RT)
mov A, AORIG
#else
mov A, AO
#endif
#ifndef RT
add C, LDC, C
#endif
sra M, 2, I
cmp I, 0
ble,pn %icc, .LL250
nop
.LL221:
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 2 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 2 + BASE_SHIFT, TEMP1
sll KK, 0 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c01
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c02
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t2
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c03
LDF [BO + 2 * SIZE], b3
FMOV FZERO, t3
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c04
LDF [BO + 3 * SIZE], b4
FMOV FZERO, t4
ble,pn %icc, .LL225
prefetch [C1 + 4 * SIZE], 2
.LL222:
FADD c01, t1, c01
add BO, 4 * SIZE, BO
FMUL a1, b1, t1
LDF [AO + 4 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b1, t2
LDF [AO + 5 * SIZE], a2
FADD c03, t3, c03
add L, -1, L
FMUL a3, b1, t3
LDF [AO + 6 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b1, t4
LDF [AO + 7 * SIZE], a4
LDF [BO + 0 * SIZE], b1
FADD c01, t1, c01
cmp L, 0
FMUL a1, b2, t1
LDF [AO + 8 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b2, t2
LDF [AO + 9 * SIZE], a2
FADD c03, t3, c03
FMUL a3, b2, t3
LDF [AO + 10 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b2, t4
LDF [AO + 11 * SIZE], a4
LDF [BO + 1 * SIZE], b2
FADD c01, t1, c01
FMUL a1, b3, t1
LDF [AO + 12 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b3, t2
LDF [AO + 13 * SIZE], a2
FADD c03, t3, c03
FMUL a3, b3, t3
LDF [AO + 14 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b3, t4
LDF [AO + 15 * SIZE], a4
LDF [BO + 2 * SIZE], b3
FADD c01, t1, c01
FMUL a1, b4, t1
LDF [AO + 16 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b4, t2
LDF [AO + 17 * SIZE], a2
FADD c03, t3, c03
FMUL a3, b4, t3
LDF [AO + 18 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b4, t4
LDF [AO + 19 * SIZE], a4
add AO, 16 * SIZE, AO
bg,pt %icc, .LL222
LDF [BO + 3 * SIZE], b4
.LL225:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL229
nop
.LL226:
FADD c01, t1, c01
add BO, 1 * SIZE, BO
FMUL a1, b1, t1
LDF [AO + 4 * SIZE], a1
FADD c02, t2, c02
add L, -1, L
FMUL a2, b1, t2
LDF [AO + 5 * SIZE], a2
FADD c03, t3, c03
cmp L, 0
FMUL a3, b1, t3
LDF [AO + 6 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b1, t4
LDF [AO + 7 * SIZE], a4
add AO, 4 * SIZE, AO
bg,pt %icc, .LL226
LDF [BO + 0 * SIZE], b1
.LL229:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 4, TEMP1
#else
sub KK, 1, TEMP1
#endif
sll TEMP1, 2 + BASE_SHIFT, TEMP2
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
#endif
#ifdef LN
LDF [AO + 15 * SIZE], a1
LDF [AO + 14 * SIZE], a2
LDF [AO + 13 * SIZE], a3
LDF [AO + 12 * SIZE], a4
FMUL a1, c04, c04
FMUL a2, c04, t1
FSUB c03, t1, c03
FMUL a3, c04, t1
FSUB c02, t1, c02
FMUL a4, c04, t1
FSUB c01, t1, c01
LDF [AO + 10 * SIZE], a1
LDF [AO + 9 * SIZE], a2
LDF [AO + 8 * SIZE], a3
FMUL a1, c03, c03
FMUL a2, c03, t1
FSUB c02, t1, c02
FMUL a3, c03, t1
FSUB c01, t1, c01
LDF [AO + 5 * SIZE], a1
LDF [AO + 4 * SIZE], a2
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c01, t1, c01
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a2, c01, t1
FSUB c02, t1, c02
FMUL a3, c01, t1
FSUB c03, t1, c03
FMUL a4, c01, t1
FSUB c04, t1, c04
LDF [AO + 5 * SIZE], a1
LDF [AO + 6 * SIZE], a2
LDF [AO + 7 * SIZE], a3
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c03, t1, c03
FMUL a3, c02, t1
FSUB c04, t1, c04
LDF [AO + 10 * SIZE], a1
LDF [AO + 11 * SIZE], a2
FMUL a1, c03, c03
FMUL a2, c03, t1
FSUB c04, t1, c04
LDF [AO + 15 * SIZE], a1
FMUL a1, c04, c04
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
#endif
#ifdef RT
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
#endif
#ifdef LN
add C1, -4 * SIZE, C1
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c02, [BO + 1 * SIZE]
STF c03, [BO + 2 * SIZE]
STF c04, [BO + 3 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
STF c03, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C1 + 1 * SIZE]
STF c03, [C1 + 2 * SIZE]
STF c04, [C1 + 3 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 4 * SIZE, C1
#endif
#ifdef RT
sll K, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 2 + BASE_SHIFT, TEMP2
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 4, KK
#endif
#ifdef LN
sub KK, 4, KK
#endif
add I, -1, I
cmp I, 0
bg,pt %icc, .LL221
nop
.LL250:
and M, 2, I
cmp I, 0
ble,pn %icc, .LL270
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 1 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 1 + BASE_SHIFT, TEMP1
sll KK, 0 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c01
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c02
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t2
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c03
LDF [BO + 2 * SIZE], b3
FMOV FZERO, t3
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c04
LDF [BO + 3 * SIZE], b4
FMOV FZERO, t4
ble,pn %icc, .LL255
nop
.LL252:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
LDF [AO + 4 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b1, t2
LDF [AO + 5 * SIZE], a2
LDF [BO + 4 * SIZE], b1
FADD c03, t3, c03
cmp L, 0
FMUL a3, b2, t3
LDF [AO + 6 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b2, t4
LDF [AO + 7 * SIZE], a4
LDF [BO + 5 * SIZE], b2
FADD c01, t1, c01
FMUL a1, b3, t1
LDF [AO + 8 * SIZE], a1
FADD c02, t2, c02
FMUL a2, b3, t2
LDF [AO + 9 * SIZE], a2
LDF [BO + 6 * SIZE], b3
FADD c03, t3, c03
FMUL a3, b4, t3
LDF [AO + 10 * SIZE], a3
FADD c04, t4, c04
FMUL a4, b4, t4
LDF [AO + 11 * SIZE], a4
add AO, 8 * SIZE, AO
LDF [BO + 7 * SIZE], b4
bg,pt %icc, .LL252
add BO, 4 * SIZE, BO
.LL255:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL259
nop
.LL256:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
LDF [AO + 2 * SIZE], a1
FADD c02, t2, c02
cmp L, 0
FMUL a2, b1, t2
LDF [AO + 3 * SIZE], a2
LDF [BO + 1 * SIZE], b1
add AO, 2 * SIZE, AO
bg,pt %icc, .LL256
add BO, 1 * SIZE, BO
.LL259:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
FADD c01, c03, c01
FADD c02, c04, c02
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 2, TEMP1
#else
sub KK, 1, TEMP1
#endif
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
FSUB a1, c01, c01
FSUB a2, c02, c02
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
FSUB a1, c01, c01
FSUB a2, c02, c02
#endif
#ifdef LN
LDF [AO + 3 * SIZE], a1
LDF [AO + 2 * SIZE], a2
LDF [AO + 0 * SIZE], a3
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c01, t1, c01
FMUL a3, c01, c01
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a2, c01, t1
FSUB c02, t1, c02
FMUL a3, c02, c02
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
#endif
#ifdef RT
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
#endif
#ifdef LN
add C1, -2 * SIZE, C1
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c02, [BO + 1 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C1 + 1 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 2 * SIZE, C1
#endif
#ifdef RT
sll K, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 2, KK
#endif
#ifdef LN
sub KK, 2, KK
#endif
.LL270:
and M, 1, I
cmp I, 0
ble,pn %icc, .LL299
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 0 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AO
add B, TEMP1, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c01
LDF [AO + 2 * SIZE], a3
FMOV FZERO, t2
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c02
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t3
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t4
LDF [BO + 2 * SIZE], b3
ble,pn %icc, .LL275
LDF [BO + 3 * SIZE], b4
.LL272:
FADD c01, t1, c01
add L, -1, L
add AO, 4 * SIZE, AO
FMUL a1, b1, t1
add BO, 4 * SIZE, BO
LDF [AO + 0 * SIZE], a1
FADD c02, t2, c02
cmp L, 0
LDF [BO + 0 * SIZE], b1
FMUL a2, b2, t2
LDF [AO + 1 * SIZE], a2
FADD c01, t3, c01
LDF [BO + 1 * SIZE], b2
FMUL a3, b3, t3
LDF [AO + 2 * SIZE], a3
FADD c02, t4, c02
LDF [BO + 2 * SIZE], b3
FMUL a4, b4, t4
LDF [AO + 3 * SIZE], a4
bg,pt %icc, .LL272
LDF [BO + 3 * SIZE], b4
.LL275:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL279
nop
.LL276:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
LDF [AO + 1 * SIZE], a1
LDF [BO + 1 * SIZE], b1
add BO, 1 * SIZE, BO
cmp L, 0
bg,pt %icc, .LL276
add AO, 1 * SIZE, AO
.LL279:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c01, t3, c01
FADD c02, t4, c02
FADD c01, c02, c01
#if defined(LN) || defined(RT)
sub KK, 1, TEMP1
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
FSUB a1, c01, c01
#else
LDF [AO + 0 * SIZE], a1
FSUB a1, c01, c01
#endif
#ifdef LN
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef RT
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef LN
add C1, -1 * SIZE, C1
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 1 * SIZE, C1
#endif
#ifdef RT
sll K, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 0 + BASE_SHIFT, TEMP1
add AO, TEMP1, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 1, KK
#endif
#ifdef LN
sub KK, 1, KK
#endif
.LL299:
#ifdef LN
sll K, 0 + BASE_SHIFT, TEMP1
add B, TEMP1, B
#endif
#if defined(LT) || defined(RN)
mov BO, B
#endif
#ifdef RN
add KK, 1, KK
#endif
#ifdef RT
sub KK, 1, KK
#endif
.LL100: /* n & 2 */
and N, 2, J
cmp J, 0
ble,pn %icc, .LL200
nop
#ifdef RT
sll K, 1 + BASE_SHIFT, TEMP1
sub B, TEMP1, B
sll LDC, 1, TEMP1
sub C, TEMP1, C
#endif
mov C, C1
add C, LDC, C2
#ifdef LN
add M, OFFSET, KK
#endif
#ifdef LT
mov OFFSET, KK
#endif
#if defined(LN) || defined(RT)
mov A, AORIG
#else
mov A, AO
#endif
#ifndef RT
add C2, LDC, C
#endif
sra M, 2, I
cmp I, 0
ble,pn %icc, .LL150
FMOV FZERO, c03
.LL121:
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 2 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 2 + BASE_SHIFT, TEMP1
sll KK, 1 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, t1
LDF [BO + 0 * SIZE], b1
FMOV FZERO, c07
LDF [AO + 1 * SIZE], a2
FMOV FZERO, t2
LDF [BO + 1 * SIZE], b2
FMOV FZERO, c04
LDF [AO + 2 * SIZE], a3
FMOV FZERO, t3
LDF [BO + 2 * SIZE], b3
FMOV FZERO, c08
LDF [AO + 3 * SIZE], a4
FMOV FZERO, t4
LDF [BO + 3 * SIZE], b4
FMOV FZERO, c01
prefetch [C1 + 3 * SIZE], 2
FMOV FZERO, c05
prefetch [C2 + 3 * SIZE], 2
FMOV FZERO, c02
ble,pn %icc, .LL125
FMOV FZERO, c06
.LL122:
FADD c03, t1, c03
add L, -1, L
FMUL a1, b1, t1
prefetch [AO + APREFETCHSIZE * SIZE], 0
FADD c07, t2, c07
add BO, 8 * SIZE, BO
FMUL a1, b2, t2
LDF [AO + 4 * SIZE], a1
FADD c04, t3, c04
add AO, 16 * SIZE, AO
FMUL a2, b1, t3
cmp L, 0
FADD c08, t4, c08
nop
FMUL a2, b2, t4
LDF [AO - 11 * SIZE], a2
FADD c01, t1, c01
nop
FMUL a3, b1, t1
nop
FADD c05, t2, c05
nop
FMUL a3, b2, t2
LDF [AO - 10 * SIZE], a3
FADD c02, t3, c02
nop
FMUL a4, b1, t3
LDF [BO - 4 * SIZE], b1
FADD c06, t4, c06
nop
FMUL a4, b2, t4
LDF [BO - 3 * SIZE], b2
FADD c03, t1, c03
nop
FMUL a1, b3, t1
LDF [AO - 9 * SIZE], a4
FADD c07, t2, c07
nop
FMUL a1, b4, t2
LDF [AO - 8 * SIZE], a1
FADD c04, t3, c04
nop
FMUL a2, b3, t3
nop
FADD c08, t4, c08
nop
FMUL a2, b4, t4
LDF [AO - 7 * SIZE], a2
FADD c01, t1, c01
nop
FMUL a3, b3, t1
nop
FADD c05, t2, c05
nop
FMUL a3, b4, t2
LDF [AO - 6 * SIZE], a3
FADD c02, t3, c02
nop
FMUL a4, b3, t3
LDF [BO - 2 * SIZE], b3
FADD c06, t4, c06
nop
FMUL a4, b4, t4
LDF [BO - 1 * SIZE], b4
FADD c03, t1, c03
nop
FMUL a1, b1, t1
LDF [AO - 5 * SIZE], a4
FADD c07, t2, c07
nop
FMUL a1, b2, t2
LDF [AO - 4 * SIZE], a1
FADD c04, t3, c04
nop
FMUL a2, b1, t3
nop
FADD c08, t4, c08
nop
FMUL a2, b2, t4
LDF [AO - 3 * SIZE], a2
FADD c01, t1, c01
nop
FMUL a3, b1, t1
nop
FADD c05, t2, c05
nop
FMUL a3, b2, t2
LDF [AO - 2 * SIZE], a3
FADD c02, t3, c02
nop
FMUL a4, b1, t3
LDF [BO + 0 * SIZE], b1
FADD c06, t4, c06
nop
FMUL a4, b2, t4
LDF [BO + 1 * SIZE], b2
FADD c03, t1, c03
nop
FMUL a1, b3, t1
LDF [AO - 1 * SIZE], a4
FADD c07, t2, c07
nop
FMUL a1, b4, t2
LDF [AO + 0 * SIZE], a1
FADD c04, t3, c04
nop
FMUL a2, b3, t3
nop
FADD c08, t4, c08
nop
FMUL a2, b4, t4
LDF [AO + 1 * SIZE], a2
FADD c01, t1, c01
nop
FMUL a3, b3, t1
nop
FADD c05, t2, c05
nop
FMUL a3, b4, t2
LDF [AO + 2 * SIZE], a3
FADD c02, t3, c02
nop
FMUL a4, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c06, t4, c06
FMUL a4, b4, t4
LDF [AO + 3 * SIZE], a4
bg,pt %icc, .LL122
LDF [BO + 3 * SIZE], b4
.LL125:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL129
nop
.LL126:
FADD c03, t1, c03
add AO, 4 * SIZE, AO
FMUL a1, b1, t1
add BO, 2 * SIZE, BO
FADD c07, t2, c07
add L, -1, L
FMUL a1, b2, t2
LDF [AO + 0 * SIZE], a1
FADD c04, t3, c04
cmp L, 0
FMUL a2, b1, t3
FADD c08, t4, c08
FMUL a2, b2, t4
LDF [AO + 1 * SIZE], a2
FADD c01, t1, c01
FMUL a3, b1, t1
FADD c05, t2, c05
FMUL a3, b2, t2
LDF [AO + 2 * SIZE], a3
FADD c02, t3, c02
FMUL a4, b1, t3
LDF [BO + 0 * SIZE], b1
FADD c06, t4, c06
FMUL a4, b2, t4
LDF [BO + 1 * SIZE], b2
bg,pt %icc, .LL126
LDF [AO + 3 * SIZE], a4
.LL129:
FADD c03, t1, c03
FADD c07, t2, c07
FADD c04, t3, c04
FADD c08, t4, c08
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 4, TEMP1
#else
sub KK, 2, TEMP1
#endif
sll TEMP1, 2 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
LDF [BO + 4 * SIZE], b1
LDF [BO + 5 * SIZE], b2
LDF [BO + 6 * SIZE], b3
LDF [BO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c05, c05
FSUB a3, c02, c02
FSUB a4, c06, c06
FSUB b1, c03, c03
FSUB b2, c07, c07
FSUB b3, c04, c04
FSUB b4, c08, c08
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
LDF [AO + 4 * SIZE], b1
LDF [AO + 5 * SIZE], b2
LDF [AO + 6 * SIZE], b3
LDF [AO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
FSUB b1, c05, c05
FSUB b2, c06, c06
FSUB b3, c07, c07
FSUB b4, c08, c08
#endif
#ifdef LN
LDF [AO + 15 * SIZE], a1
LDF [AO + 14 * SIZE], a2
LDF [AO + 13 * SIZE], a3
LDF [AO + 12 * SIZE], a4
FMUL a1, c04, c04
FMUL a1, c08, c08
FMUL a2, c04, t1
FMUL a2, c08, t2
FSUB c03, t1, c03
FSUB c07, t2, c07
FMUL a3, c04, t1
FMUL a3, c08, t2
FSUB c02, t1, c02
FSUB c06, t2, c06
FMUL a4, c04, t1
FMUL a4, c08, t2
FSUB c01, t1, c01
FSUB c05, t2, c05
LDF [AO + 10 * SIZE], a1
LDF [AO + 9 * SIZE], a2
LDF [AO + 8 * SIZE], a3
FMUL a1, c03, c03
FMUL a1, c07, c07
FMUL a2, c03, t1
FMUL a2, c07, t2
FSUB c02, t1, c02
FSUB c06, t2, c06
FMUL a3, c03, t1
FMUL a3, c07, t2
FSUB c01, t1, c01
FSUB c05, t2, c05
LDF [AO + 5 * SIZE], a1
LDF [AO + 4 * SIZE], a2
FMUL a1, c02, c02
FMUL a1, c06, c06
FMUL a2, c02, t1
FMUL a2, c06, t2
FSUB c01, t1, c01
FSUB c05, t2, c05
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c05, c05
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a1, c05, c05
FMUL a2, c01, t1
FMUL a2, c05, t2
FSUB c02, t1, c02
FSUB c06, t2, c06
FMUL a3, c01, t1
FMUL a3, c05, t2
FSUB c03, t1, c03
FSUB c07, t2, c07
FMUL a4, c01, t1
FMUL a4, c05, t2
FSUB c04, t1, c04
FSUB c08, t2, c08
LDF [AO + 5 * SIZE], a1
LDF [AO + 6 * SIZE], a2
LDF [AO + 7 * SIZE], a3
FMUL a1, c02, c02
FMUL a1, c06, c06
FMUL a2, c02, t1
FMUL a2, c06, t2
FSUB c03, t1, c03
FSUB c07, t2, c07
FMUL a3, c02, t1
FMUL a3, c06, t2
FSUB c04, t1, c04
FSUB c08, t2, c08
LDF [AO + 10 * SIZE], a1
LDF [AO + 11 * SIZE], a2
FMUL a1, c03, c03
FMUL a1, c07, c07
FMUL a2, c03, t1
FMUL a2, c07, t2
FSUB c04, t1, c04
FSUB c08, t2, c08
LDF [AO + 15 * SIZE], a1
FMUL a1, c04, c04
FMUL a1, c08, c08
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
FMUL a2, c01, t1
FMUL a2, c02, t2
FMUL a2, c03, t3
FMUL a2, c04, t4
FSUB c05, t1, c05
FSUB c06, t2, c06
FSUB c07, t3, c07
FSUB c08, t4, c08
FMUL a3, c05, c05
FMUL a3, c06, c06
FMUL a3, c07, c07
FMUL a3, c08, c08
#endif
#ifdef RT
LDF [BO + 3 * SIZE], a1
LDF [BO + 2 * SIZE], a2
LDF [BO + 0 * SIZE], a3
FMUL a1, c05, c05
FMUL a1, c06, c06
FMUL a1, c07, c07
FMUL a1, c08, c08
FMUL a2, c05, t1
FMUL a2, c06, t2
FMUL a2, c07, t3
FMUL a2, c08, t4
FSUB c01, t1, c01
FSUB c02, t2, c02
FSUB c03, t3, c03
FSUB c04, t4, c04
FMUL a3, c01, c01
FMUL a3, c02, c02
FMUL a3, c03, c03
FMUL a3, c04, c04
#endif
#ifdef LN
add C1, -4 * SIZE, C1
add C2, -4 * SIZE, C2
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c05, [BO + 1 * SIZE]
STF c02, [BO + 2 * SIZE]
STF c06, [BO + 3 * SIZE]
STF c03, [BO + 4 * SIZE]
STF c07, [BO + 5 * SIZE]
STF c04, [BO + 6 * SIZE]
STF c08, [BO + 7 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
STF c03, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
STF c05, [AO + 4 * SIZE]
STF c06, [AO + 5 * SIZE]
STF c07, [AO + 6 * SIZE]
STF c08, [AO + 7 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C1 + 1 * SIZE]
STF c03, [C1 + 2 * SIZE]
STF c04, [C1 + 3 * SIZE]
STF c05, [C2 + 0 * SIZE]
STF c06, [C2 + 1 * SIZE]
STF c07, [C2 + 2 * SIZE]
STF c08, [C2 + 3 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 4 * SIZE, C1
add C2, 4 * SIZE, C2
#endif
#ifdef RT
sll K, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 2 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 4, KK
#endif
#ifdef LN
sub KK, 4, KK
#endif
add I, -1, I
cmp I, 0
bg,pt %icc, .LL121
FMOV FZERO, c03
.LL150:
and M, 2, I
cmp I, 0
ble,pn %icc, .LL170
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 1 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 1 + BASE_SHIFT, TEMP1
sll KK, 1 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c01
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
cmp L, 0
FMOV FZERO, c02
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t2
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c03
LDF [BO + 2 * SIZE], b3
FMOV FZERO, t3
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c04
LDF [BO + 3 * SIZE], b4
FMOV FZERO, t4
ble,pn %icc, .LL155
nop
.LL152:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
prefetch [AO + APREFETCHSIZE * SIZE], 0
FADD c02, t2, c02
add BO, 8 * SIZE, BO
FMUL a1, b2, t2
LDF [AO + 4 * SIZE], a1
FADD c03, t3, c03
cmp L, 0
FMUL a2, b1, t3
LDF [BO - 4 * SIZE], b1
FADD c04, t4, c04
nop
FMUL a2, b2, t4
LDF [AO + 5 * SIZE], a2
FADD c01, t1, c01
nop
FMUL a3, b3, t1
LDF [BO - 3 * SIZE], b2
FADD c02, t2, c02
nop
FMUL a3, b4, t2
LDF [AO + 6 * SIZE], a3
FADD c03, t3, c03
nop
FMUL a4, b3, t3
LDF [BO - 2 * SIZE], b3
FADD c04, t4, c04
nop
FMUL a4, b4, t4
LDF [AO + 7 * SIZE], a4
FADD c01, t1, c01
nop
FMUL a1, b1, t1
LDF [BO - 1 * SIZE], b4
FADD c02, t2, c02
FMUL a1, b2, t2
LDF [AO + 8 * SIZE], a1
FADD c03, t3, c03
FMUL a2, b1, t3
LDF [BO + 0 * SIZE], b1
FADD c04, t4, c04
FMUL a2, b2, t4
LDF [AO + 9 * SIZE], a2
FADD c01, t1, c01
FMUL a3, b3, t1
LDF [BO + 1 * SIZE], b2
FADD c02, t2, c02
FMUL a3, b4, t2
LDF [AO + 10 * SIZE], a3
FADD c03, t3, c03
FMUL a4, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c04, t4, c04
FMUL a4, b4, t4
LDF [AO + 11 * SIZE], a4
add AO, 8 * SIZE, AO
bg,pt %icc, .LL152
LDF [BO + 3 * SIZE], b4
.LL155:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL159
nop
.LL156:
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [BO + 0 * SIZE], b1
LDF [BO + 1 * SIZE], b2
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
FMUL a1, b1, t1
FMUL a1, b2, t2
FMUL a2, b1, t3
FMUL a2, b2, t4
add AO, 2 * SIZE, AO
add BO, 2 * SIZE, BO
add L, -1, L
cmp L, 0
bg,pt %icc, .LL156
nop
.LL159:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 2, TEMP1
#else
sub KK, 2, TEMP1
#endif
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c03, c03
FSUB a3, c02, c02
FSUB a4, c04, c04
#endif
#ifdef LN
LDF [AO + 3 * SIZE], a1
LDF [AO + 2 * SIZE], a2
LDF [AO + 0 * SIZE], a3
FMUL a1, c03, c03
FMUL a1, c04, c04
FMUL a2, c03, t1
FMUL a2, c04, t2
FSUB c01, t1, c01
FSUB c02, t2, c02
FMUL a3, c01, c01
FMUL a3, c02, c02
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a2, c01, t1
FMUL a2, c02, t2
FSUB c03, t1, c03
FSUB c04, t2, c04
FMUL a3, c03, c03
FMUL a3, c04, c04
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a1, c03, c03
FMUL a2, c01, t1
FMUL a2, c03, t2
FSUB c02, t1, c02
FSUB c04, t2, c04
FMUL a3, c02, c02
FMUL a3, c04, c04
#endif
#ifdef RT
LDF [BO + 3 * SIZE], a1
LDF [BO + 2 * SIZE], a2
LDF [BO + 0 * SIZE], a3
FMUL a1, c02, c02
FMUL a1, c04, c04
FMUL a2, c02, t1
FMUL a2, c04, t2
FSUB c01, t1, c01
FSUB c03, t2, c03
FMUL a3, c01, c01
FMUL a3, c03, c03
#endif
#ifdef LN
add C1, -2 * SIZE, C1
add C2, -2 * SIZE, C2
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c02, [BO + 1 * SIZE]
STF c03, [BO + 2 * SIZE]
STF c04, [BO + 3 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c03, [AO + 1 * SIZE]
STF c02, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c03, [C1 + 1 * SIZE]
STF c02, [C2 + 0 * SIZE]
STF c04, [C2 + 1 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 2 * SIZE, C1
add C2, 2 * SIZE, C2
#endif
#ifdef RT
sll K, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 2, KK
#endif
#ifdef LN
sub KK, 2, KK
#endif
.LL170:
and M, 1, I
cmp I, 0
ble,pn %icc, .LL199
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 0 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 0 + BASE_SHIFT, TEMP1
sll KK, 1 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c01
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c02
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t2
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c03
LDF [BO + 2 * SIZE], b3
FMOV FZERO, t3
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c04
LDF [BO + 3 * SIZE], b4
FMOV FZERO, t4
ble,pn %icc, .LL175
nop
.LL172:
FADD c01, t1, c01
add AO, 4 * SIZE, AO
FMUL a1, b1, t1
LDF [BO + 4 * SIZE], b1
FADD c02, t2, c02
FMUL a1, b2, t2
LDF [BO + 5 * SIZE], b2
add L, -1, L
LDF [AO + 0 * SIZE], a1
FADD c03, t3, c03
cmp L, 0
FMUL a2, b3, t3
LDF [BO + 6 * SIZE], b3
FADD c04, t4, c04
FMUL a2, b4, t4
LDF [BO + 7 * SIZE], b4
LDF [AO + 1 * SIZE], a2
FADD c01, t1, c01
FMUL a3, b1, t1
LDF [BO + 8 * SIZE], b1
FADD c02, t2, c02
FMUL a3, b2, t2
LDF [BO + 9 * SIZE], b2
LDF [AO + 2 * SIZE], a3
FADD c03, t3, c03
FMUL a4, b3, t3
LDF [BO + 10 * SIZE], b3
FADD c04, t4, c04
FMUL a4, b4, t4
LDF [BO + 11 * SIZE], b4
add BO, 8 * SIZE, BO
bg,pt %icc, .LL172
LDF [AO + 3 * SIZE], a4
.LL175:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL179
nop
.LL176:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
add AO, 1 * SIZE, AO
LDF [BO + 2 * SIZE], b1
FADD c02, t2, c02
cmp L, 0
FMUL a1, b2, t2
LDF [BO + 3 * SIZE], b2
add BO, 2 * SIZE, BO
bg,pt %icc, .LL176
LDF [AO + 0 * SIZE], a1
.LL179:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
FADD c01, c03, c01
FADD c02, c04, c02
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 1, TEMP1
#else
sub KK, 2, TEMP1
#endif
sll TEMP1, 0 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
FSUB a1, c01, c01
FSUB a2, c02, c02
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
FSUB a1, c01, c01
FSUB a2, c02, c02
#endif
#ifdef LN
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a2, c01, t1
FSUB c02, t1, c02
FMUL a3, c02, c02
#endif
#ifdef RT
LDF [BO + 3 * SIZE], a1
LDF [BO + 2 * SIZE], a2
LDF [BO + 0 * SIZE], a3
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c01, t1, c01
FMUL a3, c01, c01
#endif
#ifdef LN
add C1, -1 * SIZE, C1
add C2, -1 * SIZE, C2
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c02, [BO + 1 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C2 + 0 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 1 * SIZE, C1
add C2, 1 * SIZE, C2
#endif
#ifdef RT
sll K, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 0 + BASE_SHIFT, TEMP2
sll TEMP1, 1 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 1, KK
#endif
#ifdef LN
sub KK, 1, KK
#endif
.LL199:
#ifdef LN
sll K, 1 + BASE_SHIFT, TEMP1
add B, TEMP1, B
#endif
#if defined(LT) || defined(RN)
mov BO, B
#endif
#ifdef RN
add KK, 2, KK
#endif
#ifdef RT
sub KK, 2, KK
#endif
.LL200:
sra N, 2, J
cmp J, 0
ble,pn %icc, .LL999
nop
.LL11:
#ifdef RT
sll K, 2 + BASE_SHIFT, TEMP1
sub B, TEMP1, B
sll LDC, 2, TEMP1
sub C, TEMP1, C
#endif
add C, LDC, C2
FMOV FZERO, t1
nop
mov C, C1
add C2, LDC, C3
FMOV FZERO, t2
nop
mov A, AO
sra M, 2, I
add C3, LDC, C4
FMOV FZERO, t3
#ifdef LN
add M, OFFSET, KK
#endif
#ifdef LT
mov OFFSET, KK
#endif
#if defined(LN) || defined(RT)
mov A, AORIG
#else
mov A, AO
#endif
cmp I, 0
#ifndef RT
add C4, LDC, C
#endif
FMOV FZERO, t4
ble,pn %icc, .LL50
FMOV FZERO, c01
.LL21:
FMOV FZERO, c02
FMOV FZERO, c03
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 2 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AO
add B, TEMP1, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c04
LDF [BO + 0 * SIZE], b1
FMOV FZERO, c05
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c06
LDF [BO + 1 * SIZE], b2
FMOV FZERO, c07
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c08
LDF [BO + 2 * SIZE], b3
FMOV FZERO, c09
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c10
LDF [BO + 3 * SIZE], b4
FMOV FZERO, c11
LDF [BO + 4 * SIZE], b5 /* ***** */
LDF [AO + 4 * SIZE], a5 /* ***** */
prefetch [C1 + 3 * SIZE], 3
FMOV FZERO, c12
prefetch [C2 + 3 * SIZE], 3
FMOV FZERO, c13
prefetch [C3 + 3 * SIZE], 3
FMOV FZERO, c14
prefetch [C4 + 3 * SIZE], 3
FMOV FZERO, c15
ble,pn %icc, .LL25
FMOV FZERO, c16
.LL22:
FADD c04, t1, c04
prefetch [AO + APREFETCHSIZE * SIZE], APREFETCH_CATEGORY
FMUL a1, b1, t1
nop
FADD c08, t2, c08
prefetch [BO + BPREFETCHSIZE * SIZE], BPREFETCH_CATEGORY
FMUL a1, b2, t2
add AO, 16 * SIZE, AO
FADD c12, t3, c12
LDF [AO - 13 * SIZE], a4
FMUL a1, b3, t3
add BO, 16 * SIZE, BO
FADD c16, t4, c16
nop
FMUL a1, b4, t4
LDF [AO - 8 * SIZE], a1
FADD c01, t1, c01
nop
FMUL a2, b1, t1
nop
FADD c05, t2, c05
nop
FMUL a2, b2, t2
nop
FADD c09, t3, c09
nop
FMUL a2, b3, t3
nop
FADD c13, t4, c13
add L, -1, L
FMUL a2, b4, t4
LDF [AO - 11 * SIZE], a2
FADD c02, t1, c02
nop
FMUL a3, b1, t1
nop
FADD c06, t2, c06
nop
FMUL a3, b2, t2
nop
FADD c10, t3, c10
nop
FMUL a3, b3, t3
nop
FADD c14, t4, c14
nop
FMUL a3, b4, t4
LDF [AO - 10 * SIZE], a3
FADD c03, t1, c03
nop
FMUL a4, b1, t1
LDF [BO - 8 * SIZE], b1
FADD c07, t2, c07
nop
FMUL a4, b2, t2
LDF [BO - 11 * SIZE], b2
FADD c11, t3, c11
nop
FMUL a4, b3, t3
LDF [BO - 10 * SIZE], b3
FADD c15, t4, c15
nop
FMUL a4, b4, t4
LDF [BO - 9 * SIZE], b4
FADD c04, t1, c04
nop
FMUL a5, b5, t1
LDF [AO - 9 * SIZE], a4
FADD c08, t2, c08
nop
FMUL a5, b2, t2
nop
FADD c12, t3, c12
nop
FMUL a5, b3, t3
nop
FADD c16, t4, c16
nop
FMUL a5, b4, t4
LDF [AO - 4 * SIZE], a5
FADD c01, t1, c01
nop
FMUL a2, b5, t1
nop
FADD c05, t2, c05
nop
FMUL a2, b2, t2
nop
FADD c09, t3, c09
nop
FMUL a2, b3, t3
nop
FADD c13, t4, c13
nop
FMUL a2, b4, t4
LDF [AO - 7 * SIZE], a2
FADD c02, t1, c02
nop
FMUL a3, b5, t1
nop
FADD c06, t2, c06
nop
FMUL a3, b2, t2
nop
FADD c10, t3, c10
nop
FMUL a3, b3, t3
nop
FADD c14, t4, c14
nop
FMUL a3, b4, t4
LDF [AO - 6 * SIZE], a3
FADD c03, t1, c03
nop
FMUL a4, b5, t1
LDF [BO - 4 * SIZE], b5
FADD c07, t2, c07
nop
FMUL a4, b2, t2
LDF [BO - 7 * SIZE], b2
FADD c11, t3, c11
nop
FMUL a4, b3, t3
LDF [BO - 6 * SIZE], b3
FADD c15, t4, c15
nop
FMUL a4, b4, t4
LDF [BO - 5 * SIZE], b4
FADD c04, t1, c04
nop
FMUL a1, b1, t1
LDF [AO - 5 * SIZE], a4
FADD c08, t2, c08
nop
FMUL a1, b2, t2
nop
FADD c12, t3, c12
nop
FMUL a1, b3, t3
nop
FADD c16, t4, c16
nop
FMUL a1, b4, t4
LDF [AO - 0 * SIZE], a1
FADD c01, t1, c01
nop
FMUL a2, b1, t1
nop
#ifdef DOUBLE
prefetch [AO + (APREFETCHSIZE + 8) * SIZE], APREFETCH_CATEGORY
#else
nop
#endif
FADD c05, t2, c05
nop
FMUL a2, b2, t2
FADD c09, t3, c09
nop
FMUL a2, b3, t3
nop
FADD c13, t4, c13
nop
FMUL a2, b4, t4
nop
FADD c02, t1, c02
nop
FMUL a3, b1, t1
LDF [AO - 3 * SIZE], a2
FADD c06, t2, c06
#ifdef DOUBLE
prefetch [BO + (BPREFETCHSIZE + 8) * SIZE], BPREFETCH_CATEGORY
#else
nop
#endif
FMUL a3, b2, t2
nop
FADD c10, t3, c10
nop
FMUL a3, b3, t3
nop
FADD c14, t4, c14
nop
FMUL a3, b4, t4
LDF [AO - 2 * SIZE], a3
FADD c03, t1, c03
nop
FMUL a4, b1, t1
LDF [BO - 0 * SIZE], b1
FADD c07, t2, c07
nop
FMUL a4, b2, t2
LDF [BO - 3 * SIZE], b2
FADD c11, t3, c11
nop
FMUL a4, b3, t3
LDF [BO - 2 * SIZE], b3
FADD c15, t4, c15
nop
FMUL a4, b4, t4
LDF [BO - 1 * SIZE], b4
FADD c04, t1, c04
nop
FMUL a5, b5, t1
LDF [AO - 1 * SIZE], a4
FADD c08, t2, c08
FMUL a5, b2, t2
FADD c12, t3, c12
FMUL a5, b3, t3
FADD c16, t4, c16
nop
FMUL a5, b4, t4
LDF [AO + 4 * SIZE], a5
FADD c01, t1, c01
nop
FMUL a2, b5, t1
nop
FADD c05, t2, c05
nop
FMUL a2, b2, t2
nop
FADD c09, t3, c09
nop
FMUL a2, b3, t3
nop
FADD c13, t4, c13
nop
FMUL a2, b4, t4
LDF [AO + 1 * SIZE], a2
FADD c02, t1, c02
nop
FMUL a3, b5, t1
nop
FADD c06, t2, c06
nop
FMUL a3, b2, t2
nop
FADD c10, t3, c10
nop
FMUL a3, b3, t3
nop
FADD c14, t4, c14
nop
FMUL a3, b4, t4
LDF [AO + 2 * SIZE], a3
FADD c03, t1, c03
cmp L, 0
FMUL a4, b5, t1
LDF [BO + 4 * SIZE], b5
FADD c07, t2, c07
nop
FMUL a4, b2, t2
LDF [BO + 1 * SIZE], b2
FADD c11, t3, c11
nop
FMUL a4, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c15, t4, c15
FMUL a4, b4, t4
bg,pt %icc, .LL22
LDF [BO + 3 * SIZE], b4
.LL25:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL29
nop
.LL26:
FADD c04, t1, c04
LDF [AO + 3 * SIZE], a4
FMUL a1, b1, t1
add AO, 4 * SIZE, AO
FADD c08, t2, c08
add BO, 4 * SIZE, BO
FMUL a1, b2, t2
add L, -1, L
FADD c12, t3, c12
nop
FMUL a1, b3, t3
cmp L, 0
FADD c16, t4, c16
nop
FMUL a1, b4, t4
LDF [AO + 0 * SIZE], a1
FADD c01, t1, c01
nop
FMUL a2, b1, t1
nop
FADD c05, t2, c05
nop
FMUL a2, b2, t2
nop
FADD c09, t3, c09
nop
FMUL a2, b3, t3
nop
FADD c13, t4, c13
nop
FMUL a2, b4, t4
LDF [AO + 1 * SIZE], a2
FADD c02, t1, c02
nop
FMUL a3, b1, t1
nop
FADD c06, t2, c06
nop
FMUL a3, b2, t2
nop
FADD c10, t3, c10
nop
FMUL a3, b3, t3
nop
FADD c14, t4, c14
nop
FMUL a3, b4, t4
LDF [AO + 2 * SIZE], a3
FADD c03, t1, c03
nop
FMUL a4, b1, t1
LDF [BO + 0 * SIZE], b1
FADD c07, t2, c07
nop
FMUL a4, b2, t2
LDF [BO + 1 * SIZE], b2
FADD c11, t3, c11
nop
FMUL a4, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c15, t4, c15
FMUL a4, b4, t4
bg,pt %icc, .LL26
LDF [BO + 3 * SIZE], b4
.LL29:
#if defined(LN) || defined(RT)
sub KK, 4, TEMP1
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AO
add B, TEMP1, BO
#endif
FADD c04, t1, c04
FADD c08, t2, c08
FADD c12, t3, c12
FADD c16, t4, c16
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
LDF [BO + 4 * SIZE], b1
LDF [BO + 5 * SIZE], b2
LDF [BO + 6 * SIZE], b3
LDF [BO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c05, c05
FSUB a3, c09, c09
FSUB a4, c13, c13
FSUB b1, c02, c02
FSUB b2, c06, c06
FSUB b3, c10, c10
FSUB b4, c14, c14
LDF [BO + 8 * SIZE], a1
LDF [BO + 9 * SIZE], a2
LDF [BO + 10 * SIZE], a3
LDF [BO + 11 * SIZE], a4
LDF [BO + 12 * SIZE], b1
LDF [BO + 13 * SIZE], b2
LDF [BO + 14 * SIZE], b3
LDF [BO + 15 * SIZE], b4
FSUB a1, c03, c03
FSUB a2, c07, c07
FSUB a3, c11, c11
FSUB a4, c15, c15
FSUB b1, c04, c04
FSUB b2, c08, c08
FSUB b3, c12, c12
FSUB b4, c16, c16
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
LDF [AO + 4 * SIZE], b1
LDF [AO + 5 * SIZE], b2
LDF [AO + 6 * SIZE], b3
LDF [AO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
FSUB b1, c05, c05
FSUB b2, c06, c06
FSUB b3, c07, c07
FSUB b4, c08, c08
LDF [AO + 8 * SIZE], a1
LDF [AO + 9 * SIZE], a2
LDF [AO + 10 * SIZE], a3
LDF [AO + 11 * SIZE], a4
LDF [AO + 12 * SIZE], b1
LDF [AO + 13 * SIZE], b2
LDF [AO + 14 * SIZE], b3
LDF [AO + 15 * SIZE], b4
FSUB a1, c09, c09
FSUB a2, c10, c10
FSUB a3, c11, c11
FSUB a4, c12, c12
FSUB b1, c13, c13
FSUB b2, c14, c14
FSUB b3, c15, c15
FSUB b4, c16, c16
#endif
#ifdef LN
LDF [AO + 15 * SIZE], a1
LDF [AO + 14 * SIZE], a2
LDF [AO + 13 * SIZE], a3
LDF [AO + 12 * SIZE], a4
FMUL a1, c04, c04
FMUL a1, c08, c08
FMUL a1, c12, c12
FMUL a1, c16, c16
FMUL a2, c04, t1
FMUL a2, c08, t2
FMUL a2, c12, t3
FMUL a2, c16, t4
FSUB c03, t1, c03
FSUB c07, t2, c07
FSUB c11, t3, c11
FSUB c15, t4, c15
FMUL a3, c04, t1
FMUL a3, c08, t2
FMUL a3, c12, t3
FMUL a3, c16, t4
FSUB c02, t1, c02
FSUB c06, t2, c06
FSUB c10, t3, c10
FSUB c14, t4, c14
FMUL a4, c04, t1
FMUL a4, c08, t2
FMUL a4, c12, t3
FMUL a4, c16, t4
FSUB c01, t1, c01
FSUB c05, t2, c05
FSUB c09, t3, c09
FSUB c13, t4, c13
LDF [AO + 10 * SIZE], a1
LDF [AO + 9 * SIZE], a2
LDF [AO + 8 * SIZE], a3
FMUL a1, c03, c03
FMUL a1, c07, c07
FMUL a1, c11, c11
FMUL a1, c15, c15
FMUL a2, c03, t1
FMUL a2, c07, t2
FMUL a2, c11, t3
FMUL a2, c15, t4
FSUB c02, t1, c02
FSUB c06, t2, c06
FSUB c10, t3, c10
FSUB c14, t4, c14
FMUL a3, c03, t1
FMUL a3, c07, t2
FMUL a3, c11, t3
FMUL a3, c15, t4
FSUB c01, t1, c01
FSUB c05, t2, c05
FSUB c09, t3, c09
FSUB c13, t4, c13
LDF [AO + 5 * SIZE], a1
LDF [AO + 4 * SIZE], a2
FMUL a1, c02, c02
FMUL a1, c06, c06
FMUL a1, c10, c10
FMUL a1, c14, c14
FMUL a2, c02, t1
FMUL a2, c06, t2
FMUL a2, c10, t3
FMUL a2, c14, t4
FSUB c01, t1, c01
FSUB c05, t2, c05
FSUB c09, t3, c09
FSUB c13, t4, c13
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c05, c05
FMUL a1, c09, c09
FMUL a1, c13, c13
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a1, c05, c05
FMUL a1, c09, c09
FMUL a1, c13, c13
FMUL a2, c01, t1
FMUL a2, c05, t2
FMUL a2, c09, t3
FMUL a2, c13, t4
FSUB c02, t1, c02
FSUB c06, t2, c06
FSUB c10, t3, c10
FSUB c14, t4, c14
FMUL a3, c01, t1
FMUL a3, c05, t2
FMUL a3, c09, t3
FMUL a3, c13, t4
FSUB c03, t1, c03
FSUB c07, t2, c07
FSUB c11, t3, c11
FSUB c15, t4, c15
FMUL a4, c01, t1
FMUL a4, c05, t2
FMUL a4, c09, t3
FMUL a4, c13, t4
FSUB c04, t1, c04
FSUB c08, t2, c08
FSUB c12, t3, c12
FSUB c16, t4, c16
LDF [AO + 5 * SIZE], a1
LDF [AO + 6 * SIZE], a2
LDF [AO + 7 * SIZE], a3
FMUL a1, c02, c02
FMUL a1, c06, c06
FMUL a1, c10, c10
FMUL a1, c14, c14
FMUL a2, c02, t1
FMUL a2, c06, t2
FMUL a2, c10, t3
FMUL a2, c14, t4
FSUB c03, t1, c03
FSUB c07, t2, c07
FSUB c11, t3, c11
FSUB c15, t4, c15
FMUL a3, c02, t1
FMUL a3, c06, t2
FMUL a3, c10, t3
FMUL a3, c14, t4
FSUB c04, t1, c04
FSUB c08, t2, c08
FSUB c12, t3, c12
FSUB c16, t4, c16
LDF [AO + 10 * SIZE], a1
LDF [AO + 11 * SIZE], a2
FMUL a1, c03, c03
FMUL a1, c07, c07
FMUL a1, c11, c11
FMUL a1, c15, c15
FMUL a2, c03, t1
FMUL a2, c07, t2
FMUL a2, c11, t3
FMUL a2, c15, t4
FSUB c04, t1, c04
FSUB c08, t2, c08
FSUB c12, t3, c12
FSUB c16, t4, c16
LDF [AO + 15 * SIZE], a1
FMUL a1, c04, c04
FMUL a1, c08, c08
FMUL a1, c12, c12
FMUL a1, c16, c16
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
FMUL a2, c01, t1
FMUL a2, c02, t2
FMUL a2, c03, t3
FMUL a2, c04, t4
FSUB c05, t1, c05
FSUB c06, t2, c06
FSUB c07, t3, c07
FSUB c08, t4, c08
FMUL a3, c01, t1
FMUL a3, c02, t2
FMUL a3, c03, t3
FMUL a3, c04, t4
FSUB c09, t1, c09
FSUB c10, t2, c10
FSUB c11, t3, c11
FSUB c12, t4, c12
FMUL a4, c01, t1
FMUL a4, c02, t2
FMUL a4, c03, t3
FMUL a4, c04, t4
FSUB c13, t1, c13
FSUB c14, t2, c14
FSUB c15, t3, c15
FSUB c16, t4, c16
LDF [BO + 5 * SIZE], a1
LDF [BO + 6 * SIZE], a2
LDF [BO + 7 * SIZE], a3
FMUL a1, c05, c05
FMUL a1, c06, c06
FMUL a1, c07, c07
FMUL a1, c08, c08
FMUL a2, c05, t1
FMUL a2, c06, t2
FMUL a2, c07, t3
FMUL a2, c08, t4
FSUB c09, t1, c09
FSUB c10, t2, c10
FSUB c11, t3, c11
FSUB c12, t4, c12
FMUL a3, c05, t1
FMUL a3, c06, t2
FMUL a3, c07, t3
FMUL a3, c08, t4
FSUB c13, t1, c13
FSUB c14, t2, c14
FSUB c15, t3, c15
FSUB c16, t4, c16
LDF [BO + 10 * SIZE], a1
LDF [BO + 11 * SIZE], a2
FMUL a1, c09, c09
FMUL a1, c10, c10
FMUL a1, c11, c11
FMUL a1, c12, c12
FMUL a2, c09, t1
FMUL a2, c10, t2
FMUL a2, c11, t3
FMUL a2, c12, t4
FSUB c13, t1, c13
FSUB c14, t2, c14
FSUB c15, t3, c15
FSUB c16, t4, c16
LDF [BO + 15 * SIZE], a1
FMUL a1, c13, c13
FMUL a1, c14, c14
FMUL a1, c15, c15
FMUL a1, c16, c16
#endif
#ifdef RT
LDF [BO + 15 * SIZE], a1
LDF [BO + 14 * SIZE], a2
LDF [BO + 13 * SIZE], a3
LDF [BO + 12 * SIZE], a4
FMUL a1, c13, c13
FMUL a1, c14, c14
FMUL a1, c15, c15
FMUL a1, c16, c16
FMUL a2, c13, t1
FMUL a2, c14, t2
FMUL a2, c15, t3
FMUL a2, c16, t4
FSUB c09, t1, c09
FSUB c10, t2, c10
FSUB c11, t3, c11
FSUB c12, t4, c12
FMUL a3, c13, t1
FMUL a3, c14, t2
FMUL a3, c15, t3
FMUL a3, c16, t4
FSUB c05, t1, c05
FSUB c06, t2, c06
FSUB c07, t3, c07
FSUB c08, t4, c08
FMUL a4, c13, t1
FMUL a4, c14, t2
FMUL a4, c15, t3
FMUL a4, c16, t4
FSUB c01, t1, c01
FSUB c02, t2, c02
FSUB c03, t3, c03
FSUB c04, t4, c04
LDF [BO + 10 * SIZE], a1
LDF [BO + 9 * SIZE], a2
LDF [BO + 8 * SIZE], a3
FMUL a1, c09, c09
FMUL a1, c10, c10
FMUL a1, c11, c11
FMUL a1, c12, c12
FMUL a2, c09, t1
FMUL a2, c10, t2
FMUL a2, c11, t3
FMUL a2, c12, t4
FSUB c05, t1, c05
FSUB c06, t2, c06
FSUB c07, t3, c07
FSUB c08, t4, c08
FMUL a3, c09, t1
FMUL a3, c10, t2
FMUL a3, c11, t3
FMUL a3, c12, t4
FSUB c01, t1, c01
FSUB c02, t2, c02
FSUB c03, t3, c03
FSUB c04, t4, c04
LDF [BO + 5 * SIZE], a1
LDF [BO + 4 * SIZE], a2
FMUL a1, c05, c05
FMUL a1, c06, c06
FMUL a1, c07, c07
FMUL a1, c08, c08
FMUL a2, c05, t1
FMUL a2, c06, t2
FMUL a2, c07, t3
FMUL a2, c08, t4
FSUB c01, t1, c01
FSUB c02, t2, c02
FSUB c03, t3, c03
FSUB c04, t4, c04
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
#endif
#ifdef LN
add C1, -4 * SIZE, C1
add C2, -4 * SIZE, C2
add C3, -4 * SIZE, C3
add C4, -4 * SIZE, C4
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c05, [BO + 1 * SIZE]
STF c09, [BO + 2 * SIZE]
STF c13, [BO + 3 * SIZE]
STF c02, [BO + 4 * SIZE]
STF c06, [BO + 5 * SIZE]
STF c10, [BO + 6 * SIZE]
STF c14, [BO + 7 * SIZE]
STF c03, [BO + 8 * SIZE]
STF c07, [BO + 9 * SIZE]
STF c11, [BO + 10 * SIZE]
STF c15, [BO + 11 * SIZE]
STF c04, [BO + 12 * SIZE]
STF c08, [BO + 13 * SIZE]
STF c12, [BO + 14 * SIZE]
STF c16, [BO + 15 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
STF c03, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
STF c05, [AO + 4 * SIZE]
STF c06, [AO + 5 * SIZE]
STF c07, [AO + 6 * SIZE]
STF c08, [AO + 7 * SIZE]
STF c09, [AO + 8 * SIZE]
STF c10, [AO + 9 * SIZE]
STF c11, [AO + 10 * SIZE]
STF c12, [AO + 11 * SIZE]
STF c13, [AO + 12 * SIZE]
STF c14, [AO + 13 * SIZE]
STF c15, [AO + 14 * SIZE]
STF c16, [AO + 15 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C1 + 1 * SIZE]
STF c03, [C1 + 2 * SIZE]
STF c04, [C1 + 3 * SIZE]
STF c05, [C2 + 0 * SIZE]
STF c06, [C2 + 1 * SIZE]
STF c07, [C2 + 2 * SIZE]
STF c08, [C2 + 3 * SIZE]
STF c09, [C3 + 0 * SIZE]
STF c10, [C3 + 1 * SIZE]
STF c11, [C3 + 2 * SIZE]
STF c12, [C3 + 3 * SIZE]
STF c13, [C4 + 0 * SIZE]
STF c14, [C4 + 1 * SIZE]
STF c15, [C4 + 2 * SIZE]
STF c16, [C4 + 3 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 4 * SIZE, C1
add C2, 4 * SIZE, C2
add C3, 4 * SIZE, C3
add C4, 4 * SIZE, C4
#endif
#ifdef RT
sll K, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AO, TEMP1, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 4, KK
#endif
#ifdef LN
sub KK, 4, KK
#endif
add I, -1, I
cmp I, 0
sra K, 2, L
bg,pt %icc, .LL21
FMOV FZERO, c01
.LL50:
and M, 2, I
cmp I, 0
ble,pn %icc, .LL70
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 1 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 1 + BASE_SHIFT, TEMP1
sll KK, 2 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
FMOV FZERO, c02
FMOV FZERO, t1
FMOV FZERO, c04
LDF [AO + 0 * SIZE], a1
FMOV FZERO, t2
LDF [BO + 0 * SIZE], b1
FMOV FZERO, c06
LDF [AO + 1 * SIZE], a2
FMOV FZERO, t3
LDF [BO + 1 * SIZE], b2
FMOV FZERO, c08
LDF [AO + 2 * SIZE], a3
FMOV FZERO, t4
LDF [BO + 2 * SIZE], b3
FMOV FZERO, c01
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c03
LDF [BO + 3 * SIZE], b4
FMOV FZERO, c05
ble,pn %icc, .LL55
FMOV FZERO, c07
.LL52:
FADD c02, t1, c02
add AO, 8 * SIZE, AO
prefetch [AO + APREFETCHSIZE * SIZE], 0
FMUL a1, b1, t1
add BO, 16 * SIZE, BO
FADD c04, t2, c04
add L, -1, L
FMUL a1, b2, t2
FADD c06, t3, c06
cmp L, 0
FMUL a1, b3, t3
FADD c08, t4, c08
FMUL a1, b4, t4
LDF [AO - 4 * SIZE], a1
FADD c01, t1, c01
FMUL a2, b1, t1
LDF [BO - 12 * SIZE], b1
FADD c03, t2, c03
FMUL a2, b2, t2
LDF [BO - 11 * SIZE], b2
FADD c05, t3, c05
FMUL a2, b3, t3
LDF [BO - 10 * SIZE], b3
FADD c07, t4, c07
FMUL a2, b4, t4
LDF [BO - 9 * SIZE], b4
FADD c02, t1, c02
FMUL a3, b1, t1
LDF [AO - 3 * SIZE], a2
FADD c04, t2, c04
FMUL a3, b2, t2
FADD c06, t3, c06
FMUL a3, b3, t3
FADD c08, t4, c08
FMUL a3, b4, t4
LDF [AO - 2 * SIZE], a3
FADD c01, t1, c01
FMUL a4, b1, t1
LDF [BO - 8 * SIZE], b1
FADD c03, t2, c03
FMUL a4, b2, t2
LDF [BO - 7 * SIZE], b2
FADD c05, t3, c05
FMUL a4, b3, t3
LDF [BO - 6 * SIZE], b3
FADD c07, t4, c07
FMUL a4, b4, t4
LDF [BO - 5 * SIZE], b4
FADD c02, t1, c02
FMUL a1, b1, t1
LDF [AO - 1 * SIZE], a4
FADD c04, t2, c04
FMUL a1, b2, t2
FADD c06, t3, c06
FMUL a1, b3, t3
FADD c08, t4, c08
FMUL a1, b4, t4
LDF [AO + 0 * SIZE], a1
FADD c01, t1, c01
FMUL a2, b1, t1
LDF [BO - 4 * SIZE], b1
FADD c03, t2, c03
FMUL a2, b2, t2
LDF [BO - 3 * SIZE], b2
FADD c05, t3, c05
FMUL a2, b3, t3
LDF [BO - 2 * SIZE], b3
FADD c07, t4, c07
FMUL a2, b4, t4
LDF [BO - 1 * SIZE], b4
FADD c02, t1, c02
FMUL a3, b1, t1
LDF [AO + 1 * SIZE], a2
FADD c04, t2, c04
FMUL a3, b2, t2
FADD c06, t3, c06
FMUL a3, b3, t3
FADD c08, t4, c08
FMUL a3, b4, t4
LDF [AO + 2 * SIZE], a3
FADD c01, t1, c01
FMUL a4, b1, t1
LDF [BO + 0 * SIZE], b1
FADD c03, t2, c03
FMUL a4, b2, t2
LDF [BO + 1 * SIZE], b2
FADD c05, t3, c05
FMUL a4, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c07, t4, c07
FMUL a4, b4, t4
LDF [BO + 3 * SIZE], b4
bg,pt %icc, .LL52
LDF [AO + 3 * SIZE], a4
.LL55:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL59
nop
.LL56:
FADD c02, t1, c02
add AO, 2 * SIZE, AO
FMUL a1, b1, t1
add L, -1, L
add BO, 4 * SIZE, BO
FADD c04, t2, c04
cmp L, 0
FMUL a1, b2, t2
FADD c06, t3, c06
FMUL a1, b3, t3
FADD c08, t4, c08
FMUL a1, b4, t4
LDF [AO + 0 * SIZE], a1
FADD c01, t1, c01
FMUL a2, b1, t1
LDF [BO + 0 * SIZE], b1
FADD c03, t2, c03
FMUL a2, b2, t2
LDF [BO + 1 * SIZE], b2
FADD c05, t3, c05
FMUL a2, b3, t3
LDF [BO + 2 * SIZE], b3
FADD c07, t4, c07
FMUL a2, b4, t4
LDF [BO + 3 * SIZE], b4
bg,pt %icc, .LL56
LDF [AO + 1 * SIZE], a2
.LL59:
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 2, TEMP1
#else
sub KK, 4, TEMP1
#endif
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
FADD c02, t1, c02
FADD c04, t2, c04
FADD c06, t3, c06
FADD c08, t4, c08
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
LDF [BO + 4 * SIZE], b1
LDF [BO + 5 * SIZE], b2
LDF [BO + 6 * SIZE], b3
LDF [BO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c03, c03
FSUB a3, c05, c05
FSUB a4, c07, c07
FSUB b1, c02, c02
FSUB b2, c04, c04
FSUB b3, c06, c06
FSUB b4, c08, c08
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
LDF [AO + 4 * SIZE], b1
LDF [AO + 5 * SIZE], b2
LDF [AO + 6 * SIZE], b3
LDF [AO + 7 * SIZE], b4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
FSUB b1, c05, c05
FSUB b2, c06, c06
FSUB b3, c07, c07
FSUB b4, c08, c08
#endif
#ifdef LN
LDF [AO + 3 * SIZE], a1
LDF [AO + 2 * SIZE], a2
LDF [AO + 0 * SIZE], a3
FMUL a1, c02, c02
FMUL a1, c04, c04
FMUL a1, c06, c06
FMUL a1, c08, c08
FMUL a2, c02, t1
FMUL a2, c04, t2
FMUL a2, c06, t3
FMUL a2, c08, t4
FSUB c01, t1, c01
FSUB c03, t2, c03
FSUB c05, t3, c05
FSUB c07, t4, c07
FMUL a3, c01, c01
FMUL a3, c03, c03
FMUL a3, c05, c05
FMUL a3, c07, c07
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 3 * SIZE], a3
FMUL a1, c01, c01
FMUL a1, c03, c03
FMUL a1, c05, c05
FMUL a1, c07, c07
FMUL a2, c01, t1
FMUL a2, c03, t2
FMUL a2, c05, t3
FMUL a2, c07, t4
FSUB c02, t1, c02
FSUB c04, t2, c04
FSUB c06, t3, c06
FSUB c08, t4, c08
FMUL a3, c02, c02
FMUL a3, c04, c04
FMUL a3, c06, c06
FMUL a3, c08, c08
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a2, c01, t1
FMUL a2, c02, t2
FSUB c03, t1, c03
FSUB c04, t2, c04
FMUL a3, c01, t1
FMUL a3, c02, t2
FSUB c05, t1, c05
FSUB c06, t2, c06
FMUL a4, c01, t1
FMUL a4, c02, t2
FSUB c07, t1, c07
FSUB c08, t2, c08
LDF [BO + 5 * SIZE], a1
LDF [BO + 6 * SIZE], a2
LDF [BO + 7 * SIZE], a3
FMUL a1, c03, c03
FMUL a1, c04, c04
FMUL a2, c03, t1
FMUL a2, c04, t2
FSUB c05, t1, c05
FSUB c06, t2, c06
FMUL a3, c03, t1
FMUL a3, c04, t2
FSUB c07, t1, c07
FSUB c08, t2, c08
LDF [BO + 10 * SIZE], a1
LDF [BO + 11 * SIZE], a2
FMUL a1, c05, c05
FMUL a1, c06, c06
FMUL a2, c05, t1
FMUL a2, c06, t2
FSUB c07, t1, c07
FSUB c08, t2, c08
LDF [BO + 15 * SIZE], a1
FMUL a1, c07, c07
FMUL a1, c08, c08
#endif
#ifdef RT
LDF [BO + 15 * SIZE], a1
LDF [BO + 14 * SIZE], a2
LDF [BO + 13 * SIZE], a3
LDF [BO + 12 * SIZE], a4
FMUL a1, c07, c07
FMUL a1, c08, c08
FMUL a2, c07, t1
FMUL a2, c08, t2
FSUB c05, t1, c05
FSUB c06, t2, c06
FMUL a3, c07, t1
FMUL a3, c08, t2
FSUB c03, t1, c03
FSUB c04, t2, c04
FMUL a4, c07, t1
FMUL a4, c08, t2
FSUB c01, t1, c01
FSUB c02, t2, c02
LDF [BO + 10 * SIZE], a1
LDF [BO + 9 * SIZE], a2
LDF [BO + 8 * SIZE], a3
FMUL a1, c05, c05
FMUL a1, c06, c06
FMUL a2, c05, t1
FMUL a2, c06, t2
FSUB c03, t1, c03
FSUB c04, t2, c04
FMUL a3, c05, t1
FMUL a3, c06, t2
FSUB c01, t1, c01
FSUB c02, t2, c02
LDF [BO + 5 * SIZE], a1
LDF [BO + 4 * SIZE], a2
FMUL a1, c03, c03
FMUL a1, c04, c04
FMUL a2, c03, t1
FMUL a2, c04, t2
FSUB c01, t1, c01
FSUB c02, t2, c02
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
#endif
#ifdef LN
add C1, -2 * SIZE, C1
add C2, -2 * SIZE, C2
add C3, -2 * SIZE, C3
add C4, -2 * SIZE, C4
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c03, [BO + 1 * SIZE]
STF c05, [BO + 2 * SIZE]
STF c07, [BO + 3 * SIZE]
STF c02, [BO + 4 * SIZE]
STF c04, [BO + 5 * SIZE]
STF c06, [BO + 6 * SIZE]
STF c08, [BO + 7 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
STF c03, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
STF c05, [AO + 4 * SIZE]
STF c06, [AO + 5 * SIZE]
STF c07, [AO + 6 * SIZE]
STF c08, [AO + 7 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C1 + 1 * SIZE]
STF c03, [C2 + 0 * SIZE]
STF c04, [C2 + 1 * SIZE]
STF c05, [C3 + 0 * SIZE]
STF c06, [C3 + 1 * SIZE]
STF c07, [C4 + 0 * SIZE]
STF c08, [C4 + 1 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 2 * SIZE, C1
add C2, 2 * SIZE, C2
add C3, 2 * SIZE, C3
add C4, 2 * SIZE, C4
#endif
#ifdef RT
sll K, 1 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 1 + BASE_SHIFT, TEMP2
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 2, KK
#endif
#ifdef LN
sub KK, 2, KK
#endif
.LL70:
and M, 1, I
cmp I, 0
ble,pn %icc, .LL99
nop
#if defined(LT) || defined(RN)
sra KK, 2, L
mov B, BO
cmp L, 0
#else
#ifdef LN
sll K, 0 + BASE_SHIFT, TEMP1
sub AORIG, TEMP1, AORIG
#endif
sll KK, 0 + BASE_SHIFT, TEMP1
sll KK, 2 + BASE_SHIFT, TEMP2
add AORIG, TEMP1, AO
add B, TEMP2, BO
sub K, KK, TEMP1
sra TEMP1, 2, L
cmp L, 0
#endif
LDF [AO + 0 * SIZE], a1
FMOV FZERO, c01
LDF [BO + 0 * SIZE], b1
FMOV FZERO, t1
LDF [AO + 1 * SIZE], a2
FMOV FZERO, c02
LDF [BO + 1 * SIZE], b2
FMOV FZERO, t2
LDF [AO + 2 * SIZE], a3
FMOV FZERO, c03
LDF [BO + 2 * SIZE], b3
FMOV FZERO, t3
LDF [AO + 3 * SIZE], a4
FMOV FZERO, c04
LDF [BO + 3 * SIZE], b4
FMOV FZERO, t4
ble,pn %icc, .LL75
nop
.LL72:
FADD c01, t1, c01
add L, -1, L
FMUL a1, b1, t1
LDF [BO + 4 * SIZE], b1
FADD c02, t2, c02
cmp L, 0
FMUL a1, b2, t2
LDF [BO + 5 * SIZE], b2
FADD c03, t3, c03
FMUL a1, b3, t3
LDF [BO + 6 * SIZE], b3
FADD c04, t4, c04
FMUL a1, b4, t4
LDF [BO + 7 * SIZE], b4
LDF [AO + 4 * SIZE], a1
FADD c01, t1, c01
add AO, 4 * SIZE, AO
FMUL a2, b1, t1
LDF [BO + 8 * SIZE], b1
FADD c02, t2, c02
FMUL a2, b2, t2
LDF [BO + 9 * SIZE], b2
FADD c03, t3, c03
FMUL a2, b3, t3
LDF [BO + 10 * SIZE], b3
FADD c04, t4, c04
FMUL a2, b4, t4
LDF [BO + 11 * SIZE], b4
LDF [AO + 1 * SIZE], a2
FADD c01, t1, c01
FMUL a3, b1, t1
LDF [BO + 12 * SIZE], b1
FADD c02, t2, c02
FMUL a3, b2, t2
LDF [BO + 13 * SIZE], b2
FADD c03, t3, c03
FMUL a3, b3, t3
LDF [BO + 14 * SIZE], b3
FADD c04, t4, c04
FMUL a3, b4, t4
LDF [BO + 15 * SIZE], b4
LDF [AO + 2 * SIZE], a3
FADD c01, t1, c01
FMUL a4, b1, t1
LDF [BO + 16 * SIZE], b1
FADD c02, t2, c02
FMUL a4, b2, t2
LDF [BO + 17 * SIZE], b2
FADD c03, t3, c03
FMUL a4, b3, t3
LDF [BO + 18 * SIZE], b3
FADD c04, t4, c04
FMUL a4, b4, t4
LDF [BO + 19 * SIZE], b4
add BO, 16 * SIZE, BO
bg,pt %icc, .LL72
LDF [AO + 3 * SIZE], a4
.LL75:
#if defined(LT) || defined(RN)
and KK, 3, L
#else
and TEMP1, 3, L
#endif
cmp L, 0
ble,a,pn %icc, .LL79
nop
.LL76:
FADD c01, t1, c01
add AO, 1 * SIZE, AO
FMUL a1, b1, t1
LDF [BO + 4 * SIZE], b1
FADD c02, t2, c02
add L, -1, L
FMUL a1, b2, t2
LDF [BO + 5 * SIZE], b2
FADD c03, t3, c03
cmp L, 0
FMUL a1, b3, t3
LDF [BO + 6 * SIZE], b3
FADD c04, t4, c04
add BO, 4 * SIZE, BO
FMUL a1, b4, t4
LDF [AO + 0 * SIZE], a1
bg,pt %icc, .LL76
LDF [BO + 3 * SIZE], b4
.LL79:
FADD c01, t1, c01
FADD c02, t2, c02
FADD c03, t3, c03
FADD c04, t4, c04
#if defined(LN) || defined(RT)
#ifdef LN
sub KK, 1, TEMP1
#else
sub KK, 4, TEMP1
#endif
sll TEMP1, 0 + BASE_SHIFT, TEMP2
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AORIG, TEMP2, AO
add B, TEMP1, BO
#endif
#if defined(LN) || defined(LT)
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
#else
LDF [AO + 0 * SIZE], a1
LDF [AO + 1 * SIZE], a2
LDF [AO + 2 * SIZE], a3
LDF [AO + 3 * SIZE], a4
FSUB a1, c01, c01
FSUB a2, c02, c02
FSUB a3, c03, c03
FSUB a4, c04, c04
#endif
#ifdef LN
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
#endif
#ifdef LT
LDF [AO + 0 * SIZE], a1
FMUL a1, c01, c01
FMUL a1, c02, c02
FMUL a1, c03, c03
FMUL a1, c04, c04
#endif
#ifdef RN
LDF [BO + 0 * SIZE], a1
LDF [BO + 1 * SIZE], a2
LDF [BO + 2 * SIZE], a3
LDF [BO + 3 * SIZE], a4
FMUL a1, c01, c01
FMUL a2, c01, t1
FSUB c02, t1, c02
FMUL a3, c01, t1
FSUB c03, t1, c03
FMUL a4, c01, t1
FSUB c04, t1, c04
LDF [BO + 5 * SIZE], a1
LDF [BO + 6 * SIZE], a2
LDF [BO + 7 * SIZE], a3
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c03, t1, c03
FMUL a3, c02, t1
FSUB c04, t1, c04
LDF [BO + 10 * SIZE], a1
LDF [BO + 11 * SIZE], a2
FMUL a1, c03, c03
FMUL a2, c03, t1
FSUB c04, t1, c04
LDF [BO + 15 * SIZE], a1
FMUL a1, c04, c04
#endif
#ifdef RT
LDF [BO + 15 * SIZE], a1
LDF [BO + 14 * SIZE], a2
LDF [BO + 13 * SIZE], a3
LDF [BO + 12 * SIZE], a4
FMUL a1, c04, c04
FMUL a2, c04, t1
FSUB c03, t1, c03
FMUL a3, c04, t1
FSUB c02, t1, c02
FMUL a4, c04, t1
FSUB c01, t1, c01
LDF [BO + 10 * SIZE], a1
LDF [BO + 9 * SIZE], a2
LDF [BO + 8 * SIZE], a3
FMUL a1, c03, c03
FMUL a2, c03, t1
FSUB c02, t1, c02
FMUL a3, c03, t1
FSUB c01, t1, c01
LDF [BO + 5 * SIZE], a1
LDF [BO + 4 * SIZE], a2
FMUL a1, c02, c02
FMUL a2, c02, t1
FSUB c01, t1, c01
LDF [BO + 0 * SIZE], a1
FMUL a1, c01, c01
#endif
#ifdef LN
add C1, -1 * SIZE, C1
add C2, -1 * SIZE, C2
add C3, -1 * SIZE, C3
add C4, -1 * SIZE, C4
#endif
#if defined(LN) || defined(LT)
STF c01, [BO + 0 * SIZE]
STF c02, [BO + 1 * SIZE]
STF c03, [BO + 2 * SIZE]
STF c04, [BO + 3 * SIZE]
#else
STF c01, [AO + 0 * SIZE]
STF c02, [AO + 1 * SIZE]
STF c03, [AO + 2 * SIZE]
STF c04, [AO + 3 * SIZE]
#endif
STF c01, [C1 + 0 * SIZE]
STF c02, [C2 + 0 * SIZE]
STF c03, [C3 + 0 * SIZE]
STF c04, [C4 + 0 * SIZE]
FMOV FZERO, t1
FMOV FZERO, t2
FMOV FZERO, t3
FMOV FZERO, t4
#ifndef LN
add C1, 1 * SIZE, C1
add C2, 1 * SIZE, C2
add C3, 1 * SIZE, C3
add C4, 1 * SIZE, C4
#endif
#ifdef RT
sll K, 0 + BASE_SHIFT, TEMP1
add AORIG, TEMP1, AORIG
#endif
#if defined(LT) || defined(RN)
sub K, KK, TEMP1
sll TEMP1, 0 + BASE_SHIFT, TEMP2
sll TEMP1, 2 + BASE_SHIFT, TEMP1
add AO, TEMP2, AO
add BO, TEMP1, BO
#endif
#ifdef LT
add KK, 1, KK
#endif
#ifdef LN
sub KK, 1, KK
#endif
.LL99:
#ifdef LN
sll K, 2 + BASE_SHIFT, TEMP1
add B, TEMP1, B
#endif
#if defined(LT) || defined(RN)
mov BO, B
#endif
#ifdef RN
add KK, 4, KK
#endif
#ifdef RT
sub KK, 4, KK
#endif
add J, -1, J
cmp J, 0
bg,pt %icc, .LL11
nop
.LL999:
return %i7 + 8
clr %o0
EPILOGUE