/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define PREFETCH_SIZE (32 * 16)
#ifndef CONJ
#define FMA1 FNMA
#define FMA2 FMA
#else
#define FMA1 FMA
#define FMA2 FNMA
#endif
#define SP r12
#define N r32
#define X1 r37
#define INCX r38
#define Y1 r39
#define INCY r36
#define PREX1 r2
#define PREY1 r3
#define I r33
#define J r34
#define Y2 r35
#define X2 r14
#define YY1 r15
#define YY2 r16
#define YY3 r17
#define YY4 r18
#define INCXM1 r19
#define INCYM1 r20
#define INCX3M1 r21
#define INCY3M1 r22
#define INCX7M1 r23
#define INCY7M1 r24
#define X3 r8
#define Y3 r9
#define X4 r10
#define Y4 r11
#define INCX8 r25
#define INCY8 r26
#define ARLC r29
#define PR r30
#define ALPHA_R f8
#define ALPHA_I f9
PROLOGUE
.prologue
PROFCODE
{ .mmi
adds r14 = 16, SP
and J = 7, N
.save ar.lc, ARLC
mov ARLC = ar.lc
}
{ .mib
cmp.gt p15, p0 = r0, N
shr I = N, 3
(p15) br.ret.sptk.many b0
}
;;
{ .mmi
ld8 INCY = [r14]
nop __LINE__
mov PR = pr
}
{ .mmi
adds PREX1 = (PREFETCH_SIZE + 0) * SIZE, X1
adds PREY1 = (PREFETCH_SIZE + 0) * SIZE, Y1
shl INCX = INCX, ZBASE_SHIFT
}
;;
{ .mii
adds I = -1, I
mov pr.rot= 0
shl INCY = INCY, ZBASE_SHIFT
}
;;
{ .mmi
adds INCXM1 = -SIZE, INCX
adds INCYM1 = -SIZE, INCY
mov ar.ec = 3
}
{ .mmi
shladd X2 = INCX, 1, X1
shladd Y2 = INCY, 1, Y1
cmp.eq p16, p0 = r0, r0
}
;;
{ .mmi
shladd INCX3M1 = INCX, 1, INCXM1
shladd INCY3M1 = INCY, 1, INCYM1
shladd INCX8 = INCX, 3, r0
}
{ .mmi
shladd X3 = INCX, 1, X2
shladd Y3 = INCY, 1, Y2
shladd INCY8 = INCY, 3, r0
}
;;
{ .mmi
shladd X4 = INCX, 1, X3
shladd Y4 = INCY, 1, Y3
shladd INCX7M1 = INCX, 2, INCX3M1
}
{ .mmi
mov YY1 = Y1
mov YY2 = Y2
shladd INCY7M1 = INCY, 2, INCY3M1
}
;;
{ .mmi
mov YY3 = Y3
mov YY4 = Y4
mov ar.lc = I
}
{ .mib
cmp.eq p11 ,p0 = -1, I
tbit.z p0, p13 = N, 2
(p11) br.cond.dpnt .L25
}
;;
.align 32
.L22:
{ .mmf
(p19) STFD [YY3] = f14
(p19) STFD [YY4] = f15
(p18) FMA2 f14 = ALPHA_R, f64, f112
}
{ .mmf
(p16) LDFD f80 = [Y1], 1 * SIZE
(p16) LDFD f92 = [Y2], 1 * SIZE
(p18) FMA2 f15 = ALPHA_R, f76, f124
}
;;
{ .mmf
(p16) lfetch.excl.nt1 [PREY1], INCY8
(p16) LDFD f104 = [Y3], 1 * SIZE
(p18) FMA1 f6 = ALPHA_I, f40, f6
}
{ .mmf
(p16) LDFD f116 = [Y4], 1 * SIZE
nop __LINE__
(p18) FMA1 f7 = ALPHA_I, f52, f7
}
;;
{ .mmf
(p16) LDFD f86 = [Y1], INCYM1
(p16) LDFD f98 = [Y2], INCYM1
(p18) FMA1 f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
(p18) FMA1 f11 = ALPHA_I, f76, f11
}
;;
{ .mmf
(p16) LDFD f110 = [Y3], INCYM1
(p16) LDFD f122 = [Y4], INCYM1
(p18) FMA f12 = ALPHA_I, f34, f12
}
{ .mmf
(p19) add YY1 = YY1, INCY7M1
(p19) add YY2 = YY2, INCY7M1
(p18) FMA f13 = ALPHA_I, f46, f13
}
;;
{ .mmf
(p16) LDFD f32 = [X1], 1 * SIZE
(p16) LDFD f44 = [X2], 1 * SIZE
(p18) FMA f14 = ALPHA_I, f58, f14
}
{ .mmf
(p19) add YY3 = YY3, INCY7M1
(p19) add YY4 = YY4, INCY7M1
(p18) FMA f15 = ALPHA_I, f70, f15
}
;;
{ .mmf
(p18) STFD [YY1] = f6, 1 * SIZE
(p18) STFD [YY2] = f7, 1 * SIZE
(p18) FMA f6 = ALPHA_R, f37, f85
}
{ .mmf
(p16) LDFD f56 = [X3], 1 * SIZE
(p16) LDFD f68 = [X4], 1 * SIZE
(p18) FMA f7 = ALPHA_R, f49, f97
}
;;
{ .mmf
(p18) STFD [YY3] = f10, 1 * SIZE
(p18) STFD [YY4] = f11, 1 * SIZE
(p18) FMA f10 = ALPHA_R, f61, f109
}
{ .mmf
(p16) LDFD f38 = [X1], INCXM1
(p16) LDFD f50 = [X2], INCXM1
(p18) FMA f11 = ALPHA_R, f73, f121
}
;;
{ .mmf
(p18) STFD [YY1] = f12
(p18) STFD [YY2] = f13
(p18) FMA2 f12 = ALPHA_R, f43, f91
}
{ .mmf
(p16) LDFD f62 = [X3], INCXM1
(p16) LDFD f74 = [X4], INCXM1
(p18) FMA2 f13 = ALPHA_R, f55, f103
}
;;
{ .mmf
(p18) STFD [YY3] = f14
(p18) STFD [YY4] = f15
(p18) FMA2 f14 = ALPHA_R, f67, f115
}
{ .mmf
(p16) LDFD f83 = [Y1], 1 * SIZE
(p16) LDFD f95 = [Y2], 1 * SIZE
(p18) FMA2 f15 = ALPHA_R, f79, f127
}
;;
{ .mmf
(p16) LDFD f107 = [Y3], 1 * SIZE
(p16) LDFD f119 = [Y4], 1 * SIZE
(p18) FMA1 f6 = ALPHA_I, f43, f6
}
{ .mmf
nop __LINE__
nop __LINE__
(p18) FMA1 f7 = ALPHA_I, f55, f7
}
;;
{ .mmf
(p16) LDFD f89 = [Y1], INCY7M1
(p16) LDFD f101 = [Y2], INCY7M1
(p18) FMA1 f10 = ALPHA_I, f67, f10
}
{ .mmf
(p18) add YY1 = YY1, INCYM1
(p18) add YY2 = YY2, INCYM1
(p18) FMA1 f11 = ALPHA_I, f79, f11
}
;;
{ .mmf
(p16) LDFD f113 = [Y3], INCY7M1
(p16) LDFD f125 = [Y4], INCY7M1
(p18) FMA f12 = ALPHA_I, f37, f12
}
{ .mmf
(p18) add YY3 = YY3, INCYM1
(p18) add YY4 = YY4, INCYM1
(p18) FMA f13 = ALPHA_I, f49, f13
}
;;
{ .mmf
(p16) LDFD f35 = [X1], 1 * SIZE
(p16) LDFD f47 = [X2], 1 * SIZE
(p18) FMA f14 = ALPHA_I, f61, f14
}
{ .mmf
(p16) LDFD f59 = [X3], 1 * SIZE
(p16) LDFD f71 = [X4], 1 * SIZE
(p18) FMA f15 = ALPHA_I, f73, f15
}
;;
{ .mmf
(p18) STFD [YY1] = f6, 1 * SIZE
(p18) STFD [YY2] = f7, 1 * SIZE
(p17) FMA f6 = ALPHA_R, f33, f81
}
{ .mmf
(p16) LDFD f41 = [X1], INCX7M1
(p16) LDFD f53 = [X2], INCX7M1
(p17) FMA f7 = ALPHA_R, f45, f93
}
;;
{ .mmf
(p18) STFD [YY3] = f10, 1 * SIZE
(p18) STFD [YY4] = f11, 1 * SIZE
(p17) FMA f10 = ALPHA_R, f57, f105
}
{ .mmf
(p16) LDFD f65 = [X3], INCX7M1
(p16) LDFD f77 = [X4], INCX7M1
(p17) FMA f11 = ALPHA_R, f69, f117
}
;;
{ .mmf
(p18) STFD [YY1] = f12
(p18) STFD [YY2] = f13
(p17) FMA2 f12 = ALPHA_R, f39, f87
}
{ .mfb
(p16) lfetch.nt1 [PREX1], INCX8
(p17) FMA2 f13 = ALPHA_R, f51, f99
br.ctop.sptk.few .L22
}
;;
(p19) add YY1 = YY1, INCY7M1
(p19) add YY2 = YY2, INCY7M1
;;
{ .mmf
(p19) STFD [YY3] = f14
(p19) STFD [YY4] = f15
}
{ .mmf
(p19) add YY3 = YY3, INCY7M1
(p19) add YY4 = YY4, INCY7M1
}
;;
.align 32
.L25:
{ .mmi
(p13) LDFD f32 = [X1], 1 * SIZE
(p13) LDFD f36 = [X2], 1 * SIZE
mov ar.lc = ARLC
}
;;
{ .mmi
(p13) LDFD f80 = [Y1], 1 * SIZE
(p13) LDFD f84 = [Y2], 1 * SIZE
mov pr = PR, -65474
}
;;
{ .mmi
(p13) LDFD f33 = [X1], INCXM1
(p13) LDFD f37 = [X2], INCXM1
cmp.eq p12, p0 = r0, J
}
;;
{ .mmb
(p13) LDFD f81 = [Y1], INCYM1
(p13) LDFD f85 = [Y2], INCYM1
(p12) br.ret.sptk.many b0
}
;;
{ .mmi
(p13) LDFD f34 = [X1], 1 * SIZE
(p13) LDFD f38 = [X2], 1 * SIZE
tbit.z p0, p14 = N, 1
}
;;
{ .mmi
(p13) LDFD f82 = [Y1], 1 * SIZE
(p13) LDFD f86 = [Y2], 1 * SIZE
tbit.z p0, p15 = N, 0
}
;;
{ .mmf
(p13) LDFD f35 = [X1], INCX3M1
(p13) LDFD f39 = [X2], INCX3M1
(p13) FMA f80 = ALPHA_R, f32, f80
}
;;
{ .mmf
(p13) LDFD f83 = [Y1], INCY3M1
(p13) LDFD f87 = [Y2], INCY3M1
(p13) FMA f84 = ALPHA_R, f36, f84
}
;;
{ .mmf
(p14) LDFD f40 = [X1], 1 * SIZE
(p14) LDFD f88 = [Y1], 1 * SIZE
(p13) FMA2 f81 = ALPHA_R, f33, f81
}
;;
{ .mmf
(p14) LDFD f41 = [X1], INCXM1
(p14) LDFD f89 = [Y1], INCYM1
(p13) FMA2 f85 = ALPHA_R, f37, f85
}
;;
{ .mmf
(p14) LDFD f42 = [X1], 1 * SIZE
(p14) LDFD f90 = [Y1], 1 * SIZE
(p13) FMA f82 = ALPHA_R, f34, f82
}
;;
{ .mmf
(p14) LDFD f43 = [X1], INCXM1
(p14) LDFD f91 = [Y1], INCYM1
(p13) FMA f86 = ALPHA_R, f38, f86
}
;;
{ .mmf
(p15) LDFD f44 = [X1], 1 * SIZE
(p15) LDFD f92 = [Y1], 1 * SIZE
(p13) FMA2 f83 = ALPHA_R, f35, f83
}
;;
{ .mmf
(p15) LDFD f45 = [X1]
(p15) LDFD f93 = [Y1]
(p13) FMA2 f87 = ALPHA_R, f39, f87
}
;;
(p13) FMA1 f80 = ALPHA_I, f33, f80
(p13) FMA1 f84 = ALPHA_I, f37, f84
(p13) FMA f81 = ALPHA_I, f32, f81
(p13) FMA f85 = ALPHA_I, f36, f85
(p13) FMA1 f82 = ALPHA_I, f35, f82
(p13) FMA1 f86 = ALPHA_I, f39, f86
(p13) FMA f83 = ALPHA_I, f34, f83
(p13) FMA f87 = ALPHA_I, f38, f87
;;
{ .mmf
(p13) STFD [YY1] = f80, 1 * SIZE
(p13) STFD [YY2] = f84, 1 * SIZE
(p14) FMA f88 = ALPHA_R, f40, f88
}
;;
{ .mmf
(p13) STFD [YY1] = f81
(p13) STFD [YY2] = f85
(p14) FMA2 f89 = ALPHA_R, f41, f89
}
{ .mmf
(p13) add YY1 = YY1, INCYM1
(p13) add YY2 = YY2, INCYM1
(p14) FMA f90 = ALPHA_R, f42, f90
}
;;
{ .mmf
(p13) STFD [YY1] = f82, 1 * SIZE
(p13) STFD [YY2] = f86, 1 * SIZE
(p14) FMA2 f91 = ALPHA_R, f43, f91
}
;;
{ .mmf
(p13) STFD [YY1] = f83
(p13) STFD [YY2] = f87
(p15) FMA f92 = ALPHA_R, f44, f92
}
{ .mmf
(p13) add YY1 = YY1, INCY3M1
nop __LINE__
(p15) FMA2 f93 = ALPHA_R, f45, f93
}
;;
(p14) FMA1 f88 = ALPHA_I, f41, f88
(p14) FMA f89 = ALPHA_I, f40, f89
(p14) FMA1 f90 = ALPHA_I, f43, f90
(p14) FMA f91 = ALPHA_I, f42, f91
;;
{ .mmf
(p14) STFD [YY1] = f88, 1 * SIZE
(p15) FMA1 f92 = ALPHA_I, f45, f92
}
;;
{ .mmf
(p14) STFD [YY1] = f89
(p14) add YY1 = YY1, INCYM1
(p15) FMA f93 = ALPHA_I, f44, f93
}
;;
(p14) STFD [YY1] = f90, 1 * SIZE
;;
(p14) STFD [YY1] = f91
(p14) add YY1 = YY1, INCYM1
;;
(p15) STFD [YY1] = f92, 1 * SIZE
;;
{ .mmb
(p15) STFD [YY1] = f93
nop __LINE__
br.ret.sptk.many b0
}
;;
EPILOGUE