/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#ifdef DOUBLE
#define PREFETCHSIZE (16 * 8)
#else
#define PREFETCHSIZE (32 * 8)
#endif
#define CPREFETCHSIZE 15
#define CPREFETCH lfetch.excl.nt1
#define M r32
#define N r33
#define K r34
#define A r37
#define B r38
#define C r39
#define LDC r35
#define I r15
#define J r16
#define AOFFSET r17
#define BOFFSET r18
#define L r20
#define C1 r21
#define C2 r22
#define C3 r23
#define C4 r24
#define C5 r25
#define C6 r26
#define C7 r27
#define C8 r28
#define C9 loc0
#define C10 loc1
#define C11 loc2
#define C12 loc3
#define C13 loc4
#define C14 loc5
#define C15 loc6
#define C16 loc7
#define PREA r8
#define PREB r9
#define PREC r10
#define SP r12
#define ARLC r29
#define PR r30
#define ARPFS r31
#define ALPHA_R f8
#define ALPHA_I f9
PROLOGUE
.prologue
PROFCODE
{ .mmi
.save ar.pfs, ARPFS
alloc ARPFS = ar.pfs, 8, 16, 0, 0
adds r14 = 16, SP
mov ARLC = ar.lc
}
{ .mmi
adds r8 = -16 * 16, SP
adds r9 = -15 * 16, SP
adds SP = -16 * 16, SP
}
;;
{ .mmi
stf.spill [r8] = f16, 32
stf.spill [r9] = f17, 32
mov PR = pr
}
{ .mmi
ld8 LDC = [r14], 8
nop __LINE__
nop __LINE__
}
;;
stf.spill [r8] = f18, 32
stf.spill [r9] = f19, 32
shr J = N, 3
;;
stf.spill [r8] = f20, 32
stf.spill [r9] = f21, 32
shladd LDC = LDC, ZBASE_SHIFT, r0
;;
stf.spill [r8] = f22, 32
stf.spill [r9] = f23, 32
mov AOFFSET = A
;;
stf.spill [r8] = f24, 32
stf.spill [r9] = f25, 32
cmp.ge p6, p0 = 0, J
;;
stf.spill [r8] = f26, 32
stf.spill [r9] = f27, 32
;;
stf.spill [r8] = f28, 32
stf.spill [r9] = f29, 32
;;
stf.spill [r8] = f30
stf.spill [r9] = f31
(p6) br.cond.dpnt .L050
.body
;;
.align 32
.L010:
{ .mfi
adds J = -1, J
mov f64 = f0
shr I = M, 3
}
{ .mfi
mov C1 = C // coffset1 = c + 0 * ldc
mov f72 = f0
}
;;
{ .mmf
cmp.eq p6, p7 = 0, I
nop __LINE__
mov f80 = f0
}
{ .mmf
add C2 = LDC, C // coffset2 = c + 1 * ldc
shladd C3 = LDC, 1, C // coffset3 = c + 2 * ldc
mov f88 = f0
}
;;
{ .mmf
shladd C5 = LDC, 2, C // coffset5 = c + 4 * ldc
shladd C = LDC, 3, C // coffset += 8 * ldc
mov f96 = f0
}
{ .mmf
shladd C4 = LDC, 1, C2 // coffset4 = c + 3 * ldc
shladd C6 = LDC, 2, C2 // coffset6 = c + 5 * ldc
mov f104 = f0
}
;;
{ .mfi
shladd C7 = LDC, 2, C3 // coffset7 = c + 6 * ldc
mov f112 = f0
nop __LINE__
}
{ .mfb
sub C8 = C, LDC // coffset8 = c + 7 * ldc
mov f120 = f0
(p6) br.cond.dpnt .L020
}
;;
.align 16
.L011:
{ .mfb
LDFPD f48, f49 = [B]
mov f65 = f0
nop __LINE__
}
{ .mfb
adds BOFFSET = 2 * SIZE, B
mov f73 = f0
nop __LINE__
}
;;
{ .mfb
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
mov f81 = f0
nop __LINE__
}
{ .mfb
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
mov f89 = f0
nop __LINE__
}
;;
{ .mmf
LDFPD f52, f53 = [BOFFSET], 2 * SIZE
setf.d f97 = r0
mov f105 = f0
}
{ .mfb
setf.d f113 = r0
mov f121 = f0
nop __LINE__
}
;;
{ .mmf
LDFPD f54, f55 = [BOFFSET], 2 * SIZE
setf.d f66 = r0
mov f74 = f0
}
{ .mfb
setf.d f82 = r0
mov f90 = f0
nop __LINE__
}
;;
{ .mmf
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
setf.d f98 = r0
mov f106 = f0
}
{ .mfb
setf.d f114 = r0
mov f122 = f0
nop __LINE__
}
;;
{ .mmf
LDFPD f36, f37 = [AOFFSET], 2 * SIZE
setf.d f67 = r0
mov f75 = f0
}
{ .mfi
setf.d f83 = r0
mov f91 = f0
nop __LINE__
}
;;
{ .mmf
LDFPD f38, f39 = [AOFFSET], 2 * SIZE
setf.d f99 = r0
mov f107 = f0
}
{ .mfi
setf.d f115 = r0
mov f123 = f0
adds PREC = CPREFETCHSIZE * SIZE, C1
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f68 = r0
mov f76 = f0
}
{ .mfi
setf.d f84 = r0
mov f92 = f0
adds L = 1, K
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f100 = r0
mov f108 = f0
}
{ .mfi
setf.d f116 = r0
mov f124 = f0
adds PREA = (PREFETCHSIZE + 8) * SIZE, AOFFSET
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f69 = r0
mov f77 = f0
}
{ .mfi
setf.d f85 = r0
mov f93 = f0
adds PREB = (PREFETCHSIZE - 8) * SIZE, BOFFSET
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f101 = r0
mov f109 = f0
}
{ .mfi
setf.d f117 = r0
mov f125 = f0
tbit.z p12, p0 = L, 0
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f70 = r0
mov f78 = f0
}
{ .mfi
setf.d f86 = r0
mov f94 = f0
shr L = L, 1
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f102 = r0
mov f110 = f0
}
{ .mfi
setf.d f118 = r0
mov f126 = f0
adds L = -1, L
}
;;
{ .mmf
CPREFETCH [PREC], LDC
setf.d f71 = r0
mov f79 = f0
}
{ .mfi
setf.d f87 = r0
mov f95 = f0
mov ar.lc = L
}
;;
{ .mmf
CPREFETCH [PREC]
setf.d f103 = r0
mov f111 = f0
}
{ .mfi
setf.d f119 = r0
mov f127 = f0
cmp.eq p3, p0 = r0, r0
}
;;
.align 16
.L012:
/* 1 */
{ .mfi
lfetch.nt1 [PREA], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfi
(p12) cmp.ne p3, p0 = 0, L
FMA f72 = f32, f49, f72 // A1 * B2
nop __LINE__
}
;;
/* 2 */
{ .mfi
lfetch.nt1 [PREB], 16 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
nop __LINE__
}
{ .mfi
cmp.ne p4, p5 = 0, L
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
/* 3 */
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f96 = f32, f52, f96 // A1 * B5
nop __LINE__
}
{ .mfi
adds C9 = 4 * SIZE, C1
FMA f104 = f32, f53, f104 // A1 * B6
nop __LINE__
}
;;
/* 4 */
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f112 = f32, f54, f112 // A1 * B7
nop __LINE__
}
{ .mfi
adds C10 = 4 * SIZE, C2
FMA f120 = f32, f55, f120 // A1 * B8
nop __LINE__
}
;;
/* 5 */
{ .mfi
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
nop __LINE__
}
{ .mfi
adds C11 = 4 * SIZE, C3
FMA f73 = f33, f49, f73 // A2 * B2
nop __LINE__
}
;;
/* 6 */
{ .mfi
(p3) LDFPD f60, f61 = [BOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
nop __LINE__
}
{ .mfi
adds C12 = 4 * SIZE, C4
FMA f89 = f33, f51, f89 // A2 * B4
nop __LINE__
}
;;
/* 7 */
{ .mfi
(p3) LDFPD f62, f63 = [BOFFSET], 2 * SIZE
FMA f97 = f33, f52, f97 // A2 * B5
nop __LINE__
}
{ .mfi
adds C13 = 4 * SIZE, C5
FMA f105 = f33, f53, f105 // A2 * B6
nop __LINE__
}
;;
/* 8 */
{ .mfi
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f113 = f33, f54, f113 // A2 * B7
nop __LINE__
}
{ .mfi
adds C14 = 4 * SIZE, C6
FMA f121 = f33, f55, f121 // A2 * B8
nop __LINE__
}
;;
/* 9 */
{ .mfi
(p3) LDFPD f44, f45 = [AOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
nop __LINE__
}
{ .mfi
adds C15 = 4 * SIZE, C7
FMA f74 = f34, f49, f74 // A3 * B2
nop __LINE__
}
;;
/* 10 */
{ .mfi
(p3) LDFPD f46, f47 = [AOFFSET], 2 * SIZE
FMA f82 = f34, f50, f82 // A3 * B3
nop __LINE__
}
{ .mfi
adds C16 = 4 * SIZE, C8
FMA f90 = f34, f51, f90 // A3 * B4
nop __LINE__
}
;;
/* 11 */
{ .mfi
FMA f98 = f34, f52, f98 // A3 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f106 = f34, f53, f106 // A3 * B6
nop __LINE__
}
;;
/* 12 */
{ .mfi
FMA f114 = f34, f54, f114 // A3 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f122 = f34, f55, f122 // A3 * B8
nop __LINE__
}
;;
/* 13 */
{ .mfi
nop __LINE__
FMA f67 = f35, f48, f67 // A4 * B1
}
{ .mfi
nop __LINE__
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
;;
/* 14 */
{ .mfi
FMA f83 = f35, f50, f83 // A4 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f91 = f35, f51, f91 // A4 * B4
nop __LINE__
}
;;
/* 15 */
{ .mfi
FMA f99 = f35, f52, f99 // A4 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f107 = f35, f53, f107 // A4 * B6
nop __LINE__
}
;;
/* 16 */
{ .mfi
FMA f115 = f35, f54, f115 // A4 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f123 = f35, f55, f123 // A4 * B8
nop __LINE__
}
;;
/* 17 */
{ .mfi
nop __LINE__
FMA f68 = f36, f48, f68 // A5 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f76 = f36, f49, f76 // A5 * B2
nop __LINE__
}
;;
/* 18 */
{ .mfi
nop __LINE__
FMA f84 = f36, f50, f84 // A5 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f92 = f36, f51, f92 // A5 * B4
nop __LINE__
}
;;
/* 19 */
{ .mfi
nop __LINE__
FMA f100 = f36, f52, f100 // A5 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f108 = f36, f53, f108 // A5 * B6
nop __LINE__
}
;;
/* 20 */
{ .mfi
nop __LINE__
FMA f116 = f36, f54, f116 // A5 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f124 = f36, f55, f124 // A5 * B8
nop __LINE__
}
;;
/* 21 */
{ .mfi
nop __LINE__
FMA f69 = f37, f48, f69 // A6 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f77 = f37, f49, f77 // A6 * B2
nop __LINE__
}
;;
/* 22 */
{ .mfi
nop __LINE__
FMA f85 = f37, f50, f85 // A6 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f93 = f37, f51, f93 // A6 * B4
nop __LINE__
}
;;
/* 23 */
{ .mfi
nop __LINE__
FMA f101 = f37, f52, f101 // A6 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f109 = f37, f53, f109 // A6 * B6
nop __LINE__
}
;;
/* 24 */
{ .mfi
nop __LINE__
FMA f117 = f37, f54, f117 // A6 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f125 = f37, f55, f125 // A6 * B8
nop __LINE__
}
;;
/* 25 */
{ .mfi
nop __LINE__
FMA f70 = f38, f48, f70 // A7 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f78 = f38, f49, f78 // A7 * B2
nop __LINE__
}
;;
/* 26 */
{ .mfi
nop __LINE__
FMA f86 = f38, f50, f86 // A7 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f94 = f38, f51, f94 // A7 * B4
nop __LINE__
}
;;
/* 27 */
{ .mfi
nop __LINE__
FMA f102 = f38, f52, f102 // A7 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f110 = f38, f53, f110 // A7 * B6
nop __LINE__
}
;;
/* 28 */
{ .mfi
nop __LINE__
FMA f118 = f38, f54, f118 // A7 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f126 = f38, f55, f126 // A7 * B8
nop __LINE__
}
;;
/* 29 */
{ .mfi
nop __LINE__
FMA f71 = f39, f48, f71 // A8 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f79 = f39, f49, f79 // A8 * B2
nop __LINE__
}
;;
/* 30 */
{ .mfi
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
FMA f87 = f39, f50, f87 // A8 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f95 = f39, f51, f95 // A8 * B4
nop __LINE__
}
;;
/* 31 */
{ .mfi
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
FMA f103 = f39, f52, f103 // A8 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f111 = f39, f53, f111 // A8 * B6
nop __LINE__
}
;;
/* 32 */
{ .mfi
nop __LINE__
FMA f119 = f39, f54, f119 // A8 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f127 = f39, f55, f127 // A8 * B8
nop __LINE__
}
;;
/* 33 */
{ .mfi
nop __LINE__
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
/* 34 */
{ .mfi
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f88 = f40, f59, f88 // A1 * B4
nop __LINE__
}
;;
/* 35 */
{ .mfi
(p4) LDFPD f52, f53 = [BOFFSET], 2 * SIZE
(p3) FMA f96 = f40, f60, f96 // A1 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f104 = f40, f61, f104 // A1 * B6
nop __LINE__
}
;;
/* 36 */
{ .mfi
(p4) LDFPD f54, f55 = [BOFFSET], 2 * SIZE
(p3) FMA f112 = f40, f62, f112 // A1 * B7
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f120 = f40, f63, f120 // A1 * B8
nop __LINE__
}
;;
/* 37 */
{ .mfi
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
/* 38 */
{ .mfi
(p4) LDFPD f36, f37 = [AOFFSET], 2 * SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f89 = f41, f59, f89 // A2 * B4
nop __LINE__
}
;;
/* 39 */
{ .mfi
(p4) LDFPD f38, f39 = [AOFFSET], 2 * SIZE
(p3) FMA f97 = f41, f60, f97 // A2 * B5
nop __LINE__
}
{ .mfi
nop __LINE__
(p3) FMA f105 = f41, f61, f105 // A2 * B6
nop __LINE__
}
;;
/* 40 */
{ .mfi
(p5) LDFD f6 = [C1 ], SIZE
(p3) FMA f113 = f41, f62, f113 // A2 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f7 = [C9 ], SIZE
(p3) FMA f121 = f41, f63, f121 // A2 * B8
nop __LINE__
}
;;
/* 41 */
{ .mfi
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
/* 42 */
{ .mfi
(p5) LDFD f12 = [C1 ], SIZE
(p3) FMA f82 = f42, f58, f82 // A3 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f13 = [C9 ], SIZE
(p3) FMA f90 = f42, f59, f90 // A3 * B4
nop __LINE__
}
;;
/* 43 */
{ .mfi
(p5) LDFD f14 = [C1 ], 5 * SIZE
(p3) FMA f98 = f42, f60, f98 // A3 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f15 = [C9 ], 5 * SIZE
(p3) FMA f106 = f42, f61, f106 // A3 * B6
nop __LINE__
}
;;
/* 44 */
{ .mfi
(p5) LDFD f16 = [C1 ], SIZE
(p3) FMA f114 = f42, f62, f114 // A3 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f17 = [C9 ], SIZE
(p3) FMA f122 = f42, f63, f122 // A3 * B8
nop __LINE__
}
;;
/* 45 */
{ .mfi
(p5) LDFD f18 = [C1 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f19 = [C9 ], SIZE
(p3) FMA f75 = f43, f57, f75 // A4 * B2
nop __LINE__
}
;;
/* 46 */
{ .mfi
(p5) LDFD f20 = [C1 ], SIZE
(p3) FMA f83 = f43, f58, f83 // A4 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f21 = [C9 ], SIZE
(p3) FMA f91 = f43, f59, f91 // A4 * B4
nop __LINE__
}
;;
/* 47 */
{ .mfi
(p5) LDFD f22 = [C1 ], - 11 * SIZE
(p3) FMA f99 = f43, f60, f99 // A4 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f23 = [C9 ], - 11 * SIZE
(p3) FMA f107 = f43, f61, f107 // A4 * B6
nop __LINE__
}
;;
/* 48 */
{ .mfi
(p5) LDFD f24 = [C2 ], SIZE
(p3) FMA f115 = f43, f62, f115 // A4 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f25 = [C10], SIZE
(p3) FMA f123 = f43, f63, f123 // A4 * B8
nop __LINE__
}
;;
/* 49 */
{ .mfi
(p5) LDFD f26 = [C2 ], SIZE
(p3) FMA f68 = f44, f56, f68 // A5 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f27 = [C10], SIZE
(p3) FMA f76 = f44, f57, f76 // A5 * B2
nop __LINE__
}
;;
/* 50 */
{ .mfi
(p5) LDFD f28 = [C2 ], SIZE
(p3) FMA f84 = f44, f58, f84 // A5 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f29 = [C10], SIZE
(p3) FMA f92 = f44, f59, f92 // A5 * B4
nop __LINE__
}
;;
/* 51 */
{ .mfi
(p5) LDFD f30 = [C2 ], 5 * SIZE
(p3) FMA f100 = f44, f60, f100 // A5 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f31 = [C10], 5 * SIZE
(p3) FMA f108 = f44, f61, f108 // A5 * B6
nop __LINE__
}
;;
/* 52 */
{ .mfi
(p5) LDFD f32 = [C2 ], SIZE
(p3) FMA f116 = f44, f62, f116 // A5 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f33 = [C10], SIZE
(p3) FMA f124 = f44, f63, f124 // A5 * B8
nop __LINE__
}
;;
/* 53 */
{ .mfi
(p5) LDFD f34 = [C2 ], SIZE
(p3) FMA f69 = f45, f56, f69 // A6 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f35 = [C10], SIZE
(p3) FMA f77 = f45, f57, f77 // A6 * B2
nop __LINE__
}
;;
/* 54 */
{ .mfi
(p5) LDFD f36 = [C2 ], SIZE
(p3) FMA f85 = f45, f58, f85 // A6 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f37 = [C10], SIZE
(p3) FMA f93 = f45, f59, f93 // A6 * B4
nop __LINE__
}
;;
/* 55 */
{ .mfi
(p5) LDFD f38 = [C2 ], - 11 * SIZE
(p3) FMA f101 = f45, f60, f101 // A6 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f39 = [C10], - 11 * SIZE
(p3) FMA f109 = f45, f61, f109 // A6 * B6
nop __LINE__
}
;;
/* 56 */
{ .mfi
(p5) LDFD f48 = [C3 ], SIZE
(p3) FMA f117 = f45, f62, f117 // A6 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f49 = [C11], SIZE
(p3) FMA f125 = f45, f63, f125 // A6 * B8
nop __LINE__
}
;;
/* 57 */
{ .mfi
(p5) LDFD f50 = [C3 ], SIZE
(p3) FMA f70 = f46, f56, f70 // A7 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f51 = [C11], SIZE
(p3) FMA f78 = f46, f57, f78 // A7 * B2
nop __LINE__
}
;;
/* 58 */
{ .mfi
(p5) LDFD f52 = [C3 ], SIZE
(p3) FMA f86 = f46, f58, f86 // A7 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f53 = [C11], SIZE
(p3) FMA f94 = f46, f59, f94 // A7 * B4
nop __LINE__
}
;;
/* 59 */
{ .mfi
(p5) LDFD f54 = [C3 ], 5 * SIZE
(p3) FMA f102 = f46, f60, f102 // A7 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f55 = [C11], 5 * SIZE
(p3) FMA f110 = f46, f61, f110 // A7 * B6
nop __LINE__
}
;;
/* 60 */
{ .mfi
(p5) LDFD f40 = [C3 ], SIZE
(p3) FMA f118 = f46, f62, f118 // A7 * B7
nop __LINE__
}
{ .mfi
(p5) LDFD f41 = [C11], SIZE
(p3) FMA f126 = f46, f63, f126 // A7 * B8
nop __LINE__
}
;;
/* 61 */
{ .mfi
(p5) LDFD f42 = [C3 ], SIZE
(p3) FMA f71 = f47, f56, f71 // A8 * B1
nop __LINE__
}
{ .mfi
(p5) LDFD f43 = [C11], SIZE
(p3) FMA f79 = f47, f57, f79 // A8 * B2
nop __LINE__
}
;;
/* 62 */
{ .mfi
(p5) LDFD f44 = [C3 ], SIZE
(p3) FMA f87 = f47, f58, f87 // A8 * B3
nop __LINE__
}
{ .mfi
(p5) LDFD f45 = [C11], SIZE
(p3) FMA f95 = f47, f59, f95 // A8 * B4
nop __LINE__
}
;;
/* 63 */
{ .mfi
(p5) LDFD f46 = [C3 ], - 11 * SIZE
(p3) FMA f103 = f47, f60, f103 // A8 * B5
nop __LINE__
}
{ .mfi
(p5) LDFD f56 = [C11], - 11 * SIZE
(p3) FMA f111 = f47, f61, f111 // A8 * B6
nop __LINE__
}
;;
/* 64 */
{ .mfi
(p5) LDFD f57 = [C4 ], SIZE
(p3) FMA f119 = f47, f62, f119 // A8 * B7
adds L = -1, L
}
{ .mfb
(p5) LDFD f58 = [C12], SIZE
(p3) FMA f127 = f47, f63, f127 // A8 * B8
br.cloop.sptk.few .L012
}
;;
.L013:
{ .mmf
(p5) LDFD f59 = [C4 ], SIZE
(p5) LDFD f60 = [C12], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
cmp.ne p6, p0 = 1, I
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
(p5) LDFD f61 = [C4 ], SIZE
(p5) LDFD f62 = [C12], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
(p5) LDFD f63 = [C4 ], 5 * SIZE
(p5) LDFD f47 = [C12], 5 * SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mfi
(p5) LDFD f64 = [C4 ], SIZE
FMA f14 = ALPHA_I, f65, f14
nop __LINE__
}
{ .mfi
(p5) LDFD f65 = [C12], SIZE
FMA f15 = ALPHA_I, f67, f15
nop __LINE__
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f68, f16
}
{ .mmf
(p5) LDFD f6 = [C4 ], SIZE
(p5) LDFD f7 = [C12], SIZE
FMA f17 = ALPHA_R, f70, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f68, f18
}
{ .mmf
(p5) LDFD f10 = [C4 ], SIZE
(p5) LDFD f11 = [C12], SIZE
FMA f19 = ALPHA_I, f70, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f69, f20
}
{ .mmf
(p5) LDFD f12 = [C4 ], - 11 * SIZE
(p5) LDFD f13 = [C12], - 11 * SIZE
FMA f21 = ALPHA_R, f71, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f69, f22
}
{ .mmf
(p5) LDFD f14 = [C5 ], SIZE
(p5) LDFD f15 = [C13], SIZE
FMA f23 = ALPHA_I, f71, f23
}
;;
{ .mmf
STFD [C1 ] = f16, SIZE
STFD [C9 ] = f17, SIZE
FMA f24 = ALPHA_R, f72, f24
}
{ .mmf
(p5) LDFD f16 = [C5 ], SIZE
(p5) LDFD f17 = [C13], SIZE
FMA f25 = ALPHA_R, f74, f25
}
;;
{ .mmf
STFD [C1 ] = f18, SIZE
STFD [C9 ] = f19, SIZE
FMA f26 = ALPHA_I, f72, f26
}
{ .mmf
(p5) LDFD f18 = [C5 ], SIZE
(p5) LDFD f19 = [C13], SIZE
FMA f27 = ALPHA_I, f74, f27
}
;;
{ .mmf
STFD [C1 ] = f20, SIZE
STFD [C9 ] = f21, SIZE
FMA f28 = ALPHA_R, f73, f28
}
{ .mmf
(p5) LDFD f20 = [C5 ], 5 * SIZE
(p5) LDFD f21 = [C13], 5 * SIZE
FMA f29 = ALPHA_R, f75, f29
}
;;
{ .mmf
STFD [C1 ] = f22, 5 * SIZE
STFD [C9 ] = f23, 5 * SIZE
FMA f30 = ALPHA_I, f73, f30
}
{ .mmf
(p5) LDFD f22 = [C5 ], SIZE
(p5) LDFD f23 = [C13], SIZE
FMA f31 = ALPHA_I, f75, f31
}
;;
{ .mmf
STFD [C2 ] = f24, SIZE
STFD [C10] = f25, SIZE
FMA f32 = ALPHA_R, f76, f32
}
{ .mmf
(p5) LDFD f24 = [C5 ], SIZE
(p5) LDFD f25 = [C13], SIZE
FMA f33 = ALPHA_R, f78, f33
}
;;
{ .mmf
STFD [C2 ] = f26, SIZE
STFD [C10] = f27, SIZE
FMA f34 = ALPHA_I, f76, f34
}
{ .mmf
(p5) LDFD f26 = [C5 ], SIZE
(p5) LDFD f27 = [C13], SIZE
FMA f35 = ALPHA_I, f78, f35
}
;;
{ .mmf
STFD [C2 ] = f28, SIZE
STFD [C10] = f29, SIZE
FMA f36 = ALPHA_R, f77, f36
}
{ .mmf
(p5) LDFD f28 = [C5 ], - 11 * SIZE
(p5) LDFD f29 = [C13], - 11 * SIZE
FMA f37 = ALPHA_R, f79, f37
}
;;
{ .mmf
STFD [C2 ] = f30, 5 * SIZE
STFD [C10] = f31, 5 * SIZE
FMA f38 = ALPHA_I, f77, f38
}
{ .mmf
(p5) LDFD f30 = [C6 ], SIZE
(p5) LDFD f31 = [C14], SIZE
FMA f39 = ALPHA_I, f79, f39
}
;;
{ .mmf
STFD [C2 ] = f32, SIZE
STFD [C10] = f33, SIZE
FMA f48 = ALPHA_R, f80, f48
}
{ .mmf
(p5) LDFD f32 = [C6 ], SIZE
(p5) LDFD f33 = [C14], SIZE
FMA f49 = ALPHA_R, f82, f49
}
;;
{ .mmf
STFD [C2 ] = f34, SIZE
STFD [C10] = f35, SIZE
FMA f50 = ALPHA_I, f80, f50
}
{ .mmf
(p5) LDFD f34 = [C6 ], SIZE
(p5) LDFD f35 = [C14], SIZE
FMA f51 = ALPHA_I, f82, f51
}
;;
{ .mmf
STFD [C2 ] = f36, SIZE
STFD [C10] = f37, SIZE
FMA f52 = ALPHA_R, f81, f52
}
{ .mmf
(p5) LDFD f36 = [C6 ], 5 * SIZE
(p5) LDFD f37 = [C14], 5 * SIZE
FMA f53 = ALPHA_R, f83, f53
}
;;
{ .mmf
STFD [C2 ] = f38, 5 * SIZE
STFD [C10] = f39, 5 * SIZE
FMA f54 = ALPHA_I, f81, f54
}
{ .mmf
(p5) LDFD f38 = [C6 ], SIZE
(p5) LDFD f39 = [C14], SIZE
FMA f55 = ALPHA_I, f83, f55
}
;;
{ .mmf
STFD [C3 ] = f48, SIZE
STFD [C11] = f49, SIZE
FMA f40 = ALPHA_R, f84, f40
}
{ .mmf
(p5) LDFD f48 = [C6 ], SIZE
(p5) LDFD f49 = [C14], SIZE
FMA f41 = ALPHA_R, f86, f41
}
;;
{ .mmf
STFD [C3 ] = f50, SIZE
STFD [C11] = f51, SIZE
FMA f42 = ALPHA_I, f84, f42
}
{ .mmf
(p5) LDFD f50 = [C6 ], SIZE
(p5) LDFD f51 = [C14], SIZE
FMA f43 = ALPHA_I, f86, f43
}
;;
{ .mmf
STFD [C3 ] = f52, SIZE
STFD [C11] = f53, SIZE
FMA f44 = ALPHA_R, f85, f44
}
{ .mmf
(p5) LDFD f52 = [C6 ], - 11 * SIZE
(p5) LDFD f53 = [C14], - 11 * SIZE
FMA f45 = ALPHA_R, f87, f45
}
;;
{ .mmf
STFD [C3 ] = f54, 5 * SIZE
STFD [C11] = f55, 5 * SIZE
FMA f46 = ALPHA_I, f85, f46
}
{ .mmf
(p5) LDFD f54 = [C7 ], SIZE
(p5) LDFD f55 = [C15], SIZE
FMA f56 = ALPHA_I, f87, f56
}
;;
{ .mmf
STFD [C3 ] = f40, SIZE
STFD [C11] = f41, SIZE
FMA f57 = ALPHA_R, f88, f57
}
{ .mmf
(p5) LDFD f40 = [C7 ], SIZE
(p5) LDFD f41 = [C15], SIZE
FMA f58 = ALPHA_R, f90, f58
}
;;
{ .mmf
STFD [C3 ] = f42, SIZE
STFD [C11] = f43, SIZE
FMA f59 = ALPHA_I, f88, f59
}
{ .mmf
(p5) LDFD f42 = [C7 ], SIZE
(p5) LDFD f43 = [C15], SIZE
FMA f60 = ALPHA_I, f90, f60
}
;;
{ .mmf
STFD [C3 ] = f44, SIZE
STFD [C11] = f45, SIZE
FMA f61 = ALPHA_R, f89, f61
}
{ .mmf
(p5) LDFD f44 = [C7 ], 5 * SIZE
(p5) LDFD f45 = [C15], 5 * SIZE
FMA f62 = ALPHA_R, f91, f62
}
;;
{ .mmf
STFD [C3 ] = f46, 5 * SIZE
STFD [C11] = f56, 5 * SIZE
FMA f63 = ALPHA_I, f89, f63
}
{ .mmf
(p5) LDFD f46 = [C7 ], SIZE
(p5) LDFD f56 = [C15], SIZE
FMA f47 = ALPHA_I, f91, f47
}
;;
{ .mmf
STFD [C4 ] = f57, SIZE
STFD [C12] = f58, SIZE
FMA f64 = ALPHA_R, f92, f64
}
{ .mmf
(p5) LDFD f57 = [C7 ], SIZE
(p5) LDFD f58 = [C15], SIZE
FMA f65 = ALPHA_R, f94, f65
}
;;
{ .mmf
STFD [C4 ] = f59, SIZE
STFD [C12] = f60, SIZE
FMA f6 = ALPHA_I, f92, f6
}
{ .mmf
(p5) LDFD f59 = [C7 ], SIZE
(p5) LDFD f60 = [C15], SIZE
FMA f7 = ALPHA_I, f94, f7
}
;;
{ .mmf
STFD [C4 ] = f61, SIZE
STFD [C12] = f62, SIZE
FMA f10 = ALPHA_R, f93, f10
}
{ .mmf
(p5) LDFD f61 = [C7 ], - 11 * SIZE
(p5) LDFD f62 = [C15], - 11 * SIZE
FMA f11 = ALPHA_R, f95, f11
}
;;
{ .mmf
STFD [C4 ] = f63, 5 * SIZE
STFD [C12] = f47, 5 * SIZE
FMA f12 = ALPHA_I, f93, f12
}
{ .mmf
(p5) LDFD f63 = [C8 ], SIZE
(p5) LDFD f47 = [C16], SIZE
FMA f13 = ALPHA_I, f95, f13
}
;;
{ .mmf
STFD [C4 ] = f64, SIZE
STFD [C12] = f65, SIZE
FMA f14 = ALPHA_R, f96, f14
}
{ .mmf
(p5) LDFD f64 = [C8 ], SIZE
(p5) LDFD f65 = [C16], SIZE
FMA f15 = ALPHA_R, f98, f15
}
;;
{ .mmf
STFD [C4 ] = f6, SIZE
STFD [C12] = f7, SIZE
FMA f16 = ALPHA_I, f96, f16
}
{ .mmf
(p5) LDFD f6 = [C8 ], SIZE
(p5) LDFD f7 = [C16], SIZE
FMA f17 = ALPHA_I, f98, f17
}
;;
{ .mmf
STFD [C4 ] = f10, SIZE
STFD [C12] = f11, SIZE
FMA f18 = ALPHA_R, f97, f18
}
{ .mmf
(p5) LDFD f10 = [C8 ], 5 * SIZE
(p5) LDFD f11 = [C16], 5 * SIZE
FMA f19 = ALPHA_R, f99, f19
}
;;
{ .mmf
STFD [C4 ] = f12, 5 * SIZE
STFD [C12] = f13, 5 * SIZE
FMA f20 = ALPHA_I, f97, f20
}
{ .mmf
(p5) LDFD f12 = [C8 ], SIZE
(p5) LDFD f13 = [C16], SIZE
FMA f21 = ALPHA_I, f99, f21
}
;;
{ .mmf
STFD [C5 ] = f14, SIZE
STFD [C13] = f15, SIZE
FMA f22 = ALPHA_R, f100, f22
}
{ .mmf
(p5) LDFD f14 = [C8 ], SIZE
(p5) LDFD f15 = [C16], SIZE
FMA f23 = ALPHA_R, f102, f23
}
;;
{ .mmf
STFD [C5 ] = f16, SIZE
STFD [C13] = f17, SIZE
FMA f24 = ALPHA_I, f100, f24
}
{ .mmf
(p5) LDFD f16 = [C8 ], SIZE
(p5) LDFD f17 = [C16], SIZE
FMA f25 = ALPHA_I, f102, f25
}
;;
{ .mmf
STFD [C5 ] = f18, SIZE
STFD [C13] = f19, SIZE
FMA f26 = ALPHA_R, f101, f26
}
{ .mmf
(p5) LDFD f18 = [C8 ], - 11 * SIZE
(p5) LDFD f19 = [C16], - 11 * SIZE
FMA f27 = ALPHA_R, f103, f27
}
;;
{ .mmf
STFD [C5 ] = f20, 5 * SIZE
STFD [C13] = f21, 5 * SIZE
FMA f28 = ALPHA_I, f101, f28
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f29 = ALPHA_I, f103, f29
}
;;
{ .mmf
STFD [C5 ] = f22, SIZE
STFD [C13] = f23, SIZE
FMA f30 = ALPHA_R, f104, f30
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f31 = ALPHA_R, f106, f31
}
;;
{ .mmf
STFD [C5 ] = f24, SIZE
STFD [C13] = f25, SIZE
FMA f32 = ALPHA_I, f104, f32
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f33 = ALPHA_I, f106, f33
}
;;
{ .mmf
STFD [C5 ] = f26, SIZE
STFD [C13] = f27, SIZE
FMA f34 = ALPHA_R, f105, f34
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f35 = ALPHA_R, f107, f35
}
;;
{ .mmf
STFD [C5 ] = f28, 5 * SIZE
STFD [C13] = f29, 5 * SIZE
FMA f36 = ALPHA_I, f105, f36
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f37 = ALPHA_I, f107, f37
}
;;
{ .mmf
STFD [C6 ] = f30, SIZE
STFD [C14] = f31, SIZE
FMA f38 = ALPHA_R, f108, f38
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f39 = ALPHA_R, f110, f39
}
;;
{ .mmf
STFD [C6 ] = f32, SIZE
STFD [C14] = f33, SIZE
FMA f48 = ALPHA_I, f108, f48
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f49 = ALPHA_I, f110, f49
}
;;
{ .mmf
STFD [C6 ] = f34, SIZE
STFD [C14] = f35, SIZE
FMA f50 = ALPHA_R, f109, f50
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f51 = ALPHA_R, f111, f51
}
;;
{ .mmf
STFD [C6 ] = f36, 5 * SIZE
STFD [C14] = f37, 5 * SIZE
FMA f52 = ALPHA_I, f109, f52
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f53 = ALPHA_I, f111, f53
}
;;
{ .mmf
STFD [C6 ] = f38, SIZE
STFD [C14] = f39, SIZE
FMA f54 = ALPHA_R, f112, f54
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f55 = ALPHA_R, f114, f55
}
;;
{ .mmf
STFD [C6 ] = f48, SIZE
STFD [C14] = f49, SIZE
FMA f40 = ALPHA_I, f112, f40
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f41 = ALPHA_I, f114, f41
}
;;
{ .mmf
STFD [C6 ] = f50, SIZE
STFD [C14] = f51, SIZE
FMA f42 = ALPHA_R, f113, f42
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f43 = ALPHA_R, f115, f43
}
;;
{ .mmf
STFD [C6 ] = f52, 5 * SIZE
STFD [C14] = f53, 5 * SIZE
FMA f44 = ALPHA_I, f113, f44
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f45 = ALPHA_I, f115, f45
}
;;
{ .mmf
STFD [C7 ] = f54, SIZE
STFD [C15] = f55, SIZE
FMA f46 = ALPHA_R, f116, f46
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f56 = ALPHA_R, f118, f56
}
;;
{ .mmf
STFD [C7 ] = f40, SIZE
STFD [C15] = f41, SIZE
FMA f57 = ALPHA_I, f116, f57
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f58 = ALPHA_I, f118, f58
}
;;
{ .mmf
STFD [C7 ] = f42, SIZE
STFD [C15] = f43, SIZE
FMA f59 = ALPHA_R, f117, f59
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f60 = ALPHA_R, f119, f60
}
;;
{ .mmf
STFD [C7 ] = f44, 5 * SIZE
STFD [C15] = f45, 5 * SIZE
FMA f61 = ALPHA_I, f117, f61
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f62 = ALPHA_I, f119, f62
}
;;
{ .mmf
STFD [C7 ] = f46, SIZE
STFD [C15] = f56, SIZE
FMA f63 = ALPHA_R, f120, f63
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f47 = ALPHA_R, f122, f47
}
;;
{ .mmf
STFD [C7 ] = f57, SIZE
STFD [C15] = f58, SIZE
FMA f64 = ALPHA_I, f120, f64
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f65 = ALPHA_I, f122, f65
}
;;
{ .mmf
STFD [C7 ] = f59, SIZE
STFD [C15] = f60, SIZE
FMA f6 = ALPHA_R, f121, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f123, f7
}
;;
{ .mmf
STFD [C7 ] = f61, 5 * SIZE
STFD [C15] = f62, 5 * SIZE
FMA f10 = ALPHA_I, f121, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f123, f11
}
;;
{ .mmf
STFD [C8 ] = f63, SIZE
STFD [C16] = f47, SIZE
FMA f12 = ALPHA_R, f124, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f126, f13
}
;;
{ .mmf
STFD [C8 ] = f64, SIZE
STFD [C16] = f65, SIZE
FMA f14 = ALPHA_I, f124, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f126, f15
}
;;
{ .mmf
STFD [C8 ] = f6, SIZE
STFD [C16] = f7, SIZE
FMA f16 = ALPHA_R, f125, f16
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f17 = ALPHA_R, f127, f17
}
;;
{ .mmf
STFD [C8 ] = f10, 5 * SIZE
STFD [C16] = f11, 5 * SIZE
FMA f18 = ALPHA_I, f125, f18
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f19 = ALPHA_I, f127, f19
}
;;
{ .mmf
STFD [C8 ] = f12, SIZE
STFD [C16] = f13, SIZE
mov f64 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C8 ] = f14, SIZE
STFD [C16] = f15, SIZE
mov f80 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f88 = f0
}
;;
{ .mmf
STFD [C8 ] = f16, SIZE
STFD [C16] = f17, SIZE
mov f96 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f104 = f0
}
;;
{ .mmf
STFD [C8 ] = f18, 5 * SIZE
STFD [C16] = f19, 5 * SIZE
mov f112 = f0
}
{ .mfb
adds I = -1, I
mov f120 = f0
(p6) br.cond.dptk .L011
}
;;
.L020:
{ .mfi
cmp.eq p3, p0 = r0, r0
mov f89 = f0
tbit.z p6, p7 = M, 2
}
{ .mfb
nop __LINE__
mov f81 = f0
(p6) br.cond.dptk .L030
}
;;
{ .mfi
LDFPD f48, f49 = [B]
mov f65 = f0
nop __LINE__
}
{ .mfi
adds BOFFSET = 2 * SIZE, B
mov f73 = f0
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
}
;;
{ .mmf
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
setf.d f97 = r0
mov f105 = f0
}
{ .mfi
setf.d f113 = r0
mov f121 = f0
adds L = 1, K
}
;;
{ .mmf
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
setf.d f66 = r0
mov f74 = f0
}
{ .mfi
setf.d f82 = r0
mov f90 = f0
tbit.z p12, p0 = L, 0
}
;;
{ .mmf
LDFPD f52, f53 = [BOFFSET], 2 * SIZE
setf.d f98 = r0
mov f106 = f0
}
{ .mfi
setf.d f114 = r0
mov f122 = f0
shr L = L, 1
}
;;
{ .mfi
LDFPD f54, f55 = [BOFFSET], 2 * SIZE
mov f75 = f0
adds L = -1, L
}
{ .mmf
setf.d f67 = r0
setf.d f83 = r0
mov f91 = f0
}
;;
{ .mfi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
mov f107 = f0
mov ar.lc = L
}
{ .mmf
setf.d f99 = r0
setf.d f115 = r0
mov f123 = f0
}
;;
.align 32
.L022:
{ .mfi
lfetch.nt1 [PREA], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 16 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
cmp.ne p4, p5 = 0, L
}
{ .mfb
nop __LINE__
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f96 = f32, f52, f96 // A1 * B5
(p5) adds C9 = 4 * SIZE, C1
}
{ .mfi
nop __LINE__
FMA f104 = f32, f53, f104 // A1 * B6
(p5) adds C10 = 4 * SIZE, C2
}
;;
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f112 = f32, f54, f112 // A1 * B7
(p5) adds C11 = 4 * SIZE, C3
}
{ .mfi
nop __LINE__
FMA f120 = f32, f55, f120 // A1 * B8
(p5) adds C12 = 4 * SIZE, C4
}
;;
{ .mfi
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
(p5) adds C13 = 4 * SIZE, C5
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
(p5) adds C14 = 4 * SIZE, C6
}
;;
{ .mfi
(p3) LDFPD f60, f61 = [BOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
(p5) adds C15 = 4 * SIZE, C7
}
{ .mfi
nop __LINE__
FMA f89 = f33, f51, f89 // A2 * B4
(p5) adds C16 = 4 * SIZE, C8
}
;;
{ .mfb
(p3) LDFPD f62, f63 = [BOFFSET], 2 * SIZE
FMA f97 = f33, f52, f97 // A2 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f105 = f33, f53, f105 // A2 * B6
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f113 = f33, f54, f113 // A2 * B7
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f121 = f33, f55, f121 // A2 * B8
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f66 = f34, f48, f66 // A3 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f74 = f34, f49, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f82 = f34, f50, f82 // A3 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f90 = f34, f51, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f98 = f34, f52, f98 // A3 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f106 = f34, f53, f106 // A3 * B6
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f114 = f34, f54, f114 // A3 * B7
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f122 = f34, f55, f122 // A3 * B8
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f67 = f35, f48, f67 // A4 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f83 = f35, f50, f83 // A4 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f91 = f35, f51, f91 // A4 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
FMA f99 = f35, f52, f99 // A4 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f107 = f35, f53, f107 // A4 * B6
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
FMA f115 = f35, f54, f115 // A4 * B7
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f123 = f35, f55, f123 // A4 * B8
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f52, f53 = [BOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f88 = f40, f59, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f6 = [C1 ], SIZE
(p3) FMA f96 = f40, f60, f96 // A1 * B5
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
(p3) FMA f104 = f40, f61, f104 // A1 * B6
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f112 = f40, f62, f112 // A1 * B7
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f120 = f40, f63, f120 // A1 * B8
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f54, f55 = [BOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
{ .mfb
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
nop __LINE__
}
{ .mfb
(p3) FMA f89 = f41, f59, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f12 = [C1 ], SIZE
(p3) FMA f97 = f41, f60, f97 // A2 * B5
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C9 ], SIZE
(p3) FMA f105 = f41, f61, f105 // A2 * B6
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f14 = [C1 ], - 3 * SIZE
(p3) FMA f113 = f41, f62, f113 // A2 * B7
nop __LINE__
}
{ .mfb
(p5) LDFD f15 = [C9 ], - 3 * SIZE
(p3) FMA f121 = f41, f63, f121 // A2 * B8
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f16 = [C2 ], SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f17 = [C10], SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f18 = [C2 ], SIZE
(p3) FMA f82 = f42, f58, f82 // A3 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f19 = [C10], SIZE
(p3) FMA f90 = f42, f59, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f20 = [C2 ], SIZE
(p3) FMA f98 = f42, f60, f98 // A3 * B5
nop __LINE__
}
{ .mfb
(p5) LDFD f21 = [C10], SIZE
(p3) FMA f106 = f42, f61, f106 // A3 * B6
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f22 = [C2 ], - 3 * SIZE
(p3) FMA f114 = f42, f62, f114 // A3 * B7
nop __LINE__
}
{ .mfb
(p5) LDFD f23 = [C10], - 3 * SIZE
(p3) FMA f122 = f42, f63, f122 // A3 * B8
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f24 = [C3 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f25 = [C11], SIZE
(p3) FMA f75 = f43, f57, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f26 = [C3 ], SIZE
(p3) FMA f83 = f43, f58, f83 // A4 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f27 = [C11], SIZE
(p3) FMA f91 = f43, f59, f91 // A4 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f28 = [C3 ], SIZE
(p3) FMA f99 = f43, f60, f99 // A4 * B5
nop __LINE__
}
{ .mfb
(p5) LDFD f29 = [C11], SIZE
(p3) FMA f107 = f43, f61, f107 // A4 * B6
nop __LINE__
}
;;
{ .mfi
(p5) LDFD f30 = [C3 ], - 3 * SIZE
(p3) FMA f115 = f43, f62, f115 // A4 * B7
adds L = -1, L
}
{ .mfb
(p5) LDFD f31 = [C11], - 3 * SIZE
(p3) FMA f123 = f43, f63, f123 // A4 * B8
br.cloop.sptk.few .L022
}
;;
.L028:
{ .mmf
LDFD f68 = [C4 ], SIZE
LDFD f69 = [C12], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f70 = [C4 ], SIZE
LDFD f71 = [C12], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f76 = [C4 ], SIZE
LDFD f77 = [C12], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f78 = [C4 ], -3 * SIZE
LDFD f79 = [C12], -3 * SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f72, f16
}
{ .mmf
LDFD f84 = [C5 ], SIZE
LDFD f85 = [C13], SIZE
FMA f17 = ALPHA_R, f74, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f72, f18
}
{ .mmf
LDFD f86 = [C5 ], SIZE
LDFD f87 = [C13], SIZE
FMA f19 = ALPHA_I, f74, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f73, f20
}
{ .mmf
LDFD f92 = [C5 ], SIZE
LDFD f93 = [C13], SIZE
FMA f21 = ALPHA_R, f75, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f73, f22
}
{ .mmf
LDFD f94 = [C5 ], -3 * SIZE
LDFD f95 = [C13], -3 * SIZE
FMA f23 = ALPHA_I, f75, f23
}
;;
{ .mmf
STFD [C2 ] = f16, SIZE
STFD [C10] = f17, SIZE
FMA f24 = ALPHA_R, f80, f24
}
{ .mmf
LDFD f100 = [C6 ], SIZE
LDFD f101 = [C14], SIZE
FMA f25 = ALPHA_R, f82, f25
}
;;
{ .mmf
STFD [C2 ] = f18, SIZE
STFD [C10] = f19, SIZE
FMA f26 = ALPHA_I, f80, f26
}
{ .mmf
LDFD f102 = [C6 ], SIZE
LDFD f103 = [C14], SIZE
FMA f27 = ALPHA_I, f82, f27
}
;;
{ .mmf
STFD [C2 ] = f20, SIZE
STFD [C10] = f21, SIZE
FMA f28 = ALPHA_R, f81, f28
}
{ .mmf
LDFD f108 = [C6 ], SIZE
LDFD f109 = [C14], SIZE
FMA f29 = ALPHA_R, f83, f29
}
;;
{ .mmf
STFD [C2 ] = f22, 5 * SIZE
STFD [C10] = f23, 5 * SIZE
FMA f30 = ALPHA_I, f81, f30
}
{ .mmf
LDFD f110 = [C6 ], -3 * SIZE
LDFD f111 = [C14], -3 * SIZE
FMA f31 = ALPHA_I, f83, f31
}
;;
{ .mmf
STFD [C3 ] = f24, SIZE
STFD [C11] = f25, SIZE
FMA f68 = ALPHA_R, f88, f68
}
{ .mmf
LDFD f116 = [C7 ], SIZE
LDFD f117 = [C15], SIZE
FMA f69 = ALPHA_R, f90, f69
}
;;
{ .mmf
STFD [C3 ] = f26, SIZE
STFD [C11] = f27, SIZE
FMA f70 = ALPHA_I, f88, f70
}
{ .mmf
LDFD f118 = [C7 ], SIZE
LDFD f119 = [C15], SIZE
FMA f71 = ALPHA_I, f90, f71
}
;;
{ .mmf
STFD [C3 ] = f28, SIZE
STFD [C11] = f29, SIZE
FMA f76 = ALPHA_R, f89, f76
}
{ .mmf
LDFD f124 = [C7 ], SIZE
LDFD f125 = [C15], SIZE
FMA f77 = ALPHA_R, f91, f77
}
;;
{ .mmf
STFD [C3 ] = f30, 5 * SIZE
STFD [C11] = f31, 5 * SIZE
FMA f78 = ALPHA_I, f89, f78
}
{ .mmf
LDFD f126 = [C7 ], -3 * SIZE
LDFD f127 = [C15], -3 * SIZE
FMA f79 = ALPHA_I, f91, f79
}
;;
{ .mmf
STFD [C4 ] = f68, SIZE
STFD [C12] = f69, SIZE
FMA f84 = ALPHA_R, f96, f84
}
{ .mmf
LDFD f32 = [C8 ], SIZE
LDFD f33 = [C16], SIZE
FMA f85 = ALPHA_R, f98, f85
}
;;
{ .mmf
STFD [C4 ] = f70, SIZE
STFD [C12] = f71, SIZE
FMA f86 = ALPHA_I, f96, f86
}
{ .mmf
LDFD f34 = [C8 ], SIZE
LDFD f35 = [C16], SIZE
FMA f87 = ALPHA_I, f98, f87
}
;;
{ .mmf
STFD [C4 ] = f76, SIZE
STFD [C12] = f77, SIZE
FMA f92 = ALPHA_R, f97, f92
}
{ .mmf
LDFD f36 = [C8 ], SIZE
LDFD f37 = [C16], SIZE
FMA f93 = ALPHA_R, f99, f93
}
;;
{ .mmf
STFD [C4 ] = f78, 5 * SIZE
STFD [C12] = f79, 5 * SIZE
FMA f94 = ALPHA_I, f97, f94
}
{ .mmf
LDFD f38 = [C8 ], -3 * SIZE
LDFD f39 = [C16], -3 * SIZE
FMA f95 = ALPHA_I, f99, f95
}
;;
{ .mmf
STFD [C5 ] = f84, SIZE
STFD [C13] = f85, SIZE
FMA f100 = ALPHA_R, f104, f100
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f101 = ALPHA_R, f106, f101
}
;;
{ .mmf
STFD [C5 ] = f86, SIZE
STFD [C13] = f87, SIZE
FMA f102 = ALPHA_I, f104, f102
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f103 = ALPHA_I, f106, f103
}
;;
{ .mmf
STFD [C5 ] = f92, SIZE
STFD [C13] = f93, SIZE
FMA f108 = ALPHA_R, f105, f108
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f109 = ALPHA_R, f107, f109
}
;;
{ .mmf
STFD [C5 ] = f94, 5 * SIZE
STFD [C13] = f95, 5 * SIZE
FMA f110 = ALPHA_I, f105, f110
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f111 = ALPHA_I, f107, f111
}
;;
{ .mmf
STFD [C6 ] = f100, SIZE
STFD [C14] = f101, SIZE
FMA f116 = ALPHA_R, f112, f116
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f117 = ALPHA_R, f114, f117
}
;;
{ .mmf
STFD [C6 ] = f102, SIZE
STFD [C14] = f103, SIZE
FMA f118 = ALPHA_I, f112, f118
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f119 = ALPHA_I, f114, f119
}
;;
{ .mmf
STFD [C6 ] = f108, SIZE
STFD [C14] = f109, SIZE
FMA f124 = ALPHA_R, f113, f124
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f125 = ALPHA_R, f115, f125
}
;;
{ .mmf
STFD [C6 ] = f110, 5 * SIZE
STFD [C14] = f111, 5 * SIZE
FMA f126 = ALPHA_I, f113, f126
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f127 = ALPHA_I, f115, f127
}
;;
{ .mmf
STFD [C7 ] = f116, SIZE
STFD [C15] = f117, SIZE
FMA f32 = ALPHA_R, f120, f32
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f33 = ALPHA_R, f122, f33
}
;;
{ .mmf
STFD [C7 ] = f118, SIZE
STFD [C15] = f119, SIZE
FMA f34 = ALPHA_I, f120, f34
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f35 = ALPHA_I, f122, f35
}
;;
{ .mmf
STFD [C7 ] = f124, SIZE
STFD [C15] = f125, SIZE
FMA f36 = ALPHA_R, f121, f36
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f37 = ALPHA_R, f123, f37
}
;;
{ .mmf
STFD [C7 ] = f126, 5 * SIZE
STFD [C15] = f127, 5 * SIZE
FMA f38 = ALPHA_I, f121, f38
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f39 = ALPHA_I, f123, f39
}
;;
{ .mmf
STFD [C8 ] = f32, SIZE
STFD [C16] = f33, SIZE
mov f64 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C8 ] = f34, SIZE
STFD [C16] = f35, SIZE
mov f80 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f88 = f0
}
;;
{ .mmf
STFD [C8 ] = f36, SIZE
STFD [C16] = f37, SIZE
mov f96 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f104 = f0
}
;;
{ .mmf
STFD [C8 ] = f38, 5 * SIZE
STFD [C16] = f39, 5 * SIZE
mov f112 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f120 = f0
}
;;
.align 32
.L030:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 1
(p6) br.cond.dptk .L040
}
;;
{ .mfi
LDFPD f48, f49 = [B]
mov f65 = f0
nop __LINE__
}
{ .mfi
adds BOFFSET = 2 * SIZE, B
mov f73 = f0
adds L = 1, K
}
;;
{ .mfi
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
mov f81 = f0
tbit.z p12, p0 = L, 0
}
{ .mfi
(p7) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
mov f89 = f0
shr L = L, 1
}
;;
{ .mfi
LDFPD f52, f53 = [BOFFSET], 2 * SIZE
mov f97 = f0
adds L = -1, L
}
{ .mfi
nop __LINE__
mov f105 = f0
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
}
;;
{ .mfi
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
mov f113 = f0
mov ar.lc = L
}
{ .mfi
LDFPD f54, f55 = [BOFFSET], 2 * SIZE
mov f121 = f0
cmp.eq p3, p0 = r0, r0
}
;;
.align 32
.L032:
{ .mfb
lfetch.nt1 [PREA], 4 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 16 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
cmp.ne p4, p5 = 0, L
}
{ .mfb
nop __LINE__
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f96 = f32, f52, f96 // A1 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f104 = f32, f53, f104 // A1 * B6
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f112 = f32, f54, f112 // A1 * B7
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f120 = f32, f55, f120 // A1 * B8
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f60, f61 = [BOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f89 = f33, f51, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f62, f63 = [BOFFSET], 2 * SIZE
FMA f97 = f33, f52, f97 // A2 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f105 = f33, f53, f105 // A2 * B6
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f113 = f33, f54, f113 // A2 * B7
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f121 = f33, f55, f121 // A2 * B8
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f88 = f40, f59, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f52, f53 = [BOFFSET], 2 * SIZE
(p3) FMA f96 = f40, f60, f96 // A1 * B5
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f104 = f40, f61, f104 // A1 * B6
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f6 = [C1], SIZE
(p3) FMA f112 = f40, f62, f112 // A1 * B7
nop __LINE__
}
{ .mfb
(p5) LDFD f12 = [C2], SIZE
(p3) FMA f120 = f40, f63, f120 // A1 * B8
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f54, f55 = [BOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C1], SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C2], SIZE
(p3) FMA f89 = f41, f59, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f10 = [C1], SIZE
(p3) FMA f97 = f41, f60, f97 // A2 * B5
nop __LINE__
}
{ .mfb
(p5) LDFD f14 = [C2], SIZE
(p3) FMA f105 = f41, f61, f105 // A2 * B6
nop __LINE__
}
;;
{ .mfi
(p5) LDFD f11 = [C1], -3 * SIZE
(p3) FMA f113 = f41, f62, f113 // A2 * B7
adds L = -1, L
}
{ .mfb
(p5) LDFD f15 = [C2], -3 * SIZE
(p3) FMA f121 = f41, f63, f121 // A2 * B8
br.cloop.sptk.few .L032
}
;;
.L038:
{ .mmf
LDFD f16 = [C3], SIZE
LDFD f20 = [C4], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f72, f12
}
;;
{ .mmf
LDFD f17 = [C3], SIZE
LDFD f21 = [C4], SIZE
FMA f7 = ALPHA_I, f64, f7
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_I, f72, f13
}
;;
{ .mmf
LDFD f18 = [C3], SIZE
LDFD f22 = [C4], SIZE
FMA f10 = ALPHA_R, f65, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_R, f73, f14
}
;;
{ .mmf
LDFD f19 = [C3], - 3 * SIZE
LDFD f23 = [C4], - 3 * SIZE
FMA f11 = ALPHA_I, f65, f11
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f73, f15
}
;;
{ .mmf
STFD [C1] = f6, SIZE
STFD [C2] = f12, SIZE
FMA f16 = ALPHA_R, f80, f16
}
{ .mmf
LDFD f24 = [C5], SIZE
LDFD f28 = [C6], SIZE
FMA f20 = ALPHA_R, f88, f20
}
;;
{ .mmf
STFD [C1] = f7, SIZE
STFD [C2] = f13, SIZE
FMA f17 = ALPHA_I, f80, f17
}
{ .mmf
LDFD f25 = [C5], SIZE
LDFD f29 = [C6], SIZE
FMA f21 = ALPHA_I, f88, f21
}
;;
{ .mmf
STFD [C1] = f10, SIZE
STFD [C2] = f14, SIZE
FMA f18 = ALPHA_R, f81, f18
}
{ .mmf
LDFD f26 = [C5], SIZE
LDFD f30 = [C6], SIZE
FMA f22 = ALPHA_R, f89, f22
}
;;
{ .mmf
STFD [C1] = f11, SIZE
STFD [C2] = f15, SIZE
FMA f19 = ALPHA_I, f81, f19
}
{ .mmf
LDFD f27 = [C5], - 3 * SIZE
LDFD f31 = [C6], - 3 * SIZE
FMA f23 = ALPHA_I, f89, f23
}
;;
{ .mmf
STFD [C3] = f16, SIZE
STFD [C4] = f20, SIZE
FMA f24 = ALPHA_R, f96, f24
}
{ .mmf
LDFD f32 = [C7], SIZE
LDFD f36 = [C8], SIZE
FMA f28 = ALPHA_R, f104, f28
}
;;
{ .mmf
STFD [C3] = f17, SIZE
STFD [C4] = f21, SIZE
FMA f25 = ALPHA_I, f96, f25
}
{ .mmf
LDFD f33 = [C7], SIZE
LDFD f37 = [C8], SIZE
FMA f29 = ALPHA_I, f104, f29
}
;;
{ .mmf
STFD [C3] = f18, SIZE
STFD [C4] = f22, SIZE
FMA f26 = ALPHA_R, f97, f26
}
{ .mmf
LDFD f34 = [C7], SIZE
LDFD f38 = [C8], SIZE
FMA f30 = ALPHA_R, f105, f30
}
;;
{ .mmf
STFD [C3] = f19, SIZE
STFD [C4] = f23, SIZE
FMA f27 = ALPHA_I, f97, f27
}
{ .mmf
LDFD f35 = [C7], - 3 * SIZE
LDFD f39 = [C8], - 3 * SIZE
FMA f31 = ALPHA_I, f105, f31
}
;;
{ .mmf
STFD [C5] = f24, SIZE
STFD [C6] = f28, SIZE
FMA f32 = ALPHA_R, f112, f32
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f36 = ALPHA_R, f120, f36
}
;;
{ .mmf
STFD [C5] = f25, SIZE
STFD [C6] = f29, SIZE
FMA f33 = ALPHA_I, f112, f33
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f37 = ALPHA_I, f120, f37
}
;;
{ .mmf
STFD [C5] = f26, SIZE
STFD [C6] = f30, SIZE
FMA f34 = ALPHA_R, f113, f34
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f38 = ALPHA_R, f121, f38
}
;;
{ .mmf
STFD [C5] = f27, SIZE
STFD [C6] = f31, SIZE
FMA f35 = ALPHA_I, f113, f35
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f39 = ALPHA_I, f121, f39
}
;;
{ .mmf
STFD [C7] = f32, SIZE
STFD [C8] = f36, SIZE
mov f64 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C7] = f33, SIZE
STFD [C8] = f37, SIZE
mov f80 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f88 = f0
}
;;
{ .mmf
STFD [C7] = f34, SIZE
STFD [C8] = f38, SIZE
mov f96 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f104 = f0
}
;;
{ .mmf
STFD [C7] = f35, SIZE
STFD [C8] = f39, SIZE
mov f112 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f120 = f0
}
;;
.align 32
.L040:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 0
(p6) br.cond.dptk .L049
}
;;
{ .mmi
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
adds L = 1, K
}
;;
{ .mii
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
LDFPD f52, f53 = [BOFFSET], 2 * SIZE
LDFD f32 = [AOFFSET], 1 * SIZE
adds L = -1, L
}
;;
{ .mmi
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
cmp.eq p3, p0 = r0, r0
mov ar.lc = L
}
{ .mmi
LDFPD f54, f55 = [BOFFSET], 2 * SIZE
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
nop __LINE__
}
;;
.align 32
.L042:
{ .mfb
lfetch.nt1 [PREB], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p12) cmp.ne p3, p0 = 0, L
FMA f72 = f32, f49, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfi
(p3) LDFD f40 = [AOFFSET], 1 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
cmp.ne p4, p5 = 0, L
}
{ .mfb
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfi
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f96 = f32, f52, f96 // A1 * B5
nop __LINE__
}
{ .mmf
(p5) LDFD f6 = [C1], SIZE
(p5) LDFD f10 = [C2], SIZE
FMA f104 = f32, f53, f104 // A1 * B6
}
;;
{ .mfi
(p3) LDFPD f60, f61 = [BOFFSET], 2 * SIZE
FMA f112 = f32, f54, f112 // A1 * B7
nop __LINE__
}
{ .mmf
(p5) LDFD f7 = [C1], -SIZE
(p5) LDFD f11 = [C2], -SIZE
FMA f120 = f32, f55, f120 // A1 * B8
}
;;
{ .mmf
(p3) LDFPD f62, f63 = [BOFFSET], 2 * SIZE
(p4) LDFD f32 = [AOFFSET], 1 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
}
{ .mmf
(p5) LDFD f12 = [C3], SIZE
(p5) LDFD f14 = [C4], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
}
;;
{ .mfi
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mmf
(p5) LDFD f13 = [C3], -SIZE
(p5) LDFD f15 = [C4], -SIZE
(p3) FMA f88 = f40, f59, f88 // A1 * B4
}
;;
{ .mfi
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f96 = f40, f60, f96 // A1 * B5
nop __LINE__
}
{ .mmf
(p5) LDFD f16 = [C5], SIZE
(p5) LDFD f18 = [C6], SIZE
(p3) FMA f104 = f40, f61, f104 // A1 * B6
}
;;
{ .mfi
(p4) LDFPD f52, f53 = [BOFFSET], 2 * SIZE
(p3) FMA f112 = f40, f62, f112 // A1 * B7
adds L = -1, L
}
{ .mmb
(p5) LDFD f17 = [C5], -SIZE
(p5) LDFD f19 = [C6], -SIZE
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f54, f55 = [BOFFSET], 2 * SIZE
(p3) FMA f120 = f40, f63, f120 // A1 * B8
nop __LINE__
}
{ .mmb
(p5) LDFD f20 = [C7], SIZE
(p5) LDFD f22 = [C8], SIZE
br.cloop.sptk.few .L042
}
;;
{ .mmf
LDFD f21 = [C7], -SIZE
LDFD f23 = [C8], -SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f10 = ALPHA_R, f72, f10
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_I, f64, f7
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f72, f11
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f80, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_R, f88, f14
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_I, f80, f13
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f88, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C2 ] = f10, SIZE
FMA f16 = ALPHA_R, f96, f16
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f18 = ALPHA_R, f104, f18
}
;;
{ .mmf
STFD [C1 ] = f7, SIZE
STFD [C2 ] = f11, SIZE
FMA f17 = ALPHA_I, f96, f17
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f19 = ALPHA_I, f104, f19
}
;;
{ .mmf
STFD [C3 ] = f12, SIZE
STFD [C4 ] = f14, SIZE
FMA f20 = ALPHA_R, f112, f20
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f22 = ALPHA_R, f120, f22
}
;;
{ .mmf
STFD [C3 ] = f13, SIZE
STFD [C4 ] = f15, SIZE
FMA f21 = ALPHA_I, f112, f21
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f23 = ALPHA_I, f120, f23
}
;;
{ .mmi
STFD [C5 ] = f16, SIZE
STFD [C6 ] = f18, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C5 ] = f17, SIZE
STFD [C6 ] = f19, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C7 ] = f20, SIZE
STFD [C8 ] = f22, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C7 ] = f21, SIZE
STFD [C8 ] = f23, SIZE
nop __LINE__
}
;;
.align 32
.L049:
{ .mmi
mov B = BOFFSET
mov AOFFSET = A
nop __LINE__
}
;;
{ .mmb
nop __LINE__
cmp.lt p6, p0 = 0, J
(p6) br.cond.dptk .L010
}
;;
.align 32
.L050:
{ .mfi
mov C1 = C
mov f64 = f0
tbit.z p6, p0 = N, 2
}
{ .mfi
add C2 = LDC, C
mov f72 = f0
shr I = M, 3
}
;;
{ .mfi
shladd C3 = LDC, 1, C
mov f80 = f0
nop __LINE__
}
{ .mfb
mov AOFFSET = A
mov f88 = f0
(p6) br.cond.dpnt .L090
}
;;
{ .mfi
cmp.eq p6, p7 = 0, I
mov f65 = f0
nop __LINE__
}
{ .mfi
shladd C4 = LDC, 1, C2
mov f73 = f0
nop __LINE__
}
;;
{ .mfi
nop __LINE__
mov f81 = f0
nop __LINE__
}
{ .mfb
shladd C = LDC, 2, C
mov f89 = f0
(p6) br.cond.dpnt .L060
}
;;
.align 32
.L052:
{ .mfb
LDFPD f48, f49 = [B]
mov f66 = f0
nop __LINE__
}
{ .mfb
adds BOFFSET = 2 * SIZE, B
mov f74 = f0
nop __LINE__
}
;;
{ .mfi
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
mov f82 = f0
nop __LINE__
}
{ .mfi
setf.d f84 = r0
mov f90 = f0
nop __LINE__
}
;;
{ .mfi
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
mov f67 = f0
adds PREC = CPREFETCHSIZE * SIZE, C1
}
{ .mfi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
mov f75 = f0
adds L = 1, K
}
;;
{ .mfi
LDFPD f36, f37 = [AOFFSET], 2 * SIZE
mov f83 = f0
tbit.z p12, p0 = L, 0
}
{ .mfi
setf.d f91 = r0
mov f68 = f0
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
}
;;
{ .mfi
CPREFETCH [PREC], LDC
mov f76 = f0
adds PREA = (PREFETCHSIZE + 8) * SIZE, AOFFSET
}
{ .mfi
LDFPD f38, f39 = [AOFFSET], 2 * SIZE
mov f92 = f0
cmp.eq p3, p0 = r0, r0
}
;;
{ .mfi
CPREFETCH [PREC], LDC
mov f69 = f0
shr L = L, 1
}
{ .mmf
setf.d f77 = r0
setf.d f85 = r0
mov f93 = f0
}
;;
{ .mfi
CPREFETCH [PREC], LDC
mov f70 = f0
adds L = -1, L
}
{ .mmf
setf.d f78 = r0
setf.d f86 = r0
mov f94 = f0
}
;;
{ .mfi
CPREFETCH [PREC]
mov f71 = f0
mov ar.lc = L
}
{ .mmf
setf.d f79 = r0
setf.d f87 = r0
mov f95 = f0
}
;;
.align 32
.L053:
{ .mfb
lfetch.nt1 [PREA], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 8 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
cmp.ne p4, p5 = 0, L
}
{ .mfi
nop __LINE__
FMA f88 = f32, f51, f88 // A1 * B4
adds C9 = 4 * SIZE, C1
}
;;
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
adds C10 = 4 * SIZE, C2
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
adds C11 = 4 * SIZE, C3
}
;;
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
adds C12 = 4 * SIZE, C4
}
{ .mfb
nop __LINE__
FMA f89 = f33, f51, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f74 = f34, f49, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f82 = f34, f50, f82 // A3 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f90 = f34, f51, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f44, f45 = [AOFFSET], 2 * SIZE
FMA f67 = f35, f48, f67 // A4 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f46, f47 = [AOFFSET], 2 * SIZE
FMA f83 = f35, f50, f83 // A4 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f91 = f35, f51, f91 // A4 * B4
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f68 = f36, f48, f68 // A5 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f76 = f36, f49, f76 // A5 * B2
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f84 = f36, f50, f84 // A5 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f92 = f36, f51, f92 // A5 * B4
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f69 = f37, f48, f69 // A6 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f77 = f37, f49, f77 // A6 * B2
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f85 = f37, f50, f85 // A6 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f93 = f37, f51, f93 // A6 * B4
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f70 = f38, f48, f70 // A7 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f78 = f38, f49, f78 // A7 * B2
nop __LINE__
}
;;
{ .mfb
nop __LINE__
FMA f86 = f38, f50, f86 // A7 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f94 = f38, f51, f94 // A7 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
FMA f71 = f39, f48, f71 // A8 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f79 = f39, f49, f79 // A8 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
FMA f87 = f39, f50, f87 // A8 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f95 = f39, f51, f95 // A8 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f88 = f40, f59, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f36, f37 = [AOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f38, f39 = [AOFFSET], 2 * SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f89 = f41, f59, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f6 = [C1 ], SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f82 = f42, f58, f82 // A3 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f90 = f42, f59, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f12 = [C1 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C9 ], SIZE
(p3) FMA f75 = f43, f57, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f14 = [C1 ], 5 * SIZE
(p3) FMA f83 = f43, f58, f83 // A4 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f15 = [C9 ], 5 * SIZE
(p3) FMA f91 = f43, f59, f91 // A4 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f16 = [C1 ], SIZE
(p3) FMA f68 = f44, f56, f68 // A5 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f17 = [C9], SIZE
(p3) FMA f76 = f44, f57, f76 // A5 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f18 = [C1 ], SIZE
(p3) FMA f84 = f44, f58, f84 // A5 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f19 = [C9], SIZE
(p3) FMA f92 = f44, f59, f92 // A5 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f20 = [C1 ], SIZE
(p3) FMA f69 = f45, f56, f69 // A6 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f21 = [C9], SIZE
(p3) FMA f77 = f45, f57, f77 // A6 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f22 = [C1 ], -11 * SIZE
(p3) FMA f85 = f45, f58, f85 // A6 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f23 = [C9 ], -11 * SIZE
(p3) FMA f93 = f45, f59, f93 // A6 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f24 = [C2 ], SIZE
(p3) FMA f70 = f46, f56, f70 // A7 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f25 = [C10], SIZE
(p3) FMA f78 = f46, f57, f78 // A7 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f26 = [C2 ], SIZE
(p3) FMA f86 = f46, f58, f86 // A7 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f27 = [C10], SIZE
(p3) FMA f94 = f46, f59, f94 // A7 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f28 = [C2 ], SIZE
(p3) FMA f71 = f47, f56, f71 // A8 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f29 = [C10], SIZE
(p3) FMA f79 = f47, f57, f79 // A8 * B2
nop __LINE__
}
;;
{ .mfi
(p5) LDFD f30 = [C2 ], 5 * SIZE
(p3) FMA f87 = f47, f58, f87 // A8 * B3
adds L = -1, L
}
{ .mfb
(p5) LDFD f31 = [C10], 5 * SIZE
(p3) FMA f95 = f47, f59, f95 // A8 * B4
br.cloop.sptk.few .L053
}
;;
.align 32
.L058:
{ .mmf
LDFD f32 = [C2 ], SIZE
LDFD f33 = [C10], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f34 = [C2 ], SIZE
LDFD f35 = [C10], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f36 = [C2 ], SIZE
LDFD f37 = [C10], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f38 = [C2 ], - 11 * SIZE
LDFD f39 = [C10], - 11 * SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f68, f16
}
{ .mmf
LDFD f48 = [C3 ], SIZE
LDFD f49 = [C11], SIZE
FMA f17 = ALPHA_R, f70, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f68, f18
}
{ .mmf
LDFD f50 = [C3 ], SIZE
LDFD f51 = [C11], SIZE
FMA f19 = ALPHA_I, f70, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f69, f20
}
{ .mmf
LDFD f52 = [C3 ], SIZE
LDFD f53 = [C11], SIZE
FMA f21 = ALPHA_R, f71, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f69, f22
}
{ .mmf
LDFD f54 = [C3 ], 5 * SIZE
LDFD f55 = [C11], 5 * SIZE
FMA f23 = ALPHA_I, f71, f23
}
;;
{ .mmf
STFD [C1 ] = f16, SIZE
STFD [C9 ] = f17, SIZE
FMA f24 = ALPHA_R, f72, f24
}
{ .mmf
LDFD f40 = [C3 ], SIZE
LDFD f41 = [C11], SIZE
FMA f25 = ALPHA_R, f74, f25
}
;;
{ .mmf
STFD [C1 ] = f18, SIZE
STFD [C9 ] = f19, SIZE
FMA f26 = ALPHA_I, f72, f26
}
{ .mmf
LDFD f42 = [C3 ], SIZE
LDFD f43 = [C11], SIZE
FMA f27 = ALPHA_I, f74, f27
}
;;
{ .mmf
STFD [C1 ] = f20, SIZE
STFD [C9 ] = f21, SIZE
FMA f28 = ALPHA_R, f73, f28
}
{ .mmf
LDFD f44 = [C3 ], SIZE
LDFD f45 = [C11], SIZE
FMA f29 = ALPHA_R, f75, f29
}
;;
{ .mmf
STFD [C1 ] = f22, 5 * SIZE
STFD [C9 ] = f23, 5 * SIZE
FMA f30 = ALPHA_I, f73, f30
}
{ .mmf
LDFD f46 = [C3 ], - 11 * SIZE
LDFD f56 = [C11], - 11 * SIZE
FMA f31 = ALPHA_I, f75, f31
}
;;
{ .mmf
STFD [C2 ] = f24, SIZE
STFD [C10] = f25, SIZE
FMA f32 = ALPHA_R, f76, f32
}
{ .mmf
LDFD f57 = [C4 ], SIZE
LDFD f58 = [C12], SIZE
FMA f33 = ALPHA_R, f78, f33
}
;;
{ .mmf
STFD [C2 ] = f26, SIZE
STFD [C10] = f27, SIZE
FMA f34 = ALPHA_I, f76, f34
}
{ .mmf
LDFD f59 = [C4 ], SIZE
LDFD f60 = [C12], SIZE
FMA f35 = ALPHA_I, f78, f35
}
;;
{ .mmf
STFD [C2 ] = f28, SIZE
STFD [C10] = f29, SIZE
FMA f36 = ALPHA_R, f77, f36
}
{ .mmf
LDFD f61 = [C4 ], SIZE
LDFD f62 = [C12], SIZE
FMA f37 = ALPHA_R, f79, f37
}
;;
{ .mmf
STFD [C2 ] = f30, 5 * SIZE
STFD [C10] = f31, 5 * SIZE
FMA f38 = ALPHA_I, f77, f38
}
{ .mmf
LDFD f63 = [C4 ], 5 * SIZE
LDFD f47 = [C12], 5 * SIZE
FMA f39 = ALPHA_I, f79, f39
}
;;
{ .mmf
STFD [C2 ] = f32, SIZE
STFD [C10] = f33, SIZE
FMA f48 = ALPHA_R, f80, f48
}
{ .mmf
LDFD f64 = [C4 ], SIZE
LDFD f65 = [C12], SIZE
FMA f49 = ALPHA_R, f82, f49
}
;;
{ .mmf
STFD [C2 ] = f34, SIZE
STFD [C10] = f35, SIZE
FMA f50 = ALPHA_I, f80, f50
}
{ .mmf
LDFD f6 = [C4 ], SIZE
LDFD f7 = [C12], SIZE
FMA f51 = ALPHA_I, f82, f51
}
;;
{ .mmf
STFD [C2 ] = f36, SIZE
STFD [C10] = f37, SIZE
FMA f52 = ALPHA_R, f81, f52
}
{ .mmf
LDFD f10 = [C4 ], SIZE
LDFD f11 = [C12], SIZE
FMA f53 = ALPHA_R, f83, f53
}
;;
{ .mmf
STFD [C2 ] = f38, 5 * SIZE
STFD [C10] = f39, 5 * SIZE
FMA f54 = ALPHA_I, f81, f54
}
{ .mmf
LDFD f12 = [C4 ], - 11 * SIZE
LDFD f13 = [C12], - 11 * SIZE
FMA f55 = ALPHA_I, f83, f55
}
;;
{ .mmf
STFD [C3 ] = f48, SIZE
STFD [C11] = f49, SIZE
FMA f40 = ALPHA_R, f84, f40
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f41 = ALPHA_R, f86, f41
}
;;
{ .mmf
STFD [C3 ] = f50, SIZE
STFD [C11] = f51, SIZE
FMA f42 = ALPHA_I, f84, f42
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f43 = ALPHA_I, f86, f43
}
;;
{ .mmf
STFD [C3 ] = f52, SIZE
STFD [C11] = f53, SIZE
FMA f44 = ALPHA_R, f85, f44
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f45 = ALPHA_R, f87, f45
}
;;
{ .mmf
STFD [C3 ] = f54, 5 * SIZE
STFD [C11] = f55, 5 * SIZE
FMA f46 = ALPHA_I, f85, f46
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f56 = ALPHA_I, f87, f56
}
;;
{ .mmf
STFD [C3 ] = f40, SIZE
STFD [C11] = f41, SIZE
FMA f57 = ALPHA_R, f88, f57
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f58 = ALPHA_R, f90, f58
}
;;
{ .mmf
STFD [C3 ] = f42, SIZE
STFD [C11] = f43, SIZE
FMA f59 = ALPHA_I, f88, f59
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f60 = ALPHA_I, f90, f60
}
;;
{ .mmf
STFD [C3 ] = f44, SIZE
STFD [C11] = f45, SIZE
FMA f61 = ALPHA_R, f89, f61
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f62 = ALPHA_R, f91, f62
}
;;
{ .mmf
STFD [C3 ] = f46, 5 * SIZE
STFD [C11] = f56, 5 * SIZE
FMA f63 = ALPHA_I, f89, f63
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f47 = ALPHA_I, f91, f47
}
;;
{ .mmf
STFD [C4 ] = f57, SIZE
STFD [C12] = f58, SIZE
FMA f64 = ALPHA_R, f92, f64
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f65 = ALPHA_R, f94, f65
}
;;
{ .mmf
STFD [C4 ] = f59, SIZE
STFD [C12] = f60, SIZE
FMA f6 = ALPHA_I, f92, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_I, f94, f7
}
;;
{ .mmf
STFD [C4 ] = f61, SIZE
STFD [C12] = f62, SIZE
FMA f10 = ALPHA_R, f93, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_R, f95, f11
}
;;
{ .mmf
STFD [C4 ] = f63, 5 * SIZE
STFD [C12] = f47, 5 * SIZE
FMA f12 = ALPHA_I, f93, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_I, f95, f13
}
;;
{ .mmf
STFD [C4 ] = f64, SIZE
STFD [C12] = f65, SIZE
mov f64 = f0
}
{ .mmf
cmp.ne p6, p0 = 1, I
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C4 ] = f6, SIZE
STFD [C12] = f7, SIZE
mov f80 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f88 = f0
}
;;
{ .mmf
STFD [C4 ] = f10, SIZE
STFD [C12] = f11, SIZE
mov f65 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f73 = f0
}
;;
{ .mmf
STFD [C4 ] = f12, 5 * SIZE
STFD [C12] = f13, 5 * SIZE
mov f81 = f0
}
{ .mfb
adds I = -1, I
mov f89 = f0
(p6) br.cond.dptk .L052
}
;;
.align 32
.L060:
{ .mfi
nop __LINE__
mov f66 = f0
tbit.z p6, p7 = M, 2
}
{ .mfb
nop __LINE__
mov f74 = f0
(p6) br.cond.dptk .L070
}
;;
{ .mfb
LDFPD f48, f49 = [B]
mov f82 = f0
nop __LINE__
}
{ .mfi
adds BOFFSET = 2 * SIZE, B
mov f90 = f0
adds L = 1, K
}
;;
{ .mii
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mfi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
mov f67 = f0
adds L = -1, L
}
{ .mfi
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
mov f75 = f0
nop __LINE__
}
;;
{ .mfi
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
mov f83 = f0
mov ar.lc = L
}
{ .mfi
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
mov f91 = f0
cmp.eq p3, p0 = r0, r0
}
;;
.align 32
.L062:
{ .mfi
lfetch.nt1 [PREA], 8 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 8 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
(p5) adds C9 = 4 * SIZE, C1
}
{ .mfi
nop __LINE__
FMA f88 = f32, f51, f88 // A1 * B4
(p5) adds C10 = 4 * SIZE, C2
}
;;
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
(p5) adds C11 = 4 * SIZE, C3
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
(p5) adds C12 = 4 * SIZE, C4
}
;;
{ .mfb
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f89 = f33, f51, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f74 = f34, f49, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f82 = f34, f50, f82 // A3 * B3
nop __LINE__
}
{ .mfb
nop __LINE__
FMA f90 = f34, f51, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
FMA f67 = f35, f48, f67 // A4 * B1
}
{ .mfb
(p5) LDFD f6 = [C1 ], SIZE
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
FMA f83 = f35, f50, f83 // A4 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
FMA f91 = f35, f51, f91 // A4 * B4
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f88 = f40, f59, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f12 = [C1 ], SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C9], SIZE
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f14 = [C1 ], - 3 * SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f15 = [C9], - 3 * SIZE
(p3) FMA f89 = f41, f59, f89 // A2 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f16 = [C2 ], SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f17 = [C10], SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f18 = [C2 ], SIZE
(p3) FMA f82 = f42, f58, f82 // A3 * B3
nop __LINE__
}
{ .mfb
(p5) LDFD f19 = [C10], SIZE
(p3) FMA f90 = f42, f59, f90 // A3 * B4
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f20 = [C2 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f21 = [C10], SIZE
(p3) FMA f75 = f43, f57, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfi
(p5) LDFD f22 = [C2 ], -3 * SIZE
(p3) FMA f83 = f43, f58, f83 // A4 * B3
adds L = -1, L
}
{ .mfb
(p5) LDFD f23 = [C10], -3 * SIZE
(p3) FMA f91 = f43, f59, f91 // A4 * B4
br.cloop.sptk.few .L062
}
;;
{ .mmf
LDFD f24 = [C3 ], SIZE
LDFD f25 = [C11], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f26 = [C3 ], SIZE
LDFD f27 = [C11], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f28 = [C3 ], SIZE
LDFD f29 = [C11], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f30 = [C3 ], - 3 * SIZE
LDFD f31 = [C11], - 3 * SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f72, f16
}
{ .mmf
LDFD f32 = [C4 ], SIZE
LDFD f33 = [C12], SIZE
FMA f17 = ALPHA_R, f74, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f72, f18
}
{ .mmf
LDFD f34 = [C4 ], SIZE
LDFD f35 = [C12], SIZE
FMA f19 = ALPHA_I, f74, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f73, f20
}
{ .mmf
LDFD f36 = [C4 ], SIZE
LDFD f37 = [C12], SIZE
FMA f21 = ALPHA_R, f75, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f73, f22
}
{ .mmf
LDFD f38 = [C4 ], - 3 * SIZE
LDFD f39 = [C12], - 3 * SIZE
FMA f23 = ALPHA_I, f75, f23
}
;;
{ .mmf
STFD [C2 ] = f16, SIZE
STFD [C10] = f17, SIZE
FMA f24 = ALPHA_R, f80, f24
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f25 = ALPHA_R, f82, f25
}
;;
{ .mmf
STFD [C2 ] = f18, SIZE
STFD [C10] = f19, SIZE
FMA f26 = ALPHA_I, f80, f26
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f27 = ALPHA_I, f82, f27
}
;;
{ .mmf
STFD [C2 ] = f20, SIZE
STFD [C10] = f21, SIZE
FMA f28 = ALPHA_R, f81, f28
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f29 = ALPHA_R, f83, f29
}
;;
{ .mmf
STFD [C2 ] = f22, 5 * SIZE
STFD [C10] = f23, 5 * SIZE
FMA f30 = ALPHA_I, f81, f30
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f31 = ALPHA_I, f83, f31
}
;;
{ .mmf
STFD [C3 ] = f24, SIZE
STFD [C11] = f25, SIZE
FMA f32 = ALPHA_R, f88, f32
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f33 = ALPHA_R, f90, f33
}
;;
{ .mmf
STFD [C3 ] = f26, SIZE
STFD [C11] = f27, SIZE
FMA f34 = ALPHA_I, f88, f34
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f35 = ALPHA_I, f90, f35
}
;;
{ .mmf
STFD [C3 ] = f28, SIZE
STFD [C11] = f29, SIZE
FMA f36 = ALPHA_R, f89, f36
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f37 = ALPHA_R, f91, f37
}
;;
{ .mmf
STFD [C3 ] = f30, 5 * SIZE
STFD [C11] = f31, 5 * SIZE
FMA f38 = ALPHA_I, f89, f38
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f39 = ALPHA_I, f91, f39
}
;;
{ .mmf
STFD [C4 ] = f32, SIZE
STFD [C12] = f33, SIZE
mov f64 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C4 ] = f34, SIZE
STFD [C12] = f35, SIZE
mov f80 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f88 = f0
}
;;
{ .mmf
STFD [C4 ] = f36, SIZE
STFD [C12] = f37, SIZE
mov f81 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f65 = f0
}
;;
{ .mmf
STFD [C4 ] = f38, 5 * SIZE
STFD [C12] = f39, 5 * SIZE
mov f89 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f73 = f0
}
;;
.align 32
.L070:
{ .mib
nop __LINE__
tbit.z p6,p7 = M, 1
(p6) br.cond.dptk .L080
}
;;
{ .mmi
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
adds L = 1, K
}
;;
{ .mii
cmp.eq p3, p0 = r0, r0
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
(p7) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
adds L = -1, L
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
}
;;
{ .mmi
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
mov ar.lc = L
}
;;
.align 32
.L072:
{ .mfb
lfetch.nt1 [PREA], 4 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 8 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
cmp.ne p4, p5 = 0, L
}
{ .mfb
nop __LINE__
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
}
;;
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f81 = f33, f50, f81 // A2 * B3
}
{ .mmf
(p5) LDFD f6 = [C1 ], SIZE
(p5) LDFD f12 = [C2 ], SIZE
FMA f89 = f33, f51, f89 // A2 * B4
}
;;
{ .mfb
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mmf
(p5) LDFD f7 = [C1 ], SIZE
(p5) LDFD f13 = [C2 ], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
nop __LINE__
}
{ .mmf
(p5) LDFD f10 = [C1 ], SIZE
(p5) LDFD f14 = [C2 ], SIZE
(p3) FMA f88 = f40, f59, f88 // A1 * B4
}
;;
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C1 ], - 3 * SIZE
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfi
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
(p3) FMA f81 = f41, f58, f81 // A2 * B3
adds L = -1, L
}
{ .mfb
(p5) LDFD f15 = [C2 ], - 3 * SIZE
(p3) FMA f89 = f41, f59, f89 // A2 * B4
br.cloop.sptk.few .L072
}
;;
{ .mmf
LDFD f16 = [C3], SIZE
LDFD f20 = [C4], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f72, f12
}
;;
{ .mmf
LDFD f17 = [C3], SIZE
LDFD f21 = [C4], SIZE
FMA f7 = ALPHA_I, f64, f7
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_I, f72, f13
}
;;
{ .mmf
LDFD f18 = [C3], SIZE
LDFD f22 = [C4], SIZE
FMA f10 = ALPHA_R, f65, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_R, f73, f14
}
;;
{ .mmf
LDFD f19 = [C3], - 3 * SIZE
LDFD f23 = [C4], - 3 * SIZE
FMA f11 = ALPHA_I, f65, f11
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f73, f15
}
;;
{ .mmf
STFD [C1] = f6, SIZE
STFD [C2] = f12, SIZE
FMA f16 = ALPHA_R, f80, f16
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f20 = ALPHA_R, f88, f20
}
;;
{ .mmf
STFD [C1] = f7, SIZE
STFD [C2] = f13, SIZE
FMA f17 = ALPHA_I, f80, f17
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f21 = ALPHA_I, f88, f21
}
;;
{ .mmf
STFD [C1] = f10, SIZE
STFD [C2] = f14, SIZE
FMA f18 = ALPHA_R, f81, f18
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f22 = ALPHA_R, f89, f22
}
;;
{ .mmf
STFD [C1] = f11, SIZE
STFD [C2] = f15, SIZE
FMA f19 = ALPHA_I, f81, f19
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f23 = ALPHA_I, f89, f23
}
;;
{ .mmf
STFD [C3] = f16, SIZE
STFD [C4] = f20, SIZE
mov f64 = f0
}
;;
{ .mmf
STFD [C3] = f17, SIZE
STFD [C4] = f21, SIZE
mov f72 = f0
}
;;
{ .mmf
STFD [C3] = f18, SIZE
STFD [C4] = f22, SIZE
mov f80 = f0
}
;;
{ .mmf
STFD [C3] = f19, SIZE
STFD [C4] = f23, SIZE
mov f88 = f0
}
;;
.align 32
.L080:
{ .mib
nop __LINE__
tbit.z p6,p7 = M, 0
(p6) br.cond.dptk .L089
}
;;
{ .mmi
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
adds L = 1, K
}
;;
{ .mii
LDFD f32 = [AOFFSET], 1 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
nop __LINE__
nop __LINE__
adds L = -1, L
}
;;
{ .mmi
LDFPD f50, f51 = [BOFFSET], 2 * SIZE
cmp.eq p3, p0 = r0, r0
mov ar.lc = L
}
;;
.align 32
.L082:
{ .mfb
cmp.ne p4, p5 = 0, L
FMA f64 = f32, f48, f64 // A1 * B1
nop __LINE__
}
{ .mfi
(p12) cmp.ne p3, p0 = 0, L
FMA f72 = f32, f49, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f80 = f32, f50, f80 // A1 * B3
nop __LINE__
}
{ .mfb
(p3) LDFD f40 = [AOFFSET], 1 * SIZE
FMA f88 = f32, f51, f88 // A1 * B4
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f58, f59 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mmf
(p5) LDFD f6 = [C1], SIZE
(p5) LDFD f10 = [C2], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
}
;;
{ .mmf
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p4) LDFD f32 = [AOFFSET], 1 * SIZE
(p3) FMA f80 = f40, f58, f80 // A1 * B3
}
{ .mmf
(p5) LDFD f7 = [C1], -SIZE
(p5) LDFD f11 = [C2], -SIZE
(p3) FMA f88 = f40, f59, f88 // A1 * B4
}
;;
{ .mib
(p4) LDFPD f50, f51 = [BOFFSET], 2 * SIZE
adds L = -1, L
br.cloop.sptk.few .L082
}
;;
{ .mmf
LDFD f12 = [C3], SIZE
LDFD f14 = [C4], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f10 = ALPHA_R, f72, f10
}
;;
{ .mmf
LDFD f13 = [C3], -SIZE
LDFD f15 = [C4], -SIZE
FMA f7 = ALPHA_I, f64, f7
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f72, f11
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f80, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_R, f88, f14
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_I, f80, f13
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f88, f15
}
;;
{ .mmi
STFD [C1] = f6, SIZE
STFD [C2] = f10, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C1] = f7, SIZE
STFD [C2] = f11, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C3] = f12, SIZE
STFD [C4] = f14, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C3] = f13, SIZE
STFD [C4] = f15, SIZE
nop __LINE__
}
;;
.align 32
.L089:
{ .mmi
mov B = BOFFSET
mov AOFFSET = A
nop __LINE__
}
;;
.align 16
.L090:
{ .mfi
mov C1 = C
mov f64 = f0
tbit.z p6, p0 = N, 1
}
{ .mfi
add C2 = LDC, C
mov f72 = f0
shr I = M, 3
}
;;
{ .mfi
setf.d f66 = r0
mov f65 = f0
nop __LINE__
}
{ .mfb
mov AOFFSET = A
mov f73 = f0
(p6) br.cond.dpnt .L130
}
;;
{ .mfi
nop __LINE__
mov f67 = f0
shladd C = LDC, 1, C
}
{ .mfb
cmp.eq p6, p7 = 0, I
mov f74 = f0
(p6) br.cond.dpnt .L100
}
;;
.align 32
.L092:
{ .mfb
LDFPD f48, f49 = [B]
mov f68 = f0
nop __LINE__
}
{ .mfb
adds BOFFSET = 2 * SIZE, B
mov f79 = f0
nop __LINE__
}
;;
{ .mfi
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
mov f75 = f0
nop __LINE__
}
;;
{ .mfi
adds PREC = CPREFETCHSIZE * SIZE, C1
mov f76 = f0
adds L = 1, K
}
;;
{ .mfi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
mov f69 = f0
tbit.z p12, p0 = L, 0
}
{ .mfi
cmp.eq p3, p0 = r0, r0
mov f77 = f0
shr L = L, 1
}
;;
{ .mfi
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
adds L = -1, L
}
{ .mmf
LDFPD f36, f37 = [AOFFSET], 2 * SIZE
CPREFETCH [PREC], LDC
mov f70 = f0
}
;;
{ .mfi
LDFPD f38, f39 = [AOFFSET], 2 * SIZE
mov f78 = f0
mov ar.lc = L
}
{ .mfi
CPREFETCH [PREC]
mov f71 = f0
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
}
;;
.align 32
.L093:
/* 1 */
{ .mfi
lfetch.nt1 [PREA], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 4 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
adds C9 = 4 * SIZE, C1
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
adds C10 = 4 * SIZE, C2
}
;;
{ .mfi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
adds C11 = 4 * SIZE, C3
}
{ .mfi
nop __LINE__
FMA f74 = f34, f49, f74 // A3 * B2
adds C12 = 4 * SIZE, C4
}
;;
{ .mfb
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f67 = f35, f48, f67 // A4 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f6 = [C1 ], SIZE
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f68 = f36, f48, f68 // A5 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
FMA f76 = f36, f49, f76 // A5 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f44, f45 = [AOFFSET], 2 * SIZE
FMA f69 = f37, f48, f69 // A6 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
FMA f77 = f37, f49, f77 // A6 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f46, f47 = [AOFFSET], 2 * SIZE
FMA f70 = f38, f48, f70 // A7 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
FMA f78 = f38, f49, f78 // A7 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
FMA f71 = f39, f48, f71 // A8 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f12 = [C1 ], SIZE
FMA f79 = f39, f49, f79 // A8 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C9 ], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f14 = [C1 ], 5 * SIZE
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f36, f37 = [AOFFSET], 2 * SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f15 = [C9 ], 5 * SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f38, f39 = [AOFFSET], 2 * SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
nop __LINE__
}
{ .mfb
nop __LINE__
(p3) FMA f75 = f43, f57, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f16 = [C1 ], SIZE
(p3) FMA f68 = f44, f56, f68 // A5 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f17 = [C9 ], SIZE
(p3) FMA f76 = f44, f57, f76 // A5 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f18 = [C1 ], SIZE
(p3) FMA f69 = f45, f56, f69 // A6 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f19 = [C9 ], SIZE
(p3) FMA f77 = f45, f57, f77 // A6 * B2
nop __LINE__
}
;;
{ .mfb
(p5) LDFD f20 = [C1 ], SIZE
(p3) FMA f70 = f46, f56, f70 // A7 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f21 = [C9 ], SIZE
(p3) FMA f78 = f46, f57, f78 // A7 * B2
nop __LINE__
}
;;
{ .mfi
(p5) LDFD f22 = [C1 ], -11 * SIZE
(p3) FMA f71 = f47, f56, f71 // A8 * B1
adds L = -1, L
}
{ .mfb
(p5) LDFD f23 = [C9 ], -11 * SIZE
(p3) FMA f79 = f47, f57, f79 // A8 * B2
br.cloop.sptk.few .L093
}
;;
{ .mmf
LDFD f24 = [C2 ], SIZE
LDFD f25 = [C10], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f26 = [C2 ], SIZE
LDFD f27 = [C10], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f28 = [C2 ], SIZE
LDFD f29 = [C10], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f30 = [C2 ], 5 * SIZE
LDFD f31 = [C10], 5 * SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f68, f16
}
{ .mmf
LDFD f32 = [C2 ], SIZE
LDFD f33 = [C10], SIZE
FMA f17 = ALPHA_R, f70, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f68, f18
}
{ .mmf
LDFD f34 = [C2 ], SIZE
LDFD f35 = [C10], SIZE
FMA f19 = ALPHA_I, f70, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f69, f20
}
{ .mmf
LDFD f36 = [C2 ], SIZE
LDFD f37 = [C10], SIZE
FMA f21 = ALPHA_R, f71, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f69, f22
}
{ .mmf
LDFD f38 = [C2 ], - 11 * SIZE
LDFD f39 = [C10], - 11 * SIZE
FMA f23 = ALPHA_I, f71, f23
}
;;
{ .mmf
STFD [C1 ] = f16, SIZE
STFD [C9 ] = f17, SIZE
FMA f24 = ALPHA_R, f72, f24
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f25 = ALPHA_R, f74, f25
}
;;
{ .mmf
STFD [C1 ] = f18, SIZE
STFD [C9 ] = f19, SIZE
FMA f26 = ALPHA_I, f72, f26
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f27 = ALPHA_I, f74, f27
}
;;
{ .mmf
STFD [C1 ] = f20, SIZE
STFD [C9 ] = f21, SIZE
FMA f28 = ALPHA_R, f73, f28
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f29 = ALPHA_R, f75, f29
}
;;
{ .mmf
STFD [C1 ] = f22, 5 * SIZE
STFD [C9 ] = f23, 5 * SIZE
FMA f30 = ALPHA_I, f73, f30
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f31 = ALPHA_I, f75, f31
}
;;
{ .mmf
STFD [C2 ] = f24, SIZE
STFD [C10] = f25, SIZE
FMA f32 = ALPHA_R, f76, f32
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f33 = ALPHA_R, f78, f33
}
;;
{ .mmf
STFD [C2 ] = f26, SIZE
STFD [C10] = f27, SIZE
FMA f34 = ALPHA_I, f76, f34
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f35 = ALPHA_I, f78, f35
}
;;
{ .mmf
STFD [C2 ] = f28, SIZE
STFD [C10] = f29, SIZE
FMA f36 = ALPHA_R, f77, f36
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f37 = ALPHA_R, f79, f37
}
;;
{ .mmf
STFD [C2 ] = f30, 5 * SIZE
STFD [C10] = f31, 5 * SIZE
FMA f38 = ALPHA_I, f77, f38
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f39 = ALPHA_I, f79, f39
}
;;
{ .mmf
STFD [C2 ] = f32, SIZE
STFD [C10] = f33, SIZE
mov f64 = f0
}
{ .mmf
cmp.ne p6, p0 = 1, I
nop __LINE__
mov f72 = f0
}
;;
{ .mmf
STFD [C2 ] = f34, SIZE
STFD [C10] = f35, SIZE
mov f65 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f73 = f0
}
;;
{ .mmf
STFD [C2 ] = f36, SIZE
STFD [C10] = f37, SIZE
mov f66 = f0
}
{ .mmf
nop __LINE__
nop __LINE__
mov f74 = f0
}
;;
{ .mmf
STFD [C2 ] = f38, 5 * SIZE
STFD [C10] = f39, 5 * SIZE
mov f67 = f0
}
{ .mfb
adds I = -1, I
mov f75 = f0
(p6) br.cond.dptk .L092
}
;;
.align 32
.L100:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 2
(p6) br.cond.dptk .L110
}
;;
{ .mmf
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
mov f75 = f0
}
{ .mii
nop __LINE__
adds L = 1, K
}
;;
{ .mii
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
nop __LINE__
adds L = -1, L
}
;;
{ .mmi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
cmp.eq p3, p0 = r0, r0
mov ar.lc = L
}
;;
.align 32
.L102:
{ .mfi
lfetch.nt1 [PREA], 8 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
lfetch.nt1 [PREB], 4 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
adds C9 = 4 * SIZE, C1
}
{ .mfi
nop __LINE__
FMA f73 = f33, f49, f73 // A2 * B2
adds C10 = 4 * SIZE, C2
}
;;
{ .mfb
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f6 = [C1 ], SIZE
FMA f74 = f34, f49, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f67 = f35, f48, f67 // A4 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
FMA f75 = f35, f49, f75 // A4 * B2
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f73 = f41, f57, f73 // A2 * B2
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f12 = [C1], SIZE
(p3) FMA f74 = f42, f57, f74 // A3 * B2
nop __LINE__
}
;;
{ .mfi
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
adds L = -1, L
}
{ .mfb
(p5) LDFD f13 = [C9], SIZE
(p3) FMA f75 = f43, f57, f75 // A4 * B2
br.cloop.sptk.few .L102
}
;;
{ .mmf
LDFD f14 = [C1], - 3 * SIZE
LDFD f15 = [C9], - 3 * SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f16 = [C2 ], SIZE
LDFD f17 = [C10], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f18 = [C2 ], SIZE
LDFD f19 = [C10], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f20 = [C2 ], SIZE
LDFD f21 = [C10], SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f72, f16
}
{ .mmf
LDFD f22 = [C2 ], - 3 * SIZE
LDFD f23 = [C10], - 3 * SIZE
FMA f17 = ALPHA_R, f74, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f72, f18
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f19 = ALPHA_I, f74, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f73, f20
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f21 = ALPHA_R, f75, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f73, f22
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f23 = ALPHA_I, f75, f23
}
;;
{ .mmf
STFD [C2 ] = f16, SIZE
STFD [C10] = f17, SIZE
mov f64 = f0
}
;;
{ .mmf
STFD [C2 ] = f18, SIZE
STFD [C10] = f19, SIZE
mov f65 = f0
}
;;
{ .mmf
STFD [C2 ] = f20, SIZE
STFD [C10] = f21, SIZE
mov f72 = f0
}
;;
{ .mmf
STFD [C2 ] = f22, 5 * SIZE
STFD [C10] = f23, 5 * SIZE
mov f73 = f0
}
;;
.align 32
.L110:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 1
(p6) br.cond.dptk .L120
}
;;
{ .mmi
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
adds L = 1, K
}
;;
{ .mii
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
nop __LINE__
adds L = -1, L
}
;;
{ .mmi
cmp.eq p3, p0 = r0, r0
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
mov ar.lc = L
}
;;
.align 32
.L112:
{ .mfi
lfetch.nt1 [PREA], 4 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
lfetch.nt1 [PREB], 4 * SIZE
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mmf
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
}
{ .mmf
(p5) LDFD f6 = [C1 ], SIZE
(p5) LDFD f7 = [C2 ], SIZE
FMA f73 = f33, f49, f73 // A2 * B2
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
nop __LINE__
}
;;
{ .mfi
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
adds L = -1, L
}
{ .mfb
(p5) LDFD f11 = [C2 ], SIZE
(p3) FMA f73 = f41, f57, f73 // A2 * B2
br.cloop.sptk.few .L112
}
;;
{ .mmf
LDFD f12 = [C1], SIZE
LDFD f13 = [C2], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f72, f7
}
;;
{ .mmf
LDFD f14 = [C1], - 3 * SIZE
LDFD f15 = [C2], - 3 * SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f72, f11
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f73, f13
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f73, f15
}
;;
{ .mmf
STFD [C1] = f6, SIZE
STFD [C2] = f7, SIZE
mov f64 = f0
}
;;
{ .mmf
STFD [C1] = f10, SIZE
STFD [C2] = f11, SIZE
mov f72 = f0
}
;;
{ .mmf
STFD [C1] = f12, SIZE
STFD [C2] = f13, SIZE
mov f65 = f0
}
;;
{ .mmf
STFD [C1] = f14, SIZE
STFD [C2] = f15, SIZE
mov f73 = f0
}
;;
.align 32
.L120:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 0
(p6) br.cond.dptk .L129
}
;;
{ .mmi
LDFPD f48, f49 = [B]
adds BOFFSET = 2 * SIZE, B
adds L = 1, K
}
;;
{ .mii
nop __LINE__
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
LDFD f32 = [AOFFSET], 1 * SIZE
nop __LINE__
adds L = -1, L
}
;;
{ .mmi
cmp.eq p3, p0 = r0, r0
nop __LINE__
mov ar.lc = L
}
;;
.align 32
.L122:
{ .mfi
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
nop __LINE__
FMA f72 = f32, f49, f72 // A1 * B2
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mmi
(p3) LDFPD f56, f57 = [BOFFSET], 2 * SIZE
(p3) LDFD f40 = [AOFFSET], 1 * SIZE
nop __LINE__
}
{ .mmi
(p5) LDFD f6 = [C1], SIZE
(p5) LDFD f7 = [C2], SIZE
}
;;
{ .mfi
(p4) LDFPD f48, f49 = [BOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
adds L = -1, L
}
{ .mfb
(p4) LDFD f32 = [AOFFSET], 1 * SIZE
(p3) FMA f72 = f40, f57, f72 // A1 * B2
br.cloop.sptk.few .L122
}
;;
.L128:
{ .mmf
(p5) LDFD f10 = [C1], -SIZE
(p5) LDFD f11 = [C2], -SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f72, f7
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f72, f11
}
;;
{ .mmi
STFD [C1 ] = f6, SIZE
STFD [C2 ] = f7, SIZE
nop __LINE__
}
;;
{ .mmi
STFD [C1 ] = f10, SIZE
STFD [C2 ] = f11, SIZE
nop __LINE__
}
;;
.align 32
.L129:
{ .mmi
mov B = BOFFSET
mov AOFFSET = A
nop __LINE__
}
;;
.align 16
.L130:
{ .mfi
nop __LINE__
mov f64 = f0
tbit.z p6, p0 = N, 0
}
{ .mib
mov AOFFSET = A
shr I = M, 3
(p6) br.cond.dpnt .L999
}
;;
{ .mfi
mov C1 = C
mov f65 = f0
nop __LINE__
}
;;
{ .mfi
nop __LINE__
mov f66 = f0
nop __LINE__
}
{ .mfb
cmp.eq p7, p0 = 0, I
mov f67 = f0
(p7) br.cond.dpnt .L140
}
;;
.align 32
.L132:
{ .mfb
LDFD f48 = [B]
mov f68 = f0
nop __LINE__
}
{ .mfi
adds BOFFSET = 1 * SIZE, B
mov f69 = f0
nop __LINE__
}
;;
{ .mfi
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
mov f70 = f0
adds L = 1, K
}
;;
{ .mii
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mfi
LDFPD f36, f37 = [AOFFSET], 2 * SIZE
mov f71 = f0
adds L = -1, L
}
;;
{ .mmi
LDFPD f38, f39 = [AOFFSET], 2 * SIZE
adds PREC = CPREFETCHSIZE * SIZE, C1
cmp.eq p3, p0 = r0, r0
}
;;
{ .mmi
CPREFETCH [PREC]
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
mov ar.lc = L
}
;;
.align 32
.L133:
{ .mfi
lfetch.nt1 [PREA], 16 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
adds PREB = (PREFETCHSIZE + 0) * SIZE, BOFFSET
FMA f65 = f33, f48, f65 // A2 * B1
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
adds C9 = 4 * SIZE, C1
}
{ .mmf
(p3) LDFD f56 = [BOFFSET], 1 * SIZE
(p5) LDFD f6 = [C1 ], SIZE
FMA f67 = f35, f48, f67 // A4 * B1
}
;;
{ .mfb
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
FMA f68 = f36, f48, f68 // A5 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f7 = [C9 ], SIZE
FMA f69 = f37, f48, f69 // A6 * B1
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f44, f45 = [AOFFSET], 2 * SIZE
FMA f70 = f38, f48, f70 // A7 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f10 = [C1 ], SIZE
FMA f71 = f39, f48, f71 // A8 * B1
nop __LINE__
}
;;
{ .mfb
(p3) LDFPD f46, f47 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
nop __LINE__
}
{ .mmf
(p4) LDFD f48 = [BOFFSET], 1 * SIZE
(p5) LDFD f12 = [C1 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
}
;;
{ .mfb
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
(p3) FMA f68 = f44, f56, f68 // A5 * B1
nop __LINE__
}
{ .mfb
(p5) LDFD f13 = [C9 ], SIZE
(p3) FMA f69 = f45, f56, f69 // A6 * B1
nop __LINE__
}
;;
{ .mfi
(p4) LDFPD f36, f37 = [AOFFSET], 2 * SIZE
(p3) FMA f70 = f46, f56, f70 // A7 * B1
adds L = -1, L
}
{ .mfb
(p5) LDFD f14 = [C1 ], 5 * SIZE
(p3) FMA f71 = f47, f56, f71 // A8 * B1
nop __LINE__
}
;;
{ .mfb
(p4) LDFPD f38, f39 = [AOFFSET], 2 * SIZE
nop __LINE__
nop __LINE__
}
{ .mfb
(p5) LDFD f15 = [C9 ], 5 * SIZE
nop __LINE__
br.cloop.sptk.few .L133
}
;;
.L138:
{ .mmf
LDFD f16 = [C1 ], SIZE
LDFD f17 = [C9 ], SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
LDFD f18 = [C1 ], SIZE
LDFD f19 = [C9 ], SIZE
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
LDFD f20 = [C1 ], SIZE
LDFD f21 = [C9 ], SIZE
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
LDFD f22 = [C1 ], - 11 * SIZE
LDFD f23 = [C9 ], - 11 * SIZE
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
FMA f16 = ALPHA_R, f68, f16
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f17 = ALPHA_R, f70, f17
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
FMA f18 = ALPHA_I, f68, f18
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f19 = ALPHA_I, f70, f19
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
FMA f20 = ALPHA_R, f69, f20
}
{ .mmf
cmp.ne p6, p0 = 1, I
adds I = -1, I
FMA f21 = ALPHA_R, f71, f21
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
FMA f22 = ALPHA_I, f69, f22
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f23 = ALPHA_I, f71, f23
}
;;
{ .mmf
STFD [C1 ] = f16, SIZE
STFD [C9 ] = f17, SIZE
mov f64 = f0
}
;;
{ .mmf
STFD [C1 ] = f18, SIZE
STFD [C9 ] = f19, SIZE
mov f65 = f0
}
;;
{ .mmf
STFD [C1 ] = f20, SIZE
STFD [C9 ] = f21, SIZE
mov f66 = f0
}
;;
{ .mmf
STFD [C1 ] = f22, 5 * SIZE
STFD [C9 ] = f23, 5 * SIZE
mov f67 = f0
}
{ .mmb
nop __LINE__
nop __LINE__
(p6) br.cond.dptk .L132
}
;;
.align 32
.L140:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 2
(p6) br.cond.dptk .L150
}
;;
{ .mmi
LDFD f48 = [B]
adds BOFFSET = 1 * SIZE, B
adds L = 1, K
}
;;
{ .mii
(p7) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mmi
LDFPD f34, f35 = [AOFFSET], 2 * SIZE
adds L = -1, L
nop __LINE__
}
;;
{ .mmi
adds PREA = (PREFETCHSIZE + 0) * SIZE, AOFFSET
cmp.eq p3, p0 = r0, r0
mov ar.lc = L
}
;;
.align 32
.L142:
{ .mfi
lfetch.nt1 [PREA], 8 * SIZE
FMA f64 = f32, f48, f64 // A1 * B1
cmp.ne p4, p5 = 0, L
}
{ .mfi
nop __LINE__
FMA f65 = f33, f48, f65 // A2 * B1
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mfi
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f66 = f34, f48, f66 // A3 * B1
(p5) adds C9 = 4 * SIZE, C1
}
{ .mmf
(p3) LDFD f56 = [BOFFSET], 1 * SIZE
FMA f67 = f35, f48, f67 // A4 * B1
}
;;
{ .mfi
(p3) LDFPD f42, f43 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
(p5) adds C10 = 2 * SIZE, C2
}
{ .mmf
(p5) LDFD f6 = [C1 ], SIZE
(p5) LDFD f7 = [C9 ], SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
}
;;
{ .mmf
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p4) LDFD f48 = [BOFFSET], 1 * SIZE
(p3) FMA f66 = f42, f56, f66 // A3 * B1
}
{ .mmf
(p5) LDFD f10 = [C1 ], SIZE
(p5) LDFD f11 = [C9 ], SIZE
(p3) FMA f67 = f43, f56, f67 // A4 * B1
}
;;
{ .mfi
(p4) LDFPD f34, f35 = [AOFFSET], 2 * SIZE
nop __LINE__
adds L = -1, L
}
{ .mmb
(p5) LDFD f12 = [C1 ], SIZE
(p5) LDFD f13 = [C9 ], SIZE
br.cloop.sptk.few .L142
}
;;
.L148:
{ .mmf
LDFD f14 = [C1 ], - 3 * SIZE
LDFD f15 = [C9 ], - 3 * SIZE
FMA f6 = ALPHA_R, f64, f6
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f7 = ALPHA_R, f66, f7
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f10 = ALPHA_I, f64, f10
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f11 = ALPHA_I, f66, f11
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f12 = ALPHA_R, f65, f12
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f13 = ALPHA_R, f67, f13
}
;;
{ .mmf
nop __LINE__
nop __LINE__
FMA f14 = ALPHA_I, f65, f14
}
{ .mmf
nop __LINE__
nop __LINE__
FMA f15 = ALPHA_I, f67, f15
}
;;
{ .mmf
STFD [C1 ] = f6, SIZE
STFD [C9 ] = f7, SIZE
mov f64 = f0
}
;;
{ .mmf
STFD [C1 ] = f10, SIZE
STFD [C9 ] = f11, SIZE
mov f65 = f0
}
;;
{ .mmf
STFD [C1 ] = f12, SIZE
STFD [C9 ] = f13, SIZE
mov f66 = f0
}
;;
{ .mmf
STFD [C1 ] = f14, 5 * SIZE
STFD [C9 ] = f15, 5 * SIZE
mov f67 = f0
}
;;
.align 32
.L150:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 1
(p6) br.cond.dptk .L160
}
;;
{ .mmi
LDFD f48 = [B]
adds BOFFSET = 1 * SIZE, B
adds L = 1, K
}
;;
{ .mii
cmp.eq p3, p0 = r0, r0
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mii
LDFPD f32, f33 = [AOFFSET], 2 * SIZE
adds L = -1, L
;;
mov ar.lc = L
}
;;
.align 32
.L152:
{ .mfi
cmp.ne p4, p5 = 0, L
FMA f64 = f32, f48, f64 // A1 * B1
(p12) cmp.ne p3, p0 = 0, L
}
;;
{ .mmf
(p3) LDFD f56 = [BOFFSET], 1 * SIZE
(p3) LDFPD f40, f41 = [AOFFSET], 2 * SIZE
FMA f65 = f33, f48, f65 // A2 * B1
}
;;
{ .mfi
(p4) LDFPD f32, f33 = [AOFFSET], 2 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
adds L = -1, L
}
;;
{ .mfb
(p4) LDFD f48 = [BOFFSET], 1 * SIZE
(p3) FMA f65 = f41, f56, f65 // A2 * B1
br.cloop.sptk.few .L152
}
;;
.L158:
LDFD f68 = [C1 ], 1 * SIZE
;;
LDFD f69 = [C1 ], 1 * SIZE
;;
LDFD f70 = [C1 ], 1 * SIZE
;;
LDFD f71 = [C1 ], - 3 * SIZE
;;
FMA f68 = ALPHA_R, f64, f68
FMA f69 = ALPHA_I, f64, f69
FMA f70 = ALPHA_R, f65, f70
FMA f71 = ALPHA_I, f65, f71
;;
STFD [C1 ] = f68, SIZE
;;
STFD [C1 ] = f69, SIZE
;;
STFD [C1 ] = f70, SIZE
mov f64 = f0
;;
STFD [C1 ] = f71, SIZE
mov f65 = f0
;;
.align 32
.L160:
{ .mib
nop __LINE__
tbit.z p6, p7 = M, 0
(p6) br.cond.dptk .L169
}
;;
{ .mmi
LDFD f48 = [B]
adds BOFFSET = 1 * SIZE, B
adds L = 1, K
}
;;
{ .mii
LDFD f32 = [AOFFSET], 1 * SIZE
tbit.z p12, p0 = L, 0
shr L = L, 1
}
;;
{ .mii
adds L = -1, L
cmp.eq p3, p0 = r0, r0
;;
mov ar.lc = L
}
;;
.align 32
.L162:
{ .mmf
cmp.ne p4, p5 = 0, L
(p12) cmp.ne p3, p0 = 0, L
FMA f64 = f32, f48, f64 // A1 * B1
}
;;
{ .mmi
(p3) LDFD f56 = [BOFFSET], 1 * SIZE
(p3) LDFD f40 = [AOFFSET], 1 * SIZE
nop __LINE__
}
;;
{ .mmi
(p4) LDFD f32 = [AOFFSET], 1 * SIZE
(p5) LDFD f68 = [C1], 1 * SIZE
adds L = -1, L
}
;;
{ .mmf
(p4) LDFD f48 = [BOFFSET], 1 * SIZE
(p5) LDFD f69 = [C1], - 1 * SIZE
(p3) FMA f64 = f40, f56, f64 // A1 * B1
}
{ .mib
nop __LINE__
nop __LINE__
br.cloop.sptk.few .L162
}
;;
FMA f68 = ALPHA_R, f64, f68
FMA f69 = ALPHA_I, f64, f69
;;
STFD [C1 ] = f68, SIZE
;;
STFD [C1 ] = f69, SIZE
;;
.align 32
.L169:
{ .mmi
mov B = BOFFSET
mov AOFFSET = A
nop __LINE__
}
;;
.align 16
.L999:
mov r8 = r0
adds r9 = 1 * 16, SP
;;
ldf.fill f16 = [SP], 32
ldf.fill f17 = [r9], 32
;;
ldf.fill f18 = [SP], 32
ldf.fill f19 = [r9], 32
;;
ldf.fill f20 = [SP], 32
ldf.fill f21 = [r9], 32
;;
ldf.fill f22 = [SP], 32
ldf.fill f23 = [r9], 32
mov ar.lc = ARLC
;;
ldf.fill f24 = [SP], 32
ldf.fill f25 = [r9], 32
mov pr = PR, -1
;;
ldf.fill f26 = [SP], 32
ldf.fill f27 = [r9], 32
mov ar.pfs = ARPFS
;;
ldf.fill f28 = [SP], 32
ldf.fill f29 = [r9], 32
;;
ldf.fill f30 = [SP], 32
ldf.fill f31 = [r9]
br.ret.sptk.many b0
EPILOGUE