Blob Blame Raw
/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin.           */
/* All rights reserved.                                              */
/*                                                                   */
/* Redistribution and use in source and binary forms, with or        */
/* without modification, are permitted provided that the following   */
/* conditions are met:                                               */
/*                                                                   */
/*   1. Redistributions of source code must retain the above         */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer.                                                  */
/*                                                                   */
/*   2. Redistributions in binary form must reproduce the above      */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer in the documentation and/or other materials       */
/*      provided with the distribution.                              */
/*                                                                   */
/*    THIS  SOFTWARE IS PROVIDED  BY THE  UNIVERSITY OF  TEXAS AT    */
/*    AUSTIN  ``AS IS''  AND ANY  EXPRESS OR  IMPLIED WARRANTIES,    */
/*    INCLUDING, BUT  NOT LIMITED  TO, THE IMPLIED  WARRANTIES OF    */
/*    MERCHANTABILITY  AND FITNESS FOR  A PARTICULAR  PURPOSE ARE    */
/*    DISCLAIMED.  IN  NO EVENT SHALL THE UNIVERSITY  OF TEXAS AT    */
/*    AUSTIN OR CONTRIBUTORS BE  LIABLE FOR ANY DIRECT, INDIRECT,    */
/*    INCIDENTAL,  SPECIAL, EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES    */
/*    (INCLUDING, BUT  NOT LIMITED TO,  PROCUREMENT OF SUBSTITUTE    */
/*    GOODS  OR  SERVICES; LOSS  OF  USE,  DATA,  OR PROFITS;  OR    */
/*    BUSINESS INTERRUPTION) HOWEVER CAUSED  AND ON ANY THEORY OF    */
/*    LIABILITY, WHETHER  IN CONTRACT, STRICT  LIABILITY, OR TORT    */
/*    (INCLUDING NEGLIGENCE OR OTHERWISE)  ARISING IN ANY WAY OUT    */
/*    OF  THE  USE OF  THIS  SOFTWARE,  EVEN  IF ADVISED  OF  THE    */
/*    POSSIBILITY OF SUCH DAMAGE.                                    */
/*                                                                   */
/* The views and conclusions contained in the software and           */
/* documentation are those of the authors and should not be          */
/* interpreted as representing official policies, either expressed   */
/* or implied, of The University of Texas at Austin.                 */
/*********************************************************************/

#define ASSEMBLER
#include "common.h"

#define PREFETCHSIZE   64
#define WPREFETCHSIZE  32

#ifndef XDOUBLE
#define LD	LDF8
#define ST	STF8_NTA
#else
#define LD	LDFD
#define ST	STFD_NTA
#endif
	
#define J	r15
#define PREB	r17
#define PREA	r18

#define A1	r19
#define A2	r20
#define A3	r21
#define A4	r22
#define A5	r23
#define A6	r24
#define A7	r25
#define A8	r26
#define B1	r27
#define B2	r28

#define COUNT	r9
#define I	r10
#define II	r11

#define ARLC	r29
#define PR	r30

#define M	r32
#define N	r33
#define A	r34
#define LDA	r35
#define B	r36

	PROLOGUE
	.prologue
	PROFCODE

	.body	
	{ .mii
	shladd	LDA = LDA, BASE_SHIFT, r0
	mov	PR = pr
	shr	J = N, 3
	}
	;;
	{ .mib
	cmp.eq	p8,  p0 = 0, J
	mov	ARLC  = ar.lc
	(p8)  br.cond.dpnt .L20
	}
	;;
	.align 32

.L11:
	{ .mmi
	mov	A1 = A
	add	A2 = A, LDA
	mov	pr.rot = 0
	}
	{ .mmi
	shladd	A3 = LDA, 1, A
	shladd	A5 = LDA, 2, A
	adds	I = 1, M
	}
	;;
	{ .mmi
	shladd	A4 = LDA, 1, A2
	shladd	A6 = LDA, 2, A2
	mov	ar.ec  = 6
	}
	{ .mmi
	cmp.eq	p16, p0 = r0, r0
	shladd	A7 = LDA, 2, A3
	shr	I = I, 1
	}
	;;
	{ .mmi
	adds	B1 = 8 * SIZE, B
	shladd	A8 = LDA, 2, A4
	shladd	A = LDA, 3, A
	}
	{ .mmi
	adds	I = -1, I
	mov	COUNT = 0
	adds	J = -1, J
	}
	;;
	{ .mmi
	adds	PREA =  PREFETCHSIZE * SIZE, A
	adds	PREB = WPREFETCHSIZE * SIZE, B
	mov	ar.lc = I
	}
	{ .mmi
	mov	I  = M
	mov	II = M
	cmp.ne p14, p0 = r0, r0
	}
	;;
	.align 32

.L12:
	{ .mmi
	(p21) ST	[B ] = f37,  1 * SIZE
	(p14) ST	[B1] = f49,  1 * SIZE
	(p16) cmp.ne.unc p13, p0 = 1, I
	}
	{ .mmi
	lfetch.nt1	[PREA], LDA
	lfetch.excl.nt1 [PREB]
	adds	PREB = 16 * SIZE, PREB
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f43,  1 * SIZE
	(p14) ST	[B1] = f55,  1 * SIZE
	cmp.eq	p9, p0 = 8, COUNT
	}
	{ .mmi
	(p16) LD	f32  = [A1], SIZE
	(p16) LD	f38  = [A2], SIZE
	(p16) adds	I  = -2, I
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f61,  1 * SIZE
	(p14) ST	[B1] = f73,  1 * SIZE
	(p9) mov COUNT = 0
	}
	{ .mmi
	(p13) LD	f44  = [A1], SIZE
	(p13) LD	f50  = [A2], SIZE
	(p21) adds	II = -2, II
	}
	;;
	{ .mmb
	(p21) ST	[B ] = f67,  1 * SIZE
	(p14) ST	[B1] = f79,  1 * SIZE
	nop   __LINE__
	}
	{ .mmb
	(p16) LD	f56  = [A3], SIZE
	(p16) LD	f62  = [A4], SIZE
	nop   __LINE__
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f85,  1 * SIZE
	(p14) ST	[B1] = f97,  1 * SIZE
	(p9) adds PREA =  (PREFETCHSIZE - 2)* SIZE, A1
	}
	{ .mmb
	(p13) LD	f68  = [A3], SIZE
	(p13) LD	f74  = [A4], SIZE
	nop   __LINE__
	}
	;;
	{ .mmb
	(p21) ST	[B ] = f91,  1 * SIZE
	(p14) ST	[B1] = f103, 1 * SIZE
	nop   __LINE__
	}
	{ .mmb
	(p16) LD	f80  = [A5], SIZE
	(p16) LD	f86  = [A6], SIZE
	nop   __LINE__
	}
	;;
	{ .mmb
	(p21) ST	[B ] = f109, 1 * SIZE
	(p14) ST	[B1] = f121, 1 * SIZE
	nop   __LINE__
	}
	{ .mmb
	(p13) LD	f92  = [A5], SIZE
	(p13) LD	f98  = [A6], SIZE
	nop   __LINE__
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f115, 1 * SIZE
	(p14) ST	[B1] = f127, 9 * SIZE
	(p16) adds	COUNT = 1, COUNT
	}
	{ .mmb
	(p16) LD	f104 = [A7], SIZE
	(p16) LD	f110 = [A8], SIZE
	nop   __LINE__
	}
	;;
	{ .mmi
	(p13) LD	f116 = [A7], SIZE
	(p13) LD	f122 = [A8], SIZE
	(p14) adds	B = 8 * SIZE, B
	}
	{ .mmb
	(p20) cmp.ne.unc p14, p0 = 1, II
	nop   __LINE__
	br.ctop.sptk.few .L12
	}
	;;
	{ .mmb
	cmp.ne	p6, p0 = 0, J
	nop   __LINE__
	(p6) br.cond.dptk .L11
	}
	;;
	.align 32

.L20:
	{ .mmi
	adds	I = 1, M
	mov	A1 = A
	mov	pr.rot = 0
	}
	{ .mmi
	add	A2 = A, LDA
	shladd	A3 = LDA, 1, A
	tbit.z	p6, p0 = N, 2
	}
	;;
	{ .mmi
	shladd	A4 = LDA, 1, A2
	adds	B1 = 4 * SIZE, B
	mov	ar.ec  = 6
	}
	{ .mib
	cmp.eq	p16, p0 = r0, r0
	shr	I = I, 1
	(p6)  br.cond.dpnt .L30
	}
	;;
	{ .mmi
	shladd	A = LDA, 2, A
	nop	__LINE__
	nop	__LINE__
	}
	{ .mmi
	adds	I = -1, I
	mov	COUNT = 0
	adds	J = -1, J
	}
	;;
	{ .mmi
	adds	PREA =  PREFETCHSIZE * SIZE, A
	adds	PREB = WPREFETCHSIZE * SIZE, B
	mov	ar.lc = I
	}
	{ .mmi
	mov	I  = M
	mov	II = M
	cmp.ne p14, p0 = r0, r0
	}
	;;
	.align 32

.L22:
	{ .mmi
	(p21) ST	[B ] = f37,  1 * SIZE
	(p14) ST	[B1] = f49,  1 * SIZE
	(p16) cmp.ne.unc p13, p0 = 1, I
	}
	{ .mmi
	lfetch.nt1	[PREA], LDA
	lfetch.excl.nt1 [PREB], 8 * SIZE
	cmp.eq	p9, p0 = 4, COUNT
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f43,  1 * SIZE
	(p14) ST	[B1] = f55,  1 * SIZE
	(p16) adds	I  = -2, I
	}
	{ .mmi
	(p16) LD	f32  = [A1], SIZE
	(p16) LD	f38  = [A2], SIZE
	(p21) adds	II = -2, II
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f61,  1 * SIZE
	(p14) ST	[B1] = f73,  1 * SIZE
	(p9) mov COUNT = 0
	}
	{ .mmi
	(p13) LD	f44  = [A1], SIZE
	(p13) LD	f50  = [A2], SIZE
	nop   __LINE__
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f67,  1 * SIZE
	(p14) ST	[B1] = f79,  5 * SIZE
	(p9) adds PREA =  PREFETCHSIZE * SIZE, A1
	}
	{ .mmb
	(p16) LD	f56  = [A3], SIZE
	(p16) LD	f62  = [A4], SIZE
	nop   __LINE__
	}
	;;
	{ .mmi
	(p13) LD	f68  = [A3], SIZE
	(p13) LD	f74  = [A4], SIZE
	(p16) adds	COUNT = 1, COUNT
	}
	{ .mmb
	(p14) adds	B = 4 * SIZE, B
	(p20) cmp.ne.unc p14, p0 = 1, II
	br.ctop.sptk.few .L22
	}
	;;
	.align 32

.L30:
	{ .mmi
	adds	I = 1, M
	mov	A1 = A
	mov	pr.rot = 0
	}
	{ .mmi
	add	A2 = A, LDA
	adds	B1 = 2 * SIZE, B
	tbit.z	p6, p0 = N, 1
	}
	;;
	{ .mmi
	nop	__LINE__
	nop	__LINE__
	mov	ar.ec  = 6
	}
	{ .mib
	cmp.eq	p16, p0 = r0, r0
	shr	I = I, 1
	(p6)  br.cond.dpnt .L40
	}
	;;
	{ .mmi
	adds	I = -1, I
	;;
	shladd	A = LDA, 1, A
	mov	ar.lc = I
	}
	{ .mmi
	mov	I  = M
	mov	II = M
	cmp.ne p14, p0 = r0, r0
	}
	;;
	.align 32

.L32:
	{ .mmi
	(p21) ST	[B ] = f37,  1 * SIZE
	(p14) ST	[B1] = f49,  1 * SIZE
	(p16) cmp.ne.unc p13, p0 = 1, I
	}
	{ .mmi
	nop	__LINE__
	nop	__LINE__
	(p21) adds	II = -2, II
	}
	;;
	{ .mmi
	(p21) ST	[B ] = f43,  1 * SIZE
	(p14) ST	[B1] = f55,  3 * SIZE
	nop	__LINE__
	}
	{ .mmi
	(p16) LD	f32  = [A1], SIZE
	(p16) LD	f38  = [A2], SIZE
	nop	__LINE__
	}
	;;
	{ .mmi
	(p13) LD	f44  = [A1], SIZE
	(p13) LD	f50  = [A2], SIZE
	(p16) adds	I  = -2, I
	}
	{ .mmb
	(p14) adds	B = 2 * SIZE, B
	(p20) cmp.ne.unc p14, p0 = 1, II
	br.ctop.sptk.few .L32
	}
	;;
	.align 32

.L40:
	{ .mmi
	adds	I = 1, M
	mov	A1 = A
	mov	pr.rot = 0
	}
	{ .mmi
	tbit.z	p6, p0 = N, 0
	}
	;;
	{ .mmi
	nop	__LINE__
	nop	__LINE__
	mov	ar.ec  = 6
	}
	{ .mib
	cmp.eq	p16, p0 = r0, r0
	shr	I = I, 1
	(p6)  br.cond.dpnt .L999
	}
	;;
	{ .mmi
	adds	I = -1, I
	;;
	mov	ar.lc = I
	}
	{ .mmi
	mov	I  = M
	mov	II = M
	cmp.ne p14, p0 = r0, r0
	}
	;;
	.align 32

.L42:
	{ .mmi
	(p21) ST	[B ] = f37,  1 * SIZE
	(p16) cmp.ne.unc p13, p0 = 1, I
	(p21) adds	II = -2, II
	}
	;;
	{ .mmi
	(p14) ST	[B ] = f49,  1 * SIZE
	(p16) LD	f32  = [A1], SIZE
	(p16) adds	I  = -2, I
	}
	;;
	{ .mmb
	(p13) LD	f44  = [A1], SIZE
	(p20) cmp.ne.unc p14, p0 = 1, II
	br.ctop.sptk.few .L42
	}
	;;
	.align 32

.L999:
	mov pr    = PR, -1
	mov	 ar.lc = ARLC
	br.ret.sptk.many b0
	EPILOGUE