Blob Blame Raw
/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin.           */
/* All rights reserved.                                              */
/*                                                                   */
/* Redistribution and use in source and binary forms, with or        */
/* without modification, are permitted provided that the following   */
/* conditions are met:                                               */
/*                                                                   */
/*   1. Redistributions of source code must retain the above         */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer.                                                  */
/*                                                                   */
/*   2. Redistributions in binary form must reproduce the above      */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer in the documentation and/or other materials       */
/*      provided with the distribution.                              */
/*                                                                   */
/*    THIS  SOFTWARE IS PROVIDED  BY THE  UNIVERSITY OF  TEXAS AT    */
/*    AUSTIN  ``AS IS''  AND ANY  EXPRESS OR  IMPLIED WARRANTIES,    */
/*    INCLUDING, BUT  NOT LIMITED  TO, THE IMPLIED  WARRANTIES OF    */
/*    MERCHANTABILITY  AND FITNESS FOR  A PARTICULAR  PURPOSE ARE    */
/*    DISCLAIMED.  IN  NO EVENT SHALL THE UNIVERSITY  OF TEXAS AT    */
/*    AUSTIN OR CONTRIBUTORS BE  LIABLE FOR ANY DIRECT, INDIRECT,    */
/*    INCIDENTAL,  SPECIAL, EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES    */
/*    (INCLUDING, BUT  NOT LIMITED TO,  PROCUREMENT OF SUBSTITUTE    */
/*    GOODS  OR  SERVICES; LOSS  OF  USE,  DATA,  OR PROFITS;  OR    */
/*    BUSINESS INTERRUPTION) HOWEVER CAUSED  AND ON ANY THEORY OF    */
/*    LIABILITY, WHETHER  IN CONTRACT, STRICT  LIABILITY, OR TORT    */
/*    (INCLUDING NEGLIGENCE OR OTHERWISE)  ARISING IN ANY WAY OUT    */
/*    OF  THE  USE OF  THIS  SOFTWARE,  EVEN  IF ADVISED  OF  THE    */
/*    POSSIBILITY OF SUCH DAMAGE.                                    */
/*                                                                   */
/* The views and conclusions contained in the software and           */
/* documentation are those of the authors and should not be          */
/* interpreted as representing official policies, either expressed   */
/* or implied, of The University of Texas at Austin.                 */
/*********************************************************************/

#define ASSEMBLER
#include "common.h"

#define N	%i0
#define X	%i1
#define INCX	%i2
#define Y	%i3
#define INCY	%i4
#define I	%i5

#define XX	%l0
#define YY	%l1

#ifdef DOUBLE
#define a1	%f4
#define a2	%f6
#define a3	%f8
#define a4	%f10
#define a5	%f12
#define a6	%f14
#define a7	%f16
#define a8	%f18
#define b1	%f20
#define b2	%f22
#define b3	%f24
#define b4	%f26
#define b5	%f28
#define b6	%f30
#define b7	%f32
#define b8	%f34

#define c1	%f36
#define c2	%f38
#define c3	%f40
#define c4	%f42
#define c5	%f44
#define c6	%f46
#define c7	%f48
#define c8	%f50

#define t1	%f52
#define t2	%f54
#define t3	%f56
#define t4	%f58
#else
#define a1	%f2
#define a2	%f3
#define a3	%f4
#define a4	%f5
#define a5	%f6
#define a6	%f7
#define a7	%f8
#define a8	%f9
#define b1	%f10
#define b2	%f11
#define b3	%f12
#define b4	%f13
#define b5	%f14
#define b6	%f15
#define b7	%f16
#define b8	%f17

#define c1	%f18
#define c2	%f19
#define c3	%f20
#define c4	%f21
#define c5	%f22
#define c6	%f23
#define c7	%f24
#define c8	%f25

#define t1	%f26
#define t2	%f27
#define t3	%f28
#define t4	%f29
#endif

#ifdef DOUBLE
#define C	%f0
#define S	%f2
#else
#define C	%f0
#define S	%f1
#endif

	PROLOGUE
	SAVESP

#ifndef __64BIT__

#ifdef DOUBLE
	st	%i5, [%sp + STACK_START + 24]

	LDF	[%sp + STACK_START + 24], C
	LDF	[%sp + STACK_START + 32], S
#else
	st	%i5, [%sp + STACK_START + 24]

	LDF	[%sp + STACK_START + 24], C
	LDF	[%sp + STACK_START + 28], S
#endif
#else
#ifdef DOUBLE
	FMOV	%f10, C
	FMOV	%f12, S
#else
	FMOV	%f11, C
	FMOV	%f13, S
#endif
#endif

	cmp	N, 0
	ble	.LL19
	nop
	
	sll	INCX, BASE_SHIFT, INCX
	sll	INCY, BASE_SHIFT, INCY

	cmp	INCX, SIZE
	bne	.LL50
	nop

	cmp	INCY, SIZE
	bne	.LL50
	nop

	sra	N, 3, I
	cmp	I, 0
	ble,pn	%icc, .LL15
	nop


	LDF	[X +  0 * SIZE], a1
	LDF	[Y +  0 * SIZE], b1
	LDF	[X +  1 * SIZE], a2
	LDF	[Y +  1 * SIZE], b2
	LDF	[X +  2 * SIZE], a3
	LDF	[Y +  2 * SIZE], b3
	LDF	[X +  3 * SIZE], a4
	LDF	[Y +  3 * SIZE], b4

	LDF	[X +  4 * SIZE], a5
	LDF	[Y +  4 * SIZE], b5
	LDF	[X +  5 * SIZE], a6
	LDF	[Y +  5 * SIZE], b6
	LDF	[X +  6 * SIZE], a7
	LDF	[Y +  6 * SIZE], b7
	LDF	[X +  7 * SIZE], a8
	LDF	[Y +  7 * SIZE], b8

	FMUL	C, a1, c1
	FMUL	S, b1, c2
	FMUL	C, b1, c3
	LDF	[Y +  8 * SIZE], b1
	FMUL	S, a1, c4
	LDF	[X +  8 * SIZE], a1

	FMUL	C, a2, c5
	FMUL	S, b2, c6
	FADD	c1, c2, t1

	FMUL	C, b2, c7
	LDF	[Y +  9 * SIZE], b2
	FMUL	S, a2, c8
	LDF	[X +  9 * SIZE], a2
	FSUB	c3, c4, t2

	addcc	I, -1, I
	ble,pt	%icc, .LL12
	nop

#define PREFETCHSIZE 64

.LL11:
	FMUL	C, a3, c1
	nop
	prefetch [Y  + PREFETCHSIZE * SIZE], 1
	nop

	FMUL	S, b3, c2
	STF	t1, [X +  0 * SIZE]
	FADD	c5, c6, t3
	nop

	FMUL	C, b3, c3
	LDF	[Y + 10 * SIZE], b3
	nop
	nop

	FMUL	S, a3, c4
	STF	t2, [Y +  0 * SIZE]
	FSUB	c7, c8, t4
	nop

	FMUL	C, a4, c5
	LDF	[X + 10 * SIZE], a3
	nop
	nop

	FMUL	S, b4, c6
	STF	t3, [X +  1 * SIZE]
	FADD	c1, c2, t1
	nop

	FMUL	C, b4, c7
	LDF	[Y + 11 * SIZE], b4
	nop
	nop

	FMUL	S, a4, c8
	STF	t4, [Y +  1 * SIZE]
	FSUB	c3, c4, t2
	nop

	FMUL	C, a5, c1
	LDF	[X + 11 * SIZE], a4
	nop
	nop

	FMUL	S, b5, c2
	STF	t1, [X +  2 * SIZE]
	FADD	c5, c6, t3
	nop

	FMUL	C, b5, c3
	LDF	[Y + 12 * SIZE], b5
	nop
	nop

	FMUL	S, a5, c4
	STF	t2, [Y +  2 * SIZE]
	FSUB	c7, c8, t4
	nop

	FMUL	C, a6, c5
	LDF	[X + 12 * SIZE], a5
	nop
	nop

	FMUL	S, b6, c6
	STF	t3, [X +  3 * SIZE]
	FADD	c1, c2, t1
	nop

	FMUL	C, b6, c7
	LDF	[Y + 13 * SIZE], b6
	nop
	nop

	FMUL	S, a6, c8
	STF	t4, [Y +  3 * SIZE]
	FSUB	c3, c4, t2
	nop

	FMUL	C, a7, c1
	LDF	[X + 13 * SIZE], a6
	nop
	nop

	FMUL	S, b7, c2
	STF	t1, [X +  4 * SIZE]
	FADD	c5, c6, t3
	nop

	FMUL	C, b7, c3
	LDF	[Y + 14 * SIZE], b7
	nop
	nop

	FMUL	S, a7, c4
	STF	t2, [Y +  4 * SIZE]
	FSUB	c7, c8, t4
	nop

	FMUL	C, a8, c5
	LDF	[X + 14 * SIZE], a7
	nop
	nop

	FMUL	S, b8, c6
	STF	t3, [X +  5 * SIZE]
	FADD	c1, c2, t1
	nop

	FMUL	C, b8, c7
	LDF	[Y + 15 * SIZE], b8
	nop
	nop

	FMUL	S, a8, c8
	STF	t4, [Y +  5 * SIZE]
	FSUB	c3, c4, t2
	nop

	FMUL	C, a1, c1
	LDF	[X + 15 * SIZE], a8
	addcc	I, -1, I
	nop

	FMUL	S, b1, c2
	STF	t1, [X +  6 * SIZE]
	FADD	c5, c6, t3
	nop

	FMUL	C, b1, c3
	LDF	[Y + 16 * SIZE], b1
	nop
	nop

	FMUL	S, a1, c4
	STF	t2, [Y +  6 * SIZE]
	FSUB	c7, c8, t4
	nop

	FMUL	C, a2, c5
	LDF	[X + 16 * SIZE], a1
	add	Y, 8 * SIZE, Y
	nop

	FMUL	S, b2, c6
	STF	t3, [X +  7 * SIZE]
	FADD	c1, c2, t1
	nop

	FMUL	C, b2, c7
	LDF	[Y +  9 * SIZE], b2
	add	X, 8 * SIZE, X
	nop

	FMUL	S, a2, c8
	STF	t4, [Y -  1 * SIZE]
	FSUB	c3, c4, t2
	nop

	bg,pt	%icc, .LL11
	LDF	[X +  9 * SIZE], a2


.LL12:
	FMUL	C, a3, c1
	FMUL	S, b3, c2
	STF	t1, [X +  0 * SIZE]
	FADD	c5, c6, t3

	FMUL	C, b3, c3
	FMUL	S, a3, c4
	STF	t2, [Y +  0 * SIZE]
	FSUB	c7, c8, t4


	FMUL	C, a4, c5
	FMUL	S, b4, c6
	STF	t3, [X +  1 * SIZE]
	FADD	c1, c2, t1

	FMUL	C, b4, c7
	FMUL	S, a4, c8
	STF	t4, [Y +  1 * SIZE]
	FSUB	c3, c4, t2


	FMUL	C, a5, c1
	FMUL	S, b5, c2
	STF	t1, [X +  2 * SIZE]
	FADD	c5, c6, t3

	FMUL	C, b5, c3
	FMUL	S, a5, c4
	STF	t2, [Y +  2 * SIZE]
	FSUB	c7, c8, t4

	FMUL	C, a6, c5
	FMUL	S, b6, c6
	STF	t3, [X +  3 * SIZE]
	FADD	c1, c2, t1

	FMUL	C, b6, c7
	FMUL	S, a6, c8
	STF	t4, [Y +  3 * SIZE]
	FSUB	c3, c4, t2

	FMUL	C, a7, c1
	FMUL	S, b7, c2
	STF	t1, [X +  4 * SIZE]
	FADD	c5, c6, t3

	FMUL	C, b7, c3
	FMUL	S, a7, c4
	STF	t2, [Y +  4 * SIZE]
	FSUB	c7, c8, t4

	FMUL	C, a8, c5
	FMUL	S, b8, c6
	STF	t3, [X +  5 * SIZE]
	FADD	c1, c2, t1

	FMUL	C, b8, c7
	FMUL	S, a8, c8
	STF	t4, [Y +  5 * SIZE]
	FSUB	c3, c4, t2

	FADD	c5, c6, t3
	STF	t1, [X +  6 * SIZE]

	FSUB	c7, c8, t4
	STF	t2, [Y +  6 * SIZE]

	STF	t3, [X +  7 * SIZE]
	STF	t4, [Y +  7 * SIZE]

	add	X, 8 * SIZE, X
	add	Y, 8 * SIZE, Y


.LL15:
	andcc	N, 7, I
	nop
	ble,a,pn %icc, .LL19
	nop

.LL16:
	LDF	[X + 0 * SIZE], a1
	add	X, 1 * SIZE, X
	LDF	[Y + 0 * SIZE], b1
	add	Y, 1 * SIZE, Y

	FMUL	C, a1, c1
	FMUL	S, b1, c2
	FMUL	C, b1, c3
	FMUL	S, a1, c4

	FADD	c1, c2, c2
	addcc	I, -1, I
	FSUB	c3, c4, c4
	nop

	STF	c2, [X - 1 * SIZE]
	STF	c4, [Y - 1 * SIZE]
	bg,pt	%icc, .LL16
	nop

.LL19:
	return	%i7 + 8
	nop
	
.LL50:
	mov	X, XX
	mov	Y, YY

	sra	N, 3, I
	cmp	I, 0
	ble,pn	%icc, .LL55
	nop

.LL51:
	LDF	[X +  0 * SIZE], a1
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b1
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a2
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b2
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a3
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b3
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a4
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b4
	add	Y, INCY, Y

	LDF	[X +  0 * SIZE], a5
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b5
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a6
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b6
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a7
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b7
	add	Y, INCY, Y
	LDF	[X +  0 * SIZE], a8
	add	X, INCX, X
	LDF	[Y +  0 * SIZE], b8
	add	Y, INCY, Y

	FMUL	C, a1, c1
	FMUL	S, b1, c2
	FMUL	C, b1, c3
	FMUL	S, a1, c4

	FADD	c1, c2, t1
	FSUB	c3, c4, t2

	STF	t1, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t2, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a2, c5
	FMUL	S, b2, c6
	FMUL	C, b2, c7
	FMUL	S, a2, c8

	FADD	c5, c6, t3
	FSUB	c7, c8, t4

	STF	t3, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t4, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a3, c1
	FMUL	S, b3, c2
	FMUL	C, b3, c3
	FMUL	S, a3, c4

	FADD	c1, c2, t1
	FSUB	c3, c4, t2

	STF	t1, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t2, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a4, c5
	FMUL	S, b4, c6
	FMUL	C, b4, c7
	FMUL	S, a4, c8

	FADD	c5, c6, t3
	FSUB	c7, c8, t4

	STF	t3, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t4, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a5, c1
	FMUL	S, b5, c2
	FMUL	C, b5, c3
	FMUL	S, a5, c4

	FADD	c1, c2, t1
	FSUB	c3, c4, t2

	STF	t1, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t2, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a6, c5
	FMUL	S, b6, c6
	FMUL	C, b6, c7
	FMUL	S, a6, c8

	FADD	c5, c6, t3
	FSUB	c7, c8, t4

	STF	t3, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t4, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a7, c1
	FMUL	S, b7, c2
	FMUL	C, b7, c3
	FMUL	S, a7, c4

	FADD	c1, c2, t1
	FSUB	c3, c4, t2

	STF	t1, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t2, [YY +  0 * SIZE]
	add	YY, INCY, YY

	FMUL	C, a8, c5
	FMUL	S, b8, c6
	FMUL	C, b8, c7
	FMUL	S, a8, c8

	FADD	c5, c6, t3
	FSUB	c7, c8, t4

	STF	t3, [XX +  0 * SIZE]
	add	XX, INCX, XX
	STF	t4, [YY +  0 * SIZE]
	add	YY, INCY, YY

	addcc	I, -1, I
	bg,pt	%icc, .LL51
	nop


.LL55:
	andcc	N, 7, I
	nop
	ble	%icc, .LL59
	nop

.LL56:
	LDF	[X +  0 * SIZE], a1
	LDF	[Y +  0 * SIZE], b1

	FMUL	C, a1, c1
	FMUL	S, b1, c2
	FMUL	C, b1, c3
	FMUL	S, a1, c4

	FADD	c1, c2, c2
	FSUB	c3, c4, c4

	STF	c2, [X +  0 * SIZE]
	add	X, INCX, X
	STF	c4, [Y +  0 * SIZE]
	addcc	I, -1, I

	bg	%icc, .LL56
	add	Y, INCY, Y


.LL59:
	return	%i7 + 8
	nop

	EPILOGUE