Blob Blame Raw
/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin.           */
/* All rights reserved.                                              */
/*                                                                   */
/* Redistribution and use in source and binary forms, with or        */
/* without modification, are permitted provided that the following   */
/* conditions are met:                                               */
/*                                                                   */
/*   1. Redistributions of source code must retain the above         */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer.                                                  */
/*                                                                   */
/*   2. Redistributions in binary form must reproduce the above      */
/*      copyright notice, this list of conditions and the following  */
/*      disclaimer in the documentation and/or other materials       */
/*      provided with the distribution.                              */
/*                                                                   */
/*    THIS  SOFTWARE IS PROVIDED  BY THE  UNIVERSITY OF  TEXAS AT    */
/*    AUSTIN  ``AS IS''  AND ANY  EXPRESS OR  IMPLIED WARRANTIES,    */
/*    INCLUDING, BUT  NOT LIMITED  TO, THE IMPLIED  WARRANTIES OF    */
/*    MERCHANTABILITY  AND FITNESS FOR  A PARTICULAR  PURPOSE ARE    */
/*    DISCLAIMED.  IN  NO EVENT SHALL THE UNIVERSITY  OF TEXAS AT    */
/*    AUSTIN OR CONTRIBUTORS BE  LIABLE FOR ANY DIRECT, INDIRECT,    */
/*    INCIDENTAL,  SPECIAL, EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES    */
/*    (INCLUDING, BUT  NOT LIMITED TO,  PROCUREMENT OF SUBSTITUTE    */
/*    GOODS  OR  SERVICES; LOSS  OF  USE,  DATA,  OR PROFITS;  OR    */
/*    BUSINESS INTERRUPTION) HOWEVER CAUSED  AND ON ANY THEORY OF    */
/*    LIABILITY, WHETHER  IN CONTRACT, STRICT  LIABILITY, OR TORT    */
/*    (INCLUDING NEGLIGENCE OR OTHERWISE)  ARISING IN ANY WAY OUT    */
/*    OF  THE  USE OF  THIS  SOFTWARE,  EVEN  IF ADVISED  OF  THE    */
/*    POSSIBILITY OF SUCH DAMAGE.                                    */
/*                                                                   */
/* The views and conclusions contained in the software and           */
/* documentation are those of the authors and should not be          */
/* interpreted as representing official policies, either expressed   */
/* or implied, of The University of Texas at Austin.                 */
/*********************************************************************/

#define ASSEMBLER
#include "common.h"
		
#define N	r3
#define X	r6
#define INCX	r7	

#define INCX2	r4
#define XX	r5
#define Y	r8
#define YY	r9

#define ALPHA	f1
#define ALPHA_I	f2

#define A1	f0
#define A2	f16
#define A3	f17
#define A4	f3
#define A5	f4
#define A6	f5
#define A7	f6
#define A8	f7

#define B1	f8
#define B2	f9
#define B3	f10
#define B4	f11
#define B5	f12
#define B6	f13
#define B7	f14
#define B8	f15

	PROLOGUE
	PROFCODE

	li	r10, -16

	stfpdux	f14, SP, r10
	stfpdux	f15, SP, r10
	stfpdux	f16, SP, r10
	stfpdux	f17, SP, r10
	
	li	r10,   0
	stwu	r10,   -4(SP)
	stwu	r10,   -4(SP)
	stwu	r10,   -4(SP)
	stwu	r10,   -4(SP)

	lfpdx	A1, SP, r10		# Zero clear
	fsmfp	ALPHA, ALPHA_I

	slwi	INCX,  INCX, BASE_SHIFT
	add	INCX2, INCX, INCX

	cmpwi	cr0, N, 0
	ble	LL(999)

	cmpwi	cr0, INCX, SIZE
	bne	LL(100)

	fcmpu	cr7, ALPHA, A1
	bne	cr7, LL(50)

	fscmp	cr7, ALPHA, A1
	bne	cr7, LL(50)

	andi.	r0, X, 2 * SIZE - 1
	bne	LL(20)

	sub	X,  X, INCX2

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(15)
	.align 4

LL(12):
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	bdnz	LL(12)
	.align 4

LL(15):
	andi.	r0,  N, 3
	beq	LL(999)
	andi.	r0,  N, 2
	beq	LL(17)

	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	.align 4

LL(17):
	andi.	r0,  N, 1
	beq	LL(999)

	STFPDUX	A1,   X, INCX2
	b	LL(999)
	.align 4

LL(20):
	sub	X,  X, INCX2

	STFDX	A1, X, INCX2
	addi	X, X, SIZE
	addi	N, N, -1
	cmpwi	cr0, N, 0
	ble	LL(29)

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(25)
	.align 4

LL(22):
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	bdnz	LL(22)
	.align 4

LL(25):
	andi.	r0,  N, 3
	beq	LL(29)
	andi.	r0,  N, 2
	beq	LL(27)

	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	.align 4

LL(27):
	andi.	r0,  N, 1
	beq	LL(29)

	STFPDUX	A1,   X, INCX2
	.align 4

LL(29):
	STFDX	A1,   X, INCX2
	b	LL(999)
	.align 4

LL(50):
	sub	Y,  X, INCX2
	sub	X,  X, INCX2

	andi.	r0, X, 2 * SIZE - 1
	bne	LL(60)

	srawi.	r0, N, 3
	mtspr	CTR,  r0
	beq-	LL(55)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2
	LFPDUX	A3,  X,  INCX2
	LFPDUX	A4,  X,  INCX2

	LFPDUX	A5,  X,  INCX2
	fxpmul	B1,  ALPHA, A1
	LFPDUX	A6,  X,  INCX2
	fxpmul	B2,  ALPHA, A2
	LFPDUX	A7,  X,  INCX2
	fxpmul	B3,  ALPHA, A3
	LFPDUX	A8,  X,  INCX2
	fxpmul	B4,  ALPHA, A4
 	fxpmul	B5,  ALPHA, A5

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2
	bdz	LL(53)
	.align 4

LL(52):
	fxcxnpma B3, ALPHA, A3, B3
	LFPDUX	A1,  X,  INCX2
	fxpmul	B6,  ALPHA, A6
	STFPDUX	B1,   Y,  INCX2

	fxcxnpma B4, ALPHA, A4, B4
	LFPDUX	A2,  X,  INCX2
	fxpmul	B7,  ALPHA, A7
	STFPDUX	B2,   Y,  INCX2

	fxcxnpma B5, ALPHA, A5, B5
	LFPDUX	A3,  X,  INCX2
	fxpmul	B8,  ALPHA, A8
	STFPDUX	B3,   Y,  INCX2

	fxcxnpma B6, ALPHA, A6, B6
	LFPDUX	A4,  X,  INCX2
	fxpmul	B1,  ALPHA, A1
	STFPDUX	B4,   Y,  INCX2

	fxcxnpma B7, ALPHA, A7, B7
 	LFPDUX	A5,  X,  INCX2
	fxpmul	B2,  ALPHA, A2
	STFPDUX	B5,   Y,  INCX2

	fxcxnpma B8, ALPHA, A8, B8
	LFPDUX	A6,  X,  INCX2
	fxpmul	B3,  ALPHA, A3
	STFPDUX	B6,   Y,  INCX2

	fxcxnpma B1, ALPHA, A1, B1
	LFPDUX	A7,  X,  INCX2
	fxpmul	B4,  ALPHA, A4
	STFPDUX	B7,   Y,  INCX2

	fxcxnpma B2, ALPHA, A2, B2
	LFPDUX	A8,  X,  INCX2
	fxpmul	B5,  ALPHA, A5
	STFPDUX	B8,   Y,  INCX2
	bdnz	LL(52)
	.align 4

LL(53):
	fxcxnpma B3, ALPHA, A3, B3
	fxpmul	B6,  ALPHA, A6
	STFPDUX	B1,   Y,  INCX2

	fxcxnpma B4, ALPHA, A4, B4
	fxpmul	B7,  ALPHA, A7
	STFPDUX	B2,   Y,  INCX2

	fxcxnpma B5, ALPHA, A5, B5
	fxpmul	B8,  ALPHA, A8
	STFPDUX	B3,   Y,  INCX2

	fxcxnpma B6, ALPHA, A6, B6
	STFPDUX	B4,   Y,  INCX2
	fxcxnpma B7, ALPHA, A7, B7
	STFPDUX	B5,   Y,  INCX2
	fxcxnpma B8, ALPHA, A8, B8
	STFPDUX	B6,   Y,  INCX2
	STFPDUX	B7,   Y,  INCX2
	STFPDUX	B8,   Y,  INCX2
	.align 4

LL(55):
	andi.	r0,  N, 7
	beq	LL(999)

	andi.	r0,  N, 4
	beq	LL(56)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2
	LFPDUX	A3,  X,  INCX2
	LFPDUX	A4,  X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxpmul	B2,  ALPHA, A2
	fxpmul	B3,  ALPHA, A3
	fxpmul	B4,  ALPHA, A4

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2
	fxcxnpma B3, ALPHA, A3, B3
	fxcxnpma B4, ALPHA, A4, B4

	STFPDUX	B1,   Y,  INCX2
	STFPDUX	B2,   Y,  INCX2
	STFPDUX	B3,   Y,  INCX2
	STFPDUX	B4,   Y,  INCX2
	.align 4

LL(56):
	andi.	r0,  N, 2
	beq	LL(57)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxpmul	B2,  ALPHA, A2

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2

	STFPDUX	B1,   Y,  INCX2
	STFPDUX	B2,   Y,  INCX2
	.align 4

LL(57):
	andi.	r0,  N, 1
	beq	LL(999)

	LFPDUX	A1,   X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxcxnpma B1, ALPHA, A1, B1

	STFPDUX	B1,   Y,  INCX2
	b	LL(999)
	.align 4

LL(60):
	addi	XX, X, SIZE
	addi	YY, Y, SIZE

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(65)

	LFDUX	A1,  X,  INCX2
	LFDUX	A2,  XX, INCX2
	LFDUX	A3,  X,  INCX2
	LFDUX	A4,  XX, INCX2

	LFDUX	A5,  X,  INCX2
	fmul	B1, ALPHA,   A1
	LFDUX	A6,  XX, INCX2
	fmul	B2, ALPHA_I, A1
	LFDUX	A7,  X,  INCX2
	fmul	B3, ALPHA,   A3
	LFDUX	A8,  XX, INCX2
	fmul	B4, ALPHA_I, A3

	fmul	B5, ALPHA,   A5
	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA  , A2, B2
	bdz	LL(63)
	.align 4

LL(62):
	fnmsub	B3, ALPHA_I, A4, B3
 	LFDUX	A1,  X,  INCX2
	fmul	B6, ALPHA_I, A5
	STFDUX	B1,  Y,  INCX2

	fmadd	B4, ALPHA  , A4, B4
	LFDUX	A2,  XX, INCX2
	fmul	B7, ALPHA,   A7
	STFDUX	B2,  YY, INCX2

	fnmsub	B5, ALPHA_I, A6, B5
	LFDUX	A3,  X,  INCX2
	fmul	B8, ALPHA_I, A7
	STFDUX	B3,  Y,  INCX2

	fmadd	B6, ALPHA  , A6, B6
	LFDUX	A4,  XX, INCX2
	fmul	B1, ALPHA,   A1
	STFDUX	B4,  YY, INCX2

	fnmsub	B7, ALPHA_I, A8, B7
	LFDUX	A5,  X,  INCX2
	fmul	B2, ALPHA_I, A1
	STFDUX	B5,  Y,  INCX2

	fmadd	B8, ALPHA  , A8, B8
	LFDUX	A6,  XX, INCX2
	fmul	B3, ALPHA,   A3
	STFDUX	B6,  YY, INCX2

	fnmsub	B1, ALPHA_I, A2, B1
	LFDUX	A7,  X,  INCX2
	fmul	B4, ALPHA_I, A3
	STFDUX	B7,  Y,  INCX2

	fmadd	B2, ALPHA  , A2, B2
	LFDUX	A8,  XX, INCX2
	fmul	B5, ALPHA,   A5
	STFDUX	B8,  YY, INCX2
	bdnz	LL(62)
	.align 4

LL(63):
	fnmsub	B3, ALPHA_I, A4, B3
	fmul	B6, ALPHA_I, A5
	STFDUX	B1,  Y,  INCX2

	fmadd	B4, ALPHA  , A4, B4
	fmul	B7, ALPHA,   A7
	STFDUX	B2,  YY, INCX2

	fnmsub	B5, ALPHA_I, A6, B5
	fmul	B8, ALPHA_I, A7
	STFDUX	B3,  Y,  INCX2

	fmadd	B6, ALPHA  , A6, B6
	STFDUX	B4,  YY, INCX2
	fnmsub	B7, ALPHA_I, A8, B7
	STFDUX	B5,  Y,  INCX2
	fmadd	B8, ALPHA  , A8, B8
	STFDUX	B6,  YY, INCX2
	STFDUX	B7,  Y,  INCX2
	STFDUX	B8,  YY, INCX2
	.align 4

LL(65):
	andi.	r0,  N, 3
	beq	LL(999)
	andi.	r0,  N, 2
	beq	LL(67)

	LFDUX	A1,  X,  INCX2
	LFDUX	A2,  XX, INCX2
	LFDUX	A3,  X,  INCX2
	LFDUX	A4,  XX, INCX2

	fmul	B1, ALPHA, A1
	fmul	B2, ALPHA, A2
	fmul	B3, ALPHA, A3
	fmul	B4, ALPHA, A4

	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA_I, A1, B2
	fnmsub	B3, ALPHA_I, A4, B3
	fmadd	B4, ALPHA_I, A3, B4

	STFDUX	B1,  Y,  INCX2
	STFDUX	B2,  YY, INCX2
	STFDUX	B3,  Y,  INCX2
	STFDUX	B4,  YY, INCX2
	.align 4

LL(67):
	andi.	r0,  N, 1
	beq	LL(999)

	LFDUX	A1,   X,  INCX2
	LFDUX	A2,   XX, INCX2

	fmul	B1, ALPHA, A1
	fmul	B2, ALPHA, A2
	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA_I, A1, B2

	STFDUX	B1,   Y,  INCX2
	STFDUX	B2,   YY, INCX2
	b	LL(999)
	.align 4


LL(100):
	fcmpu	cr7, ALPHA, A1
	bne	cr7, LL(150)

	fscmp	cr7, ALPHA, A1
	bne	cr7, LL(150)

	andi.	r0, X, 2 * SIZE - 1
	bne	LL(120)

	sub	X,  X, INCX2

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(115)
	.align 4

LL(112):
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	bdnz	LL(112)
	.align 4

LL(115):
	andi.	r0,  N, 3
	beq	LL(999)
	andi.	r0,  N, 2
	beq	LL(117)

	STFPDUX	A1,   X, INCX2
	STFPDUX	A1,   X, INCX2
	.align 4

LL(117):
	andi.	r0,  N, 1
	beq	LL(999)

	STFPDUX	A1,   X, INCX2
	b	LL(999)
	.align 4

LL(120):
	subi	INCX2, INCX2, SIZE
	li	INCX, SIZE

	sub	X,  X, INCX2

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(125)
	.align 4

LL(122):
	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	bdnz	LL(122)
	.align 4

LL(125):
	andi.	r0,  N, 3
	beq	LL(999)
	andi.	r0,  N, 2
	beq	LL(127)

	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	.align 4

LL(127):
	andi.	r0,  N, 1
	beq	LL(999)

	STFDUX	A1,   X, INCX2
	STFDUX	A1,   X, INCX
	b	LL(999)
	.align 4

LL(150):
	sub	Y,  X, INCX2
	sub	X,  X, INCX2

	andi.	r0, X, 2 * SIZE - 1
	bne	LL(160)

	srawi.	r0, N, 3
	mtspr	CTR,  r0
	beq-	LL(155)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2
	LFPDUX	A3,  X,  INCX2
	LFPDUX	A4,  X,  INCX2

	LFPDUX	A5,  X,  INCX2
	fxpmul	B1,  ALPHA, A1
	LFPDUX	A6,  X,  INCX2
	fxpmul	B2,  ALPHA, A2
	LFPDUX	A7,  X,  INCX2
	fxpmul	B3,  ALPHA, A3
	LFPDUX	A8,  X,  INCX2
	fxpmul	B4,  ALPHA, A4
 	fxpmul	B5,  ALPHA, A5

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2
	bdz	LL(153)
	.align 4

LL(152):
	fxcxnpma B3, ALPHA, A3, B3
	LFPDUX	A1,  X,  INCX2
	fxpmul	B6,  ALPHA, A6
	STFPDUX	B1,   Y,  INCX2

	fxcxnpma B4, ALPHA, A4, B4
	LFPDUX	A2,  X,  INCX2
	fxpmul	B7,  ALPHA, A7
	STFPDUX	B2,   Y,  INCX2

	fxcxnpma B5, ALPHA, A5, B5
	LFPDUX	A3,  X,  INCX2
	fxpmul	B8,  ALPHA, A8
	STFPDUX	B3,   Y,  INCX2

	fxcxnpma B6, ALPHA, A6, B6
	LFPDUX	A4,  X,  INCX2
	fxpmul	B1,  ALPHA, A1
	STFPDUX	B4,   Y,  INCX2

	fxcxnpma B7, ALPHA, A7, B7
 	LFPDUX	A5,  X,  INCX2
	fxpmul	B2,  ALPHA, A2
	STFPDUX	B5,   Y,  INCX2

	fxcxnpma B8, ALPHA, A8, B8
	LFPDUX	A6,  X,  INCX2
	fxpmul	B3,  ALPHA, A3
	STFPDUX	B6,   Y,  INCX2

	fxcxnpma B1, ALPHA, A1, B1
	LFPDUX	A7,  X,  INCX2
	fxpmul	B4,  ALPHA, A4
	STFPDUX	B7,   Y,  INCX2

	fxcxnpma B2, ALPHA, A2, B2
	LFPDUX	A8,  X,  INCX2
	fxpmul	B5,  ALPHA, A5
	STFPDUX	B8,   Y,  INCX2
	bdnz	LL(152)
	.align 4

LL(153):
	fxcxnpma B3, ALPHA, A3, B3
	fxpmul	B6,  ALPHA, A6
	STFPDUX	B1,   Y,  INCX2

	fxcxnpma B4, ALPHA, A4, B4
	fxpmul	B7,  ALPHA, A7
	STFPDUX	B2,   Y,  INCX2

	fxcxnpma B5, ALPHA, A5, B5
	fxpmul	B8,  ALPHA, A8
	STFPDUX	B3,   Y,  INCX2

	fxcxnpma B6, ALPHA, A6, B6
	STFPDUX	B4,   Y,  INCX2
	fxcxnpma B7, ALPHA, A7, B7
	STFPDUX	B5,   Y,  INCX2
	fxcxnpma B8, ALPHA, A8, B8
	STFPDUX	B6,   Y,  INCX2
	STFPDUX	B7,   Y,  INCX2
	STFPDUX	B8,   Y,  INCX2
	.align 4

LL(155):
	andi.	r0,  N, 7
	beq	LL(999)

	andi.	r0,  N, 4
	beq	LL(156)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2
	LFPDUX	A3,  X,  INCX2
	LFPDUX	A4,  X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxpmul	B2,  ALPHA, A2
	fxpmul	B3,  ALPHA, A3
	fxpmul	B4,  ALPHA, A4

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2
	fxcxnpma B3, ALPHA, A3, B3
	fxcxnpma B4, ALPHA, A4, B4

	STFPDUX	B1,   Y,  INCX2
	STFPDUX	B2,   Y,  INCX2
	STFPDUX	B3,   Y,  INCX2
	STFPDUX	B4,   Y,  INCX2
	.align 4

LL(156):
	andi.	r0,  N, 2
	beq	LL(157)

	LFPDUX	A1,  X,  INCX2
	LFPDUX	A2,  X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxpmul	B2,  ALPHA, A2

	fxcxnpma B1, ALPHA, A1, B1
	fxcxnpma B2, ALPHA, A2, B2

	STFPDUX	B1,   Y,  INCX2
	STFPDUX	B2,   Y,  INCX2
	.align 4

LL(157):
	andi.	r0,  N, 1
	beq	LL(999)

	LFPDUX	A1,   X,  INCX2

	fxpmul	B1,  ALPHA, A1
	fxcxnpma B1, ALPHA, A1, B1

	STFPDUX	B1,   Y,  INCX2
	b	LL(999)
	.align 4

LL(160):
	addi	XX, X, SIZE
	addi	YY, Y, SIZE

	srawi.	r0, N, 2
	mtspr	CTR,  r0
	beq-	LL(165)

	LFDUX	A1,  X,  INCX2
	LFDUX	A2,  XX, INCX2
	LFDUX	A3,  X,  INCX2
	LFDUX	A4,  XX, INCX2

	LFDUX	A5,  X,  INCX2
	fmul	B1, ALPHA,   A1
	LFDUX	A6,  XX, INCX2
	fmul	B2, ALPHA_I, A1
	LFDUX	A7,  X,  INCX2
	fmul	B3, ALPHA,   A3
	LFDUX	A8,  XX, INCX2
	fmul	B4, ALPHA_I, A3

	fmul	B5, ALPHA,   A5
	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA  , A2, B2
	bdz	LL(163)

	.align 4

LL(162):
	fnmsub	B3, ALPHA_I, A4, B3
 	LFDUX	A1,  X,  INCX2
	fmul	B6, ALPHA_I, A5
	STFDUX	B1,  Y,  INCX2

	fmadd	B4, ALPHA  , A4, B4
	LFDUX	A2,  XX, INCX2
	fmul	B7, ALPHA,   A7
	STFDUX	B2,  YY, INCX2

	fnmsub	B5, ALPHA_I, A6, B5
	LFDUX	A3,  X,  INCX2
	fmul	B8, ALPHA_I, A7
	STFDUX	B3,  Y,  INCX2

	fmadd	B6, ALPHA  , A6, B6
	LFDUX	A4,  XX, INCX2
	fmul	B1, ALPHA,   A1
	STFDUX	B4,  YY, INCX2

	fnmsub	B7, ALPHA_I, A8, B7
	LFDUX	A5,  X,  INCX2
	fmul	B2, ALPHA_I, A1
	STFDUX	B5,  Y,  INCX2

	fmadd	B8, ALPHA  , A8, B8
	LFDUX	A6,  XX, INCX2
	fmul	B3, ALPHA,   A3
	STFDUX	B6,  YY, INCX2

	fnmsub	B1, ALPHA_I, A2, B1
	LFDUX	A7,  X,  INCX2
	fmul	B4, ALPHA_I, A3
	STFDUX	B7,  Y,  INCX2

	fmadd	B2, ALPHA  , A2, B2
	LFDUX	A8,  XX, INCX2
	fmul	B5, ALPHA,   A5
	STFDUX	B8,  YY, INCX2
	bdnz	LL(162)
	.align 4

LL(163):
	fnmsub	B3, ALPHA_I, A4, B3
	fmul	B6, ALPHA_I, A5
	STFDUX	B1,  Y,  INCX2

	fmadd	B4, ALPHA  , A4, B4
	fmul	B7, ALPHA,   A7
	STFDUX	B2,  YY, INCX2

	fnmsub	B5, ALPHA_I, A6, B5
	fmul	B8, ALPHA_I, A7
	STFDUX	B3,  Y,  INCX2

	fmadd	B6, ALPHA  , A6, B6
	STFDUX	B4,  YY, INCX2
	fnmsub	B7, ALPHA_I, A8, B7
	STFDUX	B5,  Y,  INCX2
	fmadd	B8, ALPHA  , A8, B8
	STFDUX	B6,  YY, INCX2
	STFDUX	B7,  Y,  INCX2
	STFDUX	B8,  YY, INCX2
	.align 4

LL(165):
	andi.	r0,  N, 3
	beq	LL(999)
	andi.	r0,  N, 2
	beq	LL(167)

	LFDUX	A1,  X,  INCX2
	LFDUX	A2,  XX, INCX2
	LFDUX	A3,  X,  INCX2
	LFDUX	A4,  XX, INCX2

	fmul	B1, ALPHA, A1
	fmul	B2, ALPHA, A2
	fmul	B3, ALPHA, A3
	fmul	B4, ALPHA, A4

	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA_I, A1, B2
	fnmsub	B3, ALPHA_I, A4, B3
	fmadd	B4, ALPHA_I, A3, B4

	STFDUX	B1,  Y,  INCX2
	STFDUX	B2,  YY, INCX2
	STFDUX	B3,  Y,  INCX2
	STFDUX	B4,  YY, INCX2
	.align 4

LL(167):
	andi.	r0,  N, 1
	beq	LL(999)

	LFDUX	A1,   X,  INCX2
	LFDUX	A2,   XX, INCX2

	fmul	B1, ALPHA, A1
	fmul	B2, ALPHA, A2
	fnmsub	B1, ALPHA_I, A2, B1
	fmadd	B2, ALPHA_I, A1, B2

	STFDUX	B1,   Y,  INCX2
	STFDUX	B2,   YY, INCX2
	.align 4

LL(999):
	li	r10, 16

	lfpdux	f17, SP, r10
	lfpdux	f16, SP, r10
	lfpdux	f15, SP, r10
	lfpdux	f14, SP, r10

	addi	SP, SP,  16
	blr

	EPILOGUE