/*********************************************************************/ /* Copyright 2009, 2010 The University of Texas at Austin. */ /* All rights reserved. */ /* */ /* Redistribution and use in source and binary forms, with or */ /* without modification, are permitted provided that the following */ /* conditions are met: */ /* */ /* 1. Redistributions of source code must retain the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer. */ /* */ /* 2. Redistributions in binary form must reproduce the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer in the documentation and/or other materials */ /* provided with the distribution. */ /* */ /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ /* POSSIBILITY OF SUCH DAMAGE. */ /* */ /* The views and conclusions contained in the software and */ /* documentation are those of the authors and should not be */ /* interpreted as representing official policies, either expressed */ /* or implied, of The University of Texas at Austin. */ /*********************************************************************/ #define ASSEMBLER #include "common.h" #define N r3 #define X r4 #define INCX r5 #define Y r6 #define INCY r7 #define INCX2 r8 #define INCY2 r9 #define C1 f1 #define C2 f0 #define C3 f2 #define C4 f3 #define A1 f4 #define A2 f5 #define A3 f6 #define A4 f7 #define A5 f8 #define A6 f9 #define A7 f10 #define A8 f11 #define A9 f20 #define B1 f12 #define B2 f13 #define B3 f14 #define B4 f15 #define B5 f16 #define B6 f17 #define B7 f18 #define B8 f19 #define B9 f20 PROLOGUE PROFCODE li r10, -16 stfpdux f14, SP, r10 stfpdux f15, SP, r10 stfpdux f16, SP, r10 stfpdux f17, SP, r10 stfpdux f18, SP, r10 stfpdux f19, SP, r10 stfpdux f20, SP, r10 li r10, 0 stwu r10, -4(SP) stwu r10, -4(SP) stwu r10, -4(SP) stwu r10, -4(SP) #ifdef F_INTERFACE LDINT N, 0(N) LDINT INCX, 0(INCX) LDINT INCY, 0(INCY) #endif lfpdx C1, SP, r10 # Zero clear slwi INCX, INCX, BASE_SHIFT add INCX2, INCX, INCX fpmr C2, C1 slwi INCY, INCY, BASE_SHIFT fpmr C3, C1 add INCY2, INCY, INCY fpmr C4, C1 cmpwi cr0, N, 0 ble LL(999) cmpwi cr0, INCX, SIZE bne LL(100) cmpwi cr0, INCY, SIZE bne LL(100) /* X is aligned, Y is aligned */ LL(10): andi. r0, X, 2 * SIZE - 1 bne LL(30) andi. r0, Y, 2 * SIZE - 1 bne LL(20) sub X, X, INCX2 sub Y, Y, INCY2 srawi. r0, N, 4 mtspr CTR, r0 beq- LL(15) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 LFPDUX A5, X, INCX2 LFPDUX B5, Y, INCY2 LFPDUX A6, X, INCX2 LFPDUX B6, Y, INCY2 LFPDUX A7, X, INCX2 LFPDUX B7, Y, INCY2 LFPDUX A8, X, INCX2 LFPDUX B8, Y, INCY2 bdz LL(14) .align 4 LL(13): fpmadd C1, A1, B1, C1 LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 fpmadd C2, A2, B2, C2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 fpmadd C3, A3, B3, C3 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 fpmadd C4, A4, B4, C4 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 fpmadd C1, A5, B5, C1 LFPDUX A5, X, INCX2 LFPDUX B5, Y, INCY2 fpmadd C2, A6, B6, C2 LFPDUX A6, X, INCX2 LFPDUX B6, Y, INCY2 fpmadd C3, A7, B7, C3 LFPDUX A7, X, INCX2 LFPDUX B7, Y, INCY2 fpmadd C4, A8, B8, C4 LFPDUX A8, X, INCX2 LFPDUX B8, Y, INCY2 bdnz LL(13) .align 4 LL(14): fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 fpmadd C1, A5, B5, C1 fpmadd C2, A6, B6, C2 fpmadd C3, A7, B7, C3 fpmadd C4, A8, B8, C4 .align 4 LL(15): andi. r0, N, 15 beq LL(999) andi. r0, N, 8 beq LL(16) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 .align 4 LL(16): andi. r0, N, 4 beq LL(17) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 .align 4 LL(17): andi. r0, N, 2 beq LL(18) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 fpmadd C1, A1, B1, C1 .align 4 LL(18): andi. r0, N, 1 beq LL(999) LFDUX A1, X, INCX2 LFDUX B1, Y, INCY2 fmadd C1, A1, B1, C1 b LL(999) .align 4 /* X is aligned, Y is NOT aligned */ LL(20): LFD B1, 0 * SIZE(Y) sub X, X, INCX2 sub Y, Y, INCY srawi. r0, N, 4 mtspr CTR, r0 beq- LL(25) LFPDUX A1, X, INCX2 LFXDUX B2, Y, INCY2 LFPDUX A2, X, INCX2 LFXDUX B3, Y, INCY2 LFPDUX A3, X, INCX2 LFXDUX B4, Y, INCY2 LFPDUX A4, X, INCX2 LFXDUX B5, Y, INCY2 LFPDUX A5, X, INCX2 LFXDUX B6, Y, INCY2 LFPDUX A6, X, INCX2 LFXDUX B7, Y, INCY2 LFPDUX A7, X, INCX2 fsmr B1, B2 LFXDUX B8, Y, INCY2 fsmr B2, B3 LFPDUX A8, X, INCX2 fsmr B3, B4 bdz LL(24) .align 4 LL(23): fpmadd C1, A1, B1, C1 LFPDUX A1, X, INCX2 fsmr B4, B5 LFXDUX B9, Y, INCY2 fpmadd C2, A2, B2, C2 LFPDUX A2, X, INCX2 fsmr B5, B6 LFXDUX B2, Y, INCY2 fpmadd C3, A3, B3, C3 LFXDUX B3, Y, INCY2 fsmr B6, B7 LFPDUX A3, X, INCX2 fpmadd C4, A4, B4, C4 LFXDUX B4, Y, INCY2 fsmr B7, B8 LFPDUX A4, X, INCX2 fpmadd C1, A5, B5, C1 LFXDUX B5, Y, INCY2 fsmr B8, B9 LFPDUX A5, X, INCX2 fpmadd C2, A6, B6, C2 LFXDUX B6, Y, INCY2 fpmr B1, B9 LFPDUX A6, X, INCX2 fpmadd C3, A7, B7, C3 LFXDUX B7, Y, INCY2 fsmr B1, B2 LFPDUX A7, X, INCX2 fpmadd C4, A8, B8, C4 LFXDUX B8, Y, INCY2 fsmr B2, B3 LFPDUX A8, X, INCX2 fsmr B3, B4 bdnz LL(23) .align 4 LL(24): LFXDUX B9, Y, INCY2 fpmadd C1, A1, B1, C1 fsmr B4, B5 fpmadd C2, A2, B2, C2 fsmr B5, B6 fpmadd C3, A3, B3, C3 fsmr B6, B7 fpmadd C4, A4, B4, C4 fsmr B7, B8 fpmadd C1, A5, B5, C1 fsmr B8, B9 fpmadd C2, A6, B6, C2 fpmr B1, B9 fpmadd C3, A7, B7, C3 fpmadd C4, A8, B8, C4 .align 4 LL(25): andi. r0, N, 15 beq LL(999) andi. r0, N, 8 beq LL(26) LFPDUX A1, X, INCX2 LFXDUX B2, Y, INCY2 LFPDUX A2, X, INCX2 LFXDUX B3, Y, INCY2 LFPDUX A3, X, INCX2 LFXDUX B4, Y, INCY2 LFPDUX A4, X, INCX2 LFXDUX B5, Y, INCY2 fsmr B1, B2 fsmr B2, B3 fsmr B3, B4 fsmr B4, B5 fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 fpmr B1, B5 .align 4 LL(26): andi. r0, N, 4 beq LL(27) LFPDUX A1, X, INCX2 LFXDUX B2, Y, INCY2 LFPDUX A2, X, INCX2 LFXDUX B3, Y, INCY2 fsmr B1, B2 fsmr B2, B3 fpmadd C1, A1, B1, C1 fpmr B1, B3 fpmadd C2, A2, B2, C2 .align 4 LL(27): andi. r0, N, 2 beq LL(28) LFPDUX A1, X, INCX2 LFXDUX B2, Y, INCY2 fsmr B1, B2 fpmadd C1, A1, B1, C1 fpmr B1, B2 .align 4 LL(28): andi. r0, N, 1 beq LL(999) LFDUX A1, X, INCX2 fmadd C1, A1, B1, C1 b LL(999) .align 4 /* X is not aligned, Y is aligned */ LL(30): andi. r0, Y, 2 * SIZE - 1 bne LL(40) LFD A1, 0 * SIZE(X) sub X, X, INCX sub Y, Y, INCY2 srawi. r0, N, 4 mtspr CTR, r0 beq- LL(35) LFXDUX A2, X, INCX2 LFPDUX B1, Y, INCY2 LFXDUX A3, X, INCX2 LFPDUX B2, Y, INCY2 LFXDUX A4, X, INCX2 LFPDUX B3, Y, INCY2 LFXDUX A5, X, INCX2 LFPDUX B4, Y, INCY2 LFXDUX A6, X, INCX2 LFPDUX B5, Y, INCY2 LFXDUX A7, X, INCX2 LFPDUX B6, Y, INCY2 LFXDUX A8, X, INCX2 fsmr A1, A2 LFPDUX B7, Y, INCY2 fsmr A2, A3 LFPDUX B8, Y, INCY2 fsmr A3, A4 bdz LL(34) .align 4 LL(33): fpmadd C1, A1, B1, C1 LFXDUX A9, X, INCX2 fsmr A4, A5 LFPDUX B1, Y, INCY2 fpmadd C2, A2, B2, C2 LFXDUX A2, X, INCX2 fsmr A5, A6 LFPDUX B2, Y, INCY2 fpmadd C3, A3, B3, C3 LFXDUX A3, X, INCX2 fsmr A6, A7 LFPDUX B3, Y, INCY2 fpmadd C4, A4, B4, C4 LFXDUX A4, X, INCX2 fsmr A7, A8 LFPDUX B4, Y, INCY2 fpmadd C1, A5, B5, C1 LFXDUX A5, X, INCX2 fsmr A8, A9 LFPDUX B5, Y, INCY2 fpmadd C2, A6, B6, C2 LFXDUX A6, X, INCX2 fpmr A1, A9 LFPDUX B6, Y, INCY2 fpmadd C3, A7, B7, C3 LFXDUX A7, X, INCX2 fsmr A1, A2 LFPDUX B7, Y, INCY2 fpmadd C4, A8, B8, C4 LFXDUX A8, X, INCX2 fsmr A2, A3 LFPDUX B8, Y, INCY2 fsmr A3, A4 bdnz LL(33) .align 4 LL(34): LFXDUX A9, X, INCX2 fpmadd C1, A1, B1, C1 fsmr A4, A5 fpmadd C2, A2, B2, C2 fsmr A5, A6 fpmadd C3, A3, B3, C3 fsmr A6, A7 fpmadd C4, A4, B4, C4 fsmr A7, A8 fpmadd C1, A5, B5, C1 fsmr A8, A9 fpmadd C2, A6, B6, C2 fpmr A1, A9 fpmadd C3, A7, B7, C3 fpmadd C4, A8, B8, C4 .align 4 LL(35): andi. r0, N, 15 beq LL(999) andi. r0, N, 8 beq LL(36) LFXDUX A2, X, INCX2 LFPDUX B1, Y, INCY2 LFXDUX A3, X, INCX2 LFPDUX B2, Y, INCY2 LFXDUX A4, X, INCX2 LFPDUX B3, Y, INCY2 LFXDUX A5, X, INCX2 LFPDUX B4, Y, INCY2 fsmr A1, A2 fsmr A2, A3 fsmr A3, A4 fsmr A4, A5 fpmadd C1, A1, B1, C1 fpmr A1, A5 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 .align 4 LL(36): andi. r0, N, 4 beq LL(37) LFXDUX A2, X, INCX2 LFPDUX B1, Y, INCY2 LFXDUX A3, X, INCX2 LFPDUX B2, Y, INCY2 fsmr A1, A2 fsmr A2, A3 fpmadd C1, A1, B1, C1 fpmr A1, A3 fpmadd C2, A2, B2, C2 .align 4 LL(37): andi. r0, N, 2 beq LL(38) LFXDUX A2, X, INCX2 LFPDUX B1, Y, INCY2 fsmr A1, A2 fpmadd C1, A1, B1, C1 fpmr A1, A2 .align 4 LL(38): andi. r0, N, 1 beq LL(999) LFDUX B1, Y, INCY2 fmadd C1, A1, B1, C1 b LL(999) .align 4 /* X is NOT aligned, Y is NOT aligned */ LL(40): LFD A1, 0 * SIZE(X) LFD B1, 0 * SIZE(Y) sub X, X, INCX sub Y, Y, INCY addi N, N, -1 cmpwi cr0, N, 0 fmadd C1, A1, B1, C1 ble LL(999) srawi. r0, N, 4 mtspr CTR, r0 beq- LL(45) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 LFPDUX A5, X, INCX2 LFPDUX B5, Y, INCY2 LFPDUX A6, X, INCX2 LFPDUX B6, Y, INCY2 LFPDUX A7, X, INCX2 LFPDUX B7, Y, INCY2 LFPDUX A8, X, INCX2 LFPDUX B8, Y, INCY2 bdz LL(44) .align 4 LL(43): fpmadd C1, A1, B1, C1 LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 fpmadd C2, A2, B2, C2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 fpmadd C3, A3, B3, C3 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 fpmadd C4, A4, B4, C4 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 fpmadd C1, A5, B5, C1 LFPDUX A5, X, INCX2 LFPDUX B5, Y, INCY2 fpmadd C2, A6, B6, C2 LFPDUX A6, X, INCX2 LFPDUX B6, Y, INCY2 fpmadd C3, A7, B7, C3 LFPDUX A7, X, INCX2 LFPDUX B7, Y, INCY2 fpmadd C4, A8, B8, C4 LFPDUX A8, X, INCX2 LFPDUX B8, Y, INCY2 bdnz LL(43) .align 4 LL(44): fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 fpmadd C1, A5, B5, C1 fpmadd C2, A6, B6, C2 fpmadd C3, A7, B7, C3 fpmadd C4, A8, B8, C4 .align 4 LL(45): andi. r0, N, 15 beq LL(999) andi. r0, N, 8 beq LL(46) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 LFPDUX A3, X, INCX2 LFPDUX B3, Y, INCY2 LFPDUX A4, X, INCX2 LFPDUX B4, Y, INCY2 fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 fpmadd C3, A3, B3, C3 fpmadd C4, A4, B4, C4 .align 4 LL(46): andi. r0, N, 4 beq LL(47) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 LFPDUX A2, X, INCX2 LFPDUX B2, Y, INCY2 fpmadd C1, A1, B1, C1 fpmadd C2, A2, B2, C2 .align 4 LL(47): andi. r0, N, 2 beq LL(48) LFPDUX A1, X, INCX2 LFPDUX B1, Y, INCY2 fpmadd C1, A1, B1, C1 .align 4 LL(48): andi. r0, N, 1 beq LL(999) LFDUX A1, X, INCX2 LFDUX B1, Y, INCY2 fmadd C1, A1, B1, C1 b LL(999) .align 4 LL(100): #ifdef F_INTERFACE cmpwi cr0, INCX, 0 bge+ LL(101) subi r0, N, 1 mullw r0, r0, INCX sub X, X, r0 .align 4 LL(101): cmpwi cr0, INCY, 0 bge+ LL(102) subi r0, N, 1 mullw r0, r0, INCY sub Y, Y, r0 .align 4 LL(102): #endif sub X, X, INCX sub Y, Y, INCY srawi. r0, N, 3 mtspr CTR, r0 beq- LL(105) LFDUX A1, X, INCX LFDUX B1, Y, INCY LFDUX A2, X, INCX LFDUX B2, Y, INCY LFDUX A3, X, INCX LFDUX B3, Y, INCY LFDUX A4, X, INCX LFDUX B4, Y, INCY LFDUX A5, X, INCX LFDUX B5, Y, INCY LFDUX A6, X, INCX LFDUX B6, Y, INCY LFDUX A7, X, INCX LFDUX B7, Y, INCY LFDUX A8, X, INCX LFDUX B8, Y, INCY bdz LL(104) .align 4 LL(103): fmadd C1, A1, B1, C1 LFDUX A1, X, INCX LFDUX B1, Y, INCY fmadd C2, A2, B2, C2 LFDUX A2, X, INCX LFDUX B2, Y, INCY fmadd C3, A3, B3, C3 LFDUX A3, X, INCX LFDUX B3, Y, INCY fmadd C4, A4, B4, C4 LFDUX A4, X, INCX LFDUX B4, Y, INCY fmadd C1, A5, B5, C1 LFDUX A5, X, INCX LFDUX B5, Y, INCY fmadd C2, A6, B6, C2 LFDUX A6, X, INCX LFDUX B6, Y, INCY fmadd C3, A7, B7, C3 LFDUX A7, X, INCX LFDUX B7, Y, INCY fmadd C4, A8, B8, C4 LFDUX A8, X, INCX LFDUX B8, Y, INCY bdnz LL(103) .align 4 LL(104): fmadd C1, A1, B1, C1 fmadd C2, A2, B2, C2 fmadd C3, A3, B3, C3 fmadd C4, A4, B4, C4 fmadd C1, A5, B5, C1 fmadd C2, A6, B6, C2 fmadd C3, A7, B7, C3 fmadd C4, A8, B8, C4 .align 4 LL(105): andi. r0, N, 7 beq LL(999) andi. r0, N, 4 beq LL(107) LFDUX A1, X, INCX LFDUX B1, Y, INCY LFDUX A2, X, INCX LFDUX B2, Y, INCY LFDUX A3, X, INCX LFDUX B3, Y, INCY LFDUX A4, X, INCX LFDUX B4, Y, INCY fmadd C1, A1, B1, C1 fmadd C2, A2, B2, C2 fmadd C3, A3, B3, C3 fmadd C4, A4, B4, C4 .align 4 LL(107): andi. r0, N, 2 beq LL(108) LFDUX A1, X, INCX LFDUX B1, Y, INCY LFDUX A2, X, INCX LFDUX B2, Y, INCY fmadd C1, A1, B1, C1 fmadd C2, A2, B2, C2 .align 4 LL(108): andi. r0, N, 1 beq LL(999) LFDUX A1, X, INCX LFDUX B1, Y, INCY fmadd C1, A1, B1, C1 .align 4 LL(999): li r10, 16 fpadd C1, C1, C2 fpadd C3, C3, C4 fpadd C1, C1, C3 lfpdux f20, SP, r10 lfpdux f19, SP, r10 lfpdux f18, SP, r10 lfpdux f17, SP, r10 fsmtp C2, C1 lfpdux f16, SP, r10 lfpdux f15, SP, r10 lfpdux f14, SP, r10 fadd C1, C1, C2 addi SP, SP, 16 blr EPILOGUE