/*********************************************************************/ /* Copyright 2009, 2010 The University of Texas at Austin. */ /* All rights reserved. */ /* */ /* Redistribution and use in source and binary forms, with or */ /* without modification, are permitted provided that the following */ /* conditions are met: */ /* */ /* 1. Redistributions of source code must retain the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer. */ /* */ /* 2. Redistributions in binary form must reproduce the above */ /* copyright notice, this list of conditions and the following */ /* disclaimer in the documentation and/or other materials */ /* provided with the distribution. */ /* */ /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ /* POSSIBILITY OF SUCH DAMAGE. */ /* */ /* The views and conclusions contained in the software and */ /* documentation are those of the authors and should not be */ /* interpreted as representing official policies, either expressed */ /* or implied, of The University of Texas at Austin. */ /*********************************************************************/ #define ASSEMBLER #include "common.h" #define STACK 16 #define ARGS 8 #define J 0 + STACK(%esp) #define BOFFSET2 4 + STACK(%esp) #define M 4 + STACK + ARGS(%esp) #define N 8 + STACK + ARGS(%esp) #define A 12 + STACK + ARGS(%esp) #define LDA 16 + STACK + ARGS(%esp) #define B 20 + STACK + ARGS(%esp) PROLOGUE subl $ARGS, %esp pushl %ebp pushl %edi pushl %esi pushl %ebx PROFCODE EMMS movl A, %ebp movl B, %edi movl M, %ebx movl N, %eax andl $-2, %eax imull %ebx, %eax # m * ( n & ~1) leal (%edi,%eax,SIZE), %eax # boffset2 = b + m * (n & ~1) movl %eax, BOFFSET2 movl M, %esi #ifdef DOUBLE sall $4,%esi #else sall $3,%esi #endif sarl $1, %ebx # if !(m & 1) goto L28 movl %ebx, J jle .L28 ALIGN_4 .L39: movl %ebp, %edx # aoffset1 = a movl LDA, %eax movl N, %ebx leal (%ebp, %eax,SIZE), %ecx # aoffset2 = a + lda leal (%ecx, %eax,SIZE), %ebp # aoffset += 2 * lda movl %edi, %eax # boffset1 = b_offset addl $4 * SIZE, %edi # boffset += 4 sarl $2, %ebx jle .L32 ALIGN_4 .L36: #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%edx), %mm0 MMXLOAD 1 * SIZE(%edx), %mm1 MMXLOAD 0 * SIZE(%ecx), %mm2 MMXLOAD 1 * SIZE(%ecx), %mm3 MMXLOAD 2 * SIZE(%edx), %mm4 MMXLOAD 3 * SIZE(%edx), %mm5 MMXLOAD 2 * SIZE(%ecx), %mm6 MMXLOAD 3 * SIZE(%ecx), %mm7 MMXSTORE %mm0, 0 * SIZE(%eax) MMXSTORE %mm1, 1 * SIZE(%eax) MMXSTORE %mm2, 2 * SIZE(%eax) MMXSTORE %mm3, 3 * SIZE(%eax) addl %esi, %eax MMXSTORE %mm4, 0 * SIZE(%eax) MMXSTORE %mm5, 1 * SIZE(%eax) MMXSTORE %mm6, 2 * SIZE(%eax) MMXSTORE %mm7, 3 * SIZE(%eax) #else FLD 1 * SIZE(%ecx) FLD 0 * SIZE(%ecx) FLD 1 * SIZE(%edx) FLD 0 * SIZE(%edx) FST 0 * SIZE(%eax) FST 1 * SIZE(%eax) FST 2 * SIZE(%eax) FST 3 * SIZE(%eax) addl %esi, %eax FLD 3 * SIZE(%ecx) FLD 2 * SIZE(%ecx) FLD 3 * SIZE(%edx) FLD 2 * SIZE(%edx) FST 0 * SIZE(%eax) FST 1 * SIZE(%eax) FST 2 * SIZE(%eax) FST 3 * SIZE(%eax) #endif addl $4 * SIZE, %ecx addl $4 * SIZE, %edx addl %esi, %eax decl %ebx jne .L36 ALIGN_4 .L32: movl N, %ebx test $2, %ebx je .L37 #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%edx), %mm0 MMXLOAD 1 * SIZE(%edx), %mm1 MMXLOAD 0 * SIZE(%ecx), %mm2 MMXLOAD 1 * SIZE(%ecx), %mm3 MMXSTORE %mm0, 0 * SIZE(%eax) MMXSTORE %mm1, 1 * SIZE(%eax) MMXSTORE %mm2, 2 * SIZE(%eax) MMXSTORE %mm3, 3 * SIZE(%eax) #else FLD 1 * SIZE(%ecx) FLD 0 * SIZE(%ecx) FLD 1 * SIZE(%edx) FLD 0 * SIZE(%edx) FST 0 * SIZE(%eax) FST 1 * SIZE(%eax) FST 2 * SIZE(%eax) FST 3 * SIZE(%eax) #endif addl $2 * SIZE, %ecx addl $2 * SIZE, %edx ALIGN_4 .L37: movl N, %ebx test $1, %ebx je .L38 movl BOFFSET2, %eax #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%edx), %mm0 MMXLOAD 0 * SIZE(%ecx), %mm1 MMXSTORE %mm0, 0 * SIZE(%eax) MMXSTORE %mm1, 1 * SIZE(%eax) #else FLD 0 * SIZE(%edx) FST 0 * SIZE(%eax) FLD 0 * SIZE(%ecx) FST 1 * SIZE(%eax) #endif addl $2 * SIZE, %eax movl %eax, BOFFSET2 ALIGN_4 .L38: decl J jg .L39 ALIGN_4 .L28: movl M, %eax movl N, %ebx testb $1, %al je .L40 sarl $2, %ebx jle .L41 ALIGN_4 .L45: #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%ebp), %mm0 MMXLOAD 1 * SIZE(%ebp), %mm1 MMXLOAD 2 * SIZE(%ebp), %mm2 MMXLOAD 3 * SIZE(%ebp), %mm3 MMXSTORE %mm0, 0 * SIZE(%edi) MMXSTORE %mm1, 1 * SIZE(%edi) addl %esi, %edi MMXSTORE %mm2, 0 * SIZE(%edi) MMXSTORE %mm3, 1 * SIZE(%edi) #else FLD 0 * SIZE(%ebp) FST 0 * SIZE(%edi) FLD 1 * SIZE(%ebp) FST 1 * SIZE(%edi) addl %esi, %edi FLD 2 * SIZE(%ebp) FST 0 * SIZE(%edi) FLD 3 * SIZE(%ebp) FST 1 * SIZE(%edi) #endif addl %esi,%edi addl $4 * SIZE, %ebp decl %ebx jg .L45 ALIGN_4 .L41: movl N, %ebx test $2, %ebx je .L46 #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%ebp), %mm0 MMXSTORE %mm0, 0 * SIZE(%edi) MMXLOAD 1 * SIZE(%ebp), %mm1 MMXSTORE %mm1, 1 * SIZE(%edi) #else FLD 1 * SIZE(%ebp) FLD 0 * SIZE(%ebp) FST 0 * SIZE(%edi) FST 1 * SIZE(%edi) #endif addl $2 * SIZE, %ebp ALIGN_4 .L46: movl N, %ebx test $1, %ebx je .L40 movl BOFFSET2, %eax #ifdef HAVE_MMX MMXLOAD 0 * SIZE(%ebp), %mm0 MMXSTORE %mm0, 0 * SIZE(%eax) #else FLD (%ebp) FST (%eax) #endif ALIGN_4 .L40: EMMS popl %ebx popl %esi popl %edi popl %ebp addl $ARGS,%esp ret EPILOGUE