|
|
|
@@ -0,0 +1,811 @@ |
|
|
|
/*************************************************************************** |
|
|
|
Copyright (c) 2017, The OpenBLAS Project |
|
|
|
All rights reserved. |
|
|
|
Redistribution and use in source and binary forms, with or without |
|
|
|
modification, are permitted provided that the following conditions are |
|
|
|
met: |
|
|
|
1. Redistributions of source code must retain the above copyright |
|
|
|
notice, this list of conditions and the following disclaimer. |
|
|
|
2. Redistributions in binary form must reproduce the above copyright |
|
|
|
notice, this list of conditions and the following disclaimer in |
|
|
|
the documentation and/or other materials provided with the |
|
|
|
distribution. |
|
|
|
3. Neither the name of the OpenBLAS project nor the names of |
|
|
|
its contributors may be used to endorse or promote products |
|
|
|
derived from this software without specific prior written permission. |
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
|
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
|
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
|
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE |
|
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
|
|
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR |
|
|
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER |
|
|
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
|
|
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE |
|
|
|
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
|
|
*****************************************************************************/ |
|
|
|
|
|
|
|
#include "common.h" |
|
|
|
|
|
|
|
#define NBMAX 2048 |
|
|
|
|
|
|
|
static void sgemv_kernel_4x4(BLASLONG n, FLOAT **ap, FLOAT *x, FLOAT *y) |
|
|
|
{ |
|
|
|
__asm__ volatile ( |
|
|
|
"vzero %%v0 \n\t" |
|
|
|
"vzero %%v1 \n\t" |
|
|
|
"vzero %%v2 \n\t" |
|
|
|
"vzero %%v3 \n\t" |
|
|
|
"xgr %%r1,%%r1 \n\t" |
|
|
|
|
|
|
|
"lghi %%r0,-32 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 1f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,5 \n\t" |
|
|
|
"0: \n\t" |
|
|
|
"pfd 1,1024(%%r1,%1) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%2) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%3) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%4) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%5) \n\t" |
|
|
|
|
|
|
|
"vl %%v16,0(%%r1,%5) \n\t" |
|
|
|
"vl %%v17,16(%%r1,%5) \n\t" |
|
|
|
"vl %%v18,32(%%r1,%5) \n\t" |
|
|
|
"vl %%v19,48(%%r1,%5) \n\t" |
|
|
|
"vl %%v20,64(%%r1,%5) \n\t" |
|
|
|
"vl %%v21,80(%%r1,%5) \n\t" |
|
|
|
"vl %%v22,96(%%r1,%5) \n\t" |
|
|
|
"vl %%v23,112(%%r1,%5) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,0(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t" |
|
|
|
"vl %%v26,0(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v16,%%v26,%%v2 \n\t" |
|
|
|
"vl %%v27,0(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v16,%%v27,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,16(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v17,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,16(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v17,%%v29,%%v1 \n\t" |
|
|
|
"vl %%v30,16(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v17,%%v30,%%v2 \n\t" |
|
|
|
"vl %%v31,16(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v17,%%v31,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v24,32(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v18,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,32(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v18,%%v25,%%v1 \n\t" |
|
|
|
"vl %%v26,32(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v18,%%v26,%%v2 \n\t" |
|
|
|
"vl %%v27,32(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v18,%%v27,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,48(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v19,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,48(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v19,%%v29,%%v1 \n\t" |
|
|
|
"vl %%v30,48(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v19,%%v30,%%v2 \n\t" |
|
|
|
"vl %%v31,48(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v19,%%v31,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v24,64(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v20,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,64(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v20,%%v25,%%v1 \n\t" |
|
|
|
"vl %%v26,64(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v20,%%v26,%%v2 \n\t" |
|
|
|
"vl %%v27,64(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v20,%%v27,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,80(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v21,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,80(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v21,%%v29,%%v1 \n\t" |
|
|
|
"vl %%v30,80(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v21,%%v30,%%v2 \n\t" |
|
|
|
"vl %%v31,80(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v21,%%v31,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v24,96(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v22,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,96(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v22,%%v25,%%v1 \n\t" |
|
|
|
"vl %%v26,96(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v22,%%v26,%%v2 \n\t" |
|
|
|
"vl %%v27,96(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v22,%%v27,%%v3 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,112(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v23,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,112(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v23,%%v29,%%v1 \n\t" |
|
|
|
"vl %%v30,112(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v23,%%v30,%%v2 \n\t" |
|
|
|
"vl %%v31,112(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v23,%%v31,%%v3 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,128 \n\t" |
|
|
|
"brctg %%r0,0b \n\t" |
|
|
|
|
|
|
|
"1: \n\t" |
|
|
|
"lghi %%r0,28 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 3f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,2 \n\t" |
|
|
|
"2: \n\t" |
|
|
|
"vl %%v16,0(%%r1,%5) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,0(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t" |
|
|
|
"vl %%v26,0(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v2,%%v16,%%v26,%%v2 \n\t" |
|
|
|
"vl %%v27,0(%%r1,%4) \n\t" |
|
|
|
"vfmasb %%v3,%%v16,%%v27,%%v3 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,16 \n\t" |
|
|
|
"brctg %%r0,2b \n\t" |
|
|
|
|
|
|
|
"3: \n\t" |
|
|
|
"veslg %%v4,%%v0,32 \n\t" |
|
|
|
"vfasb %%v0,%%v0,%%v4 \n\t" |
|
|
|
"vrepg %%v4,%%v0,1 \n\t" |
|
|
|
"aebr %%f0,%%f4 \n\t" |
|
|
|
"ste %%f0,0(%6) \n\t" |
|
|
|
"veslg %%v4,%%v1,32 \n\t" |
|
|
|
"vfasb %%v1,%%v1,%%v4 \n\t" |
|
|
|
"vrepg %%v4,%%v1,1 \n\t" |
|
|
|
"aebr %%f1,%%f4 \n\t" |
|
|
|
"ste %%f1,4(%6) \n\t" |
|
|
|
"veslg %%v4,%%v2,32 \n\t" |
|
|
|
"vfasb %%v2,%%v2,%%v4 \n\t" |
|
|
|
"vrepg %%v4,%%v2,1 \n\t" |
|
|
|
"aebr %%f2,%%f4 \n\t" |
|
|
|
"ste %%f2,8(%6) \n\t" |
|
|
|
"veslg %%v4,%%v3,32 \n\t" |
|
|
|
"vfasb %%v3,%%v3,%%v4 \n\t" |
|
|
|
"vrepg %%v4,%%v3,1 \n\t" |
|
|
|
"aebr %%f3,%%f4 \n\t" |
|
|
|
"ste %%f3,12(%6) " |
|
|
|
: |
|
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])ap[0]),"ZR"((const FLOAT (*)[n])ap[1]),"ZR"((const FLOAT (*)[n])ap[2]),"ZR"((const FLOAT (*)[n])ap[3]),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[4])y) |
|
|
|
:"memory","cc","r0","r1","v0","v1","v2","v3","v4","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31" |
|
|
|
); |
|
|
|
} |
|
|
|
|
|
|
|
static void sgemv_kernel_4x2(BLASLONG n, FLOAT **ap, FLOAT *x, FLOAT *y) |
|
|
|
{ |
|
|
|
__asm__ volatile ( |
|
|
|
"vzero %%v0 \n\t" |
|
|
|
"vzero %%v1 \n\t" |
|
|
|
"xgr %%r1,%%r1 \n\t" |
|
|
|
|
|
|
|
"lghi %%r0,-32 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 1f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,5 \n\t" |
|
|
|
"0: \n\t" |
|
|
|
"pfd 1,1024(%%r1,%1) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%2) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"vl %%v16,0(%%r1,%3) \n\t" |
|
|
|
"vl %%v17,16(%%r1,%3) \n\t" |
|
|
|
"vl %%v18,32(%%r1,%3) \n\t" |
|
|
|
"vl %%v19,48(%%r1,%3) \n\t" |
|
|
|
"vl %%v20,64(%%r1,%3) \n\t" |
|
|
|
"vl %%v21,80(%%r1,%3) \n\t" |
|
|
|
"vl %%v22,96(%%r1,%3) \n\t" |
|
|
|
"vl %%v23,112(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,0(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v26,16(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v17,%%v26,%%v0 \n\t" |
|
|
|
"vl %%v27,16(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v17,%%v27,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,32(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v18,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,32(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v18,%%v29,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v30,48(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v19,%%v30,%%v0 \n\t" |
|
|
|
"vl %%v31,48(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v19,%%v31,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v24,64(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v20,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,64(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v20,%%v25,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v26,80(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v21,%%v26,%%v0 \n\t" |
|
|
|
"vl %%v27,80(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v21,%%v27,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,96(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v22,%%v28,%%v0 \n\t" |
|
|
|
"vl %%v29,96(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v22,%%v29,%%v1 \n\t" |
|
|
|
|
|
|
|
"vl %%v30,112(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v23,%%v30,%%v0 \n\t" |
|
|
|
"vl %%v31,112(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v23,%%v31,%%v1 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,128 \n\t" |
|
|
|
"brctg %%r0,0b \n\t" |
|
|
|
|
|
|
|
"1: \n\t" |
|
|
|
"lghi %%r0,28 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 3f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,2 \n\t" |
|
|
|
"2: \n\t" |
|
|
|
"vl %%v16,0(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
"vl %%v25,0(%%r1,%2) \n\t" |
|
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,16 \n\t" |
|
|
|
"brctg %%r0,2b \n\t" |
|
|
|
|
|
|
|
"3: \n\t" |
|
|
|
"veslg %%v2,%%v0,32 \n\t" |
|
|
|
"vfasb %%v0,%%v0,%%v2 \n\t" |
|
|
|
"vrepg %%v2,%%v0,1 \n\t" |
|
|
|
"aebr %%f0,%%f2 \n\t" |
|
|
|
"ste %%f0,0(%4) \n\t" |
|
|
|
"veslg %%v2,%%v1,32 \n\t" |
|
|
|
"vfasb %%v1,%%v1,%%v2 \n\t" |
|
|
|
"vrepg %%v2,%%v1,1 \n\t" |
|
|
|
"aebr %%f1,%%f2 \n\t" |
|
|
|
"ste %%f1,4(%4) " |
|
|
|
: |
|
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])ap[0]),"ZR"((const FLOAT (*)[n])ap[1]),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[2])y) |
|
|
|
:"memory","cc","r0","r1","v0","v1","v2","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31" |
|
|
|
); |
|
|
|
} |
|
|
|
|
|
|
|
static void sgemv_kernel_4x1(BLASLONG n, FLOAT *a0, FLOAT *x, FLOAT *y) |
|
|
|
{ |
|
|
|
__asm__ volatile ( |
|
|
|
"vzero %%v0 \n\t" |
|
|
|
"xgr %%r1,%%r1 \n\t" |
|
|
|
|
|
|
|
"lghi %%r0,-32 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 1f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,5 \n\t" |
|
|
|
"0: \n\t" |
|
|
|
"pfd 1,1024(%%r1,%1) \n\t" |
|
|
|
"pfd 1,1024(%%r1,%2) \n\t" |
|
|
|
|
|
|
|
"vl %%v16,0(%%r1,%2) \n\t" |
|
|
|
"vl %%v17,16(%%r1,%2) \n\t" |
|
|
|
"vl %%v18,32(%%r1,%2) \n\t" |
|
|
|
"vl %%v19,48(%%r1,%2) \n\t" |
|
|
|
"vl %%v20,64(%%r1,%2) \n\t" |
|
|
|
"vl %%v21,80(%%r1,%2) \n\t" |
|
|
|
"vl %%v22,96(%%r1,%2) \n\t" |
|
|
|
"vl %%v23,112(%%r1,%2) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v25,16(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v17,%%v25,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v26,32(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v18,%%v26,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v27,48(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v19,%%v27,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v28,64(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v20,%%v28,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v29,80(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v21,%%v29,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v30,96(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v22,%%v30,%%v0 \n\t" |
|
|
|
|
|
|
|
"vl %%v31,112(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v23,%%v31,%%v0 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,128 \n\t" |
|
|
|
"brctg %%r0,0b \n\t" |
|
|
|
|
|
|
|
"1: \n\t" |
|
|
|
"lghi %%r0,28 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 3f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,2 \n\t" |
|
|
|
"2: \n\t" |
|
|
|
"vl %%v16,0(%%r1,%2) \n\t" |
|
|
|
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t" |
|
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,16 \n\t" |
|
|
|
"brctg %%r0,2b \n\t" |
|
|
|
|
|
|
|
"3: \n\t" |
|
|
|
"veslg %%v1,%%v0,32 \n\t" |
|
|
|
"vfasb %%v0,%%v0,%%v1 \n\t" |
|
|
|
"vrepg %%v1,%%v0,1 \n\t" |
|
|
|
"aebr %%f0,%%f1 \n\t" |
|
|
|
"ste %%f0,0(%3) " |
|
|
|
: |
|
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])a0),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[1])y) |
|
|
|
:"memory","cc","r0","r1","v0","v1","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31" |
|
|
|
); |
|
|
|
} |
|
|
|
|
|
|
|
static void copy_x(BLASLONG n, FLOAT *src, FLOAT *dest, BLASLONG inc_src) |
|
|
|
{ |
|
|
|
BLASLONG i; |
|
|
|
for (i = 0; i < n; i++) |
|
|
|
{ |
|
|
|
dest[i] = *src; |
|
|
|
src += inc_src; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
static void add_y_kernel_4(BLASLONG n, FLOAT da, FLOAT *src, FLOAT *dest) |
|
|
|
{ |
|
|
|
__asm__ volatile ( |
|
|
|
"vlrepf %%v0,%1 \n\t" |
|
|
|
"xgr %%r1,%%r1 \n\t" |
|
|
|
|
|
|
|
"lghi %%r0,-32 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 1f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,5 \n\t" |
|
|
|
"0: \n\t" |
|
|
|
"pfd 1,1024(%%r1,%2) \n\t" |
|
|
|
"pfd 2,1024(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"vl %%v16,0(%%r1,%2) \n\t" |
|
|
|
"vl %%v17,16(%%r1,%2) \n\t" |
|
|
|
"vl %%v18,32(%%r1,%2) \n\t" |
|
|
|
"vl %%v19,48(%%r1,%2) \n\t" |
|
|
|
"vl %%v20,64(%%r1,%2) \n\t" |
|
|
|
"vl %%v21,80(%%r1,%2) \n\t" |
|
|
|
"vl %%v22,96(%%r1,%2) \n\t" |
|
|
|
"vl %%v23,112(%%r1,%2) \n\t" |
|
|
|
|
|
|
|
"vl %%v24, 0(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v24,%%v16,%%v0,%%v24 \n\t" |
|
|
|
"vst %%v24, 0(%%r1,%3) \n\t" |
|
|
|
"vl %%v25, 16(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v25,%%v17,%%v0,%%v25 \n\t" |
|
|
|
"vst %%v25, 16(%%r1,%3) \n\t" |
|
|
|
"vl %%v26, 32(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v26,%%v18,%%v0,%%v26 \n\t" |
|
|
|
"vst %%v26, 32(%%r1,%3) \n\t" |
|
|
|
"vl %%v27, 48(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v27,%%v19,%%v0,%%v27 \n\t" |
|
|
|
"vst %%v27, 48(%%r1,%3) \n\t" |
|
|
|
"vl %%v28, 64(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v28,%%v20,%%v0,%%v28 \n\t" |
|
|
|
"vst %%v28, 64(%%r1,%3) \n\t" |
|
|
|
"vl %%v29, 80(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v29,%%v21,%%v0,%%v29 \n\t" |
|
|
|
"vst %%v29, 80(%%r1,%3) \n\t" |
|
|
|
"vl %%v30, 96(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v30,%%v22,%%v0,%%v30 \n\t" |
|
|
|
"vst %%v30, 96(%%r1,%3) \n\t" |
|
|
|
"vl %%v31, 112(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v31,%%v23,%%v0,%%v31 \n\t" |
|
|
|
"vst %%v31, 112(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,128 \n\t" |
|
|
|
"brctg %%r0,0b \n\t" |
|
|
|
|
|
|
|
"1: \n\t" |
|
|
|
"lghi %%r0,28 \n\t" |
|
|
|
"ngr %%r0,%0 \n\t" |
|
|
|
"ltgr %%r0,%%r0 \n\t" |
|
|
|
"jz 3f \n\t" |
|
|
|
|
|
|
|
"srlg %%r0,%%r0,2 \n\t" |
|
|
|
"2: \n\t" |
|
|
|
"vl %%v16,0(%%r1,%2) \n\t" |
|
|
|
|
|
|
|
"vl %%v24, 0(%%r1,%3) \n\t" |
|
|
|
"vfmasb %%v24,%%v16,%%v0,%%v24 \n\t" |
|
|
|
"vst %%v24, 0(%%r1,%3) \n\t" |
|
|
|
|
|
|
|
"agfi %%r1,16 \n\t" |
|
|
|
"brctg %%r0,2b \n\t" |
|
|
|
|
|
|
|
"3: \n\t" |
|
|
|
"nop " |
|
|
|
: |
|
|
|
:"r"(n),"m"(da),"ZR"((const FLOAT (*)[n])src),"ZR"((FLOAT (*)[n])dest) |
|
|
|
:"memory","cc","r0","r1","v0","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31" |
|
|
|
); |
|
|
|
} |
|
|
|
static void add_y(BLASLONG n, FLOAT da, FLOAT *src, FLOAT *dest, BLASLONG inc_dest) |
|
|
|
{ |
|
|
|
if (inc_dest == 1) |
|
|
|
add_y_kernel_4(n, da, src, dest); |
|
|
|
else |
|
|
|
{ |
|
|
|
BLASLONG i; |
|
|
|
for (i = 0; i < n; i++) |
|
|
|
{ |
|
|
|
*dest += src[i] * da; |
|
|
|
dest += inc_dest; |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
int CNAME(BLASLONG m, BLASLONG n, BLASLONG dummy1, FLOAT alpha, FLOAT *a, BLASLONG lda, FLOAT *x, BLASLONG inc_x, FLOAT *y, BLASLONG inc_y, FLOAT *buffer) |
|
|
|
{ |
|
|
|
BLASLONG register i; |
|
|
|
BLASLONG register j; |
|
|
|
FLOAT *a_ptr; |
|
|
|
FLOAT *x_ptr; |
|
|
|
FLOAT *y_ptr; |
|
|
|
BLASLONG n0; |
|
|
|
BLASLONG n1; |
|
|
|
BLASLONG m1; |
|
|
|
BLASLONG m2; |
|
|
|
BLASLONG m3; |
|
|
|
BLASLONG n2; |
|
|
|
FLOAT ybuffer[2] __attribute__ ((aligned(16))); |
|
|
|
FLOAT *xbuffer; |
|
|
|
FLOAT *ytemp; |
|
|
|
|
|
|
|
if ( m < 1 ) return(0); |
|
|
|
if ( n < 1 ) return(0); |
|
|
|
|
|
|
|
xbuffer = buffer; |
|
|
|
ytemp = buffer + (m < NBMAX ? m : NBMAX); |
|
|
|
|
|
|
|
n0 = n / NBMAX; |
|
|
|
n1 = (n % NBMAX) >> 2 ; |
|
|
|
n2 = n & 3 ; |
|
|
|
|
|
|
|
m3 = m & 3 ; |
|
|
|
m1 = m & -4 ; |
|
|
|
m2 = (m & (NBMAX-1)) - m3 ; |
|
|
|
|
|
|
|
|
|
|
|
BLASLONG NB = NBMAX; |
|
|
|
|
|
|
|
while ( NB == NBMAX ) |
|
|
|
{ |
|
|
|
m1 -= NB; |
|
|
|
if ( m1 < 0) |
|
|
|
{ |
|
|
|
if ( m2 == 0 ) break; |
|
|
|
NB = m2; |
|
|
|
} |
|
|
|
|
|
|
|
y_ptr = y; |
|
|
|
a_ptr = a; |
|
|
|
x_ptr = x; |
|
|
|
|
|
|
|
if ( inc_x == 1 ) |
|
|
|
xbuffer = x_ptr; |
|
|
|
else |
|
|
|
copy_x(NB,x_ptr,xbuffer,inc_x); |
|
|
|
|
|
|
|
|
|
|
|
FLOAT *ap[4]; |
|
|
|
FLOAT *yp; |
|
|
|
BLASLONG register lda4 = 4 * lda; |
|
|
|
ap[0] = a_ptr; |
|
|
|
ap[1] = a_ptr + lda; |
|
|
|
ap[2] = ap[1] + lda; |
|
|
|
ap[3] = ap[2] + lda; |
|
|
|
|
|
|
|
if ( n0 > 0 ) |
|
|
|
{ |
|
|
|
BLASLONG nb1 = NBMAX / 4; |
|
|
|
for( j=0; j<n0; j++) |
|
|
|
{ |
|
|
|
|
|
|
|
yp = ytemp; |
|
|
|
for( i = 0; i < nb1 ; i++) |
|
|
|
{ |
|
|
|
sgemv_kernel_4x4(NB,ap,xbuffer,yp); |
|
|
|
ap[0] += lda4 ; |
|
|
|
ap[1] += lda4 ; |
|
|
|
ap[2] += lda4 ; |
|
|
|
ap[3] += lda4 ; |
|
|
|
yp += 4; |
|
|
|
} |
|
|
|
add_y(nb1*4, alpha, ytemp, y_ptr, inc_y ); |
|
|
|
y_ptr += nb1 * inc_y * 4; |
|
|
|
a_ptr += nb1 * lda4 ; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
yp = ytemp; |
|
|
|
|
|
|
|
for( i = 0; i < n1 ; i++) |
|
|
|
{ |
|
|
|
sgemv_kernel_4x4(NB,ap,xbuffer,yp); |
|
|
|
ap[0] += lda4 ; |
|
|
|
ap[1] += lda4 ; |
|
|
|
ap[2] += lda4 ; |
|
|
|
ap[3] += lda4 ; |
|
|
|
yp += 4; |
|
|
|
} |
|
|
|
if ( n1 > 0 ) |
|
|
|
{ |
|
|
|
add_y(n1*4, alpha, ytemp, y_ptr, inc_y ); |
|
|
|
y_ptr += n1 * inc_y * 4; |
|
|
|
a_ptr += n1 * lda4 ; |
|
|
|
} |
|
|
|
|
|
|
|
if ( n2 & 2 ) |
|
|
|
{ |
|
|
|
|
|
|
|
sgemv_kernel_4x2(NB,ap,xbuffer,ybuffer); |
|
|
|
a_ptr += lda * 2; |
|
|
|
*y_ptr += ybuffer[0] * alpha; |
|
|
|
y_ptr += inc_y; |
|
|
|
*y_ptr += ybuffer[1] * alpha; |
|
|
|
y_ptr += inc_y; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
if ( n2 & 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
sgemv_kernel_4x1(NB,a_ptr,xbuffer,ybuffer); |
|
|
|
// a_ptr += lda; |
|
|
|
*y_ptr += ybuffer[0] * alpha; |
|
|
|
// y_ptr += inc_y; |
|
|
|
|
|
|
|
} |
|
|
|
a += NB; |
|
|
|
x += NB * inc_x; |
|
|
|
} |
|
|
|
|
|
|
|
if ( m3 == 0 ) return(0); |
|
|
|
|
|
|
|
x_ptr = x; |
|
|
|
a_ptr = a; |
|
|
|
if ( m3 == 3 ) |
|
|
|
{ |
|
|
|
FLOAT xtemp0 = *x_ptr * alpha; |
|
|
|
x_ptr += inc_x; |
|
|
|
FLOAT xtemp1 = *x_ptr * alpha; |
|
|
|
x_ptr += inc_x; |
|
|
|
FLOAT xtemp2 = *x_ptr * alpha; |
|
|
|
|
|
|
|
FLOAT *aj = a_ptr; |
|
|
|
y_ptr = y; |
|
|
|
|
|
|
|
if ( lda == 3 && inc_y == 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
for ( j=0; j< ( n & -4) ; j+=4 ) |
|
|
|
{ |
|
|
|
|
|
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 + aj[2] * xtemp2; |
|
|
|
y_ptr[j+1] += aj[3] * xtemp0 + aj[4] * xtemp1 + aj[5] * xtemp2; |
|
|
|
y_ptr[j+2] += aj[6] * xtemp0 + aj[7] * xtemp1 + aj[8] * xtemp2; |
|
|
|
y_ptr[j+3] += aj[9] * xtemp0 + aj[10] * xtemp1 + aj[11] * xtemp2; |
|
|
|
aj += 12; |
|
|
|
} |
|
|
|
|
|
|
|
for ( ; j<n; j++ ) |
|
|
|
{ |
|
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 + aj[2] * xtemp2; |
|
|
|
aj += 3; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
|
|
|
|
if ( inc_y == 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
BLASLONG register lda2 = lda << 1; |
|
|
|
BLASLONG register lda4 = lda << 2; |
|
|
|
BLASLONG register lda3 = lda2 + lda; |
|
|
|
|
|
|
|
for ( j=0; j< ( n & -4 ); j+=4 ) |
|
|
|
{ |
|
|
|
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2; |
|
|
|
y_ptr[j+1] += *(aj+lda) * xtemp0 + *(aj+lda+1) * xtemp1 + *(aj+lda+2) * xtemp2; |
|
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp0 + *(aj+lda2+1) * xtemp1 + *(aj+lda2+2) * xtemp2; |
|
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp0 + *(aj+lda3+1) * xtemp1 + *(aj+lda3+2) * xtemp2; |
|
|
|
aj += lda4; |
|
|
|
} |
|
|
|
|
|
|
|
for ( ; j< n ; j++ ) |
|
|
|
{ |
|
|
|
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2 ; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
|
|
|
|
for ( j=0; j<n; j++ ) |
|
|
|
{ |
|
|
|
*y_ptr += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2; |
|
|
|
y_ptr += inc_y; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
return(0); |
|
|
|
} |
|
|
|
|
|
|
|
if ( m3 == 2 ) |
|
|
|
{ |
|
|
|
FLOAT xtemp0 = *x_ptr * alpha; |
|
|
|
x_ptr += inc_x; |
|
|
|
FLOAT xtemp1 = *x_ptr * alpha; |
|
|
|
|
|
|
|
FLOAT *aj = a_ptr; |
|
|
|
y_ptr = y; |
|
|
|
|
|
|
|
if ( lda == 2 && inc_y == 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
for ( j=0; j< ( n & -4) ; j+=4 ) |
|
|
|
{ |
|
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 ; |
|
|
|
y_ptr[j+1] += aj[2] * xtemp0 + aj[3] * xtemp1 ; |
|
|
|
y_ptr[j+2] += aj[4] * xtemp0 + aj[5] * xtemp1 ; |
|
|
|
y_ptr[j+3] += aj[6] * xtemp0 + aj[7] * xtemp1 ; |
|
|
|
aj += 8; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
for ( ; j<n; j++ ) |
|
|
|
{ |
|
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 ; |
|
|
|
aj += 2; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
if ( inc_y == 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
BLASLONG register lda2 = lda << 1; |
|
|
|
BLASLONG register lda4 = lda << 2; |
|
|
|
BLASLONG register lda3 = lda2 + lda; |
|
|
|
|
|
|
|
for ( j=0; j< ( n & -4 ); j+=4 ) |
|
|
|
{ |
|
|
|
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 ; |
|
|
|
y_ptr[j+1] += *(aj+lda) * xtemp0 + *(aj+lda+1) * xtemp1 ; |
|
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp0 + *(aj+lda2+1) * xtemp1 ; |
|
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp0 + *(aj+lda3+1) * xtemp1 ; |
|
|
|
aj += lda4; |
|
|
|
} |
|
|
|
|
|
|
|
for ( ; j< n ; j++ ) |
|
|
|
{ |
|
|
|
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 ; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
for ( j=0; j<n; j++ ) |
|
|
|
{ |
|
|
|
*y_ptr += *aj * xtemp0 + *(aj+1) * xtemp1 ; |
|
|
|
y_ptr += inc_y; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
return(0); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
FLOAT xtemp = *x_ptr * alpha; |
|
|
|
FLOAT *aj = a_ptr; |
|
|
|
y_ptr = y; |
|
|
|
if ( lda == 1 && inc_y == 1 ) |
|
|
|
{ |
|
|
|
for ( j=0; j< ( n & -4) ; j+=4 ) |
|
|
|
{ |
|
|
|
y_ptr[j] += aj[j] * xtemp; |
|
|
|
y_ptr[j+1] += aj[j+1] * xtemp; |
|
|
|
y_ptr[j+2] += aj[j+2] * xtemp; |
|
|
|
y_ptr[j+3] += aj[j+3] * xtemp; |
|
|
|
} |
|
|
|
for ( ; j<n ; j++ ) |
|
|
|
{ |
|
|
|
y_ptr[j] += aj[j] * xtemp; |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
if ( inc_y == 1 ) |
|
|
|
{ |
|
|
|
|
|
|
|
BLASLONG register lda2 = lda << 1; |
|
|
|
BLASLONG register lda4 = lda << 2; |
|
|
|
BLASLONG register lda3 = lda2 + lda; |
|
|
|
for ( j=0; j< ( n & -4 ); j+=4 ) |
|
|
|
{ |
|
|
|
y_ptr[j] += *aj * xtemp; |
|
|
|
y_ptr[j+1] += *(aj+lda) * xtemp; |
|
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp; |
|
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp; |
|
|
|
aj += lda4 ; |
|
|
|
} |
|
|
|
|
|
|
|
for ( ; j<n; j++ ) |
|
|
|
{ |
|
|
|
y_ptr[j] += *aj * xtemp; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
else |
|
|
|
{ |
|
|
|
for ( j=0; j<n; j++ ) |
|
|
|
{ |
|
|
|
*y_ptr += *aj * xtemp; |
|
|
|
y_ptr += inc_y; |
|
|
|
aj += lda; |
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
return(0); |
|
|
|
} |