|
|
|
@@ -25,6 +25,14 @@ OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE |
|
|
|
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
|
|
*****************************************************************************/ |
|
|
|
|
|
|
|
|
|
|
|
/* Ensure that the compiler knows how to generate AVX2 instructions if it doesn't already */ |
|
|
|
#ifndef __AVX512CD_ |
|
|
|
#pragma GCC target("avx2,fma") |
|
|
|
#endif |
|
|
|
|
|
|
|
#ifdef __AVX2__ |
|
|
|
|
|
|
|
#include <immintrin.h> |
|
|
|
|
|
|
|
#define HAVE_KERNEL_4x4 1 |
|
|
|
@@ -33,13 +41,14 @@ static void dsymv_kernel_4x4(BLASLONG from, BLASLONG to, FLOAT **a, FLOAT *x, FL |
|
|
|
{ |
|
|
|
|
|
|
|
|
|
|
|
__m256d temp2_0, temp2_1, temp2_2, temp2_3; // temp2_0 temp2_1 temp2_2 temp2_3 |
|
|
|
__m256d accum_0, accum_1, accum_2, accum_3; |
|
|
|
__m256d temp1_0, temp1_1, temp1_2, temp1_3; |
|
|
|
|
|
|
|
temp2_0 = _mm256_setzero_pd(); |
|
|
|
temp2_1 = _mm256_setzero_pd(); |
|
|
|
temp2_2 = _mm256_setzero_pd(); |
|
|
|
temp2_3 = _mm256_setzero_pd(); |
|
|
|
/* the 256 bit wide acculmulator vectors start out as zero */ |
|
|
|
accum_0 = _mm256_setzero_pd(); |
|
|
|
accum_1 = _mm256_setzero_pd(); |
|
|
|
accum_2 = _mm256_setzero_pd(); |
|
|
|
accum_3 = _mm256_setzero_pd(); |
|
|
|
|
|
|
|
temp1_0 = _mm256_broadcastsd_pd(_mm_load_sd(&temp1[0])); |
|
|
|
temp1_1 = _mm256_broadcastsd_pd(_mm_load_sd(&temp1[1])); |
|
|
|
@@ -47,15 +56,16 @@ static void dsymv_kernel_4x4(BLASLONG from, BLASLONG to, FLOAT **a, FLOAT *x, FL |
|
|
|
temp1_3 = _mm256_broadcastsd_pd(_mm_load_sd(&temp1[3])); |
|
|
|
|
|
|
|
#ifdef __AVX512CD__ |
|
|
|
__m512d temp2_05, temp2_15, temp2_25, temp2_35; // temp2_0 temp2_1 temp2_2 temp2_3 |
|
|
|
__m512d accum_05, accum_15, accum_25, accum_35; |
|
|
|
__m512d temp1_05, temp1_15, temp1_25, temp1_35; |
|
|
|
BLASLONG to2; |
|
|
|
int delta; |
|
|
|
|
|
|
|
temp2_05 = _mm512_setzero_pd(); |
|
|
|
temp2_15 = _mm512_setzero_pd(); |
|
|
|
temp2_25 = _mm512_setzero_pd(); |
|
|
|
temp2_35 = _mm512_setzero_pd(); |
|
|
|
/* the 512 bit wide accumulator vectors start out as zero */ |
|
|
|
accum_05 = _mm512_setzero_pd(); |
|
|
|
accum_15 = _mm512_setzero_pd(); |
|
|
|
accum_25 = _mm512_setzero_pd(); |
|
|
|
accum_35 = _mm512_setzero_pd(); |
|
|
|
|
|
|
|
temp1_05 = _mm512_broadcastsd_pd(_mm_load_sd(&temp1[0])); |
|
|
|
temp1_15 = _mm512_broadcastsd_pd(_mm_load_sd(&temp1[1])); |
|
|
|
@@ -80,19 +90,23 @@ static void dsymv_kernel_4x4(BLASLONG from, BLASLONG to, FLOAT **a, FLOAT *x, FL |
|
|
|
|
|
|
|
_y += temp1_05 * a0 + temp1_15 * a1 + temp1_25 * a2 + temp1_35 * a3; |
|
|
|
|
|
|
|
temp2_05 += _x * a0; |
|
|
|
temp2_15 += _x * a1; |
|
|
|
temp2_25 += _x * a2; |
|
|
|
temp2_35 += _x * a3; |
|
|
|
accum_05 += _x * a0; |
|
|
|
accum_15 += _x * a1; |
|
|
|
accum_25 += _x * a2; |
|
|
|
accum_35 += _x * a3; |
|
|
|
|
|
|
|
_mm512_storeu_pd(&y[from], _y); |
|
|
|
|
|
|
|
}; |
|
|
|
|
|
|
|
temp2_0 = _mm256_add_pd(_mm512_extractf64x4_pd(temp2_05, 0), _mm512_extractf64x4_pd(temp2_05, 1)); |
|
|
|
temp2_1 = _mm256_add_pd(_mm512_extractf64x4_pd(temp2_15, 0), _mm512_extractf64x4_pd(temp2_15, 1)); |
|
|
|
temp2_2 = _mm256_add_pd(_mm512_extractf64x4_pd(temp2_25, 0), _mm512_extractf64x4_pd(temp2_25, 1)); |
|
|
|
temp2_3 = _mm256_add_pd(_mm512_extractf64x4_pd(temp2_35, 0), _mm512_extractf64x4_pd(temp2_35, 1)); |
|
|
|
/* |
|
|
|
* we need to fold our 512 bit wide accumulator vectors into 256 bit wide vectors so that the AVX2 code |
|
|
|
* below can continue using the intermediate results in its loop |
|
|
|
*/ |
|
|
|
accum_0 = _mm256_add_pd(_mm512_extractf64x4_pd(accum_05, 0), _mm512_extractf64x4_pd(accum_05, 1)); |
|
|
|
accum_1 = _mm256_add_pd(_mm512_extractf64x4_pd(accum_15, 0), _mm512_extractf64x4_pd(accum_15, 1)); |
|
|
|
accum_2 = _mm256_add_pd(_mm512_extractf64x4_pd(accum_25, 0), _mm512_extractf64x4_pd(accum_25, 1)); |
|
|
|
accum_3 = _mm256_add_pd(_mm512_extractf64x4_pd(accum_35, 0), _mm512_extractf64x4_pd(accum_35, 1)); |
|
|
|
|
|
|
|
#endif |
|
|
|
|
|
|
|
@@ -103,6 +117,7 @@ static void dsymv_kernel_4x4(BLASLONG from, BLASLONG to, FLOAT **a, FLOAT *x, FL |
|
|
|
_y = _mm256_loadu_pd(&y[from]); |
|
|
|
_x = _mm256_loadu_pd(&x[from]); |
|
|
|
|
|
|
|
/* load 4 rows of matrix data */ |
|
|
|
a0 = _mm256_loadu_pd(&a[0][from]); |
|
|
|
a1 = _mm256_loadu_pd(&a[1][from]); |
|
|
|
a2 = _mm256_loadu_pd(&a[2][from]); |
|
|
|
@@ -110,33 +125,40 @@ static void dsymv_kernel_4x4(BLASLONG from, BLASLONG to, FLOAT **a, FLOAT *x, FL |
|
|
|
|
|
|
|
_y += temp1_0 * a0 + temp1_1 * a1 + temp1_2 * a2 + temp1_3 * a3; |
|
|
|
|
|
|
|
temp2_0 += _x * a0; |
|
|
|
temp2_1 += _x * a1; |
|
|
|
temp2_2 += _x * a2; |
|
|
|
temp2_3 += _x * a3; |
|
|
|
accum_0 += _x * a0; |
|
|
|
accum_1 += _x * a1; |
|
|
|
accum_2 += _x * a2; |
|
|
|
accum_3 += _x * a3; |
|
|
|
|
|
|
|
_mm256_storeu_pd(&y[from], _y); |
|
|
|
|
|
|
|
}; |
|
|
|
|
|
|
|
__m128d xmm0, xmm1, xmm2, xmm3; |
|
|
|
/* |
|
|
|
* we now have 4 accumulator vectors. Each vector needs to be summed up element wise and stored in the temp2 |
|
|
|
* output array. There is no direct instruction for this in 256 bit space, only in 128 space. |
|
|
|
*/ |
|
|
|
|
|
|
|
__m128d half_accum0, half_accum1, half_accum2, half_accum3; |
|
|
|
|
|
|
|
xmm0 = _mm_add_pd(_mm256_extractf128_pd(temp2_0, 0), _mm256_extractf128_pd(temp2_0, 1)); |
|
|
|
xmm1 = _mm_add_pd(_mm256_extractf128_pd(temp2_1, 0), _mm256_extractf128_pd(temp2_1, 1)); |
|
|
|
xmm2 = _mm_add_pd(_mm256_extractf128_pd(temp2_2, 0), _mm256_extractf128_pd(temp2_2, 1)); |
|
|
|
xmm3 = _mm_add_pd(_mm256_extractf128_pd(temp2_3, 0), _mm256_extractf128_pd(temp2_3, 1)); |
|
|
|
|
|
|
|
xmm0 = _mm_hadd_pd(xmm0, xmm0); |
|
|
|
xmm1 = _mm_hadd_pd(xmm1, xmm1); |
|
|
|
xmm2 = _mm_hadd_pd(xmm2, xmm2); |
|
|
|
xmm3 = _mm_hadd_pd(xmm3, xmm3); |
|
|
|
/* Add upper half to lower half of each of the four 256 bit vectors to get to four 128 bit vectors */ |
|
|
|
half_accum0 = _mm_add_pd(_mm256_extractf128_pd(accum_0, 0), _mm256_extractf128_pd(accum_0, 1)); |
|
|
|
half_accum1 = _mm_add_pd(_mm256_extractf128_pd(accum_1, 0), _mm256_extractf128_pd(accum_1, 1)); |
|
|
|
half_accum2 = _mm_add_pd(_mm256_extractf128_pd(accum_2, 0), _mm256_extractf128_pd(accum_2, 1)); |
|
|
|
half_accum3 = _mm_add_pd(_mm256_extractf128_pd(accum_3, 0), _mm256_extractf128_pd(accum_3, 1)); |
|
|
|
|
|
|
|
/* in 128 bit land there is a hadd operation to do the rest of the element-wise sum in one go */ |
|
|
|
half_accum0 = _mm_hadd_pd(half_accum0, half_accum0); |
|
|
|
half_accum1 = _mm_hadd_pd(half_accum1, half_accum1); |
|
|
|
half_accum2 = _mm_hadd_pd(half_accum2, half_accum2); |
|
|
|
half_accum3 = _mm_hadd_pd(half_accum3, half_accum3); |
|
|
|
|
|
|
|
temp2[0] += xmm0[0]; |
|
|
|
temp2[1] += xmm1[0]; |
|
|
|
temp2[2] += xmm2[0]; |
|
|
|
temp2[3] += xmm3[0]; |
|
|
|
/* and store the lowest double value from each of these vectors in the temp2 output */ |
|
|
|
temp2[0] += half_accum0[0]; |
|
|
|
temp2[1] += half_accum1[0]; |
|
|
|
temp2[2] += half_accum2[0]; |
|
|
|
temp2[3] += half_accum3[0]; |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
#endif |