| @@ -187,6 +187,6 @@ ZTRMMLNCOPY_M = ztrmm_lncopy_sve_v1.c | |||
| ZTRMMUTCOPY_M = ztrmm_utcopy_sve_v1.c | |||
| ZTRMMLTCOPY_M = ztrmm_ltcopy_sve_v1.c | |||
| ZSYMMUCOPY_M = symm_ucopy_sve.c | |||
| ZSYMMLCOPY_M = symm_lcopy_sve.c | |||
| ZHEMMLTCOPY_M = zhemm_ltcopy_sve.c | |||
| ZHEMMUTCOPY_M = zhemm_utcopy_sve.c | |||
| @@ -0,0 +1,106 @@ | |||
| /*********************************************************************/ | |||
| /* Copyright 2009, 2010 The University of Texas at Austin. */ | |||
| /* All rights reserved. */ | |||
| /* */ | |||
| /* Redistribution and use in source and binary forms, with or */ | |||
| /* without modification, are permitted provided that the following */ | |||
| /* conditions are met: */ | |||
| /* */ | |||
| /* 1. Redistributions of source code must retain the above */ | |||
| /* copyright notice, this list of conditions and the following */ | |||
| /* disclaimer. */ | |||
| /* */ | |||
| /* 2. Redistributions in binary form must reproduce the above */ | |||
| /* copyright notice, this list of conditions and the following */ | |||
| /* disclaimer in the documentation and/or other materials */ | |||
| /* provided with the distribution. */ | |||
| /* */ | |||
| /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ | |||
| /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ | |||
| /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ | |||
| /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ | |||
| /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ | |||
| /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ | |||
| /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ | |||
| /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ | |||
| /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ | |||
| /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ | |||
| /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ | |||
| /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ | |||
| /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ | |||
| /* POSSIBILITY OF SUCH DAMAGE. */ | |||
| /* */ | |||
| /* The views and conclusions contained in the software and */ | |||
| /* documentation are those of the authors and should not be */ | |||
| /* interpreted as representing official policies, either expressed */ | |||
| /* or implied, of The University of Texas at Austin. */ | |||
| /*********************************************************************/ | |||
| #include <stdio.h> | |||
| #include "common.h" | |||
| #include <arm_sve.h> | |||
| int CNAME(BLASLONG m, BLASLONG n, FLOAT *a, BLASLONG lda, BLASLONG posX, BLASLONG posY, FLOAT *b){ | |||
| BLASLONG offset, i; | |||
| lda *= 2; | |||
| uint64_t sve_size = svcntd(); | |||
| svint64_t posY_vec = svdup_s64(posY); | |||
| svint64_t posX_vec = svdup_s64(posX); | |||
| svint64_t lda_vec = svdup_s64(lda); | |||
| svint64_t one_vec = svdup_s64(1LL); | |||
| int64_t j = 0; | |||
| svbool_t pg = svwhilelt_b64(j, n); | |||
| int64_t active = svcntp_b64(svptrue_b64(), pg); | |||
| svint64_t index_neg = svindex_s64(0LL, -1LL); | |||
| svint64_t index = svindex_s64(0LL, 1LL); | |||
| do { | |||
| offset = posX - posY; | |||
| svint64_t vec_off = svdup_s64(offset); | |||
| svbool_t cmp = svcmpgt(pg, vec_off, index_neg); | |||
| svint64_t temp = svadd_z(pg, posX_vec, index); | |||
| svint64_t temp1 = svmul_z(pg, temp, 2); | |||
| temp1 = svmla_z(pg, temp1, posY_vec, lda_vec); | |||
| svint64_t temp2 = svmul_z(pg, temp, lda_vec); | |||
| temp2 = svmla_z(pg, temp2, posY_vec, 2); | |||
| svint64_t gat_ind = svsel(cmp, temp1, temp2); | |||
| i = m; | |||
| while (i>0) { | |||
| svfloat64_t data_vec_real = svld1_gather_index(pg, a, gat_ind); | |||
| svfloat64_t data_vec_imag = svld1_gather_index(pg, a+1, gat_ind); | |||
| gat_ind = svadd_m(cmp, gat_ind, lda_vec); | |||
| gat_ind = svadd_m(svnot_z(pg, cmp) , gat_ind, 2); | |||
| if (offset <= 0) { | |||
| svbool_t off_g = svwhilelt_b64(offset, 0LL); | |||
| data_vec_imag = svneg_m(data_vec_imag, off_g, data_vec_imag); | |||
| } | |||
| svst2(pg, b, svcreate2(data_vec_real, data_vec_imag)); | |||
| // dealing with ZERO separately | |||
| if (offset > -active && offset < 1) | |||
| b[ -2*offset + 1 ] = ZERO; | |||
| b += active * 2; | |||
| offset --; | |||
| vec_off = svsub_z(pg, vec_off, one_vec); | |||
| cmp = svcmpgt(pg, vec_off, index_neg); | |||
| i--; | |||
| } | |||
| posX += sve_size; | |||
| posX_vec = svdup_s64(posX); | |||
| j += sve_size; | |||
| pg = svwhilelt_b64(j, n); | |||
| active = svcntp_b64(svptrue_b64(), pg); | |||
| } while (svptest_any(svptrue_b64(), pg)); | |||
| return 0; | |||
| } | |||
| @@ -0,0 +1,107 @@ | |||
| /*********************************************************************/ | |||
| /* Copyright 2009, 2010 The University of Texas at Austin. */ | |||
| /* All rights reserved. */ | |||
| /* */ | |||
| /* Redistribution and use in source and binary forms, with or */ | |||
| /* without modification, are permitted provided that the following */ | |||
| /* conditions are met: */ | |||
| /* */ | |||
| /* 1. Redistributions of source code must retain the above */ | |||
| /* copyright notice, this list of conditions and the following */ | |||
| /* disclaimer. */ | |||
| /* */ | |||
| /* 2. Redistributions in binary form must reproduce the above */ | |||
| /* copyright notice, this list of conditions and the following */ | |||
| /* disclaimer in the documentation and/or other materials */ | |||
| /* provided with the distribution. */ | |||
| /* */ | |||
| /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */ | |||
| /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */ | |||
| /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */ | |||
| /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */ | |||
| /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */ | |||
| /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */ | |||
| /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */ | |||
| /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */ | |||
| /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */ | |||
| /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */ | |||
| /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */ | |||
| /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */ | |||
| /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */ | |||
| /* POSSIBILITY OF SUCH DAMAGE. */ | |||
| /* */ | |||
| /* The views and conclusions contained in the software and */ | |||
| /* documentation are those of the authors and should not be */ | |||
| /* interpreted as representing official policies, either expressed */ | |||
| /* or implied, of The University of Texas at Austin. */ | |||
| /*********************************************************************/ | |||
| #include <stdio.h> | |||
| #include "common.h" | |||
| #include <arm_sve.h> | |||
| int CNAME(BLASLONG m, BLASLONG n, FLOAT *a, BLASLONG lda, BLASLONG posX, BLASLONG posY, FLOAT *b){ | |||
| BLASLONG offset, i; | |||
| lda *= 2; | |||
| uint64_t sve_size = svcntd(); | |||
| svint64_t posY_vec = svdup_s64(posY); | |||
| svint64_t posX_vec = svdup_s64(posX); | |||
| svint64_t lda_vec = svdup_s64(lda); | |||
| svint64_t one_vec = svdup_s64(1LL); | |||
| int64_t j = 0; | |||
| svbool_t pg = svwhilelt_b64(j, n); | |||
| int64_t active = svcntp_b64(svptrue_b64(), pg); | |||
| svint64_t index_neg = svindex_s64(0LL, -1LL); | |||
| svint64_t index = svindex_s64(0LL, 1LL); | |||
| do { | |||
| offset = posX - posY; | |||
| svint64_t vec_off = svdup_s64(offset); | |||
| svbool_t cmp = svcmpgt(pg, vec_off, index_neg); | |||
| svint64_t temp = svadd_z(pg, posX_vec, index); | |||
| svint64_t temp1 = svmul_z(pg, temp, lda); | |||
| temp1 = svmla_z(pg, temp1, posY_vec, 2); | |||
| svint64_t temp2 = svmul_z(pg, temp, 2); | |||
| temp2 = svmla_z(pg, temp2, posY_vec, lda); | |||
| svint64_t gat_ind = svsel(cmp, temp1, temp2); | |||
| i = m; | |||
| while (i>0) { | |||
| svfloat64_t data_vec_real = svld1_gather_index(pg, a, gat_ind); | |||
| svfloat64_t data_vec_imag = svld1_gather_index(pg, a+1, gat_ind); | |||
| gat_ind = svadd_m(cmp, gat_ind, 2); | |||
| gat_ind = svadd_m(svnot_z(pg, cmp) , gat_ind, lda_vec); | |||
| data_vec_imag = svneg_z(pg, data_vec_imag); | |||
| if (offset <= 0) { | |||
| svbool_t off_g = svwhilelt_b64(offset, 0LL); | |||
| data_vec_imag = svneg_m(data_vec_imag, off_g, data_vec_imag); | |||
| } | |||
| svst2(pg, b, svcreate2(data_vec_real, data_vec_imag)); | |||
| // dealing with ZERO separately | |||
| if (offset > -active && offset < 1) | |||
| b[ -2*offset + 1 ] = ZERO; | |||
| b += active * 2; | |||
| offset --; | |||
| vec_off = svsub_z(pg, vec_off, one_vec); | |||
| cmp = svcmpgt(pg, vec_off, index_neg); | |||
| i--; | |||
| } | |||
| posX += sve_size; | |||
| posX_vec = svdup_s64(posX); | |||
| j += sve_size; | |||
| pg = svwhilelt_b64(j, n); | |||
| active = svcntp_b64(svptrue_b64(), pg); | |||
| } while (svptest_any(svptrue_b64(), pg)); | |||
| return 0; | |||
| } | |||