| @@ -56,25 +56,32 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| #define VI3 $xr8 | |||
| #define VI4 $xr19 | |||
| #define VT0 $xr23 | |||
| #define VZE $xr3 | |||
| #define VT1 $xr4 | |||
| #define VT2 $xr5 | |||
| #define VC0 $xr6 | |||
| PROLOGUE | |||
| li.d i0, 0 | |||
| bge $r0, N, .L999 | |||
| bge $r0, INCX, .L999 | |||
| li.d TEMP, 1 | |||
| xvldi VZE, 0 | |||
| slli.d TEMP, TEMP, BASE_SHIFT | |||
| slli.d INCX, INCX, BASE_SHIFT | |||
| bne INCX, TEMP, .L20 | |||
| xvld VM0, X, 0 | |||
| #ifdef DOUBLE | |||
| xvfsub.d VT1, VZE, VM0 | |||
| addi.d i0, i0, 1 | |||
| srai.d I, N, 3 | |||
| bge $r0, I, .L21 | |||
| slli.d i0, i0, 2 //4 | |||
| xvfmaxa.d VM0, VM0, VT1 | |||
| bge $r0, I, .L11 | |||
| slli.d i0, i0, 1 //2 | |||
| xvreplgr2vr.d VINC4, i0 | |||
| slli.d i0, i0, 1 //8 | |||
| slli.d i0, i0, 1 //4 | |||
| xvreplgr2vr.d VINC8, i0 | |||
| addi.d i0, i0, -15 | |||
| addi.d i0, i0, -7 | |||
| xvinsgr2vr.d VI1, i0, 0 //initialize the index value for vectorization | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI1, i0, 1 | |||
| @@ -82,19 +89,23 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| xvinsgr2vr.d VI1, i0, 2 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI1, i0, 3 | |||
| addi.d i0, i0, 5 | |||
| xvinsgr2vr.d VI0, i0, 0 //1 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 1 //2 | |||
| xvinsgr2vr.d VI0, i0, 0 //initialize the index value for vectorization | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 2 //3 | |||
| xvinsgr2vr.d VI0, i0, 1 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 3 //4 | |||
| xvinsgr2vr.d VI0, i0, 2 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 3 | |||
| #else | |||
| xvfsub.s VT1, VZE, VM0 | |||
| addi.w i0, i0, 1 | |||
| srai.d I, N, 3 | |||
| xvfmaxa.s VM0, VM0, VT1 | |||
| bge $r0, I, .L21 | |||
| slli.w i0, i0, 3 //8 | |||
| slli.w i0, i0, 2 //4 | |||
| xvreplgr2vr.w VINC4, i0 | |||
| slli.w i0, i0, 1 //8 | |||
| xvreplgr2vr.w VINC8, i0 | |||
| addi.w i0, i0, -15 | |||
| xvinsgr2vr.w VI1, i0, 0 //initialize the index value for vectorization | |||
| @@ -135,73 +146,124 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| #ifdef DOUBLE | |||
| xvld VX0, X, 0 * SIZE | |||
| xvadd.d VI1, VI1, VINC8 | |||
| xvld VX1, X, 4 * SIZE | |||
| xvld VX1, X, 2 * SIZE | |||
| xvadd.d VI2, VI1, VINC4 | |||
| xvfsub.d VT1, VZE, VX0 | |||
| xvfsub.d VT2, VZE, VX1 | |||
| xvfmaxa.d VX0, VX0, VT1 | |||
| xvfmaxa.d VX1, VX1, VT2 | |||
| xvfcmp.clt.d VT0, VX0, VX1 //abx(x0) < abs(x1) | |||
| xvbitsel.v x1, VX0, VX1, VT0 //abs(maxf) | |||
| xvbitsel.v x2, VI1, VI2, VT0 //i | |||
| xvld VX0, X, 4 * SIZE | |||
| xvadd.d VI1, VI2, VINC4 | |||
| xvld VX1, X, 6 * SIZE | |||
| xvadd.d VI2, VI1, VINC4 | |||
| xvfmaxa.d VM1, VX0, VX1 | |||
| xvfcmp.ceq.d VT0, VX0, VM1 | |||
| xvfsub.d VT1, VZE, VX0 | |||
| xvfsub.d VT2, VZE, VX1 | |||
| xvfmaxa.d VX0, VX0, VT1 | |||
| xvfmaxa.d VX1, VX1, VT2 | |||
| xvfcmp.clt.d VT0, VX0, VX1 | |||
| xvbitsel.v x3, VX0, VX1, VT0 //abs(maxf) | |||
| xvbitsel.v x4, VI1, VI2, VT0 //i | |||
| xvfcmp.clt.d VC0, x1, x3 | |||
| xvbitsel.v x1, x1, x3, VC0 //abs(maxf) | |||
| xvbitsel.v x2, x2, x4, VC0 //i | |||
| xvfcmp.clt.d VT0, VM0, x1 | |||
| addi.d I, I, -1 | |||
| xvbitsel.v VI2, VI2, VI1, VT0 | |||
| xvfmaxa.d VM1, VM0, VM1 | |||
| xvfcmp.ceq.d VT0, VM0, VM1 | |||
| addi.d X, X, 8 * SIZE | |||
| xvbitsel.v VM0, VM1, VM0, VT0 | |||
| xvbitsel.v VI0, VI2, VI0, VT0 | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, x2, VT0 | |||
| #else | |||
| xvld VX0, X, 0 * SIZE | |||
| addi.d I, I, -1 | |||
| xvadd.w VI1, VI1, VINC8 | |||
| xvfmaxa.s VM1, VX0, VM0 | |||
| xvfcmp.ceq.s VT0, VM0, VM1 | |||
| xvld VX1, X, 4 * SIZE | |||
| xvadd.w VI2, VI1, VINC4 | |||
| xvfsub.s VT1, VZE, VX0 | |||
| xvfsub.s VT2, VZE, VX1 | |||
| xvfmaxa.s VX0, VX0, VT1 | |||
| xvfmaxa.s VX1, VX1, VT2 | |||
| xvfcmp.clt.s VT0, VX0, VX1 | |||
| xvbitsel.v x1, VX0, VX1, VT0 //abs(maxf) | |||
| xvbitsel.v x2, VI1, VI2, VT0 //i | |||
| addi.d I, I, -1 | |||
| xvfcmp.clt.s VT0, VM0, x1 | |||
| addi.d X, X, 8 * SIZE | |||
| xvbitsel.v VM0, VM1, VM0, VT0 | |||
| xvbitsel.v VI0, VI1, VI0, VT0 | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, x2, VT0 | |||
| #endif | |||
| blt $r0, I, .L10 | |||
| .align 3 | |||
| .L15: | |||
| #ifdef DOUBLE | |||
| xvpickve.d VI1, VI0, 0 | |||
| xvpickve.d VI2, VI0, 1 | |||
| xvpickve.d VI3, VI0, 2 | |||
| xvpickve.d VI4, VI0, 3 | |||
| xvpickve.d x1, VM0, 0 | |||
| xvpickve.d x2, VM0, 1 | |||
| xvpickve.d x3, VM0, 2 | |||
| xvpickve.d x4, VM0, 3 | |||
| vreplvei.d $vr21, $vr20, 0 | |||
| vreplvei.d $vr22, $vr20, 1 | |||
| vreplvei.d $vr9, $vr15, 0 | |||
| vreplvei.d $vr10, $vr15, 1 | |||
| fcmp.ceq.d $fcc0, $f9, $f10 | |||
| bceqz $fcc0, .L16 | |||
| xvfcmp.clt.d VT0, VI1, VI2 | |||
| xvbitsel.v VI0, VI2, VI1, VT0 | |||
| b .L17 | |||
| #else | |||
| xvxor.v VX0, VX0, VX0 | |||
| xvor.v VX0, VI0, VX0 | |||
| xvxor.v VX1, VX1, VX1 | |||
| xvor.v VX1, VM0, VX1 | |||
| xvpickve.w VI1, VI0, 0 | |||
| xvpickve.w VI2, VI0, 1 | |||
| xvpickve.w VI3, VI0, 2 | |||
| xvpickve.w VI4, VI0, 3 | |||
| xvpickve.w x1, VM0, 0 | |||
| xvpickve.w x2, VM0, 1 | |||
| xvpickve.w x3, VM0, 2 | |||
| xvpickve.w x4, VM0, 3 | |||
| vreplvei.w $vr21, $vr20, 0 | |||
| vreplvei.w $vr22, $vr20, 1 | |||
| vreplvei.w $vr8, $vr20, 2 | |||
| vreplvei.w $vr19, $vr20, 3 | |||
| vreplvei.w $vr9, $vr15, 0 | |||
| vreplvei.w $vr10, $vr15, 1 | |||
| vreplvei.w $vr11, $vr15, 2 | |||
| vreplvei.w $vr12, $vr15, 3 | |||
| b .L26 | |||
| #endif | |||
| XVFMAXA VM1, x1, x2 | |||
| XVCMPEQ VT0, x1, VM1 | |||
| xvbitsel.v VINC4, VI2, VI1, VT0 | |||
| XVFMAXA VM0, x3, x4 | |||
| XVCMPEQ VT0, x3, VM0 | |||
| xvbitsel.v VINC8, VI4, VI3, VT0 | |||
| XVFMAXA VM0, VM0, VM1 | |||
| XVCMPEQ VT0, VM0, VM1 | |||
| xvbitsel.v VI0, VINC8, VINC4, VT0 | |||
| CMPEQ $fcc0, $f15, $f9 | |||
| bceqz $fcc0, .L26 | |||
| XVCMPLT VT0, VI1, VI0 | |||
| .align 3 | |||
| #ifdef DOUBLE | |||
| .L16: | |||
| xvfcmp.clt.d VT0, x1, x2 | |||
| xvbitsel.v VI0, VI1, VI2, VT0 | |||
| xvbitsel.v VM0, x1, x2, VT0 | |||
| .align 3 | |||
| .L17: | |||
| movfr2gr.d i0, $f20 | |||
| .align 3 | |||
| .L11: //INCX==1 and N<8 | |||
| andi I, N, 7 | |||
| bge $r0, I, .L14 | |||
| srai.d i1, N, 3 | |||
| slli.d i1, i1, 3 | |||
| addi.d i1, i1, 1 //current index | |||
| movgr2fr.d $f21, i1 | |||
| movgr2fr.d $f20, i0 | |||
| .align 3 | |||
| .L13: | |||
| fld.d $f9, X, 0 | |||
| fsub.d $f10, $f3, $f9 | |||
| xvfmaxa.d x1, x1, x2 | |||
| xvfcmp.clt.d VT0, VM0, x1 | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, VI1, VT0 | |||
| b .L26 | |||
| addi.d I, I, -1 | |||
| addi.d i1, i1, 1 | |||
| addi.d X, X, SIZE | |||
| movgr2fr.d $f21, i1 | |||
| blt $r0, I, .L13 | |||
| movfr2gr.d i0, $f20 | |||
| .align 3 | |||
| .L14: | |||
| move $r4, $r17 | |||
| jirl $r0, $r1, 0x0 | |||
| .align 3 | |||
| .L20: // INCX!=1 | |||
| move TEMP, X | |||
| #ifdef DOUBLE | |||
| addi.d i0, i0, 1 | |||
| ld.d t1, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| @@ -210,34 +272,103 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| bge $r0, I, .L21 | |||
| ld.d t2, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| ld.d t3, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| ld.d t4, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| xvinsgr2vr.d VM0, t2, 1 | |||
| xvinsgr2vr.d VM0, t3, 2 | |||
| xvinsgr2vr.d VM0, t4, 3 | |||
| slli.d i0, i0, 2 //4 | |||
| slli.d i0, i0, 1 //2 | |||
| xvfsub.d VT1, VZE, VM0 | |||
| xvreplgr2vr.d VINC4, i0 | |||
| slli.d i0, i0, 1 //8 | |||
| slli.d i0, i0, 1 //4 | |||
| xvreplgr2vr.d VINC8, i0 | |||
| addi.d i0, i0, -15 | |||
| addi.d i0, i0, -7 | |||
| xvfmaxa.d VM0, VM0, VT1 | |||
| xvinsgr2vr.d VI1, i0, 0 //initialize the index value for vectorization | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI1, i0, 1 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI1, i0, 2 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI1, i0, 3 | |||
| addi.d i0, i0, 5 | |||
| addi.d i0, i0, 3 | |||
| xvinsgr2vr.d VI0, i0, 0 //1 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 1 //2 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 2 //3 | |||
| addi.d i0, i0, 1 | |||
| xvinsgr2vr.d VI0, i0, 3 //4 | |||
| .align 3 | |||
| .L24: | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX0, t1, 0 | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX0, t2, 1 | |||
| xvadd.d VI1, VI1, VINC8 | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX1, t1, 0 | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX1, t2, 1 | |||
| xvadd.d VI2, VI1, VINC4 | |||
| xvfsub.d VT1, VZE, VX0 | |||
| xvfsub.d VT2, VZE, VX1 | |||
| xvfmaxa.d VX0, VX0, VT1 | |||
| xvfmaxa.d VX1, VX1, VT2 | |||
| xvfcmp.clt.d VT0, VX0, VX1 | |||
| xvbitsel.v x1, VX0, VX1, VT0 | |||
| xvbitsel.v x2, VI1, VI2, VT0 | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX0, t1, 0 | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX0, t2, 1 | |||
| xvadd.d VI1, VI2, VINC4 | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX1, t1, 0 | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX1, t2, 1 | |||
| xvadd.d VI2, VI1, VINC4 | |||
| xvfsub.d VT1, VZE, VX0 | |||
| xvfsub.d VT2, VZE, VX1 | |||
| xvfmaxa.d VX0, VX0, VT1 | |||
| xvfmaxa.d VX1, VX1, VT2 | |||
| xvfcmp.clt.d VT0, VX0, VX1 | |||
| xvbitsel.v x3, VX0, VX1, VT0 | |||
| xvbitsel.v x4, VI1, VI2, VT0 | |||
| xvfcmp.clt.d VC0, x1, x3 | |||
| xvbitsel.v x1, x1, x3, VC0 | |||
| xvbitsel.v x2, x2, x4, VC0 | |||
| xvfcmp.clt.d VT0, VM0, x1 | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, x2, VT0 | |||
| addi.d I, I, -1 | |||
| blt $r0, I, .L24 | |||
| .align 3 | |||
| .L25: | |||
| vreplvei.d $vr21, $vr20, 0 | |||
| vreplvei.d $vr22, $vr20, 1 | |||
| vreplvei.d $vr9, $vr15, 0 | |||
| vreplvei.d $vr10, $vr15, 1 | |||
| fcmp.ceq.d $fcc0, $f10, $f9 | |||
| bceqz $fcc0, .L26 | |||
| xvfcmp.clt.d VT0, VI1, VI2 | |||
| xvbitsel.v VI0, VI2, VI1, VT0 | |||
| b .L27 | |||
| .align 3 | |||
| .L26: | |||
| xvfcmp.clt.d VT0, x1, x2 | |||
| xvbitsel.v VI0, VI1, VI2, VT0 | |||
| xvbitsel.v VM0, x1, x2, VT0 | |||
| .align 3 | |||
| .L27: | |||
| movfr2gr.d i0, $f20 | |||
| .align 3 | |||
| #else | |||
| .L20: // INCX!=1 | |||
| move TEMP, X | |||
| addi.w i0, i0, 1 | |||
| ld.w t1, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| @@ -253,19 +384,9 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| xvinsgr2vr.w VM0, t2, 1 | |||
| xvinsgr2vr.w VM0, t3, 2 | |||
| xvinsgr2vr.w VM0, t4, 3 | |||
| ld.w t1, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| ld.w t2, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| ld.w t3, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| ld.w t4, TEMP, 0 * SIZE | |||
| add.d TEMP, TEMP, INCX | |||
| xvinsgr2vr.w VM0, t1, 4 | |||
| xvinsgr2vr.w VM0, t2, 5 | |||
| xvinsgr2vr.w VM0, t3, 6 | |||
| xvinsgr2vr.w VM0, t4, 7 | |||
| slli.w i0, i0, 3 //8 | |||
| slli.w i0, i0, 2 //4 | |||
| xvreplgr2vr.w VINC4, i0 | |||
| slli.w i0, i0, 1 //8 | |||
| xvreplgr2vr.w VINC8, i0 | |||
| addi.w i0, i0, -15 | |||
| xvinsgr2vr.w VI1, i0, 0 //initialize the index value for vectorization | |||
| @@ -275,15 +396,7 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| xvinsgr2vr.w VI1, i0, 2 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI1, i0, 3 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI1, i0, 4 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI1, i0, 5 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI1, i0, 6 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI1, i0, 7 | |||
| addi.w i0, i0, 1 | |||
| addi.w i0, i0, 5 | |||
| xvinsgr2vr.w VI0, i0, 0 //1 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 1 //2 | |||
| @@ -291,54 +404,9 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| xvinsgr2vr.w VI0, i0, 2 //3 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 3 //4 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 4 //5 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 5 //6 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 6 //7 | |||
| addi.w i0, i0, 1 | |||
| xvinsgr2vr.w VI0, i0, 7 //8 | |||
| #endif | |||
| .align 3 | |||
| .L24: | |||
| #ifdef DOUBLE | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t3, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t4, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX0, t1, 0 | |||
| xvinsgr2vr.d VX0, t2, 1 | |||
| xvinsgr2vr.d VX0, t3, 2 | |||
| xvinsgr2vr.d VX0, t4, 3 | |||
| xvadd.d VI1, VI1, VINC8 | |||
| ld.d t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t2, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t3, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.d t4, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.d VX1, t1, 0 | |||
| xvinsgr2vr.d VX1, t2, 1 | |||
| xvinsgr2vr.d VX1, t3, 2 | |||
| xvinsgr2vr.d VX1, t4, 3 | |||
| xvadd.d VI2, VI1, VINC4 | |||
| xvfmaxa.d VM1, VX0, VX1 | |||
| xvfcmp.ceq.d VT0, VX0, VM1 | |||
| addi.d I, I, -1 | |||
| xvbitsel.v VI2, VI2, VI1, VT0 | |||
| xvfmaxa.d VM1, VM0, VM1 | |||
| xvfcmp.ceq.d VT0, VM0, VM1 | |||
| xvbitsel.v VM0, VM1, VM0, VT0 | |||
| xvbitsel.v VI0, VI2, VI0, VT0 | |||
| #else | |||
| ld.w t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.w t2, X, 0 * SIZE | |||
| @@ -351,6 +419,7 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| xvinsgr2vr.w VX0, t2, 1 | |||
| xvinsgr2vr.w VX0, t3, 2 | |||
| xvinsgr2vr.w VX0, t4, 3 | |||
| xvadd.w VI1, VI1, VINC8 | |||
| ld.w t1, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| ld.w t2, X, 0 * SIZE | |||
| @@ -359,158 +428,80 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| add.d X, X, INCX | |||
| ld.w t4, X, 0 * SIZE | |||
| add.d X, X, INCX | |||
| xvinsgr2vr.w VX0, t1, 4 | |||
| xvinsgr2vr.w VX0, t2, 5 | |||
| xvinsgr2vr.w VX0, t3, 6 | |||
| xvinsgr2vr.w VX0, t4, 7 | |||
| xvadd.w VI1, VI1, VINC8 | |||
| xvfmaxa.s VM1, VX0, VM0 | |||
| xvfcmp.ceq.s VT0, VM1, VM0 | |||
| xvinsgr2vr.w VX1, t1, 0 | |||
| xvinsgr2vr.w VX1, t2, 1 | |||
| xvinsgr2vr.w VX1, t3, 2 | |||
| xvinsgr2vr.w VX1, t4, 3 | |||
| xvadd.w VI2, VI1, VINC4 | |||
| xvfsub.s VT1, VZE, VX0 | |||
| xvfsub.s VT2, VZE, VX1 | |||
| xvfmaxa.s VX0, VX0, VT1 | |||
| xvfmaxa.s VX1, VX1, VT2 | |||
| xvfcmp.clt.s VT0, VX0, VX1 | |||
| xvbitsel.v x1, VX0, VX1, VT0 | |||
| xvbitsel.v x2, VI1, VI2, VT0 //i | |||
| addi.d I, I, -1 | |||
| xvbitsel.v VM0, VM1, VM0, VT0 | |||
| xvbitsel.v VI0, VI1, VI0, VT0 | |||
| #endif | |||
| xvfcmp.clt.s VT0, VM0, x1 | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, x2, VT0 | |||
| blt $r0, I, .L24 | |||
| .align 3 | |||
| .L25: | |||
| #ifdef DOUBLE | |||
| xvpickve.d VI1, VI0, 0 | |||
| xvpickve.d VI2, VI0, 1 | |||
| xvpickve.d VI3, VI0, 2 | |||
| xvpickve.d VI4, VI0, 3 | |||
| xvpickve.d x1, VM0, 0 | |||
| xvpickve.d x2, VM0, 1 | |||
| xvpickve.d x3, VM0, 2 | |||
| xvpickve.d x4, VM0, 3 | |||
| xvfmaxa.d VM1, x1, x2 | |||
| xvfcmp.ceq.d VT0, x1, VM1 | |||
| xvbitsel.v VINC4, VI2, VI1, VT0 | |||
| xvfmaxa.d VM0, x4, x3 | |||
| xvfcmp.ceq.d VT0, x3, VM0 | |||
| xvbitsel.v VINC8, VI4, VI3, VT0 | |||
| xvfmaxa.d VM0, VM0, VM1 | |||
| xvfcmp.ceq.d VT0, VM0, VM1 | |||
| xvbitsel.v VI0, VINC8, VINC4, VT0 | |||
| #else | |||
| xvxor.v VX0, VX0, VX0 | |||
| xvor.v VX0, VI0, VX0 | |||
| xvxor.v VX1, VX1, VX1 | |||
| xvor.v VX1, VM0, VX1 | |||
| xvpickve.w VI1, VI0, 0 | |||
| xvpickve.w VI2, VI0, 1 | |||
| xvpickve.w VI3, VI0, 2 | |||
| xvpickve.w VI4, VI0, 3 | |||
| xvpickve.w x1, VM0, 0 | |||
| xvpickve.w x2, VM0, 1 | |||
| xvpickve.w x3, VM0, 2 | |||
| xvpickve.w x4, VM0, 3 | |||
| xvfmaxa.s VM1, x1, x2 | |||
| xvfcmp.ceq.s VT0, x1, VM1 | |||
| xvbitsel.v VINC4, VI2, VI1, VT0 | |||
| xvfmaxa.s VM0, x3, x4 | |||
| xvfcmp.ceq.s VT0, x3, VM0 | |||
| xvbitsel.v VINC8, VI3, VI4, VT0 | |||
| xvfmaxa.s VM0, VM0, VM1 | |||
| xvfcmp.ceq.s VT0, VM0, VM1 | |||
| xvbitsel.v VM0, VM0, VM1, VT0 | |||
| xvbitsel.v VI0, VINC8, VINC4, VT0 | |||
| #endif | |||
| CMPEQ $fcc0, $f15, $f9 | |||
| bceqz $fcc0, .L26 | |||
| XVCMPLT VT0, VI1, VI0 | |||
| xvbitsel.v VI0, VI0, VI1, VT0 | |||
| vreplvei.w $vr21, $vr20, 0 | |||
| vreplvei.w $vr22, $vr20, 1 | |||
| vreplvei.w $vr8, $vr20, 2 | |||
| vreplvei.w $vr19, $vr20, 3 | |||
| vreplvei.w $vr9, $vr15, 0 | |||
| vreplvei.w $vr10, $vr15, 1 | |||
| vreplvei.w $vr11, $vr15, 2 | |||
| vreplvei.w $vr12, $vr15, 3 | |||
| .align 3 | |||
| .L26: | |||
| fcmp.ceq.d $fcc0, $f15, $f10 | |||
| bceqz $fcc0, .L27 | |||
| XVCMPLT VT0, VI2, VI0 | |||
| xvbitsel.v VI0, VI0, VI2, VT0 | |||
| fcmp.ceq.s $fcc0, $f9, $f10 | |||
| bceqz $fcc0, .L31 | |||
| xvfcmp.clt.s VT0, VI1, VI2 | |||
| xvbitsel.v VI1, VI2, VI1, VT0 | |||
| b .L32 | |||
| .align 3 | |||
| .L27: | |||
| fcmp.ceq.d $fcc0, $f15, $f11 | |||
| bceqz $fcc0, .L28 | |||
| XVCMPLT VT0, VI3, VI0 | |||
| xvbitsel.v VI0, VI0, VI3, VT0 | |||
| .L31: | |||
| xvfcmp.clt.s VT0, x1, x2 | |||
| xvbitsel.v VI1, VI1, VI2, VT0 | |||
| xvbitsel.v x1, x1, x2, VT0 | |||
| .align 3 | |||
| .L28: | |||
| fcmp.ceq.d $fcc0, $f15, $f12 | |||
| bceqz $fcc0, .L29 | |||
| XVCMPLT VT0, VI4, VI0 | |||
| xvbitsel.v VI0, VI0, VI4, VT0 | |||
| .L32: | |||
| fcmp.ceq.s $fcc0, $f11, $f12 | |||
| bceqz $fcc0, .L33 | |||
| xvfcmp.clt.s VT1, VI3, VI4 | |||
| xvbitsel.v VI3, VI4, VI3, VT1 | |||
| b .L34 | |||
| .align 3 | |||
| .L29: | |||
| #ifdef DOUBLE | |||
| movfr2gr.d i0, $f20 | |||
| #else | |||
| fmov.s $f16, $f20 | |||
| #endif | |||
| .L33: | |||
| xvfcmp.clt.s VT1, x3, x4 | |||
| xvbitsel.v x3, x3, x4, VT1 | |||
| xvbitsel.v VI3, VI3, VI4, VT1 | |||
| .align 3 | |||
| #ifdef DOUBLE | |||
| #else | |||
| .L252: | |||
| xvxor.v VI0, VI0, VI0 | |||
| xvor.v VI0, VI0, VX0 | |||
| fmov.s $f13, $f15 | |||
| xvxor.v VM0, VM0, VM0 | |||
| xvor.v VM0, VM0, VX1 | |||
| xvpickve.w VI1, VI0, 4 | |||
| xvpickve.w VI2, VI0, 5 | |||
| xvpickve.w VI3, VI0, 6 | |||
| xvpickve.w VI4, VI0, 7 | |||
| xvpickve.w x1, VM0, 4 | |||
| xvpickve.w x2, VM0, 5 | |||
| xvpickve.w x3, VM0, 6 | |||
| xvpickve.w x4, VM0, 7 | |||
| xvfmaxa.s VM1, x1, x2 | |||
| xvfcmp.ceq.s VT0, x1, VM1 | |||
| xvbitsel.v VINC4, VI2, VI1, VT0 | |||
| xvfmaxa.s VM0, x3, x4 | |||
| xvfcmp.ceq.s VT0, x3, VM0 | |||
| xvbitsel.v VINC8, VI4, VI3, VT0 | |||
| xvfmaxa.s VM0, VM0, VM1 | |||
| xvfcmp.ceq.s VT0, VM0, VM1 | |||
| xvbitsel.v VI0, VINC8, VINC4, VT0 | |||
| fcmp.ceq.d $fcc0, $f15, $f9 | |||
| bceqz $fcc0, .L262 | |||
| xvfcmp.clt.s VT0, VI1, VI0 | |||
| xvbitsel.v VI0, VI0, VI1, VT0 | |||
| .L34: | |||
| fcmp.ceq.s $fcc0, $f9, $f11 | |||
| bceqz $fcc0, .L35 | |||
| xvfcmp.clt.s VT0, VI1, VI3 | |||
| xvbitsel.v VI0, VI3, VI1, VT0 | |||
| xvxor.v VM0, x1, VZE | |||
| b .L29 | |||
| .align 3 | |||
| .L262: | |||
| fcmp.ceq.d $fcc0, $f15, $f10 | |||
| bceqz $fcc0, .L272 | |||
| xvfcmp.clt.s VT0, VI2, VI0 | |||
| xvbitsel.v VI0, VI0, VI2, VT0 | |||
| .L35: | |||
| xvfcmp.clt.s VT0, x1, x3 | |||
| xvbitsel.v VM0, x1, x3, VT0 | |||
| xvbitsel.v VI0, VI1, VI3, VT0 | |||
| .align 3 | |||
| .L272: | |||
| fcmp.ceq.d $fcc0, $f15, $f11 | |||
| bceqz $fcc0, .L282 | |||
| xvfcmp.clt.s VT0, VI3, VI0 | |||
| xvbitsel.v VI0, VI0, VI3, VT0 | |||
| .align 3 | |||
| .L282: | |||
| fcmp.ceq.d $fcc0, $f15, $f12 | |||
| bceqz $fcc0, .L292 | |||
| xvfcmp.clt.s VT0, VI4, VI0 | |||
| xvbitsel.v VI0, VI0, VI4, VT0 | |||
| .L29: | |||
| movfr2gr.s i0, $f20 | |||
| .align 3 | |||
| .L292: | |||
| xvfmaxa.s VM0, VX0, VM0 | |||
| xvfcmp.ceq.s VT0, VM0, VX0 | |||
| xvbitsel.v VI0, VI0, VI1, VT0 | |||
| movfr2gr.s i0, $f20 | |||
| #endif | |||
| .L21: //N<8 | |||
| .L21: // N<8 | |||
| andi I, N, 7 | |||
| bge $r0, I, .L999 | |||
| srai.d i1, N, 3 | |||
| @@ -521,17 +512,24 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||
| .align 3 | |||
| .L22: | |||
| LD $f9, X, 0 | |||
| LD $f9, X, 0 | |||
| #ifdef DOUBLE | |||
| fsub.d $f10, $f3, $f9 | |||
| xvfmaxa.d x1, x1, x2 | |||
| xvfcmp.clt.d VT0, VM0, x1 | |||
| #else | |||
| fsub.s $f10, $f3, $f9 | |||
| xvfmaxa.s x1, x1, x2 | |||
| xvfcmp.clt.s VT0, VM0, x1 | |||
| #endif | |||
| xvbitsel.v VM0, VM0, x1, VT0 | |||
| xvbitsel.v VI0, VI0, VI1, VT0 | |||
| addi.d I, I, -1 | |||
| XVFMAXA VM1, x1, VM0 | |||
| XVCMPEQ VT0, VM0, VM1 | |||
| add.d X, X, INCX | |||
| xvbitsel.v VM0, VM1, VM0, VT0 | |||
| xvbitsel.v VI0, VI1, VI0, VT0 | |||
| addi.d i1, i1, 1 | |||
| add.d X, X, INCX | |||
| movgr2fr.d $f21, i1 | |||
| blt $r0, I, .L22 | |||
| MTG i0, $f20 | |||
| MTG i0, $f20 | |||
| .align 3 | |||
| .L999: | |||