Lines Matching refs:reg0
68 #define VP9_DOTP_CONST_PAIR(reg0, reg1, cnst0, cnst1, out0, out1) \
76 s1_m = __lsx_vilvl_h(__lsx_vneg_h(reg1), reg0); \
77 s0_m = __lsx_vilvh_h(__lsx_vneg_h(reg1), reg0); \
78 s3_m = __lsx_vilvl_h(reg0, reg1); \
79 s2_m = __lsx_vilvh_h(reg0, reg1); \
376 __m128i reg0, reg2, reg4, reg6, reg8, reg10, reg12, reg14;
383 reg0, reg1, reg2, reg3);
412 VP9_DOTP_CONST_PAIR(reg0, reg8, cospi_16_64, cospi_16_64, reg0, reg8);
414 LSX_BUTTERFLY_4_H(reg8, reg0, reg4, reg12, reg2, reg6, reg10, reg14);
416 reg0 = __lsx_vsub_h(reg2, loc1);
445 loc0 = __lsx_vadd_h(reg0, loc1);
446 loc1 = __lsx_vsub_h(reg0, loc1);
449 reg0 = loc2;
481 DUP4_ARG2(__lsx_vsrari_h, reg0, 6, reg2, 6, reg4, 6, reg6, 6,
482 reg0, reg2, reg4, reg6);
483 VP9_ADDBLK_ST8x4_UB(dst, dst_stride, reg0, reg2, reg4, reg6);
502 __m128i reg0, reg2, reg4, reg6, reg8, reg10, reg12, reg14;
508 reg0, reg1, reg2, reg3);
537 VP9_DOTP_CONST_PAIR(reg0, reg8, cospi_16_64, cospi_16_64, reg0, reg8);
539 LSX_BUTTERFLY_4_H(reg8, reg0, reg4, reg12, reg2, reg6, reg10, reg14);
541 reg0 = __lsx_vsub_h(reg2, loc1);
570 loc0 = __lsx_vadd_h(reg0, loc1);
571 loc1 = __lsx_vsub_h(reg0, loc1);
574 reg0 = loc2;
608 LSX_TRANSPOSE8x8_H(reg0, reg2, reg4, reg6, reg8, reg10, reg12, reg14,
609 reg0, reg2, reg4, reg6, reg8, reg10, reg12, reg14);
611 __lsx_vst(reg0, output, 32*0);
883 __m128i reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
889 tmp_buf, 32 * 16, tmp_buf, 32 * 24, reg0, reg1, reg2, reg3);
912 VP9_DOTP_CONST_PAIR(reg0, reg4, cospi_16_64, cospi_16_64, reg0, reg4);
914 LSX_BUTTERFLY_4_H(reg4, reg0, reg2, reg6, vec1, vec3, vec2, vec0);
921 tmp_buf, 32 * 16, tmp_buf, 32 * 24, reg0, reg1, reg2, reg3);
934 VP9_DOTP_CONST_PAIR(reg0, reg7, cospi_30_64, cospi_2_64, reg0, reg7);
939 vec0 = __lsx_vadd_h(reg0, reg4);
940 reg0 = __lsx_vsub_h(reg0, reg4);
955 VP9_DOTP_CONST_PAIR(reg7, reg0, cospi_24_64, cospi_8_64, reg0, reg7);
959 vec0 = __lsx_vsub_h(reg0, reg6);
960 reg0 = __lsx_vadd_h(reg0, reg6);
987 LSX_BUTTERFLY_4_H(stp6, stp7, reg2, reg0, loc1, loc3, loc2, loc0);
998 __m128i reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
1002 reg0 = __lsx_vld(tmp_buf, 64);
1020 VP9_DOTP_CONST_PAIR(reg0, reg7, cospi_31_64, cospi_1_64, reg0, reg7);
1025 vec0 = __lsx_vadd_h(reg0, reg3);
1026 reg0 = __lsx_vsub_h(reg0, reg3);
1045 VP9_DOTP_CONST_PAIR(reg7, reg0, cospi_28_64, cospi_4_64, reg0, reg7);
1047 LSX_BUTTERFLY_4_H(reg0, reg7, reg6, reg1, vec0, vec1, vec2, vec3);
1056 reg0 = __lsx_vld(tmp_buf, 3 * 64);
1077 VP9_DOTP_CONST_PAIR(reg7, reg0, cospi_3_64, cospi_29_64, reg0, reg7);
1080 DUP4_ARG2(__lsx_vsub_h,reg1, reg2, reg6, reg5, reg0, reg3, reg7, reg4,
1092 DUP4_ARG2(__lsx_vadd_h, reg0, reg3, reg1, reg2, reg5, reg6, reg4, reg7,
1094 LSX_BUTTERFLY_4_H(vec0, vec3, vec2, vec1, reg0, reg1, reg3, reg2);
1095 __lsx_vst(reg0, tmp_odd_buf, 13 * 16);
1098 reg0, reg1);
1099 __lsx_vst(reg0, tmp_odd_buf, 8 * 16);
1105 tmp_odd_buf, 32, tmp_odd_buf, 48, reg0, reg1, reg2, reg3);
1110 DUP4_ARG2(__lsx_vadd_h, reg0, reg4, reg1, reg5, reg2, reg6, reg3, reg7,
1116 DUP2_ARG2(__lsx_vsub_h, reg0, reg4, reg1, reg5, vec0, vec1);
1129 reg1, reg2, reg0, reg3);
1134 DUP4_ARG2(__lsx_vadd_h, reg0, reg4, reg1, reg5, reg2, reg6, reg3, reg7,
1141 DUP2_ARG2(__lsx_vsub_h, reg0, reg4, reg3, reg7, vec0, vec1);