Lines Matching refs:reg1

68 #define VP9_DOTP_CONST_PAIR(reg0, reg1, cnst0, cnst1, out0, out1)  \
76 s1_m = __lsx_vilvl_h(__lsx_vneg_h(reg1), reg0); \
77 s0_m = __lsx_vilvh_h(__lsx_vneg_h(reg1), reg0); \
78 s3_m = __lsx_vilvl_h(reg0, reg1); \
79 s2_m = __lsx_vilvh_h(reg0, reg1); \
377 __m128i reg1, reg3, reg5, reg7, reg9, reg11, reg13, reg15;
383 reg0, reg1, reg2, reg3);
426 VP9_DOTP_CONST_PAIR(reg1, reg15, cospi_30_64, cospi_2_64, reg1, reg15);
429 reg9 = __lsx_vsub_h(reg1, loc2);
430 reg1 = __lsx_vadd_h(reg1, loc2);
443 loc1 = __lsx_vadd_h(reg1, reg13);
444 reg13 = __lsx_vsub_h(reg1, reg13);
458 reg1 = __lsx_vsub_h(reg6, loc0);
493 DUP4_ARG2(__lsx_vsrari_h, reg1, 6, reg7, 6, reg9, 6, reg15, 6,
494 reg1, reg7, reg9, reg15);
495 VP9_ADDBLK_ST8x4_UB(dst, dst_stride, reg7, reg9, reg1, reg15);
501 __m128i reg1, reg3, reg5, reg7, reg9, reg11, reg13, reg15;
508 reg0, reg1, reg2, reg3);
551 VP9_DOTP_CONST_PAIR(reg1, reg15, cospi_30_64, cospi_2_64, reg1, reg15);
554 reg9 = __lsx_vsub_h(reg1, loc2);
555 reg1 = __lsx_vadd_h(reg1, loc2);
568 loc1 = __lsx_vadd_h(reg1, reg13);
569 reg13 = __lsx_vsub_h(reg1, reg13);
583 reg1 = __lsx_vsub_h(reg6, loc0);
621 LSX_TRANSPOSE8x8_H(reg3, reg13, reg11, reg5, reg7, reg9, reg1, reg15,
622 reg3, reg13, reg11, reg5, reg7, reg9, reg1, reg15);
633 __lsx_vst(reg1, offset, 32*2);
883 __m128i reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
889 tmp_buf, 32 * 16, tmp_buf, 32 * 24, reg0, reg1, reg2, reg3);
904 VP9_DOTP_CONST_PAIR(reg1, reg7, cospi_28_64, cospi_4_64, reg1, reg7);
906 LSX_BUTTERFLY_4_H(reg1, reg7, reg3, reg5, vec1, vec3, vec2, vec0);
921 tmp_buf, 32 * 16, tmp_buf, 32 * 24, reg0, reg1, reg2, reg3);
937 VP9_DOTP_CONST_PAIR(reg6, reg1, cospi_6_64, cospi_26_64, reg6, reg1);
943 reg2 = __lsx_vadd_h(reg1, reg5);
944 reg1 = __lsx_vsub_h(reg1, reg5);
956 VP9_DOTP_CONST_PAIR(__lsx_vneg_h(reg6), reg1, cospi_24_64, cospi_8_64,
957 reg6, reg1);
961 vec1 = __lsx_vsub_h(reg7, reg1);
962 reg7 = __lsx_vadd_h(reg7, reg1);
964 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, reg6, reg1);
974 LSX_BUTTERFLY_4_H(stp2, stp3, reg4, reg1, loc1, loc3, loc2, loc0);
998 __m128i reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
1003 reg1 = __lsx_vld(tmp_buf, 7 * 64);
1023 VP9_DOTP_CONST_PAIR(reg6, reg1, cospi_7_64, cospi_25_64, reg1, reg6);
1029 reg4 = __lsx_vadd_h(reg1, reg2);
1030 reg1 = __lsx_vsub_h(reg1, reg2);
1046 VP9_DOTP_CONST_PAIR(reg6, reg1, -cospi_4_64, cospi_28_64, reg1, reg6);
1047 LSX_BUTTERFLY_4_H(reg0, reg7, reg6, reg1, vec0, vec1, vec2, vec3);
1057 reg1 = __lsx_vld(tmp_buf, 5 * 64);
1074 VP9_DOTP_CONST_PAIR(reg1, reg6, cospi_27_64, cospi_5_64, reg1, reg6);
1080 DUP4_ARG2(__lsx_vsub_h,reg1, reg2, reg6, reg5, reg0, reg3, reg7, reg4,
1092 DUP4_ARG2(__lsx_vadd_h, reg0, reg3, reg1, reg2, reg5, reg6, reg4, reg7,
1094 LSX_BUTTERFLY_4_H(vec0, vec3, vec2, vec1, reg0, reg1, reg3, reg2);
1096 __lsx_vst(reg1, tmp_odd_buf, 13 * 16 + 16);
1098 reg0, reg1);
1100 __lsx_vst(reg1, tmp_odd_buf, 8 * 16 + 16);
1105 tmp_odd_buf, 32, tmp_odd_buf, 48, reg0, reg1, reg2, reg3);
1110 DUP4_ARG2(__lsx_vadd_h, reg0, reg4, reg1, reg5, reg2, reg6, reg3, reg7,
1116 DUP2_ARG2(__lsx_vsub_h, reg0, reg4, reg1, reg5, vec0, vec1);
1129 reg1, reg2, reg0, reg3);
1134 DUP4_ARG2(__lsx_vadd_h, reg0, reg4, reg1, reg5, reg2, reg6, reg3, reg7,
1144 DUP2_ARG2(__lsx_vsub_h, reg1, reg5, reg2, reg6, vec0, vec1);