Lines Matching refs:loc2
375 __m128i loc0, loc1, loc2, loc3;
411 VP9_DOTP_CONST_PAIR(reg14, reg2, cospi_16_64, cospi_16_64, loc2, loc3);
422 reg8 = __lsx_vsub_h(reg10, loc2);
423 reg10 = __lsx_vadd_h(reg10, loc2);
427 VP9_DOTP_CONST_PAIR(reg9, reg7, cospi_14_64, cospi_18_64, loc2, loc3);
429 reg9 = __lsx_vsub_h(reg1, loc2);
430 reg1 = __lsx_vadd_h(reg1, loc2);
440 loc2 = __lsx_vadd_h(reg2, loc1);
449 reg0 = loc2;
463 loc2 = __lsx_vsub_h(reg4, loc0);
474 reg13 = loc2;
500 __m128i loc0, loc1, loc2, loc3;
536 VP9_DOTP_CONST_PAIR(reg14, reg2, cospi_16_64, cospi_16_64, loc2, loc3);
547 reg8 = __lsx_vsub_h(reg10, loc2);
548 reg10 = __lsx_vadd_h(reg10, loc2);
552 VP9_DOTP_CONST_PAIR(reg9, reg7, cospi_14_64, cospi_18_64, loc2, loc3);
554 reg9 = __lsx_vsub_h(reg1, loc2);
555 reg1 = __lsx_vadd_h(reg1, loc2);
565 loc2 = __lsx_vadd_h(reg2, loc1);
574 reg0 = loc2;
588 loc2 = __lsx_vsub_h(reg4, loc0);
600 reg13 = loc2;
741 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
751 loc2 = __lsx_vld(tmp_eve_buf, 4 * 16);
754 DUP4_ARG2(__lsx_vadd_h,loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
761 __lsx_vst(SUB(loc2, vec1), tmp_buf, 27 * 16);
771 loc2 = __lsx_vld(tmp_eve_buf, 6 * 16);
774 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
779 __lsx_vst(SUB(loc2, vec1), tmp_buf, 25 * 16);
789 loc2 = __lsx_vld(tmp_eve_buf, 5 * 16);
792 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
797 __lsx_vst(SUB(loc2, vec1), tmp_buf, 26 * 16);
807 loc2 = __lsx_vld(tmp_eve_buf, 7 * 16);
810 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
815 __lsx_vst(SUB(loc2, vec1), tmp_buf, 24 * 16);
882 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
907 VP9_DOTP_CONST_PAIR(vec2, vec0, cospi_16_64, cospi_16_64, loc2, loc3);
916 LSX_BUTTERFLY_4_H(vec2, vec3, loc3, loc2, stp2, stp1, stp6, stp5);
969 LSX_BUTTERFLY_4_H(stp0, stp1, reg7, reg5, loc1, loc3, loc2, loc0);
972 __lsx_vst(loc2, tmp_eve_buf, 14 * 16);
974 LSX_BUTTERFLY_4_H(stp2, stp3, reg4, reg1, loc1, loc3, loc2, loc0);
977 __lsx_vst(loc2, tmp_eve_buf, 12 * 16);
981 LSX_BUTTERFLY_4_H(stp4, stp5, reg6, reg3, loc1, loc3, loc2, loc0);
984 __lsx_vst(loc2, tmp_eve_buf, 10 * 16);
987 LSX_BUTTERFLY_4_H(stp6, stp7, reg2, reg0, loc1, loc3, loc2, loc0);
990 __lsx_vst(loc2, tmp_eve_buf, 8 * 16);
997 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1083 VP9_DOTP_CONST_PAIR(vec3, vec2, -cospi_20_64, cospi_12_64, loc2, loc3);
1084 LSX_BUTTERFLY_4_H(loc2, loc3, loc1, loc0, vec0, vec1, vec3, vec2);
1111 loc0, loc1, loc2, loc3);
1114 __lsx_vst(loc2, tmp_odd_buf, 32);
1120 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc2, loc3);
1123 __lsx_vst(loc2, tmp_odd_buf, 8 * 16 + 32);
1135 loc0, loc1, loc2, loc3);
1138 __lsx_vst(loc2, tmp_odd_buf, 4 * 16 + 32);
1145 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc2, loc3);
1148 __lsx_vst(loc2, tmp_odd_buf, 12 * 16 + 32);
1157 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1167 loc2 = __lsx_vld(tmp_eve_buf, 4 * 16);
1170 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1175 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1188 loc2 = __lsx_vld(tmp_eve_buf, 6 * 16);
1191 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1197 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1210 loc2 = __lsx_vld(tmp_eve_buf, 5 * 16);
1213 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1218 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1231 loc2 = __lsx_vld(tmp_eve_buf, 7 * 16);
1234 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1239 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,