Lines Matching refs:loc1
375 __m128i loc0, loc1, loc2, loc3;
410 LSX_BUTTERFLY_4_H(reg2, reg14, reg6, reg10, loc0, loc1, reg14, reg2);
416 reg0 = __lsx_vsub_h(reg2, loc1);
417 reg2 = __lsx_vadd_h(reg2, loc1);
435 VP9_DOTP_CONST_PAIR(reg13, reg3, cospi_6_64, cospi_26_64, loc0, loc1);
436 LSX_BUTTERFLY_4_H(loc0, loc1, reg11, reg5, reg13, reg3, reg11, reg5);
438 loc1 = __lsx_vadd_h(reg15, reg3);
440 loc2 = __lsx_vadd_h(reg2, loc1);
441 reg15 = __lsx_vsub_h(reg2, loc1);
443 loc1 = __lsx_vadd_h(reg1, reg13);
445 loc0 = __lsx_vadd_h(reg0, loc1);
446 loc1 = __lsx_vsub_h(reg0, loc1);
448 tmp7 = loc1;
462 loc1 = __lsx_vadd_h(reg4, loc0);
464 tmp5 = loc1;
467 LSX_BUTTERFLY_4_H(reg8, reg10, reg11, reg5, loc0, reg4, reg9, loc1);
470 reg11 = loc1;
500 __m128i loc0, loc1, loc2, loc3;
535 LSX_BUTTERFLY_4_H(reg2, reg14, reg6, reg10, loc0, loc1, reg14, reg2);
541 reg0 = __lsx_vsub_h(reg2, loc1);
542 reg2 = __lsx_vadd_h(reg2, loc1);
560 VP9_DOTP_CONST_PAIR(reg13, reg3, cospi_6_64, cospi_26_64, loc0, loc1);
561 LSX_BUTTERFLY_4_H(loc0, loc1, reg11, reg5, reg13, reg3, reg11, reg5);
563 loc1 = __lsx_vadd_h(reg15, reg3);
565 loc2 = __lsx_vadd_h(reg2, loc1);
566 reg15 = __lsx_vsub_h(reg2, loc1);
568 loc1 = __lsx_vadd_h(reg1, reg13);
570 loc0 = __lsx_vadd_h(reg0, loc1);
571 loc1 = __lsx_vsub_h(reg0, loc1);
573 tmp7 = loc1;
587 loc1 = __lsx_vadd_h(reg4, loc0);
590 tmp5 = loc1;
593 LSX_BUTTERFLY_4_H(reg8, reg10, reg11, reg5, loc0, reg4, reg9, loc1);
596 reg11 = loc1;
741 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
750 loc1 = __lsx_vld(tmp_eve_buf, 8 * 16);
754 DUP4_ARG2(__lsx_vadd_h,loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
760 __lsx_vst(SUB(loc1, vec2), tmp_buf, 23 * 16);
770 loc1 = __lsx_vld(tmp_eve_buf, 10 * 16);
774 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
778 __lsx_vst(SUB(loc1, vec2), tmp_buf, 21 * 16);
788 loc1 = __lsx_vld(tmp_eve_buf, 9 * 16);
792 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
796 __lsx_vst(SUB(loc1, vec2), tmp_buf, 22 * 16);
806 loc1 = __lsx_vld(tmp_eve_buf, 11 * 16);
810 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
814 __lsx_vst(SUB(loc1, vec2), tmp_buf, 20 * 16);
882 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
909 loc1 = vec3;
915 LSX_BUTTERFLY_4_H(vec0, vec1, loc1, loc0, stp3, stp0, stp7, stp4);
969 LSX_BUTTERFLY_4_H(stp0, stp1, reg7, reg5, loc1, loc3, loc2, loc0);
970 __lsx_vst(loc1, tmp_eve_buf, 0);
974 LSX_BUTTERFLY_4_H(stp2, stp3, reg4, reg1, loc1, loc3, loc2, loc0);
975 __lsx_vst(loc1, tmp_eve_buf, 2 * 16);
981 LSX_BUTTERFLY_4_H(stp4, stp5, reg6, reg3, loc1, loc3, loc2, loc0);
982 __lsx_vst(loc1, tmp_eve_buf, 4 * 16);
987 LSX_BUTTERFLY_4_H(stp6, stp7, reg2, reg0, loc1, loc3, loc2, loc0);
988 __lsx_vst(loc1, tmp_eve_buf, 6 * 16);
997 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1082 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_12_64, cospi_20_64, loc0, loc1);
1084 LSX_BUTTERFLY_4_H(loc2, loc3, loc1, loc0, vec0, vec1, vec3, vec2);
1111 loc0, loc1, loc2, loc3);
1113 __lsx_vst(loc1, tmp_odd_buf, 16);
1117 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc0, loc1);
1122 __lsx_vst(loc1, tmp_odd_buf, 8 * 16 + 16);
1135 loc0, loc1, loc2, loc3);
1137 __lsx_vst(loc1, tmp_odd_buf, 4 * 16 + 16);
1142 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc0, loc1);
1147 __lsx_vst(loc1, tmp_odd_buf, 12 * 16 + 16);
1157 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1166 loc1 = __lsx_vld(tmp_eve_buf, 8 * 16);
1170 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1175 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1187 loc1 = __lsx_vld(tmp_eve_buf, 10 * 16);
1191 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1197 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1209 loc1 = __lsx_vld(tmp_eve_buf, 9 * 16);
1213 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1218 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1230 loc1 = __lsx_vld(tmp_eve_buf, 11 * 16);
1234 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1239 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,