Lines Matching defs:vec1
741 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
746 vec1 = __lsx_vld(tmp_odd_buf, 9 * 16);
754 DUP4_ARG2(__lsx_vadd_h,loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
761 __lsx_vst(SUB(loc2, vec1), tmp_buf, 27 * 16);
766 vec1 = __lsx_vld(tmp_odd_buf, 13 * 16);
774 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
779 __lsx_vst(SUB(loc2, vec1), tmp_buf, 25 * 16);
784 vec1 = __lsx_vld(tmp_odd_buf, 11 * 16);
792 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
797 __lsx_vst(SUB(loc2, vec1), tmp_buf, 26 * 16);
802 vec1 = __lsx_vld(tmp_odd_buf, 15 * 16);
810 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
815 __lsx_vst(SUB(loc2, vec1), tmp_buf, 24 * 16);
882 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
906 LSX_BUTTERFLY_4_H(reg1, reg7, reg3, reg5, vec1, vec3, vec2, vec0);
910 loc0 = vec1;
914 LSX_BUTTERFLY_4_H(reg4, reg0, reg2, reg6, vec1, vec3, vec2, vec0);
915 LSX_BUTTERFLY_4_H(vec0, vec1, loc1, loc0, stp3, stp0, stp7, stp4);
949 vec1 = reg2;
952 reg4 = __lsx_vsub_h(reg5, vec1);
953 reg5 = __lsx_vadd_h(reg5, vec1);
961 vec1 = __lsx_vsub_h(reg7, reg1);
964 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, reg6, reg1);
997 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1036 DUP2_ARG2(__lsx_vadd_h, reg5, reg4, reg3, reg2, vec0, vec1);
1038 __lsx_vst(vec1, tmp_odd_buf, 4 * 16 + 16);
1039 DUP2_ARG2(__lsx_vsub_h, reg5, reg4, reg3, reg2, vec0, vec1);
1040 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_24_64, cospi_8_64, vec0, vec1);
1042 __lsx_vst(vec1, tmp_odd_buf, 16);
1047 LSX_BUTTERFLY_4_H(reg0, reg7, reg6, reg1, vec0, vec1, vec2, vec3);
1049 __lsx_vst(vec1, tmp_odd_buf, 6 * 16 + 16);
1081 vec0, vec1, vec2, vec3);
1082 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_12_64, cospi_20_64, loc0, loc1);
1084 LSX_BUTTERFLY_4_H(loc2, loc3, loc1, loc0, vec0, vec1, vec3, vec2);
1086 __lsx_vst(vec1, tmp_odd_buf, 12 * 16 + 3 * 16);
1087 VP9_DOTP_CONST_PAIR(vec3, vec2, -cospi_8_64, cospi_24_64, vec0, vec1);
1089 __lsx_vst(vec1, tmp_odd_buf, 10 * 16 + 16);
1093 vec0, vec1, vec2, vec3);
1094 LSX_BUTTERFLY_4_H(vec0, vec3, vec2, vec1, reg0, reg1, reg3, reg2);
1116 DUP2_ARG2(__lsx_vsub_h, reg0, reg4, reg1, reg5, vec0, vec1);
1117 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc0, loc1);
1119 DUP2_ARG2(__lsx_vsub_h, reg2, reg6, reg3, reg7, vec0, vec1);
1120 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc2, loc3);
1141 DUP2_ARG2(__lsx_vsub_h, reg0, reg4, reg3, reg7, vec0, vec1);
1142 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc0, loc1);
1144 DUP2_ARG2(__lsx_vsub_h, reg1, reg5, reg2, reg6, vec0, vec1);
1145 VP9_DOTP_CONST_PAIR(vec1, vec0, cospi_16_64, cospi_16_64, loc2, loc3);
1157 __m128i vec0, vec1, vec2, vec3, loc0, loc1, loc2, loc3;
1162 vec1 = __lsx_vld(tmp_odd_buf, 9 * 16);
1170 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1175 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1183 vec1 = __lsx_vld(tmp_odd_buf, 13 * 16);
1191 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1197 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1205 vec1 = __lsx_vld(tmp_odd_buf, 11 * 16);
1213 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1218 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1226 vec1 = __lsx_vld(tmp_odd_buf, 15 * 16);
1234 DUP4_ARG2(__lsx_vadd_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,
1239 DUP4_ARG2(__lsx_vsub_h, loc0, vec3, loc1, vec2, loc2, vec1, loc3, vec0,