Lines Matching defs:vec9

749     v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
772 VSHF_B2_SB(src2, src2, src2, src3, mask0, mask3, vec6, vec9);
777 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
783 VSHF_B2_SB(src6, src6, src6, src7, mask0, mask3, vec6, vec9);
788 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
817 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
840 VSHF_B2_SB(src2, src2, src2, src3, mask0, mask3, vec6, vec9);
845 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
851 VSHF_B2_SB(src6, src6, src6, src7, mask0, mask3, vec6, vec9);
856 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
884 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
899 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
901 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
910 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
912 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
939 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
954 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
956 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
965 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
967 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
1056 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
1077 VSHF_B2_SB(src2, src2, src3, src3, mask0, mask0, vec6, vec9);
1082 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
1088 VSHF_B2_SB(src6, src6, src7, src7, mask0, mask0, vec6, vec9);
1093 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
1115 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
1131 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
1133 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b,
1142 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
1144 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b,
3152 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3176 VSHF_B2_SB(src2, src2, src2, src3, mask0, mask3, vec6, vec9);
3181 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
3187 VSHF_B2_SB(src6, src6, src6, src7, mask0, mask3, vec6, vec9);
3192 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
3224 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3248 VSHF_B2_SB(src2, src2, src2, src3, mask0, mask3, vec6, vec9);
3253 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
3259 VSHF_B2_SB(src6, src6, src6, src7, mask0, mask3, vec6, vec9);
3264 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
3296 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3311 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
3313 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
3322 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
3324 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
3361 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3376 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
3378 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
3387 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
3389 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
3497 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3519 VSHF_B2_SB(src2, src2, src3, src3, mask0, mask0, vec6, vec9);
3524 HADD_SB4_SH(vec0, vec3, vec6, vec9, res0, res1, res2, res3);
3530 VSHF_B2_SB(src6, src6, src7, src7, mask0, mask0, vec6, vec9);
3535 HADD_SB4_SH(vec0, vec3, vec6, vec9, res4, res5, res6, res7);
3561 v16i8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9, vec10;
3578 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec8, vec9);
3580 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,
3589 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec8, vec9);
3591 DPADD_SB4_SH(vec8, vec9, vec10, vec11, plus20b, plus20b, plus20b, plus20b,