Lines Matching refs:mask2
455 v16i8 mask1, mask2, mask3;
469 mask2 = mask0 + 4;
477 VSHF_B4_SB(src0, src1, mask0, mask1, mask2, mask3,
482 VSHF_B4_SB(src2, src3, mask0, mask1, mask2, mask3,
487 VSHF_B4_SB(src4, src5, mask0, mask1, mask2, mask3,
492 VSHF_B4_SB(src6, src7, mask0, mask1, mask2, mask3,
510 v16i8 mask1, mask2, mask3;
524 mask2 = mask0 + 4;
532 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3,
537 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3,
542 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3,
547 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3,
565 v16i8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask7;
579 mask2 = mask0 + 4;
610 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0, vec1);
611 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2, vec3);
640 v16i8 mask1, mask2, mask3;
654 mask2 = mask0 + 4;
675 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0, vec1);
676 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2, vec3);
697 v16i8 mask1, mask2, mask3, mask4, mask5, mask6, mask7;
708 mask2 = mask0 + 4;
743 VSHF_B2_SB(src0, src0, src0, src1, mask2, mask6, vec0, vec1);
744 VSHF_B2_SB(src1, src1, src2, src2, mask2, mask2, vec2, vec3);
745 VSHF_B2_SB(src2, src3, src3, src3, mask6, mask2, vec4, vec5);
772 v16i8 mask1, mask2, mask3, mask4, mask5, mask6, mask7;
783 mask2 = mask0 + 4;
799 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3,
809 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3,
814 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3,
832 v16i8 mask1, mask2, mask3, mask4, mask5, mask6, mask7;
843 mask2 = mask0 + 4;
873 VSHF_B2_SB(src0, src0, src0, src1, mask2, mask6, vec0, vec1);
874 VSHF_B2_SB(src1, src1, src1, src2, mask2, mask6, vec2, vec3);
887 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec4, vec5);
903 v16i8 mask1, mask2, mask3, mask4, mask5, mask6, mask7;
915 mask2 = mask0 + 4;
931 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3,
945 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3,
959 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3,
973 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3,
980 VSHF_B4_SB(src4, src4, mask0, mask1, mask2, mask3,
1369 v16i8 mask1, mask2, mask3;
1389 mask2 = mask0 + 4;
1399 VSHF_B4_SB(src0, src3, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1400 VSHF_B4_SB(src1, src4, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1401 VSHF_B4_SB(src2, src5, mask0, mask1, mask2, mask3,
1403 VSHF_B4_SB(src3, src6, mask0, mask1, mask2, mask3,
1428 VSHF_B4_SB(src7, src9, mask0, mask1, mask2, mask3,
1430 VSHF_B4_SB(src8, src10, mask0, mask1, mask2, mask3,
1481 v16i8 mask1, mask2, mask3;
1501 mask2 = mask0 + 4;
1516 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3,
1518 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3,
1520 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3,
1522 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3,
1538 VSHF_B4_SB(src4, src4, mask0, mask1, mask2, mask3,
1540 VSHF_B4_SB(src5, src5, mask0, mask1, mask2, mask3,
1542 VSHF_B4_SB(src6, src6, mask0, mask1, mask2, mask3,
1559 VSHF_B4_SB(src7, src7, mask0, mask1, mask2, mask3,
1612 v16i8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask7;
1634 mask2 = mask0 + 4;
1648 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1649 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1650 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3, vec8, vec9, vec10,
1652 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3, vec12, vec13, vec14,
1668 VSHF_B4_SB(src4, src4, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1669 VSHF_B4_SB(src5, src5, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1670 VSHF_B4_SB(src6, src6, mask0, mask1, mask2, mask3, vec8, vec9, vec10,
1687 VSHF_B4_SB(src7, src7, mask0, mask1, mask2, mask3, vec0, vec1, vec2,
2179 v16i8 mask2 = {
2189 mask3 = mask2 + 2;
2211 VSHF_B2_SB(src0, src1, src0, src1, mask2, mask3, vec0, vec1);
2214 VSHF_B2_SB(src2, src3, src2, src3, mask2, mask3, vec0, vec1);
2408 v16i8 mask1, mask2, mask3;
2422 mask2 = mask0 + 8;
2436 VSHF_B2_SB(src0, src0, src0, src1, mask0, mask2, vec0, vec1);
4009 v16i8 mask0, mask1, mask2, mask3;
4106 mask2 = LD_SB(ff_hevc_mask_arr + 16);
4107 mask3 = mask2 + 2;
4112 VSHF_B2_SB(src0, src1, src0, src1, mask2, mask3, vec0, vec1);
4113 VSHF_B2_SB(src1, src2, src1, src2, mask2, mask3, vec2, vec3);
4126 VSHF_B2_SB(src3, src7, src3, src7, mask2, mask3, vec0, vec1);
4127 VSHF_B2_SB(src4, src8, src4, src8, mask2, mask3, vec2, vec3);
4128 VSHF_B2_SB(src5, src9, src5, src9, mask2, mask3, vec4, vec5);
4129 VSHF_B2_SB(src6, src10, src6, src10, mask2, mask3, vec6, vec7);