Lines Matching refs:mask2

35                                    mask0, mask1, mask2, mask3,              \
45 VSHF_B2_SB(src0, src1, src2, src3, mask2, mask2, vec4_m, vec5_m); \
52 mask0, mask1, mask2, mask3, \
62 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0_m, vec1_m); \
63 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2_m, vec3_m); \
290 v16u8 mask0, mask1, mask2, mask3, out;
302 mask2 = mask0 + 4;
307 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
321 v16u8 mask0, mask1, mask2, mask3, out;
332 mask2 = mask0 + 4;
338 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
342 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
356 v16u8 mask0, mask1, mask2, mask3, out;
368 mask2 = mask0 + 4;
374 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
379 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
392 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
397 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
427 v16u8 mask0, mask1, mask2, mask3, tmp0, tmp1;
439 mask2 = mask0 + 4;
451 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0_m, vec1_m);
452 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2_m, vec3_m);
478 v16u8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask00;
495 mask2 = mask00 + 4;
515 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0, vec1);
516 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2, vec3);
557 v16u8 mask0, mask1, mask2, mask3, out;
570 mask2 = mask0 + 4;
584 HORIZ_8TAP_8WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
596 HORIZ_8TAP_8WID_4VECS_FILT(src4, src5, src6, src7, mask0, mask1, mask2,
616 v16u8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask7, out;
629 mask2 = mask0 + 4;
647 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0, vec8);
648 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2, vec9);
685 v16u8 mask0, mask1, mask2, mask3, out;
698 mask2 = mask0 + 4;
716 HORIZ_8TAP_8WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2,
728 HORIZ_8TAP_8WID_4VECS_FILT(src4, src5, src6, src7, mask0, mask1, mask2,
748 v16u8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask7, out;
759 mask2 = mask0 + 4;
784 VSHF_B3_SB(src0, src0, src1, src1, src2, src2, mask2, mask2, mask2,
807 VSHF_B3_SB(src2, src3, src3, src3, src4, src4, mask6, mask2, mask2,
832 v16u8 mask0, mask1, mask2, mask3, out;
846 mask2 = mask0 + 4;
859 VSHF_B2_SB(src0, src0, src1, src1, mask2, mask2, vec0, vec1);
860 VSHF_B2_SB(src2, src2, src3, src3, mask2, mask2, vec2, vec3);
883 VSHF_B2_SB(src4, src4, src5, src5, mask2, mask2, vec0, vec1);
884 VSHF_B2_SB(src6, src6, src7, src7, mask2, mask2, vec2, vec3);
1349 v16i8 mask1, mask2, mask3;
1369 mask2 = mask0 + 4;
1376 VSHF_B4_SB(src0, src3, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1377 VSHF_B4_SB(src1, src4, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1378 VSHF_B4_SB(src2, src5, mask0, mask1, mask2, mask3,
1380 VSHF_B4_SB(src3, src6, mask0, mask1, mask2, mask3,
1404 VSHF_B4_SB(src7, src11, mask0, mask1, mask2, mask3,
1406 VSHF_B4_SB(src8, src12, mask0, mask1, mask2, mask3,
1408 VSHF_B4_SB(src9, src13, mask0, mask1, mask2, mask3,
1410 VSHF_B4_SB(src10, src14, mask0, mask1, mask2, mask3,
1484 v16i8 mask1, mask2, mask3;
1507 mask2 = mask0 + 4;
1519 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3,
1521 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3,
1523 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3,
1525 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3,
1536 VSHF_B4_SB(src4, src4, mask0, mask1, mask2, mask3,
1538 VSHF_B4_SB(src5, src5, mask0, mask1, mask2, mask3,
1540 VSHF_B4_SB(src6, src6, mask0, mask1, mask2, mask3,
1561 VSHF_B4_SB(src7, src7, mask0, mask1, mask2, mask3,
1574 VSHF_B4_SB(src8, src8, mask0, mask1, mask2, mask3,
1633 v16i8 mask0, mask1, mask2, mask3, mask4, mask5, mask6, mask7;
1658 mask2 = mask0 + 4;
1669 VSHF_B4_SB(src0, src0, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1670 VSHF_B4_SB(src1, src1, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1671 VSHF_B4_SB(src2, src2, mask0, mask1, mask2, mask3, vec8, vec9, vec10,
1673 VSHF_B4_SB(src3, src3, mask0, mask1, mask2, mask3, vec12, vec13, vec14,
1684 VSHF_B4_SB(src4, src4, mask0, mask1, mask2, mask3, vec0, vec1, vec2, vec3);
1685 VSHF_B4_SB(src5, src5, mask0, mask1, mask2, mask3, vec4, vec5, vec6, vec7);
1686 VSHF_B4_SB(src6, src6, mask0, mask1, mask2, mask3, vec8, vec9, vec10,
1707 VSHF_B4_SB(src7, src7, mask0, mask1, mask2, mask3, vec0, vec1, vec2,
1720 VSHF_B4_SB(src8, src8, mask0, mask1, mask2, mask3, vec0, vec1, vec2,
2201 v16i8 src0, src1, src2, src3, filt0, filt1, mask0, mask1, mask2, mask3;
2208 mask2 = LD_SB(&ff_hevc_mask_arr[32]);
2217 mask3 = mask2 + 2;
2224 VSHF_B2_SB(src0, src1, src2, src3, mask2, mask2, vec0, vec1);
3838 v16i8 mask0, mask1, mask2, mask3;
3931 mask2 = LD_SB(ff_hevc_mask_arr + 16);
3932 mask3 = mask2 + 2;
3937 VSHF_B2_SB(src0, src1, src0, src1, mask2, mask3, vec0, vec1);
3938 VSHF_B2_SB(src1, src2, src1, src2, mask2, mask3, vec2, vec3);
3951 VSHF_B2_SB(src3, src7, src3, src7, mask2, mask3, vec0, vec1);
3952 VSHF_B2_SB(src4, src8, src4, src8, mask2, mask3, vec2, vec3);
3953 VSHF_B2_SB(src5, src9, src5, src9, mask2, mask3, vec4, vec5);
3954 VSHF_B2_SB(src6, src10, src6, src10, mask2, mask3, vec6, vec7);