Lines Matching refs:q3

238     __m128i p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out;
244 q3 = __lsx_vldx(dst, stride3);
250 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
272 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
278 q3 = __lsx_vldx(dst, stride3);
292 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit0, b_limit0, thresh0,
311 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
322 q3 = __lsx_vldx(dst, stride3);
328 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
330 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
345 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
382 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
398 q3 = __lsx_vldx(dst, stride3);
413 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
415 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
428 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
435 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
473 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
486 q3 = __lsx_vldx(dst, stride3);
501 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
503 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
518 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
555 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
568 q3 = __lsx_vldx(dst, stride3);
583 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
585 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
600 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
638 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
654 q3 = __lsx_vldx(dst, stride3);
661 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
663 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
677 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
684 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
724 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
748 q3 = __lsx_vldx(dst, stride3);
866 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
873 q3_h_in = (v8u16)__lsx_vilvh_b(zero, q3);
1007 /* q3 */
1021 q3 = __lsx_vbitsel_v(q3, out_l, flat2);
1022 __lsx_vst(q3, dst, 0);
1108 __m128i p3, p2, p1, p0, q3, q2, q1, q0, p7, p6, p5, p4, q4, q5, q6, q7;
1122 q3 = __lsx_vldx(dst, stride3);
1128 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
1130 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1146 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1308 /* calculation of q3 and q4 */
1321 p0_filter16 = __lsx_vbitsel_v(q3, p0_filter16, flat2);
1361 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1369 q3 = __lsx_vldx(dst_tmp2, stride3);
1375 LSX_TRANSPOSE8x8_B(p3, p2, p1, p0, q0, q1, q2, q3,
1376 p3, p2, p1, p0, q0, q1, q2, q3);
1377 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
1407 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1430 p3, p2, p1, p0, q0, q1, q2, q3);
1444 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit0, b_limit0, thresh0,
1485 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1501 q3 = __lsx_vldx(dst_tmp, stride3);
1503 LSX_TRANSPOSE8x8_B(p3, p2, p1, p0, q0, q1, q2, q3,
1504 p3, p2, p1, p0, q0, q1, q2, q3);
1511 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
1514 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1541 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1603 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1624 q3 = __lsx_vld(dst_tmp, 0);
1634 q3, q2, q1, q0, row12, row13, row14, row15,
1635 p3, p2, p1, p0, q0, q1, q2, q3);
1650 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
1653 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1690 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1697 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
1788 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1806 q3 = __lsx_vld(dst_tmp, 0);
1816 q3, q2, q1, q0, row12, row13, row14, row15,
1817 p3, p2, p1, p0, q0, q1, q2, q3);
1832 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
1835 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1874 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1963 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
1981 q3 = __lsx_vld(dst_tmp, 0);
1991 q3, q2, q1, q0, row12, row13, row14, row15,
1992 p3, p2, p1, p0, q0, q1, q2, q3);
2007 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
2010 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
2049 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
2135 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
2152 DUP4_ARG2(__lsx_vbsrl_v, q0, 8, q2, 8, q4, 8, q6, 8, q1, q3, q5, q7);
2165 __lsx_vst(q3, output, 176);
2176 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
2186 q0, q1, q2, q3);
2189 LSX_TRANSPOSE16x8_B(p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5,
2202 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
2226 q3 = __lsx_vpackod_d(row12, row4);
2234 DUP2_ARG2(__lsx_vpackev_b, q2, q3, q0, q1, q5, q7);
2235 DUP2_ARG2(__lsx_vpackod_b, q2, q3, q0, q1, tmp6, tmp7);
2252 q3 = __lsx_vpackev_w(tmp3, tmp2);
2258 LSX_ST_8(q0, q1, q2, q3, q4, q5, q6, q7, output, out_stride,
2268 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
2280 DUP4_ARG2(__lsx_vld, src, 0, src, 16, src, 32, src, 48, q0, q1, q2, q3);
2287 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
2290 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
2323 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
2362 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
2376 DUP4_ARG2(__lsx_vld, dst, 0, dst, 16, dst, 32, dst, 48, q0, q1, q2, q3);
2478 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
2565 /* q3 */
2572 q3 = __lsx_vbitsel_v(q3, out_l, flat2);
2573 __lsx_vstelm_d(q3, dst, 0, 0);
2646 __m128i p3, p2, p1, p0, q3, q2, q1, q0;
2661 DUP4_ARG2(__lsx_vld, dst, 0, dst, 16, dst, 32, dst, 48, q0, q1, q2, q3);
2668 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh,
2671 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
2710 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
2716 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
2755 __m128i p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7;
2774 DUP4_ARG2(__lsx_vld, dst, 0, dst, 16, dst, 32, dst, 48, q0, q1, q2, q3);
2934 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
2940 q3_h_in = (v8u16)__lsx_vilvh_b(zero, q3);
3054 /* q3 */
3066 q3 = __lsx_vbitsel_v(q3, out_l, flat2);
3067 __lsx_vst(q3, dst, 16*10);