Lines Matching refs:flat
202 v16u8 mask, hev, flat, thresh, b_limit, limit;
213 hev, mask, flat);
230 v16u8 mask, hev, flat, thresh0, b_limit0, limit0, thresh1, b_limit1, limit1;
249 hev, mask, flat);
261 v16u8 mask, hev, flat, thresh, b_limit, limit;
277 hev, mask, flat);
278 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
282 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
284 /* if flat is zero for all pixels, then no need to calculate other filter */
285 if (__msa_test_bz_v(flat)) {
305 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filter8, flat);
306 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filter8, flat);
307 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filter8, flat);
308 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filter8, flat);
309 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filter8, flat);
310 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filter8, flat);
336 v16u8 flat, mask, hev, tmp, thresh, b_limit, limit;
362 hev, mask, flat);
363 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
367 /* if flat is zero for all pixels, then no need to calculate other filter */
368 if (__msa_test_bz_v(flat)) {
392 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
393 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
394 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
395 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
396 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
397 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
415 v16u8 flat, mask, hev, tmp, thresh, b_limit, limit;
438 hev, mask, flat);
439 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
443 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
445 /* if flat is zero for all pixels, then no need to calculate other filter */
446 if (__msa_test_bz_v(flat)) {
463 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
464 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
465 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
466 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
467 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
468 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
486 v16u8 flat, mask, hev, tmp, thresh, b_limit, limit;
509 hev, mask, flat);
510 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
514 flat = (v16u8) __msa_insve_d((v2i64) flat, 0, (v2i64) zero);
516 /* if flat is zero for all pixels, then no need to calculate other filter */
517 if (__msa_test_bz_v(flat)) {
535 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filt8_l, flat);
536 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_l, flat);
537 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_l, flat);
538 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_l, flat);
539 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_l, flat);
540 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filt8_l, flat);
559 v16u8 flat, mask, hev, thresh, b_limit, limit;
577 hev, mask, flat);
578 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
582 /* if flat is zero for all pixels, then no need to calculate other filter */
583 if (__msa_test_bz_v(flat)) {
609 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
610 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
611 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
612 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
613 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
614 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
620 ST_UB(flat, filter48);
628 v16u8 flat, flat2, filter8;
642 flat = LD_UB(filter48 + 96);
646 VP9_FLAT5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, flat, flat2);
981 v16u8 flat2, mask, hev, flat, thresh, b_limit, limit;
1000 hev, mask, flat);
1001 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1005 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
1007 /* if flat is zero for all pixels, then no need to calculate other filter */
1008 if (__msa_test_bz_v(flat)) {
1031 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filter8, flat);
1032 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filter8, flat);
1033 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filter8, flat);
1034 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filter8, flat);
1035 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filter8, flat);
1036 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filter8, flat);
1042 VP9_FLAT5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, flat, flat2);
1203 v16u8 mask, hev, flat, limit, thresh, b_limit;
1216 hev, mask, flat);
1230 v16u8 mask, hev, flat;
1258 hev, mask, flat);
1278 v16u8 flat, mask, hev, thresh, b_limit, limit;
1297 hev, mask, flat);
1299 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1304 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
1306 /* if flat is zero for all pixels, then no need to calculate other filter */
1307 if (__msa_test_bz_v(flat)) {
1328 p2 = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
1329 p1 = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
1330 p0 = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
1331 q0 = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
1332 q1 = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
1333 q2 = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
1357 v16u8 flat, mask, hev, thresh, b_limit, limit;
1393 hev, mask, flat);
1395 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1400 /* if flat is zero for all pixels, then no need to calculate other filter */
1401 if (__msa_test_bz_v(flat)) {
1434 p2 = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
1435 p1 = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
1436 p0 = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
1437 q0 = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
1438 q1 = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
1439 q2 = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
1470 v16u8 flat, mask, hev, thresh, b_limit, limit;
1503 hev, mask, flat);
1505 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1510 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
1512 /* if flat is zero for all pixels, then no need to calculate other filter */
1513 if (__msa_test_bz_v(flat)) {
1537 p2 = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
1538 p1 = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
1539 p0 = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
1540 q0 = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
1541 q1 = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
1542 q2 = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
1573 v16u8 flat, mask, hev, thresh, b_limit, limit;
1606 hev, mask, flat);
1608 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1613 flat = (v16u8) __msa_insve_d((v2i64) flat, 0, (v2i64) zero);
1615 /* if flat is zero for all pixels, then no need to calculate other filter */
1616 if (__msa_test_bz_v(flat)) {
1642 p2 = __msa_bmnz_v(p2, (v16u8) p2_filt8_l, flat);
1643 p1 = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_l, flat);
1644 p0 = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_l, flat);
1645 q0 = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_l, flat);
1646 q1 = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_l, flat);
1647 q2 = __msa_bmnz_v(q2, (v16u8) q2_filt8_l, flat);
1778 v16u8 flat, mask, hev, thresh, b_limit, limit;
1794 hev, mask, flat);
1796 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
1801 flat = (v16u8) __msa_ilvr_d((v2i64) zero, (v2i64) flat);
1803 /* if flat is zero for all pixels, then no need to calculate other filter */
1804 if (__msa_test_bz_v(flat)) {
1825 p2_out = __msa_bmnz_v(p2, (v16u8) p2_r, flat);
1826 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_r, flat);
1827 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_r, flat);
1828 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_r, flat);
1829 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_r, flat);
1830 q2_out = __msa_bmnz_v(q2, (v16u8) q2_r, flat);
1836 ST_UB(flat, filter48);
1846 v16u8 filter8, flat, flat2;
1855 flat = LD_UB(filter48 + 6 * 16);
1860 VP9_FLAT5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, flat, flat2);
2099 v16u8 flat, mask, hev, thresh, b_limit, limit;
2118 hev, mask, flat);
2120 VP9_FLAT4(p3, p2, p0, q0, q2, q3, flat);
2125 /* if flat is zero for all pixels, then no need to calculate other filter */
2126 if (__msa_test_bz_v(flat)) {
2158 p2_out = __msa_bmnz_v(p2, (v16u8) p2_filt8_r, flat);
2159 p1_out = __msa_bmnz_v(p1_out, (v16u8) p1_filt8_r, flat);
2160 p0_out = __msa_bmnz_v(p0_out, (v16u8) p0_filt8_r, flat);
2161 q0_out = __msa_bmnz_v(q0_out, (v16u8) q0_filt8_r, flat);
2162 q1_out = __msa_bmnz_v(q1_out, (v16u8) q1_filt8_r, flat);
2163 q2_out = __msa_bmnz_v(q2, (v16u8) q2_filt8_r, flat);
2169 ST_UB(flat, filter48);
2178 v16u8 flat, flat2, filter8;
2192 flat = LD_UB(filter48 + 6 * 16);
2197 VP9_FLAT5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, flat, flat2);