Lines Matching refs:zero
316 __m128i zero = __lsx_vldi(0);
334 flat = __lsx_vilvl_d(zero, flat);
336 /* if flat is zero for all pixels, then no need to calculate other filter */
343 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
345 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
351 DUP4_ARG2(__lsx_vpickev_b, zero, p2_filter8, zero, p1_filter8,
352 zero, p0_filter8, zero, q0_filter8, p2_filter8,
354 DUP2_ARG2(__lsx_vpickev_b, zero, q1_filter8, zero, q2_filter8,
391 __m128i zero = __lsx_vldi(0);
419 /* if flat is zero for all pixels, then no need to calculate other filter */
426 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
428 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
433 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
435 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
479 __m128i zero = __lsx_vldi(0);
507 flat = __lsx_vilvl_d(zero, flat);
509 /* if flat is zero for all pixels, then no need to calculate other filter */
516 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
518 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
561 __m128i zero = { 0 };
589 flat = __lsx_vilvh_d(flat, zero);
591 /* if flat is zero for all pixels, then no need to calculate other filter */
598 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
600 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
647 __m128i zero = __lsx_vldi(0);
667 /* if flat is zero for all pixels, then no need to calculate other filter */
675 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
677 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
682 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
684 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
726 __m128i zero = __lsx_vldi(0);
755 /* if flat2 is zero for all pixels, then no need to calculate other filter */
770 p7_l_in = (v8u16)__lsx_vilvl_b(zero, p7);
771 p6_l_in = (v8u16)__lsx_vilvl_b(zero, p6);
772 p5_l_in = (v8u16)__lsx_vilvl_b(zero, p5);
773 p4_l_in = (v8u16)__lsx_vilvl_b(zero, p4);
774 p3_l_in = (v8u16)__lsx_vilvl_b(zero, p3);
775 p2_l_in = (v8u16)__lsx_vilvl_b(zero, p2);
776 p1_l_in = (v8u16)__lsx_vilvl_b(zero, p1);
777 p0_l_in = (v8u16)__lsx_vilvl_b(zero, p0);
779 q0_l_in = (v8u16)__lsx_vilvl_b(zero, q0);
795 p7_h_in = (v8u16)__lsx_vilvh_b(zero, p7);
796 p6_h_in = (v8u16)__lsx_vilvh_b(zero, p6);
797 p5_h_in = (v8u16)__lsx_vilvh_b(zero, p5);
798 p4_h_in = (v8u16)__lsx_vilvh_b(zero, p4);
800 p3_h_in = (v8u16)__lsx_vilvh_b(zero, p3);
801 p2_h_in = (v8u16)__lsx_vilvh_b(zero, p2);
802 p1_h_in = (v8u16)__lsx_vilvh_b(zero, p1);
803 p0_h_in = (v8u16)__lsx_vilvh_b(zero, p0);
804 q0_h_in = (v8u16)__lsx_vilvh_b(zero, q0);
826 q1_l_in = (v8u16)__lsx_vilvl_b(zero, q1);
833 q1_h_in = (v8u16)__lsx_vilvh_b(zero, q1);
846 q2_l_in = (v8u16)__lsx_vilvl_b(zero, q2);
853 q2_h_in = (v8u16)__lsx_vilvh_b(zero, q2);
866 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
873 q3_h_in = (v8u16)__lsx_vilvh_b(zero, q3);
886 q4_l_in = (v8u16)__lsx_vilvl_b(zero, q4);
894 q4_h_in = (v8u16)__lsx_vilvh_b(zero, q4);
907 q5_l_in = (v8u16)__lsx_vilvl_b(zero, q5);
915 q5_h_in = (v8u16)__lsx_vilvh_b(zero, q5);
928 q6_l_in = (v8u16)__lsx_vilvl_b(zero, q6);
936 q6_h_in = (v8u16)__lsx_vilvh_b(zero, q6);
949 q7_l_in = (v8u16)__lsx_vilvl_b(zero, q7);
957 q7_h_in = (v8u16)__lsx_vilvh_b(zero, q7);
1106 __m128i zero = __lsx_vldi(0);
1134 flat = __lsx_vilvl_d(zero, flat);
1136 /* if flat is zero for all pixels, then no need to calculate other filter */
1144 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
1146 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1153 DUP4_ARG2(__lsx_vpickev_b, zero, p2_filter8, zero, p1_filter8,
1154 zero, p0_filter8, zero, q0_filter8, p2_filter8,
1156 DUP2_ARG2(__lsx_vpickev_b, zero, q1_filter8, zero, q2_filter8,
1175 /* if flat2 is zero for all pixels, then no need to calculate other filter */
1191 DUP4_ARG2(__lsx_vilvl_b, zero, p7, zero, p6, zero, p5, zero, p4,
1193 DUP4_ARG2(__lsx_vilvl_b, zero, q4, zero, q5, zero, q6, zero, q7,
1219 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1239 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1259 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1279 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1299 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1319 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1339 DUP2_ARG2(__lsx_vpickev_b, zero, p0_filter16, zero,
1492 __m128i zero = __lsx_vldi(0);
1519 flat = __lsx_vilvl_d(zero, flat);
1521 /* if flat is zero for all pixels, then no need to calculate other filter */
1539 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
1541 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1614 __m128i zero = __lsx_vldi(0);
1658 /* if flat is zero for all pixels, then no need to calculate other filter */
1688 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
1690 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1695 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
1697 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
1796 __m128i zero = __lsx_vldi(0);
1840 flat = __lsx_vilvl_d(zero, flat);
1842 /* if flat is zero for all pixels, then no need to calculate other filter */
1872 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
1874 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
1971 __m128i zero = __lsx_vldi(0);
2015 flat = __lsx_vilvh_d(flat, zero);
2017 /* if flat is zero for all pixels, then no need to calculate other filter */
2047 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
2049 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
2275 __m128i zero = __lsx_vldi(0);
2295 flat = __lsx_vilvl_d(zero, flat);
2297 /* if flat is zero for all pixels, then no need to calculate other filter */
2321 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
2323 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
2360 __m128i zero = __lsx_vldi(0);
2384 /* if flat2 is zero for all pixels, then no need to calculate other filter */
2425 p7_l_in = (v8u16)__lsx_vilvl_b(zero, p7);
2426 p6_l_in = (v8u16)__lsx_vilvl_b(zero, p6);
2427 p5_l_in = (v8u16)__lsx_vilvl_b(zero, p5);
2428 p4_l_in = (v8u16)__lsx_vilvl_b(zero, p4);
2429 p3_l_in = (v8u16)__lsx_vilvl_b(zero, p3);
2430 p2_l_in = (v8u16)__lsx_vilvl_b(zero, p2);
2431 p1_l_in = (v8u16)__lsx_vilvl_b(zero, p1);
2432 p0_l_in = (v8u16)__lsx_vilvl_b(zero, p0);
2433 q0_l_in = (v8u16)__lsx_vilvl_b(zero, q0);
2454 q1_l_in = (v8u16)__lsx_vilvl_b(zero, q1);
2466 q2_l_in = (v8u16)__lsx_vilvl_b(zero, q2);
2478 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
2490 q4_l_in = (v8u16)__lsx_vilvl_b(zero, q4);
2503 q5_l_in = (v8u16)__lsx_vilvl_b(zero, q5);
2516 q6_l_in = (v8u16)__lsx_vilvl_b(zero, q6);
2529 q7_l_in = (v8u16)__lsx_vilvl_b(zero, q7);
2656 __m128i zero = __lsx_vldi(0);
2676 /* if flat is zero for all pixels, then no need to calculate other filter */
2708 DUP4_ARG2(__lsx_vilvl_b, zero, p3, zero, p2, zero, p1, zero, p0,
2710 DUP4_ARG2(__lsx_vilvl_b, zero, q0, zero, q1, zero, q2, zero, q3,
2714 DUP4_ARG2(__lsx_vilvh_b, zero, p3, zero, p2, zero, p1, zero, p0,
2716 DUP4_ARG2(__lsx_vilvh_b, zero, q0, zero, q1, zero, q2, zero, q3,
2753 __m128i zero = __lsx_vldi(0);
2779 /* if flat2 is zero for all pixels, then no need to calculate other filter */
2849 p7_l_in = (v8u16)__lsx_vilvl_b(zero, p7);
2850 p6_l_in = (v8u16)__lsx_vilvl_b(zero, p6);
2851 p5_l_in = (v8u16)__lsx_vilvl_b(zero, p5);
2852 p4_l_in = (v8u16)__lsx_vilvl_b(zero, p4);
2853 p3_l_in = (v8u16)__lsx_vilvl_b(zero, p3);
2854 p2_l_in = (v8u16)__lsx_vilvl_b(zero, p2);
2855 p1_l_in = (v8u16)__lsx_vilvl_b(zero, p1);
2856 p0_l_in = (v8u16)__lsx_vilvl_b(zero, p0);
2857 q0_l_in = (v8u16)__lsx_vilvl_b(zero, q0);
2872 p7_h_in = (v8u16)__lsx_vilvh_b(zero, p7);
2873 p6_h_in = (v8u16)__lsx_vilvh_b(zero, p6);
2874 p5_h_in = (v8u16)__lsx_vilvh_b(zero, p5);
2875 p4_h_in = (v8u16)__lsx_vilvh_b(zero, p4);
2876 p3_h_in = (v8u16)__lsx_vilvh_b(zero, p3);
2877 p2_h_in = (v8u16)__lsx_vilvh_b(zero, p2);
2878 p1_h_in = (v8u16)__lsx_vilvh_b(zero, p1);
2879 p0_h_in = (v8u16)__lsx_vilvh_b(zero, p0);
2880 q0_h_in = (v8u16)__lsx_vilvh_b(zero, q0);
2900 q1_l_in = (v8u16)__lsx_vilvl_b(zero, q1);
2906 q1_h_in = (v8u16)__lsx_vilvh_b(zero, q1);
2917 q2_l_in = (v8u16)__lsx_vilvl_b(zero, q2);
2923 q2_h_in = (v8u16)__lsx_vilvh_b(zero, q2);
2934 q3_l_in = (v8u16)__lsx_vilvl_b(zero, q3);
2940 q3_h_in = (v8u16)__lsx_vilvh_b(zero, q3);
2951 q4_l_in = (v8u16)__lsx_vilvl_b(zero, q4);
2958 q4_h_in = (v8u16)__lsx_vilvh_b(zero, q4);
2969 q5_l_in = (v8u16)__lsx_vilvl_b(zero, q5);
2976 q5_h_in = (v8u16)__lsx_vilvh_b(zero, q5);
2987 q6_l_in = (v8u16)__lsx_vilvl_b(zero, q6);
2994 q6_h_in = (v8u16)__lsx_vilvh_b(zero, q6);
3005 q7_l_in = (v8u16)__lsx_vilvl_b(zero, q7);
3012 q7_h_in = (v8u16)__lsx_vilvh_b(zero, q7);