Lines Matching refs:src
85 uint8_t *src = data - 4;
95 uint8_t *src_tmp = src + img_width_8x;
99 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
100 src, img_width_3x, row0, row1, row2, row3);
101 src += img_width_4x;
102 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
103 src, img_width_3x, row4, row5, row6, row7);
104 src -= img_width_4x;
216 __lasx_xvstelm_d(row4, src, 0, 0);
217 __lasx_xvstelm_d(row4, src + img_width, 0, 1);
218 src += img_width_2x;
219 __lasx_xvstelm_d(row4, src, 0, 2);
220 __lasx_xvstelm_d(row4, src + img_width, 0, 3);
221 src += img_width_2x;
222 __lasx_xvstelm_d(row5, src, 0, 0);
223 __lasx_xvstelm_d(row5, src + img_width, 0, 1);
224 src += img_width_2x;
225 __lasx_xvstelm_d(row5, src, 0, 2);
226 __lasx_xvstelm_d(row5, src + img_width, 0, 3);
227 src += img_width_2x;
228 __lasx_xvstelm_d(row6, src, 0, 0);
229 __lasx_xvstelm_d(row6, src + img_width, 0, 1);
230 src += img_width_2x;
231 __lasx_xvstelm_d(row6, src, 0, 2);
232 __lasx_xvstelm_d(row6, src + img_width, 0, 3);
233 src += img_width_2x;
234 __lasx_xvstelm_d(row7, src, 0, 0);
235 __lasx_xvstelm_d(row7, src + img_width, 0, 1);
236 src += img_width_2x;
237 __lasx_xvstelm_d(row7, src, 0, 2);
238 __lasx_xvstelm_d(row7, src + img_width, 0, 3);
384 uint8_t *src = data - 2;
395 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
396 src, img_width_3x, row0, row1, row2, row3);
397 src += img_width_4x;
398 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
399 src, img_width_3x, row4, row5, row6, row7);
400 src -= img_width_4x;
454 src = data - 1;
455 __lasx_xvstelm_h(p0_org, src, 0, 0);
456 src += img_width;
457 __lasx_xvstelm_h(p0_org, src, 0, 1);
458 src += img_width;
459 __lasx_xvstelm_h(p0_org, src, 0, 2);
460 src += img_width;
461 __lasx_xvstelm_h(p0_org, src, 0, 3);
462 src += img_width;
463 __lasx_xvstelm_h(p0_org, src, 0, 4);
464 src += img_width;
465 __lasx_xvstelm_h(p0_org, src, 0, 5);
466 src += img_width;
467 __lasx_xvstelm_h(p0_org, src, 0, 6);
468 src += img_width;
469 __lasx_xvstelm_h(p0_org, src, 0, 7);
590 uint8_t *src = data - 4;
600 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
601 src, img_width_3x, row0, row1, row2, row3);
602 src += img_width_4x;
603 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
604 src, img_width_3x, row4, row5, row6, row7);
605 src += img_width_4x;
606 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
607 src, img_width_3x, row8, row9, row10, row11);
608 src += img_width_4x;
609 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
610 src, img_width_3x, row12, row13, row14, row15);
611 src += img_width_4x;
732 src = data - 4;
733 __lasx_xvstelm_d(row4, src, 0, 0);
734 __lasx_xvstelm_d(row4, src + img_width, 0, 1);
735 src += img_width_2x;
736 __lasx_xvstelm_d(row4, src, 0, 2);
737 __lasx_xvstelm_d(row4, src + img_width, 0, 3);
738 src += img_width_2x;
739 __lasx_xvstelm_d(row5, src, 0, 0);
740 __lasx_xvstelm_d(row5, src + img_width, 0, 1);
741 src += img_width_2x;
742 __lasx_xvstelm_d(row5, src, 0, 2);
743 __lasx_xvstelm_d(row5, src + img_width, 0, 3);
744 src += img_width_2x;
745 __lasx_xvstelm_d(row6, src, 0, 0);
746 __lasx_xvstelm_d(row6, src + img_width, 0, 1);
747 src += img_width_2x;
748 __lasx_xvstelm_d(row6, src, 0, 2);
749 __lasx_xvstelm_d(row6, src + img_width, 0, 3);
750 src += img_width_2x;
751 __lasx_xvstelm_d(row7, src, 0, 0);
752 __lasx_xvstelm_d(row7, src + img_width, 0, 1);
753 src += img_width_2x;
754 __lasx_xvstelm_d(row7, src, 0, 2);
755 __lasx_xvstelm_d(row7, src + img_width, 0, 3);
765 uint8_t *src = data - img_width_2x;
771 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x,
772 src, img_width_3x, p1_org, p0_org, q0_org, q1_org);
789 __m256i p2_org = __lasx_xvldx(src, -img_width);
811 __m256i p3_org = __lasx_xvldx(src, -img_width_2x);
827 __lasx_xvst(p1_org, src, 0);
828 __lasx_xvst(p2_org, src - img_width, 0);
883 uint8_t *src = data - 2;
894 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x, src,
896 src += img_width_4x;
897 DUP4_ARG2(__lasx_xvldx, src, 0, src, img_width, src, img_width_2x, src,
942 src = data - 1;
943 __lasx_xvstelm_h(p0_org, src, 0, 0);
944 src += img_width;
945 __lasx_xvstelm_h(p0_org, src, 0, 1);
946 src += img_width;
947 __lasx_xvstelm_h(p0_org, src, 0, 2);
948 src += img_width;
949 __lasx_xvstelm_h(p0_org, src, 0, 3);
950 src += img_width;
951 __lasx_xvstelm_h(p0_org, src, 0, 4);
952 src += img_width;
953 __lasx_xvstelm_h(p0_org, src, 0, 5);
954 src += img_width;
955 __lasx_xvstelm_h(p0_org, src, 0, 6);
956 src += img_width;
957 __lasx_xvstelm_h(p0_org, src, 0, 7);
1004 void ff_biweight_h264_pixels16_8_lasx(uint8_t *dst, uint8_t *src,
1029 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1030 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1031 src += stride_4x;
1032 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1033 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1034 src += stride_4x;
1101 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1102 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1103 src += stride_4x;
1104 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1105 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1106 src += stride_4x;
1173 static void avc_biwgt_8x4_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1193 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1194 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1216 static void avc_biwgt_8x8_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1238 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1239 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1240 src += stride_4x;
1243 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1244 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1282 static void avc_biwgt_8x16_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1304 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1305 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1306 src += stride_4x;
1309 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1310 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1311 src += stride_4x;
1314 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1315 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1316 src += stride_4x;
1319 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1320 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1391 void ff_biweight_h264_pixels8_8_lasx(uint8_t *dst, uint8_t *src,
1397 avc_biwgt_8x4_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1400 avc_biwgt_8x8_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1403 avc_biwgt_8x16_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1408 static void avc_biwgt_4x2_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1426 DUP2_ARG2(__lasx_xvldx, src, 0, src, stride, tmp0, tmp1);
1440 static void avc_biwgt_4x4_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1460 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1461 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1482 static void avc_biwgt_4x8_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1503 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1504 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1505 src += stride_4x;
1506 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1507 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1542 void ff_biweight_h264_pixels4_8_lasx(uint8_t *dst, uint8_t *src,
1548 avc_biwgt_4x2_lasx(src, dst, stride, log2_denom, weight_src,
1551 avc_biwgt_4x4_lasx(src, dst, stride, log2_denom, weight_src,
1554 avc_biwgt_4x8_lasx(src, dst, stride, log2_denom, weight_src,
1559 void ff_weight_h264_pixels16_8_lasx(uint8_t *src, ptrdiff_t stride,
1579 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1580 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1581 src += stride_4x;
1582 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1583 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1584 src -= stride_4x;
1619 __lasx_xvstelm_d(src0_l, src, 0, 0);
1620 __lasx_xvstelm_d(src0_h, src, 8, 0);
1621 src += stride;
1622 __lasx_xvstelm_d(src0_l, src, 0, 2);
1623 __lasx_xvstelm_d(src0_h, src, 8, 2);
1624 src += stride;
1625 __lasx_xvstelm_d(src1_l, src, 0, 0);
1626 __lasx_xvstelm_d(src1_h, src, 8, 0);
1627 src += stride;
1628 __lasx_xvstelm_d(src1_l, src, 0, 2);
1629 __lasx_xvstelm_d(src1_h, src, 8, 2);
1630 src += stride;
1631 __lasx_xvstelm_d(src2_l, src, 0, 0);
1632 __lasx_xvstelm_d(src2_h, src, 8, 0);
1633 src += stride;
1634 __lasx_xvstelm_d(src2_l, src, 0, 2);
1635 __lasx_xvstelm_d(src2_h, src, 8, 2);
1636 src += stride;
1637 __lasx_xvstelm_d(src3_l, src, 0, 0);
1638 __lasx_xvstelm_d(src3_h, src, 8, 0);
1639 src += stride;
1640 __lasx_xvstelm_d(src3_l, src, 0, 2);
1641 __lasx_xvstelm_d(src3_h, src, 8, 2);
1642 src += stride;
1645 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1646 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1647 src += stride_4x;
1648 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1649 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1650 src -= stride_4x;
1685 __lasx_xvstelm_d(src0_l, src, 0, 0);
1686 __lasx_xvstelm_d(src0_h, src, 8, 0);
1687 src += stride;
1688 __lasx_xvstelm_d(src0_l, src, 0, 2);
1689 __lasx_xvstelm_d(src0_h, src, 8, 2);
1690 src += stride;
1691 __lasx_xvstelm_d(src1_l, src, 0, 0);
1692 __lasx_xvstelm_d(src1_h, src, 8, 0);
1693 src += stride;
1694 __lasx_xvstelm_d(src1_l, src, 0, 2);
1695 __lasx_xvstelm_d(src1_h, src, 8, 2);
1696 src += stride;
1697 __lasx_xvstelm_d(src2_l, src, 0, 0);
1698 __lasx_xvstelm_d(src2_h, src, 8, 0);
1699 src += stride;
1700 __lasx_xvstelm_d(src2_l, src, 0, 2);
1701 __lasx_xvstelm_d(src2_h, src, 8, 2);
1702 src += stride;
1703 __lasx_xvstelm_d(src3_l, src, 0, 0);
1704 __lasx_xvstelm_d(src3_h, src, 8, 0);
1705 src += stride;
1706 __lasx_xvstelm_d(src3_l, src, 0, 2);
1707 __lasx_xvstelm_d(src3_h, src, 8, 2);
1711 static void avc_wgt_8x4_lasx(uint8_t *src, ptrdiff_t stride,
1728 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1729 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1744 __lasx_xvstelm_d(src0, src, 0, 0);
1745 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1746 __lasx_xvstelm_d(src0, src + stride_2x, 0, 2);
1747 __lasx_xvstelm_d(src0, src + stride_3x, 0, 3);
1750 static void avc_wgt_8x8_lasx(uint8_t *src, ptrdiff_t stride, int32_t log2_denom,
1756 uint8_t* src_tmp = src;
1794 __lasx_xvstelm_d(src0, src, 0, 0);
1795 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1796 __lasx_xvstelm_d(src0, src + stride_2x, 0, 2);
1797 __lasx_xvstelm_d(src0, src + stride_3x, 0, 3);
1798 src += stride_4x;
1799 __lasx_xvstelm_d(src1, src, 0, 0);
1800 __lasx_xvstelm_d(src1, src + stride, 0, 1);
1801 __lasx_xvstelm_d(src1, src + stride_2x, 0, 2);
1802 __lasx_xvstelm_d(src1, src + stride_3x, 0, 3);
1805 static void avc_wgt_8x16_lasx(uint8_t *src, ptrdiff_t stride,
1814 uint8_t* src_tmp = src;
1882 __lasx_xvstelm_d(src0, src, 0, 0);
1883 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1884 __lasx_xvstelm_d(src0, src + stride_2x, 0, 2);
1885 __lasx_xvstelm_d(src0, src + stride_3x, 0, 3);
1886 src += stride_4x;
1887 __lasx_xvstelm_d(src1, src, 0, 0);
1888 __lasx_xvstelm_d(src1, src + stride, 0, 1);
1889 __lasx_xvstelm_d(src1, src + stride_2x, 0, 2);
1890 __lasx_xvstelm_d(src1, src + stride_3x, 0, 3);
1891 src += stride_4x;
1892 __lasx_xvstelm_d(src2, src, 0, 0);
1893 __lasx_xvstelm_d(src2, src + stride, 0, 1);
1894 __lasx_xvstelm_d(src2, src + stride_2x, 0, 2);
1895 __lasx_xvstelm_d(src2, src + stride_3x, 0, 3);
1896 src += stride_4x;
1897 __lasx_xvstelm_d(src3, src, 0, 0);
1898 __lasx_xvstelm_d(src3, src + stride, 0, 1);
1899 __lasx_xvstelm_d(src3, src + stride_2x, 0, 2);
1900 __lasx_xvstelm_d(src3, src + stride_3x, 0, 3);
1903 void ff_weight_h264_pixels8_8_lasx(uint8_t *src, ptrdiff_t stride,
1908 avc_wgt_8x4_lasx(src, stride, log2_denom, weight_src, offset);
1910 avc_wgt_8x8_lasx(src, stride, log2_denom, weight_src, offset);
1912 avc_wgt_8x16_lasx(src, stride, log2_denom, weight_src, offset);
1916 static void avc_wgt_4x2_lasx(uint8_t *src, ptrdiff_t stride,
1930 DUP2_ARG2(__lasx_xvldx, src, 0, src, stride, tmp0, tmp1);
1937 __lasx_xvstelm_w(src0, src, 0, 0);
1938 __lasx_xvstelm_w(src0, src + stride, 0, 1);
1941 static void avc_wgt_4x4_lasx(uint8_t *src, ptrdiff_t stride,
1957 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1958 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1966 __lasx_xvstelm_w(src0, src, 0, 0);
1967 __lasx_xvstelm_w(src0, src + stride, 0, 1);
1968 __lasx_xvstelm_w(src0, src + stride_2x, 0, 4);
1969 __lasx_xvstelm_w(src0, src + stride_3x, 0, 5);
1972 static void avc_wgt_4x8_lasx(uint8_t *src, ptrdiff_t stride,
1990 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1991 src, stride_3x, tmp0, tmp1, tmp2, tmp3);
1992 src += stride_4x;
1993 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1994 src, stride_3x, tmp4, tmp5, tmp6, tmp7);
1995 src -= stride_4x;
2010 __lasx_xvstelm_w(src0_l, src, 0, 0);
2011 __lasx_xvstelm_w(src0_l, src + stride, 0, 1);
2012 __lasx_xvstelm_w(src0_h, src + stride_2x, 0, 0);
2013 __lasx_xvstelm_w(src0_h, src + stride_3x, 0, 1);
2014 src += stride_4x;
2015 __lasx_xvstelm_w(src0_l, src, 0, 4);
2016 __lasx_xvstelm_w(src0_l, src + stride, 0, 5);
2017 __lasx_xvstelm_w(src0_h, src + stride_2x, 0, 4);
2018 __lasx_xvstelm_w(src0_h, src + stride_3x, 0, 5);
2021 void ff_weight_h264_pixels4_8_lasx(uint8_t *src, ptrdiff_t stride,
2026 avc_wgt_4x2_lasx(src, stride, log2_denom, weight_src, offset);
2028 avc_wgt_4x4_lasx(src, stride, log2_denom, weight_src, offset);
2030 avc_wgt_4x8_lasx(src, stride, log2_denom, weight_src, offset);