Lines Matching refs:stride
67 uint8_t *dst, ptrdiff_t stride)
73 ptrdiff_t stride_2x = stride << 1;
74 ptrdiff_t stride_3x = stride_2x + stride;
75 ptrdiff_t stride_4x = stride << 2;
94 DUP4_ARG2(__lasx_xvldx, src_y, stride, src_y, stride_2x, src_y, stride_3x,
104 DUP2_ARG2(__lasx_xvldx, src_x, stride, src_x, stride_2x,
122 DUP4_ARG2(__lasx_xvldx, src_y, stride, src_y, stride_2x,
166 DUP2_ARG2(__lasx_xvldx, dst, stride, dst, stride_2x, out1, out2);
178 __lasx_xvstelm_d(tmp0, dst + stride, 0, 1);
183 __lasx_xvstelm_d(tmp0, dst + stride, 8, 3);
198 uint8_t *dst, ptrdiff_t stride)
204 ptrdiff_t stride_2x = stride << 1;
205 ptrdiff_t stride_3x = stride_2x + stride;
206 ptrdiff_t stride_4x = stride << 2;
225 DUP4_ARG2(__lasx_xvldx, src_y, stride, src_y, stride_2x, src_y, stride_3x,
235 DUP2_ARG2(__lasx_xvldx, src_x, stride, src_x, stride_2x,
253 DUP4_ARG2(__lasx_xvldx, src_y, stride, src_y, stride_2x,
297 __lasx_xvstelm_d(tmp0, dst + stride, 0, 1);
302 __lasx_xvstelm_d(tmp0, dst + stride, 8, 3);
317 put_pixels8_8_inline_asm(uint8_t *dst, const uint8_t *src, ptrdiff_t stride)
322 "slli.d %[stride_2], %[stride], 1 \n\t"
323 "add.d %[stride_3], %[stride_2], %[stride] \n\t"
326 "ldx.d %[tmp1], %[src], %[stride] \n\t"
331 "ldx.d %[tmp5], %[src], %[stride] \n\t"
336 "stx.d %[tmp1], %[dst], %[stride] \n\t"
341 "stx.d %[tmp5], %[dst], %[stride] \n\t"
351 : [stride]"r"(stride)
357 * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
358 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
360 avg_pixels8_8_lsx(uint8_t *dst, const uint8_t *src, ptrdiff_t stride)
366 "slli.d %[stride_2], %[stride], 1 \n\t"
367 "add.d %[stride_3], %[stride_2], %[stride] \n\t"
370 "vldx $vr1, %[src], %[stride] \n\t"
375 "vldx $vr5, %[src], %[stride] \n\t"
380 "vldx $vr9, %[tmp], %[stride] \n\t"
385 "vldx $vr13, %[tmp], %[stride] \n\t"
399 "add.d %[dst], %[dst], %[stride] \n\t"
401 "add.d %[dst], %[dst], %[stride] \n\t"
403 "add.d %[dst], %[dst], %[stride] \n\t"
405 "add.d %[dst], %[dst], %[stride] \n\t"
407 "add.d %[dst], %[dst], %[stride] \n\t"
409 "add.d %[dst], %[dst], %[stride] \n\t"
411 "add.d %[dst], %[dst], %[stride] \n\t"
416 : [stride]"r"(stride)
422 * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
423 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
486 * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
487 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
574 put_pixels16_8_lsx(uint8_t *dst, const uint8_t *src, ptrdiff_t stride)
578 "slli.d %[stride_2], %[stride], 1 \n\t"
579 "add.d %[stride_3], %[stride_2], %[stride] \n\t"
582 "vldx $vr1, %[src], %[stride] \n\t"
587 "vldx $vr5, %[src], %[stride] \n\t"
593 "vstx $vr1, %[dst], %[stride] \n\t"
598 "vstx $vr5, %[dst], %[stride] \n\t"
604 "vldx $vr1, %[src], %[stride] \n\t"
609 "vldx $vr5, %[src], %[stride] \n\t"
614 "vstx $vr1, %[dst], %[stride] \n\t"
619 "vstx $vr5, %[dst], %[stride] \n\t"
625 : [stride]"r"(stride)
631 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
632 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
634 avg_pixels16_8_lsx(uint8_t *dst, const uint8_t *src, ptrdiff_t stride)
640 "slli.d %[stride_2], %[stride], 1 \n\t"
641 "add.d %[stride_3], %[stride_2], %[stride] \n\t"
644 "vldx $vr1, %[src], %[stride] \n\t"
649 "vldx $vr5, %[src], %[stride] \n\t"
655 "vldx $vr9, %[tmp], %[stride] \n\t"
660 "vldx $vr13, %[tmp], %[stride] \n\t"
675 "vstx $vr1, %[dst], %[stride] \n\t"
680 "vstx $vr5, %[dst], %[stride] \n\t"
687 "vldx $vr1, %[src], %[stride] \n\t"
692 "vldx $vr5, %[src], %[stride] \n\t"
697 "vldx $vr9, %[tmp], %[stride] \n\t"
702 "vldx $vr13, %[tmp], %[stride] \n\t"
716 "vstx $vr1, %[dst], %[stride] \n\t"
721 "vstx $vr5, %[dst], %[stride] \n\t"
727 : [stride]"r"(stride)
733 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
734 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
837 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
838 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
1470 ptrdiff_t stride)
1474 put_pixels8_8_inline_asm(dst, src, stride);
1478 ptrdiff_t stride)
1482 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1483 /* in qpel8, the stride of half and height of block is 8 */
1484 put_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1488 ptrdiff_t stride)
1490 put_h264_qpel8_h_lowpass_lasx(dst, src, stride, stride);
1494 ptrdiff_t stride)
1498 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1499 put_pixels8_l2_8_lsx(dst, src+1, half, stride, stride);
1503 ptrdiff_t stride)
1507 put_h264_qpel8_v_lowpass_lasx(half, (uint8_t*)src, 8, stride);
1508 put_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1512 ptrdiff_t stride)
1517 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1518 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src, 8, stride);
1519 put_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1523 ptrdiff_t stride)
1529 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1530 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1531 put_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1535 ptrdiff_t stride)
1540 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1541 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src + 1, 8, stride);
1542 put_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1546 ptrdiff_t stride)
1548 put_h264_qpel8_v_lowpass_lasx(dst, (uint8_t*)src, stride, stride);
1552 ptrdiff_t stride)
1558 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1559 put_h264_qpel8_v_lowpass_lasx(halfH, (uint8_t*)src, 8, stride);
1560 put_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1564 ptrdiff_t stride)
1566 put_h264_qpel8_hv_lowpass_lasx(dst, src, stride, stride);
1570 ptrdiff_t stride)
1576 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1577 put_h264_qpel8_v_lowpass_lasx(halfH, (uint8_t*)src + 1, 8, stride);
1578 put_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1582 ptrdiff_t stride)
1586 put_h264_qpel8_v_lowpass_lasx(half, (uint8_t*)src, 8, stride);
1587 put_pixels8_l2_8_lsx(dst, src + stride, half, stride, stride);
1591 ptrdiff_t stride)
1596 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1597 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src, 8, stride);
1598 put_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1602 ptrdiff_t stride)
1608 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1609 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1610 put_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1614 ptrdiff_t stride)
1619 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1620 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src + 1, 8, stride);
1621 put_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1625 ptrdiff_t stride)
1629 avg_pixels8_8_lsx(dst, src, stride);
1633 ptrdiff_t stride)
1637 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1638 avg_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1642 ptrdiff_t stride)
1644 avg_h264_qpel8_h_lowpass_lasx(dst, src, stride, stride);
1648 ptrdiff_t stride)
1652 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1653 avg_pixels8_l2_8_lsx(dst, src+1, half, stride, stride);
1657 ptrdiff_t stride)
1662 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1663 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src, 8, stride);
1664 avg_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1668 ptrdiff_t stride)
1674 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1675 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1676 avg_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1680 ptrdiff_t stride)
1685 put_h264_qpel8_h_lowpass_lasx(halfH, src, 8, stride);
1686 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src + 1, 8, stride);
1687 avg_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1691 ptrdiff_t stride)
1693 avg_h264_qpel8_v_lowpass_lasx(dst, (uint8_t*)src, stride, stride);
1697 ptrdiff_t stride)
1703 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1704 put_h264_qpel8_v_lowpass_lasx(halfH, (uint8_t*)src, 8, stride);
1705 avg_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1709 ptrdiff_t stride)
1711 avg_h264_qpel8_hv_lowpass_lasx(dst, src, stride, stride);
1715 ptrdiff_t stride)
1721 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1722 put_h264_qpel8_v_lowpass_lasx(halfH, (uint8_t*)src + 1, 8, stride);
1723 avg_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1727 ptrdiff_t stride)
1732 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1733 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src, 8, stride);
1734 avg_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1738 ptrdiff_t stride)
1744 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1745 put_h264_qpel8_hv_lowpass_lasx(halfHV, src, 8, stride);
1746 avg_pixels8_l2_8_lsx(dst, halfH, halfHV, stride, 8);
1750 ptrdiff_t stride)
1755 put_h264_qpel8_h_lowpass_lasx(halfH, src + stride, 8, stride);
1756 put_h264_qpel8_v_lowpass_lasx(halfV, (uint8_t*)src + 1, 8, stride);
1757 avg_pixels8_l2_8_lsx(dst, halfH, halfV, stride, 8);
1761 ptrdiff_t stride)
1765 put_pixels16_8_lsx(dst, src, stride);
1769 ptrdiff_t stride)
1773 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1774 put_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1778 ptrdiff_t stride)
1780 put_h264_qpel16_h_lowpass_lasx(dst, src, stride, stride);
1784 ptrdiff_t stride)
1788 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1789 put_pixels16_l2_8_lsx(dst, src+1, half, stride, stride);
1793 ptrdiff_t stride)
1797 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1798 put_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1802 ptrdiff_t stride)
1804 avc_luma_hv_qrt_16x16_lasx((uint8_t*)src - 2, (uint8_t*)src - (stride * 2),
1805 dst, stride);
1809 ptrdiff_t stride)
1815 put_h264_qpel16_h_lowpass_lasx(halfH, src, 16, stride);
1816 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1817 put_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1821 ptrdiff_t stride)
1823 avc_luma_hv_qrt_16x16_lasx((uint8_t*)src - 2, (uint8_t*)src - (stride * 2) + 1,
1824 dst, stride);
1828 ptrdiff_t stride)
1830 put_h264_qpel16_v_lowpass_lasx(dst, src, stride, stride);
1834 ptrdiff_t stride)
1840 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1841 put_h264_qpel16_v_lowpass_lasx(halfH, src, 16, stride);
1842 put_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1846 ptrdiff_t stride)
1848 put_h264_qpel16_hv_lowpass_lasx(dst, src, stride, stride);
1852 ptrdiff_t stride)
1858 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1859 put_h264_qpel16_v_lowpass_lasx(halfH, src + 1, 16, stride);
1860 put_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1864 ptrdiff_t stride)
1868 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1869 put_pixels16_l2_8_lsx(dst, src+stride, half, stride, stride);
1873 ptrdiff_t stride)
1875 avc_luma_hv_qrt_16x16_lasx((uint8_t*)src + stride - 2, (uint8_t*)src - (stride * 2),
1876 dst, stride);
1880 ptrdiff_t stride)
1886 put_h264_qpel16_h_lowpass_lasx(halfH, src + stride, 16, stride);
1887 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1888 put_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1892 ptrdiff_t stride)
1894 avc_luma_hv_qrt_16x16_lasx((uint8_t*)src + stride - 2,
1895 (uint8_t*)src - (stride * 2) + 1, dst, stride);
1899 ptrdiff_t stride)
1903 avg_pixels16_8_lsx(dst, src, stride);
1907 ptrdiff_t stride)
1911 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1912 avg_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1916 ptrdiff_t stride)
1918 avg_h264_qpel16_h_lowpass_lasx(dst, src, stride, stride);
1922 ptrdiff_t stride)
1926 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1927 avg_pixels16_l2_8_lsx(dst, src+1, half, stride, stride);
1931 ptrdiff_t stride)
1935 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1936 avg_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1940 ptrdiff_t stride)
1943 (uint8_t*)src - (stride * 2),
1944 dst, stride);
1948 ptrdiff_t stride)
1954 put_h264_qpel16_h_lowpass_lasx(halfH, src, 16, stride);
1955 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1956 avg_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1960 ptrdiff_t stride)
1963 (uint8_t*)src - (stride * 2) + 1,
1964 dst, stride);
1968 ptrdiff_t stride)
1970 avg_h264_qpel16_v_lowpass_lasx(dst, src, stride, stride);
1974 ptrdiff_t stride)
1980 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1981 put_h264_qpel16_v_lowpass_lasx(halfH, src, 16, stride);
1982 avg_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
1986 ptrdiff_t stride)
1988 avg_h264_qpel16_hv_lowpass_lasx(dst, src, stride, stride);
1992 ptrdiff_t stride)
1998 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
1999 put_h264_qpel16_v_lowpass_lasx(halfH, src + 1, 16, stride);
2000 avg_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
2004 ptrdiff_t stride)
2008 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
2009 avg_pixels16_l2_8_lsx(dst, src + stride, half, stride, stride);
2013 ptrdiff_t stride)
2015 avc_luma_hv_qrt_and_aver_dst_16x16_lasx((uint8_t*)src + stride - 2,
2016 (uint8_t*)src - (stride * 2),
2017 dst, stride);
2021 ptrdiff_t stride)
2027 put_h264_qpel16_h_lowpass_lasx(halfH, src + stride, 16, stride);
2028 put_h264_qpel16_hv_lowpass_lasx(halfHV, src, 16, stride);
2029 avg_pixels16_l2_8_lsx(dst, halfH, halfHV, stride, 16);
2033 ptrdiff_t stride)
2035 avc_luma_hv_qrt_and_aver_dst_16x16_lasx((uint8_t*)src + stride - 2,
2036 (uint8_t*)src - (stride * 2) + 1,
2037 dst, stride);