Lines Matching defs:stride
1005 ptrdiff_t stride, int height,
1015 int stride_2x = stride << 1;
1016 int stride_4x = stride << 2;
1017 int stride_3x = stride_2x + stride;
1029 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1032 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1037 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1040 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1077 dst += stride;
1080 dst += stride;
1083 dst += stride;
1086 dst += stride;
1089 dst += stride;
1092 dst += stride;
1095 dst += stride;
1098 dst += stride;
1101 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1104 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1109 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1112 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1149 dst += stride;
1152 dst += stride;
1155 dst += stride;
1158 dst += stride;
1161 dst += stride;
1164 dst += stride;
1167 dst += stride;
1173 static void avc_biwgt_8x4_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1180 ptrdiff_t stride_2x = stride << 1;
1181 ptrdiff_t stride_3x = stride_2x + stride;
1193 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1197 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1211 __lasx_xvstelm_d(dst0, dst + stride, 0, 1);
1216 static void avc_biwgt_8x8_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1223 ptrdiff_t stride_2x = stride << 1;
1224 ptrdiff_t stride_4x = stride << 2;
1225 ptrdiff_t stride_3x = stride_2x + stride;
1238 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1243 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1248 DUP2_ARG2(__lasx_xvldx, dst_tmp, stride, dst_tmp, stride_2x, tmp1, tmp2);
1253 DUP4_ARG2(__lasx_xvldx, dst_tmp, 0, dst_tmp, stride, dst_tmp, stride_2x,
1272 __lasx_xvstelm_d(dst0, dst + stride, 0, 1);
1277 __lasx_xvstelm_d(dst1, dst + stride, 0, 1);
1282 static void avc_biwgt_8x16_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1289 ptrdiff_t stride_2x = stride << 1;
1290 ptrdiff_t stride_4x = stride << 2;
1291 ptrdiff_t stride_3x = stride_2x + stride;
1304 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1309 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1314 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1319 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1324 DUP4_ARG2(__lasx_xvldx, dst_tmp, 0, dst_tmp, stride, dst_tmp, stride_2x,
1329 DUP4_ARG2(__lasx_xvldx, dst_tmp, 0, dst_tmp, stride, dst_tmp, stride_2x,
1334 DUP4_ARG2(__lasx_xvldx, dst_tmp, 0, dst_tmp, stride, dst_tmp, stride_2x,
1339 DUP4_ARG2(__lasx_xvldx, dst_tmp, 0, dst_tmp, stride, dst_tmp, stride_2x,
1371 __lasx_xvstelm_d(dst0, dst + stride, 0, 1);
1376 __lasx_xvstelm_d(dst1, dst + stride, 0, 1);
1381 __lasx_xvstelm_d(dst2, dst + stride, 0, 1);
1386 __lasx_xvstelm_d(dst3, dst + stride, 0, 1);
1392 ptrdiff_t stride, int height,
1397 avc_biwgt_8x4_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1400 avc_biwgt_8x8_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1403 avc_biwgt_8x16_lasx(src, dst, stride, log2_denom, weight_src, weight_dst,
1408 static void avc_biwgt_4x2_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1426 DUP2_ARG2(__lasx_xvldx, src, 0, src, stride, tmp0, tmp1);
1428 DUP2_ARG2(__lasx_xvldx, dst, 0, dst, stride, tmp0, tmp1);
1437 __lasx_xvstelm_w(tmp0, dst + stride, 0, 1);
1440 static void avc_biwgt_4x4_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1447 ptrdiff_t stride_2x = stride << 1;
1448 ptrdiff_t stride_3x = stride_2x + stride;
1460 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1464 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1477 __lasx_xvstelm_w(tmp0, dst + stride, 0, 1);
1482 static void avc_biwgt_4x8_lasx(uint8_t *src, uint8_t *dst, ptrdiff_t stride,
1489 ptrdiff_t stride_2x = stride << 1;
1490 ptrdiff_t stride_4x = stride << 2;
1491 ptrdiff_t stride_3x = stride_2x + stride;
1503 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1506 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1512 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1515 DUP4_ARG2(__lasx_xvldx, dst, 0, dst, stride, dst, stride_2x,
1532 __lasx_xvstelm_w(tmp0, dst + stride, 0, 1);
1537 __lasx_xvstelm_w(tmp0, dst + stride, 0, 5);
1543 ptrdiff_t stride, int height,
1548 avc_biwgt_4x2_lasx(src, dst, stride, log2_denom, weight_src,
1551 avc_biwgt_4x4_lasx(src, dst, stride, log2_denom, weight_src,
1554 avc_biwgt_4x8_lasx(src, dst, stride, log2_denom, weight_src,
1559 void ff_weight_h264_pixels16_8_lasx(uint8_t *src, ptrdiff_t stride,
1564 ptrdiff_t stride_2x = stride << 1;
1565 ptrdiff_t stride_4x = stride << 2;
1566 ptrdiff_t stride_3x = stride_2x + stride;
1579 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1582 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1621 src += stride;
1624 src += stride;
1627 src += stride;
1630 src += stride;
1633 src += stride;
1636 src += stride;
1639 src += stride;
1642 src += stride;
1645 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1648 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1687 src += stride;
1690 src += stride;
1693 src += stride;
1696 src += stride;
1699 src += stride;
1702 src += stride;
1705 src += stride;
1711 static void avc_wgt_8x4_lasx(uint8_t *src, ptrdiff_t stride,
1716 ptrdiff_t stride_2x = stride << 1;
1717 ptrdiff_t stride_3x = stride_2x + stride;
1728 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1745 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1750 static void avc_wgt_8x8_lasx(uint8_t *src, ptrdiff_t stride, int32_t log2_denom,
1757 ptrdiff_t stride_2x = stride << 1;
1758 ptrdiff_t stride_4x = stride << 2;
1759 ptrdiff_t stride_3x = stride_2x + stride;
1767 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1772 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1795 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1800 __lasx_xvstelm_d(src1, src + stride, 0, 1);
1805 static void avc_wgt_8x16_lasx(uint8_t *src, ptrdiff_t stride,
1815 ptrdiff_t stride_2x = stride << 1;
1816 ptrdiff_t stride_4x = stride << 2;
1817 ptrdiff_t stride_3x = stride_2x + stride;
1825 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1830 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1835 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1840 DUP4_ARG2(__lasx_xvldx, src_tmp, 0, src_tmp, stride, src_tmp, stride_2x,
1883 __lasx_xvstelm_d(src0, src + stride, 0, 1);
1888 __lasx_xvstelm_d(src1, src + stride, 0, 1);
1893 __lasx_xvstelm_d(src2, src + stride, 0, 1);
1898 __lasx_xvstelm_d(src3, src + stride, 0, 1);
1903 void ff_weight_h264_pixels8_8_lasx(uint8_t *src, ptrdiff_t stride,
1908 avc_wgt_8x4_lasx(src, stride, log2_denom, weight_src, offset);
1910 avc_wgt_8x8_lasx(src, stride, log2_denom, weight_src, offset);
1912 avc_wgt_8x16_lasx(src, stride, log2_denom, weight_src, offset);
1916 static void avc_wgt_4x2_lasx(uint8_t *src, ptrdiff_t stride,
1930 DUP2_ARG2(__lasx_xvldx, src, 0, src, stride, tmp0, tmp1);
1938 __lasx_xvstelm_w(src0, src + stride, 0, 1);
1941 static void avc_wgt_4x4_lasx(uint8_t *src, ptrdiff_t stride,
1948 ptrdiff_t stride_2x = stride << 1;
1949 ptrdiff_t stride_3x = stride_2x + stride;
1957 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1967 __lasx_xvstelm_w(src0, src + stride, 0, 1);
1972 static void avc_wgt_4x8_lasx(uint8_t *src, ptrdiff_t stride,
1980 ptrdiff_t stride_2x = stride << 1;
1981 ptrdiff_t stride_4x = stride << 2;
1982 ptrdiff_t stride_3x = stride_2x + stride;
1990 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
1993 DUP4_ARG2(__lasx_xvldx, src, 0, src, stride, src, stride_2x,
2011 __lasx_xvstelm_w(src0_l, src + stride, 0, 1);
2016 __lasx_xvstelm_w(src0_l, src + stride, 0, 5);
2021 void ff_weight_h264_pixels4_8_lasx(uint8_t *src, ptrdiff_t stride,
2026 avc_wgt_4x2_lasx(src, stride, log2_denom, weight_src, offset);
2028 avc_wgt_4x4_lasx(src, stride, log2_denom, weight_src, offset);
2030 avc_wgt_4x8_lasx(src, stride, log2_denom, weight_src, offset);
2034 void ff_h264_add_pixels4_8_lasx(uint8_t *_dst, int16_t *_src, int stride)
2038 uint8_t* _dst1 = _dst + stride;
2039 uint8_t* _dst2 = _dst1 + stride;
2040 uint8_t* _dst3 = _dst2 + stride;
2061 void ff_h264_add_pixels8_8_lasx(uint8_t *_dst, int16_t *_src, int stride)
2067 uint8_t *_dst1 = _dst + stride;
2068 uint8_t *_dst2 = _dst1 + stride;
2069 uint8_t *_dst3 = _dst2 + stride;
2070 uint8_t *_dst4 = _dst3 + stride;
2071 uint8_t *_dst5 = _dst4 + stride;
2072 uint8_t *_dst6 = _dst5 + stride;
2073 uint8_t *_dst7 = _dst6 + stride;