Lines Matching refs:stride2
234 ptrdiff_t stride2 = stride << 1;
235 ptrdiff_t stride3 = stride2 + stride;
236 ptrdiff_t stride4 = stride2 << 1;
240 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
243 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
256 __lsx_vstelm_d(p1_out, dst - stride2, 0, 0);
267 ptrdiff_t stride2 = stride << 1;
268 ptrdiff_t stride3 = stride2 + stride;
269 ptrdiff_t stride4 = stride2 << 1;
274 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
277 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
296 __lsx_vst(p1, dst - stride2, 0);
307 ptrdiff_t stride2 = stride << 1;
308 ptrdiff_t stride3 = stride2 + stride;
309 ptrdiff_t stride4 = stride2 << 1;
318 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
321 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
338 __lsx_vstelm_d(p1_out, dst - stride2, 0, 0);
366 __lsx_vstelm_d(p1_out, dst - stride2, 0, 0);
370 __lsx_vstelm_d(q2_out, dst + stride2, 0, 0);
379 ptrdiff_t stride2 = stride << 1;
380 ptrdiff_t stride3 = stride2 + stride;
381 ptrdiff_t stride4 = stride2 << 1;
394 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
397 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
421 __lsx_vst(p1_out, dst - stride2, 0);
457 __lsx_vstx(p1_out, dst, -stride2);
461 __lsx_vstx(q2_out, dst, stride2);
470 ptrdiff_t stride2 = stride << 1;
471 ptrdiff_t stride3 = stride2 + stride;
472 ptrdiff_t stride4 = stride2 << 1;
482 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
485 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
511 __lsx_vstx(p1_out, dst, -stride2);
539 __lsx_vstx(p1_out, dst, -stride2);
543 __lsx_vstx(q2_out, dst, stride2);
552 ptrdiff_t stride2 = stride << 1;
553 ptrdiff_t stride3 = stride2 + stride;
554 ptrdiff_t stride4 = stride2 << 1;
564 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
567 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
593 __lsx_vstx(p1_out, dst, -stride2);
621 __lsx_vstx(p1_out, dst, -stride2);
625 __lsx_vstx(q2_out, dst, stride2);
635 ptrdiff_t stride2 = stride << 1;
636 ptrdiff_t stride3 = stride2 + stride;
637 ptrdiff_t stride4 = stride2 << 1;
650 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
653 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
669 __lsx_vstx(p1_out, dst, -stride2);
719 ptrdiff_t stride2 = stride << 1;
720 ptrdiff_t stride3 = stride2 + stride;
721 ptrdiff_t stride4 = stride2 << 1;
741 -stride2, dst_tmp, -stride, p7, p6, p5, p4);
743 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, p2, p1);
747 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
751 DUP2_ARG2(__lsx_vldx, dst_tmp1, stride, dst_tmp1, stride2, q5, q6);
762 __lsx_vstx(p1, dst, -stride2);
766 __lsx_vstx(q2, dst, stride2);
1101 ptrdiff_t stride2 = stride << 1;
1102 ptrdiff_t stride3 = stride2 + stride;
1103 ptrdiff_t stride4 = stride2 << 1;
1118 DUP4_ARG2(__lsx_vldx, dst, -stride4, dst, -stride3, dst, -stride2,
1121 DUP2_ARG2(__lsx_vldx, dst, stride, dst, stride2, q1, q2);
1138 __lsx_vstelm_d(p1_out, dst - stride2, 0, 0);
1169 dst_tmp - stride2, 0, dst_tmp - stride, 0, p7, p6, p5, p4);
1171 dst_tmp1 + stride2, 0, dst_tmp1 + stride3, 0, q4, q5, q6, q7);
1355 ptrdiff_t stride2 = stride << 1;
1356 ptrdiff_t stride3 = stride2 + stride;
1357 ptrdiff_t stride4 = stride2 << 1;
1365 DUP2_ARG2(__lsx_vldx, dst_tmp1, stride, dst_tmp1, stride2, p2, p1);
1368 DUP2_ARG2(__lsx_vldx, dst_tmp2, stride, dst_tmp2, stride2, q1, q2);
1387 __lsx_vstelm_w(vec2, dst + stride2, 0, 2);
1392 __lsx_vstelm_w(vec3, dst + stride2, 0, 2);
1401 ptrdiff_t stride2 = stride << 1;
1402 ptrdiff_t stride3 = stride2 + stride;
1403 ptrdiff_t stride4 = stride2 << 1;
1413 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row1, row2);
1417 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row5, row6);
1421 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row9, row10);
1425 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row13, row14);
1457 __lsx_vstelm_w(tmp2, dst + stride2, 0, 2);
1462 __lsx_vstelm_w(tmp3, dst + stride2, 0, 2);
1467 __lsx_vstelm_w(tmp4, dst + stride2, 0, 2);
1472 __lsx_vstelm_w(tmp5, dst + stride2, 0, 2);
1481 ptrdiff_t stride2 = stride << 1;
1482 ptrdiff_t stride3 = stride2 + stride;
1483 ptrdiff_t stride4 = stride2 << 1;
1496 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, p2, p1);
1500 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, q1, q2);
1531 __lsx_vstelm_w(vec2, dst + stride2, 0, 2);
1536 __lsx_vstelm_w(vec3, dst + stride2, 0, 2);
1599 ptrdiff_t stride2 = stride << 1;
1600 ptrdiff_t stride3 = stride2 + stride;
1601 ptrdiff_t stride4 = stride2 << 1;
1617 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, p1, p2);
1621 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row5, row6);
1625 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, q2, q1);
1629 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row13, row14);
1670 __lsx_vstelm_w(vec2, dst + stride2, 0, 2);
1675 __lsx_vstelm_w(vec3, dst + stride2, 0, 2);
1680 __lsx_vstelm_w(vec4, dst + stride2, 0, 2);
1685 __lsx_vstelm_w(vec5, dst + stride2, 0, 2);
1784 ptrdiff_t stride2 = stride << 1;
1785 ptrdiff_t stride3 = stride2 + stride;
1786 ptrdiff_t stride4 = stride2 << 1;
1799 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, p1, p2);
1803 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row5, row6);
1807 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, q2, q1);
1811 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row13, row14);
1854 __lsx_vstelm_w(vec2, dst + stride2, 0, 2);
1859 __lsx_vstelm_w(vec3, dst + stride2, 0, 2);
1864 __lsx_vstelm_w(vec4, dst + stride2, 0, 2);
1869 __lsx_vstelm_w(vec5, dst + stride2, 0, 2);
1959 ptrdiff_t stride2 = stride << 1;
1960 ptrdiff_t stride3 = stride2 + stride;
1961 ptrdiff_t stride4 = stride2 << 1;
1974 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, p1, p2);
1978 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row5, row6);
1982 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, q2, q1);
1986 DUP2_ARG2(__lsx_vldx, dst_tmp, stride, dst_tmp, stride2, row13, row14);
2029 __lsx_vstelm_w(vec2, dst + stride2, 0, 2);
2034 __lsx_vstelm_w(vec3, dst + stride2, 0, 2);
2039 __lsx_vstelm_w(vec4, dst + stride2, 0, 2);
2044 __lsx_vstelm_w(vec5, dst + stride2, 0, 2);
2643 ptrdiff_t stride2 = stride << 1;
2644 ptrdiff_t stride3 = stride2 + stride;
2645 ptrdiff_t stride4 = stride2 << 1;
2688 __lsx_vstelm_w(vec2, dst_org + stride2, 0, 2);
2693 __lsx_vstelm_w(vec3, dst_org + stride2, 0, 2);
2698 __lsx_vstelm_w(vec4, dst_org + stride2, 0, 2);
2703 __lsx_vstelm_w(vec5, dst_org + stride2, 0, 2);