Lines Matching refs:src
29 static inline void copy_block4_mmi(uint8_t *dst, const uint8_t *src,
37 MMI_ULWC1(%[ftmp0], %[src], 0x00)
40 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
44 [dst]"+&r"(dst), [src]"+&r"(src),
53 static inline void copy_block8_mmi(uint8_t *dst, const uint8_t *src,
61 MMI_ULDC1(%[ftmp0], %[src], 0x00)
64 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
69 [dst]"+&r"(dst), [src]"+&r"(src),
77 static inline void copy_block16_mmi(uint8_t *dst, const uint8_t *src,
86 MMI_ULDC1(%[ftmp0], %[src], 0x00)
87 "ldl %[tmp0], 0x0f(%[src]) \n\t"
88 "ldr %[tmp0], 0x08(%[src]) \n\t"
93 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
99 [dst]"+&r"(dst), [src]"+&r"(src),
109 static void put_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
120 MMI_ULWC1(%[ftmp1], %[src], -0x02)
121 MMI_ULWC1(%[ftmp2], %[src], -0x01)
122 MMI_ULWC1(%[ftmp3], %[src], 0x00)
123 MMI_ULWC1(%[ftmp4], %[src], 0x01)
124 MMI_ULWC1(%[ftmp5], %[src], 0x02)
125 MMI_ULWC1(%[ftmp6], %[src], 0x03)
146 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
155 [dst]"+&r"(dst), [src]"+&r"(src)
164 static void put_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
175 MMI_ULDC1(%[ftmp1], %[src], -0x02)
176 MMI_ULDC1(%[ftmp2], %[src], -0x01)
177 MMI_ULDC1(%[ftmp3], %[src], 0x00)
178 MMI_ULDC1(%[ftmp4], %[src], 0x01)
179 MMI_ULDC1(%[ftmp5], %[src], 0x02)
180 MMI_ULDC1(%[ftmp6], %[src], 0x03)
214 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
225 [dst]"+&r"(dst), [src]"+&r"(src)
234 static void put_h264_qpel16_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
237 put_h264_qpel8_h_lowpass_mmi(dst, src, dstStride, srcStride);
238 put_h264_qpel8_h_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
239 src += 8*srcStride;
241 put_h264_qpel8_h_lowpass_mmi(dst, src, dstStride, srcStride);
242 put_h264_qpel8_h_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
245 static void avg_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
256 MMI_ULWC1(%[ftmp1], %[src], -0x02)
257 MMI_ULWC1(%[ftmp2], %[src], -0x01)
258 MMI_ULWC1(%[ftmp3], %[src], 0x00)
259 MMI_ULWC1(%[ftmp4], %[src], 0x01)
260 MMI_ULWC1(%[ftmp5], %[src], 0x02)
261 MMI_ULWC1(%[ftmp6], %[src], 0x03)
282 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
293 [dst]"+&r"(dst), [src]"+&r"(src)
302 static void avg_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
313 MMI_ULDC1(%[ftmp1], %[src], -0x02)
314 MMI_ULDC1(%[ftmp2], %[src], -0x01)
315 MMI_ULDC1(%[ftmp3], %[src], 0x00)
316 MMI_ULDC1(%[ftmp4], %[src], 0x01)
317 MMI_ULDC1(%[ftmp5], %[src], 0x02)
318 MMI_ULDC1(%[ftmp6], %[src], 0x03)
354 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
365 [dst]"+&r"(dst), [src]"+&r"(src)
374 static void avg_h264_qpel16_h_lowpass_mmi(uint8_t *dst, const uint8_t *src,
377 avg_h264_qpel8_h_lowpass_mmi(dst, src, dstStride, srcStride);
378 avg_h264_qpel8_h_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
379 src += 8*srcStride;
381 avg_h264_qpel8_h_lowpass_mmi(dst, src, dstStride, srcStride);
382 avg_h264_qpel8_h_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
385 static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
392 src -= 2 * srcStride;
399 MMI_LWC1(%[ftmp1], %[src], 0x00)
401 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
403 MMI_LWC1(%[ftmp2], %[src], 0x00)
405 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
406 MMI_LWC1(%[ftmp3], %[src], 0x00)
407 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
408 MMI_LWC1(%[ftmp4], %[src], 0x00)
409 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
410 MMI_LWC1(%[ftmp5], %[src], 0x00)
411 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
417 MMI_LWC1(%[ftmp6], %[src], 0x00)
425 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
432 MMI_LWC1(%[ftmp1], %[src], 0x00)
440 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
447 MMI_LWC1(%[ftmp2], %[src], 0x00)
455 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
462 MMI_LWC1(%[ftmp3], %[src], 0x00)
470 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
486 [dst]"+&r"(dst), [src]"+&r"(src)
494 static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
503 src -= 2 * srcStride;
510 MMI_LWC1(%[ftmp0], %[src], 0x00)
512 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
514 MMI_LWC1(%[ftmp1], %[src], 0x00)
516 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
517 MMI_LWC1(%[ftmp2], %[src], 0x00)
518 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
520 MMI_LWC1(%[ftmp3], %[src], 0x00)
521 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
522 MMI_LWC1(%[ftmp4], %[src], 0x00)
523 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
526 MMI_LWC1(%[ftmp5], %[src], 0x00)
537 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
544 MMI_LWC1(%[ftmp0], %[src], 0x00)
551 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
560 MMI_LWC1(%[ftmp1], %[src], 0x00)
568 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
576 MMI_LWC1(%[ftmp2], %[src], 0x00)
582 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
591 MMI_LWC1(%[ftmp3], %[src], 0x00)
597 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
606 MMI_LWC1(%[ftmp4], %[src], 0x00)
612 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
621 MMI_LWC1(%[ftmp5], %[src], 0x00)
627 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
636 MMI_LWC1(%[ftmp0], %[src], 0x00)
642 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
651 MMI_LWC1(%[ftmp1], %[src], 0x00)
660 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
667 MMI_LWC1(%[ftmp2], %[src], 0x00)
673 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
682 MMI_LWC1(%[ftmp3], %[src], 0x00)
688 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
697 MMI_LWC1(%[ftmp4], %[src], 0x00)
703 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
712 MMI_LWC1(%[ftmp5], %[src], 0x00)
718 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
727 MMI_LWC1(%[ftmp0], %[src], 0x00)
733 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
742 MMI_LWC1(%[ftmp1], %[src], 0x00)
748 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
757 MMI_LWC1(%[ftmp2], %[src], 0x00)
763 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
779 [src]"+&r"(src), [dst]"+&r"(dst),
787 src += 4 - (h + 5) * srcStride;
792 static void put_h264_qpel16_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
795 put_h264_qpel8_v_lowpass_mmi(dst, src, dstStride, srcStride);
796 put_h264_qpel8_v_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
797 src += 8*srcStride;
799 put_h264_qpel8_v_lowpass_mmi(dst, src, dstStride, srcStride);
800 put_h264_qpel8_v_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
803 static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
809 src -= 2 * srcStride;
818 MMI_LWC1(%[ftmp0], %[src], 0x00)
820 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
821 MMI_LWC1(%[ftmp1], %[src], 0x00)
822 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
823 MMI_LWC1(%[ftmp2], %[src], 0x00)
824 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
825 MMI_LWC1(%[ftmp3], %[src], 0x00)
826 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
827 MMI_LWC1(%[ftmp4], %[src], 0x00)
828 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
834 MMI_LWC1(%[ftmp5], %[src], 0x00)
842 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
851 MMI_LWC1(%[ftmp0], %[src], 0x00)
859 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
868 MMI_LWC1(%[ftmp1], %[src], 0x00)
876 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
885 MMI_LWC1(%[ftmp2], %[src], 0x00)
893 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
909 [src]"+&r"(src), [dst]"+&r"(dst)
917 static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
926 src -= 2 * srcStride;
936 MMI_LWC1(%[ftmp0], %[src], 0x00)
938 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
939 MMI_LWC1(%[ftmp1], %[src], 0x00)
940 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
941 MMI_LWC1(%[ftmp2], %[src], 0x00)
942 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
943 MMI_LWC1(%[ftmp3], %[src], 0x00)
944 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
945 MMI_LWC1(%[ftmp4], %[src], 0x00)
946 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
952 MMI_LWC1(%[ftmp5], %[src], 0x00)
960 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
969 MMI_LWC1(%[ftmp0], %[src], 0x00)
977 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
986 MMI_LWC1(%[ftmp1], %[src], 0x00)
994 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1003 MMI_LWC1(%[ftmp2], %[src], 0x00)
1011 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1020 MMI_LWC1(%[ftmp3], %[src], 0x00)
1028 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1037 MMI_LWC1(%[ftmp4], %[src], 0x00)
1045 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1054 MMI_LWC1(%[ftmp5], %[src], 0x00)
1062 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1071 MMI_LWC1(%[ftmp0], %[src], 0x00)
1079 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1089 MMI_LWC1(%[ftmp1], %[src], 0x00)
1097 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1106 MMI_LWC1(%[ftmp2], %[src], 0x00)
1114 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1123 MMI_LWC1(%[ftmp3], %[src], 0x00)
1131 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1140 MMI_LWC1(%[ftmp4], %[src], 0x00)
1148 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1157 MMI_LWC1(%[ftmp5], %[src], 0x00)
1165 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1174 MMI_LWC1(%[ftmp0], %[src], 0x00)
1182 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1191 MMI_LWC1(%[ftmp1], %[src], 0x00)
1199 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1208 MMI_LWC1(%[ftmp2], %[src], 0x00)
1216 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1234 [src]"+&r"(src), [dst]"+&r"(dst),
1242 src += 4 - (h + 5) * srcStride;
1247 static void avg_h264_qpel16_v_lowpass_mmi(uint8_t *dst, const uint8_t *src,
1250 avg_h264_qpel8_v_lowpass_mmi(dst, src, dstStride, srcStride);
1251 avg_h264_qpel8_v_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
1252 src += 8*srcStride;
1254 avg_h264_qpel8_v_lowpass_mmi(dst, src, dstStride, srcStride);
1255 avg_h264_qpel8_v_lowpass_mmi(dst+8, src+8, dstStride, srcStride);
1258 static void put_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src,
1269 src -= 2*srcStride;
1275 MMI_ULWC1(%[ftmp1], %[src], -0x02)
1276 MMI_ULWC1(%[ftmp2], %[src], -0x01)
1277 MMI_ULWC1(%[ftmp3], %[src], 0x00)
1278 MMI_ULWC1(%[ftmp4], %[src], 0x01)
1279 MMI_ULWC1(%[ftmp5], %[src], 0x02)
1280 MMI_ULWC1(%[ftmp6], %[src], 0x03)
1296 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1306 [tmp]"+&r"(tmp), [src]"+&r"(src)
1335 const uint8_t *src, ptrdiff_t tmpStride, ptrdiff_t srcStride, int size)
1342 src -= 2 * srcStride + 2;
1347 MMI_ULWC1(%[ftmp0], %[src], 0x00)
1349 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1351 MMI_ULWC1(%[ftmp1], %[src], 0x00)
1352 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1353 MMI_ULWC1(%[ftmp2], %[src], 0x00)
1354 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1355 MMI_ULWC1(%[ftmp3], %[src], 0x00)
1356 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1357 MMI_ULWC1(%[ftmp4], %[src], 0x00)
1358 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1364 MMI_ULWC1(%[ftmp5], %[src], 0x00)
1373 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1376 MMI_ULWC1(%[ftmp0], %[src], 0x00)
1385 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1388 MMI_ULWC1(%[ftmp1], %[src], 0x00)
1397 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1400 MMI_ULWC1(%[ftmp2], %[src], 0x00)
1409 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1412 MMI_ULWC1(%[ftmp3], %[src], 0x00)
1421 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1424 MMI_ULWC1(%[ftmp4], %[src], 0x00)
1433 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1436 MMI_ULWC1(%[ftmp5], %[src], 0x00)
1445 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1448 MMI_ULWC1(%[ftmp0], %[src], 0x00)
1457 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1462 MMI_ULWC1(%[ftmp1], %[src], 0x00)
1471 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1474 MMI_ULWC1(%[ftmp2], %[src], 0x00)
1483 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1486 MMI_ULWC1(%[ftmp3], %[src], 0x00)
1495 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1498 MMI_ULWC1(%[ftmp4], %[src], 0x00)
1507 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1510 MMI_ULWC1(%[ftmp5], %[src], 0x00)
1519 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1522 MMI_ULWC1(%[ftmp0], %[src], 0x00)
1531 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1534 MMI_ULWC1(%[ftmp1], %[src], 0x00)
1543 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1546 MMI_ULWC1(%[ftmp2], %[src], 0x00)
1555 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1567 [src]"+&r"(src)
1575 src += 4 - (size + 5) * srcStride;
1651 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1654 put_h264_qpel8or16_hv1_lowpass_mmi(tmp, src, tmpStride, srcStride, size);
1659 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1662 put_h264_qpel8or16_hv_lowpass_mmi(dst, tmp, src, dstStride, tmpStride,
1667 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1670 put_h264_qpel8or16_hv_lowpass_mmi(dst, tmp, src, dstStride, tmpStride,
1674 static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src,
1690 MMI_ULDC1(%[ftmp1], %[src], 0x00)
1691 MMI_ULDC1(%[ftmp3], %[src], 0x01)
1700 MMI_ULDC1(%[ftmp3], %[src], -0x01)
1701 MMI_ULDC1(%[ftmp5], %[src], 0x02)
1712 MMI_ULWC1(%[ftmp3], %[src], -0x02)
1713 MMI_ULWC1(%[ftmp6], %[src], 0x07)
1726 PTR_ADDU "%[src], %[src], %[dstStride] \n\t"
1741 [src]"+&r"(src), [dst]"+&r"(dst),
1798 static void put_h264_qpel16_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src,
1801 put_h264_qpel8_h_lowpass_l2_mmi(dst, src, src2, dstStride, src2Stride);
1802 put_h264_qpel8_h_lowpass_l2_mmi(dst + 8, src + 8, src2 + 8, dstStride,
1805 src += 8 * dstStride;
1809 put_h264_qpel8_h_lowpass_l2_mmi(dst, src, src2, dstStride, src2Stride);
1810 put_h264_qpel8_h_lowpass_l2_mmi(dst + 8, src + 8, src2 + 8, dstStride,
1822 static void avg_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src,
1833 src -= 2*srcStride;
1839 MMI_ULWC1(%[ftmp1], %[src], -0x02)
1840 MMI_ULWC1(%[ftmp2], %[src], -0x01)
1841 MMI_ULWC1(%[ftmp3], %[src], 0x00)
1842 MMI_ULWC1(%[ftmp4], %[src], 0x01)
1843 MMI_ULWC1(%[ftmp5], %[src], 0x02)
1844 MMI_ULWC1(%[ftmp6], %[src], 0x03)
1860 PTR_ADDU "%[src], %[src], %[srcStride] \n\t"
1870 [tmp]"+&r"(tmp), [src]"+&r"(src)
1972 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1975 put_h264_qpel8or16_hv1_lowpass_mmi(tmp, src, tmpStride, srcStride, size);
1980 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1983 avg_h264_qpel8or16_hv_lowpass_mmi(dst, tmp, src, dstStride, tmpStride,
1988 const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t tmpStride,
1991 avg_h264_qpel8or16_hv_lowpass_mmi(dst, tmp, src, dstStride, tmpStride,
1995 static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src,
2011 MMI_ULDC1(%[ftmp1], %[src], 0x00)
2012 MMI_ULDC1(%[ftmp2], %[src], 0x01)
2021 MMI_ULDC1(%[ftmp2], %[src], -0x01)
2022 MMI_ULDC1(%[ftmp5], %[src], 0x02)
2033 MMI_ULWC1(%[ftmp2], %[src], -0x02)
2034 MMI_ULWC1(%[ftmp6], %[src], 0x07)
2050 PTR_ADDU "%[src], %[src], %[dstStride] \n\t"
2064 [dst]"+&r"(dst), [src]"+&r"(src),
2073 static void avg_h264_qpel16_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src,
2076 avg_h264_qpel8_h_lowpass_l2_mmi(dst, src, src2, dstStride, src2Stride);
2077 avg_h264_qpel8_h_lowpass_l2_mmi(dst + 8, src + 8, src2 + 8, dstStride,
2080 src += 8 * dstStride;
2084 avg_h264_qpel8_h_lowpass_l2_mmi(dst, src, src2, dstStride, src2Stride);
2085 avg_h264_qpel8_h_lowpass_l2_mmi(dst + 8, src + 8, src2 + 8, dstStride,
2150 void ff_put_h264_qpel4_mc00_mmi(uint8_t *dst, const uint8_t *src,
2153 ff_put_pixels4_8_mmi(dst, src, stride, 4);
2156 void ff_put_h264_qpel4_mc10_mmi(uint8_t *dst, const uint8_t *src,
2160 put_h264_qpel4_h_lowpass_mmi(half, src, 4, stride);
2161 ff_put_pixels4_l2_8_mmi(dst, src, half, stride, stride, 4, 4);
2164 void ff_put_h264_qpel4_mc20_mmi(uint8_t *dst, const uint8_t *src,
2167 put_h264_qpel4_h_lowpass_mmi(dst, src, stride, stride);
2170 void ff_put_h264_qpel4_mc30_mmi(uint8_t *dst, const uint8_t *src,
2174 put_h264_qpel4_h_lowpass_mmi(half, src, 4, stride);
2175 ff_put_pixels4_l2_8_mmi(dst, src+1, half, stride, stride, 4, 4);
2178 void ff_put_h264_qpel4_mc01_mmi(uint8_t *dst, const uint8_t *src,
2184 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2189 void ff_put_h264_qpel4_mc02_mmi(uint8_t *dst, const uint8_t *src,
2194 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2198 void ff_put_h264_qpel4_mc03_mmi(uint8_t *dst, const uint8_t *src,
2204 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2209 void ff_put_h264_qpel4_mc11_mmi(uint8_t *dst, const uint8_t *src,
2216 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2217 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2222 void ff_put_h264_qpel4_mc31_mmi(uint8_t *dst, const uint8_t *src,
2229 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2230 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2235 void ff_put_h264_qpel4_mc13_mmi(uint8_t *dst, const uint8_t *src,
2242 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2243 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2248 void ff_put_h264_qpel4_mc33_mmi(uint8_t *dst, const uint8_t *src,
2255 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2256 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2261 void ff_put_h264_qpel4_mc22_mmi(uint8_t *dst, const uint8_t *src,
2264 put_h264_qpel4_hv_lowpass_mmi(dst, src, stride, stride);
2267 void ff_put_h264_qpel4_mc21_mmi(uint8_t *dst, const uint8_t *src,
2272 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2273 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2277 void ff_put_h264_qpel4_mc23_mmi(uint8_t *dst, const uint8_t *src,
2282 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2283 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2287 void ff_put_h264_qpel4_mc12_mmi(uint8_t *dst, const uint8_t *src,
2294 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2296 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2300 void ff_put_h264_qpel4_mc32_mmi(uint8_t *dst, const uint8_t *src,
2307 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2309 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2314 void ff_avg_h264_qpel4_mc00_mmi(uint8_t *dst, const uint8_t *src,
2317 ff_avg_pixels4_8_mmi(dst, src, stride, 4);
2320 void ff_avg_h264_qpel4_mc10_mmi(uint8_t *dst, const uint8_t *src,
2324 put_h264_qpel4_h_lowpass_mmi(half, src, 4, stride);
2325 ff_avg_pixels4_l2_8_mmi(dst, src, half, stride, stride, 4, 4);
2328 void ff_avg_h264_qpel4_mc20_mmi(uint8_t *dst, const uint8_t *src,
2331 avg_h264_qpel4_h_lowpass_mmi(dst, src, stride, stride);
2334 void ff_avg_h264_qpel4_mc30_mmi(uint8_t *dst, const uint8_t *src,
2338 put_h264_qpel4_h_lowpass_mmi(half, src, 4, stride);
2339 ff_avg_pixels4_l2_8_mmi(dst, src+1, half, stride, stride, 4, 4);
2342 void ff_avg_h264_qpel4_mc01_mmi(uint8_t *dst, const uint8_t *src,
2348 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2353 void ff_avg_h264_qpel4_mc02_mmi(uint8_t *dst, const uint8_t *src,
2358 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2362 void ff_avg_h264_qpel4_mc03_mmi(uint8_t *dst, const uint8_t *src,
2368 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2373 void ff_avg_h264_qpel4_mc11_mmi(uint8_t *dst, const uint8_t *src,
2380 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2381 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2386 void ff_avg_h264_qpel4_mc31_mmi(uint8_t *dst, const uint8_t *src,
2393 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2394 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2399 void ff_avg_h264_qpel4_mc13_mmi(uint8_t *dst, const uint8_t *src,
2406 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2407 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2412 void ff_avg_h264_qpel4_mc33_mmi(uint8_t *dst, const uint8_t *src,
2419 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2420 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2425 void ff_avg_h264_qpel4_mc22_mmi(uint8_t *dst, const uint8_t *src,
2428 avg_h264_qpel4_hv_lowpass_mmi(dst, src, stride, stride);
2431 void ff_avg_h264_qpel4_mc21_mmi(uint8_t *dst, const uint8_t *src,
2436 put_h264_qpel4_h_lowpass_mmi(halfH, src, 4, stride);
2437 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2441 void ff_avg_h264_qpel4_mc23_mmi(uint8_t *dst, const uint8_t *src,
2446 put_h264_qpel4_h_lowpass_mmi(halfH, src + stride, 4, stride);
2447 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2451 void ff_avg_h264_qpel4_mc12_mmi(uint8_t *dst, const uint8_t *src,
2458 copy_block4_mmi(full, src - stride*2, 4, stride, 9);
2460 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2464 void ff_avg_h264_qpel4_mc32_mmi(uint8_t *dst, const uint8_t *src,
2471 copy_block4_mmi(full, src - stride*2 + 1, 4, stride, 9);
2473 put_h264_qpel4_hv_lowpass_mmi(halfHV, src, 4, stride);
2478 void ff_put_h264_qpel8_mc00_mmi(uint8_t *dst, const uint8_t *src,
2481 ff_put_pixels8_8_mmi(dst, src, stride, 8);
2484 void ff_put_h264_qpel8_mc10_mmi(uint8_t *dst, const uint8_t *src,
2488 put_h264_qpel8_h_lowpass_mmi(half, src, 8, stride);
2489 ff_put_pixels8_l2_8_mmi(dst, src, half, stride, stride, 8, 8);
2492 void ff_put_h264_qpel8_mc20_mmi(uint8_t *dst, const uint8_t *src,
2495 put_h264_qpel8_h_lowpass_mmi(dst, src, stride, stride);
2498 void ff_put_h264_qpel8_mc30_mmi(uint8_t *dst, const uint8_t *src,
2502 put_h264_qpel8_h_lowpass_mmi(half, src, 8, stride);
2503 ff_put_pixels8_l2_8_mmi(dst, src+1, half, stride, stride, 8, 8);
2506 void ff_put_h264_qpel8_mc01_mmi(uint8_t *dst, const uint8_t *src,
2512 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2517 void ff_put_h264_qpel8_mc02_mmi(uint8_t *dst, const uint8_t *src,
2522 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2526 void ff_put_h264_qpel8_mc03_mmi(uint8_t *dst, const uint8_t *src,
2532 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2537 void ff_put_h264_qpel8_mc11_mmi(uint8_t *dst, const uint8_t *src,
2544 put_h264_qpel8_h_lowpass_mmi(halfH, src, 8, stride);
2545 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2550 void ff_put_h264_qpel8_mc31_mmi(uint8_t *dst, const uint8_t *src,
2557 put_h264_qpel8_h_lowpass_mmi(halfH, src, 8, stride);
2558 copy_block8_mmi(full, src - stride*2 + 1, 8, stride, 13);
2563 void ff_put_h264_qpel8_mc13_mmi(uint8_t *dst, const uint8_t *src,
2570 put_h264_qpel8_h_lowpass_mmi(halfH, src + stride, 8, stride);
2571 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2576 void ff_put_h264_qpel8_mc33_mmi(uint8_t *dst, const uint8_t *src,
2583 put_h264_qpel8_h_lowpass_mmi(halfH, src + stride, 8, stride);
2584 copy_block8_mmi(full, src - stride*2 + 1, 8, stride, 13);
2589 void ff_put_h264_qpel8_mc22_mmi(uint8_t *dst, const uint8_t *src,
2594 put_h264_qpel8_hv_lowpass_mmi(dst, temp, src, stride, 8, stride);
2597 void ff_put_h264_qpel8_mc21_mmi(uint8_t *dst, const uint8_t *src,
2604 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2605 put_h264_qpel8_h_lowpass_l2_mmi(dst, src, halfHV, stride, 8);
2608 void ff_put_h264_qpel8_mc23_mmi(uint8_t *dst, const uint8_t *src,
2615 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2616 put_h264_qpel8_h_lowpass_l2_mmi(dst, src + stride, halfHV, stride, 8);
2619 void ff_put_h264_qpel8_mc12_mmi(uint8_t *dst, const uint8_t *src,
2626 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2630 void ff_put_h264_qpel8_mc32_mmi(uint8_t *dst, const uint8_t *src,
2637 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2642 void ff_avg_h264_qpel8_mc00_mmi(uint8_t *dst, const uint8_t *src,
2645 ff_avg_pixels8_8_mmi(dst, src, stride, 8);
2648 void ff_avg_h264_qpel8_mc10_mmi(uint8_t *dst, const uint8_t *src,
2652 put_h264_qpel8_h_lowpass_mmi(half, src, 8, stride);
2653 ff_avg_pixels8_l2_8_mmi(dst, src, half, stride, stride, 8, 8);
2656 void ff_avg_h264_qpel8_mc20_mmi(uint8_t *dst, const uint8_t *src,
2659 avg_h264_qpel8_h_lowpass_mmi(dst, src, stride, stride);
2662 void ff_avg_h264_qpel8_mc30_mmi(uint8_t *dst, const uint8_t *src,
2666 put_h264_qpel8_h_lowpass_mmi(half, src, 8, stride);
2667 ff_avg_pixels8_l2_8_mmi(dst, src+1, half, stride, stride, 8, 8);
2670 void ff_avg_h264_qpel8_mc01_mmi(uint8_t *dst, const uint8_t *src,
2676 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2681 void ff_avg_h264_qpel8_mc02_mmi(uint8_t *dst, const uint8_t *src,
2686 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2690 void ff_avg_h264_qpel8_mc03_mmi(uint8_t *dst, const uint8_t *src,
2696 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2701 void ff_avg_h264_qpel8_mc11_mmi(uint8_t *dst, const uint8_t *src,
2708 put_h264_qpel8_h_lowpass_mmi(halfH, src, 8, stride);
2709 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2714 void ff_avg_h264_qpel8_mc31_mmi(uint8_t *dst, const uint8_t *src,
2721 put_h264_qpel8_h_lowpass_mmi(halfH, src, 8, stride);
2722 copy_block8_mmi(full, src - stride*2 + 1, 8, stride, 13);
2727 void ff_avg_h264_qpel8_mc13_mmi(uint8_t *dst, const uint8_t *src,
2734 put_h264_qpel8_h_lowpass_mmi(halfH, src + stride, 8, stride);
2735 copy_block8_mmi(full, src - stride*2, 8, stride, 13);
2740 void ff_avg_h264_qpel8_mc33_mmi(uint8_t *dst, const uint8_t *src,
2747 put_h264_qpel8_h_lowpass_mmi(halfH, src + stride, 8, stride);
2748 copy_block8_mmi(full, src - stride*2 + 1, 8, stride, 13);
2753 void ff_avg_h264_qpel8_mc22_mmi(uint8_t *dst, const uint8_t *src,
2758 avg_h264_qpel8_hv_lowpass_mmi(dst, temp, src, stride, 8, stride);
2761 void ff_avg_h264_qpel8_mc21_mmi(uint8_t *dst, const uint8_t *src,
2768 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2769 avg_h264_qpel8_h_lowpass_l2_mmi(dst, src, halfHV, stride, 8);
2772 void ff_avg_h264_qpel8_mc23_mmi(uint8_t *dst, const uint8_t *src,
2779 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2780 avg_h264_qpel8_h_lowpass_l2_mmi(dst, src + stride, halfHV, stride, 8);
2783 void ff_avg_h264_qpel8_mc12_mmi(uint8_t *dst, const uint8_t *src,
2790 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2794 void ff_avg_h264_qpel8_mc32_mmi(uint8_t *dst, const uint8_t *src,
2801 put_h264_qpel8_hv_lowpass_mmi(halfHV, halfV, src, 8, 8, stride);
2806 void ff_put_h264_qpel16_mc00_mmi(uint8_t *dst, const uint8_t *src,
2809 ff_put_pixels16_8_mmi(dst, src, stride, 16);
2812 void ff_put_h264_qpel16_mc10_mmi(uint8_t *dst, const uint8_t *src,
2816 put_h264_qpel16_h_lowpass_mmi(half, src, 16, stride);
2817 ff_put_pixels16_l2_8_mmi(dst, src, half, stride, stride, 16, 16);
2820 void ff_put_h264_qpel16_mc20_mmi(uint8_t *dst, const uint8_t *src,
2823 put_h264_qpel16_h_lowpass_mmi(dst, src, stride, stride);
2826 void ff_put_h264_qpel16_mc30_mmi(uint8_t *dst, const uint8_t *src,
2830 put_h264_qpel16_h_lowpass_mmi(half, src, 16, stride);
2831 ff_put_pixels16_l2_8_mmi(dst, src+1, half, stride, stride, 16, 16);
2834 void ff_put_h264_qpel16_mc01_mmi(uint8_t *dst, const uint8_t *src,
2840 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
2845 void ff_put_h264_qpel16_mc02_mmi(uint8_t *dst, const uint8_t *src,
2850 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
2854 void ff_put_h264_qpel16_mc03_mmi(uint8_t *dst, const uint8_t *src,
2860 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
2865 void ff_put_h264_qpel16_mc11_mmi(uint8_t *dst, const uint8_t *src,
2872 put_h264_qpel16_h_lowpass_mmi(halfH, src, 16, stride);
2873 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
2878 void ff_put_h264_qpel16_mc31_mmi(uint8_t *dst, const uint8_t *src,
2885 put_h264_qpel16_h_lowpass_mmi(halfH, src, 16, stride);
2886 copy_block16_mmi(full, src - stride*2 + 1, 16, stride, 21);
2891 void ff_put_h264_qpel16_mc13_mmi(uint8_t *dst, const uint8_t *src,
2898 put_h264_qpel16_h_lowpass_mmi(halfH, src + stride, 16, stride);
2899 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
2904 void ff_put_h264_qpel16_mc33_mmi(uint8_t *dst, const uint8_t *src,
2911 put_h264_qpel16_h_lowpass_mmi(halfH, src + stride, 16, stride);
2912 copy_block16_mmi(full, src - stride*2 + 1, 16, stride, 21);
2917 void ff_put_h264_qpel16_mc22_mmi(uint8_t *dst, const uint8_t *src,
2922 put_h264_qpel16_hv_lowpass_mmi(dst, temp, src, stride, 16, stride);
2925 void ff_put_h264_qpel16_mc21_mmi(uint8_t *dst, const uint8_t *src,
2932 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
2933 put_h264_qpel16_h_lowpass_l2_mmi(dst, src, halfHV, stride, 16);
2936 void ff_put_h264_qpel16_mc23_mmi(uint8_t *dst, const uint8_t *src,
2943 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
2944 put_h264_qpel16_h_lowpass_l2_mmi(dst, src + stride, halfHV, stride, 16);
2947 void ff_put_h264_qpel16_mc12_mmi(uint8_t *dst, const uint8_t *src,
2954 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
2958 void ff_put_h264_qpel16_mc32_mmi(uint8_t *dst, const uint8_t *src,
2965 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
2970 void ff_avg_h264_qpel16_mc00_mmi(uint8_t *dst, const uint8_t *src,
2973 ff_avg_pixels16_8_mmi(dst, src, stride, 16);
2976 void ff_avg_h264_qpel16_mc10_mmi(uint8_t *dst, const uint8_t *src,
2980 put_h264_qpel16_h_lowpass_mmi(half, src, 16, stride);
2981 ff_avg_pixels16_l2_8_mmi(dst, src, half, stride, stride, 16, 16);
2984 void ff_avg_h264_qpel16_mc20_mmi(uint8_t *dst, const uint8_t *src,
2987 avg_h264_qpel16_h_lowpass_mmi(dst, src, stride, stride);
2990 void ff_avg_h264_qpel16_mc30_mmi(uint8_t *dst, const uint8_t *src,
2994 put_h264_qpel16_h_lowpass_mmi(half, src, 16, stride);
2995 ff_avg_pixels16_l2_8_mmi(dst, src+1, half, stride, stride, 16, 16);
2998 void ff_avg_h264_qpel16_mc01_mmi(uint8_t *dst, const uint8_t *src,
3004 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
3009 void ff_avg_h264_qpel16_mc02_mmi(uint8_t *dst, const uint8_t *src,
3014 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
3018 void ff_avg_h264_qpel16_mc03_mmi(uint8_t *dst, const uint8_t *src,
3024 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
3029 void ff_avg_h264_qpel16_mc11_mmi(uint8_t *dst, const uint8_t *src,
3036 put_h264_qpel16_h_lowpass_mmi(halfH, src, 16, stride);
3037 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
3042 void ff_avg_h264_qpel16_mc31_mmi(uint8_t *dst, const uint8_t *src,
3049 put_h264_qpel16_h_lowpass_mmi(halfH, src, 16, stride);
3050 copy_block16_mmi(full, src - stride*2 + 1, 16, stride, 21);
3055 void ff_avg_h264_qpel16_mc13_mmi(uint8_t *dst, const uint8_t *src,
3062 put_h264_qpel16_h_lowpass_mmi(halfH, src + stride, 16, stride);
3063 copy_block16_mmi(full, src - stride*2, 16, stride, 21);
3068 void ff_avg_h264_qpel16_mc33_mmi(uint8_t *dst, const uint8_t *src,
3075 put_h264_qpel16_h_lowpass_mmi(halfH, src + stride, 16, stride);
3076 copy_block16_mmi(full, src - stride*2 + 1, 16, stride, 21);
3081 void ff_avg_h264_qpel16_mc22_mmi(uint8_t *dst, const uint8_t *src,
3086 avg_h264_qpel16_hv_lowpass_mmi(dst, temp, src, stride, 16, stride);
3089 void ff_avg_h264_qpel16_mc21_mmi(uint8_t *dst, const uint8_t *src,
3096 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
3097 avg_h264_qpel16_h_lowpass_l2_mmi(dst, src, halfHV, stride, 16);
3100 void ff_avg_h264_qpel16_mc23_mmi(uint8_t *dst, const uint8_t *src,
3107 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
3108 avg_h264_qpel16_h_lowpass_l2_mmi(dst, src + stride, halfHV, stride, 16);
3111 void ff_avg_h264_qpel16_mc12_mmi(uint8_t *dst, const uint8_t *src,
3118 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);
3122 void ff_avg_h264_qpel16_mc32_mmi(uint8_t *dst, const uint8_t *src,
3129 put_h264_qpel16_hv_lowpass_mmi(halfHV, halfV, src, 16, 16, stride);