Lines Matching refs:stride
137 static void intra_predict_plane_8x8_msa(uint8_t *src, int32_t stride)
150 src_top = LD_UB(src - (stride + 1));
160 res1 = (src[4 * stride - 1] - src[2 * stride - 1]) +
161 2 * (src[5 * stride - 1] - src[stride - 1]) +
162 3 * (src[6 * stride - 1] - src[-1]) +
163 4 * (src[7 * stride - 1] - src[-stride - 1]);
171 res2 = 16 * (src[7 * stride - 1] + src[-stride + 7] + 1);
197 src += stride;
199 src += stride;
205 static void intra_predict_plane_16x16_msa(uint8_t *src, int32_t stride)
219 load0 = LD(src - (stride + 1));
220 load1 = LD(src - (stride + 1) + 9);
233 res1 = (src[8 * stride - 1] - src[6 * stride - 1]) +
234 2 * (src[9 * stride - 1] - src[5 * stride - 1]) +
235 3 * (src[10 * stride - 1] - src[4 * stride - 1]) +
236 4 * (src[11 * stride - 1] - src[3 * stride - 1]) +
237 5 * (src[12 * stride - 1] - src[2 * stride - 1]) +
238 6 * (src[13 * stride - 1] - src[stride - 1]) +
239 7 * (src[14 * stride - 1] - src[-1]) +
240 8 * (src[15 * stride - 1] - src[-1 * stride - 1]);
248 res2 = 16 * (src[15 * stride - 1] + src[-stride + 15] + 1);
277 ST_UB2(store0, store1, src, stride);
278 src += 2 * stride;
284 static void intra_predict_dc_4blk_8x8_msa(uint8_t *src, int32_t stride)
293 src_top = LD_UB(src - stride);
298 src0 += src[0 * stride - 1];
299 src0 += src[1 * stride - 1];
300 src0 += src[2 * stride - 1];
301 src0 += src[3 * stride - 1];
302 src2 = src[4 * stride - 1];
303 src2 += src[5 * stride - 1];
304 src2 += src[6 * stride - 1];
305 src2 += src[7 * stride - 1];
317 SD4(store0, store0, store0, store0, src, stride);
318 src += (4 * stride);
319 SD4(store1, store1, store1, store1, src, stride);
322 static void intra_predict_hor_dc_8x8_msa(uint8_t *src, int32_t stride)
327 src0 = src[0 * stride - 1];
328 src0 += src[1 * stride - 1];
329 src0 += src[2 * stride - 1];
330 src0 += src[3 * stride - 1];
331 src1 = src[4 * stride - 1];
332 src1 += src[5 * stride - 1];
333 src1 += src[6 * stride - 1];
334 src1 += src[7 * stride - 1];
340 SD4(out0, out0, out0, out0, src, stride);
341 src += (4 * stride);
342 SD4(out1, out1, out1, out1, src, stride);
345 static void intra_predict_vert_dc_8x8_msa(uint8_t *src, int32_t stride)
353 src_top = LD_UB(src - stride);
360 SD4(out0, out0, out0, out0, src, stride);
361 src += (4 * stride);
362 SD4(out0, out0, out0, out0, src, stride);
365 static void intra_predict_mad_cow_dc_l0t_8x8_msa(uint8_t *src, int32_t stride)
374 src_top = LD_UB(src - stride);
380 src2 = src[0 * stride - 1];
381 src2 += src[1 * stride - 1];
382 src2 += src[2 * stride - 1];
383 src2 += src[3 * stride - 1];
394 SD4(store0, store0, store0, store0, src, stride);
395 src += (4 * stride);
396 SD4(store1, store1, store1, store1, src, stride);
399 static void intra_predict_mad_cow_dc_0lt_8x8_msa(uint8_t *src, int32_t stride)
408 src_top = LD_UB(src - stride);
414 src2 = src[4 * stride - 1];
415 src2 += src[5 * stride - 1];
416 src2 += src[6 * stride - 1];
417 src2 += src[7 * stride - 1];
430 SD4(store0, store0, store0, store0, src, stride);
431 src += (4 * stride);
432 SD4(store1, store1, store1, store1, src, stride);
435 static void intra_predict_mad_cow_dc_l00_8x8_msa(uint8_t *src, int32_t stride)
440 src0 = src[0 * stride - 1];
441 src0 += src[1 * stride - 1];
442 src0 += src[2 * stride - 1];
443 src0 += src[3 * stride - 1];
448 SD4(out0, out0, out0, out0, src, stride);
449 src += (4 * stride);
450 SD4(out1, out1, out1, out1, src, stride);
453 static void intra_predict_mad_cow_dc_0l0_8x8_msa(uint8_t *src, int32_t stride)
458 src0 = src[4 * stride - 1];
459 src0 += src[5 * stride - 1];
460 src0 += src[6 * stride - 1];
461 src0 += src[7 * stride - 1];
467 SD4(out0, out0, out0, out0, src, stride);
468 src += (4 * stride);
469 SD4(out1, out1, out1, out1, src, stride);
472 void ff_h264_intra_predict_plane_8x8_msa(uint8_t *src, ptrdiff_t stride)
474 intra_predict_plane_8x8_msa(src, stride);
477 void ff_h264_intra_predict_dc_4blk_8x8_msa(uint8_t *src, ptrdiff_t stride)
479 intra_predict_dc_4blk_8x8_msa(src, stride);
482 void ff_h264_intra_predict_hor_dc_8x8_msa(uint8_t *src, ptrdiff_t stride)
484 intra_predict_hor_dc_8x8_msa(src, stride);
487 void ff_h264_intra_predict_vert_dc_8x8_msa(uint8_t *src, ptrdiff_t stride)
489 intra_predict_vert_dc_8x8_msa(src, stride);
493 ptrdiff_t stride)
495 intra_predict_mad_cow_dc_l0t_8x8_msa(src, stride);
499 ptrdiff_t stride)
501 intra_predict_mad_cow_dc_0lt_8x8_msa(src, stride);
505 ptrdiff_t stride)
507 intra_predict_mad_cow_dc_l00_8x8_msa(src, stride);
511 ptrdiff_t stride)
513 intra_predict_mad_cow_dc_0l0_8x8_msa(src, stride);
516 void ff_h264_intra_predict_plane_16x16_msa(uint8_t *src, ptrdiff_t stride)
518 intra_predict_plane_16x16_msa(src, stride);
521 void ff_h264_intra_pred_vert_8x8_msa(uint8_t *src, ptrdiff_t stride)
525 intra_predict_vert_8x8_msa(src - stride, dst, stride);
528 void ff_h264_intra_pred_horiz_8x8_msa(uint8_t *src, ptrdiff_t stride)
532 intra_predict_horiz_8x8_msa(src - 1, stride, dst, stride);
535 void ff_h264_intra_pred_dc_16x16_msa(uint8_t *src, ptrdiff_t stride)
537 uint8_t *src_top = src - stride;
554 addition += src_left[ 0 * stride];
555 addition += src_left[ 1 * stride];
556 addition += src_left[ 2 * stride];
557 addition += src_left[ 3 * stride];
558 addition += src_left[ 4 * stride];
559 addition += src_left[ 5 * stride];
560 addition += src_left[ 6 * stride];
561 addition += src_left[ 7 * stride];
562 addition += src_left[ 8 * stride];
563 addition += src_left[ 9 * stride];
564 addition += src_left[10 * stride];
565 addition += src_left[11 * stride];
566 addition += src_left[12 * stride];
567 addition += src_left[13 * stride];
568 addition += src_left[14 * stride];
569 addition += src_left[15 * stride];
573 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
574 dst += (8 * stride);
575 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
578 void ff_h264_intra_pred_vert_16x16_msa(uint8_t *src, ptrdiff_t stride)
582 intra_predict_vert_16x16_msa(src - stride, dst, stride);
585 void ff_h264_intra_pred_horiz_16x16_msa(uint8_t *src, ptrdiff_t stride)
589 intra_predict_horiz_16x16_msa(src - 1, stride, dst, stride);
592 void ff_h264_intra_pred_dc_left_16x16_msa(uint8_t *src, ptrdiff_t stride)
599 addition = src_left[ 0 * stride];
600 addition += src_left[ 1 * stride];
601 addition += src_left[ 2 * stride];
602 addition += src_left[ 3 * stride];
603 addition += src_left[ 4 * stride];
604 addition += src_left[ 5 * stride];
605 addition += src_left[ 6 * stride];
606 addition += src_left[ 7 * stride];
607 addition += src_left[ 8 * stride];
608 addition += src_left[ 9 * stride];
609 addition += src_left[10 * stride];
610 addition += src_left[11 * stride];
611 addition += src_left[12 * stride];
612 addition += src_left[13 * stride];
613 addition += src_left[14 * stride];
614 addition += src_left[15 * stride];
619 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
620 dst += (8 * stride);
621 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
624 void ff_h264_intra_pred_dc_top_16x16_msa(uint8_t *src, ptrdiff_t stride)
626 uint8_t *src_top = src - stride;
643 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
644 dst += (8 * stride);
645 ST_UB8(out, out, out, out, out, out, out, out, dst, stride);
648 void ff_h264_intra_pred_dc_128_8x8_msa(uint8_t *src, ptrdiff_t stride)
656 SD4(out, out, out, out, src, stride);
657 src += (4 * stride);
658 SD4(out, out, out, out, src, stride);
661 void ff_h264_intra_pred_dc_128_16x16_msa(uint8_t *src, ptrdiff_t stride)
667 ST_UB8(out, out, out, out, out, out, out, out, src, stride);
668 src += (8 * stride);
669 ST_UB8(out, out, out, out, out, out, out, out, src, stride);
672 void ff_vp8_pred8x8_127_dc_8_msa(uint8_t *src, ptrdiff_t stride)
674 intra_predict_127dc_8x8_msa(src, stride);
677 void ff_vp8_pred8x8_129_dc_8_msa(uint8_t *src, ptrdiff_t stride)
679 intra_predict_129dc_8x8_msa(src, stride);
682 void ff_vp8_pred16x16_127_dc_8_msa(uint8_t *src, ptrdiff_t stride)
684 intra_predict_127dc_16x16_msa(src, stride);
687 void ff_vp8_pred16x16_129_dc_8_msa(uint8_t *src, ptrdiff_t stride)
689 intra_predict_129dc_16x16_msa(src, stride);