Lines Matching refs:half

357  * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
358 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
422 * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
423 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
425 put_pixels8_l2_8_lsx(uint8_t *dst, const uint8_t *src, const uint8_t *half,
444 "vld $vr8, %[half], 0x00 \n\t"
445 "vld $vr9, %[half], 0x08 \n\t"
446 "vld $vr10, %[half], 0x10 \n\t"
447 "vld $vr11, %[half], 0x18 \n\t"
448 "vld $vr12, %[half], 0x20 \n\t"
449 "vld $vr13, %[half], 0x28 \n\t"
450 "vld $vr14, %[half], 0x30 \n\t"
451 "vld $vr15, %[half], 0x38 \n\t"
477 : [dst]"+&r"(dst), [half]"+&r"(half), [src]"+&r"(src),
486 * put_pixels8_l2_8_lsx: dst = avg(src, half) , half stride is 8.
487 * avg_pixels8_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
489 avg_pixels8_l2_8_lsx(uint8_t *dst, const uint8_t *src, const uint8_t *half,
509 "vld $vr8, %[half], 0x00 \n\t"
510 "vld $vr9, %[half], 0x08 \n\t"
511 "vld $vr10, %[half], 0x10 \n\t"
512 "vld $vr11, %[half], 0x18 \n\t"
513 "vld $vr12, %[half], 0x20 \n\t"
514 "vld $vr13, %[half], 0x28 \n\t"
515 "vld $vr14, %[half], 0x30 \n\t"
516 "vld $vr15, %[half], 0x38 \n\t"
564 : [dst]"+&r"(dst), [tmp]"+&r"(tmp), [half]"+&r"(half),
631 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
632 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
733 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
734 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
736 put_pixels16_l2_8_lsx(uint8_t *dst, const uint8_t *src, uint8_t *half,
760 "vld $vr8, %[half], 0x00 \n\t"
761 "vld $vr9, %[half], 0x10 \n\t"
762 "vld $vr10, %[half], 0x20 \n\t"
763 "vld $vr11, %[half], 0x30 \n\t"
764 "vld $vr12, %[half], 0x40 \n\t"
765 "vld $vr13, %[half], 0x50 \n\t"
766 "vld $vr14, %[half], 0x60 \n\t"
767 "vld $vr15, %[half], 0x70 \n\t"
800 "vld $vr8, %[half], 0x80 \n\t"
801 "vld $vr9, %[half], 0x90 \n\t"
802 "vld $vr10, %[half], 0xa0 \n\t"
803 "vld $vr11, %[half], 0xb0 \n\t"
804 "vld $vr12, %[half], 0xc0 \n\t"
805 "vld $vr13, %[half], 0xd0 \n\t"
806 "vld $vr14, %[half], 0xe0 \n\t"
807 "vld $vr15, %[half], 0xf0 \n\t"
827 : [dst]"+&r"(dst), [half]"+&r"(half), [src]"+&r"(src),
837 * put_pixels16_l2_8_lsx: dst = avg(src, half) , half stride is 8.
838 * avg_pixels16_l2_8_lsx: dst = avg(avg(src, half), dst) , half stride is 8.*/
840 avg_pixels16_l2_8_lsx(uint8_t *dst, const uint8_t *src, uint8_t *half,
865 "vld $vr8, %[half], 0x00 \n\t"
866 "vld $vr9, %[half], 0x10 \n\t"
867 "vld $vr10, %[half], 0x20 \n\t"
868 "vld $vr11, %[half], 0x30 \n\t"
869 "vld $vr12, %[half], 0x40 \n\t"
870 "vld $vr13, %[half], 0x50 \n\t"
871 "vld $vr14, %[half], 0x60 \n\t"
872 "vld $vr15, %[half], 0x70 \n\t"
925 "vld $vr8, %[half], 0x80 \n\t"
926 "vld $vr9, %[half], 0x90 \n\t"
927 "vld $vr10, %[half], 0xa0 \n\t"
928 "vld $vr11, %[half], 0xb0 \n\t"
929 "vld $vr12, %[half], 0xc0 \n\t"
930 "vld $vr13, %[half], 0xd0 \n\t"
931 "vld $vr14, %[half], 0xe0 \n\t"
932 "vld $vr15, %[half], 0xf0 \n\t"
971 : [dst]"+&r"(dst), [tmp]"+&r"(tmp), [half]"+&r"(half), [src]"+&r"(src),
1480 uint8_t half[64];
1482 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1483 /* in qpel8, the stride of half and height of block is 8 */
1484 put_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1496 uint8_t half[64];
1498 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1499 put_pixels8_l2_8_lsx(dst, src+1, half, stride, stride);
1505 uint8_t half[64];
1507 put_h264_qpel8_v_lowpass_lasx(half, (uint8_t*)src, 8, stride);
1508 put_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1584 uint8_t half[64];
1586 put_h264_qpel8_v_lowpass_lasx(half, (uint8_t*)src, 8, stride);
1587 put_pixels8_l2_8_lsx(dst, src + stride, half, stride, stride);
1635 uint8_t half[64];
1637 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1638 avg_pixels8_l2_8_lsx(dst, src, half, stride, stride);
1650 uint8_t half[64];
1652 put_h264_qpel8_h_lowpass_lasx(half, src, 8, stride);
1653 avg_pixels8_l2_8_lsx(dst, src+1, half, stride, stride);
1771 uint8_t half[256];
1773 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1774 put_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1786 uint8_t half[256];
1788 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1789 put_pixels16_l2_8_lsx(dst, src+1, half, stride, stride);
1795 uint8_t half[256];
1797 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1798 put_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1866 uint8_t half[256];
1868 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1869 put_pixels16_l2_8_lsx(dst, src+stride, half, stride, stride);
1909 uint8_t half[256];
1911 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1912 avg_pixels16_l2_8_lsx(dst, src, half, stride, stride);
1924 uint8_t half[256];
1926 put_h264_qpel16_h_lowpass_lasx(half, src, 16, stride);
1927 avg_pixels16_l2_8_lsx(dst, src+1, half, stride, stride);
1933 uint8_t half[256];
1935 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
1936 avg_pixels16_l2_8_lsx(dst, src, half, stride, stride);
2006 uint8_t half[256];
2008 put_h264_qpel16_v_lowpass_lasx(half, src, 16, stride);
2009 avg_pixels16_l2_8_lsx(dst, src + stride, half, stride, stride);