/third_party/ffmpeg/libavcodec/mips/ |
H A D | h264qpel_msa.c | 433 v16u8 out0, out1, dst0 = { 0 }, dst1 = { 0 }; in avc_luma_hv_qrt_and_aver_dst_8x8_msa() local 494 INSERT_D2_UB(tp2, tp3, dst1); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 499 AVER_UB2_UB(out0, dst0, out1, dst1, dst0, dst1); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 500 ST_D4(dst0, dst1, 0, 1, 0, 1, dst, stride); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 535 INSERT_D2_UB(tp2, tp3, dst1); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 540 AVER_UB2_UB(out0, dst0, out1, dst1, dst0, dst1); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 541 ST_D4(dst0, dst1, 0, 1, 0, 1, dst, stride); in avc_luma_hv_qrt_and_aver_dst_8x8_msa() 557 v16u8 tmp0, tmp1, dst0 = { 0 }, dst1 in avc_luma_hv_qrt_and_aver_dst_16x16_msa() local 676 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_avg_h264_qpel16_mc00_msa() local 704 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc00_msa() local 747 v16i8 dst0, dst1, dst2, dst3, src0, src1, src2, src3, src4, src5, src6; ff_put_h264_qpel16_mc10_msa() local 815 v16i8 dst0, dst1, dst2, dst3, src0, src1, src2, src3, src4, src5, src6; ff_put_h264_qpel16_mc30_msa() local 1607 v8i16 hz_out7, hz_out8, dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_put_h264_qpel16_mc21_msa() local 1720 v8i16 hz_out7, hz_out8, dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_put_h264_qpel16_mc23_msa() local 1833 v8i16 hz_out1110_r, hz_out1211_r, dst0, dst1, dst2, dst3; ff_put_h264_qpel8_mc21_msa() local 1968 v8i16 hz_out1110_r, hz_out1211_r, dst0, dst1, dst2, dst3; ff_put_h264_qpel8_mc23_msa() local 2100 v8i16 hz_out7, hz_out8, dst0, dst1, filt0, filt1, filt2; ff_put_h264_qpel4_mc21_msa() local 2164 v8i16 hz_out7, hz_out8, dst0, dst1, filt0, filt1, filt2; ff_put_h264_qpel4_mc23_msa() local 2380 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, mask3; ff_put_h264_qpel16_mc12_msa() local 2462 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, mask3; ff_put_h264_qpel16_mc32_msa() local 2544 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3; ff_put_h264_qpel8_mc12_msa() local 2619 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3; ff_put_h264_qpel8_mc32_msa() local 2700 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, shf_vec7; ff_put_h264_qpel4_mc12_msa() local 2792 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, shf_vec7; ff_put_h264_qpel4_mc32_msa() local 2889 v8i16 hz_out7, hz_out8, dst0, dst1, dst2, dst3; ff_put_h264_qpel16_mc22_msa() local 2989 v8i16 hz_out1110_r, hz_out1211_r, dst0, dst1, dst2, dst3; ff_put_h264_qpel8_mc22_msa() local 3100 v8i16 hz_out7, hz_out8, dst0, dst1, filt0, filt1, filt2; ff_put_h264_qpel4_mc22_msa() local 3149 v16u8 dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc10_msa() local 3221 v16u8 dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc30_msa() local 3293 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc10_msa() local 3358 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc30_msa() local 3495 v16u8 dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc20_msa() local 3648 v16u8 res0, res1, res2, res3, dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc01_msa() local 3725 v16u8 res0, res1, res2, res3, dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc03_msa() local 3801 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc01_msa() local 3867 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc03_msa() local 4126 v16u8 out0, out1, dst0 = { 0 }, dst1 = { 0 }; ff_avg_h264_qpel16_mc21_msa() local 4258 v16u8 out0, out1, dst0 = { 0 }, dst1 = { 0 }; ff_avg_h264_qpel16_mc23_msa() local 4384 v16u8 dst0 = { 0 }, dst1 = { 0 }, out0, out1; ff_avg_h264_qpel8_mc21_msa() local 4530 v16u8 dst0 = { 0 }, dst1 = { 0 }, out0, out1; ff_avg_h264_qpel8_mc23_msa() local 4680 v8i16 hz_out7, hz_out8, dst0, dst1, filt0, filt1, filt2; ff_avg_h264_qpel4_mc21_msa() local 4747 v8i16 hz_out7, hz_out8, dst0, dst1, filt0, filt1, filt2; ff_avg_h264_qpel4_mc23_msa() local 4811 v16u8 res0, res1, res2, res3, dst0, dst1, dst2, dst3; ff_avg_h264_qpel16_mc02_msa() local 4881 v16u8 dst0 = { 0 }, dst1 = { 0 }, dst2 = { 0 }, dst3 = { 0 }; ff_avg_h264_qpel8_mc02_msa() local 5321 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, shf_vec7; ff_avg_h264_qpel4_mc12_msa() local 5417 v8i16 vt_res0, vt_res1, vt_res2, vt_res3, dst0, dst1, dst2, dst3, shf_vec7; ff_avg_h264_qpel4_mc32_msa() local 5514 v16u8 dst0, dst1, out0, out1; ff_avg_h264_qpel16_mc22_msa() local 5615 v16u8 out0, out1, dst0 = { 0 }, dst1 = { 0 }; ff_avg_h264_qpel8_mc22_msa() local [all...] |
H A D | vp9_mc_msa.c | 148 #define PCKEV_AVG_ST8x4_UB(in0, in1, in2, in3, dst0, dst1, \ 155 AVER_UB2_UB(tmp0_m, dst0, tmp1_m, dst1, tmp0_m, tmp1_m); \ 1029 v16u8 dst0, dst1; in common_hz_8t_and_aver_dst_4x8_msa() local 1049 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst1); in common_hz_8t_and_aver_dst_4x8_msa() 1062 AVER_UB2_UB(res0, dst0, res2, dst1, res0, res2); in common_hz_8t_and_aver_dst_4x8_msa() 1090 v16u8 mask0, mask1, mask2, mask3, dst0, dst1; in common_hz_8t_and_aver_dst_8w_msa() local 1113 INSERT_D2_UB(tp2, tp3, dst1); in common_hz_8t_and_aver_dst_8w_msa() 1116 CONVERT_UB_AVG_ST8x4_UB(out0, out1, out2, out3, dst0, dst1, in common_hz_8t_and_aver_dst_8w_msa() 1130 v16u8 mask0, mask1, mask2, mask3, dst0, dst1; in common_hz_8t_and_aver_dst_16w_msa() local 1170 LD_UB2(dst, dst_stride, dst0, dst1); in common_hz_8t_and_aver_dst_16w_msa() 1188 v16u8 dst1, dst2, mask0, mask1, mask2, mask3; common_hz_8t_and_aver_dst_32w_msa() local 1247 v16u8 dst1, dst2, mask0, mask1, mask2, mask3; common_hz_8t_and_aver_dst_64w_msa() local 1369 v16u8 dst0, dst1; common_vt_8t_and_aver_dst_8w_msa() local 1437 v16u8 dst0, dst1, dst2, dst3, tmp0, tmp1, tmp2, tmp3; common_vt_8t_and_aver_dst_16w_mult_msa() local 1644 v16u8 dst0, dst1, mask0, mask1, mask2, mask3; common_hv_8ht_8vt_and_aver_dst_8w_msa() local 2855 v16u8 dst0, dst1; common_hz_2t_and_aver_dst_4x8_msa() local 2903 v16u8 filt0, dst0, dst1; common_hz_2t_and_aver_dst_8x4_msa() local 2933 v16u8 filt0, dst0, dst1; common_hz_2t_and_aver_dst_8x8mult_msa() local 3018 v16u8 filt0, dst0, dst1, dst2, dst3; ff_avg_bilin_16h_msa() local 3086 v16u8 filt0, dst0, dst1, dst2, dst3; ff_avg_bilin_32h_msa() local 3136 v16u8 filt0, dst0, dst1, dst2, dst3; ff_avg_bilin_64h_msa() local 3213 v16u8 dst0, dst1; common_vt_2t_and_aver_dst_4x8_msa() local 3269 v16u8 dst0, dst1, vec0, vec1, vec2, vec3, filt0; common_vt_2t_and_aver_dst_8x4_msa() local 3300 v16u8 dst0, dst1, dst2, dst3; common_vt_2t_and_aver_dst_8x8mult_msa() local 3366 v16u8 src0, src1, src2, src3, src4, dst0, dst1, dst2, dst3, filt0; ff_avg_bilin_16v_msa() local 3421 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_avg_bilin_32v_msa() local 3502 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_avg_bilin_64v_msa() local 3635 v16u8 dst0, dst1; common_hv_2ht_2vt_and_aver_dst_4x8_msa() local 3702 v16u8 filt_hz, filt_vt, dst0, dst1, vec0, vec1, vec2, vec3; common_hv_2ht_2vt_and_aver_dst_8x4_msa() local 3754 v16u8 filt_hz, filt_vt, vec0, dst0, dst1; common_hv_2ht_2vt_and_aver_dst_8x8mult_msa() local 3831 v16u8 filt_hz, filt_vt, vec0, vec1, dst0, dst1, dst2, dst3; ff_avg_bilin_16hv_msa() local 4060 v16u8 src0 = { 0 }, src1 = { 0 }, dst0 = { 0 }, dst1 = { 0 }; avg_width4_msa() local 4091 v16u8 dst0, dst1, dst2, dst3; avg_width8_msa() local 4132 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; avg_width16_msa() local 4169 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; avg_width32_msa() local 4232 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; avg_width64_msa() local [all...] |
H A D | hevc_lpf_sao_msa.c | 43 v16u8 dst0, dst1, dst2, dst3, dst4, dst5; in hevc_loopfilter_luma_hor_msa() local 150 dst1 = (v16u8) (temp2 + (v8i16) p1_src); in hevc_loopfilter_luma_hor_msa() 159 dst1 = __msa_bmz_v(dst1, (v16u8) p1_src, (v16u8) p_is_pcm_vec); in hevc_loopfilter_luma_hor_msa() 188 PCKEV_B2_UB(dst1, dst0, dst3, dst2, dst0, dst1); in hevc_loopfilter_luma_hor_msa() 196 dst1 = __msa_bmz_v(dst1, dst4, (v16u8) cmp3); in hevc_loopfilter_luma_hor_msa() 202 ST_D4(dst0, dst1, 0, 1, 0, 1, p2, stride); in hevc_loopfilter_luma_hor_msa() 271 dst1 in hevc_loopfilter_luma_hor_msa() 466 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; hevc_loopfilter_luma_ver_msa() local 1111 v16i8 offset0, offset1, dst0, dst1; hevc_sao_band_filter_8width_msa() local 1458 v16u8 src10, src11, src12, src13, dst0, dst1, dst2, dst3; hevc_sao_edge_filter_0degree_16multiple_msa() local 1782 v16u8 src10, src_minus10, dst0, src11, src_minus11, dst1; hevc_sao_edge_filter_90degree_16multiple_msa() local 2114 v16u8 src10, src_minus10, dst0, src11, src_minus11, dst1; hevc_sao_edge_filter_45degree_16multiple_msa() local 2455 v16u8 dst0, dst1, dst2, dst3; hevc_sao_edge_filter_135degree_16multiple_msa() local [all...] |
H A D | hpeldsp_msa.c | 43 #define PCKEV_AVG_ST8x4_UB(in1, dst0, in2, dst1, in3, dst2, in4, dst3, \ 50 PCKEV_D2_UB(dst1, dst0, dst3, dst2, tmp2_m, tmp3_m); \ 219 uint32_t dst0, dst1, out0, out1; in common_hz_bil_and_aver_dst_4w_msa() local 232 dst1 = LW(dst + dst_stride); in common_hz_bil_and_aver_dst_4w_msa() 234 tmp1 = (v16u8) __msa_insert_w((v4i32) tmp1, 0, dst1); in common_hz_bil_and_aver_dst_4w_msa() 446 uint32_t out0, out1, dst0, dst1; in common_vt_bil_and_aver_dst_4w_msa() local 459 dst1 = LW(dst + dst_stride); in common_vt_bil_and_aver_dst_4w_msa() 461 tmp1 = (v16u8) __msa_insert_w((v4i32) tmp1, 0, dst1); in common_vt_bil_and_aver_dst_4w_msa() 504 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in common_vt_bil_and_aver_dst_16w_msa() local 517 LD_UB8(dst, dst_stride, dst0, dst1, dst in common_vt_bil_and_aver_dst_16w_msa() 934 v16u8 dst0, dst1, res0, res1; common_hv_bil_and_aver_dst_4w_msa() local 974 v16u8 dst0, dst1, dst2, dst3; common_hv_bil_and_aver_dst_8w_msa() local 1017 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; common_hv_bil_and_aver_dst_16w_msa() local 1233 v16u8 dst0, dst1, dst2, dst3; avg_width4_msa() local 1278 v16u8 dst0, dst1, dst2, dst3; avg_width8_msa() local 1303 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; avg_width16_msa() local [all...] |
H A D | hevc_idct_msa.c | 194 v8i16 filt0, filt1, dst0, dst1; \ 247 PCKEV_H2_SH(res0_l, res0_r, res1_l, res1_r, dst0, dst1); \ 249 ST_SH(dst1, (buf_ptr + ((15 - (j * 2)) * 16))); \ 255 PCKEV_H2_SH(res0_l, res0_r, res1_l, res1_r, dst0, dst1); \ 257 ST_SH(dst1, (ptr1 - (((j / 2 + j % 2) * 2 * k) * 16))); \ 718 uint32_t dst0, dst1, dst2, dst3; in hevc_addblk_4x4_msa() local 724 LW4(dst, stride, dst0, dst1, dst2, dst3); in hevc_addblk_4x4_msa() 725 INSERT_W4_SW(dst0, dst1, dst2, dst3, dst_vec); in hevc_addblk_4x4_msa() 736 uint64_t dst0, dst1, dst2, dst3; in hevc_addblk_8x8_msa() local 744 LD4(temp_dst, stride, dst0, dst1, dst in hevc_addblk_8x8_msa() 773 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; hevc_addblk_16x16_msa() local 840 v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; hevc_addblk_32x32_msa() local 943 v8i16 in0, in1, dst0, dst1; hevc_idct_luma_4x4_msa() local [all...] |
H A D | aacdec_mips.c | 152 float *dst1 = dst0 + 64 + 63; in imdct_and_windowing_mips() local 172 dst1[0] = temp0 * wi + temp1 * wj; in imdct_and_windowing_mips() 180 dst1[128] = temp2 * wi + temp3 * wj; in imdct_and_windowing_mips() 185 dst1[256] = temp0 * wi + temp1 * wj; in imdct_and_windowing_mips() 187 dst1[384] = temp2 * wi + temp3 * wj; in imdct_and_windowing_mips() 203 dst1--; in imdct_and_windowing_mips()
|
H A D | h264idct_msa.c | 122 v16i8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avc_idct8_addblk_msa() local 227 LD_SB8(dst, dst_stride, dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7); in avc_idct8_addblk_msa() 228 ILVR_B4_SH(zeros, dst0, zeros, dst1, zeros, dst2, zeros, dst3, in avc_idct8_addblk_msa() 238 dst0, dst1, dst2, dst3); in avc_idct8_addblk_msa() 239 ST_D8(dst0, dst1, dst2, dst3, 0, 1, 0, 1, 0, 1, 0, 1, dst, dst_stride) in avc_idct8_addblk_msa() 246 v16i8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; in avc_idct8_dc_addblk_msa() local 256 LD_SB8(dst, dst_stride, dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7); in avc_idct8_dc_addblk_msa() 257 ILVR_B4_SH(zeros, dst0, zeros, dst1, zeros, dst2, zeros, dst3, in avc_idct8_dc_addblk_msa() 268 dst0, dst1, dst2, dst3); in avc_idct8_dc_addblk_msa() 269 ST_D8(dst0, dst1, dst in avc_idct8_dc_addblk_msa() 282 const uint8_t *dst1 = dst + dst_stride; ff_h264_idct_add_msa() local [all...] |
/third_party/ltp/testcases/kernel/syscalls/string/ |
H A D | string01.c | 71 char dst1[LONGSTR + 1]; variable 146 dst1, longstr, longstr, 0}, { 203 dst1, longstr, LONGSTR, longstr, 0}, 230 dst1, tiat, 0, "", 0}, { 231 dst1, longstr, 5, "ttttt", 0}, { 345 memset(dst1, 0, LONGSTR + 1); /* clean slate */ in main() 378 memset(dst1, 0, LONGSTR + 1); /* clean slate */ in main() 417 memset(dst1, 0, LONGSTR + 1); /* clean slate */ in main()
|
/third_party/ffmpeg/libavcodec/ppc/ |
H A D | fmtconvert_altivec.c | 39 vector float src1, src2, dst1, dst2, mul_v, zero; in int32_to_float_fmul_scalar_altivec() local 48 dst1 = vec_madd(src1, mul_v, zero); in int32_to_float_fmul_scalar_altivec() 50 vec_st(dst1, 0, dst+i); in int32_to_float_fmul_scalar_altivec()
|
/third_party/ffmpeg/libavfilter/ |
H A D | vf_yadif.c | 92 static void filter_line_c(void *dst1, in filter_line_c() argument 96 uint8_t *dst = dst1; in filter_line_c() 112 static void filter_edges(void *dst1, void *prev1, void *cur1, void *next1, in filter_edges() argument 115 uint8_t *dst = dst1; in filter_edges() 130 dst = (uint8_t*)dst1 + offset; in filter_edges() 143 static void filter_line_c_16bit(void *dst1, in filter_line_c_16bit() argument 148 uint16_t *dst = dst1; in filter_line_c_16bit() 161 static void filter_edges_16bit(void *dst1, void *prev1, void *cur1, void *next1, in filter_edges_16bit() argument 164 uint16_t *dst = dst1; in filter_edges_16bit() 180 dst = (uint16_t*)dst1 in filter_edges_16bit() [all...] |
/third_party/ffmpeg/libavcodec/loongarch/ |
H A D | h264dsp_lasx.c | 1011 __m256i dst0, dst1, dst2, dst3; in ff_biweight_h264_pixels16_8_lasx() local 1044 0x20, tmp7, tmp6, 0x20, dst0, dst1, dst2, dst3); in ff_biweight_h264_pixels16_8_lasx() 1048 DUP4_ARG2(__lasx_xvxori_b, dst0, 128, dst1, 128, dst2, 128, dst3, 128, in ff_biweight_h264_pixels16_8_lasx() 1049 dst0, dst1, dst2, dst3); in ff_biweight_h264_pixels16_8_lasx() 1050 DUP4_ARG2(__lasx_xvilvl_b, dst0, src0, dst1, src1, dst2, src2, in ff_biweight_h264_pixels16_8_lasx() 1052 DUP4_ARG2(__lasx_xvilvh_b, dst0, src0, dst1, src1, dst2, src2, in ff_biweight_h264_pixels16_8_lasx() 1074 dst0, dst1, dst2, dst3); in ff_biweight_h264_pixels16_8_lasx() 1081 __lasx_xvstelm_d(dst1, dst, 0, 0); in ff_biweight_h264_pixels16_8_lasx() 1082 __lasx_xvstelm_d(dst1, dst, 8, 1); in ff_biweight_h264_pixels16_8_lasx() 1084 __lasx_xvstelm_d(dst1, ds in ff_biweight_h264_pixels16_8_lasx() 1221 __m256i src0, src1, dst0, dst1; avc_biwgt_8x8_lasx() local 1287 __m256i src0, src1, src2, src3, dst0, dst1, dst2, dst3; avc_biwgt_8x16_lasx() local 2036 __m256i src0, dst0, dst1, dst2, dst3, zero; ff_h264_add_pixels4_8_lasx() local 2064 __m256i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_h264_add_pixels8_8_lasx() local [all...] |
H A D | hevc_lpf_sao_lsx.c | 47 __m128i dst0, dst1, dst2, dst3, dst4, dst5; in ff_hevc_loop_filter_luma_h_8_lsx() local 140 dst1 = __lsx_vadd_h(temp2, p1_src); in ff_hevc_loop_filter_luma_h_8_lsx() 151 DUP2_ARG3(__lsx_vbitsel_v, dst0, p2_src, p_is_pcm_vec, dst1, in ff_hevc_loop_filter_luma_h_8_lsx() 152 p1_src, p_is_pcm_vec, dst0, dst1); in ff_hevc_loop_filter_luma_h_8_lsx() 186 DUP2_ARG2(__lsx_vpickev_b, dst1, dst0, dst3, dst2, dst0, dst1); in ff_hevc_loop_filter_luma_h_8_lsx() 195 DUP2_ARG3(__lsx_vbitsel_v, dst0, dst3, cmp3, dst1, dst4, cmp3, in ff_hevc_loop_filter_luma_h_8_lsx() 196 dst0, dst1); in ff_hevc_loop_filter_luma_h_8_lsx() 201 __lsx_vstelm_d(dst1, p2 + stride_2x, 0, 0); in ff_hevc_loop_filter_luma_h_8_lsx() 202 __lsx_vstelm_d(dst1, p in ff_hevc_loop_filter_luma_h_8_lsx() 456 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_hevc_loop_filter_luma_v_8_lsx() local 1192 __m128i src10, src11, src12, src13, dst0, dst1, dst2, dst3; hevc_sao_edge_filter_0degree_16multiple_lsx() local 1516 __m128i src10, src_minus10, dst0, src11, src_minus11, dst1; hevc_sao_edge_filter_90degree_16multiple_lsx() local 1871 __m128i src10, src_minus10, dst0, src11, src_minus11, dst1; hevc_sao_edge_filter_45degree_16multiple_lsx() local 2241 __m128i dst0, dst1, dst2, dst3; hevc_sao_edge_filter_135degree_16multiple_lsx() local [all...] |
H A D | hevc_mc_uniw_lsx.c | 54 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7, dst8; in hevc_hv_8t_8x2_lsx() local 100 vec12, filt0, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 101 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec1, filt1, dst1, vec5, filt1, in hevc_hv_8t_8x2_lsx() 102 dst2, vec9, filt1, dst3, vec13, filt1, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 103 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec2, filt2, dst1, vec6, filt2, in hevc_hv_8t_8x2_lsx() 104 dst2, vec10, filt2, dst3, vec14, filt2, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 105 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec3, filt3, dst1, vec7, filt3, in hevc_hv_8t_8x2_lsx() 106 dst2, vec11, filt3, dst3, vec15, filt3, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 121 DUP4_ARG2(__lsx_vilvl_h, dst1, dst0, dst3, dst2, dst5, dst4, dst2, in hevc_hv_8t_8x2_lsx() 122 dst1, dst10_ in hevc_hv_8t_8x2_lsx() [all...] |
H A D | hevc_mc_uni_lsx.c | 358 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7, dst8; in hevc_hv_8t_8x2_lsx() local 402 vec12, filt0, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 403 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec1, filt1, dst1, vec5, filt1, in hevc_hv_8t_8x2_lsx() 404 dst2, vec9, filt1, dst3, vec13, filt1, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 405 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec2, filt2, dst1, vec6, filt2, in hevc_hv_8t_8x2_lsx() 406 dst2, vec10, filt2, dst3, vec14, filt2, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 407 DUP4_ARG3(__lsx_vdp2add_h_bu_b, dst0, vec3, filt3, dst1, vec7, filt3, in hevc_hv_8t_8x2_lsx() 408 dst2, vec11, filt3, dst3, vec15, filt3, dst0, dst1, dst2, dst3); in hevc_hv_8t_8x2_lsx() 423 DUP4_ARG2(__lsx_vilvl_h, dst1, dst0, dst3, dst2, dst5, dst4, dst2, in hevc_hv_8t_8x2_lsx() 424 dst1, dst10_ in hevc_hv_8t_8x2_lsx() 762 __m128i dst0, dst1, dst2, dst3, dst4; hevc_hv_4t_8x2_lsx() local 824 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, tmp0, tmp1, tmp2, tmp3; hevc_hv_4t_8multx4_lsx() local 923 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7, dst8; hevc_hv_4t_8x6_lsx() local 1034 __m128i dst0, dst1, dst2, dst3, dst4, dst5; hevc_hv_4t_8multx4mult_lsx() local 1172 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; hevc_hv_4t_12w_lsx() local [all...] |
H A D | vp9_intra_lsx.c | 437 __m128i dst0, dst1, dst2, dst3; in ff_tm_4x4_lsx() local 446 src3, dst0, dst1, dst2, dst3); in ff_tm_4x4_lsx() 447 DUP4_ARG2(__lsx_vssub_hu, dst0, reg0, dst1, reg0, dst2, reg0, dst3, reg0, in ff_tm_4x4_lsx() 448 dst0, dst1, dst2, dst3); in ff_tm_4x4_lsx() 449 DUP4_ARG2(__lsx_vsat_hu, dst0, 7, dst1, 7, dst2, 7, dst3, 7, in ff_tm_4x4_lsx() 450 dst0, dst1, dst2, dst3); in ff_tm_4x4_lsx() 451 DUP2_ARG2(__lsx_vpickev_b, dst1, dst0, dst3, dst2, dst0, dst1); in ff_tm_4x4_lsx() 456 __lsx_vstelm_w(dst1, dst, 0, 0); in ff_tm_4x4_lsx() 458 __lsx_vstelm_w(dst1, ds in ff_tm_4x4_lsx() 602 __m128i dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; ff_tm_32x32_lsx() local [all...] |
/third_party/ffmpeg/libswscale/ |
H A D | hscale_fast_bilinear.c | 38 void ff_hcscale_fast_c(SwsContext *c, int16_t *dst1, int16_t *dst2, in ff_hcscale_fast_c() argument 47 dst1[i] = (src1[xx] * (xalpha ^ 127) + src1[xx + 1] * xalpha); in ff_hcscale_fast_c() 52 dst1[i] = src1[srcW-1]*128; in ff_hcscale_fast_c()
|
/third_party/skia/tests/ |
H A D | EncodeTest.cpp | 71 SkDynamicMemoryWStream dst0, dst1, dst2, dst3; in test_encode() local 75 auto encoder1 = make(format, &dst1, src); in test_encode() 92 sk_sp<SkData> data1 = dst1.detachAsData(); in test_encode() 194 SkDynamicMemoryWStream dst0, dst1, dst2; in DEF_TEST() local 200 success = SkJpegEncoder::Encode(&dst1, src, options); in DEF_TEST() 208 sk_sp<SkData> data1 = dst1.detachAsData(); in DEF_TEST() 306 SkDynamicMemoryWStream dst0, dst1, dst2; in DEF_TEST() local 312 success = SkPngEncoder::Encode(&dst1, src, options); in DEF_TEST() 322 sk_sp<SkData> data1 = dst1.detachAsData(); in DEF_TEST() 397 SkDynamicMemoryWStream dst0, dst1, dst in DEF_TEST() local [all...] |
/third_party/ffmpeg/tests/checkasm/ |
H A D | videodsp.c | 40 call_new((type *) dst1, (const type *) (src1 + y * pw + x), \ 43 if (memcmp(dst0, dst1, bw * bh * sizeof(type))) \ 45 bench_new((type *) dst1, (const type *) (src1 + y * pw + x),\ 68 LOCAL_ALIGNED_16(type, dst1, [64 * 64]); \
|
/third_party/musl/libc-test/src/functionalext/supplement/linux/ |
H A D | process_vm.c | 50 char dst1[__VALUE_BUFFER_SIZE__] = ""; in process_vm_writev_0200() local 53 {.iov_base = dst1, .iov_len = sizeof(dst1)}, in process_vm_writev_0200() 62 EXPECT_EQ("process_vm_writev_0200", rev, sizeof(dst1) + sizeof(dst2)); in process_vm_writev_0200()
|
/third_party/ffmpeg/libavcodec/aarch64/ |
H A D | vp9mc_neon.S | 163 // for size >= 16), and multiply-accumulate into dst1 and dst3 (or 164 // dst1-dst2 and dst3-dst4 for size >= 16) 165 .macro extmla dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size 169 mla \dst1\().8h, v20.8h, v0.h[\offset] 176 mla \dst1\().8h, v20.8h, v0.h[\offset] 179 mla \dst1\().4h, v20.4h, v0.h[\offset] 185 .macro extmulqadd dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size 203 sqadd \dst1\().4h, \dst1\().4h, v20.4h 206 sqadd \dst1\() [all...] |
H A D | vp9mc_16bpp_neon.S | 120 // for size >= 16), and multiply-accumulate into dst1 and dst5 (or 121 // dst1-dst2 and dst5-dst6 for size >= 8 and dst1-dst4 and dst5-dst8 123 .macro extmlal dst1, dst2, dst3, dst4, dst5, dst6, dst7, dst8, src1, src2, src3, src4, src5, src6, offset, size 126 smlal \dst1\().4s, v20.4h, v0.h[\offset] 386 // Evaluate the filter twice in parallel, from the inputs src1-src9 into dst1-dst2 387 // (src1-src8 into dst1, src2-src9 into dst2). 388 .macro convolve4 dst1, dst2, src1, src2, src3, src4, src5, src6, src7, src8, src9, tmp1, tmp2 389 smull \dst1\().4s, \src1\().4h, v0.h[0] 393 smlal \dst1\() [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | gdv.c | 143 uint8_t *dst1 = dst + PREAMBLE_SIZE + y * w; in rescale() local 146 scaleup_rev(dst1, src1, w); in rescale() 151 uint8_t *dst1 = dst + PREAMBLE_SIZE + y * w; in rescale() local 153 memcpy(dst1, src1, w); in rescale() 159 uint8_t *dst1 = dst + PREAMBLE_SIZE + y * (w>>1); in rescale() local 161 scaledown(dst1, src1, w>>1); in rescale() 165 uint8_t *dst1 = dst + PREAMBLE_SIZE + y * w; in rescale() local 167 memcpy(dst1, src1, w); in rescale() 171 uint8_t *dst1 = dst + PREAMBLE_SIZE + y * w; in rescale() local 172 scaledown(dst1, dst in rescale() [all...] |
/third_party/skia/samplecode/ |
H A D | SamplePolyToPoly.cpp | 46 const SkScalar dst1[] = { in PolyToPolyView() local 51 (void) m2.setPolyToPoly((const SkPoint*)src1, (SkPoint*)dst1, 4); in PolyToPolyView() 118 const int dst1[] = { 5, 5 }; variable 119 doDraw(canvas, &paint, font, src1, dst1, 1);
|
/third_party/ffmpeg/libavcodec/arm/ |
H A D | vp9mc_neon.S | 210 @ for size >= 16), and multiply-accumulate into dst1 and dst3 (or 211 @ dst1-dst2 and dst3-dst4 for size >= 16) 212 .macro extmla dst1, dst2, dst3, dst4, dst1d, dst3d, src1, src2, src3, src4, src5, src6, offset, size 216 vmla_lane \dst1, q14, \offset 223 vmla_lane \dst1, q14, \offset 232 .macro extmulqadd dst1, dst2, dst3, dst4, dst1d, dst3d, src1, src2, src3, src4, src5, src6, offset, size 253 vqadd.s16 \dst1, \dst1, q14 493 @ Evaluate the filter twice in parallel, from the inputs src1-src9 into dst1-dst2 494 @ (src1-src8 into dst1, src [all...] |
/third_party/skia/third_party/externals/libpng/mips/ |
H A D | filter_msa_intrinsics.c | 467 v16u8 dst0, dst1; in png_read_filter_row_sub4_msa() local 489 ILVEV_W2_UB(src1, src2, src3, src4, dst0, dst1); in png_read_filter_row_sub4_msa() 490 dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); in png_read_filter_row_sub4_msa() 506 v16u8 src0, src1, src2, src3, src4, dst0, dst1; in png_read_filter_row_sub3_msa() local 530 VSHF_B2_UB(src1, src2, src3, src4, mask0, mask0, dst0, dst1); in png_read_filter_row_sub3_msa() 531 dst0 = (v16u8) __msa_vshf_b(mask1, (v16i8) dst1, (v16i8) dst0); in png_read_filter_row_sub3_msa() 551 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; in png_read_filter_row_avg4_msa() local 585 ILVEV_W2_UB(src6, src7, src8, src9, dst0, dst1); in png_read_filter_row_avg4_msa() 586 dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); in png_read_filter_row_avg4_msa() 604 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; in png_read_filter_row_avg3_msa() local 663 v16u8 src10, src11, src12, src13, dst0, dst1; png_read_filter_row_paeth4_msa() local 735 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; png_read_filter_row_paeth3_msa() local [all...] |