/third_party/ffmpeg/libavcodec/ |
H A D | hq_hqadsp.c | 37 int tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8, tmp9, tmpA; in idct_row() local 48 tmp8 = tmp3 + tmp1; in idct_row() 49 tmp9 = tmp7 * 4 - tmp8; in idct_row() 63 blk[0] = tmp14 + tmp8; in idct_row() 70 blk[7] = tmp14 - tmp8; in idct_row() 75 int tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8, tmp9, tmpA; in idct_col() local 86 tmp8 = (tmp3 + tmp1) >> 1; in idct_col() 87 tmp9 = tmp7 * 2 - tmp8; in idct_col() 101 blk[0 * 8] = (tmp14 + tmp8) >> 6; in idct_col() 108 blk[7 * 8] = (tmp14 - tmp8) >> in idct_col() [all...] |
H A D | fft_template.c | 326 unsigned tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in fft_calc_c() local 342 tmp8 = tmpz[2].im - (unsigned)tmpz[3].im; in fft_calc_c() 350 tmpz[1].re = tmp3 + tmp8; in fft_calc_c() 351 tmpz[3].re = tmp3 - tmp8; in fft_calc_c() 372 tmp8 = tmp2 - tmp4; in fft_calc_c() 383 tmpz[6].re = tmpz[2].re - tmp8; in fft_calc_c() 384 tmpz[2].re = tmpz[2].re + tmp8; in fft_calc_c() 395 tmp8 = (int32_t)((accu + 0x40000000) >> 31); in fft_calc_c() 398 tmp2 = tmp6 + tmp8; in fft_calc_c() 399 tmp4 = tmp6 - tmp8; in fft_calc_c() [all...] |
H A D | cavsdsp.c | 382 const int tmp8= tmp[8 *8]; \ 390 OP(dst[5*dstStride], AV*tmp3 + BV*tmp4 + CV*tmp5 + DV*tmp6 + EV*tmp7 + FV*tmp8 + 64*src2[5*srcStride]); \ 391 OP(dst[6*dstStride], AV*tmp4 + BV*tmp5 + CV*tmp6 + DV*tmp7 + EV*tmp8 + FV*tmp9 + 64*src2[6*srcStride]); \ 392 OP(dst[7*dstStride], AV*tmp5 + BV*tmp6 + CV*tmp7 + DV*tmp8 + EV*tmp9 + FV*tmp10 + 64*src2[7*srcStride]); \ 411 const int tmp8= tmp[8 *8]; \ 419 OP(dst[5*dstStride], AV*tmp3 + BV*tmp4 + CV*tmp5 + DV*tmp6 + EV*tmp7 + FV*tmp8); \ 420 OP(dst[6*dstStride], AV*tmp4 + BV*tmp5 + CV*tmp6 + DV*tmp7 + EV*tmp8 + FV*tmp9); \ 421 OP(dst[7*dstStride], AV*tmp5 + BV*tmp6 + CV*tmp7 + DV*tmp8 + EV*tmp9 + FV*tmp10); \
|
H A D | h264qpel_template.c | 337 const int tmp8= tmp[8 *tmpStride] - pad;\ 345 OP2(dst[5*dstStride], (tmp5+tmp6)*20 - (tmp4+tmp7)*5 + (tmp3+tmp8));\ 346 OP2(dst[6*dstStride], (tmp6+tmp7)*20 - (tmp5+tmp8)*5 + (tmp4+tmp9));\ 347 OP2(dst[7*dstStride], (tmp7+tmp8)*20 - (tmp6+tmp9)*5 + (tmp5+tmp10));\
|
/third_party/ffmpeg/libavcodec/mips/ |
H A D | fft_mips.c | 66 FFTSample tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in ff_fft_calc_mips() local 88 tmp8 = tmpz[2].im - tmpz[3].im; in ff_fft_calc_mips() 96 tmpz[1].re = tmp3 + tmp8; in ff_fft_calc_mips() 97 tmpz[3].re = tmp3 - tmp8; in ff_fft_calc_mips() 131 "sub.s %[tmp8], %[tmp2], %[tmp4] \n\t" // tmp8 = tmp2 - tmp4; in ff_fft_calc_mips() 154 "sub.s %[temp], %[pom1], %[tmp8] \n\t" in ff_fft_calc_mips() 158 "swc1 %[temp], 48(%[tmpz]) \n\t" // tmpz[6].re = tmpz[2].re - tmp8; in ff_fft_calc_mips() 160 "add.s %[pom1], %[pom1], %[tmp8] \n\t" in ff_fft_calc_mips() 166 "mul.s %[tmp8], in ff_fft_calc_mips() [all...] |
H A D | h264dsp_msa.c | 2223 v8i16 tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; in ff_weight_h264_pixels16_8_msa() local 2245 MUL4(wgt, src4_r, wgt, src4_l, wgt, src5_r, wgt, src5_l, tmp8, tmp9, tmp10, in ff_weight_h264_pixels16_8_msa() 2253 ADDS_SH4_SH(tmp8, offset, tmp9, offset, tmp10, offset, tmp11, offset, tmp8, in ff_weight_h264_pixels16_8_msa() 2258 MAXI_SH8_SH(tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, 0); in ff_weight_h264_pixels16_8_msa() 2260 SRLR_H8_SH(tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, denom); in ff_weight_h264_pixels16_8_msa() 2262 SAT_UH8_SH(tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, 7); in ff_weight_h264_pixels16_8_msa() 2265 PCKEV_B4_UB(tmp9, tmp8, tmp11, tmp10, tmp13, tmp12, tmp15, tmp14, dst4, in ff_weight_h264_pixels16_8_msa() 2284 MUL4(wgt, src4_r, wgt, src4_l, wgt, src5_r, wgt, src5_l, tmp8, tmp9, in ff_weight_h264_pixels16_8_msa() 2292 ADDS_SH4_SH(tmp8, offse in ff_weight_h264_pixels16_8_msa() 2347 v8i16 tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; ff_biweight_h264_pixels16_8_msa() local [all...] |
H A D | vp8_mc_msa.c | 2085 v8u16 hz_out0, hz_out1, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in common_hv_2ht_2vt_8x8mult_msa() local 2146 tmp8 = __msa_dotp_u_h(vec0, filt_vt); in common_hv_2ht_2vt_8x8mult_msa() 2148 SRARI_H4_UH(tmp5, tmp6, tmp7, tmp8, 7); in common_hv_2ht_2vt_8x8mult_msa() 2149 SAT_UH4_UH(tmp5, tmp6, tmp7, tmp8, 7); in common_hv_2ht_2vt_8x8mult_msa() 2150 PCKEV_B2_SB(tmp6, tmp5, tmp8, tmp7, out0, out1); in common_hv_2ht_2vt_8x8mult_msa()
|
H A D | vp9_mc_msa.c | 2632 v8u16 hz_out0, hz_out1, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in common_hv_2ht_2vt_8x8mult_msa() local 2693 tmp8 = __msa_dotp_u_h(vec0, filt_vt); in common_hv_2ht_2vt_8x8mult_msa() 2695 SRARI_H4_UH(tmp5, tmp6, tmp7, tmp8, 7); in common_hv_2ht_2vt_8x8mult_msa() 2696 SAT_UH4_UH(tmp5, tmp6, tmp7, tmp8, 7); in common_hv_2ht_2vt_8x8mult_msa() 2697 PCKEV_B2_SB(tmp6, tmp5, tmp8, tmp7, out0, out1); in common_hv_2ht_2vt_8x8mult_msa()
|
/third_party/ffmpeg/libswscale/ppc/ |
H A D | swscale_vsx.c | 486 tmp8 = (vec_u8) vec_mergeh((vec_u16) out0, (vec_u16) out1); \ 487 vec_vsx_st(tmp8, 0, dest); \ 488 tmp8 = (vec_u8) vec_mergel((vec_u16) out0, (vec_u16) out1); \ 489 vec_vsx_st(tmp8, 16, dest); \ 497 tmp8 = (vec_u8) vec_mergeh((vec_u16) out0, (vec_u16) out1); \ 498 vec_vsx_st(tmp8, 0, dest); \ 499 tmp8 = (vec_u8) vec_mergel((vec_u16) out0, (vec_u16) out1); \ 500 vec_vsx_st(tmp8, 16, dest); \ 508 tmp8 = (vec_u8) vec_mergeh((vec_u16) out0, (vec_u16) out1); \ 509 vec_vsx_st(tmp8, 541 vec_u8 rd, bd, gd, ad, out0, out1, tmp8; yuv2rgb_full_X_vsx_template() local 698 vec_u8 rd, bd, gd, ad, out0, out1, tmp8; yuv2rgb_full_2_vsx_template() local 820 vec_u8 rd, bd, gd, ad, out0, out1, tmp8; yuv2rgb_2_vsx_template() local 990 vec_u8 rd, bd, gd, ad, out0, out1, tmp8; yuv2rgb_full_1_vsx_template() local 1120 vec_u8 rd, bd, gd, ad, out0, out1, tmp8; yuv2rgb_1_vsx_template() local [all...] |
/third_party/ffmpeg/libavcodec/loongarch/ |
H A D | h264qpel_lasx.c | 1219 __m256i tmp7, tmp8, tmp9, tmp10, tmp11, tmp12; in put_h264_qpel8_hv_lowpass_lasx() local 1238 QPEL8_HV_LOWPASS_H(tmp8) in put_h264_qpel8_hv_lowpass_lasx() 1242 tmp9 = __lasx_xvpermi_q(tmp10, tmp8, 0x21); in put_h264_qpel8_hv_lowpass_lasx() 1243 tmp7 = __lasx_xvpermi_q(tmp8, tmp6, 0x21); in put_h264_qpel8_hv_lowpass_lasx() 1252 QPEL8_HV_LOWPASS_V(tmp4, tmp5, tmp6, tmp7, tmp8, tmp9, src00, src01, in put_h264_qpel8_hv_lowpass_lasx() 1254 QPEL8_HV_LOWPASS_V(tmp6, tmp7, tmp8, tmp9, tmp10, tmp11, src00, src01, in put_h264_qpel8_hv_lowpass_lasx() 1328 __m256i tmp7, tmp8, tmp9, tmp10, tmp11, tmp12; in avg_h264_qpel8_hv_lowpass_lasx() local 1350 QPEL8_HV_LOWPASS_H(tmp8) in avg_h264_qpel8_hv_lowpass_lasx() 1354 tmp9 = __lasx_xvpermi_q(tmp10, tmp8, 0x21); in avg_h264_qpel8_hv_lowpass_lasx() 1355 tmp7 = __lasx_xvpermi_q(tmp8, tmp in avg_h264_qpel8_hv_lowpass_lasx() [all...] |
H A D | vp9_intra_lsx.c | 515 __m128i tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; in ff_tm_16x16_lsx() local 527 7, tmp11, tmp10, tmp9, tmp8); in ff_tm_16x16_lsx() 560 DUP4_ARG2(__lsx_vaddwev_h_bu, tmp8, reg1, tmp9, reg1, tmp10, reg1, tmp11, in ff_tm_16x16_lsx() 562 DUP4_ARG2(__lsx_vaddwod_h_bu, tmp8, reg1, tmp9, reg1, tmp10, reg1, tmp11, in ff_tm_16x16_lsx() 573 tmp8, tmp9, tmp10, tmp11); in ff_tm_16x16_lsx() 591 LSX_ST_8(tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, dst, in ff_tm_16x16_lsx()
|
H A D | vp8_lpf_lsx.c | 544 __m128i tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; in ff_vp8_h_loop_filter16_inner_lsx() local 558 src + stride3, 0, tmp8, tmp9, tmp10, tmp11); in ff_vp8_h_loop_filter16_inner_lsx() 565 tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, in ff_vp8_h_loop_filter16_inner_lsx()
|
H A D | vp9_mc_lsx.c | 1377 __m128i tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; in common_hz_8t_and_aver_dst_16w_lsx() local 1401 mask2, src3, src3, mask2, tmp8, tmp9, tmp10, tmp11); in common_hz_8t_and_aver_dst_16w_lsx() 1406 DUP4_ARG2(__lsx_vdp2_h_b, tmp8, filter2, tmp9, filter2, tmp10, filter2, tmp11, in common_hz_8t_and_aver_dst_16w_lsx() 1407 filter2, tmp8, tmp9, tmp10, tmp11); in common_hz_8t_and_aver_dst_16w_lsx() 1410 DUP4_ARG3(__lsx_vdp2add_h_b, tmp8, tmp12, filter3, tmp9, tmp13, filter3, tmp10, in common_hz_8t_and_aver_dst_16w_lsx() 1434 __m128i tmp8, tmp9, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15; in common_hz_8t_and_aver_dst_32w_lsx() local 1458 src2, mask2, src3, src3, mask2, tmp8, tmp9, tmp10, tmp11); in common_hz_8t_and_aver_dst_32w_lsx() 1463 DUP4_ARG2(__lsx_vdp2_h_b, tmp8, filter2, tmp9, filter2, tmp10, filter2, in common_hz_8t_and_aver_dst_32w_lsx() 1464 tmp11, filter2, tmp8, tmp9, tmp10, tmp11); in common_hz_8t_and_aver_dst_32w_lsx() 1467 DUP4_ARG3(__lsx_vdp2add_h_b, tmp8, tmp1 in common_hz_8t_and_aver_dst_32w_lsx() [all...] |
H A D | vp8_mc_lsx.c | 740 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in ff_put_vp8_epel8_h4v6_lsx() local 796 tmp8 = HORIZ_4TAP_FILT(src8, src8, mask0, mask1, filt_hz0, filt_hz1); in ff_put_vp8_epel8_h4v6_lsx() 797 out7 = __lsx_vpackev_b(tmp8, tmp7); in ff_put_vp8_epel8_h4v6_lsx() 812 tmp4 = tmp8; in ff_put_vp8_epel8_h4v6_lsx()
|
H A D | hevc_mc_bi_lsx.c | 1506 __m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8; in hevc_hv_4t_6w_lsx() local 1576 DUP2_ARG2(__lsx_vpickev_d, tmp3, tmp1, tmp7, tmp5, tmp0, tmp8); in hevc_hv_4t_6w_lsx() 1578 dst0_l = __lsx_vdp2add_w_h(dst0_l, tmp8, filt_h1); in hevc_hv_4t_6w_lsx() 1592 DUP2_ARG2(__lsx_vdp2_w_h, tmp8, filt_h0, tmp0, filt_h0, dst1_l, dst2_l); in hevc_hv_4t_6w_lsx()
|
/third_party/mesa3d/src/amd/compiler/tests/ |
H A D | test_optimizer_postRA.cpp | 359 Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1, reg_v2), a, dpp_row_mirror); variable 360 Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1, reg_v2), Operand(tmp8, reg_v2), b, c);
|
H A D | test_optimizer.cpp | 1029 Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); variable 1030 Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), tmp8, b, c);
|
/third_party/ffmpeg/libavcodec/aarch64/ |
H A D | vp9lpf_neon.S | 159 .macro loop_filter wd, sz, mix, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8 347 uaddl_sz \tmp7\().8h, \tmp8\().8h, v23, v26, \sz 353 sub_sz \tmp7\().8h, \tmp8\().8h, \tmp7\().8h, \tmp8\().8h, \tmp5\().8h, \tmp6\().8h, \sz 361 add_sz v0.8h, v1.8h, v0.8h, v1.8h, \tmp7\().8h, \tmp8\().8h, \sz 364 uaddl_sz \tmp7\().8h, \tmp8\().8h, v25, v27, \sz 368 sub_sz \tmp7\().8h, \tmp8\().8h, \tmp7\().8h, \tmp8\().8h, \tmp5\().8h, \tmp6\().8h, \sz 373 add_sz v0.8h, v1.8h, v0.8h, v1.8h, \tmp7\().8h, \tmp8\().8h, \sz
|
H A D | vp9lpf_16bpp_neon.S | 30 .macro loop_filter wd, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8
|
/third_party/skia/third_party/externals/libwebp/src/enc/ |
H A D | predictor_enc.c | 448 uint8_t* const tmp8 = current_max_diffs; in CopyImageWithPrediction() local 450 lower_max_diffs = tmp8; in CopyImageWithPrediction()
|
/third_party/ffmpeg/libavcodec/arm/ |
H A D | vp9lpf_neon.S | 155 @ tmp1,tmp2 = tmpq1, tmp3,tmp4 = tmpq2, tmp5,tmp6 = tmpq3, tmp7,tmp8 = tmpq4 156 .macro loop_filter wd, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp8, tmpq1, tmpq2, tmpq3, tmpq4
|