/third_party/ffmpeg/libavcodec/arm/ |
H A D | aacpsdsp_neon.S | 27 vmul.f32 q0, q0, q0 29 vmul.f32 q2, q2, q2 35 vmul.f32 q0, q0, q0 39 vmul.f32 q2, q2, q2 57 vmul.f32 d4, d0, d6[0] 58 vmul.f32 d5, d1, d6[1] 60 vmul.f32 d6, d2, d7[0] 61 vmul.f32 d7, d3, d7[1] 67 vmul.f32 d4, d0, d6[0] 68 vmul [all...] |
H A D | mdct_neon.S | 43 vmul.f32 d6, d17, d2 44 vmul.f32 d7, d0, d2 48 vmul.f32 d4, d0, d3 49 vmul.f32 d5, d17, d3 61 vmul.f32 d6, d17, d2 62 vmul.f32 d7, d0, d2 94 vmul.f32 d7, d0, d18 96 vmul.f32 d4, d1, d18 97 vmul.f32 d5, d21, d19 98 vmul [all...] |
H A D | fmtconvert_neon.S | 38 vmul.f32 q9, q3, q0 39 vmul.f32 q10, q8, q0 69 vmul.f32 q0, q0, q8 71 vmul.f32 q1, q1, q8 72 vmul.f32 q2, q2, q9 73 vmul.f32 q3, q3, q9 84 vmul.f32 q0, q0, q8 85 vmul.f32 q1, q1, q8
|
H A D | fft_neon.S | 69 vmul.f32 d26, d17, d28 @ -a2r*w,a2i*w 71 vmul.f32 d27, d19, d29 @ a3r*w,-a3i*w 74 vmul.f32 d24, d17, d31 @ a2r*w,a2i*w 75 vmul.f32 d25, d19, d31 @ a3r*w,a3i*w 120 vmul.f32 d24, d22, d2 122 vmul.f32 d25, d23, d3 124 vmul.f32 q1, q11, d2[1] 162 vmul.f32 q14, q14, d4[1] 163 vmul.f32 q1, q1, q3 177 vmul [all...] |
H A D | fmtconvert_vfp.S | 63 vmul.f32 s16, s16, s2 76 vmul.f32 s24, s24, s3 90 vmul.f32 s8, s8, s1 104 vmul.f32 s16, s16, s2 119 vmul.f32 s24, s24, s3 141 vmul.f32 s8, s8, s1 157 vmul.f32 s24, s24, s3 175 vmul.f32 s8, s8, s0 211 vmul.f32 s8, s8, s0
|
H A D | mdct_vfp.S | 53 vmul.f s8, s0, s16 @ vector operation 61 vmul.f s12, s0, s20 @ vector operation 91 vmul.f s8, s0, s16 @ vector operation 98 vmul.f s12, s0, s20 @ vector operation 167 vmul.f s8, s4, s16 @ vector operation 176 vmul.f s12, s0, s16 @ vector operation 216 vmul.f s8, s4, s16 @ vector operation 225 vmul.f s12, s0, s16 @ vector operation
|
H A D | mpegvideo_neon.S | 47 vmul.s16 q2, q0, q15 49 vmul.s16 q10, q8, q15 68 vmul.s16 d2, d0, d30
|
H A D | lossless_audiodsp_neon.S | 39 vmul.s16 q10, q10, q14 40 vmul.s16 q13, q13, q14
|
H A D | rdft_neon.S | 79 vmul.f32 q10, q0, q9 @ ev.re, ev.im, od.im, od.re 83 vmul.f32 q11, q0, q9 @ ev.re, ev.im, od.im, od.re 118 vmul.f32 q10, q0, q9 @ ev.re, ev.im, od.im, od.re 146 vmul.f32 d22, d22, d18
|
H A D | fft_vfp.S | 120 vmul.f s20, s20, s0 @ vector x scalar op 238 vmul.f s20, s12, s3 @ vector op 243 vmul.f s24, s4, s2 @ vector * scalar op 244 vmul.f s28, s12, s1 @ vector * scalar op 245 vmul.f s12, s8, s1 @ vector * scalar op 250 vmul.f s8, s8, s3 @ vector * scalar op 347 vmul.f s20, s16, s2 @ vector * scalar 351 vmul.f s16, s16, s1 @ vector * scalar 406 vmul.f s12, s8, s3 @ vector * scalar 410 vmul [all...] |
H A D | vp9mc_neon.S | 193 @ Helper macros for vmul/vmla with a constant from either d0 or d1 depending on index 196 vmul.s16 \dst, \src, d0[\idx] 198 vmul.s16 \dst, \src, d1[\idx - 4] 316 vmul.s16 q1, q8, d0[0] 317 vmul.s16 q3, q11, d0[0] 319 vmul.s16 q2, q9, d0[0] 320 vmul.s16 q4, q12, d0[0] 499 vmul.s16 \dst1, \src2, d0[1] 500 vmul.s16 \dst2, \src3, d0[1] 501 vmul [all...] |
H A D | h264pred_neon.S | 128 vmul.s16 q2, q2, q0 129 vmul.s16 q3, q3, q0 150 vmul.i16 q0, q0, d4[0] 209 vmul.s16 d4, d4, d0 210 vmul.s16 d5, d5, d0 229 vmul.i16 q0, q0, d4[0]
|
H A D | synth_filter_neon.S | 91 vmul.f32 q8, q10, d0[0] 92 vmul.f32 q9, q1, d0[0]
|
H A D | sbrdsp_neon.S | 199 vmul.f32 q3, q0, q1 207 vmul.f32 d0, d0, d2 215 vmul.f32 d0, d1, d1 218 vmul.f32 q0, q0, q1 253 vmul.f32 d21, d1, d1
|
H A D | vc1dsp_neon.S | 99 vmul.i16 q10, q1, d0[1] @ t3 = 22 * (src[1]) 100 vmul.i16 q11, q3, d0[1] @ t4 = 22 * (src[3]) 170 vmul.i16 q12, q3, d0[3] @ temp3|temp4 = 9 * src[8]|src[24] 244 vmul.i16 q2, q10, q0 @ t4 = 6/2 * src[16] 282 vmul.i16 q14, q9, d0[0] @ t2 = 15 * src[8] 283 vmul.i16 q9, q9, d0[1] @ t3 = 9 * src[8] 585 vmul.i16 q3, q14, q1 @ t3|t4 = 22 * (src[1]|src[3]) 1220 vmul.i16 d0, d4, d0[1] @ a0 >= a3 ? 5*(a0-a3) : 0 1285 vmul.i16 d0, d3, d0[1] @ a0 >= a3 ? 5*(a0-a3) : 0 1360 vmul [all...] |
H A D | vp8dsp_neon.S | 328 vmul.i16 q10, q10, q12 @ w = 3 * (QS0 - PS0) 329 vmul.i16 q11, q11, q12 729 vmul.u16 q10, q10, d0[2] 731 vmul.u16 q11, q11, d0[3] 756 vmul.u16 q11, q11, d0[3] 757 vmul.u16 q10, q10, d0[2] 758 vmul.u16 q3, q3, d0[2] 759 vmul.u16 q14, q14, d0[3] 784 vmul.u16 q10, q10, d0[0] 785 vmul [all...] |
H A D | rv34dsp_neon.S | 54 vmul.s32 q13, q11, d0[0] 60 vmul.s32 q14, q10, d0[0]
|
H A D | synth_filter_vfp.S | 167 vmul.f VB0, VB0, SCALE @ SCALE treated as scalar 170 vmul.f VA0, VA0, SCALE
|
H A D | hevcdsp_qpel_neon.S | 116 vmul.s32 q11, q8 // 11 * (c0 + f0) 117 vmul.s32 q12, q8 // 11 * (c1 + f1) 120 vmul.s32 q9, q8 // 40 * (d0 + e0) 121 vmul.s32 q10, q8 // 40 * (d1 + e1) 191 vmul.u16 q13, q12 // 10 * (d+e) 192 vmul.u16 q15, q14 // 11 * ( c + f)
|
/third_party/ffmpeg/libavutil/arm/ |
H A D | float_dsp_neon.S | 29 vmul.f32 q8, q0, q2 30 vmul.f32 q9, q1, q3 37 vmul.f32 q10, q0, q2 40 vmul.f32 q11, q1, q3 44 vmul.f32 q8, q0, q2 47 vmul.f32 q9, q1, q3 55 vmul.f32 q8, q0, q2 59 vmul.f32 q9, q1, q3 121 1: vmul.f32 q0, q0, q8 123 vmul [all...] |
H A D | float_dsp_vfp.S | 38 vmul.f32 s8, s0, s8 41 vmul.f32 s12, s4, s12 105 vmul.f s24, s0, s8 107 vmul.f s8, s16, s8 119 vmul.f s24, s0, s8 120 vmul.f s25, s1, s9 123 vmul.f s8, s16, s8 124 vmul.f s9, s17, s9 141 vmul.f s24, s0, s8 142 vmul [all...] |
/third_party/ffmpeg/libswscale/arm/ |
H A D | rgb2yuv_neon_16.S | 59 compute_y_16x1_step vmul, r8x16, CO_RY 74 vmul c16x8, r16x8, CO_R\C
|
/third_party/ffmpeg/libswscale/ppc/ |
H A D | swscale_altivec.c | 159 const vec_f vmul = (vec_f) {fmult, fmult, fmult, fmult}; in yuv2plane1_float_altivec() local 174 vd = vec_madd(vd, vmul, vzero); in yuv2plane1_float_altivec() 192 const vec_f vmul = (vec_f) {fmult, fmult, fmult, fmult}; in yuv2plane1_float_bswap_altivec() local 209 vd = vec_madd(vd, vmul, vzero); in yuv2plane1_float_bswap_altivec()
|
/third_party/mesa3d/src/panfrost/midgard/ |
H A D | midgard_schedule.c | 1002 * condition must be in the former pipeline stage (vmul/sadd), in mir_schedule_condition() 1174 midgard_instruction *vmul = NULL; in mir_schedule_alu() local 1217 * since we might not have room for a conditional in vmul/sadd. in mir_schedule_alu() 1286 vmul = cond; in mir_schedule_alu() 1293 /* Stage 2, let's schedule sadd before vmul for writeout */ in mir_schedule_alu() 1315 /* It's possible we'll be able to schedule something into vmul in mir_schedule_alu() 1317 * vmul specially that way. */ in mir_schedule_alu() 1330 vmul = peaked; in mir_schedule_alu() 1331 vmul->unit = UNIT_VMUL; in mir_schedule_alu() 1344 vmul in mir_schedule_alu() [all...] |
/third_party/ffmpeg/libavresample/arm/ |
H A D | resample_neon.S | 126 vmul.f32 d16, d0, d4 127 vmul.f32 d17, d1, d5 333 vmul.f32 q9, q0, q2 334 vmul.f32 q8, q1, q2 351 vmul.f32 s2, s2, s0 /* (v2 - val) * frac */
|