/third_party/ffmpeg/libavcodec/arm/ |
H A D | vp3dsp_neon.S | 127 vmull.s16 q2, d18, xC1S7 // (ip[1] * C1) << 16 128 vmull.s16 q3, d19, xC1S7 129 vmull.s16 q4, d2, xC4S4 // ((ip[0] + ip[4]) * C4) << 16 130 vmull.s16 q5, d3, xC4S4 131 vmull.s16 q6, d16, xC4S4 // ((ip[0] - ip[4]) * C4) << 16 132 vmull.s16 q7, d17, xC4S4 143 vmull.s16 q2, d30, xC1S7 // (ip[7] * C1) << 16 144 vmull.s16 q3, d31, xC1S7 145 vmull.s16 q4, d30, xC7S1 // (ip[7] * C7) << 16 146 vmull [all...] |
H A D | h264cmc_neon.S | 67 vmull.u8 q8, d4, d0 75 vmull.u8 q9, d6, d0 112 vmull.u8 q8, d4, d0 115 vmull.u8 q9, d6, d0 146 vmull.u8 q8, d4, d0 148 vmull.u8 q9, d6, d0 175 vmull.u8 q8, d4, d0 176 vmull.u8 q9, d5, d0 249 vmull.u8 q8, d4, d0 255 vmull [all...] |
H A D | simple_idct_neon.S | 49 vmull.s16 q7, d6, w2 /* q9 = W2 * col[2] */ 50 vmull.s16 q8, d6, w6 /* q10 = W6 * col[2] */ 51 vmull.s16 q9, d4, w1 /* q9 = W1 * col[1] */ 53 vmull.s16 q10, d4, w3 /* q10 = W3 * col[1] */ 55 vmull.s16 q5, d4, w5 /* q5 = W5 * col[1] */ 57 vmull.s16 q6, d4, w7 /* q6 = W7 * col[1] */ 100 vmull.s16 q7, d3, w4 /* q7 = W4 * col[4] */ 103 vmull.s16 q8, d7, w2 /* q8 = W2 * col[6] */ 110 vmull.s16 q7, d7, w6 /* q7 = W6 * col[6] */ 159 vmull [all...] |
H A D | hevcdsp_qpel_neon.S | 80 vmull.s16 q9, d6, d16 // 58 * d0 81 vmull.s16 q10, d7, d16 // 58 * d1 83 vmull.s16 q11, d4, d17 // 10 * c0 84 vmull.s16 q12, d5, d17 // 10 * c1 86 vmull.s16 q13, d8, d16 // 17 * e0 87 vmull.s16 q14, d9, d16 // 17 * e1 88 vmull.s16 q15, d10, d17 // 5 * f0 89 vmull.s16 q8, d11, d17 // 5 * f1 140 vmull.s16 q9, d8, d16 // 58 * d0 141 vmull [all...] |
H A D | vp8dsp_neon.S | 86 vmull.s16 q12, d1, d4[0] 87 vmull.s16 q13, d3, d4[0] 108 vmull.s16 q12, d1, d4[0] 110 vmull.s16 q13, d2, d4[0] 1573 vmull.u8 q8, d2, d1 1576 vmull.u8 q3, d3, d1 1579 vmull.u8 q11, d18, d1 1581 vmull.u8 q12, d19, d1 1604 vmull.u8 q3, d2, d1 1606 vmull [all...] |
H A D | rv34dsp_neon.S | 31 vmull.s16 q10, d28, d0 33 vmull.s16 q11, d28, d0 130 vmull.s16 q1, d0, d1 @ dc * 13 * 13 150 vmull.s16 q3, d28, d4
|
H A D | sbcdsp_neon.S | 44 vmull.s16 q0, d16, d20 46 vmull.s16 q1, d17, d21 77 vmull.s16 q10, d16, d0 78 vmull.s16 q11, d17, d0 97 vmull.s16 q12, d16, d20 99 vmull.s16 q13, d17, d21 101 vmull.s16 q14, d18, d22 103 vmull.s16 q15, d19, d23 157 vmull.s16 q12, d16, d0 159 vmull [all...] |
H A D | h264dsp_neon.S | 449 vmull.u8 q2, d0, d20 451 vmull.u8 q3, d0, d21 453 vmull.u8 q12, d0, d28 455 vmull.u8 q13, d0, d29 478 vmull.u8 q1, d0, d4 481 vmull.u8 q10, d0, d6 502 vmull.u8 q1, d0, d4 507 vmull.u8 q10, d0, d6
|
H A D | vp9mc_16bpp_neon.S | 148 @ Helper macros for vmull/vmlal with a constant from either d0 or d1 depending on index 151 vmull.s16 \dst, \src, d0[\idx] 153 vmull.s16 \dst, \src, d1[\idx - 4] 218 vmull.s16 q1, d16, d0[0] 219 vmull.s16 q12, d20, d0[0] 221 vmull.s16 q2, d17, d0[0] 222 vmull.s16 q13, d21, d0[0] 391 vmull.s16 \dst1, \src1, d0[0] 392 vmull.s16 \dst2, \src2, d0[0] 393 vmull [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm/crypto/modes/ |
H A D | ghash-armv4.S | 416 vmull.p8 q8, d16, d6 @ F = A1*B 418 vmull.p8 q0, d26, d0 @ E = A*B1 420 vmull.p8 q9, d18, d6 @ H = A2*B 422 vmull.p8 q11, d26, d22 @ G = A*B2 425 vmull.p8 q10, d20, d6 @ J = A3*B 428 vmull.p8 q0, d26, d0 @ I = A*B3 434 vmull.p8 q11, d26, d22 @ K = A*B4 445 vmull.p8 q0, d26, d6 @ D = A*B 454 vmull.p8 q8, d16, d6 @ F = A1*B 456 vmull [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm_avx2/crypto/modes/ |
H A D | ghash-armv4.S | 416 vmull.p8 q8, d16, d6 @ F = A1*B 418 vmull.p8 q0, d26, d0 @ E = A*B1 420 vmull.p8 q9, d18, d6 @ H = A2*B 422 vmull.p8 q11, d26, d22 @ G = A*B2 425 vmull.p8 q10, d20, d6 @ J = A3*B 428 vmull.p8 q0, d26, d0 @ I = A*B3 434 vmull.p8 q11, d26, d22 @ K = A*B4 445 vmull.p8 q0, d26, d6 @ D = A*B 454 vmull.p8 q8, d16, d6 @ F = A1*B 456 vmull [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm_avx2/crypto/bn/ |
H A D | armv4-gf2m.S | 179 vmull.p8 q1, d2, d27 @ F = A1*B 181 vmull.p8 q0, d26, d0 @ E = A*B1 183 vmull.p8 q2, d4, d27 @ H = A2*B 185 vmull.p8 q8, d26, d16 @ G = A*B2 188 vmull.p8 q3, d6, d27 @ J = A3*B 191 vmull.p8 q0, d26, d0 @ I = A*B3 197 vmull.p8 q8, d26, d16 @ K = A*B4 208 vmull.p8 q0, d26, d27 @ D = A*B
|
/third_party/node/deps/openssl/config/archs/linux-armv4/asm/crypto/bn/ |
H A D | armv4-gf2m.S | 179 vmull.p8 q1, d2, d27 @ F = A1*B 181 vmull.p8 q0, d26, d0 @ E = A*B1 183 vmull.p8 q2, d4, d27 @ H = A2*B 185 vmull.p8 q8, d26, d16 @ G = A*B2 188 vmull.p8 q3, d6, d27 @ J = A3*B 191 vmull.p8 q0, d26, d0 @ I = A*B3 197 vmull.p8 q8, d26, d16 @ K = A*B4 208 vmull.p8 q0, d26, d27 @ D = A*B
|
/third_party/node/deps/openssl/openssl/crypto/bn/asm/ |
H A D | armv4-gf2m.pl | 271 vmull.p8 $t0, $t0#lo, $b @ F = A1*B 273 vmull.p8 $r, $a, $r#lo @ E = A*B1 275 vmull.p8 $t1, $t1#lo, $b @ H = A2*B 277 vmull.p8 $t3, $a, $t3#lo @ G = A*B2 280 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 283 vmull.p8 $r, $a, $r#lo @ I = A*B3 289 vmull.p8 $t3, $a, $t3#lo @ K = A*B4 300 vmull.p8 $r, $a, $b @ D = A*B
|
H A D | armv4-mont.pl | 342 vmull.u32 @ACC[0],$Bi,${A0}[0] 343 vmull.u32 @ACC[1],$Bi,${A0}[1] 344 vmull.u32 @ACC[2],$Bi,${A1}[0] 346 vmull.u32 @ACC[3],$Bi,${A1}[1] 352 vmull.u32 @ACC[4],$Bi,${A2}[0] 354 vmull.u32 @ACC[5],$Bi,${A2}[1] 355 vmull.u32 @ACC[6],$Bi,${A3}[0] 357 vmull.u32 @ACC[7],$Bi,${A3}[1]
|
/third_party/openssl/crypto/bn/asm/ |
H A D | armv4-gf2m.pl | 271 vmull.p8 $t0, $t0#lo, $b @ F = A1*B 273 vmull.p8 $r, $a, $r#lo @ E = A*B1 275 vmull.p8 $t1, $t1#lo, $b @ H = A2*B 277 vmull.p8 $t3, $a, $t3#lo @ G = A*B2 280 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 283 vmull.p8 $r, $a, $r#lo @ I = A*B3 289 vmull.p8 $t3, $a, $t3#lo @ K = A*B4 300 vmull.p8 $r, $a, $b @ D = A*B
|
/third_party/ffmpeg/libswscale/arm/ |
H A D | output.S | 54 vmull.s16 q9, d14, d30 @ A*X,I*Y,B*X,J*Y 55 vmull.s16 q10, d15, d31 @ C*X,K*Y,D*X,L*Y 56 vmull.s16 q11, d16, d30 @ E*X,M*Y,F*X,N*Y 57 vmull.s16 q12, d17, d31 @ G*X,O*Y,H*X,P*Y
|
H A D | hscale.S | 45 vmull.s16 q8, d0, d4 @ srcp[filterPos[0] + {0..7}] * filter[{0..7}] (part 1) 46 vmull.s16 q9, d1, d5 @ srcp[filterPos[0] + {0..7}] * filter[{0..7}] (part 2) 47 vmull.s16 q10, d2, d6 @ srcp[filterPos[1] + {0..7}] * filter[{0..7}] (part 1) 48 vmull.s16 q11, d3, d7 @ srcp[filterPos[1] + {0..7}] * filter[{0..7}] (part 2)
|
H A D | rgb2yuv_neon_32.S | 80 compute_y_16x1_step vmull, r8x16, CO_RY 110 compute_chroma_8x1_step vmull, r16x8, CO_R\C
|
/third_party/node/deps/openssl/openssl/crypto/modes/asm/ |
H A D | ghash-armv4.pl | 389 vmull.p8 $t0, $t0#lo, $b @ F = A1*B 391 vmull.p8 $r, $a, $r#lo @ E = A*B1 393 vmull.p8 $t1, $t1#lo, $b @ H = A2*B 395 vmull.p8 $t3, $a, $t3#lo @ G = A*B2 398 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 401 vmull.p8 $r, $a, $r#lo @ I = A*B3 407 vmull.p8 $t3, $a, $t3#lo @ K = A*B4 418 vmull.p8 $r, $a, $b @ D = A*B
|
/third_party/openssl/crypto/modes/asm/ |
H A D | ghash-armv4.pl | 389 vmull.p8 $t0, $t0#lo, $b @ F = A1*B 391 vmull.p8 $r, $a, $r#lo @ E = A*B1 393 vmull.p8 $t1, $t1#lo, $b @ H = A2*B 395 vmull.p8 $t3, $a, $t3#lo @ G = A*B2 398 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 401 vmull.p8 $r, $a, $r#lo @ I = A*B3 407 vmull.p8 $t3, $a, $t3#lo @ K = A*B4 418 vmull.p8 $r, $a, $b @ D = A*B
|
/third_party/node/deps/openssl/config/archs/linux-armv4/asm/crypto/poly1305/ |
H A D | poly1305-armv4.S | 424 vmull.u32 q5,d0,d0[1] 425 vmull.u32 q6,d1,d0[1] 426 vmull.u32 q7,d3,d0[1] 427 vmull.u32 q8,d5,d0[1] 428 vmull.u32 q9,d7,d0[1] 781 vmull.u32 q7,d25,d0[1] 783 vmull.u32 q5,d21,d0[1] 785 vmull.u32 q8,d27,d0[1] 788 vmull.u32 q6,d23,d0[1] 791 vmull [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm_avx2/crypto/poly1305/ |
H A D | poly1305-armv4.S | 424 vmull.u32 q5,d0,d0[1] 425 vmull.u32 q6,d1,d0[1] 426 vmull.u32 q7,d3,d0[1] 427 vmull.u32 q8,d5,d0[1] 428 vmull.u32 q9,d7,d0[1] 781 vmull.u32 q7,d25,d0[1] 783 vmull.u32 q5,d21,d0[1] 785 vmull.u32 q8,d27,d0[1] 788 vmull.u32 q6,d23,d0[1] 791 vmull [all...] |
/third_party/node/deps/openssl/openssl/crypto/poly1305/asm/ |
H A D | poly1305-armv4.pl | 497 vmull.u32 $D0,$R0,${R0}[1] 498 vmull.u32 $D1,$R1,${R0}[1] 499 vmull.u32 $D2,$R2,${R0}[1] 500 vmull.u32 $D3,$R3,${R0}[1] 501 vmull.u32 $D4,$R4,${R0}[1] 854 vmull.u32 $D2,$H2#hi,${R0}[1] 856 vmull.u32 $D0,$H0#hi,${R0}[1] 858 vmull.u32 $D3,$H3#hi,${R0}[1] 861 vmull.u32 $D1,$H1#hi,${R0}[1] 864 vmull [all...] |
/third_party/openssl/crypto/poly1305/asm/ |
H A D | poly1305-armv4.pl | 497 vmull.u32 $D0,$R0,${R0}[1] 498 vmull.u32 $D1,$R1,${R0}[1] 499 vmull.u32 $D2,$R2,${R0}[1] 500 vmull.u32 $D3,$R3,${R0}[1] 501 vmull.u32 $D4,$R4,${R0}[1] 854 vmull.u32 $D2,$H2#hi,${R0}[1] 856 vmull.u32 $D0,$H0#hi,${R0}[1] 858 vmull.u32 $D3,$H3#hi,${R0}[1] 861 vmull.u32 $D1,$H1#hi,${R0}[1] 864 vmull [all...] |