/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm/crypto/sha/ |
H A D | sha512-armv8.S | 1123 ld1 {v25.2d},[x3],#16 1134 add v25.2d,v25.2d,v17.2d 1136 ext v25.16b,v25.16b,v25.16b,#8 1139 add v2.2d,v2.2d,v25.2d // "T1 + H + K512[i]" 1147 ld1 {v25.2d},[x3],#16 1158 add v25.2d,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm_avx2/crypto/sha/ |
H A D | sha512-armv8.S | 1123 ld1 {v25.2d},[x3],#16 1134 add v25.2d,v25.2d,v17.2d 1136 ext v25.16b,v25.16b,v25.16b,#8 1139 add v2.2d,v2.2d,v25.2d // "T1 + H + K512[i]" 1147 ld1 {v25.2d},[x3],#16 1158 add v25.2d,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm/crypto/sha/ |
H A D | sha512-armv8.S | 1123 ld1 {v25.2d},[x3],#16 1134 add v25.2d,v25.2d,v17.2d 1136 ext v25.16b,v25.16b,v25.16b,#8 1139 add v2.2d,v2.2d,v25.2d // "T1 + H + K512[i]" 1147 ld1 {v25.2d},[x3],#16 1158 add v25.2d,v25 [all...] |
/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm_avx2/crypto/sha/ |
H A D | sha512-armv8.S | 1123 ld1 {v25.2d},[x3],#16 1134 add v25.2d,v25.2d,v17.2d 1136 ext v25.16b,v25.16b,v25.16b,#8 1139 add v2.2d,v2.2d,v25.2d // "T1 + H + K512[i]" 1147 ld1 {v25.2d},[x3],#16 1158 add v25.2d,v25 [all...] |
/third_party/ffmpeg/libavcodec/aarch64/ |
H A D | fft_neon.S | 82 fmul v25.2s, v19.2s, v28.s[1] // a3r*w,a3i*w 91 fadd v25.2s, v25.2s, v27.2s // a3r-a3i,a3i+a3r t5,t6 94 ext v26.8b, v24.8b, v25.8b, #4 95 ext v27.8b, v25.8b, v24.8b, #4 98 fadd v5.2s, v25.2s, v24.2s 133 fmul v25.2s, v19.2s, v28.s[1] // a3r*w,a3i*w 142 fadd v25.2s, v25.2s, v27.2s // a3r-a3i,a3i+a3r t5,t6 147 ext v26.8b, v24.8b, v25 [all...] |
H A D | vp9lpf_neon.S | 179 uabd v7\sz, v24\sz, v25\sz // abs(q0 - q1) 180 uabd \tmp1\sz, v25\sz, v26\sz // abs(q1 - q2) 188 uabd v5\sz, v22\sz, v25\sz // abs(p1 - q1) 212 uabd \tmp1\sz, v25\sz, v24\sz // abs(q1 - q0) 251 uabd v1\sz, v25\sz, v24\sz // abs(q1 - q0) 257 usubl_sz \tmp1\().8h, \tmp2\().8h, v22, v25, \sz // p1 - q1 315 uxtl_sz v2.8h, v3.8h, v25, \sz // q1 332 bit v25\sz, v2\sz, v5\sz 345 uaddl_sz \tmp3\().8h, \tmp4\().8h, v22, v25, \sz 364 uaddl_sz \tmp7\().8h, \tmp8\().8h, v25, v2 [all...] |
H A D | vp9lpf_16bpp_neon.S | 38 uabd v7.8h, v24.8h, v25.8h // abs(q0 - q1) 39 uabd \tmp1\().8h, v25.8h, v26.8h // abs(q1 - q2) 47 uabd v5.8h, v22.8h, v25.8h // abs(p1 - q1) 69 uabd \tmp1\().8h, v25.8h, v24.8h // abs(q1 - q0) 108 uabd v1.8h, v25.8h, v24.8h // abs(q1 - q0) 115 sub \tmp1\().8h, v22.8h, v25.8h // p1 - q1 172 sub v2.8h, v25.8h, \tmp3\().8h // q1 - f 187 bit v25.16b, v2.16b, v5.16b 202 add \tmp3\().8h, v22.8h, v25.8h 221 add \tmp7\().8h, v25 [all...] |
H A D | vp9mc_16bpp_neon.S | 189 smull2 v25.4s, v16.8h, v0.h[0] 197 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 1, \size 198 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 2, \size 199 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 3, \size 200 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 4, \size 201 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 5, \size 202 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 6, \size 203 extmlal v1, v2, v3, v4, v24, v25, v26, v27, v5, v6, v7, v16, v17, v18, 7, \size 212 sqrshrun2 v24.8h, v25.4s, #7 217 sqrshrun v25 [all...] |
H A D | h264idct_neon.S | 223 add v17.8H, v31.8H, v25.8H 224 sub va.8H, v31.8H, v25.8H 229 add v19.8H, v19.8H, v25.8H 230 sshr v25.8H, v25.8H, #1 237 add v19.8H, v19.8H, v25.8H 238 sshr v25.8H, v16.8H, #2 242 sub v19.8H, v19.8H, v25.8H 249 add v25.8H, v26.8H, v18.8H 258 add v25 [all...] |
H A D | vp9itxfm_neon.S | 333 butterfly_8h v24, v25, v16, v22 // v24 = t0, v25 = t3 343 butterfly_8h v19, v20, v25, v28 // v17 = out[3], q12 = out[4] 347 dmbutterfly_l v24, v25, v26, v27, v23, v16, v1.h[1], v1.h[0] // v24,v25 = t1a, v26,v27 = t0a 353 dbutterfly_n v2, v3, v24, v25, v2, v3, v6, v7, v26, v27 // v2 = t1, v3 = t5 354 dbutterfly_n v24, v25, v30, v31, v21, v23, v6, v7, v26, v27 // v24 = t2, v25 = t6 365 dmbutterfly_l v2, v3, v4, v5, v31, v25, v0.h[3], v0.h[2] // v2,v3 = t6a, v4,v5 = t7a 367 dbutterfly_n v17, v30, v28, v29, v2, v3, v6, v7, v24, v25 // v1 [all...] |
H A D | simple_idct_neon.S | 167 idct_col4_top v24, v25, v26, v27, \i, \l 228 idct_row4_neon v24, v25, v26, v27, 1 266 idct_row4_neon v24, v25, v26, v27, 1 287 zip1 v25.2D, v2.2D, v16.2D 302 uaddw v24.8H, v25.8H, v20.8B 303 uaddw2 v25.8H, v26.8H, v20.16B 306 sqxtun2 v24.16B, v25.8H 308 uaddw v25.8H, v27.8H, v21.8B 311 sqxtun v25.8B, v25 [all...] |
H A D | vp9itxfm_16bpp_neon.S | 278 smull2 v25.2d, v17.4s, v0.s[0] 286 rshrn2 v24.4s, v25.2d, #14 336 add v25.2d, v17.2d, v23.2d 340 rshrn2 \c0\().4s, v25.2d, #14 581 ld1 {v24.4s,v25.4s,v26.4s,v27.4s}, [x2], #64 591 idct8 v17, v19, v21, v23, v25, v27, v29, v31, v2, v3, v4, v5, v6, v7 594 \txfm1\()8 v17, v19, v21, v23, v25, v27, v29, v31, v4, v5, v6, v7, v8, v9 598 transpose_8x8s v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27, v28, v29, v30, v31, v4, v5, v6, v7 602 idct8 v17, v19, v21, v23, v25, v27, v29, v31, v2, v3, v4, v5, v6, v7 605 \txfm2\()8 v17, v19, v21, v23, v25, v2 [all...] |
/third_party/json/include/nlohmann/detail/ |
H A D | macro_scope.hpp | 343 #define NLOHMANN_JSON_PASTE26(func, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25) NLOHMANN_JSON_PASTE2(func, v1) NLOHMANN_JSON_PASTE25(func, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25) 344 #define NLOHMANN_JSON_PASTE27(func, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26) NLOHMANN_JSON_PASTE2(func, v1) NLOHMANN_JSON_PASTE26(func, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26) 345 #define NLOHMANN_JSON_PASTE28(func, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27) NLOHMANN_JSON_PASTE2(func, v1) NLOHMANN_JSON_PASTE27(func, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27) 346 #define NLOHMANN_JSON_PASTE29(func, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27, v28) NLOHMANN_JSON_PASTE2(func, v1) NLOHMANN_JSON_PASTE28(func, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27, v28) 347 #define NLOHMANN_JSON_PASTE30(func, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v26, v27, v28, v29) NLOHMANN_JSON_PASTE2(func, v1) NLOHMANN_JSON_PASTE29(func, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, v13, v14, v15, v16, v17, v18, v19, v20, v21, v22, v23, v24, v25, v2 [all...] |
/third_party/ffmpeg/libavfilter/aarch64/ |
H A D | vf_nlmeans_neon.S | 26 ext v25.16B, v26.16B, \xb, #12 // ext(0000,ABCD,12)=0ABC 28 add v24.4S, v24.4S, v25.4S // {X+A,X+B+A,X+C+B,X+D+C} (+0ABC) 29 ext v25.16B, v26.16B, v25.16B, #12 // ext(0000,0ABC,12)=00AB 30 add v24.4S, v24.4S, v25.4S // {X+A,X+B+A,X+C+B+A,X+D+C+B} (+00AB) 31 ext v25.16B, v26.16B, v25.16B, #12 // ext(0000,00AB,12)=000A 32 add v24.4S, v24.4S, v25.4S // {X+A,X+B+A,X+C+B+A,X+D+C+B+A} (+000A)
|
/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm_avx2/crypto/aes/ |
H A D | aesv8-armx.S | 413 ld1 {v25.16b},[x0],#16 425 aese v25.16b,v16.16b 426 aesmc v25.16b,v25.16b 437 aese v25.16b,v17.16b 438 aesmc v25.16b,v25.16b 450 aese v25.16b,v16.16b 451 aesmc v25.16b,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm/crypto/aes/ |
H A D | aesv8-armx.S | 413 ld1 {v25.16b},[x0],#16 425 aese v25.16b,v16.16b 426 aesmc v25.16b,v25.16b 437 aese v25.16b,v17.16b 438 aesmc v25.16b,v25.16b 450 aese v25.16b,v16.16b 451 aesmc v25.16b,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm_avx2/crypto/aes/ |
H A D | aesv8-armx.S | 413 ld1 {v25.16b},[x0],#16 425 aese v25.16b,v16.16b 426 aesmc v25.16b,v25.16b 437 aese v25.16b,v17.16b 438 aesmc v25.16b,v25.16b 450 aese v25.16b,v16.16b 451 aesmc v25.16b,v25 [all...] |
/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm/crypto/aes/ |
H A D | aesv8-armx.S | 413 ld1 {v25.16b},[x0],#16 425 aese v25.16b,v16.16b 426 aesmc v25.16b,v25.16b 437 aese v25.16b,v17.16b 438 aesmc v25.16b,v25.16b 450 aese v25.16b,v16.16b 451 aesmc v25.16b,v25 [all...] |
/third_party/node/deps/openssl/openssl/crypto/aes/asm/ |
H A D | aesp8-ppc.pl | 680 my $rndkey0="v23"; # v24-v25 rotating buffer for first found keys 700 stvx v25,r11,$sp 745 ?vperm v25,v31,v30,$keyperm 747 stvx v25,$x10,$key_ # off-load round[2] 755 ?vperm v25,v31,v26,$keyperm 757 stvx v25,$x10,$key_ # off-load round[4] 770 lvx v25,$x10,$key_ # pre-load round[2] 821 vncipher $out0,$out0,v25 822 vncipher $out1,$out1,v25 823 vncipher $out2,$out2,v25 [all...] |
/third_party/openssl/crypto/aes/asm/ |
H A D | aesp8-ppc.pl | 680 my $rndkey0="v23"; # v24-v25 rotating buffer for first found keys 700 stvx v25,r11,$sp 745 ?vperm v25,v31,v30,$keyperm 747 stvx v25,$x10,$key_ # off-load round[2] 755 ?vperm v25,v31,v26,$keyperm 757 stvx v25,$x10,$key_ # off-load round[4] 770 lvx v25,$x10,$key_ # pre-load round[2] 821 vncipher $out0,$out0,v25 822 vncipher $out1,$out1,v25 823 vncipher $out2,$out2,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm/crypto/chacha/ |
H A D | chacha-armv8.S | 362 dup v25.4s,v1.s[2] 392 add v24.4s,v24.4s,v25.4s 424 eor v6.16b,v25.16b,v26.16b 432 ushr v25.4s,v6.4s,#20 440 sli v25.4s,v6.4s,#12 448 add v24.4s,v24.4s,v25.4s 480 eor v6.16b,v25.16b,v26.16b 486 ushr v25.4s,v6.4s,#25 490 sli v25.4s,v6.4s,#7 494 add v20.4s,v20.4s,v25 [all...] |
/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm/crypto/chacha/ |
H A D | chacha-armv8.S | 362 dup v25.4s,v1.s[2] 392 add v24.4s,v24.4s,v25.4s 424 eor v6.16b,v25.16b,v26.16b 432 ushr v25.4s,v6.4s,#20 440 sli v25.4s,v6.4s,#12 448 add v24.4s,v24.4s,v25.4s 480 eor v6.16b,v25.16b,v26.16b 486 ushr v25.4s,v6.4s,#25 490 sli v25.4s,v6.4s,#7 494 add v20.4s,v20.4s,v25 [all...] |
/third_party/node/deps/openssl/config/archs/darwin64-arm64-cc/asm_avx2/crypto/chacha/ |
H A D | chacha-armv8.S | 362 dup v25.4s,v1.s[2] 392 add v24.4s,v24.4s,v25.4s 424 eor v6.16b,v25.16b,v26.16b 432 ushr v25.4s,v6.4s,#20 440 sli v25.4s,v6.4s,#12 448 add v24.4s,v24.4s,v25.4s 480 eor v6.16b,v25.16b,v26.16b 486 ushr v25.4s,v6.4s,#25 490 sli v25.4s,v6.4s,#7 494 add v20.4s,v20.4s,v25 [all...] |
/third_party/node/deps/openssl/config/archs/linux-aarch64/asm_avx2/crypto/chacha/ |
H A D | chacha-armv8.S | 362 dup v25.4s,v1.s[2] 392 add v24.4s,v24.4s,v25.4s 424 eor v6.16b,v25.16b,v26.16b 432 ushr v25.4s,v6.4s,#20 440 sli v25.4s,v6.4s,#12 448 add v24.4s,v24.4s,v25.4s 480 eor v6.16b,v25.16b,v26.16b 486 ushr v25.4s,v6.4s,#25 490 sli v25.4s,v6.4s,#7 494 add v20.4s,v20.4s,v25 [all...] |
/third_party/vixl/test/aarch64/ |
H A D | test-trace-aarch64.cc | 628 __ abs(v27.V2S(), v25.V2S()); in GenerateTestSequenceNEON() 638 __ add(v25.V4S(), v28.V4S(), v29.V4S()); in GenerateTestSequenceNEON() 693 __ cmeq(v16.V2S(), v25.V2S(), 0); in GenerateTestSequenceNEON() 697 __ cmeq(v6.V4S(), v25.V4S(), 0); in GenerateTestSequenceNEON() 700 __ cmeq(v20.V8H(), v7.V8H(), v25.V8H()); in GenerateTestSequenceNEON() 708 __ cmge(v25.V2S(), v22.V2S(), v3.V2S()); in GenerateTestSequenceNEON() 720 __ cmgt(v20.V16B(), v25.V16B(), v27.V16B()); in GenerateTestSequenceNEON() 721 __ cmgt(v0.V16B(), v25.V16B(), 0); in GenerateTestSequenceNEON() 722 __ cmgt(v22.V2D(), v25.V2D(), v1.V2D()); in GenerateTestSequenceNEON() 741 __ cmhi(v11.V8H(), v10.V8H(), v25 in GenerateTestSequenceNEON() [all...] |