/third_party/ffmpeg/libavcodec/aarch64/ |
H A D | vc1dsp_neon.S | 86 srshr v3.8h, v22.8h, #2 // (t5 + t1 + 4) >> 3 87 srshr v4.8h, v5.8h, #2 // (t6 + t2 + 4) >> 3 88 srshr v5.8h, v16.8h, #2 // (t7 + t3 + 4) >> 3 89 srshr v6.8h, v17.8h, #2 // (t8 + t4 + 4) >> 3 90 srshr v2.8h, v2.8h, #2 // (t8 - t4 + 4) >> 3 91 srshr v1.8h, v1.8h, #2 // (t7 - t3 + 4) >> 3 92 srshr v7.8h, v21.8h, #2 // (t6 - t2 + 4) >> 3 93 srshr v16.8h, v23.8h, #2 // (t5 - t1 + 4) >> 3 167 srshr v2.8h, v18.8h, #6 // (t5 + t1 + 64) >> 7 168 srshr v [all...] |
H A D | h264idct_neon.S | 63 srshr v0.8H, v0.8H, #6 64 srshr v1.8H, v1.8H, #6 88 srshr v2.8H, v2.8H, #6 289 srshr v24.8H, v24.8H, #6 291 srshr v25.8H, v25.8H, #6 293 srshr v26.8H, v26.8H, #6 295 srshr v27.8H, v27.8H, #6 297 srshr v28.8H, v28.8H, #6 299 srshr v29.8H, v29.8H, #6 301 srshr v3 [all...] |
H A D | vp9itxfm_16bpp_neon.S | 256 srshr v22.4s, v22.4s, #14 257 srshr v18.4s, v18.4s, #14 258 srshr v24.4s, v24.4s, #14 259 srshr v20.4s, v20.4s, #14 308 srshr \c0\().4s, v24.4s, #14 310 srshr \c1\().4s, v26.4s, #14 312 srshr \c2\().4s, v20.4s, #14 313 srshr \c3\().4s, v16.4s, #14 418 srshr v4.4s, v4.4s, #4 419 srshr v [all...] |
H A D | vp9itxfm_neon.S | 293 srshr v4.4h, v4.4h, #4 294 srshr v5.4h, v5.4h, #4 295 srshr v6.4h, v6.4h, #4 296 srshr v7.4h, v7.4h, #4 430 srshr v16.8h, v16.8h, #5 432 srshr v17.8h, v17.8h, #5 434 srshr v18.8h, v18.8h, #5 437 srshr v19.8h, v19.8h, #5 440 srshr v20.8h, v20.8h, #5 444 srshr v2 [all...] |
H A D | vp8dsp_neon.S | 137 srshr v0.4h, v0.4h, #3 138 srshr v1.4h, v1.4h, #3 139 srshr v2.4h, v2.4h, #3 140 srshr v3.4h, v3.4h, #3 177 srshr v16.8h, v16.8h, #3 // dc >>= 3 179 srshr v18.8h, v18.8h, #3 228 srshr v16.8h, v16.8h, #3 // dc >>= 3 230 srshr v18.8h, v18.8h, #3 263 srshr v2.8h, v2.8h, #3 401 srshr v1 [all...] |
H A D | hevcdsp_idct_neon.S | 577 srshr v4.8h, v4.8h, #1 578 srshr v0.8h, v4.8h, #(14 - \bitdepth) 579 srshr v1.8h, v4.8h, #(14 - \bitdepth) 581 srshr v2.8h, v4.8h, #(14 - \bitdepth) 582 srshr v3.8h, v4.8h, #(14 - \bitdepth)
|
H A D | vp9lpf_16bpp_neon.S | 165 srshr \tmp3\().8h, \tmp3\().8h, #1 // f = (f1 + 1) >> 1
|
H A D | h264dsp_neon.S | 880 srshr v4.8h, v4.8h, #3
|
H A D | vp9lpf_neon.S | 310 srshr \tmp3\sz, \tmp3\sz, #1 // f = (f1 + 1) >> 1
|
/third_party/vixl/test/aarch64/ |
H A D | test-trace-aarch64.cc | 1747 __ srshr(d21, d18, 45); in GenerateTestSequenceNEON() 1748 __ srshr(v3.V16B(), v11.V16B(), 7); in GenerateTestSequenceNEON() 1749 __ srshr(v21.V2D(), v26.V2D(), 53); in GenerateTestSequenceNEON() 1750 __ srshr(v11.V2S(), v5.V2S(), 28); in GenerateTestSequenceNEON() 1751 __ srshr(v7.V4H(), v18.V4H(), 12); in GenerateTestSequenceNEON() 1752 __ srshr(v7.V4S(), v3.V4S(), 30); in GenerateTestSequenceNEON() 1753 __ srshr(v14.V8B(), v2.V8B(), 6); in GenerateTestSequenceNEON() 1754 __ srshr(v21.V8H(), v20.V8H(), 3); in GenerateTestSequenceNEON()
|
H A D | test-cpu-features-aarch64.cc | 1957 TEST_NEON(srshr_0, srshr(v0.V8B(), v1.V8B(), 2)) 1958 TEST_NEON(srshr_1, srshr(v0.V16B(), v1.V16B(), 3)) 1959 TEST_NEON(srshr_2, srshr(v0.V4H(), v1.V4H(), 2)) 1960 TEST_NEON(srshr_3, srshr(v0.V8H(), v1.V8H(), 7)) 1961 TEST_NEON(srshr_4, srshr(v0.V2S(), v1.V2S(), 25)) 1962 TEST_NEON(srshr_5, srshr(v0.V4S(), v1.V4S(), 27)) 1963 TEST_NEON(srshr_6, srshr(v0.V2D(), v1.V2D(), 43)) 1964 TEST_NEON(srshr_7, srshr(d0, d1, 28))
|
H A D | test-disasm-sve-aarch64.cc | 6543 COMPARE(srshr(z12.VnB(), p0.Merging(), z12.VnB(), 1), in TEST() 6544 "srshr z12.b, p0/m, z12.b, #1"); in TEST() 6545 COMPARE(srshr(z12.VnB(), p0.Merging(), z12.VnB(), 2), in TEST() 6546 "srshr z12.b, p0/m, z12.b, #2"); in TEST() 6547 COMPARE(srshr(z12.VnB(), p0.Merging(), z12.VnB(), 5), in TEST() 6548 "srshr z12.b, p0/m, z12.b, #5"); in TEST() 6549 COMPARE(srshr(z12.VnB(), p0.Merging(), z12.VnB(), 8), in TEST() 6550 "srshr z12.b, p0/m, z12.b, #8"); in TEST() 6551 COMPARE(srshr(z12.VnH(), p0.Merging(), z12.VnH(), 1), in TEST() 6552 "srshr z1 in TEST() [all...] |
H A D | test-simulator-aarch64.cc | 4775 DEFINE_TEST_NEON_2OPIMM(srshr, Basic, TypeWidth) 4810 DEFINE_TEST_NEON_2OPIMM_SCALAR_D(srshr, Basic, TypeWidth)
|
H A D | test-api-movprfx-aarch64.cc | 2274 __ srshr(z12.VnB(), p0.Merging(), z12.VnB(), 1); in TEST() 3652 __ srshr(z12.VnB(), p0.Merging(), z12.VnB(), 1); in TEST()
|
/third_party/node/deps/v8/src/codegen/arm64/ |
H A D | assembler-arm64.h | 1354 void srshr(const VRegister& vd, const VRegister& vn, int shift);
|
H A D | macro-assembler-arm64.h | 1234 V(srshr, Srshr) \
|
H A D | assembler-arm64.cc | 1709 void Assembler::srshr(const VRegister& vd, const VRegister& vn, int shift) { in srshr() function in v8::internal::Assembler
|
/third_party/vixl/src/aarch64/ |
H A D | assembler-aarch64.h | 3225 void srshr(const VRegister& vd, const VRegister& vn, int shift); 6568 void srshr(const ZRegister& zd,
|
H A D | macro-assembler-aarch64.h | 3177 V(srshr, Srshr) \ 7254 srshr(zd, pg, zd, shift); in Srshr()
|
H A D | assembler-aarch64.cc | 5647 void Assembler::srshr(const VRegister& vd, const VRegister& vn, int shift) {
|
H A D | assembler-sve-aarch64.cc | 8698 void Assembler::srshr(const ZRegister& zd, in srshr() function in vixl::aarch64::Assembler
|