/third_party/ffmpeg/libavcodec/aarch64/ |
H A D | vp8dsp_neon.S | 290 uqadd v17.16b, v17.16b, v17.16b // abs(P0-Q0) * 2 292 uqadd v19.16b, v17.16b, v18.16b // (abs(P0-Q0)*2) + (abs(P1-Q1)/2) 315 uqadd v17.16b, v17.16b, v17.16b // abs(P0-Q0) * 2 319 uqadd v19.16b, v17.16b, v18.16b // (abs(P0-Q0)*2) + (abs(P1-Q1)/2)
|
H A D | h264dsp_neon.S | 69 uqadd v23.16B, v18.16B, v24.16B 75 uqadd v4.16B, v2.16B, v24.16B
|
H A D | vp9lpf_neon.S | 187 uqadd v6\sz, v6\sz, v6\sz // abs(p0 - q0) * 2 192 uqadd v6\sz, v6\sz, v5\sz // abs(p0 - q0) * 2 + abs(p1 - q1) >> 1
|
/third_party/vixl/test/aarch64/ |
H A D | test-trace-aarch64.cc | 2312 __ uqadd(b30, b4, b28); in GenerateTestSequenceNEON() 2313 __ uqadd(d27, d20, d16); in GenerateTestSequenceNEON() 2314 __ uqadd(h7, h14, h28); in GenerateTestSequenceNEON() 2315 __ uqadd(s28, s17, s4); in GenerateTestSequenceNEON() 2316 __ uqadd(v19.V16B(), v22.V16B(), v21.V16B()); in GenerateTestSequenceNEON() 2317 __ uqadd(v16.V2D(), v4.V2D(), v11.V2D()); in GenerateTestSequenceNEON() 2318 __ uqadd(v20.V2S(), v14.V2S(), v4.V2S()); in GenerateTestSequenceNEON() 2319 __ uqadd(v5.V4H(), v0.V4H(), v16.V4H()); in GenerateTestSequenceNEON() 2320 __ uqadd(v21.V4S(), v31.V4S(), v9.V4S()); in GenerateTestSequenceNEON() 2321 __ uqadd(v2 in GenerateTestSequenceNEON() [all...] |
H A D | test-cpu-features-aarch64.cc | 2575 TEST_NEON(uqadd_0, uqadd(v0.V8B(), v1.V8B(), v2.V8B())) 2576 TEST_NEON(uqadd_1, uqadd(v0.V16B(), v1.V16B(), v2.V16B())) 2577 TEST_NEON(uqadd_2, uqadd(v0.V4H(), v1.V4H(), v2.V4H())) 2578 TEST_NEON(uqadd_3, uqadd(v0.V8H(), v1.V8H(), v2.V8H())) 2579 TEST_NEON(uqadd_4, uqadd(v0.V2S(), v1.V2S(), v2.V2S())) 2580 TEST_NEON(uqadd_5, uqadd(v0.V4S(), v1.V4S(), v2.V4S())) 2581 TEST_NEON(uqadd_6, uqadd(v0.V2D(), v1.V2D(), v2.V2D())) 2582 TEST_NEON(uqadd_7, uqadd(b0, b1, b2)) 2583 TEST_NEON(uqadd_8, uqadd(h0, h1, h2)) 2584 TEST_NEON(uqadd_9, uqadd(s [all...] |
H A D | test-disasm-sve-aarch64.cc | 2208 COMPARE(uqadd(z13.VnB(), z15.VnB(), z3.VnB()), "uqadd z13.b, z15.b, z3.b"); in TEST() 2209 COMPARE(uqadd(z12.VnH(), z16.VnH(), z2.VnH()), "uqadd z12.h, z16.h, z2.h"); in TEST() 2210 COMPARE(uqadd(z11.VnS(), z17.VnS(), z1.VnS()), "uqadd z11.s, z17.s, z1.s"); in TEST() 2211 COMPARE(uqadd(z10.VnD(), z18.VnD(), z0.VnD()), "uqadd z10.d, z18.d, z0.d"); in TEST() 3128 COMPARE(uqadd(z21.VnB(), z21.VnB(), 246), "uqadd z2 in TEST() [all...] |
H A D | test-api-movprfx-aarch64.cc | 1227 __ uqadd(z19.VnB(), z19.VnB(), 42); in TEST() 1654 __ uqadd(z9.VnD(), z9.VnD(), 42); in TEST() 2349 __ uqadd(z24.VnB(), p7.Merging(), z24.VnB(), z1.VnB()), in TEST() 3530 __ uqadd(z24.VnB(), p7.Merging(), z24.VnB(), z24.VnB()), in TEST() 3670 __ uqadd(z24.VnB(), p7.Merging(), z24.VnB(), z1.VnB()), in TEST()
|
H A D | test-simulator-aarch64.cc | 4638 DEFINE_TEST_NEON_3SAME(uqadd, Basic) 4697 DEFINE_TEST_NEON_3SAME_SCALAR_D(uqadd, Basic)
|
/third_party/vixl/src/aarch64/ |
H A D | assembler-aarch64.h | 2524 void uqadd(const VRegister& vd, const VRegister& vn, const VRegister& vm); 5751 void uqadd(const ZRegister& zd, const ZRegister& zn, const ZRegister& zm); 5754 void uqadd(const ZRegister& zd, 6742 void uqadd(const ZRegister& zd,
|
H A D | macro-assembler-aarch64.h | 2970 V(uqadd, Uqadd) \ 3238 V(uqadd, Uqadd) \ 6299 uqadd(zd, zn, zm); in Uqadd() 6306 uqadd(zd, zd, imm.AsUint16()); in Uqadd()
|
H A D | assembler-sve-aarch64.cc | 2306 void Assembler::uqadd(const ZRegister& zd, in uqadd() function in vixl::aarch64::Assembler 3785 void Assembler::uqadd(const ZRegister& zd, in uqadd() function in vixl::aarch64::Assembler 9323 void Assembler::uqadd(const ZRegister& zd, in uqadd() function in vixl::aarch64::Assembler
|
H A D | assembler-aarch64.cc | 4207 V(uqadd, NEON_UQADD, true) \
|
/third_party/node/deps/v8/src/codegen/arm64/ |
H A D | assembler-arm64.h | 1982 void uqadd(const VRegister& vd, const VRegister& vn, const VRegister& vm);
|
H A D | macro-assembler-arm64.h | 468 V(uqadd, Uqadd) \
|
H A D | assembler-arm64.cc | 3103 V(uqadd, NEON_UQADD, true) \
|