/third_party/optimized-routines/string/aarch64/ |
H A D | memchr-mte.S | 84 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */ 93 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */
|
H A D | memrchr.S | 85 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */ 94 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */
|
H A D | strnlen.S | 80 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */ 88 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b /* 128->64 */
|
H A D | strrchr-mte.S | 76 umaxp vend.16b, vhas_nul.16b, vhas_nul.16b 107 umaxp vend.16b, vhas_nul.16b, vhas_nul.16b
|
H A D | strlen-mte.S | 64 umaxp vend.16b, vhas_nul.16b, vhas_nul.16b
|
H A D | strchr-mte.S | 80 umaxp vend.16b, vhas_nul.16b, vhas_nul.16b
|
H A D | strchrnul.S | 91 umaxp vend1.16b, vend1.16b, vend1.16b
|
H A D | strchrnul-mte.S | 69 umaxp vend.16b, vhas_chr.16b, vhas_chr.16b
|
H A D | strchr.S | 99 umaxp vend1.16b, vend1.16b, vend1.16b
|
H A D | strcpy-mte.S | 143 umaxp vend.16b, vhas_nul.16b, vhas_nul.16b
|
H A D | strlen.S | 167 umaxp maskv.16b, maskv.16b, maskv.16b
|
/third_party/vixl/test/aarch64/ |
H A D | test-trace-aarch64.cc | 2253 __ umaxp(v1.V16B(), v6.V16B(), v29.V16B()); in GenerateTestSequenceNEON() 2254 __ umaxp(v19.V2S(), v17.V2S(), v27.V2S()); in GenerateTestSequenceNEON() 2255 __ umaxp(v21.V4H(), v16.V4H(), v7.V4H()); in GenerateTestSequenceNEON() 2256 __ umaxp(v9.V4S(), v20.V4S(), v29.V4S()); in GenerateTestSequenceNEON() 2257 __ umaxp(v13.V8B(), v1.V8B(), v16.V8B()); in GenerateTestSequenceNEON() 2258 __ umaxp(v19.V8H(), v23.V8H(), v26.V8H()); in GenerateTestSequenceNEON()
|
H A D | test-cpu-features-aarch64.cc | 2507 TEST_NEON(umaxp_0, umaxp(v0.V8B(), v1.V8B(), v2.V8B())) 2508 TEST_NEON(umaxp_1, umaxp(v0.V16B(), v1.V16B(), v2.V16B())) 2509 TEST_NEON(umaxp_2, umaxp(v0.V4H(), v1.V4H(), v2.V4H())) 2510 TEST_NEON(umaxp_3, umaxp(v0.V8H(), v1.V8H(), v2.V8H())) 2511 TEST_NEON(umaxp_4, umaxp(v0.V2S(), v1.V2S(), v2.V2S())) 2512 TEST_NEON(umaxp_5, umaxp(v0.V4S(), v1.V4S(), v2.V4S()))
|
H A D | test-api-movprfx-aarch64.cc | 2307 __ umaxp(z7.VnB(), p2.Merging(), z7.VnB(), z23.VnB()); in TEST() 3093 __ umaxp(z7.VnB(), p2.Merging(), z7.VnB(), z23.VnB()); in TEST() 3488 __ umaxp(z7.VnB(), p2.Merging(), z7.VnB(), z7.VnB()); in TEST()
|
H A D | test-disasm-sve-aarch64.cc | 6870 COMPARE(umaxp(z7.VnB(), p2.Merging(), z7.VnB(), z23.VnB()), in TEST() 6871 "umaxp z7.b, p2/m, z7.b, z23.b"); in TEST() 6872 COMPARE(umaxp(z7.VnD(), p2.Merging(), z7.VnD(), z23.VnD()), in TEST() 6873 "umaxp z7.d, p2/m, z7.d, z23.d"); in TEST() 6874 COMPARE(umaxp(z7.VnH(), p2.Merging(), z7.VnH(), z23.VnH()), in TEST() 6875 "umaxp z7.h, p2/m, z7.h, z23.h"); in TEST() 6876 COMPARE(umaxp(z7.VnS(), p2.Merging(), z7.VnS(), z23.VnS()), in TEST() 6877 "umaxp z7.s, p2/m, z7.s, z23.s"); in TEST() 6907 "umaxp z4.b, p1/m, z4.b, z31.b"); in TEST()
|
H A D | test-simulator-aarch64.cc | 4657 DEFINE_TEST_NEON_3SAME_NO2D(umaxp, Basic)
|
/third_party/node/deps/v8/src/codegen/arm64/ |
H A D | assembler-arm64.h | 1315 void umaxp(const VRegister& vd, const VRegister& vn, const VRegister& vm);
|
H A D | macro-assembler-arm64.h | 458 V(umaxp, Umaxp) \
|
H A D | assembler-arm64.cc | 3084 V(umaxp, NEON_UMAXP, vd.IsVector() && !vd.IsLaneSizeD()) \
|
/third_party/vixl/src/aarch64/ |
H A D | macro-assembler-sve-aarch64.cc | 640 V(Umaxp, umaxp) \
|
H A D | assembler-aarch64.h | 3186 void umaxp(const VRegister& vd, const VRegister& vn, const VRegister& vm); 6673 void umaxp(const ZRegister& zd,
|
H A D | simulator-aarch64.h | 4085 LogicVRegister umaxp(VectorFormat vform,
|
H A D | simulator-aarch64.cc | 3470 umaxp(vform, result, zdn, zm); in Simulator() 7517 umaxp(vf, rd, rn, rm); in Simulator()
|
/third_party/node/deps/v8/src/execution/arm64/ |
H A D | simulator-arm64.h | 1867 LogicVRegister umaxp(VectorFormat vform, LogicVRegister dst,
|
H A D | simulator-arm64.cc | 4347 umaxp(vf, rd, rn, rm);
|