/third_party/vixl/test/aarch64/ |
H A D | test-assembler-fp-aarch64.cc | 299 __ Fmov(tgt2, fp_tgt); in LoadFPValueHelper() 331 __ Fmov(s1, 255.0); in TEST() 332 __ Fmov(d2, 12.34567); in TEST() 333 __ Fmov(s3, 0.0); in TEST() 334 __ Fmov(d4, 0.0); in TEST() 335 __ Fmov(s5, kFP32PositiveInfinity); in TEST() 336 __ Fmov(d6, kFP64NegativeInfinity); in TEST() 337 __ Fmov(h7, RawbitsToFloat16(0x6400U)); in TEST() 338 __ Fmov(h8, kFP16PositiveInfinity); in TEST() 339 __ Fmov(s1 in TEST() [all...] |
H A D | test-assembler-neon-aarch64.cc | 3680 __ Fmov(v0.V4H(), 24.0); in TEST() 3681 __ Fmov(v1.V4H(), 1024.0); in TEST() 3682 __ Fmov(v2.V8H(), 5.5); in TEST() 3683 __ Fmov(v3.V8H(), 2048.0); in TEST() 3684 __ Fmov(v4.V8H(), kFP16PositiveInfinity); in TEST() 3685 __ Fmov(v5.V8H(), kFP16NegativeInfinity); in TEST() 3686 __ Fmov(v6.V4H(), RawbitsToFloat16(0x7c2f)); in TEST() 3687 __ Fmov(v7.V8H(), RawbitsToFloat16(0xfe0f)); in TEST() 3722 __ Fmov(v0.V4H(), 24.0); in TEST() 3723 __ Fmov(v in TEST() [all...] |
H A D | test-disasm-neon-aarch64.cc | 3206 COMPARE_MACRO(Fmov(v0.V2S(), 1.0f), "fmov v0.2s, #0x70 (1.0000)"); in TEST() 3207 COMPARE_MACRO(Fmov(v31.V2S(), -13.0f), "fmov v31.2s, #0xaa (-13.0000)"); in TEST() 3208 COMPARE_MACRO(Fmov(v0.V4S(), 1.0f), "fmov v0.4s, #0x70 (1.0000)"); in TEST() 3209 COMPARE_MACRO(Fmov(v31.V4S(), -13.0f), "fmov v31.4s, #0xaa (-13.0000)"); in TEST() 3210 COMPARE_MACRO(Fmov(v1.V2D(), 1.0), "fmov v1.2d, #0x70 (1.0000)"); in TEST() 3211 COMPARE_MACRO(Fmov(v29.V2D(), -13.0), "fmov v29.2d, #0xaa (-13.0000)"); in TEST() 3213 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0f)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 3214 COMPARE_MACRO(Fmov(v31.V8H(), Float16(29.0f)), in TEST() 3216 COMPARE_MACRO(Fmov(v0.V4H(), Float16(-5.0)), "fmov v0.4h, #0x94 (-5.0000)"); in TEST() 3217 COMPARE_MACRO(Fmov(v3 in TEST() [all...] |
H A D | test-assembler-aarch64.cc | 8199 __ Fmov(d1, x1); 8200 __ Fmov(d2, x2); 8201 __ Fmov(d3, x3); 8202 __ Fmov(d4, x4); 8267 __ Fmov(d1, x1); 8268 __ Fmov(d2, x2); 8269 __ Fmov(d3, x3); 8270 __ Fmov(d4, x4); 8271 __ Fmov(d5, x1); 8272 __ Fmov(d [all...] |
H A D | test-utils-aarch64.cc | 581 __ Fmov(dn, value); in ClobberFP() 586 __ Fmov(dn, first); in ClobberFP()
|
H A D | test-disasm-sve-aarch64.cc | 3065 COMPARE_MACRO(Fmov(z13.VnD(), p0.Merging(), 1.0), in TEST() 3067 COMPARE_MACRO(Fmov(z13.VnD(), p0.Merging(), 0.0), "mov z13.d, p0/m, #0"); in TEST() 3072 COMPARE_MACRO(Fmov(z13.VnD(), 1.0), "fmov z13.d, #0x70 (1.0000)"); in TEST() 3073 COMPARE_MACRO(Fmov(z13.VnD(), 0.0), "mov z13.d, #0"); in TEST() 3076 COMPARE_MACRO(Fmov(z13.VnD(), p0.Merging(), -0.0), in TEST() 3079 COMPARE_MACRO(Fmov(z13.VnD(), -0.0), "mov z13.d, #0x8000000000000000"); in TEST() 3299 COMPARE_MACRO(Fmov(z26.VnH(), Float16(0.0)), "mov z26.h, #0"); in TEST() 3300 COMPARE_MACRO(Fmov(z26.VnH(), Float16(0.0)), "mov z26.h, #0"); in TEST() 3301 COMPARE_MACRO(Fmov(z27.VnS(), 255.0f), in TEST() 3304 COMPARE_MACRO(Fmov(z2 in TEST() [all...] |
H A D | test-assembler-sve-aarch64.cc | 300 __ Fmov(d30, 42.0); 319 __ Fmov(s2, 1.5f); 320 __ Fmov(d3, d30); 5602 __ Fmov(d0, RawbitsToDouble(0x7ffaaaaa22223456)); 6650 __ Fmov(z9.VnD(), pg.Merging(), -9.0); 6658 // Fmov alias. 6659 __ Fmov(z14.VnS(), pg.Merging(), 0.0); 6660 __ Fmov(z15.VnH(), pg.Merging(), Float16(42.0)); 6661 __ Fmov(z16.VnD(), pg.Merging(), RawbitsToDouble(0x7ff0000012340000)); // NaN 6662 __ Fmov(z1 [all...] |
/third_party/vixl/src/aarch64/ |
H A D | macro-assembler-aarch64.cc | 1620 Fmov(tmp, value); in Emit() 1633 void MacroAssembler::Fmov(VRegister vd, double imm) { in Emit() function in vixl::aarch64::MacroAssembler 1645 Fmov(vd, Float16(imm)); in Emit() 1650 Fmov(vd, static_cast<float>(imm)); in Emit() 1673 void MacroAssembler::Fmov(VRegister vd, float imm) { in Emit() function in vixl::aarch64::MacroAssembler 1685 Fmov(vd, Float16(imm)); in Emit() 1690 Fmov(vd, static_cast<double>(imm)); in Emit() 1713 void MacroAssembler::Fmov(VRegister vd, Float16 imm) { in Emit() function in vixl::aarch64::MacroAssembler 1718 Fmov(vd, FPToFloat(imm, kIgnoreDefaultNaN)); in Emit() 1723 Fmov(v in Emit() [all...] |
H A D | macro-assembler-aarch64.h | 1648 void Fmov(const VRegister& vd, const VRegister& vn) { in Fmov() function in vixl::aarch64::MacroAssembler 1654 void Fmov(const VRegister& vd, const Register& rn) { in Fmov() function in vixl::aarch64::MacroAssembler 1660 void Fmov(const VRegister& vd, int index, const Register& rn) { in Fmov() function in vixl::aarch64::MacroAssembler 1669 void Fmov(const Register& rd, const VRegister& vn, int index) { in Fmov() function in vixl::aarch64::MacroAssembler 1683 void Fmov(VRegister vd, double imm); 1684 void Fmov(VRegister vd, float imm); 1685 void Fmov(VRegister vd, const Float16 imm); 1688 void Fmov(VRegister vd, T imm) { in Fmov() function in vixl::aarch64::MacroAssembler 1690 Fmov(vd, static_cast<double>(imm)); in Fmov() 1692 void Fmov(Registe in Fmov() function in vixl::aarch64::MacroAssembler 4665 void Fmov(const ZRegister& zd, double imm) { Fmov() function in vixl::aarch64::MacroAssembler 4669 void Fmov(const ZRegister& zd, float imm) { Fmov() function in vixl::aarch64::MacroAssembler 4673 void Fmov(const ZRegister& zd, Float16 imm) { Fmov() function in vixl::aarch64::MacroAssembler 4677 void Fmov(const ZRegister& zd, const PRegisterM& pg, double imm) { Fmov() function in vixl::aarch64::MacroAssembler 4681 void Fmov(const ZRegister& zd, const PRegisterM& pg, float imm) { Fmov() function in vixl::aarch64::MacroAssembler 4685 void Fmov(const ZRegister& zd, const PRegisterM& pg, Float16 imm) { Fmov() function in vixl::aarch64::MacroAssembler [all...] |
/third_party/node/deps/v8/src/codegen/arm64/ |
H A D | macro-assembler-arm64-inl.h | 544 Fmov(tmp, value); in Fcmp() 646 void TurboAssembler::Fmov(VRegister fd, VRegister fn) { in Fmov() function in v8::internal::TurboAssembler 657 void TurboAssembler::Fmov(VRegister fd, Register rn) { in Fmov() function in v8::internal::TurboAssembler 662 void TurboAssembler::Fmov(VRegister vd, double imm) { in Fmov() function in v8::internal::TurboAssembler 666 Fmov(vd, static_cast<float>(imm)); in Fmov() 690 void TurboAssembler::Fmov(VRegister vd, float imm) { in Fmov() function in v8::internal::TurboAssembler 693 Fmov(vd, static_cast<double>(imm)); in Fmov() 709 Fmov(vd, tmp); in Fmov() 717 void TurboAssembler::Fmov(Register rd, VRegister fn) { in Fmov() function in v8::internal::TurboAssembler
|
H A D | macro-assembler-arm64.h | 922 inline void Fmov(VRegister fd, VRegister fn); 923 inline void Fmov(VRegister fd, Register rn); 928 inline void Fmov(VRegister fd, double imm); 929 inline void Fmov(VRegister fd, float imm); 932 void Fmov(VRegister fd, T imm) { in Fmov() function in v8::internal::TurboAssembler 934 Fmov(fd, static_cast<double>(imm)); in Fmov() 936 inline void Fmov(Register rd, VRegister fn);
|
H A D | macro-assembler-arm64.cc | 3508 Fmov(new_arg, old_arg); in TruncateDoubleToI() 3525 Fmov(pcs[i].Reg(), temp0); in TruncateDoubleToI() 3527 Fmov(pcs[i].Reg(), args[i].VReg()); in TruncateDoubleToI() 3539 Fmov(pcs[i].VReg(), args[i].VReg()); in TruncateDoubleToI() 3739 Fmov(tmp, src); in TruncateDoubleToI() 3742 Fmov(dst, tmp); in TruncateDoubleToI()
|
/third_party/vixl/examples/aarch64/ |
H A D | simulator_interception.cc | 74 __ Fmov(s0, 3.5); in GenerateInterceptionExamples()
|
/third_party/node/deps/v8/src/wasm/baseline/arm64/ |
H A D | liftoff-assembler-arm64.h | 418 Fmov(reg.fp().S(), value.to_f32_boxed().get_scalar()); in LoadConstant() 421 Fmov(reg.fp().D(), value.to_f64_boxed().get_scalar()); in LoadConstant() 923 Fmov(dst.S(), src.S()); in Move() 925 Fmov(dst.D(), src.D()); in Move() 1353 Fmov(dst.S(), lhs.S()); in emit_f32_copysign() 1364 Fmov(dst.D(), lhs.D()); in emit_f64_copysign() 1447 Fmov(dst.gp().W(), src.fp().S()); in emit_type_conversion() 1488 Fmov(dst.gp().X(), src.fp().D()); in emit_type_conversion() 1506 Fmov(dst.fp().S(), src.gp().W()); in emit_type_conversion() 1524 Fmov(ds in emit_type_conversion() [all...] |
/third_party/vixl/benchmarks/aarch64/ |
H A D | bench-utils.cc | 373 __ Fmov(PickV(size), 1.25 * GetRandomBits(2)); in GenerateFPSequence()
|
/third_party/node/deps/v8/src/compiler/backend/arm64/ |
H A D | code-generator-arm64.cc | 1819 __ Fmov(i.OutputRegister32(), i.InputFloat32Register(0)); in AssembleArchInstruction() 1833 __ Fmov(i.OutputFloat64Register(), i.InputRegister(0)); in AssembleArchInstruction() 1839 __ Fmov(i.OutputRegister(), i.InputDoubleRegister(0)); in AssembleArchInstruction() 3430 __ Fmov(dst.S(), src.ToFloat32()); in AssembleArchInstruction() 3433 __ Fmov(dst, src.ToFloat64().value()); in AssembleArchInstruction() 3452 __ Fmov(temp, src.ToFloat32()); in AssembleArchInstruction() 3462 __ Fmov(temp, src.ToFloat64().value()); in AssembleArchInstruction()
|
/third_party/node/deps/v8/src/builtins/arm64/ |
H A D | builtins-arm64.cc | 666 __ Fmov(fp_zero, 0.0); in Generate_JSEntryVariant() 3375 __ Fmov(result, double_scratch); in Generate_DoubleToI()
|