/third_party/node/deps/v8/src/wasm/baseline/riscv64/ |
H A D | liftoff-assembler-riscv64.h | 354 patching_assembler.GenPCRelativeJump(kScratchReg, imm32); in PatchPrepareStackFrame() 362 Register stack_limit = kScratchReg; in PatchPrepareStackFrame() 387 GenPCRelativeJump(kScratchReg, imm32); in PatchPrepareStackFrame() 561 VU.set(kScratchReg, E8, m1); in Load() 562 Register src_reg = src_op.offset() == 0 ? src_op.rm() : kScratchReg; in Load() 625 VU.set(kScratchReg, E8, m1); in Store() 626 Register dst_reg = dst_op.offset() == 0 ? dst_op.rm() : kScratchReg; in Store() 628 Add64(kScratchReg, dst_op.rm(), dst_op.offset()); in Store() 1008 VU.set(kScratchReg, E8, m1); in Spill() 1009 Register dst_reg = dst.offset() == 0 ? dst.rm() : kScratchReg; in Spill() [all...] |
/third_party/node/deps/v8/src/compiler/backend/riscv64/ |
H A D | code-generator-riscv64.cc | 462 __ PrepareCallCFunction(0, 2, kScratchReg); \ 473 __ PrepareCallCFunction(0, 1, kScratchReg); \ 488 __ VU.set(kScratchReg, E8, m1); \ 494 __ VU.set(kScratchReg, E16, m1); \ 500 __ VU.set(kScratchReg, E32, m1); \ 508 __ VU.set(kScratchReg, E8, m1); \ 513 __ VU.set(kScratchReg, E16, m1); \ 518 __ VU.set(kScratchReg, E32, m1); \ 523 __ VU.set(kScratchReg, E64, m1); \ 530 __ VU.set(kScratchReg, E 624 RelocInfo::CODE_TARGET, ne, kScratchReg, Operand(zero_reg)); BailoutIfDeoptimized() local [all...] |
/third_party/node/deps/v8/src/compiler/backend/mips/ |
H A D | code-generator-mips.cc | 369 __ PrepareCallCFunction(3, 0, kScratchReg); \ 390 kScratchReg, kScratchReg2); \ 399 __ PrepareCallCFunction(3, 0, kScratchReg); \ 505 __ PrepareCallCFunction(0, 2, kScratchReg); \ 516 __ PrepareCallCFunction(0, 1, kScratchReg); \ 584 __ ComputeCodeStartAddress(kScratchReg); in AssembleCodeStartRegisterCheck() 586 kJavaScriptCallCodeStartRegister, Operand(kScratchReg)); in AssembleCodeStartRegisterCheck() 598 __ lw(kScratchReg, MemOperand(kJavaScriptCallCodeStartRegister, offset)); in BailoutIfDeoptimized() 599 __ lw(kScratchReg, in BailoutIfDeoptimized() 600 FieldMemOperand(kScratchReg, in BailoutIfDeoptimized() 605 RelocInfo::CODE_TARGET, ne, kScratchReg, Operand(zero_reg)); BailoutIfDeoptimized() local [all...] |
/third_party/node/deps/v8/src/compiler/backend/ppc/ |
H A D | code-generator-ppc.cc | 28 #define kScratchReg r11 macro 349 i.InputRegister(1), kScratchReg, r0); \ 352 i.InputInt32(1), kScratchReg, r0); \ 360 i.InputRegister(1), kScratchReg, r0); \ 363 -i.InputInt32(1), kScratchReg, r0); \ 371 __ extsw(kScratchReg, kScratchReg, SetRC); \ 377 __ extsw(kScratchReg, kScratchReg, SetRC); \ 412 const Register scratch = kScratchReg; \ 2073 kScratchReg); AssembleArchInstruction() local 2079 kScratchReg); AssembleArchInstruction() local [all...] |
/third_party/node/deps/v8/src/compiler/backend/mips64/ |
H A D | code-generator-mips64.cc | 476 __ PrepareCallCFunction(0, 2, kScratchReg); \ 487 __ PrepareCallCFunction(0, 1, kScratchReg); \ 547 __ ComputeCodeStartAddress(kScratchReg); in AssembleCodeStartRegisterCheck() 549 kJavaScriptCallCodeStartRegister, Operand(kScratchReg)); in AssembleCodeStartRegisterCheck() 561 __ Ld(kScratchReg, MemOperand(kJavaScriptCallCodeStartRegister, offset)); in BailoutIfDeoptimized() 562 __ Lw(kScratchReg, in BailoutIfDeoptimized() 563 FieldMemOperand(kScratchReg, in BailoutIfDeoptimized() 565 __ And(kScratchReg, kScratchReg, in BailoutIfDeoptimized() 568 RelocInfo::CODE_TARGET, ne, kScratchReg, Operan in BailoutIfDeoptimized() local [all...] |
/third_party/node/deps/v8/src/wasm/baseline/mips64/ |
H A D | liftoff-assembler-mips64.h | 87 assm->daddu(kScratchReg, addr, offset); in GetMemOp() 88 return MemOperand(kScratchReg, offset_imm32); in GetMemOp() 91 assm->li(kScratchReg, offset_imm); in GetMemOp() 92 assm->daddu(kScratchReg, kScratchReg, addr); in GetMemOp() 94 assm->daddu(kScratchReg, kScratchReg, offset); in GetMemOp() 96 return MemOperand(kScratchReg, 0); in GetMemOp() 378 Register stack_limit = kScratchReg; in PatchPrepareStackFrame() 1120 TurboAssembler::li(kScratchReg, in emit_i32_divs() [all...] |
/third_party/node/deps/v8/src/wasm/baseline/loong64/ |
H A D | liftoff-assembler-loong64.h | 87 assm->add_d(kScratchReg, addr, offset); in GetMemOp() 88 return MemOperand(kScratchReg, offset_imm32); in GetMemOp() 91 assm->li(kScratchReg, Operand(offset_imm)); in GetMemOp() 92 assm->add_d(kScratchReg, kScratchReg, addr); in GetMemOp() 94 assm->add_d(kScratchReg, kScratchReg, offset); in GetMemOp() 96 return MemOperand(kScratchReg, 0); in GetMemOp() 261 Register stack_limit = kScratchReg; in PatchPrepareStackFrame() 1018 TurboAssembler::li(kScratchReg, in emit_i32_divs() [all...] |
/third_party/node/deps/v8/src/compiler/backend/s390/ |
H A D | code-generator-s390.cc | 27 #define kScratchReg ip macro 630 __ PrepareCallCFunction(0, 2, kScratchReg); \ 642 __ PrepareCallCFunction(0, 1, kScratchReg); \ 654 __ PrepareCallCFunction(0, 2, kScratchReg); \ 752 Register addr = kScratchReg; \ 768 Register addr = kScratchReg; \ 800 Register addr = kScratchReg; \ 950 Register temp = kScratchReg; \ 973 Register temp = kScratchReg; \ 1003 Register addr = kScratchReg; \ [all...] |
/third_party/node/deps/v8/src/wasm/baseline/mips/ |
H A D | liftoff-assembler-mips.h | 212 assm->TurboAssembler::Move(kScratchReg, tmp.low_gp()); in ChangeEndiannessLoad() 214 assm->TurboAssembler::ByteSwapSigned(tmp.high_gp(), kScratchReg, 4); in ChangeEndiannessLoad() 276 assm->TurboAssembler::Move(kScratchReg, tmp.low_gp()); in ChangeEndiannessStore() 278 assm->TurboAssembler::ByteSwapSigned(tmp.high_gp(), kScratchReg, 4); in ChangeEndiannessStore() 391 Register stack_limit = kScratchReg; in PatchPrepareStackFrame() 929 TurboAssembler::li(kScratchReg, 1); in emit_i32_divs() 931 TurboAssembler::LoadZeroOnCondition(kScratchReg, lhs, Operand(kMinInt), eq); in emit_i32_divs() 933 addu(kScratchReg, kScratchReg, kScratchReg2); in emit_i32_divs() 934 TurboAssembler::Branch(trap_div_unrepresentable, eq, kScratchReg, in emit_i32_divs() [all...] |
/third_party/node/deps/v8/src/codegen/riscv64/ |
H A D | macro-assembler-riscv64.cc | 1870 feq_d(kScratchReg, fs, fs); 1871 bnez(kScratchReg, &no_nan); 1878 feq_s(kScratchReg, fs, fs); 1879 bnez(kScratchReg, &no_nan); 3925 VU.set(kScratchReg, sew, lmul); in WasmRvvEq() 3927 li(kScratchReg, -1); in WasmRvvEq() 3929 vmerge_vx(dst, kScratchReg, dst); in WasmRvvEq() 3934 VU.set(kScratchReg, sew, lmul); in WasmRvvNe() 3936 li(kScratchReg, -1); in WasmRvvNe() 3938 vmerge_vx(dst, kScratchReg, ds in WasmRvvNe() [all...] |
H A D | register-riscv64.h | 262 constexpr Register kScratchReg = s3; member
|
/third_party/node/deps/v8/src/builtins/mips/ |
H A D | builtins-mips.cc | 709 __ LoadStackLimit(kScratchReg, in Generate_ResumeGeneratorTrampoline() 711 __ Branch(&stack_overflow, lo, sp, Operand(kScratchReg)); in Generate_ResumeGeneratorTrampoline() 733 __ Lsa(kScratchReg, t1, a3, kPointerSizeLog2); in Generate_ResumeGeneratorTrampoline() 734 __ Lw(kScratchReg, FieldMemOperand(kScratchReg, FixedArray::kHeaderSize)); in Generate_ResumeGeneratorTrampoline() 735 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 739 __ Lw(kScratchReg, FieldMemOperand(a1, JSGeneratorObject::kReceiverOffset)); in Generate_ResumeGeneratorTrampoline() 740 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 1194 temps.Exclude({kScratchReg, kScratchReg2}); in Generate_BaselineOutOfLinePrologue() 1218 __ lw(kScratchReg, in Generate_InterpreterEntryTrampoline() [all...] |
/third_party/node/deps/v8/src/builtins/riscv64/ |
H A D | builtins-riscv64.cc | 140 __ Ld(kScratchReg, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSBuiltinsConstructStubHelper() 145 __ DropArguments(kScratchReg, MacroAssembler::kCountIsSmi, in Generate_JSBuiltinsConstructStubHelper() 146 MacroAssembler::kCountIncludesReceiver, kScratchReg); in Generate_JSBuiltinsConstructStubHelper() 428 __ LoadStackLimit(kScratchReg, in Generate_ResumeGeneratorTrampoline() 430 __ Branch(&stack_overflow, Uless, sp, Operand(kScratchReg)); in Generate_ResumeGeneratorTrampoline() 456 __ CalcScaledAddress(kScratchReg, t1, a3, kTaggedSizeLog2); in Generate_ResumeGeneratorTrampoline() 458 kScratchReg, FieldMemOperand(kScratchReg, FixedArray::kHeaderSize)); in Generate_ResumeGeneratorTrampoline() 459 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 464 kScratchReg, FieldMemOperan in Generate_ResumeGeneratorTrampoline() [all...] |
/third_party/node/deps/v8/src/baseline/riscv64/ |
H A D | baseline-compiler-riscv64-inl.h | 66 __ masm()->Add64(kScratchReg, sp, in VerifyFrameSize() 69 __ masm()->Assert(eq, AbortReason::kUnexpectedStackPointer, kScratchReg, in VerifyFrameSize()
|
H A D | baseline-assembler-riscv64-inl.h | 24 wrapped_scope_.Include(kScratchReg, kScratchReg2); in ScratchRegisterScope()
|
/third_party/node/deps/v8/src/baseline/mips64/ |
H A D | baseline-compiler-mips64-inl.h | 67 __ masm()->Daddu(kScratchReg, sp, in VerifyFrameSize() 70 __ masm()->Assert(eq, AbortReason::kUnexpectedStackPointer, kScratchReg, in VerifyFrameSize()
|
/third_party/node/deps/v8/src/baseline/mips/ |
H A D | baseline-compiler-mips-inl.h | 67 __ masm()->Addu(kScratchReg, sp, in VerifyFrameSize() 70 __ masm()->Assert(eq, AbortReason::kUnexpectedStackPointer, kScratchReg, in VerifyFrameSize()
|
/third_party/node/deps/v8/src/builtins/mips64/ |
H A D | builtins-mips64.cc | 403 __ LoadStackLimit(kScratchReg, in Generate_ResumeGeneratorTrampoline() 405 __ Branch(&stack_overflow, lo, sp, Operand(kScratchReg)); in Generate_ResumeGeneratorTrampoline() 429 __ Dlsa(kScratchReg, t1, a3, kPointerSizeLog2); in Generate_ResumeGeneratorTrampoline() 430 __ Ld(kScratchReg, FieldMemOperand(kScratchReg, FixedArray::kHeaderSize)); in Generate_ResumeGeneratorTrampoline() 431 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 435 __ Ld(kScratchReg, FieldMemOperand(a1, JSGeneratorObject::kReceiverOffset)); in Generate_ResumeGeneratorTrampoline() 436 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 1202 temps.Exclude({kScratchReg, kScratchReg2}); in Generate_BaselineOutOfLinePrologue() 1226 __ Ld(kScratchReg, in Generate_InterpreterEntryTrampoline() [all...] |
/third_party/node/deps/v8/src/builtins/loong64/ |
H A D | builtins-loong64.cc | 404 __ LoadStackLimit(kScratchReg, in Generate_ResumeGeneratorTrampoline() 406 __ Branch(&stack_overflow, lo, sp, Operand(kScratchReg)); in Generate_ResumeGeneratorTrampoline() 430 __ Alsl_d(kScratchReg, a3, t1, kPointerSizeLog2, t7); in Generate_ResumeGeneratorTrampoline() 431 __ Ld_d(kScratchReg, FieldMemOperand(kScratchReg, FixedArray::kHeaderSize)); in Generate_ResumeGeneratorTrampoline() 432 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 436 __ Ld_d(kScratchReg, in Generate_ResumeGeneratorTrampoline() 438 __ Push(kScratchReg); in Generate_ResumeGeneratorTrampoline() 1222 __ Ld_d(kScratchReg, in Generate_InterpreterEntryTrampoline() 1226 FieldMemOperand(kScratchReg, SharedFunctionInf in Generate_InterpreterEntryTrampoline() [all...] |
/third_party/node/deps/v8/src/regexp/riscv64/ |
H A D | regexp-macro-assembler-riscv64.cc | 24 * - s3 : kScratchReg. Temporarily stores the index of capture start after a matching pass 1191 __ li(kScratchReg, Operand(entry, RelocInfo::OFF_HEAP_TARGET)); in CallCheckStackGuardState() 1192 __ Call(kScratchReg); in CallCheckStackGuardState() 1343 __ Add64(kScratchReg, end_of_input_address(), offset); in LoadCurrentCharacterUnchecked() 1344 __ Lwu(current_character(), MemOperand(kScratchReg)); in LoadCurrentCharacterUnchecked() 1346 __ Add64(kScratchReg, end_of_input_address(), offset); in LoadCurrentCharacterUnchecked() 1347 __ Lhu(current_character(), MemOperand(kScratchReg)); in LoadCurrentCharacterUnchecked() 1350 __ Add64(kScratchReg, end_of_input_address(), offset); in LoadCurrentCharacterUnchecked() 1351 __ Lbu(current_character(), MemOperand(kScratchReg)); in LoadCurrentCharacterUnchecked() 1356 __ Add64(kScratchReg, end_of_input_addres in LoadCurrentCharacterUnchecked() [all...] |
/third_party/node/deps/v8/src/codegen/loong64/ |
H A D | register-loong64.h | 151 constexpr Register kScratchReg = s3; member
|
/third_party/node/deps/v8/src/codegen/mips64/ |
H A D | macro-assembler-mips64.h | 858 li(kScratchReg, Operand(b)); in JumpIfEqual() 859 Branch(dest, eq, a, Operand(kScratchReg)); in JumpIfEqual() 863 li(kScratchReg, Operand(b)); in JumpIfLessThan() 864 Branch(dest, lt, a, Operand(kScratchReg)); in JumpIfLessThan()
|
H A D | register-mips64.h | 212 constexpr Register kScratchReg = s3; member
|
/third_party/node/deps/v8/src/codegen/mips/ |
H A D | macro-assembler-mips.h | 814 li(kScratchReg, Operand(b)); in JumpIfEqual() 815 Branch(dest, eq, a, Operand(kScratchReg)); in JumpIfEqual() 819 li(kScratchReg, Operand(b)); in JumpIfLessThan() 820 Branch(dest, lt, a, Operand(kScratchReg)); in JumpIfLessThan()
|
H A D | register-mips.h | 204 constexpr Register kScratchReg = s3; member
|