/arkcompiler/runtime_core/static_core/runtime/bridge/arch/amd64/ |
H A D | interpreter_to_compiled_code_bridge_dyn_amd64.S | 27 CFI_DEF_CFA(rsp, 8) 32 movq %rsp, %rax 61 subq $16, %rsp 62 // %rsp should be 16-byte aligned here 110 leaq -56(%rbp), %rsp 123 CFI_DEF_CFA(rsp, (3 * 8)) 124 addq $16, %rsp 139 CFI_DEF_CFA(rsp, 8) 144 movq %rsp, %rax 168 subq $8, %rsp [all...] |
H A D | compiled_code_to_interpreter_bridge_dyn_amd64.S | 33 CFI_DEF_CFA(rsp, 8) 40 movq (%rsp), %rax 50 movq %rsp, %rbp // set frame pointer 53 subq $8, %rsp 68 subq $8, %rsp 89 movq %rbp, %rsp 90 addq $16, %rsp 94 CFI_DEF_CFA(rsp, 8) 198 leaq -48(%rbp), %rsp 209 addq $8, %rsp [all...] |
H A D | compiled_code_to_interpreter_bridge_amd64.S | 36 CFI_DEF_CFA(rsp, 8) 39 movq (%rsp), %rax 49 movq %rsp, %rbp // set frame pointer 52 subq $8, %rsp 67 subq $8, %rsp 90 movq %rbp, %rsp 91 addq $16, %rsp 95 CFI_DEF_CFA(rsp, 8) 111 // %rsp % 16 == 0 here 131 leaq 8(%rsp), [all...] |
H A D | compiled_code_to_runtime_bridge_amd64.S | 154 CFI_DEF_CFA(rsp, 8) 177 subq $(BRIDGE_FRAME_SIZE - 8), %rsp 180 SAVE_CALLEE_GP_REGS rsp, BRIDGE_FRAME_SIZE 188 mov (BRIDGE_FRAME_SIZE - 1 * 8)(%rsp), %r14 193 movq $COMPILED_CODE_TO_INTERPRETER_BRIDGE, (BRIDGE_FRAME_SIZE - 2 * 8)(%rsp) 195 mov %rbp, (BRIDGE_FRAME_SIZE - 3 * 8)(%rsp) 198 leaq (BRIDGE_FRAME_SIZE - 3 * 8)(%rsp), %r13 227 RESTORE_CALLEE_GP_REGS rsp, BRIDGE_FRAME_SIZE 239 addq $(BRIDGE_FRAME_SIZE - 8), %rsp 245 addq $(BRIDGE_FRAME_SIZE - 8), %rsp [all...] |
H A D | proxy_entrypoint_amd64.S | 35 CFI_DEF_CFA(rsp, 8) 37 // %rsp % 16 == 8 44 movq %rsp, %rbp // frame pointer 47 subq $8, %rsp 62 subq $8, %rsp 63 // %rsp % 16 == 0 77 // %rsp % 16 == 0 79 movq %rsp, %rsi 121 movq %rbp, %rsp 124 CFI_DEF_CFA(rsp, ( [all...] |
H A D | interpreter_to_compiled_code_bridge_amd64.S | 131 // +--------------+ <- %rsp 158 // %rsp - Points to the last stack argument. 164 movq %rsp, %rdx 208 // make %rsp 16 bytes aligned 211 movq %rdx, %rsp 214 leaq -(6 * 8)(%rsp), %r8 222 CFI_DEF_CFA(rsp, 8) 227 movq %rsp, %rax 256 subq $16, %rsp // do not delete, used to copy args 257 // %rsp shoul [all...] |
H A D | deoptimization_amd64.S | 32 // 8(%rsp) - callee vreg_mask 48 leaq -(((CFRAME_HEADER_SIZE - 2) * 8) + CALLEE_SAVED_SIZE)(%rcx), %rsp 107 movq %rcx, %rsp 127 CFI_DEF_CFA(rsp, (2 * 8)) 128 addq $8, %rsp 157 movq %rcx, %rsp 179 CFI_DEF_CFA(rsp, (1 * 8)) 200 movq %rdi, %rsp 221 CFI_DEF_CFA(rsp, (1 * 8))
|
H A D | expand_compiled_code_args_dyn_amd64.S | 31 // %r8 - %rsp before moving 37 movq %rsp, %r8 40 addq %rdx, %rsp
|
/arkcompiler/runtime_core/static_core/runtime/tests/arch/amd64/ |
H A D | invokation_helper.S | 23 CFI_DEF_CFA(rsp, 8) 33 movq %rsp, %rbp 59 sub %r8, %rsp 60 movq %rsp, %r8 86 leaq -16(%rbp), %rsp 93 CFI_DEF_CFA(rsp, (1 * 8)) 106 CFI_DEF_CFA(rsp, 8) 111 movq %rsp, %rbp 120 movq %rbp, %rsp 123 CFI_DEF_CFA(rsp, [all...] |
H A D | stack_walker_hook_bridge.S | 22 CFI_DEF_CFA(rsp, 8) 23 movq (%rsp), %rdi 25 subq $8, %rsp 28 addq $8, %rsp
|
/arkcompiler/runtime_core/static_core/plugins/ets/runtime/napi/arch/amd64/ |
H A D | ets_napi_entry_point_amd64.S | 46 CFI_DEF_CFA(rsp, 8) 48 // %rsp % 16 == 8 53 movq %rsp, %rbp // frame pointer 66 subq $(CFRAME_LOCALS_COUNT * 8), %rsp 68 // %rsp % 16 == 0 74 // %rsp % 16 == 0 97 movq %rsp, %rsi 99 subq %rax, %rsp // %rsp % 16 == 0 100 movq %rsp, [all...] |
H A D | ets_async_entry_point_amd64.S | 29 CFI_DEF_CFA(rsp, 8) 31 // %rsp % 16 == 8 36 movq %rsp, %rbp // frame pointer 43 subq $(CFRAME_LOCALS_COUNT * 8), %rsp 58 subq $8, %rsp 59 // %rsp % 16 == 0 64 // %rsp % 16 == 0 65 movq %rsp, %rdx 92 movq %rbp, %rsp 95 CFI_DEF_CFA(rsp, ( [all...] |
/arkcompiler/ets_runtime/ecmascript/compiler/trampoline/x64/ |
H A D | optimized_call.cpp | 79 __ Addq(FRAME_SLOT_SIZE, rsp); // 8: frame type in JSFunctionEntry() 128 __ Movq(Operand(rsp, funcSlotOffset * FRAME_SLOT_SIZE), jsFuncReg); // sp + 24 get jsFunc in OptimizedCallAndPushArgv() 138 __ Movl(Operand(rsp, FRAME_SLOT_SIZE), rdx); // argc rdx in OptimizedCallAndPushArgv() 139 __ Movq(rsp, r8); in OptimizedCallAndPushArgv() 149 __ Leaq(Operand(rsp, FRAME_SLOT_SIZE), rbp); in OptimizedCallAndPushArgv() 199 __ Pushq(rsp); // actual argv in OptimizedCallAndPushArgv() 204 __ Addq(codeAddrReg, rsp); in OptimizedCallAndPushArgv() 205 __ Addq(FRAME_SLOT_SIZE, rsp); // skip actualNumArgsReg in OptimizedCallAndPushArgv() 206 __ Addq(FRAME_SLOT_SIZE, rsp); // skip argvReg in OptimizedCallAndPushArgv() 209 __ Addq(8, rsp); // in OptimizedCallAndPushArgv() 1448 __ Movq(Operand(context, AsmStackContext::GetCallFrameTopOffset(false)), rsp); DeoptHandlerAsm() local [all...] |
H A D | common_call.cpp | 49 __ Leaq(Operand(rsp, 24), rbp); // 24: skip pc, prevSp and frame type in PushAsmInterpBridgeFrame() 70 __ Addq(8, rsp); // 8: skip pc in PopAsmInterpBridgeFrame() 72 __ Addq(8, rsp); // 8: skip frame type in PopAsmInterpBridgeFrame() 115 __ Movq(rsp, temp2); in StackOverflowCheck() 130 __ Leaq(Operand(rsp, FRAME_SLOT_SIZE), rbp); in PushAsmBridgeFrame() 135 __ Addq(FRAME_SLOT_SIZE, rsp); // skip type in PopAsmBridgeFrame()
|
H A D | asm_interpreter_call.cpp | 101 __ Movq(rsp, fpRegister); in GeneratorReEnterAsmInterpDispatch() 115 __ Movq(rsp, newSpRegister); in GeneratorReEnterAsmInterpDispatch() 185 __ Addq(16, rsp); // 16: skip argc and runtime_id in AsmInterpEntryDispatch() 249 __ Leaq(Operand(rsp, 3 * FRAME_SLOT_SIZE), rbp); // 3: 24 means skip frame type, prevSp and pc in PushAsmInterpEntryFrame() 254 __ Addq(8, rsp); // 8: skip pc in PopAsmInterpEntryFrame() 257 __ Addq(FRAME_SLOT_SIZE, rsp); // 8: skip frame type in PopAsmInterpEntryFrame() 298 __ Movq(rsp, fpRegister); in JSCallCommonEntry() 321 __ Movq(fpRegister, rsp); in JSCallCommonEntry() 714 __ Movq(rsp, newSpRegister); in PushVregs() 732 __ Testq(15, rsp); // 1 in PushVregs() 891 __ Movq(Operand(glue, JSThread::GlueData::GetLeaveFrameOffset(false)), rsp); CallNativeWithArgv() local 1160 __ Movq(Operand(frameStateBaseRegister, AsmInterpretedFrame::GetFpOffset(false)), rsp); // resume rsp ResumeRspAndDispatch() local 1179 __ Movq(Operand(frameStateBaseRegister, AsmInterpretedFrame::GetFpOffset(false)), rsp); // resume rsp ResumeRspAndDispatch() local 1502 __ Movq(Operand(frameStateBaseRegister, AsmInterpretedFrame::GetFpOffset(false)), rsp); // resume rsp ResumeRspAndRollback() local [all...] |
H A D | common_call.h | 283 assembler_->Leaq(Operand(rsp, FRAME_SLOT_SIZE), rbp); in JsFunctionArgsConfigFrameScope() 293 assembler_->Movq(rbp, rsp); in ~JsFunctionArgsConfigFrameScope() 294 assembler_->Addq(-5 * FRAME_SLOT_SIZE, rsp); // -5: get r12 r13 r14 rbx in ~JsFunctionArgsConfigFrameScope() 299 assembler_->Addq(FRAME_SLOT_SIZE, rsp); // skip frame type in ~JsFunctionArgsConfigFrameScope() 319 assembler_->Leaq(Operand(rsp, 2 * FRAME_SLOT_SIZE), rbp); in OptimizedUnfoldArgVFrameFrameScope() 325 assembler_->Movq(rbp, rsp); in ~OptimizedUnfoldArgVFrameFrameScope() 326 assembler_->Addq(-4 * FRAME_SLOT_SIZE, rsp); // -4: get r12 rbx in ~OptimizedUnfoldArgVFrameFrameScope() 329 assembler_->Addq(2 * FRAME_SLOT_SIZE, rsp); // 2: skip frame type and sp in ~OptimizedUnfoldArgVFrameFrameScope() 349 assembler_->Leaq(Operand(rsp, 2 * FRAME_SLOT_SIZE), rbp); in OptimizedUnfoldArgVFrameFrame1Scope() 357 assembler_->Movq(rbp, rsp); in ~OptimizedUnfoldArgVFrameFrame1Scope() [all...] |
H A D | optimized_fast_call.cpp | 57 __ Addq(FRAME_SLOT_SIZE, rsp); // 8: frame type
in OptimizedFastCallEntry() 116 __ Leaq(Operand(rsp, 8 * FRAME_SLOT_SIZE), argvReg); // 8: skip 8 frames to get argv
in OptimizedFastCallAndPushArgv() 238 Register sp(rsp);
in JSFastCallWithArgV() 299 Register sp(rsp);
in JSFastCallWithArgVAndPushArgv()
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/x64/ |
H A D | macro_assembler_x64.cpp | 24 x64::Register baseReg = (dstStackSlot.IsFrameBase()) ? x64::rbp : x64::rsp; in Move() 33 x64::Register dstBaseReg = (dstStackSlot.IsFrameBase()) ? x64::rbp : x64::rsp; in Move() 34 x64::Register srcBaseReg = (srcStackSlot.IsFrameBase()) ? x64::rbp : x64::rsp; in Move() 43 x64::Register baseReg = (stackSlot.IsFrameBase()) ? x64::rbp : x64::rsp; in Cmp() 71 x64::Register dstBaseReg = (dstStackSlot.IsFrameBase()) ? x64::rbp : x64::rsp; in SaveReturnRegister() 133 x64::Register dstBaseReg = (stackSlotOpnd.IsFrameBase()) ? x64::rbp : x64::rsp; in MovParameterIntoParamReg()
|
H A D | extended_assembler_x64.cpp | 30 Subq(8, rsp); // 8: 8 bytes in PushAlignBytes() 35 Addq(8, rsp); // 8: 8 bytes in PopAlignBytes() 59 Addq(8 * 5, rsp); // 8: 8 bytes, 5: number of CalleeSaveRegisters in UpdateCalleeSaveRegisters()
|
H A D | assembler_x64.cpp | 770 if (base == rsp || base == r12) { in Operand() 771 BuildSIB(Times1, rsp, base); in Operand() 792 BuildModerm(0, rsp); in Operand() 795 BuildModerm(1, rsp); in Operand() 799 BuildModerm(2, rsp); in Operand() 807 ASSERT(index != rsp); in Operand() 808 BuildModerm(0, rsp); in Operand() 903 if (dst >= rsp) { in Testb()
|
/arkcompiler/runtime_core/static_core/runtime/fibers/arch/amd64/ |
H A D | helpers.S | 31 // context->RIP = return address from stack, i.e. *%rsp 32 movq (%rsp), \tmp_reg 34 // context->RSP = %rsp + sizeof(void*), i.e. like we just returned from GetCurrentContext 35 leaq 8(%rsp), \tmp_reg
|
H A D | switch.S | 59 CFI_OFFSET(%rsp, GPR_O(RSP)) 60 movq GPR_O(RSP)(%rsi), %rsp
|
/arkcompiler/runtime_core/static_core/runtime/arch/amd64/ |
H A D | interpreter_support.S | 23 .cfi_def_cfa rsp, __SIZEOF_POINTER__
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/code_generator/target/amd64/ |
H A D | callconv.cpp | 69 GetMasm()->sub(asmjit::x86::rsp, asmjit::imm(DOUBLE_WORD_SIZE_BYTES)); in PushRegs() 70 GetMasm()->movsd(asmjit::x86::ptr(asmjit::x86::rsp), asmjit::x86::xmm(ii)); in PushRegs() 100 GetMasm()->movsd(asmjit::x86::xmm(i), asmjit::x86::ptr(asmjit::x86::rsp)); in PopRegs() 101 GetMasm()->add(asmjit::x86::rsp, asmjit::imm(DOUBLE_WORD_SIZE_BYTES)); in PopRegs()
|
/arkcompiler/runtime_core/static_core/plugins/ets/tests/napi/sampler_sig_stack/amd64/ |
H A D | SamplerSigStackAmd64.S | 37 movq %rsp, %rbp 38 // subq $184, %rsp // Move stack upper than is should be
|