/arkcompiler/runtime_core/static_core/runtime/bridge/arch/amd64/ |
H A D | deoptimization_amd64.S | 45 CFI_REL_OFFSET(r12, -((CFRAME_CALLEE_REGS_START_SLOT + 3) * 8)) 50 movq (%rcx), %r12 51 movq %r12, (-8)(%rcx) 68 leaq -((CFRAME_CALLEE_REGS_START_SLOT - 1) * 8)(%rcx), %r12 71 movq (%r12), %rbx 74 movq -8(%r12), %rbx 77 movq -16(%r12), %rbx 80 movq -24(%r12), %rbx 82 CFI_REL_OFFSET(r12, -((BOUNDARY_FRAME_SLOT + 3) * 8)) 83 movq -32(%r12), [all...] |
H A D | interpreter_to_compiled_code_bridge_dyn_amd64.S | 55 pushq %r12 56 CFI_REL_OFFSET(r12, -(6 * 8)) 65 // %rax - insn_ptr, %rbx - frame.vregs, %r12 - method, %r13, %r14 - temp, %r15 - acc 69 movq %rdx, %r12 // method 102 // %rax, %rdx - result, %r12 - frame.acc, %rcx - temp 103 movq (%rbp), %r12 // load iframe from the stack 104 addq $FRAME_ACC_OFFSET, %r12 107 movq %rax, (%r12) 113 popq %r12 114 CFI_RESTORE(r12) [all...] |
H A D | compiled_code_to_interpreter_bridge_dyn_amd64.S | 63 pushq %r12 64 CFI_REL_OFFSET(r12, -(5 * 8)) 105 // %r12 - iframe* 110 movq %rax, %r12 115 // %r12 - iframe* 124 movl FRAME_NUM_VREGS_OFFSET(%r12), %eax 126 leaq FRAME_VREGS_OFFSET(%r12, %rax, FRAME_VREGISTER_SIZE), %rax 164 leaq (EXTFRAME_EXT_DATA_OFFSET-EXTFRAME_FRAME_OFFSET)(%r12), %rax 179 movq %r12, %rsi 185 // %r12 [all...] |
H A D | compiled_code_to_interpreter_bridge_amd64.S | 62 pushq %r12 63 CFI_REL_OFFSET(r12, -(5 * 8)) 122 // %rax - SHORTY_PTR_REG, %r10d - SHORTY_REG, %r12 - iframe.vregs_ + num_vregs_, 129 shlq $3, %r12 130 leaq FRAME_VREGS_OFFSET(%r13, %r12), %r12 169 movq %rcx, (%r12, %r15, 8) 199 movq %r15, (%r12) 200 addq $FRAME_VREGISTER_SIZE, %r12 212 // %r12, [all...] |
H A D | compiled_code_to_runtime_bridge_amd64.S | 26 movq %r12, (\offset - CALLEE_REG0_OFFSET + 8*1)(%\base_reg) 27 CFI_REL_OFFSET(r12, (\offset - CALLEE_REG0_OFFSET+8*1)) 39 movq (\offset - CALLEE_REG0_OFFSET + 8*1)(%\base_reg), %r12 40 CFI_RESTORE(r12) 156 pushq %r12 158 CFI_REL_OFFSET(r12, 0) 160 movq MANAGED_THREAD_RUNTIME_CALL_ENABLED_OFFSET(%THREAD_REG), %r12 166 movq %r12, MANAGED_THREAD_RUNTIME_CALL_ENABLED_OFFSET(%THREAD_REG) 167 popq %r12 169 CFI_RESTORE(r12) [all...] |
H A D | interpreter_to_compiled_code_bridge_amd64.S | 250 pushq %r12 251 CFI_REL_OFFSET(r12, -(6 * 8)) 261 // %r9 - temp (used by PREPARE_ARG_STACK), %r12 - frame.vregs, %r13 - method, %r14 - method.shorty 263 leaq FRAME_VREGS_OFFSET(%rsi), %r12 // frame.vregs 290 // %rdi - float arg ptr, %rbx - insn_ptr, %r12 - frame.vregs, %r8 - arg base ptr 327 // %rax, %rdx / %xmm0 - result, %r14d - shorty[0] & 0xF, %r12 - frame.acc, %rcx - temp 334 movq (%rbp), %r12 // load iframe from the stack 335 addq $FRAME_ACC_OFFSET, %r12 344 movq %rdx, FRAME_ACC_MIRROR_OFFSET(%r12) 368 movsd %xmm0, (%r12) [all...] |
H A D | proxy_entrypoint_amd64.S | 57 pushq %r12 58 CFI_REL_OFFSET(r12, -(5 * 8)) 116 movq -((CFRAME_HEADER_SIZE - C2I_FP_OFFSET + 3) * 8)(%rbp), %r12 117 CFI_RESTORE(r12)
|
/arkcompiler/runtime_core/static_core/runtime/bridge/arch/arm/ |
H A D | deoptimization_arm.S | 67 sub r12, r3, #((CFRAME_CALLEE_REGS_START_SLOT - 1) * 4 - 4) 71 ldr r10, [r12, #-4]! 74 ldr r10, [r12, #-4]! 77 ldr r10, [r12, #-4]! 80 ldr r10, [r12, #-4]! 83 ldr r10, [r12, #-4]! 86 ldr r10, [r12, #-4]! 89 ldr r10, [r12, #-4]! 93 ldr r10, [r12, #-4]! 96 ldr r10, [r12, # [all...] |
H A D | compiled_code_to_runtime_bridge_arm.S | 159 ldrb r12, [THREAD_REG, #MANAGED_THREAD_FRAME_KIND_OFFSET] 160 tst r12, r12 173 ldr r12, [THREAD_REG, #MANAGED_THREAD_RUNTIME_CALL_ENABLED_OFFSET] 178 stm sp, {r12, lr} 181 mov r12, #0 182 str r12, [THREAD_REG, #MANAGED_THREAD_RUNTIME_CALL_ENABLED_OFFSET] 186 ldm sp, {r12, lr} 192 str r12, [THREAD_REG, #MANAGED_THREAD_RUNTIME_CALL_ENABLED_OFFSET]
|
H A D | compiled_code_to_interpreter_bridge_armhf.S | 124 // r7 - stack args, r8 - iframe, r9, r10, r12 - temps, lr - method 127 mov r12, r5, lsl #3 165 str r10, [r2, r12]
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/x64/ |
H A D | extended_assembler_x64.cpp | 22 r13, rbp, r12, rbx, r14, rsi, rdi, r8, r9 41 Pushq(r12); in PushCppCalleeSaveRegisters() 54 Popq(r12); in PopCppCalleeSaveRegisters() 66 Pushq(r12); in PushGhcCalleeSaveRegisters() 75 Popq(r12); in PopGhcCalleeSaveRegisters()
|
H A D | assembler_x64.h | 34 r12,
|
H A D | assembler_x64.cpp | 770 if (base == rsp || base == r12) { in Operand()
|
/arkcompiler/runtime_core/static_core/plugins/ets/runtime/napi/arch/amd64/ |
H A D | ets_async_entry_point_amd64.S | 53 pushq %r12 54 CFI_REL_OFFSET(r12, -((CFRAME_CALLEE_REGS_START_SLOT + 3) * 8)) 88 movq -((CFRAME_CALLEE_REGS_START_SLOT + 3) * 8)(%rbp), %r12 89 CFI_RESTORE(r12)
|
H A D | ets_napi_entry_point_amd64.S | 62 pushq %r12 63 CFI_REL_OFFSET(r12, -(4 * 8)) 122 popq %r12 123 CFI_RESTORE(r12) 204 // | N | m | | r12 | 3 279 pushq %r12 280 CFI_REL_OFFSET(r12, -((CFRAME_CALLEE_REGS_START_SLOT + 3) * 8)) 414 movq -((CFRAME_CALLEE_REGS_START_SLOT + 3) * 8)(%rbp), %r12 415 CFI_RESTORE(r12)
|
/arkcompiler/runtime_core/static_core/runtime/fibers/arch/amd64/ |
H A D | switch.S | 47 CFI_OFFSET(%r12, GPR_O(R12)) 48 movq GPR_O(R12)(%rsi), %r12
|
H A D | helpers.S | 27 movq %r12, GPR_O(R12)(\ctx_baseaddr_reg)
|
/arkcompiler/ets_runtime/ecmascript/compiler/trampoline/x64/ |
H A D | optimized_fast_call.cpp | 115 Register argvReg = r12;
in OptimizedFastCallAndPushArgv() 246 __ Movq(rcx, r12);
in JSFastCallWithArgV() 248 Register actualNumArgs(r12);
in JSFastCallWithArgV() 280 Register method = r12;
in JSFastCallWithArgV() 317 __ Movq(rcx, r12);
in JSFastCallWithArgVAndPushArgv() 320 Register actualNumArgsReg(r12);
in JSFastCallWithArgVAndPushArgv() 412 Register method = r12;
in JSFastCallWithArgVAndPushArgv()
|
H A D | common_call.h | 285 assembler_->Pushq(r12); in JsFunctionArgsConfigFrameScope() 294 assembler_->Addq(-5 * FRAME_SLOT_SIZE, rsp); // -5: get r12 r13 r14 rbx in ~JsFunctionArgsConfigFrameScope() 298 assembler_->Popq(r12); in ~JsFunctionArgsConfigFrameScope() 321 assembler_->Pushq(r12); // callee save in OptimizedUnfoldArgVFrameFrameScope() 326 assembler_->Addq(-4 * FRAME_SLOT_SIZE, rsp); // -4: get r12 rbx in ~OptimizedUnfoldArgVFrameFrameScope() 327 assembler_->Popq(r12); in ~OptimizedUnfoldArgVFrameFrameScope() 351 assembler_->Pushq(r12); // callee save in OptimizedUnfoldArgVFrameFrame1Scope() 358 assembler_->Addq(-6 * FRAME_SLOT_SIZE, rsp); // -6: get r12 r13 r14 rbx in ~OptimizedUnfoldArgVFrameFrame1Scope() 361 assembler_->Popq(r12); in ~OptimizedUnfoldArgVFrameFrame1Scope()
|
H A D | asm_interpreter_call.cpp | 103 Register regsArrayRegister = r12; in GeneratorReEnterAsmInterpDispatch() 118 Register pcRegister = r12; in GeneratorReEnterAsmInterpDispatch() 146 Register bitFieldRegister = r12; in AsmInterpEntryDispatch() 176 Register runtimeIdRegister = r12; in AsmInterpEntryDispatch() 179 Register trampolineIdRegister = r12; in AsmInterpEntryDispatch() 333 __ Movq(Operand(methodRegister, Method::NATIVE_POINTER_OR_BYTECODE_ARRAY_OFFSET), r12); // pc: r12 in JSCallCommonEntry() local 356 // %r12 - callTarget // %r12 - callTarget 680 // %r12 1781 __ Movl(Operand(r11, 0), r12); // r12: oldsetValue ASMFastSharedWriteBarrier() local [all...] |
H A D | optimized_call.cpp | 63 __ Movq(needPushArgv, r12); in JSFunctionEntry() 69 __ Cmp(1, r12); in JSFunctionEntry() 1369 __ Movq(Immediate(0), r12); in DeoptEnterAsmInterp() 1373 __ Cmpq(0, r12); in DeoptEnterAsmInterp() 1388 __ Addq(1, r12); in DeoptEnterAsmInterp() 1389 __ Cmpq(r12, depth); in DeoptEnterAsmInterp() 1394 // r13, rbp, r12, rbx, r14, rsi, rdi in DeoptEnterAsmInterp() 1397 __ Movq(Operand(frameStateBase, AsmInterpretedFrame::GetPcOffset(false)), r12); in DeoptEnterAsmInterp() local 1402 AsmInterpreterCall::DispatchCall(assembler, r12, opRegister, callTargetRegister, methodRegister, rsi); in DeoptEnterAsmInterp()
|
/arkcompiler/runtime_core/static_core/runtime/fibers/arch/arm/ |
H A D | helpers.S | 25 stm r\tmp_regid, {r4-r12}
|
H A D | switch.S | 35 ldm r3, {r4-r12}
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/code_generator/ |
H A D | target_info.h | 105 REG(r12) /* 12 */ \
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/code_generator/target/aarch32/ |
H A D | target.h | 37 // Temporary registers used (r12 already used by vixl) 40 vixl::aarch32::r8.GetCode(), vixl::aarch32::r9.GetCode(), vixl::aarch32::r12.GetCode()};
|