/arkcompiler/ets_runtime/ecmascript/compiler/trampoline/x64/ |
H A D | optimized_fast_call.cpp | 127 __ Cmp(Immediate(3), actualNumArgsReg); // 3: func new this
in OptimizedFastCallAndPushArgv() 139 __ Cmp(Immediate(4), actualNumArgsReg); // 4: func new this arg0
in OptimizedFastCallAndPushArgv() 151 __ Cmp(Immediate(5), actualNumArgsReg); // 5: 5 args
in OptimizedFastCallAndPushArgv() 162 __ Cmp(Immediate(6), actualNumArgsReg); // 6: 6 args
in OptimizedFastCallAndPushArgv() 207 __ Cmp(Immediate(3), expectedNumArgsReg); // 3: expectedNumArgsReg <= 3 jump
in OptimizedFastCallAndPushArgv() 217 __ Cmp(0, expectedNumArgsReg);
in OptimizedFastCallAndPushArgv() 251 __ Cmp(0, actualNumArgs);
in JSFastCallWithArgV() 257 __ Cmp(0, actualNumArgs);
in JSFastCallWithArgV() 263 __ Cmp(0, actualNumArgs);
in JSFastCallWithArgV() 269 __ Cmp( in JSFastCallWithArgV() [all...] |
/arkcompiler/ets_runtime/ecmascript/compiler/trampoline/aarch64/ |
H A D | optimized_fast_call.cpp | 138 __ Cmp(actualNumArgsReg, Immediate(3)); // 3: 3 args
in OptimizedFastCallAndPushArgv() 150 __ Cmp(actualNumArgsReg, Immediate(4)); // 4: 4 args
in OptimizedFastCallAndPushArgv() 162 __ Cmp(actualNumArgsReg, Immediate(5)); // 5: 5 args
in OptimizedFastCallAndPushArgv() 175 __ Cmp(actualNumArgsReg, Immediate(6)); // 6: 6 args
in OptimizedFastCallAndPushArgv() 187 __ Cmp(actualNumArgsReg, Immediate(7)); // 7: 7 args
in OptimizedFastCallAndPushArgv() 198 __ Cmp(actualNumArgsReg, Immediate(8)); // 8: 8 args
in OptimizedFastCallAndPushArgv() 210 __ Cmp(expectedNumArgs, actualNumArgsReg);
in OptimizedFastCallAndPushArgv() 230 __ Cmp(expectedNumArgs, Immediate(8)); // 8 : register save 8 arg
in OptimizedFastCallAndPushArgv() 280 __ Cmp(argC, Immediate(0));
in JSFastCallWithArgV() 287 __ Cmp(arg in JSFastCallWithArgV() [all...] |
H A D | common_call.cpp | 102 __ Cmp(argc.W(), Immediate(0)); in PushArgsWithArgv() 121 __ Cmp(argc.W(), Immediate(0)); in PushArgsWithArgvInPair() 141 __ Cmp(argc.W(), Immediate(1)); // 1: argc is odd number in copyArgs in PushArgsWithArgvInPair() 148 __ Cmp(argc.W(), Immediate(1)); // 1: argc is odd number in copyArgs in PushArgsWithArgvInPair() 164 __ Cmp(argc.W(), Immediate(0)); in PushUndefinedWithArgc() 184 __ Cmp(op, Operand(numArgs, LSL, 3)); // 3: each args occupies 8 bytes in StackOverflowCheck()
|
H A D | optimized_call.cpp | 159 __ Cmp(Register(X6), Immediate(1)); in JSFunctionEntry() 237 __ Cmp(expectedNumArgs, actualNumArgs); in OptimizedCallAndPushArgv() 498 __ Cmp(actualArgC, Immediate(kungfu::ArgumentAccessor::GetFixArgsNum() + 3)); in JSCallInternal() 515 __ Cmp(Register(X5).W(), Immediate(kungfu::BuiltinsStubCSigns::BUILTINS_CONSTRUCTOR_STUB_FIRST)); in JSCallInternal() 534 __ Cmp(Register(X5), Immediate(0)); in JSCallInternal() 536 __ Cmp(Register(X5), Immediate(1)); in JSCallInternal() 538 __ Cmp(Register(X5), Immediate(2)); // 2: 2 args in JSCallInternal() 540 __ Cmp(Register(X5), Immediate(3)); // 3: 3 args in JSCallInternal() 600 __ Cmp(jstype2, Immediate(static_cast<int64_t>(JSType::JS_BOUND_FUNCTION))); in JSCallInternal() 602 __ Cmp(jstype in JSCallInternal() [all...] |
H A D | asm_interpreter_call.cpp | 88 __ Cmp(functionTypeRegister, tempRegister.W()); in AsmInterpEntryDispatch() 91 __ Cmp(functionTypeRegister, tempRegister.W()); in AsmInterpEntryDispatch() 155 __ Cmp(declaredNumArgsRegister, Immediate(argc)); in JSCallCommonEntry() 157 __ Cmp(declaredNumArgsRegister, argcRegister); in JSCallCommonEntry() 293 __ Cmp(declaredNumArgsRegister, Immediate(0)); in JSCallCommonSlowPath() 306 __ Cmp(declaredNumArgsRegister, Immediate(1)); in JSCallCommonSlowPath() 728 __ Cmp(jumpSizeRegister, Immediate(0)); in ResumeRspAndDispatch() 746 __ Cmp(ret, Immediate(JSTaggedValue::VALUE_UNDEFINED)); in ResumeRspAndDispatch() 764 __ Cmp(temp, Immediate(0)); in ResumeRspAndDispatch() 770 __ Cmp(tem in ResumeRspAndDispatch() [all...] |
/arkcompiler/runtime_core/compiler/optimizer/templates/ |
H A D | codegen_arm64.rb | 85 "__ Cmp(#{a}), #{b});" 89 %Q(__ Cmp(#{a}, #{b}); 94 %Q(__ Cmp(#{a}, #{b});
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/templates/ |
H A D | codegen_arm64.rb | 85 "__ Cmp(#{a}), #{b});" 89 %Q(__ Cmp(#{a}, #{b}); 94 %Q(__ Cmp(#{a}, #{b});
|
/arkcompiler/runtime_core/static_core/libpandafile/ |
H A D | pgo.cpp | 80 static bool Cmp(const std::unique_ptr<BaseItem> &item1, const std::unique_ptr<BaseItem> &item2) in Cmp() function 125 items.sort(Cmp); in ProfileGuidedRelayout()
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/tests/ |
H A D | assembler_aarch64_test.cpp | 233 __ Cmp(Register(X1), Register(X2)); in HWTEST_F_L0() 234 __ Cmp(Register(X1), Immediate(8)); in HWTEST_F_L0() 253 __ Cmp(Register(X1), Register(X2)); in HWTEST_F_L0() 255 __ Cmp(Register(X1), Immediate(8)); in HWTEST_F_L0() 289 __ Cmp(count, Immediate(0)); in HWTEST_F_L0()
|
/arkcompiler/ets_runtime/ecmascript/compiler/baseline/ |
H A D | baseline_assembler.cpp | 96 void BaselineAssembler::Cmp(SpecialRegister reg, Immediate value) in Cmp() function in panda::ecmascript::kungfu::BaselineAssembler 101 macroAssembler->Cmp(stackSlotOperand, value); in Cmp()
|
H A D | baseline_assembler.h | 117 void Cmp(SpecialRegister reg, Immediate value);
|
/arkcompiler/runtime_core/compiler/tests/ |
H A D | vn_test.cpp | 95 // Remove duplicate Cast, AddI, Fcmp and Cmp instructions in TEST_F() 109 INST(8, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT64).Fcmpg(true).Inputs(2, 3); in TEST_F() 110 INST(9, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT32).Fcmpg(false).Inputs(4, 5); in TEST_F() 124 INST(16, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT64).Fcmpg(true).Inputs(2, 3); in TEST_F() 125 INST(17, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT32).Fcmpg(false).Inputs(4, 5); in TEST_F() 154 INST(8, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT64).Fcmpg(true).Inputs(2, 3); in TEST_F() 155 INST(9, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT32).Fcmpg(false).Inputs(4, 5); in TEST_F() 169 INST(16, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT64).Fcmpg(true).Inputs(2, 3); in TEST_F() 170 INST(17, Opcode::Cmp).s32().SrcType(DataType::Type::FLOAT32).Fcmpg(false).Inputs(4, 5); in TEST_F() 260 // - Cmp instruction in TEST_F() [all...] |
H A D | graph_comparator.h | 116 CHECK_INST(Cmp, GetOperandsType) in Compare() 240 if (inst1->GetOpcode() != Opcode::Cmp || !IsFloatType(inst1->GetInput(0).GetInst()->GetType())) { in CompareFcmpgInst()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/src/cg/x86_64/ |
H A D | x64_emitter.cpp | 998 assmbler.Cmp(kB, TransferReg(opnd0), TransferReg(opnd1)); in EmitInsn() 1001 assmbler.Cmp(kW, TransferReg(opnd0), TransferReg(opnd1)); in EmitInsn() 1004 assmbler.Cmp(kL, TransferReg(opnd0), TransferReg(opnd1)); in EmitInsn() 1007 assmbler.Cmp(kQ, TransferReg(opnd0), TransferReg(opnd1)); in EmitInsn() 1010 assmbler.Cmp(kB, TransferImm(opnd0), TransferReg(opnd1)); in EmitInsn() 1013 assmbler.Cmp(kW, TransferImm(opnd0), TransferReg(opnd1)); in EmitInsn() 1016 assmbler.Cmp(kL, TransferImm(opnd0), TransferReg(opnd1)); in EmitInsn() 1019 assmbler.Cmp(kQ, TransferImm(opnd0), TransferReg(opnd1)); in EmitInsn() 1022 assmbler.Cmp(kB, TransferMem(opnd0, funcUniqueId), TransferReg(opnd1)); in EmitInsn() 1025 assmbler.Cmp(k in EmitInsn() [all...] |
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/x64/ |
H A D | macro_assembler_x64.cpp | 41 void MacroAssemblerX64::Cmp(const StackSlotOperand &stackSlot, Immediate value) in Cmp() function in panda::ecmascript::kungfu::MacroAssemblerX64 46 assembler.Cmp(x64::Immediate(value.GetValue()), LOCAL_SCOPE_REGISTER); in Cmp()
|
H A D | macro_assembler_x64.h | 39 void Cmp(const StackSlotOperand &stackSlot, Immediate value) override;
|
/arkcompiler/runtime_core/bytecode_optimizer/tests/ |
H A D | bitops_bitwise_and_test.cpp | 103 INST(13, Opcode::Cmp).s32().Inputs(4, 22); in TEST_F()
|
H A D | common.h | 366 CHECK(Cmp, GetOperandsType) in Compare() 401 if (inst1->GetOpcode() == Opcode::Cmp && IsFloatType(inst1->GetInput(0).GetInst()->GetType())) { in Compare() 537 INST(2, Opcode::Cmp).s32().Inputs(0, 1); 569 INST(2, Opcode::Cmp).s32().SrcType(data_type).Fcmpg(fcmpg).Inputs(0, 1);
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/ |
H A D | macro_assembler.h | 78 virtual void Cmp(const StackSlotOperand &stackSlot, Immediate value) = 0;
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/code_generator/target/aarch32/ |
H A D | encode.cpp | 533 GetMasm()->Cmp(VixlReg(src), VixlImm(value)); in ComparePosImmHelper() 536 GetMasm()->Cmp(VixlRegU(src), VixlImm(0x0)); in ComparePosImmHelper() 537 GetMasm()->Cmp(Convert(Condition::EQ), VixlReg(src), VixlImm(value)); in ComparePosImmHelper() 547 GetMasm()->Cmp(VixlReg(src), VixlImm(value)); in ComparePosImmHelper() 584 GetMasm()->Cmp(VixlReg(src), vixl::aarch32::Operand(0x0)); in CompareZeroHelper() 600 GetMasm()->Cmp(VixlReg(src), vixl::aarch32::Operand(imm)); in CompareZeroHelper() 609 GetMasm()->Cmp(VixlReg(src), vixl::aarch32::Operand(imm)); in CompareZeroHelper() 1070 GetMasm()->Cmp(VixlReg(src), VixlImm(0x0)); in EncodeAbs() 1082 GetMasm()->Cmp(VixlRegU(src), VixlImm(0x0)); in EncodeAbs() 1116 GetMasm()->Cmp(VixlRe in EncodeIsInf() [all...] |
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/include/cg/x86_64/assembler/ |
H A D | asm_assembler.h | 272 void Cmp(InsnSize insnSize, Reg srcReg, Reg destReg) override; 273 void Cmp(InsnSize insnSize, const Mem &mem, Reg reg) override; 274 void Cmp(InsnSize insnSize, Reg reg, const Mem &mem) override; 275 void Cmp(InsnSize insnSize, const ImmOpnd &immOpnd, Reg reg) override; 276 void Cmp(InsnSize insnSize, const ImmOpnd &immOpnd, const Mem &mem) override;
|
H A D | assembler.h | 285 virtual void Cmp(InsnSize insnSize, Reg srcReg, Reg destReg) = 0; 286 virtual void Cmp(InsnSize insnSize, const Mem &mem, Reg reg) = 0; 287 virtual void Cmp(InsnSize insnSize, Reg reg, const Mem &mem) = 0; 288 virtual void Cmp(InsnSize insnSize, const ImmOpnd &immOpnd, Reg reg) = 0; 289 virtual void Cmp(InsnSize insnSize, const ImmOpnd &immOpnd, const Mem &mem) = 0;
|
/arkcompiler/runtime_core/static_core/bytecode_optimizer/tests/ |
H A D | bitops_bitwise_and_test.cpp | 108 INST(13U, Opcode::Cmp).s32().Inputs(4U, 22U); in TEST_F()
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/aarch64/ |
H A D | macro_assembler_aarch64.h | 44 void Cmp(const StackSlotOperand &stackSlot, Immediate value) override;
|
/arkcompiler/runtime_core/static_core/compiler/optimizer/code_generator/target/aarch64/ |
H A D | encode.cpp | 296 GetMasm()->Cmp(VixlReg(src0), VixlReg(src1)); in EncodeJump() 316 GetMasm()->Cmp(VixlReg(src), VixlImm(value)); in EncodeJump() 385 GetMasm()->Cmp(VixlReg(src), VixlReg(rzero)); in EncodeJump() 606 GetMasm()->Cmp(VixlReg(src), vixl::aarch64::xzr); in EncodeAbs() 608 GetMasm()->Cmp(VixlReg(src), vixl::aarch64::wzr); in EncodeAbs() 631 GetMasm()->Cmp(tmp, VixlReg(dst, WORD_SIZE)); in EncodeIsInf() 640 GetMasm()->Cmp(tmp, VixlReg(dst, DOUBLE_WORD_SIZE)); in EncodeIsInf() 861 GetMasm()->Cmp(VixlReg(src), VixlImm(0)); in EncodeSignum() 1061 GetMasm()->Cmp(sreg, VixlImm(I16)); in EncodeGetTypeSize() 1063 GetMasm()->Cmp(sre in EncodeGetTypeSize() [all...] |