/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/src/cg/aarch64/ |
H A D | aarch64_cfi_generator.cpp | 30 curInsn = bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiOffsetInsn(stackBaseReg, -cfiOffset, k64BitSize)); in GenerateRegisterSaveDirective() 34 *curInsn, aarchCGFunc.CreateCfiOffsetInsn(RLR, static_cast<int64>(-cfiOffset + RLROffset), k64BitSize)); in GenerateRegisterSaveDirective() 44 k64BitSize)); in GenerateRegisterSaveDirective() 49 .AddOpndChain(cgFunc.CreateCfiRegOperand(stackBaseReg, k64BitSize))); in GenerateRegisterSaveDirective() 76 curInsn = bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiOffsetInsn(reg, -cfiOffset, k64BitSize)); in GenerateRegisterSaveDirective() 101 curInsn = bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiRestoreInsn(reg, k64BitSize)); in GenerateRegisterRestoreDirective() 106 curInsn = bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiRestoreInsn(stackBaseReg, k64BitSize)); in GenerateRegisterRestoreDirective() 108 curInsn = bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiRestoreInsn(RLR, k64BitSize)); in GenerateRegisterRestoreDirective() 111 bb.InsertInsnAfter(*curInsn, aarchCGFunc.CreateCfiDefCfaInsn(R31, 0, k64BitSize)); in GenerateRegisterRestoreDirective()
|
H A D | aarch64_live.cpp | 26 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(static_cast<AArch64reg>(V0), k64BitSize, kRegTyFloat); in GenerateReturnBBDefUse() 30 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(static_cast<AArch64reg>(R0), k64BitSize, kRegTyInt); in GenerateReturnBBDefUse() 52 static_cast<AArch64reg>(preg), k64BitSize, in ProcessCallInsnParam() 61 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(static_cast<AArch64reg>(R0 + i), k64BitSize, kRegTyInt); in ProcessCallInsnParam() 64 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(static_cast<AArch64reg>(V0 + i), k64BitSize, kRegTyFloat); in ProcessCallInsnParam()
|
H A D | aarch64_isa.cpp | 213 for (; bitNum < k64BitSize; bitNum++) { in GetTail0BitNum() 218 if (bitNum == k64BitSize) { in GetTail0BitNum() 229 for (; bitNum < k64BitSize; bitNum++) { in GetHead0BitNum() 234 if (bitNum == k64BitSize) { in GetHead0BitNum()
|
H A D | aarch64_cgfunc.cpp | 192 DEBUG_ASSERT(((dsize == k8BitSize) || (dsize == k16BitSize) || (dsize == k32BitSize) || (dsize == k64BitSize)), in SelectCopyImm() 194 if (src.GetSize() == k32BitSize && dsize == k64BitSize && src.IsSingleInstructionMovable()) { in SelectCopyImm() 315 regTy = dsize == k64BitSize ? dtype : PTY_i32; in SelectCopyMemOpnd() 383 bool is64Bits = (dest.GetSize() == k64BitSize) ? true : false; in SelectCopyRegOpnd() 651 ImmOperand &immAddend = CreateImmOperand(addend, k64BitSize, true); in SplitAndGetRemained() 857 Operand &dest = GetZeroOpnd(k64BitSize); in SelectDread() 858 uint32 numLoads = static_cast<uint32>(RoundUp(aggSize, k64BitSize) / k64BitSize); in SelectDread() 1029 bool is64Bits = (GetPrimTypeBitSize(stype) == k64BitSize); in HandleFmovImm() 1082 for (; bitNum < k64BitSize; bitNu in GetHead0BitNum() [all...] |
H A D | aarch64_proepilog.cpp | 79 ImmOperand &immAddEnd = aarchCGFunc.CreateImmOperand(offsetVal, k64BitSize, true); in SplitStpLdpOffsetForCalleeSavedWithAddInstruction() 160 ImmOperand &io1 = aarchCGFunc.CreateImmOperand(fpToSpDistance, k64BitSize, true); in AppendInstructionForAllocateOrDeallocateCallFrame() 169 ImmOperand &io2 = aarchCGFunc.CreateImmOperand(size, k64BitSize, true); in AppendInstructionForAllocateOrDeallocateCallFrame() 231 Operand &spOpnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(RSP, k64BitSize, kRegTyInt); in AppendInstructionAllocateCallFrame() 262 Operand &spOpnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(RSP, k64BitSize, kRegTyInt); in AppendInstructionAllocateCallFrame() 287 Operand &spOpnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(RSP, k64BitSize, kRegTyInt); in AppendInstructionAllocateCallFrameDebug() 299 RegOperand &spOpnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(RSP, k64BitSize, kRegTyInt); in AppendInstructionAllocateCallFrameDebug() 339 auto &x10Opnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(R10, k64BitSize, kRegTyInt); in GenerateFrameTypeSave() 340 auto immOpnd = &aarchCGFunc.CreateImmOperand(frameTypeInfo.idx, k64BitSize, true); in GenerateFrameTypeSave() 365 auto &x10Opnd = aarchCGFunc.GetOrCreatePhysicalRegisterOperand(R10, k64BitSize, kRegTyIn in GenerateFunctionSave() [all...] |
H A D | aarch64_offset_adjust.cpp | 89 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(stackBaseReg, k64BitSize, kRegTyInt)); in AdjustmentOffsetForOpnd() 120 aarchCGFunc->GetOrCreatePhysicalRegisterOperand(stackBaseReg, k64BitSize, kRegTyInt); in AdjustMemBaseReg() 215 RegOperand &tempReg = aarchCGFunc->GetOrCreatePhysicalRegisterOperand(R16, k64BitSize, kRegTyInt); in AdjustmentStackPointer() 216 ImmOperand &offsetReg = aarchCGFunc->CreateImmOperand(offset, k64BitSize, false); in AdjustmentStackPointer()
|
H A D | aarch64_fixshortbranch.cpp | 165 R16, (ubfxOp == MOP_wubfxrri5i5) ? k32BitSize : k64BitSize, kRegTyInt); in FixShortBranches() 232 Operand &immOpnd = aarch64CGFunc->CreateImmOperand(labelOpnd.GetLabelIndex(), k64BitSize, false); in FixLdr() 237 k64BitSize, regOpnd, aarch64CGFunc->CreateImmOperand(0, k32BitSize, false), false); in FixLdr()
|
H A D | aarch64_insn.cpp | 30 DEBUG_ASSERT((opndSize == k32BitSize || opndSize == k64BitSize), "illegal register size"); in EmitIntReg() 53 opndSize == k64BitSize || opndSize == k128BitSize), in Visit() 80 .Emit((v->GetSize() == k64BitSize) ? value : static_cast<int64>(static_cast<int32>(value))); in Visit() 133 if (baseSize != k64BitSize) { in Visit() 134 baseReg->SetSize(k64BitSize); in Visit() 189 baseReg->SetSize(k64BitSize); in Visit() 376 .Emit((v->GetSize() == k64BitSize) ? value : static_cast<int64>(static_cast<int32>(value))); in Visit()
|
H A D | aarch64_obj_emitter.cpp | 639 DEBUG_ASSERT(immSize == k64BitSize, "support 64 bit only!"); 656 DEBUG_ASSERT(immSize == k64BitSize, "support 64 bit only!"); 661 uint32 size = (immSize == k32BitSize) ? k32BitSize : k64BitSize; 752 opnd |= ((regSize == k64BitSize ? 0b11 : 0b10) << kShiftThirteen); // option 1126 (regSize == k64BitSize && shift == k3BitSize) || (regSize == k128BitSize && shift == k4BitSize), 1182 ((size == k64BitSize) && (offsetValue % k8BitSize) != 0)) && 1310 if (insn.GetDesc()->GetOpndDes(kInsnThirdOpnd)->GetSize() == k64BitSize) { 1347 if (insn.GetDesc()->GetOpndDes(kInsnThirdOpnd)->GetSize() == k64BitSize) { 1389 if (insn.GetDesc()->GetOpndDes(kInsnThirdOpnd)->GetSize() == k64BitSize) { 1426 if (insn.GetDesc()->GetOpndDes(kInsnThirdOpnd)->GetSize() == k64BitSize) { [all...] |
H A D | aarch64_peep.cpp | 380 auto &newImm = static_cast<AArch64CGFunc &>(cgFunc).GetOrCreateOfstOpnd(val, k64BitSize); in GetInsnAddOrSubNewOffset() 433 (regSize == k64BitSize && shift == k3BitSize)) { in OptimizeAddrBOrXShiftExtend() 724 static_cast<AArch64reg>(curDestRegNo), k64BitSize, curDestOpnd.GetRegisterType()); in Run() 726 static_cast<AArch64reg>(prevDestRegNo), k64BitSize, prevDestOpnd.GetRegisterType()); in Run() 928 static_cast<AArch64reg>(dstMovOpnd.GetRegisterNumber()), k64BitSize, in Run() 1078 doOpt = k64BitSize; in Run() 1551 MOperator mopCode = (reg.GetSize() == k64BitSize) ? MOP_xcsetrc : MOP_wcsetrc; in Run() 1751 MOperator cmpOpCode = (cmpThirdOpnd.GetSize() == k64BitSize) ? MOP_xcmpri : MOP_wcmpri; in Run() 2049 (regOpnd->GetSize() <= k32BitSize) ? (k32BitSize - 1) : (k64BitSize - 1), k8BitSize, false); in Run() 2126 (defMop == MOP_xlsrrri6 && shiftBits == (k64BitSize in OpndDefByOneValidBit() [all...] |
H A D | aarch64_args.cpp | 172 bool is64Bit = (srcOpnd.GetSize() == maplebe::k64BitSize); in CreateMoveArgsToVRegInsn() 184 destOpnd.SetValidBitsNum(maplebe::k64BitSize); in CreateMoveArgsToVRegInsn()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/src/cg/x86_64/ |
H A D | x64_proepilog.cpp | 39 RegOperand &baseReg = cgFunc.GetOpndBuilder()->CreatePReg(x64::RBP, k64BitSize, kRegTyInt); in GenerateCalleeSavedRegs() 45 DEBUG_ASSERT((regSize == k32BitSize || regSize == k64BitSize), "only supported 32/64-bits"); in GenerateCalleeSavedRegs() 92 RegOperand &opndFpReg = cgFunc.GetOpndBuilder()->CreatePReg(x64::RBP, k64BitSize, kRegTyInt); in GenerateProlog() 99 RegOperand &opndSpReg = cgFunc.GetOpndBuilder()->CreatePReg(x64::RSP, k64BitSize, kRegTyInt); in GenerateProlog() 145 RegOperand &opndFpReg = cgFunc.GetOpndBuilder()->CreatePReg(x64::RBP, k64BitSize, kRegTyInt); in GenerateEpilog()
|
H A D | x64_standardize.cpp | 86 srcSize = k64BitSize; in StdzCvtOp() 89 destSize = k64BitSize; in StdzCvtOp() 126 uint32 bitSize = mOp == abstract::MOP_neg_f_32 ? k32BitSize : k64BitSize; in StdzFloatingNeg() 141 Operand *tmpOperand1 = &cgFunc.GetOpndBuilder()->CreateVReg(k64BitSize, kRegTyInt); in StdzFloatingNeg()
|
H A D | x64_reg_info.cpp | 87 case k64BitSize: in BuildStrInsn() 113 case k64BitSize: in BuildLdrInsn()
|
H A D | x64_MPIsel.cpp | 47 result->SetOffsetOperand(GetCurFunc()->GetOpndBuilder()->CreateImm(k64BitSize, stOfst + offset)); in GetOrCreateMemOpndFromSymbol() 53 stackBaseReg = &GetCurFunc()->GetOpndBuilder()->CreatePReg(x64::RIP, k64BitSize, kRegTyInt); in GetOrCreateMemOpndFromSymbol() 162 cgFunc->GetOpndBuilder()->CreatePReg(x64::RSP, k64BitSize, cgFunc->GetRegTyFromPrimTy(primType)); in SelectParmList() 183 return cgFunc->GetOpndBuilder()->CreatePReg(x64::RFP, k64BitSize, cgFunc->GetRegTyFromPrimTy(primType)); in SelectSpecialRegread() 186 return cgFunc->GetOpndBuilder()->CreatePReg(x64::RSP, k64BitSize, cgFunc->GetRegTyFromPrimTy(primType)); in SelectSpecialRegread() 226 retOpnds.PushOpnd(cgFunc->GetOpndBuilder()->CreatePReg(x64::RAX, k64BitSize, kRegTyInt)); in SelectCalleeReturn() 228 if (retSize > k64BitSize) { in SelectCalleeReturn() 229 retOpnds.PushOpnd(cgFunc->GetOpndBuilder()->CreatePReg(x64::RDX, k64BitSize, kRegTyInt)); in SelectCalleeReturn() 245 ImmOperand &fpNumImm = cgFunc->GetOpndBuilder()->CreateImm(k64BitSize, fpNum); in SelectCall() 246 RegOperand &raxOpnd = cgFunc->GetOpndBuilder()->CreatePReg(x64::RAX, k64BitSize, kRegTyIn in SelectCall() [all...] |
H A D | x64_args.cpp | 131 if (opndSize == k64BitSize) { in GenerateMovInsn() 197 if (opndSize == k64BitSize) { in LoadStackArgsToVReg() 233 if (srcBitSize == k64BitSize) { in MoveArgsToVReg()
|
/arkcompiler/ets_runtime/ecmascript/compiler/assembler/aarch64/ |
H A D | macro_assembler_aarch64.cpp | 26 constexpr uint32_t k64BitSize = 64; member 168 CopyImm(paramReg, num, k64BitSize); in MovParameterIntoParamReg() 216 if (bitLen == k64BitSize) { in IsMoveWidableImmediate() 327 if (size != k32BitSize && size != k64BitSize) { in CopyImm() 397 uint32_t immr = -k32BitSize % k64BitSize; // immr = -shift % size in CopyImmSize64()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/src/cg/ |
H A D | standardize.cpp | 77 case k64BitSize: in AddressMapping()
|
H A D | cfi_generator.cpp | 50 .AddOpndChain(cgFunc.CreateCfiImmOperand(cfiOffset, k64BitSize)); in InsertCFIDefCfaOffset()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/test/maple_ir/ |
H A D | mir_type_test.cpp | 42 size_t input_ls[] = {k1BitSize, k8BitSize, k16BitSize, k32BitSize, k64BitSize, static_cast<size_t>(2048U)}; in TEST()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/include/cg/aarch64/ |
H A D | aarch64_mop_split.h | 32 resOpnd = &opndBuilder->CreateVReg((is64Bits ? k64BitSize : k32BitSize), kRegTyInt); in GetSplitBaseReg() 34 resOpnd = &opndBuilder->CreatePReg(R16, (is64Bits ? k64BitSize : k32BitSize), kRegTyInt); in GetSplitBaseReg() 78 ImmOperand &immAddend = opndBuilder->CreateImm(k64BitSize, addend, true); in SplitGetRemained() 126 BitShiftOperand &shiftopnd = opndBuilder->CreateBitShift(BitShiftOperand::kLSL, k12BitSize, k64BitSize); in AddInsnSplit() 233 BitShiftOperand &shiftopnd = opndBuilder->CreateBitShift(BitShiftOperand::kLSL, k12BitSize, k64BitSize); in SubInsnSplit() 335 uint32 size = is64Bits ? k64BitSize : k32BitSize; in AddSubWithLslSplit() 435 ImmOperand &immOpnd = opndBuilder->CreateImm(k64BitSize, immVal, true); in MOP_wmovri32Split()
|
H A D | aarch64_imm_valid.h | 40 if (bitLen == k64BitSize) { in IsMoveWidableImmediateCopy() 117 return IsBitmaskImmediate(static_cast<uint64>(value), k64BitSize); in Imm13BitMaskValid()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/include/cg/x86_64/ |
H A D | x64_isa_tbl.h | 51 case k64BitSize: in GetCmpMop()
|
/arkcompiler/ets_runtime/ecmascript/compiler/codegen/maple/maple_be/include/cg/ |
H A D | isa.h | 444 return GetOperandSize() == k64BitSize; in Is64Bit() 525 case k64BitSize:
|
H A D | datainfo.h | 193 if (index == k64BitSize) {
|