Home
last modified time | relevance | path

Searched refs:Offset1 (Results 1 - 22 of 22) sorted by relevance

/third_party/json/tests/thirdparty/Fuzzer/test/
H A DCustomCrossOverTest.cpp49 size_t Offset1 = 0; in LLVMFuzzerCustomCrossOver() local
50 size_t Len1 = R() % (Size1 - Offset1); in LLVMFuzzerCustomCrossOver()
58 memcpy(Out, Data1 + Offset1, Len1); in LLVMFuzzerCustomCrossOver()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/SelectionDAG/
H A DScheduleDAGSDNodes.cpp240 int64_t Offset1, Offset2; in ClusterNeighboringLoads() local
241 if (!TII->areLoadsFromSameBasePtr(Base, User, Offset1, Offset2) || in ClusterNeighboringLoads()
242 Offset1 == Offset2 || in ClusterNeighboringLoads()
248 if (O2SMap.insert(std::make_pair(Offset1, Base)).second) in ClusterNeighboringLoads()
249 Offsets.push_back(Offset1); in ClusterNeighboringLoads()
252 if (Offset2 < Offset1) in ClusterNeighboringLoads()
H A DDAGCombiner.cpp13962 APInt Offset1 = cast<ConstantSDNode>(Offset)->getAPIntValue(); in CombineToPreIndexedLoadStore() local
13973 if (X1 * Y0 * Y1 < 0) CNV = CNV + Offset1; in CombineToPreIndexedLoadStore()
13974 else CNV = CNV - Offset1; in CombineToPreIndexedLoadStore()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Hexagon/
H A DHexagonSubtarget.cpp290 int64_t Offset1; in apply() local
292 MachineOperand *BaseOp1 = HII.getBaseAndOffset(L1, Offset1, Size1); in apply()
298 if (((Offset0 ^ Offset1) & 0x18) != 0) in apply()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/
H A DX86InstrInfo.h381 bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, int64_t &Offset1,
392 bool shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2, int64_t Offset1,
H A DX86InstrInfo.cpp5747 int64_t &Offset1, int64_t &Offset2) const { in areLoadsFromSameBasePtr()
5939 Offset1 = Disp1->getSExtValue(); in areLoadsFromSameBasePtr()
5945 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear()
5947 assert(Offset2 > Offset1); in shouldScheduleLoadsNear()
5948 if ((Offset2 - Offset1) / 8 > 64) in shouldScheduleLoadsNear()
5746 areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, int64_t &Offset1, int64_t &Offset2) const areLoadsFromSameBasePtr() argument
5944 shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2, int64_t Offset1, int64_t Offset2, unsigned NumLoads) const shouldScheduleLoadsNear() argument
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Mips/
H A DMicroMipsSizeReduction.cpp400 int64_t Offset1, Offset2; in ConsecutiveInstr() local
401 if (!GetImm(MI1, 2, Offset1)) in ConsecutiveInstr()
409 return ((Offset1 == (Offset2 - 4)) && (ConsecutiveRegisters(Reg1, Reg2))); in ConsecutiveInstr()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/ARM/
H A DARMBaseInstrInfo.h245 bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, int64_t &Offset1,
257 int64_t Offset1, int64_t Offset2,
H A DARMBaseInstrInfo.cpp1836 int64_t &Offset1, in areLoadsFromSameBasePtr()
1897 Offset1 = cast<ConstantSDNode>(Load1->getOperand(1))->getSExtValue(); in areLoadsFromSameBasePtr()
1917 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear()
1922 assert(Offset2 > Offset1); in shouldScheduleLoadsNear()
1924 if ((Offset2 - Offset1) / 8 > 64) in shouldScheduleLoadsNear()
1835 areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, int64_t &Offset1, int64_t &Offset2) const areLoadsFromSameBasePtr() argument
1916 shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2, int64_t Offset1, int64_t Offset2, unsigned NumLoads) const shouldScheduleLoadsNear() argument
/third_party/node/deps/v8/src/utils/
H A Dutils.h233 #define STATIC_ASSERT_FIELD_OFFSETS_EQUAL(Offset1, Offset2) \
234 STATIC_ASSERT(static_cast<int>(Offset1) == Offset2)
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/CodeGen/
H A DTargetInstrInfo.h1219 int64_t &Offset1, in areLoadsFromSameBasePtr()
1233 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear()
1218 areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2, int64_t &Offset1, int64_t &Offset2) const areLoadsFromSameBasePtr() argument
1232 shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2, int64_t Offset1, int64_t Offset2, unsigned NumLoads) const shouldScheduleLoadsNear() argument
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/
H A DSeparateConstOffsetFromGEP.cpp1294 Value *Offset1 = First->getOperand(1); in swapGEPOperand() local
1297 Second->setOperand(1, Offset1); in swapGEPOperand()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/
H A DAMDGPUISelDAGToDAG.cpp207 SDValue &Offset1) const;
1260 SDValue &Offset1) const { in SelectDS64Bit4ByteAligned()
1273 Offset1 = CurDAG->getTargetConstant(DWordOffset1, DL, MVT::i8); in SelectDS64Bit4ByteAligned()
1308 Offset1 = CurDAG->getTargetConstant(DWordOffset1, DL, MVT::i8); in SelectDS64Bit4ByteAligned()
1325 Offset1 = CurDAG->getTargetConstant(DWordOffset1, DL, MVT::i8); in SelectDS64Bit4ByteAligned()
1334 Offset1 = CurDAG->getTargetConstant(1, DL, MVT::i8); in SelectDS64Bit4ByteAligned()
H A DSIInstrInfo.cpp148 int64_t &Offset1) const { in areLoadsFromSameBasePtr()
184 Offset1 = cast<ConstantSDNode>(Load1->getOperand(Offset1Idx))->getZExtValue(); in areLoadsFromSameBasePtr()
209 Offset1 = Load1Offset->getZExtValue(); in areLoadsFromSameBasePtr()
242 Offset1 = cast<ConstantSDNode>(Off1)->getZExtValue(); in areLoadsFromSameBasePtr()
295 uint8_t Offset1 = Offset1Imm->getImm(); in getMemOperandWithOffset() local
297 if (Offset1 > Offset0 && Offset1 - Offset0 == 1) { in getMemOperandWithOffset()
505 int64_t Offset0, int64_t Offset1, in shouldScheduleLoadsNear()
507 assert(Offset1 > Offset0 && in shouldScheduleLoadsNear()
513 return (NumLoads <= 16 && (Offset1 in shouldScheduleLoadsNear()
504 shouldScheduleLoadsNear(SDNode *Load0, SDNode *Load1, int64_t Offset0, int64_t Offset1, unsigned NumLoads) const shouldScheduleLoadsNear() argument
[all...]
H A DSIInstrInfo.h181 int64_t &Offset1,
194 int64_t Offset1, unsigned NumLoads) const override;
H A DAMDGPUInstructionSelector.cpp1081 unsigned Offset1 = WaveRelease | (WaveDone << 1) | (ShaderType << 2) |
1085 Offset1 |= (CountDw - 1) << 6;
1087 unsigned Offset = Offset0 | (Offset1 << 8);
H A DSILoadStoreOptimizer.cpp1724 uint64_t Offset1 = Src1->getImm(); in processBaseWithConstOffset()
1731 Addr.Offset = (*Offset0P & 0x00000000ffffffff) | (Offset1 << 32); in processBaseWithConstOffset()
H A DSIISelLowering.cpp6200 unsigned Offset1 = WaveRelease | (WaveDone << 1) | (ShaderType << 2) | in LowerINTRINSIC_W_CHAIN() local
6204 Offset1 |= (CountDw - 1) << 6; in LowerINTRINSIC_W_CHAIN()
6206 unsigned Offset = Offset0 | (Offset1 << 8); in LowerINTRINSIC_W_CHAIN()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/
H A DAArch64InstrInfo.cpp2338 int64_t Offset1, unsigned Opcode1, int FI2, in shouldClusterFI()
2355 ObjectOffset1 += Offset1; in shouldClusterFI()
2401 int64_t Offset1 = FirstLdSt.getOperand(2).getImm(); in shouldClusterMemOps()
2402 if (isUnscaledLdSt(FirstOpc) && !scaleOffset(FirstOpc, Offset1)) in shouldClusterMemOps()
2410 if (Offset1 > 63 || Offset1 < -64) in shouldClusterMemOps()
2416 assert((!BaseOp1.isIdenticalTo(BaseOp2) || Offset1 <= Offset2) && in shouldClusterMemOps()
2421 return shouldClusterFI(MFI, BaseOp1.getIndex(), Offset1, FirstOpc, in shouldClusterMemOps()
2425 assert(Offset1 <= Offset2 && "Caller should have ordered offsets."); in shouldClusterMemOps()
2427 return Offset1 in shouldClusterMemOps()
2337 shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2) shouldClusterFI() argument
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/
H A DMachinePipeliner.cpp695 int64_t Offset1, Offset2; in addLoopCarriedDependences() local
696 if (TII->getMemOperandWithOffset(LdMI, BaseOp1, Offset1, TRI) && in addLoopCarriedDependences()
699 (int)Offset1 < (int)Offset2) { in addLoopCarriedDependences()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/
H A DValueTracking.cpp5992 auto Offset1 = getOffsetFromIndex(GEP1, Idx, DL); in isPointerOffset() local
5994 if (!Offset1 || !Offset2) in isPointerOffset()
5996 return *Offset2 - *Offset1; in isPointerOffset()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/PowerPC/
H A DPPCISelLowering.cpp12057 int64_t Offset1 = 0, Offset2 = 0; in isConsecutiveLSLoc() local
12058 getBaseWithConstantOffset(Loc, Base1, Offset1, DAG); in isConsecutiveLSLoc()
12060 if (Base1 == Base2 && Offset1 == (Offset2 + Dist * Bytes)) in isConsecutiveLSLoc()
12066 Offset1 = 0; in isConsecutiveLSLoc()
12068 bool isGA1 = TLI.isGAPlusOffset(Loc.getNode(), GV1, Offset1); in isConsecutiveLSLoc()
12071 return Offset1 == (Offset2 + Dist*Bytes); in isConsecutiveLSLoc()

Completed in 130 milliseconds