Home
last modified time | relevance | path

Searched refs:Ld (Results 1 - 25 of 41) sorted by relevance

12

/third_party/node/deps/v8/src/builtins/mips64/
H A Dbuiltins-mips64.cc88 __ Ld(scratch2, MemOperand(scratch2)); in Generate_PushArguments()
90 __ Ld(scratch2, MemOperand(scratch2)); in Generate_PushArguments()
135 __ Ld(cp, MemOperand(fp, ConstructFrameConstants::kContextOffset)); in Generate_JSBuiltinsConstructStubHelper()
137 __ Ld(t3, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSBuiltinsConstructStubHelper()
179 __ Ld(t2, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); in Generate_JSConstructStubGeneric()
238 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kConstructorOffset)); in Generate_JSConstructStubGeneric()
239 __ Ld(a0, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSConstructStubGeneric()
289 __ Ld(v0, MemOperand(sp, 0 * kPointerSize)); in Generate_JSConstructStubGeneric()
294 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSConstructStubGeneric()
316 __ Ld(c in Generate_JSConstructStubGeneric()
[all...]
/third_party/node/deps/v8/src/regexp/riscv64/
H A Dregexp-macro-assembler-riscv64.cc149 __ Ld(a0, register_location(reg)); in AdvanceRegister()
159 __ Ld(a0, MemOperand(frame_pointer(), kBacktrackCount)); in Backtrack()
193 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckAtStart()
201 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotAtStart()
274 __ Ld(a0, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase()
275 __ Ld(a1, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase()
284 __ Ld(t1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotBackReferenceIgnoreCase()
344 __ Ld(t1, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase()
345 __ Ld(a2, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase()
392 __ Ld(end_of_input_addres in CheckNotBackReferenceIgnoreCase()
[all...]
/third_party/node/deps/v8/src/regexp/mips64/
H A Dregexp-macro-assembler-mips64.cc183 __ Ld(a0, register_location(reg)); in AdvanceRegister()
194 __ Ld(a0, MemOperand(frame_pointer(), kBacktrackCount)); in Backtrack()
231 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckAtStart()
240 __ Ld(a1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotAtStart()
265 __ Ld(a0, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase()
266 __ Ld(a1, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase()
275 __ Ld(t1, MemOperand(frame_pointer(), kStringStartMinusOne)); in CheckNotBackReferenceIgnoreCase()
335 __ Ld(t1, register_location(start_reg)); // Index of start of capture. in CheckNotBackReferenceIgnoreCase()
336 __ Ld(a2, register_location(start_reg + 1)); // Index of end of capture. in CheckNotBackReferenceIgnoreCase()
398 __ Ld(a in CheckNotBackReference()
[all...]
/third_party/node/deps/v8/src/builtins/riscv64/
H A Dbuiltins-riscv64.cc87 __ Ld(scratch2, MemOperand(scratch2)); in Generate_PushArguments()
89 __ Ld(scratch2, MemOperand(scratch2)); in Generate_PushArguments()
138 __ Ld(cp, MemOperand(fp, ConstructFrameConstants::kContextOffset)); in Generate_JSBuiltinsConstructStubHelper()
140 __ Ld(kScratchReg, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSBuiltinsConstructStubHelper()
249 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kConstructorOffset)); in Generate_JSConstructStubGeneric()
250 __ Ld(a0, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSConstructStubGeneric()
294 __ Ld(cp, MemOperand(fp, ConstructFrameConstants::kContextOffset)); in Generate_JSConstructStubGeneric()
310 __ Ld(a0, MemOperand(sp, 0 * kSystemPointerSize)); in Generate_JSConstructStubGeneric()
315 __ Ld(a1, MemOperand(fp, ConstructFrameConstants::kLengthOffset)); in Generate_JSConstructStubGeneric()
341 __ Ld(c in Generate_JSConstructStubGeneric()
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/ARM/
H A DARMParallelDSP.cpp340 if (auto *Ld = dyn_cast<LoadInst>(SExt->getOperand(0))) { in IsNarrowSequence()
342 return LoadPairs.count(Ld) || OffsetLoads.count(Ld); in IsNarrowSequence()
363 auto *Ld = dyn_cast<LoadInst>(&I); in RecordMemoryOps() local
364 if (!Ld || !Ld->isSimple() || in RecordMemoryOps()
365 !Ld->hasOneUse() || !isa<SExtInst>(Ld->user_back())) in RecordMemoryOps()
367 Loads.push_back(Ld); in RecordMemoryOps()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/
H A DLoopVersioningLICM.cpp353 LoadInst *Ld = dyn_cast<LoadInst>(I); in instructionSafeForVersioning() local
354 if (!Ld || !Ld->isSimple()) { in instructionSafeForVersioning()
359 Value *Ptr = Ld->getPointerOperand(); in instructionSafeForVersioning()
H A DLoopUnrollAndJamPass.cpp260 if (auto *Ld = dyn_cast<LoadInst>(&I)) { in computeUnrollAndJamCount()
261 Value *V = Ld->getPointerOperand(); in computeUnrollAndJamCount()
/third_party/node/deps/v8/src/baseline/mips64/
H A Dbaseline-assembler-mips64-inl.h169 __ Ld(type, FieldMemOperand(map, Map::kInstanceTypeOffset)); in JumpIfInstanceType()
177 __ Ld(scratch, operand); in JumpIfPointer()
199 __ Ld(scratch, operand); in JumpIfTagged()
207 __ Ld(scratch, operand); in JumpIfTagged()
352 __ Ld(output, FieldMemOperand(source, offset)); in LoadTaggedPointerField()
356 __ Ld(output, FieldMemOperand(source, offset)); in LoadTaggedSignedField()
360 __ Ld(output, FieldMemOperand(source, offset)); in LoadTaggedAnyField()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/
H A DGCRootLowering.cpp217 Value *Ld = new LoadInst(CI->getType(), CI->getArgOperand(1), "", CI); in DoLowering() local
218 Ld->takeName(CI); in DoLowering()
219 CI->replaceAllUsesWith(Ld); in DoLowering()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/SystemZ/
H A DSystemZTargetTransformInfo.cpp866 if (LoadInst *Ld = dyn_cast<LoadInst>(I->getOperand(0))) in getCmpSelInstrCost()
868 if (!Ld->hasOneUse() && Ld->getParent() == I->getParent() && in getCmpSelInstrCost()
909 isFoldableLoad(const LoadInst *Ld, const Instruction *&FoldedValue) { in isFoldableLoad() argument
910 if (!Ld->hasOneUse()) in isFoldableLoad()
912 FoldedValue = Ld; in isFoldableLoad()
913 const Instruction *UserI = cast<Instruction>(*Ld->user_begin()); in isFoldableLoad()
914 unsigned LoadedBits = getScalarSizeInBits(Ld->getType()); in isFoldableLoad()
H A DSystemZTargetTransformInfo.h90 bool isFoldableLoad(const LoadInst *Ld, const Instruction *&FoldedValue);
/third_party/node/deps/v8/src/codegen/mips64/
H A Dmacro-assembler-mips64.h224 inline void Move(Register output, MemOperand operand) { Ld(output, operand); } in Move()
297 void Ld(Register rd, const MemOperand& rs);
395 Ld(dst, MemOperand(sp, 0)); in pop()
403 Ld(src2, MemOperand(sp, 0 * kPointerSize)); in Pop()
404 Ld(src1, MemOperand(sp, 1 * kPointerSize)); in Pop()
410 Ld(src3, MemOperand(sp, 0 * kPointerSize)); in Pop()
411 Ld(src2, MemOperand(sp, 1 * kPointerSize)); in Pop()
412 Ld(src1, MemOperand(sp, 2 * kPointerSize)); in Pop()
973 Ld(dest, MemOperand(sp, 0)); in LoadReceiver()
1265 Ld(scratc in GenerateSwitchTable()
[all...]
H A Dmacro-assembler-mips64.cc100 Ld(destination, MemOperand(s6, RootRegisterOffsetForRootIndex(index))); in LoadRoot()
107 Ld(destination, MemOperand(s6, RootRegisterOffsetForRootIndex(index))); in LoadRoot()
283 Ld(scratch, MemOperand(address)); in CallRecordWriteStub()
1277 Ld(rd, rs); in CallRecordWriteStub()
1432 void TurboAssembler::Ld(Register rd, const MemOperand& rs) { in CallRecordWriteStub() function in v8::internal::TurboAssembler
1937 Ld(ToRegister(i), MemOperand(sp, stack_offset)); in CallRecordWriteStub()
2632 Ld(scratch, src); in CallRecordWriteStub()
2730 Ld(scratch, src); in CallRecordWriteStub()
3349 Ld(result, MemOperand(sp, 0)); in CallRecordWriteStub()
4231 Ld(destinatio in CallRecordWriteStub()
[all...]
/third_party/node/deps/v8/src/codegen/riscv64/
H A Dmacro-assembler-riscv64.h202 inline void Move(Register output, MemOperand operand) { Ld(output, operand); } in Move()
286 void Ld(Register rd, const MemOperand& rs);
382 Ld(dst, MemOperand(sp, 0)); in pop()
390 Ld(src2, MemOperand(sp, 0 * kSystemPointerSize)); in Pop()
391 Ld(src1, MemOperand(sp, 1 * kSystemPointerSize)); in Pop()
397 Ld(src3, MemOperand(sp, 0 * kSystemPointerSize)); in Pop()
398 Ld(src2, MemOperand(sp, 1 * kSystemPointerSize)); in Pop()
399 Ld(src1, MemOperand(sp, 2 * kSystemPointerSize)); in Pop()
1049 Ld(dest, MemOperand(sp, 0)); in LoadReceiver()
H A Dmacro-assembler-riscv64.cc97 Ld(destination, in LoadRoot()
106 Ld(destination, in LoadRoot()
1426 void TurboAssembler::Ld(Register rd, const MemOperand& rs) { in Ld() function in v8::internal::TurboAssembler
1621 // Ld a Address from a constant pool. in li()
1709 Ld(reg, MemOperand(sp, stack_offset)); \ in MultiPop()
3189 Ld(destination, MemOperand(kRootRegister, offset)); in LoadRootRelative()
3260 Ld(t6, MemOperand(kRootRegister, offset)); in Jump()
3367 Ld(builtin, MemOperand(builtin, IsolateData::builtin_entry_table_offset())); in LoadEntryFromBuiltinIndex()
3397 Ld(destination, EntryFromBuiltinAsOperand(builtin)); in LoadEntryFromBuiltin()
3410 Ld(t in PatchAndJump()
[all...]
/third_party/node/deps/v8/src/baseline/riscv64/
H A Dbaseline-assembler-riscv64-inl.h167 __ Ld(type, FieldMemOperand(map, Map::kInstanceTypeOffset)); in JumpIfInstanceType()
175 __ Ld(temp, operand); in JumpIfPointer()
199 __ Ld(scratch, operand); in JumpIfTagged()
208 __ Ld(scratch, operand); in JumpIfTagged()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/
H A DLoopAccessAnalysis.cpp1851 auto *Ld = dyn_cast<LoadInst>(&I); in analyzeLoop() local
1852 if (!Ld) { in analyzeLoop()
1853 recordAnalysis("CantVectorizeInstruction", Ld) in analyzeLoop()
1858 if (!Ld->isSimple() && !IsAnnotatedParallel) { in analyzeLoop()
1859 recordAnalysis("NonSimpleLoad", Ld) in analyzeLoop()
1866 Loads.push_back(Ld); in analyzeLoop()
1867 DepChecker->addAccess(Ld); in analyzeLoop()
1869 collectStridedAccess(Ld); in analyzeLoop()
/third_party/node/deps/v8/src/wasm/baseline/mips64/
H A Dliftoff-assembler-mips64.h109 assm->Ld(dst.gp(), src); in Load()
314 Ld(scratch, MemOperand(fp, 8)); in PrepareTailCall()
316 Ld(scratch, MemOperand(fp, 0)); in PrepareTailCall()
322 Ld(scratch, MemOperand(sp, i * 8)); in PrepareTailCall()
379 Ld(stack_limit, in PatchPrepareStackFrame()
382 Ld(stack_limit, MemOperand(stack_limit)); in PatchPrepareStackFrame()
449 Ld(dst, liftoff::GetInstanceOperand()); in LoadInstanceFromFrame()
463 Ld(dst, MemOperand(instance, offset)); in LoadFromInstance()
474 Ld(dst, MemOperand(instance, offset)); in LoadTaggedPointerFromInstance()
489 Ld(ds in LoadTaggedPointer()
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AArch64/
H A DAArch64ISelDAGToDAG.cpp1290 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoad() local
1291 SDValue SuperReg = SDValue(Ld, 0); in SelectLoad()
1296 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 1)); in SelectLoad()
1300 CurDAG->setNodeMemRefs(cast<MachineSDNode>(Ld), {MemOp}); in SelectLoad()
1318 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectPostLoad() local
1321 ReplaceUses(SDValue(N, NumVecs), SDValue(Ld, 0)); in SelectPostLoad()
1324 SDValue SuperReg = SDValue(Ld, 1); in SelectPostLoad()
1333 ReplaceUses(SDValue(N, NumVecs + 1), SDValue(Ld, 2)); in SelectPostLoad()
1435 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); in SelectLoadLane() local
1436 SDValue SuperReg = SDValue(Ld, in SelectLoadLane()
1479 SDNode *Ld = CurDAG->getMachineNode(Opc, dl, ResTys, Ops); SelectPostLoadLane() local
3149 SDNode *Ld = CurDAG->getMachineNode(Op, DL, MVT::i64, MVT::i64, Select() local
[all...]
/third_party/node/deps/v8/src/compiler/backend/mips64/
H A Dcode-generator-mips64.cc507 __ Ld(ra, MemOperand(fp, StandardFrameConstants::kCallerPCOffset)); in AssemblePrepareTailCall()
508 __ Ld(fp, MemOperand(fp, StandardFrameConstants::kCallerFPOffset)); in AssemblePrepareTailCall()
561 __ Ld(kScratchReg, MemOperand(kJavaScriptCallCodeStartRegister, offset)); in BailoutIfDeoptimized()
659 __ Ld(kScratchReg, FieldMemOperand(func, JSFunction::kContextOffset)); in AssembleArchInstruction()
664 __ Ld(a2, FieldMemOperand(func, JSFunction::kCodeOffset)); in AssembleArchInstruction()
823 __ Ld(i.OutputRegister(), MemOperand(fp, 0)); in AssembleArchInstruction()
1648 __ Ld(i.OutputRegister(), i.MemoryOperand()); in AssembleArchInstruction()
1746 __ Ld(i.OutputRegister(0), MemOperand(fp, offset)); in AssembleArchInstruction()
1787 __ Ld(kScratchReg, i.MemoryOperand()); in AssembleArchInstruction()
1797 __ Ld(kScratchRe in AssembleArchInstruction()
[all...]
/third_party/node/deps/v8/src/wasm/baseline/riscv64/
H A Dliftoff-assembler-riscv64.h108 assm->Ld(dst.gp(), src); in Load()
299 Ld(scratch, MemOperand(fp, 8)); in PrepareTailCall()
301 Ld(scratch, MemOperand(fp, 0)); in PrepareTailCall()
307 Ld(scratch, MemOperand(sp, i * 8)); in PrepareTailCall()
363 Ld(stack_limit, in PatchPrepareStackFrame()
366 Ld(stack_limit, MemOperand(stack_limit)); in PatchPrepareStackFrame()
441 Ld(dst, liftoff::GetInstanceOperand()); in LoadInstanceFromFrame()
456 Ld(dst, MemOperand(src)); in LoadFromInstance()
487 Ld(dst, src_op); in LoadFullPointer()
552 TurboAssembler::Ld(ds in Load()
[all...]
/third_party/node/deps/v8/src/compiler/backend/riscv64/
H A Dcode-generator-riscv64.cc557 __ Ld(ra, MemOperand(fp, StandardFrameConstants::kCallerPCOffset)); in AssemblePrepareTailCall()
558 __ Ld(fp, MemOperand(fp, StandardFrameConstants::kCallerFPOffset)); in AssemblePrepareTailCall()
854 __ Ld(i.OutputRegister(), MemOperand(fp, 0)); in AssembleArchInstruction()
1617 __ Ld(i.OutputRegister(), i.MemoryOperand()); in AssembleArchInstruction()
1713 __ Ld(i.OutputRegister(0), MemOperand(fp, offset)); in AssembleArchInstruction()
1768 ASSEMBLE_ATOMIC_LOAD_INTEGER(Ld); in AssembleArchInstruction()
1980 __ Ld(kScratchReg, i.MemoryOperand()); in AssembleArchInstruction()
2000 __ Ld(kScratchReg, i.MemoryOperand()); in AssembleArchInstruction()
2008 __ Ld(kScratchReg, i.MemoryOperand()); in AssembleArchInstruction()
2027 __ Ld(kScratchRe in AssembleArchInstruction()
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/
H A DX86ISelLowering.cpp4417 } else if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(Arg)) {
4425 SDValue Ptr = Ld->getBasePtr();
8011 static bool findEltLoadSrc(SDValue Elt, LoadSDNode *&Ld, int64_t &ByteOffset) {
8016 Ld = BaseLd;
8025 return findEltLoadSrc(Elt.getOperand(0), Ld, ByteOffset);
8029 if ((Idx % 8) == 0 && findEltLoadSrc(Elt.getOperand(0), Ld, ByteOffset)) {
8041 findEltLoadSrc(Src, Ld, ByteOffset)) {
8138 LoadSDNode *Ld = Loads[EltIdx];
8143 Loads[BaseIdx] == Ld && ByteOffsets[BaseIdx] == 0);
8145 return DAG.areNonVolatileConsecutiveLoads(Ld, Bas
[all...]
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Utils/
H A DLoopUnrollAndJam.cpp596 if (auto *Ld = dyn_cast<LoadInst>(&I)) { in getLoadsAndStores()
597 if (!Ld->isSimple()) in getLoadsAndStores()
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Sparc/
H A DSparcISelLowering.cpp3388 LoadSDNode *Ld = cast<LoadSDNode>(N); in ReplaceNodeResults() local
3391 if (Ld->getValueType(0) != MVT::i64 || Ld->getMemoryVT() != MVT::i64) in ReplaceNodeResults()
3396 Ld->getExtensionType(), dl, MVT::v2i32, Ld->getChain(), in ReplaceNodeResults()
3397 Ld->getBasePtr(), Ld->getPointerInfo(), MVT::v2i32, Ld->getAlignment(), in ReplaceNodeResults()
3398 Ld->getMemOperand()->getFlags(), Ld in ReplaceNodeResults()
[all...]

Completed in 71 milliseconds

12