/third_party/vk-gl-cts/framework/delibs/decpp/ |
H A D | deArrayBuffer.hpp | 47 template <typename T, size_t Alignment = (sizeof(T) > 4 ? 4 : sizeof(T)), size_t Stride = sizeof(T)> member 51 DE_STATIC_ASSERT(Stride >= sizeof(T)); 76 template <typename T, size_t Alignment, size_t Stride> 77 ArrayBuffer<T,Alignment,Stride>::ArrayBuffer (void) throw() in ArrayBuffer() 83 template <typename T, size_t Alignment, size_t Stride> 84 ArrayBuffer<T,Alignment,Stride>::ArrayBuffer (size_t numElements) in ArrayBuffer() 90 // \note no need to allocate stride for the last element, sizeof(T) is enough. Also handles cases where sizeof(T) > Stride in ArrayBuffer() 91 const size_t storageSize = (numElements - 1) * Stride + sizeof(T); in ArrayBuffer() 102 template <typename T, size_t Alignment, size_t Stride> 103 ArrayBuffer<T,Alignment,Stride> [all...] |
/third_party/vk-gl-cts/framework/randomshaders/ |
H A D | rsgVariableValue.hpp | 73 template <int Stride> 87 template <int Stride> 97 ConstStridedValueAccess component (int compNdx) const { return ConstStridedValueAccess(getType().getElementType(), m_value + Stride*compNdx); } in component() 98 ConstStridedValueAccess arrayElement (int elementNdx) const { return ConstStridedValueAccess(getType().getElementType(), m_value + Stride*getType().getElementScalarOffset(elementNdx)); } in arrayElement() 99 ConstStridedValueAccess member (int memberNdx) const { return ConstStridedValueAccess(getType().getMembers()[memberNdx].getType(), m_value + Stride*getType().getMemberScalarOffset(memberNdx)); } in member() 101 float asFloat (void) const { DE_STATIC_ASSERT(Stride == 1); return m_value->floatVal; } in asFloat() 102 int asInt (void) const { DE_STATIC_ASSERT(Stride == 1); return m_value->intVal; } in asInt() 103 bool asBool (void) const { DE_STATIC_ASSERT(Stride == 1); return m_value->boolVal; } in asBool() 104 Scalar asScalar (void) const { DE_STATIC_ASSERT(Stride == 1); return *m_value; } in asScalar() 106 float asFloat (int ndx) const { DE_ASSERT(de::inBounds(ndx, 0, Stride)); retur [all...] |
H A D | rsgProgramExecutor.cpp | 137 template <int Stride> 138 void interpolateVertexInput (StridedValueAccess<Stride> dst, int dstComp, const ConstValueRangeAccess valueRange, float x, float y) in interpolateVertexInput() 150 template <int Stride> 151 void interpolateFragmentInput (StridedValueAccess<Stride> dst, int dstComp, ConstValueAccess vtx0, ConstValueAccess vtx1, ConstValueAccess vtx2, ConstValueAccess vtx3, float x, float y) in interpolateFragmentInput() 159 template <int Stride> 160 void copyVarying (ValueAccess dst, ConstStridedValueAccess<Stride> src, int compNdx) in copyVarying()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Analysis/ |
H A D | VectorUtils.h | 342 /// are incremented by \p Stride. The mask can be used to deinterleave an 346 /// <Start, Start + Stride, ..., Start + Stride * (VF - 1)> 348 /// For example, the mask for Start = 0, Stride = 2, and VF = 4 is: 352 unsigned Stride, unsigned VF); 423 InterleaveGroup(InstTy *Instr, int32_t Stride, Align Alignment) in InterleaveGroup() argument 425 Factor = std::abs(Stride); in InterleaveGroup() 428 Reverse = Stride < 0; in InterleaveGroup() 646 StrideDescriptor(int64_t Stride, const SCEV *Scev, uint64_t Size, in StrideDescriptor() 648 : Stride(Strid in StrideDescriptor() 651 int64_t Stride = 0; global() member 671 createInterleaveGroup(Instruction *Instr, int Stride, Align Alignment) createInterleaveGroup() argument [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Object/ |
H A D | Minidump.h | 110 MemoryInfoIterator(ArrayRef<uint8_t> Storage, size_t Stride) in MemoryInfoIterator() argument 111 : Storage(Storage), Stride(Stride) { in MemoryInfoIterator() 112 assert(Storage.size() % Stride == 0); in MemoryInfoIterator() 125 Storage = Storage.drop_front(Stride); in operator ++() 131 size_t Stride; member in llvm::object::MinidumpFile::MemoryInfoIterator
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Scalar/ |
H A D | StraightLineStrengthReduce.cpp | 113 : CandidateKind(CT), Base(B), Index(Idx), Stride(S), Ins(I) {} in Candidate() 119 // Note that Index and Stride of a GEP candidate do not necessarily have the 120 // same integer type. In that case, during rewriting, Stride will be 124 Value *Stride = nullptr; member 135 // <Base: a, Index: 1, Stride: b + 2> 139 // <Base: b, Index: 2, Stride: a + 1> 269 Basis.Base == C.Base && Basis.Stride == C.Stride && in isBasisFor() 282 // Returns whether (Base + Index * Stride) can be folded to an addressing mode. 283 static bool isAddFoldable(const SCEV *Base, ConstantInt *Index, Value *Stride, in isAddFoldable() argument [all...] |
H A D | LowerMatrixIntrinsics.cpp | 52 // assuming \p Stride elements between start two consecutive columns. 53 // \p Stride must be >= \p NumRows. 85 Value *computeColumnAddr(Value *BasePtr, Value *Col, Value *Stride, in computeColumnAddr() argument 89 assert((!isa<ConstantInt>(Stride) || in computeColumnAddr() 90 cast<ConstantInt>(Stride)->getZExtValue() >= NumRows) && in computeColumnAddr() 91 "Stride must be >= the number of rows."); in computeColumnAddr() 94 // Compute the start of the column with index Col as Col * Stride. in computeColumnAddr() 95 Value *ColumnStart = Builder.CreateMul(Col, Stride, "col.start"); in computeColumnAddr() 560 void LowerLoad(Instruction *Inst, Value *Ptr, Value *Stride, in LowerLoad() argument 569 computeColumnAddr(EltPtr, Builder.getInt32(C), Stride, Shap in LowerLoad() 583 Value *Stride = Inst->getArgOperand(1); LowerColumnwiseLoad() local 588 LowerStore(Instruction *Inst, Value *Matrix, Value *Ptr, Value *Stride, ShapeInfo Shape) LowerStore() argument 610 Value *Stride = Inst->getArgOperand(2); LowerColumnwiseStore() local [all...] |
H A D | LoopIdiomRecognize.cpp | 481 APInt Stride = getStoreStride(StoreEv); in isLegalStore() local 483 if (StoreSize != Stride && StoreSize != -Stride) in isLegalStore() 725 APInt Stride = getStoreStride(StoreEv); in processLoopStores() local 729 if (StoreSize != Stride && StoreSize != -Stride) in processLoopStores() 732 bool NegStride = StoreSize == -Stride; in processLoopStores() 777 APInt Stride = ConstStride->getAPInt(); in processLoopMemSet() 778 if (SizeInBytes != Stride && SizeInBytes != -Stride) in processLoopMemSet() 1002 APInt Stride = getStoreStride(StoreEv); processLoopStoreOfLoopLoad() local [all...] |
/third_party/skia/experimental/graphite/src/ |
H A D | UniformManager.cpp | 72 static constexpr size_t Stride(int count) { in Stride() function 79 return Rules140<BaseType, RowsOrVecLength>::Stride(1); in Stride() 82 // Stride doesn't matter for a non-array. in Stride() 107 static constexpr size_t Stride(int count) { in Stride() function 115 return Rules430<BaseType, RowsOrVecLength>::Stride(1); in Stride() 118 // Stride doesn't matter for a non-array. in Stride() 131 static constexpr size_t Stride(int count) { in Stride() function 138 return RulesMetal<BaseType, RowsOrVecLength>::Stride(1); in Stride() 141 // Stride doesn't matter for a non-array. in Stride() 175 size_t stride = Rules<UniformType, RowsOrVecLength, Cols>::Stride( in Write() [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Analysis/ |
H A D | VectorUtils.cpp | 247 Value *Stride = U->getValue(); in getStrideFromPointer() 248 if (!Lp->isLoopInvariant(Stride)) in getStrideFromPointer() 254 Stride = getUniqueCastUse(Stride, Lp, StripedOffRecurrenceCast); in getStrideFromPointer() 256 return Stride; in getStrideFromPointer() 679 unsigned Stride, unsigned VF) { in createStrideMask() 682 Mask.push_back(Builder.getInt32(Start + i * Stride)); in createStrideMask() 799 bool InterleavedAccessInfo::isStrided(int Stride) { in isStrided() argument 800 unsigned Factor = std::abs(Stride); in isStrided() 832 int64_t Stride in collectConstStrideAccesses() local 678 createStrideMask(IRBuilder< &Builder, unsigned Start, unsigned Stride, unsigned VF) createStrideMask() argument [all...] |
H A D | LoopCacheAnalysis.cpp | 279 // (TripCount*Stride)/CLS, otherwise the cost is TripCount. 285 const SCEV *Stride = SE.getMulExpr(Coeff, ElemSize); 286 const SCEV *CacheLineSize = SE.getConstant(Stride->getType(), CLS); 287 Type *WiderType = SE.getWiderType(Stride->getType(), TripCount->getType()); 288 Stride = SE.getNoopOrSignExtend(Stride, WiderType); 290 const SCEV *Numerator = SE.getMulExpr(Stride, TripCount); 293 << "Access is consecutive: RefCost=(TripCount*Stride)/CLS=" 396 const SCEV *Stride = SE.getMulExpr(Coeff, ElemSize); 397 const SCEV *CacheLineSize = SE.getConstant(Stride [all...] |
H A D | LoopAccessAnalysis.cpp | 647 int64_t Stride = getPtrStride(PSE, Ptr, L, Strides); in isNoWrap() 648 if (Stride == 1 || PSE.hasNoOverflow(Ptr, SCEVWrapPredicate::IncrementNUSW)) in isNoWrap() 1075 int64_t Stride = StepVal / Size; in getPtrStride() local 1083 if (!IsNoWrapAddRec && Stride != 1 && Stride != -1 && in getPtrStride() 1098 return Stride; in getPtrStride() 1341 /// in \p Stride, and that have the same type size \p TypeByteSize, 1353 const SCEV &Dist, uint64_t Stride, in isSafeDependenceDistance() 1373 const uint64_t ByteStride = Stride * TypeByteSize; in isSafeDependenceDistance() 1406 /// Check the dependence for two accesses with the same stride \p Stride 1351 isSafeDependenceDistance(const DataLayout &DL, ScalarEvolution &SE, const SCEV &BackedgeTakenCount, const SCEV &Dist, uint64_t Stride, uint64_t TypeByteSize) isSafeDependenceDistance() argument 1411 areStridedAccessesIndependent(uint64_t Distance, uint64_t Stride, uint64_t TypeByteSize) areStridedAccessesIndependent() argument 1497 uint64_t Stride = std::abs(StrideAPtr); isDependent() local 2297 LLVM_DEBUG(dbgs() << " Ptr: " << *Ptr << " Stride: " << *Stride << "\\n"); collectStridedAccess() local [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/X86/ |
H A D | X86InterleavedAccess.cpp | 135 // Stride 4: in isSupported() 138 // Stride 3: in isSupported() 286 unsigned VecElems, unsigned Stride, in reorderSubVector() 290 for (unsigned i = 0; i < Stride; i++) in reorderSubVector() 299 for (unsigned i = 0; i < (VecElems / 16) * Stride; i += 2) { in reorderSubVector() 300 genShuffleBland(VT, VPShuf, OptimizeShuf, (i / Stride) * 16, in reorderSubVector() 301 (i + 1) / Stride * 16); in reorderSubVector() 303 Vec[i % Stride], Vec[(i + 1) % Stride], OptimizeShuf); in reorderSubVector() 308 std::copy(Temp, Temp + Stride, TransposedMatri in reorderSubVector() 284 reorderSubVector(MVT VT, SmallVectorImpl<Value *> &TransposedMatrix, ArrayRef<Value *> Vec, ArrayRef<uint32_t> VPShuf, unsigned VecElems, unsigned Stride, IRBuilder<> Builder) reorderSubVector() argument 435 createShuffleStride(MVT VT, int Stride, SmallVectorImpl<uint32_t> &Mask) createShuffleStride() argument [all...] |
/third_party/mesa3d/src/mesa/main/ |
H A D | arrayobj.c | 642 if (binding->Stride != binding2->Stride) in _mesa_update_vao_derived_arrays() 734 if (binding->Stride != binding2->Stride) in _mesa_update_vao_derived_arrays() 739 if (ptr + binding->Stride < attrib2->Ptr + in _mesa_update_vao_derived_arrays() 745 if (attrib2->Ptr + binding->Stride < ptr + vertex_end) in _mesa_update_vao_derived_arrays() 766 assert(attrib2->_EffRelativeOffset <= binding->Stride); in _mesa_update_vao_derived_arrays() 792 assert(binding->Stride == binding2->Stride); in _mesa_update_vao_derived_arrays() 800 assert(attrib->_EffRelativeOffset < binding->Stride); in _mesa_update_vao_derived_arrays() [all...] |
H A D | varray.c | 200 * and sets the Offset and Stride fields. 230 binding->Stride != stride) { in _mesa_bind_vertex_buffer() 240 binding->Stride = stride; in _mesa_bind_vertex_buffer() 927 /* The Stride and Ptr fields are not set by update_array_format() */ in update_array() 929 if ((array->Stride != stride) || (array->Ptr != ptr)) { in update_array() 930 array->Stride = stride; in update_array() 2135 return array->Stride; in get_vertex_array_attrib() 2426 params[0] = vao->BufferBinding[VERT_ATTRIB_GENERIC(index)].Stride; in _mesa_GetVertexArrayIndexediv() 3315 vao->BufferBinding[i].Stride, false, false); in _mesa_InternalBindVertexBuffers() 3327 vao->BufferBinding[i].Stride, tru in _mesa_InternalBindVertexBuffers() [all...] |
H A D | glthread_varray.c | 64 vao->Attrib[i].Stride = elem_size; in _mesa_glthread_reset_vao() 356 vao->Attrib[attrib].Stride = stride ? stride : elem_size; in attrib_pointer() 444 vao->Attrib[i].Stride = stride; in bind_vertex_buffer()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/ADT/ |
H A D | SparseSet.h | 206 const unsigned Stride = std::numeric_limits<SparseT>::max() + 1u; in findIndex() 207 for (unsigned i = Sparse[Idx], e = size(); i < e; i += Stride) { in findIndex() 212 // Stride is 0 when SparseT >= unsigned. We don't need to loop. in findIndex() 213 if (!Stride) in findIndex()
|
H A D | SparseMultiSet.h | 356 const unsigned Stride = std::numeric_limits<SparseT>::max() + 1u; in findIndex() 357 for (unsigned i = Sparse[Idx], e = Dense.size(); i < e; i += Stride) { in findIndex() 363 // Stride is 0 when SparseT >= unsigned. We don't need to loop. in findIndex() 364 if (!Stride) in findIndex()
|
/third_party/skia/third_party/externals/tint/src/sem/ |
H A D | sem_array_test.cc | 31 EXPECT_EQ(arr->Stride(), 32u); in TEST_F() 44 EXPECT_EQ(arr->Stride(), 32u); in TEST_F()
|
H A D | array.h | 74 uint32_t Stride() const { return stride_; } in Stride() function in tint::sem::Array 81 /// @returns true if the value returned by Stride() matches the element's
|
/third_party/mesa3d/src/compiler/glsl/ |
H A D | gl_nir_link_xfb.c | 188 linked_xfb->Buffers[buf].Stride = xfb_info->buffers[buf].stride / 4; in gl_nir_link_assign_xfb_resources() 211 xfb->buffers[i].stride = info->Buffers[i].Stride * 4; in gl_to_nir_xfb_info()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/ |
H A D | MachineRegisterInfo.cpp | 339 int Stride = 1; in moveOperands() local 341 Stride = -1; in moveOperands() 370 Dst += Stride; in moveOperands() 371 Src += Stride; in moveOperands()
|
/third_party/mesa3d/src/mesa/vbo/ |
H A D | vbo_save.h | 107 return node->cold->VAO[0]->BufferBinding[0].Stride; in _vbo_save_get_stride()
|
/third_party/mesa3d/src/gallium/targets/d3d10sw/tests/ |
H A D | tri.cpp | 241 UINT Stride = sizeof(Vertex); in main() local 243 pDeviceContext->IASetVertexBuffers(0, 1, pVertexBuffer.GetAddressOf(), &Stride, &Offset); in main()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Transforms/Vectorize/ |
H A D | LoopVectorizationLegality.cpp | 413 int Stride = getPtrStride(PSE, Ptr, TheLoop, Strides, CanAddPredicate, false); in isConsecutivePtr() local 414 if (Stride == 1 || Stride == -1) in isConsecutivePtr() 415 return Stride; in isConsecutivePtr()
|