/third_party/skia/third_party/externals/icu/source/i18n/ |
H A D | formatted_string_builder.h | 45 } heap; member 192 return fUsingHeap ? fChars.heap.ptr : fChars.value; 196 return fUsingHeap ? fChars.heap.ptr : fChars.value; 200 return fUsingHeap ? fFields.heap.ptr : fFields.value; 204 return fUsingHeap ? fFields.heap.ptr : fFields.value; 208 return fUsingHeap ? fChars.heap.capacity : DEFAULT_CAPACITY;
|
/third_party/node/deps/v8/src/heap/ |
H A D | object-stats.h | 97 explicit ObjectStats(Heap* heap) : heap_(heap) { ClearObjectStats(true); } in ObjectStats() argument 135 Heap* heap() { return heap_; } in heap() function in v8::internal::ObjectStats 179 ObjectStatsCollector(Heap* heap, ObjectStats* live, ObjectStats* dead) in ObjectStatsCollector() argument 180 : heap_(heap), live_(live), dead_(dead) { in ObjectStatsCollector()
|
H A D | scavenger-inl.h | 8 #include "src/heap/evacuation-allocator-inl.h" 9 #include "src/heap/incremental-marking-inl.h" 10 #include "src/heap/memory-chunk.h" 11 #include "src/heap/scavenger.h" 90 heap()->CopyBlock(target.address() + kTaggedSize, in MigrateObject() 101 heap()->OnMoveEvent(target, source, size); in MigrateObject() 106 heap()->incremental_marking()->TransferColor(source, target); in MigrateObject() 108 heap()->UpdateAllocationSite(map, source, &local_pretenuring_feedback_); in MigrateObject() 119 DCHECK(heap()->AllowedToBeMigrated(map, object, NEW_SPACE)); in SemiSpaceCopyObject() 126 DCHECK(heap() in SemiSpaceCopyObject() 434 CheckAndScavengeObject(Heap* heap, TSlot slot) CheckAndScavengeObject() argument [all...] |
H A D | base-space.h | 27 Heap* heap() const { in heap() function in v8::internal::BaseSpace 65 BaseSpace(Heap* heap, AllocationSpace id) in BaseSpace() argument 66 : heap_(heap), id_(id), committed_(0), max_committed_(0) {} in BaseSpace()
|
H A D | memory-allocator.cc | 5 #include "src/heap/memory-allocator.h" 13 #include "src/heap/basic-memory-chunk.h" 14 #include "src/heap/gc-tracer-inl.h" 15 #include "src/heap/gc-tracer.h" 16 #include "src/heap/heap-inl.h" 17 #include "src/heap/heap.h" 18 #include "src/heap/memory-chunk.h" 19 #include "src/heap/rea 292 Heap* heap = isolate_->heap(); HandleAllocationFailure() local [all...] |
H A D | parked-scope.h | 10 #include "src/heap/local-heap.h" 15 // Scope that explicitly parks a thread, prohibiting access to the heap and the 20 : ParkedScope(local_isolate->heap()) {} in ParkedScope() 31 // Scope that explicitly unparks a thread, allowing access to the heap and the 36 : UnparkedScope(local_isolate->heap()) {} in UnparkedScope() 52 : ParkedMutexGuard(local_isolate->heap(), mutex) {} in ParkedMutexGuard() 77 : ParkedSharedMutexGuardIf(local_isolate->heap(), mutex, enable_mutex) {} in ParkedSharedMutexGuardIf()
|
/third_party/mesa3d/src/gallium/drivers/iris/ |
H A D | iris_bufmgr.c | 284 bucket_info_for_heap(struct iris_bufmgr *bufmgr, enum iris_heap heap, in bucket_info_for_heap() argument 287 switch (heap) { in bucket_info_for_heap() 304 unreachable("invalid heap"); in bucket_info_for_heap() 316 enum iris_heap heap) in bucket_for_size() 348 bucket_info_for_heap(bufmgr, heap, &buckets, &num_buckets); in bucket_for_size() 656 unsigned heap, in iris_slab_alloc() 662 unsigned flags = heap == IRIS_HEAP_SYSTEM_MEMORY ? BO_ALLOC_SMEM : in iris_slab_alloc() 663 heap == IRIS_HEAP_DEVICE_LOCAL ? BO_ALLOC_LMEM : 0; in iris_slab_alloc() 795 enum iris_heap heap = flags_to_heap(bufmgr, flags); in alloc_bo_from_slabs() local 821 entry = pb_slab_alloc(slabs, alloc_size, heap); in alloc_bo_from_slabs() 315 bucket_for_size(struct iris_bufmgr *bufmgr, uint64_t size, enum iris_heap heap) bucket_for_size() argument 655 iris_slab_alloc(void *priv, unsigned heap, unsigned entry_size, unsigned group_index) iris_slab_alloc() argument 1065 enum iris_heap heap = flags_to_heap(bufmgr, flags); iris_bo_alloc() local 2105 add_bucket(struct iris_bufmgr *bufmgr, int size, enum iris_heap heap) add_bucket() argument 2122 init_cache_buckets(struct iris_bufmgr *bufmgr, enum iris_heap heap) init_cache_buckets() argument [all...] |
/third_party/node/deps/v8/include/v8-include/cppgc/internal/ |
H A D | write-barrier.h | 11 #include "cppgc/heap-handle.h" 12 #include "cppgc/heap-state.h" 23 #include "cppgc/internal/caged-heap-local-data.h" 24 #include "cppgc/internal/caged-heap.h" 54 HeapHandle* heap = nullptr; member 258 params.heap = &heap_handle; in BarrierEnabledGet() 268 params.heap = &heap_handle; in BarrierEnabledGet() 289 params.heap = &handle; in Get() 302 params.heap = &handle; in Get() 379 params.heap in Get() [all...] |
/third_party/node/deps/v8/src/heap/cppgc/ |
H A D | marking-state.h | 13 #include "src/heap/cppgc/compaction-worklists.h" 14 #include "src/heap/cppgc/globals.h" 15 #include "src/heap/cppgc/heap-object-header.h" 16 #include "src/heap/cppgc/heap-page.h" 17 #include "src/heap/cppgc/liveness-broker.h" 18 #include "src/heap/cppgc/marking-worklists.h" 58 MarkingStateBase::MarkingStateBase(HeapBase& heap, in MarkingStateBase() argument 60 : heap_(heap), in MarkingStateBase() 226 BasicMarkingState(HeapBase& heap, MarkingWorklists& marking_worklists, CompactionWorklists* compaction_worklists) BasicMarkingState() argument 358 MutatorMarkingState(HeapBase& heap, MarkingWorklists& marking_worklists, CompactionWorklists* compaction_worklists) MutatorMarkingState() argument 461 ConcurrentMarkingState(HeapBase& heap, MarkingWorklists& marking_worklists, CompactionWorklists* compaction_worklists) ConcurrentMarkingState() argument [all...] |
H A D | concurrent-marker.h | 9 #include "src/heap/cppgc/incremental-marking-schedule.h" 10 #include "src/heap/cppgc/marking-state.h" 11 #include "src/heap/cppgc/marking-visitor.h" 12 #include "src/heap/cppgc/marking-worklists.h" 36 HeapBase& heap() const { return heap_; } in heap() function in cppgc::internal::ConcurrentMarkerBase 64 ConcurrentMarker(HeapBase& heap, MarkingWorklists& marking_worklists, in ConcurrentMarker() argument 67 : ConcurrentMarkerBase(heap, marking_worklists, in ConcurrentMarker()
|
H A D | testing.cc | 8 #include "src/heap/cppgc/heap-base.h" 16 auto& heap = internal::HeapBase::From(heap_handle_); in OverrideEmbedderStackStateScope() local 17 CHECK_NULL(heap.override_stack_state_.get()); in OverrideEmbedderStackStateScope() 18 heap.override_stack_state_ = std::make_unique<EmbedderStackState>(state); in OverrideEmbedderStackStateScope()
|
H A D | visitor.cc | 5 #include "src/heap/cppgc/visitor.h" 8 #include "src/heap/cppgc/caged-heap.h" 9 #include "src/heap/cppgc/gc-info-table.h" 10 #include "src/heap/cppgc/heap-base.h" 11 #include "src/heap/cppgc/heap-object-header.h" 12 #include "src/heap/cppgc/heap 27 ConservativeTracingVisitor( HeapBase& heap, PageBackend& page_backend, cppgc::Visitor& visitor) ConservativeTracingVisitor() argument [all...] |
/third_party/node/deps/v8/src/objects/ |
H A D | free-space-inl.h | 9 #include "src/heap/heap-write-barrier-inl.h" 10 #include "src/heap/heap.h" 50 Heap* heap = GetHeapFromWritableObject(*this); in IsValid() local 52 Isolate::FromHeap(heap)->root(RootIndex::kFreeSpaceMap); in IsValid() 54 !heap->deserialization_complete() && in IsValid()
|
/third_party/mesa3d/src/gallium/auxiliary/pipebuffer/ |
H A D | pb_slab.h | 35 * meaning of each heap is treated as opaque by this library. 84 * for fulfilling allocation requests of the given size from the given heap. 94 unsigned heap, 122 /* One group per (heap, order, three_fourth_allocations). */ 138 pb_slab_alloc_reclaimed(struct pb_slabs *slabs, unsigned size, unsigned heap, bool reclaim_all); 141 pb_slab_alloc(struct pb_slabs *slabs, unsigned size, unsigned heap);
|
/third_party/mesa3d/src/amd/vulkan/winsys/amdgpu/ |
H A D | radv_amdgpu_winsys.c | 76 struct amdgpu_heap_info heap; in radv_amdgpu_winsys_query_value() local 99 amdgpu_query_heap_info(ws->dev, AMDGPU_GEM_DOMAIN_VRAM, 0, &heap); in radv_amdgpu_winsys_query_value() 100 return heap.heap_usage; in radv_amdgpu_winsys_query_value() 103 &heap); in radv_amdgpu_winsys_query_value() 104 return heap.heap_usage; in radv_amdgpu_winsys_query_value() 106 amdgpu_query_heap_info(ws->dev, AMDGPU_GEM_DOMAIN_GTT, 0, &heap); in radv_amdgpu_winsys_query_value() 107 return heap.heap_usage; in radv_amdgpu_winsys_query_value()
|
/third_party/skia/src/pathops/ |
H A D | SkPathOpsTSect.h | 90 SkTSpan(const SkTCurve& curve, SkArenaAlloc& heap) { in SkTSpan() argument 91 fPart = curve.make(heap); in SkTSpan() 98 void debugInit(const SkTCurve& curve, SkArenaAlloc& heap) { in debugInit() argument 100 SkTCurve* fake = curve.make(heap); in debugInit() 195 bool split(SkTSpan* work, SkArenaAlloc* heap) { in split() argument 196 return splitAt(work, (work->fStartT + work->fEndT) * 0.5, heap); in split() 199 bool splitAt(SkTSpan* work, double t, SkArenaAlloc* heap);
|
/kernel/linux/linux-5.10/drivers/dma-buf/heaps/ |
H A D | system_heap.c | 3 * DMABUF System heap exporter 11 #include <linux/dma-heap.h> 21 #include "heap-helpers.h" 35 static int system_heap_allocate(struct dma_heap *heap, in system_heap_allocate() argument 50 helper_buffer->heap = heap; in system_heap_allocate()
|
/third_party/ltp/testcases/kernel/mem/hugetlb/hugemmap/ |
H A D | hugemmap24.c | 31 void *p, *heap; in init_slice_boundary() local 43 /* dummy malloc so we know where is heap */ in init_slice_boundary() 44 heap = malloc(1); in init_slice_boundary() 45 free(heap); in init_slice_boundary() 48 * The additionally plus heap address is to reduce the possibility in init_slice_boundary() 51 while (slice_boundary + slice_size < (unsigned long)heap + 2*hpage_size) in init_slice_boundary()
|
/third_party/node/deps/v8/src/execution/ |
H A D | local-isolate.h | 14 #include "src/heap/local-factory.h" 15 #include "src/heap/local-heap.h" 51 static LocalIsolate* FromHeap(LocalHeap* heap) { in FromHeap() argument 52 return reinterpret_cast<LocalIsolate*>(reinterpret_cast<Address>(heap) - in FromHeap() 56 bool is_main_thread() { return heap()->is_main_thread(); } in is_main_thread() 58 LocalHeap* heap() { return &heap_; } in heap() function in v8::internal::final
|
/third_party/skia/third_party/externals/dawn/src/dawn_native/d3d12/ |
H A D | CommandRecordingContext.cpp | 86 void CommandRecordingContext::TrackHeapUsage(Heap* heap, ExecutionSerial serial) { in TrackHeapUsage() argument 87 // Before tracking the heap, check the last serial it was recorded on to ensure we aren't in TrackHeapUsage() 89 if (heap->GetLastUsage() < serial) { in TrackHeapUsage() 90 heap->SetLastUsage(serial); in TrackHeapUsage() 91 mHeapsPendingUsage.push_back(heap); in TrackHeapUsage()
|
/third_party/node/deps/v8/third_party/zlib/ |
H A D | trees.c | 88 * need for the L_CODES extra codes used during heap construction. However 423 /* Index within the heap array of least frequent node in the Huffman tree */ 427 * Remove the smallest element from the heap and recreate the heap with 428 * one less element. Updates heap and heap_len. 432 top = s->heap[SMALLEST]; \ 433 s->heap[SMALLEST] = s->heap[s->heap_len--]; \ 446 * Restore the heap property by moving down the tree starting at node k, 448 * when the heap propert [all...] |
/third_party/node/deps/zlib/ |
H A D | trees.c | 88 * need for the L_CODES extra codes used during heap construction. However 479 /* Index within the heap array of least frequent node in the Huffman tree */ 483 * Remove the smallest element from the heap and recreate the heap with 484 * one less element. Updates heap and heap_len. 488 top = s->heap[SMALLEST]; \ 489 s->heap[SMALLEST] = s->heap[s->heap_len--]; \ 502 * Restore the heap property by moving down the tree starting at node k, 504 * when the heap propert [all...] |
/third_party/libwebsockets/win32port/zlib/ |
H A D | trees.c | 93 * need for the L_CODES extra codes used during heap construction. However
431 /* Index within the heap array of least frequent node in the Huffman tree */
435 * Remove the smallest element from the heap and recreate the heap with
436 * one less element. Updates heap and heap_len.
440 top = s->heap[SMALLEST]; \
441 s->heap[SMALLEST] = s->heap[s->heap_len--]; \
454 * Restore the heap property by moving down the tree starting at node k,
456 * when the heap propert [all...] |
/third_party/skia/third_party/externals/zlib/ |
H A D | trees.c | 88 * need for the L_CODES extra codes used during heap construction. However 423 /* Index within the heap array of least frequent node in the Huffman tree */ 427 * Remove the smallest element from the heap and recreate the heap with 428 * one less element. Updates heap and heap_len. 432 top = s->heap[SMALLEST]; \ 433 s->heap[SMALLEST] = s->heap[s->heap_len--]; \ 446 * Restore the heap property by moving down the tree starting at node k, 448 * when the heap propert [all...] |
/third_party/zlib/ |
H A D | trees.c | 88 * need for the L_CODES extra codes used during heap construction. However 496 /* Index within the heap array of least frequent node in the Huffman tree */ 500 * Remove the smallest element from the heap and recreate the heap with 501 * one less element. Updates heap and heap_len. 505 top = s->heap[SMALLEST]; \ 506 s->heap[SMALLEST] = s->heap[s->heap_len--]; \ 519 * Restore the heap property by moving down the tree starting at node k, 521 * when the heap propert [all...] |