/third_party/node/deps/v8/src/heap/cppgc/ |
H A D | garbage-collector.h | 29 MarkingType::kAtomic, SweepingType::kAtomic}; in ConservativeAtomicConfig() 34 MarkingType::kAtomic, SweepingType::kAtomic}; in PreciseAtomicConfig() 39 MarkingType::kIncremental, SweepingType::kAtomic}; in ConservativeIncrementalConfig() 44 MarkingType::kIncremental, SweepingType::kAtomic}; in PreciseIncrementalConfig() 56 MarkingType::kAtomic, SweepingType::kAtomic}; in MinorPreciseAtomicConfig() 61 MarkingType::kAtomic, SweepingType::kAtomic}; in MinorConservativeAtomicConfig() [all...] |
H A D | marking-state.h | 76 if (header.IsInConstruction<AccessMode::kAtomic>()) { in MarkAndPush() 77 not_fully_constructed_worklist_.Push<AccessMode::kAtomic>(&header); in MarkAndPush() 88 DCHECK(!header.IsFree<AccessMode::kAtomic>()); in MarkNoPush() 101 DCHECK(header.IsMarked<AccessMode::kAtomic>()); in PushMarked() 102 DCHECK(!header.IsInConstruction<AccessMode::kAtomic>()); in PushMarked() 258 if (!header.IsInConstruction<AccessMode::kAtomic>() && in RegisterWeakReferenceIfNeeded() 259 header.IsMarked<AccessMode::kAtomic>()) in RegisterWeakReferenceIfNeeded() 271 weak_containers_worklist_.Push<AccessMode::kAtomic>(&header); in RegisterWeakContainer() 283 if (header.IsInConstruction<AccessMode::kAtomic>()) { in ProcessWeakContainer() 284 not_fully_constructed_worklist_.Push<AccessMode::kAtomic>( in ProcessWeakContainer() [all...] |
H A D | heap.cc | 49 MarkingType::kAtomic, SweepingType::kAtomic, in ForceGarbageCollectionSlow() 87 CHECK_IMPLIES(options.marking_support != HeapBase::MarkingType::kAtomic, in Heap() 89 CHECK_IMPLIES(options.sweeping_support != HeapBase::SweepingType::kAtomic, in Heap() 99 Config::MarkingType::kAtomic, Config::SweepingType::kAtomic}); in ~Heap() 107 DCHECK_EQ(Config::MarkingType::kAtomic, config.marking_type); in CollectGarbage() 122 DCHECK_NE(Config::MarkingType::kAtomic, config.marking_type); in StartIncrementalGarbageCollection() 123 DCHECK_NE(marking_support_, Config::MarkingType::kAtomic); in StartIncrementalGarbageCollection() 140 DCHECK_NE(Config::MarkingType::kAtomic, config in FinalizeIncrementalGarbageCollectionIfRunning() [all...] |
H A D | object-size-trait.cc | 17 return ObjectView<AccessMode::kAtomic>(HeapObjectHeader::FromObject(object)) in GetObjectSizeForGarbageCollected() 28 ->ObjectHeaderFromInnerAddress<AccessMode::kAtomic>(address); in GetObjectSizeForGarbageCollectedMixin() 29 DCHECK(!header.IsLargeObject<AccessMode::kAtomic>()); in GetObjectSizeForGarbageCollectedMixin() 30 return header.ObjectSize<AccessMode::kAtomic>(); in GetObjectSizeForGarbageCollectedMixin()
|
H A D | concurrent-marker.cc | 113 DynamicallyTraceMarkedObject<AccessMode::kAtomic>( in ProcessWorklists() 129 DCHECK(!header.IsInConstruction<AccessMode::kAtomic>()); in ProcessWorklists() 130 DCHECK(header.IsMarked<AccessMode::kAtomic>()); in ProcessWorklists() 146 DynamicallyTraceMarkedObject<AccessMode::kAtomic>( in ProcessWorklists() 160 DynamicallyTraceMarkedObject<AccessMode::kAtomic>( in ProcessWorklists()
|
H A D | write-barrier.cc | 37 DCHECK(header.IsMarked<AccessMode::kAtomic>()); in ProcessMarkValue() 45 header.Unmark<AccessMode::kAtomic>(); in ProcessMarkValue() 120 if (!header.IsMarked<AccessMode::kAtomic>()) return; in SteeleMarkingBarrierSlow() 152 ->ObjectHeaderFromInnerAddress<AccessMode::kAtomic>(inner_pointer); in GenerationalBarrierForSourceObjectSlow()
|
H A D | trace-trait.cc | 20 page->ObjectHeaderFromInnerAddress<AccessMode::kAtomic>(address); in GetTraceDescriptor() 23 header.GetGCInfoIndex<AccessMode::kAtomic>()) in GetTraceDescriptor()
|
H A D | sweeper.cc | 63 CHECK(bitmap_->CheckBit<AccessMode::kAtomic>(raw_header)); in VisitHeapObjectHeader() 66 CHECK_EQ(prev_, bitmap_->FindHeader<AccessMode::kAtomic>(raw_header - 1)); in VisitHeapObjectHeader() 205 header->Unmark<AccessMode::kAtomic>(); in StickyUnmark() 295 constexpr auto kAtomicAccess = AccessMode::kAtomic; in SweepNormalPage() 309 bitmap.ClearBit<AccessMode::kAtomic>(address); in SweepNormalPage() 312 DCHECK(bitmap.CheckBit<AccessMode::kAtomic>(address)); in SweepNormalPage() 318 DCHECK(bitmap.CheckBit<AccessMode::kAtomic>(begin)); in SweepNormalPage() 344 DCHECK(bitmap.CheckBit<AccessMode::kAtomic>(start_of_gap)); in SweepNormalPage() 358 DCHECK(bitmap.CheckBit<AccessMode::kAtomic>(start_of_gap)); in SweepNormalPage() 766 if (config.sweeping_type == SweepingConfig::SweepingType::kAtomic) { in Start() [all...] |
H A D | marker.cc | 217 config_.marking_type == MarkingConfig::MarkingType::kAtomic in StartMarking() 270 config_.marking_type = MarkingConfig::MarkingType::kAtomic; in EnterAtomicPause() 329 DCHECK_EQ(MarkingConfig::MarkingType::kAtomic, config_.marking_type); in ProcessWeakness() 404 if (config_.marking_type != MarkingConfig::MarkingType::kAtomic || in VisitCrossThreadPersistentsIfNeeded() 454 if (config_.marking_type != MarkingConfig::MarkingType::kAtomic || in JoinConcurrentMarkingIfNeeded() 490 DCHECK_NE(MarkingConfig::MarkingType::kAtomic, config_.marking_type); in AdvanceMarkingWithLimits() 507 if ((config_.marking_type == MarkingConfig::MarkingType::kAtomic) || in ProcessWorklistsWithDeadline()
|
H A D | heap-growing.cc | 98 GarbageCollector::Config::MarkingType::kAtomic, sweeping_support_}); in AllocatedObjectSizeIncreased() 100 if (marking_support_ == cppgc::Heap::MarkingType::kAtomic) return; in AllocatedObjectSizeIncreased()
|
H A D | gc-invoker.cc | 91 DCHECK_EQ(config.marking_type, cppgc::Heap::MarkingType::kAtomic); in CollectGarbage() 113 DCHECK_NE(config.marking_type, cppgc::Heap::MarkingType::kAtomic); in StartIncrementalGarbageCollection()
|
H A D | marking-state.cc | 17 not_fully_constructed_worklist_.Extract<AccessMode::kAtomic>(); in FlushNotFullyConstructedObjects()
|
H A D | object-allocator.cc | 68 .SetBit<AccessMode::kAtomic>(start); in AddToFreeList() 87 page->object_start_bitmap().ClearBit<AccessMode::kAtomic>(new_buffer); in ReplaceLinearAllocationBuffer()
|
H A D | object-allocator.h | 174 .SetBit<AccessMode::kAtomic>(reinterpret_cast<ConstAddress>(&filler)); in AllocateObjectOnSpace() 211 .SetBit<AccessMode::kAtomic>(reinterpret_cast<ConstAddress>(header)); in AllocateObjectOnSpace()
|
H A D | globals.h | 25 enum class AccessMode : uint8_t { kNonAtomic, kAtomic };
|
H A D | pointer-policies.cc | 70 header = &base_page->ObjectHeaderFromInnerAddress<AccessMode::kAtomic>(ptr); in CheckPointerImpl()
|
H A D | marking-visitor.cc | 58 if (header.IsMarked<AccessMode::kAtomic>()) { in VisitFullyConstructedConservatively()
|
H A D | heap-base.cc | 187 {Sweeper::SweepingConfig::SweepingType::kAtomic, in Terminate()
|
H A D | marker.h | 208 .Push<AccessMode::kAtomic>(&header); in WriteBarrierForInConstructionObject()
|
H A D | marking-worklists.h | 144 AccessMode::kAtomic> {
|
/third_party/node/deps/v8/include/cppgc/ |
H A D | heap.h | 69 kAtomic, member in cppgc::Heap::MarkingType 88 kAtomic, member in cppgc::Heap::SweepingType
|
/third_party/node/deps/v8/include/v8-include/cppgc/ |
H A D | heap.h | 65 kAtomic, member in cppgc::Heap::MarkingType 84 kAtomic, member in cppgc::Heap::SweepingType
|
/third_party/node/deps/v8/src/heap/cppgc-js/ |
H A D | cpp-heap.cc | 518 if (*collection_type_ == CollectionType::kMinor) return MarkingType::kAtomic; in SelectMarkingType() 521 return MarkingType::kAtomic; in SelectMarkingType() 527 if (IsForceGC(current_gc_flags_)) return SweepingType::kAtomic; in SelectSweepingType() 561 (MarkingType::kAtomic == marking_config.marking_type) || in InitializeTracing() 622 compactor_.CancelIfShouldNotCompact(MarkingType::kAtomic, stack_state); in EnterFinalPause() 672 SweepingType::kAtomic == sweeping_config.sweeping_type); in TraceEpilogue()
|
/third_party/skia/third_party/externals/tint/src/reader/wgsl/ |
H A D | token.h | 125 kAtomic, member in tint::reader::wgsl::Token::Type
|
/third_party/vixl/src/aarch64/ |
H A D | cpu-aarch64.h | 131 static const Field kAtomic; member in vixl::aarch64::AA64ISAR0
|