Lines Matching refs:current_

167       current_(Event::START, Event::State::NOT_RUNNING,
169 previous_(current_),
194 current_.end_time = MonotonicallyIncreasingTimeInMs();
201 current_ = Event(Event::START, Event::State::NOT_RUNNING,
203 current_.end_time = MonotonicallyIncreasingTimeInMs();
204 previous_ = current_;
242 DCHECK_EQ(Event::SCAVENGER, current_.type);
259 DCHECK_EQ(Event::INCREMENTAL_MARK_COMPACTOR, current_.type);
260 DCHECK_EQ(Event::State::ATOMIC, current_.state);
262 current_.gc_reason = gc_reason;
263 current_.collector_reason = collector_reason;
267 current_.start_time = start_of_observable_pause_;
268 current_.reduce_memory = heap_->ShouldReduceMemory();
275 DCHECK_NE(Event::State::ATOMIC, current_.state);
280 young_gc_while_full_gc_ = current_.state != Event::State::NOT_RUNNING;
284 !Event::IsYoungGenerationEvent(current_.type));
302 current_.state == Event::State::NOT_RUNNING);
305 previous_ = current_;
306 current_ = Event(type, Event::State::MARKING, gc_reason, collector_reason);
314 current_.start_time = start_of_observable_pause_;
315 current_.reduce_memory = heap_->ShouldReduceMemory();
332 DCHECK_EQ(Event::State::MARKING, current_.state);
333 current_.state = Event::State::ATOMIC;
337 SampleAllocation(current_.start_time, heap_->NewSpaceAllocationCounter(),
341 current_.start_object_size = heap_->SizeOfObjects();
342 current_.start_memory_size = heap_->memory_allocator()->Size();
343 current_.start_holes_size = CountTotalHolesSize(heap_);
347 current_.young_object_size = new_space_size + new_lo_space_size;
359 current_.end_object_size = heap_->SizeOfObjects();
360 current_.end_memory_size = heap_->memory_allocator()->Size();
361 current_.end_holes_size = CountTotalHolesSize(heap_);
362 current_.survived_young_object_size = heap_->SurvivedYoungObjectSize();
375 current_.end_time = MonotonicallyIncreasingTimeInMs();
382 AddAllocation(current_.end_time);
384 double duration = current_.end_time - current_.start_time;
391 MakeBytesAndDuration(current_.young_object_size, duration));
393 MakeBytesAndDuration(current_.survived_young_object_size, duration));
396 if (current_.type == Event::INCREMENTAL_MARK_COMPACTOR) {
400 MakeBytesAndDuration(current_.end_object_size, duration));
403 MakeBytesAndDuration(current_.end_object_size, duration));
405 RecordMutatorUtilization(current_.end_time,
438 const bool is_young = Event::IsYoungGenerationEvent(current_.type);
443 if (current_.type == Event::INCREMENTAL_MARK_COMPACTOR) {
444 current_.incremental_marking_bytes = incremental_marking_bytes_;
445 current_.incremental_marking_duration = incremental_marking_duration_;
447 current_.incremental_scopes[i] = incremental_scopes_[i];
448 current_.scopes[i] = incremental_scopes_[i].duration;
454 DCHECK_EQ(0u, current_.incremental_marking_bytes);
455 DCHECK_EQ(0.0, current_.incremental_marking_duration);
463 DCHECK_EQ(Event::State::ATOMIC, current_.state);
464 current_.state = Event::State::SWEEPING;
468 DCHECK_EQ(Event::State::SWEEPING, current_.state);
469 current_.state = Event::State::NOT_RUNNING;
480 std::swap(current_, previous_);
487 static_cast<int>(current_.gc_reason));
498 if (current_.state != Event::State::SWEEPING) return;
509 DCHECK(Event::IsYoungGenerationEvent(current_.type));
510 if (current_.state != Event::State::SWEEPING) return;
515 StopCycle(current_.type == Event::SCAVENGER
526 DCHECK((current_.type == Event::MARK_COMPACTOR ||
527 current_.type == Event::INCREMENTAL_MARK_COMPACTOR) &&
528 (current_.state == Event::State::SWEEPING ||
529 (FLAG_verify_heap && current_.state == Event::State::ATOMIC)));
676 double duration = current_.end_time - current_.start_time;
680 if (current_.type == Event::INCREMENTAL_MARK_COMPACTOR) {
688 current_.end_time - incremental_marking_start_time_);
708 heap_->IsShared() ? "Shared " : "", current_.TypeName(false),
709 current_.reduce_memory ? " (reduce)" : "",
710 static_cast<double>(current_.start_object_size) / MB,
711 static_cast<double>(current_.start_memory_size) / MB,
712 static_cast<double>(current_.end_object_size) / MB,
713 static_cast<double>(current_.end_memory_size) / MB, duration,
717 Heap::GarbageCollectionReasonToString(current_.gc_reason),
718 current_.collector_reason != nullptr ? current_.collector_reason : "");
722 double duration = current_.end_time - current_.start_time;
723 double spent_in_mutator = current_.start_time - previous_.end_time;
725 current_.start_object_size - previous_.end_object_size;
729 if (current_.type == Event::INCREMENTAL_MARK_COMPACTOR) {
731 current_.end_time - incremental_marking_start_time_;
737 switch (current_.type) {
784 duration, spent_in_mutator, current_.TypeName(true),
785 current_.reduce_memory, current_.scopes[Scope::TIME_TO_SAFEPOINT],
808 ScavengeSpeedInBytesPerMillisecond(), current_.start_object_size,
809 current_.end_object_size, current_.start_holes_size,
810 current_.end_holes_size, allocated_since_last_gc,
849 duration, spent_in_mutator, "mmc", current_.reduce_memory,
970 duration, spent_in_mutator, current_.TypeName(true),
971 current_.reduce_memory, current_scope(Scope::TIME_TO_SAFEPOINT),
1041 current_scope(Scope::UNMAPPER), current_.start_object_size,
1042 current_.end_object_size, current_.start_holes_size,
1043 current_.end_holes_size, allocated_since_last_gc,
1294 static_cast<int>(current_.scopes[Scope::MC_BACKGROUND_MARKING]));
1296 static_cast<int>(current_.scopes[Scope::MC_BACKGROUND_SWEEPING]));
1312 current_.scopes[i] += background_counter_[i].total_duration_ms;
1322 static_cast<int>(current_.scopes[Scope::MC_CLEAR]));
1324 static_cast<int>(current_.scopes[Scope::MC_EPILOGUE]));
1326 static_cast<int>(current_.scopes[Scope::MC_EVACUATE]));
1328 static_cast<int>(current_.scopes[Scope::MC_FINISH]));
1330 static_cast<int>(current_.scopes[Scope::MC_MARK]));
1332 static_cast<int>(current_.scopes[Scope::MC_PROLOGUE]));
1334 static_cast<int>(current_.scopes[Scope::MC_SWEEP]));
1340 incremental_marking_duration_ + current_.scopes[Scope::MC_MARK];
1354 current_.scopes[Scope::MC_MARK_EMBEDDER_TRACING];
1370 static_cast<int>(current_.scopes[Scope::SCAVENGER_SCAVENGE_PARALLEL]));
1372 static_cast<int>(current_.scopes[Scope::SCAVENGER_SCAVENGE_ROOTS]));
1379 const double atomic_pause_duration = current_.scopes[Scope::MARK_COMPACTOR];
1397 current_.scopes[Scope::MC_PROLOGUE] + current_.scopes[Scope::MC_MARK];
1480 DCHECK(!Event::IsYoungGenerationEvent(current_.type));
1481 DCHECK_EQ(Event::State::NOT_RUNNING, current_.state);
1504 event.reason = static_cast<int>(current_.gc_reason);
1538 const double atomic_pause_duration = current_.scopes[Scope::MARK_COMPACTOR];
1540 current_.incremental_scopes[Scope::MC_INCREMENTAL_LAYOUT_CHANGE]
1542 current_.incremental_scopes[Scope::MC_INCREMENTAL_START].duration +
1543 current_.incremental_marking_duration +
1544 current_.incremental_scopes[Scope::MC_INCREMENTAL_FINALIZE].duration;
1546 current_.incremental_scopes[Scope::MC_INCREMENTAL_SWEEPING].duration;
1550 current_.scopes[Scope::MC_BACKGROUND_MARKING];
1552 current_.scopes[Scope::MC_BACKGROUND_SWEEPING];
1554 current_.scopes[Scope::MC_BACKGROUND_EVACUATE_COPY] +
1555 current_.scopes[Scope::MC_BACKGROUND_EVACUATE_UPDATE_POINTERS];
1560 current_.scopes[Scope::MC_PROLOGUE] + current_.scopes[Scope::MC_MARK];
1562 const double weak_duration = current_.scopes[Scope::MC_CLEAR];
1563 const double compact_duration = current_.scopes[Scope::MC_EVACUATE] +
1564 current_.scopes[Scope::MC_FINISH] +
1565 current_.scopes[Scope::MC_EPILOGUE];
1566 const double atomic_sweeping_duration = current_.scopes[Scope::MC_SWEEP];
1621 DCHECK_EQ(Event::Type::INCREMENTAL_MARK_COMPACTOR, current_.type);
1666 DCHECK(Event::IsYoungGenerationEvent(current_.type));
1667 DCHECK_EQ(Event::State::NOT_RUNNING, current_.state);
1675 event.reason = static_cast<int>(current_.gc_reason);
1709 (current_.scopes[Scope::SCAVENGER] +
1710 current_.scopes[Scope::MINOR_MARK_COMPACTOR] +
1711 current_.scopes[Scope::SCAVENGER_BACKGROUND_SCAVENGE_PARALLEL] +
1712 current_.scopes[Scope::MINOR_MC_BACKGROUND_EVACUATE_COPY] +
1713 current_.scopes[Scope::MINOR_MC_BACKGROUND_MARKING] +
1714 current_.scopes[Scope::MINOR_MC_BACKGROUND_EVACUATE_UPDATE_POINTERS]) *
1723 (current_.scopes[Scope::SCAVENGER] +
1724 current_.scopes[Scope::MINOR_MARK_COMPACTOR]) *
1730 static_cast<double>(current_.survived_young_object_size) /
1731 current_.young_object_size;
1734 current_.young_object_size - current_.survived_young_object_size;