Lines Matching defs:std
210 safepoint_(std::make_unique<IsolateSafepoint>(this)),
246 semi_space = std::min({semi_space, kMaxSemiSpaceSize});
247 semi_space = std::max({semi_space, kMinSemiSpaceSize});
258 std::min(old_generation,
261 std::max({old_generation, static_cast<uint64_t>(V8HeapTrait::kMinSize)});
310 return std::numeric_limits<size_t>::max();
325 return std::min(max_size, AllocatorLimitOnMaxOldGenerationSize());
498 gc_state_.store(state, std::memory_order_relaxed);
596 std::vector<int> categories_lengths(
598 std::vector<size_t> categories_sums(
606 std::ostringstream out_str;
609 out_str << "Page " << std::setw(4) << pageCnt;
620 out_str << "[" << cat << ": " << std::setw(4) << length << " || "
621 << std::setw(6) << sum << " ]"
651 std::ostringstream out_str;
655 << std::fixed << std::setprecision(2)
662 void Heap::DumpJSONHeapStatistics(std::stringstream& stream) {
676 std::stringstream stream;
751 allocations_count_.fetch_add(1, std::memory_order_relaxed);
762 allocations_count_.fetch_add(1, std::memory_order_relaxed);
764 allocations_count_.fetch_add(1, std::memory_order_relaxed);
773 allocations_count_.fetch_add(1, std::memory_order_relaxed);
785 allocations_count_.fetch_add(1, std::memory_order_relaxed);
817 allocations_count_.load(std::memory_order_relaxed), hash);
824 std::atomic<size_t> allocations_count_{0};
839 allocation_trackers_.erase(std::remove(allocation_trackers_.begin(),
884 std::vector<std::pair<HeapObject, bool>> retaining_path;
888 retaining_path.push_back(std::make_pair(object, ephemeron));
926 std::unordered_map<HeapObject, HeapObject, Object::Hasher>* map) {
927 std::unordered_map<HeapObject, HeapObject, Object::Hasher> updated_map;
948 *map = std::move(updated_map);
960 std::unordered_map<HeapObject, Root, Object::Hasher> updated_retaining_root;
974 retaining_root_ = std::move(updated_retaining_root);
1123 global_pretenuring_feedback_.insert(std::make_pair(site, 0));
1382 std::memory_order_relaxed);
1576 void ReportDuplicates(int size, std::vector<HeapObject>* objects) {
1585 std::vector<std::pair<int, HeapObject>> duplicates;
1593 duplicates.push_back(std::make_pair(count - 1, current));
1600 duplicates.push_back(std::make_pair(count - 1, current));
1656 std::map<int, std::vector<HeapObject>> objects_by_size;
1721 const double ms_step = std::min(
1722 kMaxStepSize, std::max(kMinStepSize, static_cast<double>(current) /
2515 std::max(v8_growing_factor, embedder_growing_factor);
2929 std::set<String> visited_map;
2930 std::map<MemoryChunk*, size_t> size_map;
2944 for (std::map<MemoryChunk*, size_t>::iterator it = size_map.begin();
2952 std::set<String> visited_map;
2953 std::map<MemoryChunk*, size_t> size_map;
2968 for (std::map<MemoryChunk*, size_t>::iterator it = size_map.begin();
3010 std::move(std::begin(young_strings_), std::end(young_strings_),
3011 std::back_inserter(old_strings_));
3099 Object list, const std::function<void(AllocationSite)>& visitor) {
3244 const std::function<void*(size_t)>& allocate, size_t byte_length) {
3278 std::max(OldGenerationSizeOfObjects() + minimum_growing_step,
3289 const size_t new_global_memory_limit = std::max(
4020 std::vector<MaybeObjectSlot> slots_;
4022 std::vector<CodeObjectSlot> code_slots_;
4229 MemoryPressureLevel::kNone, std::memory_order_relaxed);
4283 memory_pressure_level_.exchange(level, std::memory_order_relaxed);
4295 taskrunner->PostTask(std::make_unique<MemoryPressureInterruptTask>(this));
4312 near_heap_limit_callbacks_.push_back(std::make_pair(callback, data));
4359 std::min(heap_limit, AllocatorLimitOnMaxOldGenerationSize()));
4366 bool Heap::MeasureMemory(std::unique_ptr<v8::MeasureMemoryDelegate> delegate,
4369 std::vector<Handle<NativeContext>> contexts = FindAllNativeContexts();
4370 std::vector<Handle<NativeContext>> to_measure;
4377 return memory_measurement_->EnqueueRequest(std::move(delegate), execution,
4381 std::unique_ptr<v8::MeasureMemoryDelegate> Heap::MeasureMemoryDelegate(
4660 SlotVerifyingVisitor(Isolate* isolate, std::set<Address>* untyped,
4661 std::set<std::pair<SlotType, Address>>* typed)
4726 return typed_->count(std::make_pair(type, slot)) > 0;
4728 std::set<Address>* untyped_;
4729 std::set<std::pair<SlotType, Address>>* typed_;
4734 OldToNewSlotVerifyingVisitor(Isolate* isolate, std::set<Address>* untyped,
4735 std::set<std::pair<SlotType, Address>>* typed,
4771 std::set<Address>* untyped,
4772 std::set<std::pair<SlotType, Address>>* typed) {
4785 typed->insert(std::make_pair(type, slot));
4800 std::set<Address> old_to_new;
4801 std::set<std::pair<SlotType, Address>> typed_old_to_new;
5156 return std::min(static_cast<uint64_t>(std::numeric_limits<size_t>::max()),
5194 max_semi_space_size_ = std::max({max_semi_space_size_, kMinSemiSpaceSize});
5216 std::max(max_old_generation_size, MinOldGenerationSize());
5217 max_old_generation_size = std::min(max_old_generation_size,
5236 std::max(initial_semispace_size_, static_cast<size_t>(1 * MB));
5255 std::min(initial_semispace_size_, max_semi_space_size_);
5289 std::min(initial_old_generation_size_, max_old_generation_size() / 2);
5324 std::min(strlen(string), kTraceRingBufferSize - ring_buffer_end_);
5434 const size_t v8_margin = std::min(
5435 std::max(old_generation_allocation_limit() / 2, kMarginForSmallHeaps),
5438 std::min(std::max(global_allocation_limit_ / 2, kMarginForSmallHeaps),
5574 std::max(PercentToOldGenerationLimit(), PercentToGlobalMemoryLimit()));
5585 std::max<double>(max_marking_limit_reached_, current_percent);
5588 stress_marking_percentage_.load(std::memory_order_relaxed)) {
5590 std::memory_order_relaxed);
5599 std::max(PercentToOldGenerationLimit(), PercentToGlobalMemoryLimit()));
5715 code_range_ = std::make_shared<CodeRange>();
5854 std::make_unique<Heap::AllocationTrackerForDebugging>(this);
6209 std::find(gc_prologue_callbacks_.begin(), gc_prologue_callbacks_.end(),
6232 std::find(gc_epilogue_callbacks_.begin(), gc_epilogue_callbacks_.end(),
6284 std::vector<Handle<PrototypeInfo>> prototype_infos;
6524 reachable_[chunk] = new std::unordered_set<HeapObject, Object::Hasher>();
6611 std::vector<HeapObject> marking_stack_;
6624 std::unordered_map<Address, std::unordered_set<HeapObject, Object::Hasher>*>
6631 safepoint_scope_(std::make_unique<SafepointScope>(heap)),
6848 auto task = std::make_unique<FinalizationRegistryCleanupTask>(this);
6849 taskrunner->PostNonNestableTask(std::move(task));
6855 std::function<void(HeapObject object, ObjectSlot slot, Object target)>
6996 std::vector<Handle<NativeContext>> Heap::FindAllNativeContexts() {
6997 std::vector<Handle<NativeContext>> result;
7007 std::vector<WeakArrayList> Heap::FindAllRetainedMaps() {
7008 std::vector<WeakArrayList> result;
7250 std::shared_ptr<CodeRange> code_range = CodeRange::GetProcessWideCodeRange();
7295 ephemeron_remembered_set_.insert({table, std::unordered_set<int>()});