Lines Matching refs:value

63   // Returns the required write barrier for a given `slot` and `value`.
64 static V8_INLINE Type GetWriteBarrierType(const void* slot, const void* value,
70 // Returns the required write barrier for a given `value`.
71 static V8_INLINE Type GetWriteBarrierType(const void* value, Params& params);
114 static void DijkstraMarkingBarrierSlow(const void* value);
115 static void DijkstraMarkingBarrierSlowWithSentinelCheck(const void* value);
121 static void SteeleMarkingBarrierSlow(const void* value);
122 static void SteeleMarkingBarrierSlowWithSentinelCheck(const void* value);
149 static V8_INLINE WriteBarrier::Type Get(const void* slot, const void* value,
152 return ValueModeDispatch<value_mode>::Get(slot, value, params, callback);
156 static V8_INLINE WriteBarrier::Type Get(const void* value,
159 return GetNoSlot(value, params, callback);
166 static V8_INLINE WriteBarrier::Type GetNoSlot(const void* value,
169 if (!TryGetCagedHeap(value, value, params)) {
181 static V8_INLINE bool TryGetCagedHeap(const void* slot, const void* value,
185 if (!value) return false;
186 params.start = reinterpret_cast<uintptr_t>(value) &
191 // Check if slot is on stack or value is sentinel or nullptr. This relies
209 static V8_INLINE WriteBarrier::Type Get(const void* slot, const void* value,
217 bool within_cage = TryGetCagedHeap(slot, value, params);
225 params.value_offset = reinterpret_cast<uintptr_t>(value) - params.start;
275 static V8_INLINE WriteBarrier::Type Get(const void* slot, const void* value,
278 return ValueModeDispatch<value_mode>::Get(slot, value, params, callback);
282 static V8_INLINE WriteBarrier::Type Get(const void* value,
286 return Get<WriteBarrier::ValueMode::kValuePresent>(nullptr, value, params,
343 const void* slot, const void* value, WriteBarrier::Params& params) {
344 return WriteBarrierTypePolicy::Get<ValueMode::kValuePresent>(slot, value,
359 const void* value, WriteBarrier::Params& params) {
360 return WriteBarrierTypePolicy::Get<ValueMode::kValuePresent>(value, params,