/third_party/libdrm/ |
H A D | xf86atomic.h | 31 * Private definitions for atomic operations 42 int atomic; member 45 # define atomic_read(x) ((x)->atomic) 46 # define atomic_set(x, val) ((x)->atomic = (val)) 47 # define atomic_inc(x) ((void) __sync_fetch_and_add (&(x)->atomic, 1)) 48 # define atomic_inc_return(x) (__sync_add_and_fetch (&(x)->atomic, 1)) 49 # define atomic_dec_and_test(x) (__sync_add_and_fetch (&(x)->atomic, -1) == 0) 50 # define atomic_add(x, v) ((void) __sync_add_and_fetch(&(x)->atomic, (v))) 51 # define atomic_dec(x, v) ((void) __sync_sub_and_fetch(&(x)->atomic, (v))) 52 # define atomic_cmpxchg(x, oldv, newv) __sync_val_compare_and_swap (&(x)->atomic, old 63 AO_t atomic; global() member 88 typedef struct { LIBDRM_ATOMIC_TYPE atomic; } atomic_t; global() member [all...] |
/third_party/mesa3d/src/freedreno/ir3/ |
H A D | ir3_a6xx.c | 36 * atomic instructions (used for both SSBO and image) use a new instruction 93 * SSBO atomic intrinsics 95 * All of the SSBO atomic memory operations read a value from memory, 103 * 1: The offset into the SSBO buffer of the variable that the atomic 105 * 2: The data parameter to the atomic function (i.e. the value to add 113 struct ir3_instruction *atomic, *ibo, *src0, *src1, *data, *dummy; in emit_intrinsic_atomic_ssbo() local 147 atomic = ir3_ATOMIC_B_ADD(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 150 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 154 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 157 atomic in emit_intrinsic_atomic_ssbo() 258 struct ir3_instruction *atomic, *ibo, *src0, *src1, *dummy; emit_intrinsic_atomic_image() local 448 struct ir3_instruction *addr, *atomic, *src1; emit_intrinsic_atomic_global() local [all...] |
H A D | ir3_a4xx.c | 130 * SSBO atomic intrinsics 132 * All of the SSBO atomic memory operations read a value from memory, 140 * 1: The byte offset into the SSBO buffer of the variable that the atomic 142 * 2: The data parameter to the atomic function (i.e. the value to add 155 struct ir3_instruction *atomic; in emit_intrinsic_atomic_ssbo() local 169 atomic = ir3_ATOMIC_S_ADD(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 172 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 176 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 179 atomic = ir3_ATOMIC_S_MAX(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 183 atomic in emit_intrinsic_atomic_ssbo() 368 struct ir3_instruction *atomic, *src0, *src1, *src2; emit_intrinsic_atomic_image() local [all...] |
H A D | disasm-a3xx.c | 341 OPC(6, OPC_ATOMIC_ADD, atomic.add), 342 OPC(6, OPC_ATOMIC_SUB, atomic.sub), 343 OPC(6, OPC_ATOMIC_XCHG, atomic.xchg), 344 OPC(6, OPC_ATOMIC_INC, atomic.inc), 345 OPC(6, OPC_ATOMIC_DEC, atomic.dec), 346 OPC(6, OPC_ATOMIC_CMPXCHG, atomic.cmpxchg), 347 OPC(6, OPC_ATOMIC_MIN, atomic.min), 348 OPC(6, OPC_ATOMIC_MAX, atomic.max), 349 OPC(6, OPC_ATOMIC_AND, atomic.and), 350 OPC(6, OPC_ATOMIC_OR, atomic [all...] |
/third_party/skia/third_party/externals/tint/src/resolver/ |
H A D | atomics_test.cc | 31 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 36 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local 37 ASSERT_NE(atomic, nullptr); in TEST_F() 38 EXPECT_TRUE(atomic->Type()->Is<sem::I32>()); in TEST_F() 42 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.u32()), in TEST_F() 47 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local 48 ASSERT_NE(atomic, nullptr); in TEST_F() 49 EXPECT_TRUE(atomic->Type()->Is<sem::U32>()); in TEST_F() 53 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F() 67 auto* atomic in TEST_F() local [all...] |
H A D | atomics_validation_test.cc | 30 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 37 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F() 46 Global("a", ty.atomic(ty.f32(Source{{12, 34}})), in TEST_F() 50 EXPECT_EQ(r()->error(), "12:34 error: atomic only supports i32 or u32 types"); in TEST_F() 54 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 59 "12:34 error: atomic variables must have <storage> or <workgroup> " in TEST_F() 64 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 69 "12:34 error: atomic variables must have <storage> or <workgroup> " in TEST_F() 75 Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}); in TEST_F() 80 "error: atomic variable in TEST_F() [all...] |
/third_party/skia/src/gpu/ |
H A D | GrThreadSafePipelineBuilder.h | 14 #include <atomic> 78 std::atomic<int> fShaderCompilations{0}; 80 std::atomic<int> fNumInlineCompilationFailures{0}; 81 std::array<std::atomic<int>, kNumProgramCacheResults> fInlineProgramCacheStats = {}; 83 std::atomic<int> fNumPreCompilationFailures{0}; 84 std::array<std::atomic<int>, kNumProgramCacheResults> fPreProgramCacheStats = {}; 86 std::atomic<int> fNumCompilationFailures{0}; 87 std::atomic<int> fNumPartialCompilationSuccesses{0}; 88 std::atomic<int> fNumCompilationSuccesses{0};
|
/third_party/skia/third_party/externals/abseil-cpp/absl/base/internal/ |
H A D | spinlock_wait.h | 18 // Operations to make atomic transitions on a word, and to allow 22 #include <atomic> 40 // then return the old value of *w. Make any other atomic transitions 42 uint32_t SpinLockWait(std::atomic<uint32_t> *w, int n, 50 void SpinLockWake(std::atomic<uint32_t> *w, bool all); 56 void SpinLockDelay(std::atomic<uint32_t> *w, uint32_t value, int loop, 74 void ABSL_INTERNAL_C_SYMBOL(AbslInternalSpinLockWake)(std::atomic<uint32_t> *w, 77 std::atomic<uint32_t> *w, uint32_t value, int loop, 81 inline void absl::base_internal::SpinLockWake(std::atomic<uint32_t> *w, in SpinLockWake() 87 std::atomic<uint32_ in SpinLockDelay() [all...] |
H A D | thread_identity.h | 30 #include <atomic> 111 std::atomic<State> state; 150 std::atomic<int>* blocked_count_ptr; 155 std::atomic<int> ticker; // Tick counter, incremented once per second. 156 std::atomic<int> wait_start; // Ticker value when thread started waiting. 157 std::atomic<bool> is_idle; // Has thread become idle yet?
|
/third_party/node/deps/v8/include/v8-include/cppgc/internal/ |
H A D | gc-info.h | 8 #include <atomic> 28 std::atomic<GCInfoIndex>& registered_index) { in EnsureIndex() 39 EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback, 42 std::atomic<GCInfoIndex>&, TraceCallback, FinalizationCallback); 44 std::atomic<GCInfoIndex>&, TraceCallback, NameCallback); 46 EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback); 48 EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback, 51 std::atomic<GCInfoIndex>&, TraceCallback, FinalizationCallback); 53 std::atomic<GCInfoIndex>&, TraceCallback, NameCallback); 55 EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoInde [all...] |
/third_party/node/deps/v8/include/cppgc/internal/ |
H A D | gc-info.h | 8 #include <atomic> 27 std::atomic<GCInfoIndex>& registered_index) { in EnsureIndex() 37 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 41 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 44 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 46 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 48 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 53 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 56 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 59 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoInde [all...] |
/third_party/skia/third_party/externals/abseil-cpp/absl/container/internal/ |
H A D | hashtablez_sampler.h | 42 #include <atomic> 73 std::atomic<size_t> capacity; 74 std::atomic<size_t> size; 75 std::atomic<size_t> num_erases; 76 std::atomic<size_t> num_rehashes; 77 std::atomic<size_t> max_probe_length; 78 std::atomic<size_t> total_probe_length; 79 std::atomic<size_t> hashes_bitwise_or; 80 std::atomic<size_t> hashes_bitwise_and; 81 std::atomic<size_ [all...] |
/third_party/node/deps/v8/src/heap/cppgc/ |
H A D | gc-info.cc | 24 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 33 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 42 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 50 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexPolymorphic() 57 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 66 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 75 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 83 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexNonPolymorphic()
|
/third_party/ltp/lib/newlib_tests/ |
H A D | test15.c | 8 * limited check that atomic stores and loads order non-atomic memory 13 * atomic functions. X86 in particular has strong memory ordering by default 15 * (Raspberry Pi 3 Model B) has been observed to fail without the atomic 31 * from removing/reording atomic and seq_n, mark them as volatile. 42 static int atomic; variable 43 /* Instead of storing seq_n on the stack (probably next to the atomic variable 46 * seq_n will be synchronised between processors as a byproduct of the atomic 56 for (i = tst_atomic_load(&atomic); in worker_load_store() 58 i = tst_atomic_load(&atomic)) in worker_load_store() [all...] |
/third_party/skia/third_party/externals/harfbuzz/src/ |
H A D | hb-atomic.hh | 81 #include <atomic> 87 #define hb_atomic_int_impl_add(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->fetch_add ((V), std::memory_order_acq_rel)) 88 #define hb_atomic_int_impl_set_relaxed(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_relaxed)) 89 #define hb_atomic_int_impl_set(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_release)) 90 #define hb_atomic_int_impl_get_relaxed(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_relaxed)) 91 #define hb_atomic_int_impl_get(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_acquire)) 93 #define hb_atomic_ptr_impl_set_relaxed(P, V) (reinterpret_cast<std::atomic<void*> *> (P)->store ((V), std::memory_order_relaxed)) 94 #define hb_atomic_ptr_impl_get_relaxed(P) (reinterpret_cast<std::atomic<void*> const *> (P)->load (std::memory_order_relaxed)) 95 #define hb_atomic_ptr_impl_get(P) (reinterpret_cast<std::atomic<void*> *> (P)->load (std::memory_order_acquire)) 100 return reinterpret_cast<std::atomic<cons in _hb_atomic_ptr_impl_cmplexch() [all...] |
/third_party/skia/third_party/externals/abseil-cpp/absl/time/ |
H A D | clock.cc | 25 #include <atomic> 103 // An implementation of reader-write locks that use no atomic ops in the read 115 static inline uint64_t SeqAcquire(std::atomic<uint64_t> *seq) { 130 static inline void SeqRelease(std::atomic<uint64_t> *seq, uint64_t x) { 155 std::atomic<uint64_t> raw_ns{0}; // raw kernel time 156 std::atomic<uint64_t> base_ns{0}; // our estimate of time 157 std::atomic<uint64_t> base_cycles{0}; // cycle counter reading 158 std::atomic<uint64_t> nsscaled_per_cycle{0}; // cycle period 161 std::atomic<uint64_t> min_cycles_per_sample{0}; 163 // Same again, but with non-atomic type [all...] |
/third_party/skia/third_party/externals/abseil-cpp/absl/base/ |
H A D | call_once.h | 29 #include <atomic> 49 std::atomic<uint32_t>* ControlWord(absl::once_flag* flag); 92 friend std::atomic<uint32_t>* base_internal::ControlWord(once_flag* flag); 93 std::atomic<uint32_t> control_; 147 void CallOnceImpl(std::atomic<uint32_t>* control, in CallOnceImpl() 188 inline std::atomic<uint32_t>* ControlWord(once_flag* flag) { in ControlWord() 194 std::atomic<uint32_t>* once = base_internal::ControlWord(flag); in LowLevelCallOnce() 207 std::atomic<uint32_t>* once = base_internal::ControlWord(&flag); in call_once()
|
/third_party/skia/third_party/externals/oboe/src/fifo/ |
H A D | FifoControllerIndirect.h | 20 #include <atomic> 34 std::atomic<uint64_t> *readCounterAddress, 35 std::atomic<uint64_t> *writeCounterAddress); 59 std::atomic<uint64_t> *mReadCounterAddress; 60 std::atomic<uint64_t> *mWriteCounterAddress;
|
/third_party/node/deps/v8/include/cppgc/ |
H A D | member.h | 8 #include <atomic> 38 return reinterpret_cast<const std::atomic<const void*>*>(&raw_)->load( in GetRawAtomic() 42 reinterpret_cast<std::atomic<const void*>*>(&raw_)->store( in SetRawAtomic() 50 // results in a non-atomic write on some builds, even when the atomic version 71 // initialize using atomic assignments. This is required for preventing 74 BasicMember(std::nullptr_t, AtomicInitializerTag atomic) in BasicMember() argument 75 : MemberBase(nullptr, atomic) {} in BasicMember() 76 BasicMember(SentinelPointer s, AtomicInitializerTag atomic) in BasicMember() argument 77 : MemberBase(s, atomic) {} in BasicMember() 78 BasicMember(T* raw, AtomicInitializerTag atomic) BasicMember() argument 82 BasicMember(T& raw, AtomicInitializerTag atomic) BasicMember() argument [all...] |
/third_party/icu/icu4c/source/common/ |
H A D | umutex.h | 23 #include <atomic> 35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn. 40 // Export an explicit template instantiation of std::atomic<int32_t>. 44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class. 52 template struct U_COMMON_API std::atomic<int32_t>; member in std 53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std 59 template struct std::atomic<int32_t>; member in std 60 template struct std::atomic<std::mutex *>; member in std 73 typedef std::atomic<int32_t> u_atomic_int32_t; 240 std::atomic<st [all...] |
/third_party/node/deps/icu-small/source/common/ |
H A D | umutex.h | 23 #include <atomic> 35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn. 40 // Export an explicit template instantiation of std::atomic<int32_t>. 44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class. 52 template struct U_COMMON_API std::atomic<int32_t>; member in std 53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std 59 template struct std::atomic<int32_t>; member in std 60 template struct std::atomic<std::mutex *>; member in std 73 typedef std::atomic<int32_t> u_atomic_int32_t; 240 std::atomic<st [all...] |
/third_party/skia/third_party/externals/icu/source/common/ |
H A D | umutex.h | 23 #include <atomic> 35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn. 40 // Export an explicit template instantiation of std::atomic<int32_t>. 44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class. 52 template struct U_COMMON_API std::atomic<int32_t>; member in std 53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std 59 template struct std::atomic<int32_t>; member in std 60 template struct std::atomic<std::mutex *>; member in std 73 typedef std::atomic<int32_t> u_atomic_int32_t; 244 std::atomic<st [all...] |
/third_party/skia/third_party/externals/abseil-cpp/absl/flags/internal/ |
H A D | sequence_lock_test.cc | 17 #include <atomic> 48 absl::FixedArray<std::atomic<uint64_t>> protected_buf(buf_words); in TEST_P() 52 std::atomic<bool> stop{false}; in TEST_P() 53 std::atomic<int64_t> bad_reads{0}; in TEST_P() 54 std::atomic<int64_t> good_reads{0}; in TEST_P() 55 std::atomic<int64_t> unsuccessful_reads{0}; in TEST_P() 119 // With TSAN, a lot of threads contending for atomic access on the sequence 152 absl::FixedArray<std::atomic<uint64_t>> protected_buf( in TEST_P()
|
H A D | sequence_lock.h | 22 #include <atomic> 52 // `std::atomic<uint64>`. This is to comply with the C++ standard, which 53 // considers data races on non-atomic objects to be undefined behavior. See "Can 80 bool TryRead(void* dst, const std::atomic<uint64_t>* src, size_t size) const { in TryRead() 99 void Write(std::atomic<uint64_t>* dst, const void* src, size_t size) { in Write() 143 static void RelaxedCopyFromAtomic(void* dst, const std::atomic<uint64_t>* src, in RelaxedCopyFromAtomic() 161 static void RelaxedCopyToAtomic(std::atomic<uint64_t>* dst, const void* src, in RelaxedCopyToAtomic() 180 std::atomic<int64_t> lock_;
|
/third_party/skia/third_party/externals/abseil-cpp/absl/strings/internal/ |
H A D | cord_internal.cc | 16 #include <atomic> 29 ABSL_CONST_INIT std::atomic<bool> cord_btree_enabled(kCordEnableBtreeDefault); 30 ABSL_CONST_INIT std::atomic<bool> cord_ring_buffer_enabled( 32 ABSL_CONST_INIT std::atomic<bool> shallow_subcords_enabled( 34 ABSL_CONST_INIT std::atomic<bool> cord_btree_exhaustive_validation(false);
|