Home
last modified time | relevance | path

Searched refs:atomic (Results 1 - 25 of 600) sorted by relevance

12345678910>>...24

/third_party/libdrm/
H A Dxf86atomic.h31 * Private definitions for atomic operations
42 int atomic; member
45 # define atomic_read(x) ((x)->atomic)
46 # define atomic_set(x, val) ((x)->atomic = (val))
47 # define atomic_inc(x) ((void) __sync_fetch_and_add (&(x)->atomic, 1))
48 # define atomic_inc_return(x) (__sync_add_and_fetch (&(x)->atomic, 1))
49 # define atomic_dec_and_test(x) (__sync_add_and_fetch (&(x)->atomic, -1) == 0)
50 # define atomic_add(x, v) ((void) __sync_add_and_fetch(&(x)->atomic, (v)))
51 # define atomic_dec(x, v) ((void) __sync_sub_and_fetch(&(x)->atomic, (v)))
52 # define atomic_cmpxchg(x, oldv, newv) __sync_val_compare_and_swap (&(x)->atomic, old
63 AO_t atomic; global() member
88 typedef struct { LIBDRM_ATOMIC_TYPE atomic; } atomic_t; global() member
[all...]
/third_party/mesa3d/src/freedreno/ir3/
H A Dir3_a6xx.c36 * atomic instructions (used for both SSBO and image) use a new instruction
93 * SSBO atomic intrinsics
95 * All of the SSBO atomic memory operations read a value from memory,
103 * 1: The offset into the SSBO buffer of the variable that the atomic
105 * 2: The data parameter to the atomic function (i.e. the value to add
113 struct ir3_instruction *atomic, *ibo, *src0, *src1, *data, *dummy; in emit_intrinsic_atomic_ssbo() local
147 atomic = ir3_ATOMIC_B_ADD(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo()
150 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo()
154 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo()
157 atomic in emit_intrinsic_atomic_ssbo()
258 struct ir3_instruction *atomic, *ibo, *src0, *src1, *dummy; emit_intrinsic_atomic_image() local
448 struct ir3_instruction *addr, *atomic, *src1; emit_intrinsic_atomic_global() local
[all...]
H A Dir3_a4xx.c130 * SSBO atomic intrinsics
132 * All of the SSBO atomic memory operations read a value from memory,
140 * 1: The byte offset into the SSBO buffer of the variable that the atomic
142 * 2: The data parameter to the atomic function (i.e. the value to add
155 struct ir3_instruction *atomic; in emit_intrinsic_atomic_ssbo() local
169 atomic = ir3_ATOMIC_S_ADD(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo()
172 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo()
176 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo()
179 atomic = ir3_ATOMIC_S_MAX(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo()
183 atomic in emit_intrinsic_atomic_ssbo()
368 struct ir3_instruction *atomic, *src0, *src1, *src2; emit_intrinsic_atomic_image() local
[all...]
H A Ddisasm-a3xx.c341 OPC(6, OPC_ATOMIC_ADD, atomic.add),
342 OPC(6, OPC_ATOMIC_SUB, atomic.sub),
343 OPC(6, OPC_ATOMIC_XCHG, atomic.xchg),
344 OPC(6, OPC_ATOMIC_INC, atomic.inc),
345 OPC(6, OPC_ATOMIC_DEC, atomic.dec),
346 OPC(6, OPC_ATOMIC_CMPXCHG, atomic.cmpxchg),
347 OPC(6, OPC_ATOMIC_MIN, atomic.min),
348 OPC(6, OPC_ATOMIC_MAX, atomic.max),
349 OPC(6, OPC_ATOMIC_AND, atomic.and),
350 OPC(6, OPC_ATOMIC_OR, atomic
[all...]
/third_party/skia/third_party/externals/tint/src/resolver/
H A Datomics_test.cc31 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F()
36 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local
37 ASSERT_NE(atomic, nullptr); in TEST_F()
38 EXPECT_TRUE(atomic->Type()->Is<sem::I32>()); in TEST_F()
42 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.u32()), in TEST_F()
47 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local
48 ASSERT_NE(atomic, nullptr); in TEST_F()
49 EXPECT_TRUE(atomic->Type()->Is<sem::U32>()); in TEST_F()
53 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F()
67 auto* atomic in TEST_F() local
[all...]
H A Datomics_validation_test.cc30 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F()
37 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F()
46 Global("a", ty.atomic(ty.f32(Source{{12, 34}})), in TEST_F()
50 EXPECT_EQ(r()->error(), "12:34 error: atomic only supports i32 or u32 types"); in TEST_F()
54 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F()
59 "12:34 error: atomic variables must have <storage> or <workgroup> " in TEST_F()
64 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F()
69 "12:34 error: atomic variables must have <storage> or <workgroup> " in TEST_F()
75 Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}); in TEST_F()
80 "error: atomic variable in TEST_F()
[all...]
/third_party/skia/src/gpu/
H A DGrThreadSafePipelineBuilder.h14 #include <atomic>
78 std::atomic<int> fShaderCompilations{0};
80 std::atomic<int> fNumInlineCompilationFailures{0};
81 std::array<std::atomic<int>, kNumProgramCacheResults> fInlineProgramCacheStats = {};
83 std::atomic<int> fNumPreCompilationFailures{0};
84 std::array<std::atomic<int>, kNumProgramCacheResults> fPreProgramCacheStats = {};
86 std::atomic<int> fNumCompilationFailures{0};
87 std::atomic<int> fNumPartialCompilationSuccesses{0};
88 std::atomic<int> fNumCompilationSuccesses{0};
/third_party/skia/third_party/externals/abseil-cpp/absl/base/internal/
H A Dspinlock_wait.h18 // Operations to make atomic transitions on a word, and to allow
22 #include <atomic>
40 // then return the old value of *w. Make any other atomic transitions
42 uint32_t SpinLockWait(std::atomic<uint32_t> *w, int n,
50 void SpinLockWake(std::atomic<uint32_t> *w, bool all);
56 void SpinLockDelay(std::atomic<uint32_t> *w, uint32_t value, int loop,
74 void ABSL_INTERNAL_C_SYMBOL(AbslInternalSpinLockWake)(std::atomic<uint32_t> *w,
77 std::atomic<uint32_t> *w, uint32_t value, int loop,
81 inline void absl::base_internal::SpinLockWake(std::atomic<uint32_t> *w, in SpinLockWake()
87 std::atomic<uint32_ in SpinLockDelay()
[all...]
H A Dthread_identity.h30 #include <atomic>
111 std::atomic<State> state;
150 std::atomic<int>* blocked_count_ptr;
155 std::atomic<int> ticker; // Tick counter, incremented once per second.
156 std::atomic<int> wait_start; // Ticker value when thread started waiting.
157 std::atomic<bool> is_idle; // Has thread become idle yet?
/third_party/node/deps/v8/include/v8-include/cppgc/internal/
H A Dgc-info.h8 #include <atomic>
28 std::atomic<GCInfoIndex>& registered_index) { in EnsureIndex()
39 EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback,
42 std::atomic<GCInfoIndex>&, TraceCallback, FinalizationCallback);
44 std::atomic<GCInfoIndex>&, TraceCallback, NameCallback);
46 EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback);
48 EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, TraceCallback,
51 std::atomic<GCInfoIndex>&, TraceCallback, FinalizationCallback);
53 std::atomic<GCInfoIndex>&, TraceCallback, NameCallback);
55 EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoInde
[all...]
/third_party/node/deps/v8/include/cppgc/internal/
H A Dgc-info.h8 #include <atomic>
27 std::atomic<GCInfoIndex>& registered_index) { in EnsureIndex()
37 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&,
41 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&,
44 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&,
46 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&,
48 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&,
53 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&,
56 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&,
59 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoInde
[all...]
/third_party/skia/third_party/externals/abseil-cpp/absl/container/internal/
H A Dhashtablez_sampler.h42 #include <atomic>
73 std::atomic<size_t> capacity;
74 std::atomic<size_t> size;
75 std::atomic<size_t> num_erases;
76 std::atomic<size_t> num_rehashes;
77 std::atomic<size_t> max_probe_length;
78 std::atomic<size_t> total_probe_length;
79 std::atomic<size_t> hashes_bitwise_or;
80 std::atomic<size_t> hashes_bitwise_and;
81 std::atomic<size_
[all...]
/third_party/node/deps/v8/src/heap/cppgc/
H A Dgc-info.cc24 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic()
33 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic()
42 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic()
50 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexPolymorphic()
57 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic()
66 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic()
75 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic()
83 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexNonPolymorphic()
/third_party/ltp/lib/newlib_tests/
H A Dtest15.c8 * limited check that atomic stores and loads order non-atomic memory
13 * atomic functions. X86 in particular has strong memory ordering by default
15 * (Raspberry Pi 3 Model B) has been observed to fail without the atomic
31 * from removing/reording atomic and seq_n, mark them as volatile.
42 static int atomic; variable
43 /* Instead of storing seq_n on the stack (probably next to the atomic variable
46 * seq_n will be synchronised between processors as a byproduct of the atomic
56 for (i = tst_atomic_load(&atomic); in worker_load_store()
58 i = tst_atomic_load(&atomic)) in worker_load_store()
[all...]
/third_party/skia/third_party/externals/harfbuzz/src/
H A Dhb-atomic.hh81 #include <atomic>
87 #define hb_atomic_int_impl_add(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->fetch_add ((V), std::memory_order_acq_rel))
88 #define hb_atomic_int_impl_set_relaxed(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_relaxed))
89 #define hb_atomic_int_impl_set(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_release))
90 #define hb_atomic_int_impl_get_relaxed(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_relaxed))
91 #define hb_atomic_int_impl_get(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_acquire))
93 #define hb_atomic_ptr_impl_set_relaxed(P, V) (reinterpret_cast<std::atomic<void*> *> (P)->store ((V), std::memory_order_relaxed))
94 #define hb_atomic_ptr_impl_get_relaxed(P) (reinterpret_cast<std::atomic<void*> const *> (P)->load (std::memory_order_relaxed))
95 #define hb_atomic_ptr_impl_get(P) (reinterpret_cast<std::atomic<void*> *> (P)->load (std::memory_order_acquire))
100 return reinterpret_cast<std::atomic<cons in _hb_atomic_ptr_impl_cmplexch()
[all...]
/third_party/skia/third_party/externals/abseil-cpp/absl/time/
H A Dclock.cc25 #include <atomic>
103 // An implementation of reader-write locks that use no atomic ops in the read
115 static inline uint64_t SeqAcquire(std::atomic<uint64_t> *seq) {
130 static inline void SeqRelease(std::atomic<uint64_t> *seq, uint64_t x) {
155 std::atomic<uint64_t> raw_ns{0}; // raw kernel time
156 std::atomic<uint64_t> base_ns{0}; // our estimate of time
157 std::atomic<uint64_t> base_cycles{0}; // cycle counter reading
158 std::atomic<uint64_t> nsscaled_per_cycle{0}; // cycle period
161 std::atomic<uint64_t> min_cycles_per_sample{0};
163 // Same again, but with non-atomic type
[all...]
/third_party/skia/third_party/externals/abseil-cpp/absl/base/
H A Dcall_once.h29 #include <atomic>
49 std::atomic<uint32_t>* ControlWord(absl::once_flag* flag);
92 friend std::atomic<uint32_t>* base_internal::ControlWord(once_flag* flag);
93 std::atomic<uint32_t> control_;
147 void CallOnceImpl(std::atomic<uint32_t>* control, in CallOnceImpl()
188 inline std::atomic<uint32_t>* ControlWord(once_flag* flag) { in ControlWord()
194 std::atomic<uint32_t>* once = base_internal::ControlWord(flag); in LowLevelCallOnce()
207 std::atomic<uint32_t>* once = base_internal::ControlWord(&flag); in call_once()
/third_party/skia/third_party/externals/oboe/src/fifo/
H A DFifoControllerIndirect.h20 #include <atomic>
34 std::atomic<uint64_t> *readCounterAddress,
35 std::atomic<uint64_t> *writeCounterAddress);
59 std::atomic<uint64_t> *mReadCounterAddress;
60 std::atomic<uint64_t> *mWriteCounterAddress;
/third_party/node/deps/v8/include/cppgc/
H A Dmember.h8 #include <atomic>
38 return reinterpret_cast<const std::atomic<const void*>*>(&raw_)->load( in GetRawAtomic()
42 reinterpret_cast<std::atomic<const void*>*>(&raw_)->store( in SetRawAtomic()
50 // results in a non-atomic write on some builds, even when the atomic version
71 // initialize using atomic assignments. This is required for preventing
74 BasicMember(std::nullptr_t, AtomicInitializerTag atomic) in BasicMember() argument
75 : MemberBase(nullptr, atomic) {} in BasicMember()
76 BasicMember(SentinelPointer s, AtomicInitializerTag atomic) in BasicMember() argument
77 : MemberBase(s, atomic) {} in BasicMember()
78 BasicMember(T* raw, AtomicInitializerTag atomic) BasicMember() argument
82 BasicMember(T& raw, AtomicInitializerTag atomic) BasicMember() argument
[all...]
/third_party/icu/icu4c/source/common/
H A Dumutex.h23 #include <atomic>
35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn.
40 // Export an explicit template instantiation of std::atomic<int32_t>.
44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class.
52 template struct U_COMMON_API std::atomic<int32_t>; member in std
53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std
59 template struct std::atomic<int32_t>; member in std
60 template struct std::atomic<std::mutex *>; member in std
73 typedef std::atomic<int32_t> u_atomic_int32_t;
240 std::atomic<st
[all...]
/third_party/node/deps/icu-small/source/common/
H A Dumutex.h23 #include <atomic>
35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn.
40 // Export an explicit template instantiation of std::atomic<int32_t>.
44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class.
52 template struct U_COMMON_API std::atomic<int32_t>; member in std
53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std
59 template struct std::atomic<int32_t>; member in std
60 template struct std::atomic<std::mutex *>; member in std
73 typedef std::atomic<int32_t> u_atomic_int32_t;
240 std::atomic<st
[all...]
/third_party/skia/third_party/externals/icu/source/common/
H A Dumutex.h23 #include <atomic>
35 // Support for including an alternate implementation of atomic & mutex operations has been withdrawn.
40 // Export an explicit template instantiation of std::atomic<int32_t>.
44 // Similar story for std::atomic<std::mutex *>, and the exported UMutex class.
52 template struct U_COMMON_API std::atomic<int32_t>; member in std
53 template struct U_COMMON_API std::atomic<std::mutex *>; member in std
59 template struct std::atomic<int32_t>; member in std
60 template struct std::atomic<std::mutex *>; member in std
73 typedef std::atomic<int32_t> u_atomic_int32_t;
244 std::atomic<st
[all...]
/third_party/skia/third_party/externals/abseil-cpp/absl/flags/internal/
H A Dsequence_lock_test.cc17 #include <atomic>
48 absl::FixedArray<std::atomic<uint64_t>> protected_buf(buf_words); in TEST_P()
52 std::atomic<bool> stop{false}; in TEST_P()
53 std::atomic<int64_t> bad_reads{0}; in TEST_P()
54 std::atomic<int64_t> good_reads{0}; in TEST_P()
55 std::atomic<int64_t> unsuccessful_reads{0}; in TEST_P()
119 // With TSAN, a lot of threads contending for atomic access on the sequence
152 absl::FixedArray<std::atomic<uint64_t>> protected_buf( in TEST_P()
H A Dsequence_lock.h22 #include <atomic>
52 // `std::atomic<uint64>`. This is to comply with the C++ standard, which
53 // considers data races on non-atomic objects to be undefined behavior. See "Can
80 bool TryRead(void* dst, const std::atomic<uint64_t>* src, size_t size) const { in TryRead()
99 void Write(std::atomic<uint64_t>* dst, const void* src, size_t size) { in Write()
143 static void RelaxedCopyFromAtomic(void* dst, const std::atomic<uint64_t>* src, in RelaxedCopyFromAtomic()
161 static void RelaxedCopyToAtomic(std::atomic<uint64_t>* dst, const void* src, in RelaxedCopyToAtomic()
180 std::atomic<int64_t> lock_;
/third_party/skia/third_party/externals/abseil-cpp/absl/strings/internal/
H A Dcord_internal.cc16 #include <atomic>
29 ABSL_CONST_INIT std::atomic<bool> cord_btree_enabled(kCordEnableBtreeDefault);
30 ABSL_CONST_INIT std::atomic<bool> cord_ring_buffer_enabled(
32 ABSL_CONST_INIT std::atomic<bool> shallow_subcords_enabled(
34 ABSL_CONST_INIT std::atomic<bool> cord_btree_exhaustive_validation(false);

Completed in 11 milliseconds

12345678910>>...24