/kernel/linux/linux-5.10/tools/include/linux/ |
H A D | atomic.h | 7 /* atomic_cmpxchg_relaxed */ 8 #ifndef atomic_cmpxchg_relaxed 9 #define atomic_cmpxchg_relaxed atomic_cmpxchg macro 11 #endif /* atomic_cmpxchg_relaxed */
|
H A D | refcount.h | 89 old = atomic_cmpxchg_relaxed(&r->refs, val, new); in refcount_inc_not_zero()
|
/kernel/linux/linux-6.6/tools/include/linux/ |
H A D | atomic.h | 9 /* atomic_cmpxchg_relaxed */ 10 #ifndef atomic_cmpxchg_relaxed 11 #define atomic_cmpxchg_relaxed atomic_cmpxchg macro 13 #endif /* atomic_cmpxchg_relaxed */
|
H A D | refcount.h | 89 old = atomic_cmpxchg_relaxed(&r->refs, val, new); in refcount_inc_not_zero()
|
/kernel/linux/linux-5.10/arch/arm/include/asm/ |
H A D | atomic.h | 106 static inline int atomic_cmpxchg_relaxed(atomic_t *ptr, int old, int new) in atomic_cmpxchg_relaxed() function 126 #define atomic_cmpxchg_relaxed atomic_cmpxchg_relaxed macro
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | atomic-fallback.h | 956 #define arch_atomic_cmpxchg_relaxed atomic_cmpxchg_relaxed 958 #ifndef atomic_cmpxchg_relaxed 961 #define atomic_cmpxchg_relaxed atomic_cmpxchg macro 962 #else /* atomic_cmpxchg_relaxed */ 968 int ret = atomic_cmpxchg_relaxed(v, old, new); in atomic_cmpxchg_acquire() 980 return atomic_cmpxchg_relaxed(v, old, new); in atomic_cmpxchg_release() 991 ret = atomic_cmpxchg_relaxed(v, old, new); in atomic_cmpxchg() 998 #endif /* atomic_cmpxchg_relaxed */ 1056 r = atomic_cmpxchg_relaxed(v, o, new); in atomic_try_cmpxchg_relaxed()
|
/kernel/linux/linux-5.10/arch/riscv/include/asm/ |
H A D | atomic.h | 305 #define atomic_cmpxchg_relaxed atomic_cmpxchg_relaxed macro
|
/kernel/linux/linux-5.10/kernel/locking/ |
H A D | qspinlock.c | 231 old = atomic_cmpxchg_relaxed(&lock->val, val, new); in xchg_tail() 503 * atomic_cmpxchg_relaxed() calls will be safe. in queued_spin_lock_slowpath()
|
/kernel/linux/linux-6.6/kernel/locking/ |
H A D | qspinlock.c | 232 old = atomic_cmpxchg_relaxed(&lock->val, val, new); in xchg_tail() 506 * atomic_cmpxchg_relaxed() calls will be safe. in queued_spin_lock_slowpath()
|
/kernel/linux/linux-5.10/include/asm-generic/ |
H A D | atomic-instrumented.h | 682 atomic_cmpxchg_relaxed(atomic_t *v, int old, int new) in atomic_cmpxchg_relaxed() function 687 #define atomic_cmpxchg_relaxed atomic_cmpxchg_relaxed macro
|
H A D | atomic-long.h | 925 return atomic_cmpxchg_relaxed(v, old, new); in atomic_long_cmpxchg_relaxed()
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | atomic.h | 186 #define atomic_cmpxchg_relaxed(v, o, n) \ macro
|
/kernel/linux/linux-5.10/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3.c | 454 } while (atomic_cmpxchg_relaxed(&cmdq->lock, val, val + 1) != val); in arm_smmu_cmdq_shared_lock() 475 __ret = !atomic_cmpxchg_relaxed(&cmdq->lock, 0, INT_MIN); \
|
/kernel/linux/linux-6.6/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3.c | 459 } while (atomic_cmpxchg_relaxed(&cmdq->lock, val, val + 1) != val); in arm_smmu_cmdq_shared_lock() 480 __ret = !atomic_cmpxchg_relaxed(&cmdq->lock, 0, INT_MIN); \
|
/kernel/linux/linux-6.6/include/linux/atomic/ |
H A D | atomic-instrumented.h | 1238 * atomic_cmpxchg_relaxed() - atomic compare and exchange with relaxed ordering 1250 atomic_cmpxchg_relaxed(atomic_t *v, int old, int new) in atomic_cmpxchg_relaxed() function
|
/kernel/linux/linux-6.6/kernel/ |
H A D | workqueue.c | 1632 tmp = atomic_cmpxchg_relaxed(&nna->nr, old, old + 1); in tryinc_node_nr_active()
|