Lines Matching refs:name
30 #define ATOMIC_FETCH_OP(name, mb, op, asm_op, cl...) \
31 static inline int __lse_atomic_fetch_##op##name(int i, atomic_t *v) \
57 #define ATOMIC_OP_ADD_RETURN(name, mb, cl...) \
58 static inline int __lse_atomic_add_return##name(int i, atomic_t *v) \
90 #define ATOMIC_FETCH_OP_AND(name, mb, cl...) \
91 static inline int __lse_atomic_fetch_and##name(int i, atomic_t *v) \
121 #define ATOMIC_OP_SUB_RETURN(name, mb, cl...) \
122 static inline int __lse_atomic_sub_return##name(int i, atomic_t *v) \
145 #define ATOMIC_FETCH_OP_SUB(name, mb, cl...) \
146 static inline int __lse_atomic_fetch_sub##name(int i, atomic_t *v) \
183 #define ATOMIC64_FETCH_OP(name, mb, op, asm_op, cl...) \
184 static inline long __lse_atomic64_fetch_##op##name(s64 i, atomic64_t *v)\
210 #define ATOMIC64_OP_ADD_RETURN(name, mb, cl...) \
211 static inline long __lse_atomic64_add_return##name(s64 i, atomic64_t *v)\
243 #define ATOMIC64_FETCH_OP_AND(name, mb, cl...) \
244 static inline long __lse_atomic64_fetch_and##name(s64 i, atomic64_t *v) \
274 #define ATOMIC64_OP_SUB_RETURN(name, mb, cl...) \
275 static inline long __lse_atomic64_sub_return##name(s64 i, atomic64_t *v)\
298 #define ATOMIC64_FETCH_OP_SUB(name, mb, cl...) \
299 static inline long __lse_atomic64_fetch_sub##name(s64 i, atomic64_t *v) \
340 #define __CMPXCHG_CASE(w, sfx, name, sz, mb, cl...) \
342 __lse__cmpxchg_case_##name##sz(volatile void *ptr, \
383 #define __CMPXCHG_DBL(name, mb, cl...) \
385 __lse__cmpxchg_double##name(unsigned long old1, \