Lines Matching refs:s64
12 s64 __aligned(8) counter;
64 static __always_inline s64 arch_atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
70 static __always_inline s64 arch_atomic64_xchg(atomic64_t *v, s64 n)
72 s64 o;
82 static __always_inline void arch_atomic64_set(atomic64_t *v, s64 i)
91 static __always_inline s64 arch_atomic64_read(const atomic64_t *v)
93 s64 r;
98 static __always_inline s64 arch_atomic64_add_return(s64 i, atomic64_t *v)
107 static __always_inline s64 arch_atomic64_sub_return(s64 i, atomic64_t *v)
116 static __always_inline s64 arch_atomic64_inc_return(atomic64_t *v)
118 s64 a;
125 static __always_inline s64 arch_atomic64_dec_return(atomic64_t *v)
127 s64 a;
134 static __always_inline s64 arch_atomic64_add(s64 i, atomic64_t *v)
142 static __always_inline s64 arch_atomic64_sub(s64 i, atomic64_t *v)
164 static __always_inline int arch_atomic64_add_unless(atomic64_t *v, s64 a, s64 u)
184 static __always_inline s64 arch_atomic64_dec_if_positive(atomic64_t *v)
186 s64 r;
196 static __always_inline void arch_atomic64_and(s64 i, atomic64_t *v)
198 s64 old, c = 0;
204 static __always_inline s64 arch_atomic64_fetch_and(s64 i, atomic64_t *v)
206 s64 old, c = 0;
215 static __always_inline void arch_atomic64_or(s64 i, atomic64_t *v)
217 s64 old, c = 0;
223 static __always_inline s64 arch_atomic64_fetch_or(s64 i, atomic64_t *v)
225 s64 old, c = 0;
234 static __always_inline void arch_atomic64_xor(s64 i, atomic64_t *v)
236 s64 old, c = 0;
242 static __always_inline s64 arch_atomic64_fetch_xor(s64 i, atomic64_t *v)
244 s64 old, c = 0;
253 static __always_inline s64 arch_atomic64_fetch_add(s64 i, atomic64_t *v)
255 s64 old, c = 0;