Lines Matching defs:atomic64_t

13 } atomic64_t;
17 #define __ATOMIC64_DECL(sym) void atomic64_##sym(atomic64_t *, ...)
64 static __always_inline s64 arch_atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
70 static __always_inline s64 arch_atomic64_xchg(atomic64_t *v, s64 n)
82 static __always_inline void arch_atomic64_set(atomic64_t *v, s64 i)
91 static __always_inline s64 arch_atomic64_read(const atomic64_t *v)
98 static __always_inline s64 arch_atomic64_add_return(s64 i, atomic64_t *v)
107 static __always_inline s64 arch_atomic64_sub_return(s64 i, atomic64_t *v)
116 static __always_inline s64 arch_atomic64_inc_return(atomic64_t *v)
125 static __always_inline s64 arch_atomic64_dec_return(atomic64_t *v)
134 static __always_inline s64 arch_atomic64_add(s64 i, atomic64_t *v)
142 static __always_inline s64 arch_atomic64_sub(s64 i, atomic64_t *v)
150 static __always_inline void arch_atomic64_inc(atomic64_t *v)
157 static __always_inline void arch_atomic64_dec(atomic64_t *v)
164 static __always_inline int arch_atomic64_add_unless(atomic64_t *v, s64 a, s64 u)
175 static __always_inline int arch_atomic64_inc_not_zero(atomic64_t *v)
184 static __always_inline s64 arch_atomic64_dec_if_positive(atomic64_t *v)
196 static __always_inline void arch_atomic64_and(s64 i, atomic64_t *v)
204 static __always_inline s64 arch_atomic64_fetch_and(s64 i, atomic64_t *v)
215 static __always_inline void arch_atomic64_or(s64 i, atomic64_t *v)
223 static __always_inline s64 arch_atomic64_fetch_or(s64 i, atomic64_t *v)
234 static __always_inline void arch_atomic64_xor(s64 i, atomic64_t *v)
242 static __always_inline s64 arch_atomic64_fetch_xor(s64 i, atomic64_t *v)
253 static __always_inline s64 arch_atomic64_fetch_add(s64 i, atomic64_t *v)