Lines Matching refs:ret
32 static inline int CRYPTO_UP_REF(_Atomic int *val, int *ret, void *lock)
34 *ret = atomic_fetch_add_explicit(val, 1, memory_order_relaxed) + 1;
48 static inline int CRYPTO_DOWN_REF(_Atomic int *val, int *ret, void *lock)
50 *ret = atomic_fetch_sub_explicit(val, 1, memory_order_relaxed) - 1;
51 if (*ret == 0)
62 static __inline__ int CRYPTO_UP_REF(int *val, int *ret, void *lock)
64 *ret = __atomic_fetch_add(val, 1, __ATOMIC_RELAXED) + 1;
68 static __inline__ int CRYPTO_DOWN_REF(int *val, int *ret, void *lock)
70 *ret = __atomic_fetch_sub(val, 1, __ATOMIC_RELAXED) - 1;
71 if (*ret == 0)
88 static __inline int CRYPTO_UP_REF(volatile int *val, int *ret, void *lock)
90 *ret = _InterlockedExchangeAdd_nf(val, 1) + 1;
94 static __inline int CRYPTO_DOWN_REF(volatile int *val, int *ret, void *lock)
96 *ret = _InterlockedExchangeAdd_nf(val, -1) - 1;
97 if (*ret == 0)
114 static __inline int CRYPTO_UP_REF(volatile int *val, int *ret, void *lock)
116 *ret = _InterlockedExchangeAdd(val, 1) + 1;
120 static __inline int CRYPTO_DOWN_REF(volatile int *val, int *ret, void *lock)
122 *ret = _InterlockedExchangeAdd(val, -1) - 1;
131 # define CRYPTO_UP_REF(val, ret, lock) CRYPTO_atomic_add(val, 1, ret, lock)
132 # define CRYPTO_DOWN_REF(val, ret, lock) CRYPTO_atomic_add(val, -1, ret, lock)