Lines Matching refs:ret
30 static inline int CRYPTO_UP_REF(_Atomic int *val, int *ret,
33 *ret = atomic_fetch_add_explicit(val, 1, memory_order_relaxed) + 1;
47 static inline int CRYPTO_DOWN_REF(_Atomic int *val, int *ret,
50 *ret = atomic_fetch_sub_explicit(val, 1, memory_order_relaxed) - 1;
51 if (*ret == 0)
62 static __inline__ int CRYPTO_UP_REF(int *val, int *ret, ossl_unused void *lock)
64 *ret = __atomic_fetch_add(val, 1, __ATOMIC_RELAXED) + 1;
68 static __inline__ int CRYPTO_DOWN_REF(int *val, int *ret,
71 *ret = __atomic_fetch_sub(val, 1, __ATOMIC_RELAXED) - 1;
72 if (*ret == 0)
80 static __inline int CRYPTO_UP_REF(volatile int *val, int *ret,
83 *ret = _InterlockedExchangeAdd((void *)val, 1) + 1;
87 static __inline int CRYPTO_DOWN_REF(volatile int *val, int *ret,
90 *ret = _InterlockedExchangeAdd((void *)val, -1) - 1;
106 static __inline int CRYPTO_UP_REF(volatile int *val, int *ret,
109 *ret = _InterlockedExchangeAdd_nf(val, 1) + 1;
113 static __inline int CRYPTO_DOWN_REF(volatile int *val, int *ret,
116 *ret = _InterlockedExchangeAdd_nf(val, -1) - 1;
117 if (*ret == 0)
134 static __inline int CRYPTO_UP_REF(volatile int *val, int *ret,
137 *ret = _InterlockedExchangeAdd((long volatile *)val, 1) + 1;
141 static __inline int CRYPTO_DOWN_REF(volatile int *val, int *ret,
144 *ret = _InterlockedExchangeAdd((long volatile *)val, -1) - 1;
161 # define CRYPTO_UP_REF(val, ret, lock) CRYPTO_atomic_add(val, 1, ret, lock)
162 # define CRYPTO_DOWN_REF(val, ret, lock) CRYPTO_atomic_add(val, -1, ret, lock)