Lines Matching refs:a64
34 typedef unsigned long long a64; // NOLINT
216 static atomic_uint64_t *to_atomic(const volatile a64 *a) {
217 return reinterpret_cast<atomic_uint64_t *>(const_cast<a64 *>(a));
537 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo) {
564 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo) {
591 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo) {
618 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) {
645 a64 __tsan_atomic64_fetch_sub(volatile a64 *a, a64 v, morder mo) {
672 a64 __tsan_atomic64_fetch_and(volatile a64 *a, a64 v, morder mo) {
699 a64 __tsan_atomic64_fetch_or(volatile a64 *a, a64 v, morder mo) {
726 a64 __tsan_atomic64_fetch_xor(volatile a64 *a, a64 v, morder mo) {
753 a64 __tsan_atomic64_fetch_nand(volatile a64 *a, a64 v, morder mo) {
783 int __tsan_atomic64_compare_exchange_strong(volatile a64 *a, a64 *c, a64 v,
815 int __tsan_atomic64_compare_exchange_weak(volatile a64 *a, a64 *c, a64 v,
847 a64 __tsan_atomic64_compare_exchange_val(volatile a64 *a, a64 c, a64 v,
903 ATOMIC_RET(Load, *(a64*)(a+8), *(a64**)a, mo_acquire);
913 ATOMIC(Store, *(a64**)a, *(a64*)(a+8), mo_release);
923 ATOMIC_RET(FetchAdd, *(a64*)(a+16), *(a64**)a, *(a64*)(a+8), mo_acq_rel);
933 ATOMIC_RET(Exchange, *(a64*)(a+16), *(a64**)a, *(a64*)(a+8), mo_acq_rel);
948 a64 cur = 0;
949 a64 cmp = *(a64*)(a+8);
950 ATOMIC_RET(CAS, cur, *(a64**)a, cmp, *(a64*)(a+16), mo_acq_rel, mo_acquire);