/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_persistent_allocator.h | 37 uptr cmp = atomic_load(®ion_pos, memory_order_acquire); 38 uptr end = atomic_load(®ion_end, memory_order_acquire); 41 memory_order_acquire))
|
H A D | sanitizer_mutex.h | 35 return atomic_exchange(&state_, 1, memory_order_acquire) == 0; 56 && atomic_exchange(&state_, 1, memory_order_acquire) == 0) 110 memory_order_acquire)) 122 u32 prev = atomic_fetch_add(&state_, kReadLock, memory_order_acquire); 157 memory_order_acquire)) 168 u32 prev = atomic_load(&state_, memory_order_acquire);
|
H A D | sanitizer_lfstack.h | 48 u64 cmp = atomic_load(&head_, memory_order_acquire); 57 memory_order_acquire))
|
H A D | sanitizer_libignore.h | 91 const uptr n = atomic_load(&ignored_ranges_count_, memory_order_acquire); 105 const uptr n = atomic_load(&instrumented_ranges_count_, memory_order_acquire);
|
H A D | sanitizer_atomic.h | 23 memory_order_acquire = 1 << 2, enumerator in enum:__sanitizer::memory_order
|
H A D | sanitizer_atomic_clang_other.h | 28 | memory_order_acquire | memory_order_seq_cst)); 42 } else if (mo == memory_order_acquire) {
|
H A D | sanitizer_atomic_clang_x86.h | 30 | memory_order_acquire | memory_order_seq_cst)); 44 } else if (mo == memory_order_acquire) {
|
H A D | sanitizer_allocator_bytemap.h | 87 atomic_load(&map1_[idx], memory_order_acquire));
|
H A D | sanitizer_rtems.cpp | 121 if (atomic_exchange(m, MtxLocked, memory_order_acquire) == MtxUnlocked) 123 while (atomic_exchange(m, MtxSleeping, memory_order_acquire) != MtxUnlocked) {
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/xray/ |
H A D | xray_init.cpp | 66 if (atomic_load(&XRayInitialized, memory_order_acquire)) 72 if (!atomic_load(&XRayFlagsInitialized, memory_order_acquire)) {
|
H A D | xray_interface.cpp | 181 memory_order_acquire)) 232 memory_order_acquire)) 374 memory_order_acquire)) { 387 memory_order_acquire)) { 399 memory_order_acquire)) { 450 memory_order_acquire))
|
H A D | xray_buffer_queue.h | 233 return atomic_load(&Finalizing, memory_order_acquire); 237 return atomic_load(&Generation, memory_order_acquire);
|
H A D | xray_basic_logging.cpp | 94 if (atomic_load(&UseRealTSC, memory_order_acquire)) 106 Header.CycleFrequency = atomic_load(&CycleFrequency, memory_order_acquire); 424 atomic_load(&TicksPerSec, memory_order_acquire) * 427 __xray_set_handler_arg1(atomic_load(&UseRealTSC, memory_order_acquire) 430 __xray_set_handler(atomic_load(&UseRealTSC, memory_order_acquire)
|
H A D | xray_buffer_queue.cpp | 162 if (atomic_load(&Finalizing, memory_order_acquire)) 212 atomic_store(B->Buff.Extents, atomic_load(Buf.Extents, memory_order_acquire),
|
H A D | xray_fdr_logging.cpp | 252 atomic_thread_fence(memory_order_acquire); 253 auto BufferSize = atomic_load(It->Extents, memory_order_acquire); 280 if (atomic_load(&LoggingStatus, memory_order_acquire) != 367 auto BufferExtents = atomic_load(B.Extents, memory_order_acquire); 449 auto Status = atomic_load(&LoggingStatus, memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/Support/ |
H A D | ManagedStatic.h | 87 void *Tmp = Ptr.load(std::memory_order_acquire); 97 void *Tmp = Ptr.load(std::memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/lldb/source/Utility/ |
H A D | Timer.cpp | 132 uint64_t nanos = i->m_nanos.load(std::memory_order_acquire); 134 uint64_t nanos_total = i->m_nanos_total.load(std::memory_order_acquire); 135 uint64_t count = i->m_count.load(std::memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
H A D | atomic_helpers.h | 19 memory_order_acquire = 2, enumerator in enum:scudo::memory_order 26 static_assert(memory_order_acquire == __ATOMIC_ACQUIRE, ""); 133 atomic_compare_exchange_strong(A, &Cmp, Xchg, memory_order_acquire);
|
H A D | linux.cpp | 102 V = atomic_exchange(&M, Sleeping, memory_order_acquire); 106 V = atomic_exchange(&M, Sleeping, memory_order_acquire);
|
H A D | bytemap.h | 95 atomic_load(&Level1Map[Index], memory_order_acquire));
|
H A D | tsd_exclusive.h | 50 !atomic_load(&Disabled, memory_order_acquire))) {
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_mutex.cpp | 232 memory_order_acquire)) 238 memory_order_acquire)) { 261 uptr prev = atomic_fetch_add(&state_, kReadLock, memory_order_acquire); 265 prev = atomic_load(&state_, memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_stack.cpp | 26 return atomic_load(&malloc_context_size, memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/ADT/ |
H A D | Statistic.h | 125 if (!Initialized.load(std::memory_order_acquire))
|
/freebsd-11-stable/contrib/llvm-project/libcxx/src/experimental/ |
H A D | memory_resource.cpp | 123 &__res, memory_order_acquire);
|