/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/xray/ |
H A D | xray_profiling.cpp | 208 if (atomic_load(&ProfilerLogStatus, memory_order_acquire) != 279 auto Status = atomic_load(&ProfilerLogStatus, memory_order_acquire);
|
H A D | xray_profile_collector.cpp | 115 if (!atomic_load(&CollectorInitialized, memory_order_acquire)) { 233 if (!atomic_load(&CollectorInitialized, memory_order_acquire))
|
H A D | xray_log_interface.cpp | 201 atomic_load(&XRayBufferIterator, memory_order_acquire));
|
H A D | xray_interface.cpp | 180 if (!atomic_load(&XRayInitialized, 231 if (!atomic_load(&XRayInitialized, 373 if (atomic_load(&XRayInitialized, 386 if (atomic_load(&XRayInitialized, 398 if (atomic_load(&XRayInitialized, 449 if (!atomic_load(&XRayInitialized,
|
H A D | xray_init.cpp | 66 if (atomic_load(&XRayInitialized, memory_order_acquire)) 72 if (!atomic_load(&XRayFlagsInitialized, memory_order_acquire)) {
|
H A D | xray_fdr_logging.cpp | 253 auto BufferSize = atomic_load(It->Extents, memory_order_acquire); 280 if (atomic_load(&LoggingStatus, memory_order_acquire) != 367 auto BufferExtents = atomic_load(B.Extents, memory_order_acquire); 449 auto Status = atomic_load(&LoggingStatus, memory_order_acquire);
|
H A D | xray_buffer_queue.cpp | 162 if (atomic_load(&Finalizing, memory_order_acquire)) 212 atomic_store(B->Buff.Extents, atomic_load(Buf.Extents, memory_order_acquire),
|
H A D | xray_basic_logging.cpp | 94 if (atomic_load(&UseRealTSC, memory_order_acquire)) 106 Header.CycleFrequency = atomic_load(&CycleFrequency, memory_order_acquire); 219 if (Delta < atomic_load(&ThresholdTicks, memory_order_relaxed)) { 392 if (!atomic_load(&UseRealTSC, memory_order_relaxed) && Verbosity()) 424 atomic_load(&TicksPerSec, memory_order_acquire) * 427 __xray_set_handler_arg1(atomic_load(&UseRealTSC, memory_order_acquire) 430 __xray_set_handler(atomic_load(&UseRealTSC, memory_order_acquire)
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_rtl.cpp | 178 atomic_load(&ctx->stop_background_thread, memory_order_relaxed) == 0; 213 u64 last = atomic_load(&ctx->last_symbolize_time_ns, 433 if (atomic_load(&bg_thread, memory_order_relaxed) == 0 && 607 u64 raw = atomic_load((atomic_uint64_t*)p, memory_order_relaxed);
|
H A D | tsan_mutex.cpp | 223 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 235 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) { 265 prev = atomic_load(&state_, memory_order_acquire); 286 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
|
H A D | tsan_interface_atomic.cpp | 211 return atomic_load(to_atomic(a), to_mo(mo));
|
H A D | tsan_interceptors_posix.cpp | 325 if (atomic_load(&ctx->have_pending_signals, memory_order_relaxed) == 0) 502 atomic_load(&sctx->in_blocking_func, memory_order_relaxed) : 504 buf->in_signal_handler = atomic_load(&thr->in_signal_handler, 863 u32 cmp = atomic_load(g, memory_order_acquire); 945 while ((tid = atomic_load(&p->tid, memory_order_acquire)) == 0) 1009 while (atomic_load(&p.tid, memory_order_acquire) != 0) 1106 uptr cond = atomic_load(p, memory_order_acquire); 1132 CHECK_EQ(atomic_load(&ctx->in_blocking_func, memory_order_relaxed), 1); 1439 u32 v = atomic_load(a, memory_order_acquire); 1449 v = atomic_load( [all...] |
H A D | tsan_interceptors_libdispatch.cpp | 265 u32 v = atomic_load(a, memory_order_acquire); 276 v = atomic_load(a, memory_order_acquire);
|
H A D | tsan_fd.cpp | 57 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) 63 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) { 77 uptr l1 = atomic_load(pl1, memory_order_consume); 133 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); 145 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
|
H A D | tsan_external.cpp | 31 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; 50 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); 61 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); 99 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed));
|
H A D | tsan_clock.cpp | 93 u32 v = atomic_load(ref, memory_order_acquire); 301 if (atomic_load(ref, memory_order_acquire) == 1) 460 u32 v = atomic_load(ref, memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/tsan/dd/ |
H A D | dd_interceptors.cpp | 163 uptr cond = atomic_load(p, memory_order_acquire);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
H A D | tsd_exclusive.h | 50 !atomic_load(&Disabled, memory_order_acquire))) {
|
H A D | bytemap.h | 95 atomic_load(&Level1Map[Index], memory_order_acquire));
|
H A D | atomic_helpers.h | 63 inline typename T::Type atomic_load(const volatile T *A, memory_order MO) { function in namespace:scudo 121 return atomic_load(A, memory_order_relaxed);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_tls_get_addr.cpp | 102 atomic_load(&number_of_live_dtls, memory_order_relaxed));
|
H A D | sanitizer_thread_registry.cpp | 52 return !!atomic_load(&thread_destroyed, memory_order_acquire);
|
H A D | sanitizer_stoptheworld_netbsd_libcdep.cpp | 300 while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
|
H A D | sanitizer_stoptheworld_linux_libcdep.cpp | 455 while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
|
H A D | sanitizer_stackdepot.cpp | 40 atomic_load(&hash_and_use_count, memory_order_relaxed) & kHashMask; 78 return atomic_load(&node_->hash_and_use_count, memory_order_relaxed) & 128 uptr v = atomic_load(p, memory_order_consume);
|