/freebsd-13-stable/tools/regression/include/stdatomic/ |
H A D | logic.c | 59 T v1 = atomic_load(a); \ 62 assert(atomic_load(a) == (T)(result)); \ 66 T v1 = atomic_load(a); \ 73 assert(atomic_load(a) == v3); \ 80 assert(atomic_load(a) == v1); \ 85 assert(atomic_load(a) == v1); \
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_lfstack.h | 32 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; 36 u64 cmp = atomic_load(&head_, memory_order_relaxed); 48 u64 cmp = atomic_load(&head_, memory_order_acquire);
|
H A D | sanitizer_mutex.h | 43 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1); 55 if (atomic_load(&state_, memory_order_relaxed) == 0 104 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 136 CHECK_NE(atomic_load(&state_, memory_order_relaxed), kUnlocked); 154 u32 cmp = atomic_load(&state_, memory_order_relaxed); 168 u32 prev = atomic_load(&state_, memory_order_acquire);
|
H A D | sanitizer_libignore.h | 91 const uptr n = atomic_load(&ignored_ranges_count_, memory_order_acquire); 105 const uptr n = atomic_load(&instrumented_ranges_count_, memory_order_acquire);
|
H A D | sanitizer_persistent_allocator.h | 37 uptr cmp = atomic_load(®ion_pos, memory_order_acquire); 38 uptr end = atomic_load(®ion_end, memory_order_acquire);
|
H A D | sanitizer_addrhashmap.h | 187 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); 195 if (atomic_load(&b->add, memory_order_relaxed)) { 197 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); 200 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 216 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 228 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); 232 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 256 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 289 CHECK_EQ(atomic_load(&c->addr, memory_order_relaxed), 0); 300 uptr addr1 = atomic_load( [all...] |
H A D | sanitizer_allocator_stats.h | 34 v += atomic_load(&stats_[i], memory_order_relaxed); 39 v = atomic_load(&stats_[i], memory_order_relaxed) - v; 48 return atomic_load(&stats_[i], memory_order_relaxed);
|
H A D | sanitizer_atomic.h | 76 return atomic_load(a, memory_order_relaxed);
|
H A D | sanitizer_stackdepot.cpp | 40 atomic_load(&hash_and_use_count, memory_order_relaxed) & kHashMask; 78 return atomic_load(&node_->hash_and_use_count, memory_order_relaxed) & 128 uptr v = atomic_load(p, memory_order_consume);
|
H A D | sanitizer_libignore.cpp | 82 atomic_load(&ignored_ranges_count_, memory_order_relaxed); 111 atomic_load(&instrumented_ranges_count_, memory_order_relaxed);
|
H A D | sanitizer_stackdepotbase.h | 77 uptr cmp = atomic_load(p, memory_order_relaxed); 103 uptr v = atomic_load(p, memory_order_consume); 148 uptr v = atomic_load(p, memory_order_consume); 170 uptr s = atomic_load(p, memory_order_relaxed);
|
H A D | sanitizer_atomic_clang_other.h | 25 INLINE typename T::Type atomic_load( function in namespace:__sanitizer
|
H A D | sanitizer_atomic_clang_x86.h | 27 INLINE typename T::Type atomic_load( function in namespace:__sanitizer
|
H A D | sanitizer_allocator.cpp | 97 if (atomic_load(&internal_allocator_initialized, memory_order_acquire) == 0) { 99 if (atomic_load(&internal_allocator_initialized, memory_order_relaxed) == 254 return atomic_load(&allocator_may_return_null, memory_order_relaxed);
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/xray/ |
H A D | xray_init.cpp | 66 if (atomic_load(&XRayInitialized, memory_order_acquire)) 72 if (!atomic_load(&XRayFlagsInitialized, memory_order_acquire)) {
|
H A D | xray_basic_logging.cpp | 94 if (atomic_load(&UseRealTSC, memory_order_acquire)) 106 Header.CycleFrequency = atomic_load(&CycleFrequency, memory_order_acquire); 219 if (Delta < atomic_load(&ThresholdTicks, memory_order_relaxed)) { 392 if (!atomic_load(&UseRealTSC, memory_order_relaxed) && Verbosity()) 424 atomic_load(&TicksPerSec, memory_order_acquire) * 427 __xray_set_handler_arg1(atomic_load(&UseRealTSC, memory_order_acquire) 430 __xray_set_handler(atomic_load(&UseRealTSC, memory_order_acquire)
|
H A D | xray_interface.cpp | 207 if (!atomic_load(&XRayInitialized, 258 if (!atomic_load(&XRayInitialized, 403 if (atomic_load(&XRayInitialized, 416 if (atomic_load(&XRayInitialized, 428 if (atomic_load(&XRayInitialized, 479 if (!atomic_load(&XRayInitialized,
|
H A D | xray_buffer_queue.h | 233 return atomic_load(&Finalizing, memory_order_acquire); 237 return atomic_load(&Generation, memory_order_acquire);
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_external.cpp | 31 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; 50 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); 61 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); 99 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed));
|
H A D | tsan_mutex.cpp | 223 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 235 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) { 265 prev = atomic_load(&state_, memory_order_acquire); 286 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
|
H A D | tsan_fd.cpp | 57 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) 63 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) { 77 uptr l1 = atomic_load(pl1, memory_order_consume); 133 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); 145 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_stack.cpp | 26 return atomic_load(&malloc_context_size, memory_order_acquire);
|
H A D | asan_thread.h | 110 return !atomic_load(&stack_switching_, memory_order_relaxed) && 117 if (atomic_load(&stack_switching_, memory_order_relaxed))
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
H A D | atomic_helpers.h | 63 inline typename T::Type atomic_load(const volatile T *A, memory_order MO) { function in namespace:scudo 121 return atomic_load(A, memory_order_relaxed);
|
/freebsd-13-stable/contrib/llvm-project/compiler-rt/lib/dfsan/ |
H A D | dfsan.cpp | 204 label = atomic_load(table_ent, memory_order_acquire); 343 atomic_load(&__dfsan_last_label, memory_order_relaxed); 351 atomic_load(&__dfsan_last_label, memory_order_relaxed);
|