/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/ADT/ |
H A D | Statistic.h | 69 unsigned getValue() const { return Value.load(std::memory_order_relaxed); } 75 Value.store(Val, std::memory_order_relaxed); 80 Value.fetch_add(1, std::memory_order_relaxed); 86 return Value.fetch_add(1, std::memory_order_relaxed); 90 Value.fetch_sub(1, std::memory_order_relaxed); 96 return Value.fetch_sub(1, std::memory_order_relaxed); 102 Value.fetch_add(V, std::memory_order_relaxed); 109 Value.fetch_sub(V, std::memory_order_relaxed); 114 unsigned PrevMax = Value.load(std::memory_order_relaxed); 118 PrevMax, V, std::memory_order_relaxed)) { [all...] |
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_atomic.h | 21 memory_order_relaxed = 1 << 0, enumerator in enum:__sanitizer::memory_order 76 return atomic_load(a, memory_order_relaxed); 81 atomic_store(a, v, memory_order_relaxed);
|
H A D | sanitizer_allocator_stats.h | 34 v += atomic_load(&stats_[i], memory_order_relaxed); 35 atomic_store(&stats_[i], v, memory_order_relaxed); 39 v = atomic_load(&stats_[i], memory_order_relaxed) - v; 40 atomic_store(&stats_[i], v, memory_order_relaxed); 44 atomic_store(&stats_[i], v, memory_order_relaxed); 48 return atomic_load(&stats_[i], memory_order_relaxed);
|
H A D | sanitizer_atomic_clang_other.h | 27 DCHECK(mo & (memory_order_relaxed | memory_order_consume 34 if (mo == memory_order_relaxed) { 64 DCHECK(mo & (memory_order_relaxed | memory_order_release 70 if (mo == memory_order_relaxed) {
|
H A D | sanitizer_atomic_clang_x86.h | 29 DCHECK(mo & (memory_order_relaxed | memory_order_consume 36 if (mo == memory_order_relaxed) { 74 DCHECK(mo & (memory_order_relaxed | memory_order_release 80 if (mo == memory_order_relaxed) {
|
H A D | sanitizer_lfstack.h | 28 atomic_store(&head_, 0, memory_order_relaxed); 32 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; 36 u64 cmp = atomic_load(&head_, memory_order_relaxed);
|
H A D | sanitizer_mutex.h | 25 atomic_store(&state_, 0, memory_order_relaxed); 43 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1); 55 if (atomic_load(&state_, memory_order_relaxed) == 0 100 atomic_store(&state_, kUnlocked, memory_order_relaxed); 104 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 136 CHECK_NE(atomic_load(&state_, memory_order_relaxed), kUnlocked); 154 u32 cmp = atomic_load(&state_, memory_order_relaxed);
|
H A D | sanitizer_atomic_clang_mips.h | 44 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst)); 70 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst)); 93 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst)); 106 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst));
|
H A D | sanitizer_addrhashmap.h | 195 if (atomic_load(&b->add, memory_order_relaxed)) { 197 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); 200 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 216 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 228 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); 232 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 256 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); 271 atomic_store(&b->add, (uptr)add, memory_order_relaxed); 283 atomic_store(&b->add, (uptr)add1, memory_order_relaxed); 289 CHECK_EQ(atomic_load(&c->addr, memory_order_relaxed), [all...] |
H A D | sanitizer_stackdepot.cpp | 40 atomic_load(&hash_and_use_count, memory_order_relaxed) & kHashMask; 61 atomic_store(&hash_and_use_count, hash & kHashMask, memory_order_relaxed); 78 return atomic_load(&node_->hash_and_use_count, memory_order_relaxed) & 83 atomic_fetch_add(&node_->hash_and_use_count, 1, memory_order_relaxed) &
|
H A D | sanitizer_tls_get_addr.cpp | 49 atomic_fetch_sub(&number_of_live_dtls, 1, memory_order_relaxed); 59 atomic_fetch_add(&number_of_live_dtls, 1, memory_order_relaxed); 102 atomic_load(&number_of_live_dtls, memory_order_relaxed));
|
H A D | sanitizer_libignore.cpp | 82 atomic_load(&ignored_ranges_count_, memory_order_relaxed); 111 atomic_load(&instrumented_ranges_count_, memory_order_relaxed);
|
H A D | sanitizer_persistent_allocator.h | 55 atomic_store(®ion_pos, 0, memory_order_relaxed);
|
H A D | sanitizer_atomic_msvc.h | 73 DCHECK(mo & (memory_order_relaxed | memory_order_consume 78 if (mo == memory_order_relaxed) { 90 DCHECK(mo & (memory_order_relaxed | memory_order_release 94 if (mo == memory_order_relaxed) {
|
H A D | sanitizer_termination.cpp | 72 if (atomic_fetch_add(&num_calls, 1, memory_order_relaxed) > 10) {
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_external.cpp | 31 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; 50 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); 61 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); 76 uptr new_tag = atomic_fetch_add(&used_tags, 1, memory_order_relaxed); 99 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed));
|
H A D | tsan_fd.cpp | 52 atomic_store(&s->rc, 1, memory_order_relaxed); 57 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) 58 atomic_fetch_add(&s->rc, 1, memory_order_relaxed); 63 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) { 123 atomic_store(&fdctx.globsync.rc, (u64)-1, memory_order_relaxed); 124 atomic_store(&fdctx.filesync.rc, (u64)-1, memory_order_relaxed); 125 atomic_store(&fdctx.socksync.rc, (u64)-1, memory_order_relaxed); 133 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); 145 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
|
H A D | tsan_mutex.cpp | 219 atomic_store(&state_, kUnlocked, memory_order_relaxed); 223 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 235 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) { 286 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
H A D | atomic_helpers.h | 17 memory_order_relaxed = 0, enumerator in enum:scudo::memory_order 24 static_assert(memory_order_relaxed == __ATOMIC_RELAXED, ""); 121 return atomic_load(A, memory_order_relaxed); 126 atomic_store(A, V, memory_order_relaxed);
|
/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/Support/ |
H A D | ManagedStatic.h | 91 return *static_cast<C *>(Ptr.load(std::memory_order_relaxed)); 101 return *static_cast<C *>(Ptr.load(std::memory_order_relaxed));
|
/freebsd-11-stable/contrib/llvm-project/lldb/source/Utility/ |
H A D | Log.cpp | 91 uint32_t mask = m_mask.fetch_or(flags, std::memory_order_relaxed); 93 m_options.store(options, std::memory_order_relaxed); 95 m_channel.log_ptr.store(this, std::memory_order_relaxed); 102 uint32_t mask = m_mask.fetch_and(~flags, std::memory_order_relaxed); 105 m_channel.log_ptr.store(nullptr, std::memory_order_relaxed); 110 return m_options.load(std::memory_order_relaxed); 114 return m_mask.load(std::memory_order_relaxed); 276 return m_options.load(std::memory_order_relaxed) & LLDB_LOG_OPTION_VERBOSE; 354 c.second.m_channel.log_ptr.store(nullptr, std::memory_order_relaxed);
|
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Support/ |
H A D | PrettyStackTrace.cpp | 186 GlobalSigInfoGenerationCounter.load(std::memory_order_relaxed); 271 GlobalSigInfoGenerationCounter.fetch_add(1, std::memory_order_relaxed); 279 GlobalSigInfoGenerationCounter.load(std::memory_order_relaxed);
|
H A D | ManagedStatic.cpp | 39 if (!Ptr.load(std::memory_order_relaxed)) {
|
/freebsd-11-stable/contrib/netbsd-tests/lib/libc/sync/ |
H A D | cpp_atomic_ops_linkable.cc | 54 std::memory_order_release, std::memory_order_relaxed);
|
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_thread.h | 110 return !atomic_load(&stack_switching_, memory_order_relaxed) && 117 if (atomic_load(&stack_switching_, memory_order_relaxed))
|