Searched refs:memory_order_relaxed (Results 1 - 25 of 83) sorted by relevance

1234

/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/ADT/
H A DStatistic.h69 unsigned getValue() const { return Value.load(std::memory_order_relaxed); }
75 Value.store(Val, std::memory_order_relaxed);
80 Value.fetch_add(1, std::memory_order_relaxed);
86 return Value.fetch_add(1, std::memory_order_relaxed);
90 Value.fetch_sub(1, std::memory_order_relaxed);
96 return Value.fetch_sub(1, std::memory_order_relaxed);
102 Value.fetch_add(V, std::memory_order_relaxed);
109 Value.fetch_sub(V, std::memory_order_relaxed);
114 unsigned PrevMax = Value.load(std::memory_order_relaxed);
118 PrevMax, V, std::memory_order_relaxed)) {
[all...]
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/sanitizer_common/
H A Dsanitizer_atomic.h21 memory_order_relaxed = 1 << 0, enumerator in enum:__sanitizer::memory_order
76 return atomic_load(a, memory_order_relaxed);
81 atomic_store(a, v, memory_order_relaxed);
H A Dsanitizer_allocator_stats.h34 v += atomic_load(&stats_[i], memory_order_relaxed);
35 atomic_store(&stats_[i], v, memory_order_relaxed);
39 v = atomic_load(&stats_[i], memory_order_relaxed) - v;
40 atomic_store(&stats_[i], v, memory_order_relaxed);
44 atomic_store(&stats_[i], v, memory_order_relaxed);
48 return atomic_load(&stats_[i], memory_order_relaxed);
H A Dsanitizer_atomic_clang_other.h27 DCHECK(mo & (memory_order_relaxed | memory_order_consume
34 if (mo == memory_order_relaxed) {
64 DCHECK(mo & (memory_order_relaxed | memory_order_release
70 if (mo == memory_order_relaxed) {
H A Dsanitizer_atomic_clang_x86.h29 DCHECK(mo & (memory_order_relaxed | memory_order_consume
36 if (mo == memory_order_relaxed) {
74 DCHECK(mo & (memory_order_relaxed | memory_order_release
80 if (mo == memory_order_relaxed) {
H A Dsanitizer_lfstack.h28 atomic_store(&head_, 0, memory_order_relaxed);
32 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0;
36 u64 cmp = atomic_load(&head_, memory_order_relaxed);
H A Dsanitizer_mutex.h25 atomic_store(&state_, 0, memory_order_relaxed);
43 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1);
55 if (atomic_load(&state_, memory_order_relaxed) == 0
100 atomic_store(&state_, kUnlocked, memory_order_relaxed);
104 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked);
136 CHECK_NE(atomic_load(&state_, memory_order_relaxed), kUnlocked);
154 u32 cmp = atomic_load(&state_, memory_order_relaxed);
H A Dsanitizer_atomic_clang_mips.h44 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst));
70 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst));
93 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst));
106 (memory_order_relaxed | memory_order_releasae | memory_order_seq_cst));
H A Dsanitizer_addrhashmap.h195 if (atomic_load(&b->add, memory_order_relaxed)) {
197 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed);
200 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed);
216 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed);
228 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed);
232 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed);
256 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed);
271 atomic_store(&b->add, (uptr)add, memory_order_relaxed);
283 atomic_store(&b->add, (uptr)add1, memory_order_relaxed);
289 CHECK_EQ(atomic_load(&c->addr, memory_order_relaxed),
[all...]
H A Dsanitizer_stackdepot.cpp40 atomic_load(&hash_and_use_count, memory_order_relaxed) & kHashMask;
61 atomic_store(&hash_and_use_count, hash & kHashMask, memory_order_relaxed);
78 return atomic_load(&node_->hash_and_use_count, memory_order_relaxed) &
83 atomic_fetch_add(&node_->hash_and_use_count, 1, memory_order_relaxed) &
H A Dsanitizer_tls_get_addr.cpp49 atomic_fetch_sub(&number_of_live_dtls, 1, memory_order_relaxed);
59 atomic_fetch_add(&number_of_live_dtls, 1, memory_order_relaxed);
102 atomic_load(&number_of_live_dtls, memory_order_relaxed));
H A Dsanitizer_libignore.cpp82 atomic_load(&ignored_ranges_count_, memory_order_relaxed);
111 atomic_load(&instrumented_ranges_count_, memory_order_relaxed);
H A Dsanitizer_persistent_allocator.h55 atomic_store(&region_pos, 0, memory_order_relaxed);
H A Dsanitizer_atomic_msvc.h73 DCHECK(mo & (memory_order_relaxed | memory_order_consume
78 if (mo == memory_order_relaxed) {
90 DCHECK(mo & (memory_order_relaxed | memory_order_release
94 if (mo == memory_order_relaxed) {
H A Dsanitizer_termination.cpp72 if (atomic_fetch_add(&num_calls, 1, memory_order_relaxed) > 10) {
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/tsan/rtl/
H A Dtsan_external.cpp31 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr;
50 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed);
61 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed));
76 uptr new_tag = atomic_fetch_add(&used_tags, 1, memory_order_relaxed);
99 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed));
H A Dtsan_fd.cpp52 atomic_store(&s->rc, 1, memory_order_relaxed);
57 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1)
58 atomic_fetch_add(&s->rc, 1, memory_order_relaxed);
63 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) {
123 atomic_store(&fdctx.globsync.rc, (u64)-1, memory_order_relaxed);
124 atomic_store(&fdctx.filesync.rc, (u64)-1, memory_order_relaxed);
125 atomic_store(&fdctx.socksync.rc, (u64)-1, memory_order_relaxed);
133 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
145 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
H A Dtsan_mutex.cpp219 atomic_store(&state_, kUnlocked, memory_order_relaxed);
223 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked);
235 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) {
286 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/
H A Datomic_helpers.h17 memory_order_relaxed = 0, enumerator in enum:scudo::memory_order
24 static_assert(memory_order_relaxed == __ATOMIC_RELAXED, "");
121 return atomic_load(A, memory_order_relaxed);
126 atomic_store(A, V, memory_order_relaxed);
/freebsd-11-stable/contrib/llvm-project/llvm/include/llvm/Support/
H A DManagedStatic.h91 return *static_cast<C *>(Ptr.load(std::memory_order_relaxed));
101 return *static_cast<C *>(Ptr.load(std::memory_order_relaxed));
/freebsd-11-stable/contrib/llvm-project/lldb/source/Utility/
H A DLog.cpp91 uint32_t mask = m_mask.fetch_or(flags, std::memory_order_relaxed);
93 m_options.store(options, std::memory_order_relaxed);
95 m_channel.log_ptr.store(this, std::memory_order_relaxed);
102 uint32_t mask = m_mask.fetch_and(~flags, std::memory_order_relaxed);
105 m_channel.log_ptr.store(nullptr, std::memory_order_relaxed);
110 return m_options.load(std::memory_order_relaxed);
114 return m_mask.load(std::memory_order_relaxed);
276 return m_options.load(std::memory_order_relaxed) & LLDB_LOG_OPTION_VERBOSE;
354 c.second.m_channel.log_ptr.store(nullptr, std::memory_order_relaxed);
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Support/
H A DPrettyStackTrace.cpp186 GlobalSigInfoGenerationCounter.load(std::memory_order_relaxed);
271 GlobalSigInfoGenerationCounter.fetch_add(1, std::memory_order_relaxed);
279 GlobalSigInfoGenerationCounter.load(std::memory_order_relaxed);
H A DManagedStatic.cpp39 if (!Ptr.load(std::memory_order_relaxed)) {
/freebsd-11-stable/contrib/netbsd-tests/lib/libc/sync/
H A Dcpp_atomic_ops_linkable.cc54 std::memory_order_release, std::memory_order_relaxed);
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/asan/
H A Dasan_thread.h110 return !atomic_load(&stack_switching_, memory_order_relaxed) &&
117 if (atomic_load(&stack_switching_, memory_order_relaxed))

Completed in 140 milliseconds

1234