/fuchsia/zircon/system/ulib/fbl/include/fbl/ |
H A D | atomic.h | 12 // fbl::atomic<T> provides typesafe C++ atomics on integral types, 23 // supported, as fbl::atomic<T> is not an aggregate. 32 // fbl::atomic<T>: 33 // - atomic_init value initializes a default constructed atomic<T> 34 // - atomic_OP are nonmember versions of atomic<T>::OP 41 // In addition, fbl::atomic does not provide the same compatibility 42 // guarantees with C11's <stdatomic.h> is std::atomic does 43 // (std::atomic is designed to allow interop with C by #define 44 // _Atomic(T) std::atomic<T> and so on). The types are not guaranteed 46 // make any of the fbl::atomic function 64 struct atomic { struct in namespace:fbl 76 constexpr atomic(T value) function in struct:fbl::atomic 204 struct atomic<T*> { struct in namespace:fbl 208 constexpr atomic(T* value) function in struct:fbl::atomic [all...] |
/fuchsia/zircon/system/utest/fbl/ |
H A D | atomic_tests.cpp | 5 #include <fbl/atomic.h> 22 [[gnu::unused]] fbl::atomic<char> zero_char(0); 23 [[gnu::unused]] fbl::atomic<signed char> zero_schar(0); 24 [[gnu::unused]] fbl::atomic<unsigned char> zero_uchar(0); 25 [[gnu::unused]] fbl::atomic<short> zero_short(0); 26 [[gnu::unused]] fbl::atomic<unsigned short> zero_ushort(0); 27 [[gnu::unused]] fbl::atomic<int> zero_int(0); 28 [[gnu::unused]] fbl::atomic<unsigned int> zero_uint(0); 29 [[gnu::unused]] fbl::atomic<long> zero_long(0); 30 [[gnu::unused]] fbl::atomic<unsigne [all...] |
/fuchsia/zircon/system/ulib/trace-engine/ |
H A D | nonce.cpp | 7 #include <fbl/atomic.h>
|
H A D | context_impl.h | 9 #include <fbl/atomic.h> 270 fbl::atomic<uint64_t> durable_buffer_current_; 276 fbl::atomic<uint64_t> durable_buffer_full_mark_; 294 fbl::atomic<uint64_t> rolling_buffer_current_; 299 fbl::atomic<uint64_t> rolling_buffer_full_mark_[2]; 302 fbl::atomic<uint64_t> num_records_dropped_{0}; 305 fbl::atomic<uint64_t> num_records_dropped_after_buffer_switch_{0}; 319 fbl::atomic<trace_thread_index_t> next_thread_index_{ 323 fbl::atomic<trace_string_index_t> next_string_index_{
|
/fuchsia/zircon/third_party/ulib/jemalloc/test/unit/ |
H A D | prng.c | 4 test_prng_lg_range_u32(bool atomic) argument 10 ra = prng_lg_range_u32(&sa, 32, atomic); 12 rb = prng_lg_range_u32(&sa, 32, atomic); 17 rb = prng_lg_range_u32(&sb, 32, atomic); 22 ra = prng_lg_range_u32(&sa, 32, atomic); 23 rb = prng_lg_range_u32(&sa, 32, atomic); 28 ra = prng_lg_range_u32(&sa, 32, atomic); 31 rb = prng_lg_range_u32(&sb, lg_range, atomic); 78 test_prng_lg_range_zu(bool atomic) argument 84 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 146 test_prng_range_u32(bool atomic) argument 188 test_prng_range_zu(bool atomic) argument [all...] |
/fuchsia/zircon/third_party/ulib/jemalloc/include/jemalloc/internal/ |
H A D | prng_inlines.h | 10 bool atomic); 12 size_t prng_lg_range_zu(size_t *state, unsigned lg_range, bool atomic); 14 uint32_t prng_range_u32(uint32_t *state, uint32_t range, bool atomic); 16 size_t prng_range_zu(size_t *state, size_t range, bool atomic); 45 prng_lg_range_u32(uint32_t *state, unsigned lg_range, bool atomic) argument 52 if (atomic) { 68 /* 64-bit atomic operations cannot be supported on all relevant platforms. */ 85 prng_lg_range_zu(size_t *state, unsigned lg_range, bool atomic) argument 92 if (atomic) { 109 prng_range_u32(uint32_t *state, uint32_t range, bool atomic) argument 147 prng_range_zu(size_t *state, size_t range, bool atomic) argument [all...] |
/fuchsia/zircon/kernel/arch/arm64/include/arch/arm64/ |
H A D | interrupt.h | 11 #include <kernel/atomic.h>
|
/fuchsia/zircon/kernel/lib/libc/ |
H A D | rand.c | 10 #include <kernel/atomic.h>
|
/fuchsia/zircon/kernel/platform/pc/ |
H A D | power.cpp | 13 #include <fbl/atomic.h> 40 static fbl::atomic<cpu_mask_t> halted_cpus(0); 43 static fbl::atomic<int> halted(0); 75 static fbl::atomic<int> panic_started(0);
|
/fuchsia/zircon/kernel/arch/arm64/ |
H A D | spinlock.cpp | 9 #include <kernel/atomic.h>
|
/fuchsia/zircon/kernel/arch/x86/include/arch/ |
H A D | arch_ops.h | 17 #include <kernel/atomic.h>
|
/fuchsia/zircon/system/dev/misc/pty/ |
H A D | pty-fifo.h | 24 size_t pty_fifo_write(pty_fifo_t* fifo, const void* data, size_t len, bool atomic);
|
H A D | pty-fifo.c | 14 size_t pty_fifo_write(pty_fifo_t* fifo, const void* data, size_t len, bool atomic) { argument 17 if (atomic) {
|
/fuchsia/zircon/system/ulib/lockdep/include/lockdep/ |
H A D | lock_dependency_set.h | 9 #include <fbl/atomic.h> 19 // Implementation note: This hash set makes use of relaxed atomic operations. 21 // threads are the values of the atomic variables themselves, other loads and 23 // thread is ensured due to control dependencies only on the atomic variables. 141 fbl::atomic<LockClassId>& GetEntry(LockClassId id, size_t offset) { 145 const fbl::atomic<LockClassId>& GetEntry(LockClassId id, size_t offset) const { 150 // The list of atomic variables that make up the hash set. Initialized to 152 fbl::atomic<LockClassId> list_[kMaxLockDependencies]{};
|
/fuchsia/zircon/kernel/lib/crypto/entropy/ |
H A D | jitterentropy_collector.cpp | 11 #include <fbl/atomic.h> 23 // Note: this would be fbl::atomic<bool>, except that fbl doesn't support 25 static fbl::atomic<int> initialized = {0};
|
/fuchsia/zircon/kernel/lib/crypto/include/lib/crypto/ |
H A D | prng.h | 12 #include <fbl/atomic.h> 98 fbl::atomic<size_t> accumulated_;
|
/fuchsia/zircon/kernel/lib/hypervisor/ |
H A D | cpu.cpp | 8 #include <fbl/atomic.h> 17 fbl::atomic<cpu_mask_t> cpu_mask;
|
/fuchsia/zircon/system/dev/thermal/aml-thermal-s905d2g/ |
H A D | aml-tsensor.h | 10 #include <fbl/atomic.h> 51 fbl::atomic<bool> running_;
|
/fuchsia/zircon/system/ulib/cobalt-client/include/cobalt-client/cpp/ |
H A D | collector.h | 9 #include <fbl/atomic.h> 89 fbl::atomic<bool> flushing_;
|
H A D | types-internal.h | 10 #include <fbl/atomic.h> 84 fbl::atomic<bool> flushing_;
|
/fuchsia/zircon/system/ulib/syslog/ |
H A D | fx_logger.h | 87 fbl::atomic<fx_log_severity_t> severity_; 88 fbl::atomic<uint32_t> dropped_logs_; 89 fbl::atomic<int> logger_fd_;
|
/fuchsia/zircon/kernel/include/lib/ |
H A D | counters.h | 10 #include <kernel/atomic.h> 42 // atomic operations are never used to set their value so 77 // per-cpu table, therefore to add an atomic is not required. 85 // use a relaxed atomic load/store for arm64 to avoid a potentially nasty 87 // atomic load/stores are about as efficient as a regular load/store. 90 // x86 can do the add in a single non atomic instruction, so the data loss
|
/fuchsia/zircon/kernel/tests/ |
H A D | preempt_disable_tests.cpp | 7 #include <fbl/atomic.h> 225 auto* timer_ran = reinterpret_cast<fbl::atomic<bool>*>(arg); 247 fbl::atomic<bool> timer_ran(false); 271 fbl::atomic<bool> timer_ran(false);
|
/fuchsia/zircon/system/dev/ethernet/aml-ethernet-s912/ |
H A D | aml-ethernet.h | 12 #include <fbl/atomic.h>
|
/fuchsia/zircon/system/ulib/fs/ |
H A D | synchronous-vfs.cpp | 8 #include <fbl/atomic.h>
|