Searched refs:chunk_state (Results 1 - 5 of 5) sorted by relevance
/freebsd-current/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_allocator.cpp | 92 atomic_uint8_t chunk_state; member in class:__asan::ChunkHeader 236 if (!atomic_compare_exchange_strong(&m->chunk_state, &old_chunk_state, 405 if (ac && atomic_load(&ac->chunk_state, memory_order_acquire) == 494 u8 left_state = atomic_load(&left_chunk->chunk_state, memory_order_relaxed); 496 atomic_load(&right_chunk->chunk_state, memory_order_relaxed); 507 // Same chunk_state: choose based on offset. 519 if (atomic_load(&m->chunk_state, memory_order_acquire) != CHUNK_ALLOCATED) 653 atomic_store(&m->chunk_state, CHUNK_ALLOCATED, memory_order_release); 667 // Flip the chunk_state atomically to avoid race on double-free. 668 if (!atomic_compare_exchange_strong(&m->chunk_state, 760 u8 chunk_state = atomic_load(&m->chunk_state, memory_order_acquire); local 787 ReportInvalidFree(void *ptr, u8 chunk_state, BufferedStackTrace *stack) argument [all...] |
/freebsd-current/contrib/llvm-project/llvm/lib/Support/BLAKE3/ |
H A D | blake3.c | 201 blake3_chunk_state chunk_state; local 202 chunk_state_init(&chunk_state, key, flags); 203 chunk_state.chunk_counter = counter; 204 chunk_state_update(&chunk_state, &input[input_position], 206 output_t output = chunk_state_output(&chunk_state); 427 // compress_subtree_wide(), without setting aside anything for the chunk_state. 474 // If we have some partial chunk bytes in the internal chunk_state, we need 497 // Now the chunk_state is clear, and we have more input. If there's more than 535 blake3_chunk_state chunk_state; local 536 chunk_state_init(&chunk_state, sel [all...] |
/freebsd-current/sys/contrib/openzfs/module/icp/algs/blake3/ |
H A D | blake3.c | 255 blake3_chunk_state_t chunk_state; local 256 chunk_state_init(&chunk_state, key, flags); 257 chunk_state.chunk_counter = counter; 258 chunk_state_update(ops, &chunk_state, &input[input_position], 260 output_t output = chunk_state_output(&chunk_state); 471 * compress_subtree_wide(), without setting aside anything for the chunk_state. 538 * If we have some partial chunk bytes in the internal chunk_state, we 567 * Now the chunk_state is clear, and we have more input. If there's 616 blake3_chunk_state_t chunk_state; local 617 chunk_state_init(&chunk_state, ct [all...] |
/freebsd-current/contrib/llvm-project/compiler-rt/lib/hwasan/ |
H A D | hwasan_allocator.h | 39 atomic_uint8_t chunk_state; member in struct:__hwasan::Metadata
|
H A D | hwasan_allocator.cpp | 91 atomic_store(&chunk_state, CHUNK_ALLOCATED, memory_order_release); 95 atomic_store(&chunk_state, CHUNK_INVALID, memory_order_release); 102 return atomic_load(&chunk_state, memory_order_relaxed) == CHUNK_ALLOCATED;
|
Completed in 122 milliseconds