Searched refs:arena (Results 1 - 25 of 25) sorted by last modified time

/freebsd-11-stable/contrib/unbound/validator/
H A Dval_secalgo.c921 PLArenaPool* arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE); local
922 if(!arena) {
926 key = PORT_ArenaZNew(arena, SECKEYPublicKey);
929 PORT_FreeArena(arena, PR_FALSE);
932 key->arena = arena;
981 if(SECITEM_CopyItem(pk->arena, &pk->u.ec.publicValue, &pub)) {
985 if(SECITEM_CopyItem(pk->arena, &pk->u.ec.DEREncodedParams, &params)) {
1036 if(SECITEM_CopyItem(pk->arena, &pk->u.dsa.params.prime, &P)) {
1040 if(SECITEM_CopyItem(pk->arena,
[all...]
/freebsd-11-stable/lib/libc/stdlib/jemalloc/
H A DMakefile.inc5 JEMALLOCSRCS:= jemalloc.c arena.c atomic.c base.c bitmap.c chunk.c \
/freebsd-11-stable/contrib/llvm-project/llvm/lib/Support/Unix/
H A DProcess.inc102 // the value returned by mallinfo in the arena field.
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/scudo/standalone/
H A Dwrappers_c.h23 __scudo_mallinfo_data_t arena; member in struct:__scudo_mallinfo
/freebsd-11-stable/contrib/tcsh/
H A Dtc.alloc.c646 xprintf(CGETS(19, 14, "Total space allocated from system: %d\n"), mi.arena);
/freebsd-11-stable/contrib/jemalloc/src/
H A Djemalloc.c391 arena_set(unsigned ind, arena_t *arena) argument
394 atomic_write_p((void **)&arenas[ind], arena);
418 /* Create a new arena and insert it into the arenas array at index ind. */
422 arena_t *arena; local
432 * auto arena.
434 arena = arena_get(tsdn, ind, false);
435 if (arena != NULL) {
437 return (arena);
440 /* Actually initialize the arena. */
441 arena
449 arena_t *arena; local
460 arena_t *arena; local
488 arena_t *arena; local
638 arena_t *arena; local
691 arena_t *arena; local
745 arena_t *arena = arena_get(tsdn, i, false); local
2003 imallocx_flags_decode(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument
2035 imallocx_flags(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument
2049 imallocx_prof_sample(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument
2077 arena_t *arena; local
2110 arena_t *arena; local
2187 irallocx_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx) argument
2210 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena) argument
2262 arena_t *arena; local
2809 arena_t *arena; local
2832 arena_prefork0(tsd_tsdn(tsd), arena); local
2835 arena_prefork1(tsd_tsdn(tsd), arena); local
2838 arena_prefork2(tsd_tsdn(tsd), arena); local
2849 arena_prefork3(tsd_tsdn(tsd), arena); local
2878 arena_t *arena; local
2881 arena_postfork_parent(tsd_tsdn(tsd), arena); local
2903 arena_t *arena; local
2906 arena_postfork_child(tsd_tsdn(tsd), arena); local
[all...]
H A Darena.c40 static void arena_purge_to_limit(tsdn_t *tsdn, arena_t *arena,
42 static void arena_run_dalloc(tsdn_t *tsdn, arena_t *arena, arena_run_t *run,
44 static void arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena,
46 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk,
204 arena_runs_avail_get(arena_t *arena, szind_t ind) argument
210 return (&arena->runs_avail[ind - runs_avail_bias]);
214 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument
221 arena_run_heap_insert(arena_runs_avail_get(arena, ind),
226 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument
233 arena_run_heap_remove(arena_runs_avail_get(arena, in
238 arena_run_dirty_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
256 arena_run_dirty_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
281 arena_chunk_cache_maybe_insert(arena_t *arena, extent_node_t *node, bool cache) argument
293 arena_chunk_cache_maybe_remove(arena_t *arena, extent_node_t *node, bool dirty) argument
379 arena_nactive_add(arena_t *arena, size_t add_pages) argument
393 arena_nactive_sub(arena_t *arena, size_t sub_pages) argument
406 arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind, size_t flag_dirty, size_t flag_decommitted, size_t need_pages) argument
448 arena_run_split_large_helper(arena_t *arena, arena_run_t *run, size_t size, bool remove, bool zero) argument
521 arena_run_split_large(arena_t *arena, arena_run_t *run, size_t size, bool zero) argument
528 arena_run_init_large(arena_t *arena, arena_run_t *run, size_t size, bool zero) argument
535 arena_run_split_small(arena_t *arena, arena_run_t *run, size_t size, szind_t binind) argument
573 arena_chunk_init_spare(arena_t *arena) argument
595 arena_chunk_register(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, bool zero) argument
611 arena_chunk_alloc_internal_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, bool *zero, bool *commit) argument
645 arena_chunk_alloc_internal(tsdn_t *tsdn, arena_t *arena, bool *zero, bool *commit) argument
675 arena_chunk_init_hard(tsdn_t *tsdn, arena_t *arena) argument
730 arena_chunk_alloc(tsdn_t *tsdn, arena_t *arena) argument
750 arena_chunk_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument
780 arena_spare_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *spare) argument
794 arena_chunk_dalloc(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument
820 arena_huge_malloc_stats_update(arena_t *arena, size_t usize) argument
833 arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize) argument
846 arena_huge_dalloc_stats_update(arena_t *arena, size_t usize) argument
859 arena_huge_reset_stats_cancel(arena_t *arena, size_t usize) argument
870 arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize) argument
883 arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize) argument
891 arena_huge_ralloc_stats_update_undo(arena_t *arena, size_t oldsize, size_t usize) argument
900 arena_node_alloc(tsdn_t *tsdn, arena_t *arena) argument
916 arena_node_dalloc(tsdn_t *tsdn, arena_t *arena, extent_node_t *node) argument
926 arena_chunk_alloc_huge_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, size_t usize, size_t alignment, bool *zero, size_t csize) argument
950 arena_chunk_alloc_huge(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool *zero) argument
978 arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t usize) argument
996 arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
1014 arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
1040 arena_chunk_ralloc_huge_expand_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument
1069 arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument
1110 arena_run_first_best_fit(arena_t *arena, size_t size) argument
1126 arena_run_alloc_large_helper(arena_t *arena, size_t size, bool zero) argument
1137 arena_run_alloc_large(tsdn_t *tsdn, arena_t *arena, size_t size, bool zero) argument
1170 arena_run_alloc_small_helper(arena_t *arena, size_t size, szind_t binind) argument
1181 arena_run_alloc_small(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t binind) argument
1223 arena_lg_dirty_mult_get(tsdn_t *tsdn, arena_t *arena) argument
1235 arena_lg_dirty_mult_set(tsdn_t *tsdn, arena_t *arena, ssize_t lg_dirty_mult) argument
1250 arena_decay_deadline_init(arena_t *arena) argument
1271 arena_decay_deadline_reached(const arena_t *arena, const nstime_t *time) argument
1280 arena_decay_backlog_npages_limit(const arena_t *arena) argument
1308 arena_decay_epoch_advance(arena_t *arena, const nstime_t *time) argument
1355 arena_decay_npages_limit(arena_t *arena) argument
1371 arena_decay_init(arena_t *arena, ssize_t decay_time) argument
1401 arena_decay_time_get(tsdn_t *tsdn, arena_t *arena) argument
1413 arena_decay_time_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_time) argument
1436 arena_maybe_purge_ratio(tsdn_t *tsdn, arena_t *arena) argument
1464 arena_maybe_purge_decay(tsdn_t *tsdn, arena_t *arena) argument
1499 arena_maybe_purge(tsdn_t *tsdn, arena_t *arena) argument
1513 arena_dirty_count(arena_t *arena) argument
1547 arena_stash_dirty(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, size_t ndirty_limit, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1637 arena_purge_stashed(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1730 arena_unstash_purged(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1780 arena_purge_to_limit(tsdn_t *tsdn, arena_t *arena, size_t ndirty_limit) argument
1821 arena_purge(tsdn_t *tsdn, arena_t *arena, bool all) argument
1833 arena_achunk_prof_reset(tsd_t *tsd, arena_t *arena, arena_chunk_t *chunk) argument
1873 arena_reset(tsd_t *tsd, arena_t *arena) argument
1957 arena_chunk_discard(tsd_tsdn(tsd), arena, local
1963 arena_chunk_discard(tsd_tsdn(tsd), arena, arena->spare); local
1977 arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size, size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty, size_t flag_decommitted) argument
2069 arena_run_size_get(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t run_ind) argument
2090 arena_run_dalloc(tsdn_t *tsdn, arena_t *arena, arena_run_t *run, bool dirty, bool cleaned, bool decommitted) argument
2166 arena_run_trim_head(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument
2207 arena_run_trim_tail(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument
2274 arena_bin_nonfull_run_get(tsdn_t *tsdn, arena_t *arena, arena_bin_t *bin) argument
2325 arena_bin_malloc_hard(tsdn_t *tsdn, arena_t *arena, arena_bin_t *bin) argument
2375 arena_tcache_fill_small(tsdn_t *tsdn, arena_t *arena, tcache_bin_t *tbin, szind_t binind, uint64_t prof_accumbytes) argument
2535 arena_malloc_small(tsdn_t *tsdn, arena_t *arena, szind_t binind, bool zero) argument
2589 arena_malloc_large(tsdn_t *tsdn, arena_t *arena, szind_t binind, bool zero) argument
2651 arena_malloc_hard(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero) argument
2671 arena_palloc_large(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument
2759 arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
2845 arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2863 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2884 arena_dalloc_bin_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm, bool junked) argument
2917 arena_dalloc_bin_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm) argument
2925 arena_dalloc_bin(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_bits_t *bitselm) argument
2941 arena_dalloc_small(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument
2975 arena_dalloc_large_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, bool junked) argument
3003 arena_dalloc_large_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
3011 arena_dalloc_large(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
3022 arena_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument
3059 arena_ralloc_large_grow(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument
3187 arena_t *arena; local
3266 arena_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
3280 arena_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t size, size_t alignment, bool zero, tcache_t *tcache) argument
3325 arena_dss_prec_get(tsdn_t *tsdn, arena_t *arena) argument
3336 arena_dss_prec_set(tsdn_t *tsdn, arena_t *arena, dss_prec_t dss_prec) argument
3386 arena_basic_stats_merge_locked(arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty) argument
3400 arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty) argument
3412 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty, arena_stats_t *astats, malloc_bin_stats_t *bstats, malloc_large_stats_t *lstats, malloc_huge_stats_t *hstats) argument
3475 arena_nthreads_get(arena_t *arena, bool internal) argument
3482 arena_nthreads_inc(arena_t *arena, bool internal) argument
3489 arena_nthreads_dec(arena_t *arena, bool internal) argument
3498 arena_t *arena; local
3837 arena_prefork0(tsdn_t *tsdn, arena_t *arena) argument
3844 arena_prefork1(tsdn_t *tsdn, arena_t *arena) argument
3851 arena_prefork2(tsdn_t *tsdn, arena_t *arena) argument
3858 arena_prefork3(tsdn_t *tsdn, arena_t *arena) argument
3868 arena_postfork_parent(tsdn_t *tsdn, arena_t *arena) argument
3881 arena_postfork_child(tsdn_t *tsdn, arena_t *arena) argument
[all...]
H A Dchunk.c52 static void chunk_record(tsdn_t *tsdn, arena_t *arena,
60 chunk_hooks_get_locked(arena_t *arena) argument
63 return (arena->chunk_hooks);
67 chunk_hooks_get(tsdn_t *tsdn, arena_t *arena) argument
71 malloc_mutex_lock(tsdn, &arena->chunks_mtx);
72 chunk_hooks = chunk_hooks_get_locked(arena);
73 malloc_mutex_unlock(tsdn, &arena->chunks_mtx);
79 chunk_hooks_set(tsdn_t *tsdn, arena_t *arena, const chunk_hooks_t *chunk_hooks) argument
83 malloc_mutex_lock(tsdn, &arena->chunks_mtx);
84 old_chunk_hooks = arena
114 chunk_hooks_assure_initialized_impl(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, bool locked) argument
128 chunk_hooks_assure_initialized_locked(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks) argument
136 chunk_hooks_assure_initialized(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks) argument
190 chunk_first_best_fit(arena_t *arena, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, size_t size) argument
202 chunk_recycle(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, bool dalloc_node) argument
334 chunk_alloc_core(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, dss_prec_t dss_prec) argument
386 chunk_alloc_cache(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool dalloc_node) argument
412 arena_t *arena; local
424 chunk_alloc_default_impl(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument
444 arena_t *arena; local
454 chunk_alloc_retained(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument
475 chunk_alloc_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument
506 chunk_record(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *chunk, size_t size, bool zeroed, bool committed) argument
598 chunk_dalloc_cache(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed) argument
633 chunk_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed, bool committed) argument
701 chunk_purge_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset, size_t length) argument
[all...]
H A Dchunk_dss.c69 chunk_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, argument
137 chunk_dalloc_wrapper(tsdn, arena,
H A Dctl.c55 arena_t *arena);
58 static void ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, unsigned i);
240 {NAME("arena"), CTL(thread_arena)},
501 {NAME("arena"), CHILD(indexed, arena)},
563 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_stats_t *cstats, arena_t *arena) argument
568 arena_stats_merge(tsdn, arena, &cstats->nthreads, &cstats->dss,
581 arena_basic_stats_merge(tsdn, arena, &cstats->nthreads,
659 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, unsigned i) argument
665 ctl_arena_stats_amerge(tsdn, astats, arena);
1620 arena_t *arena; local
1679 arena_t *arena = arena_get(tsd_tsdn(tsd), arena_ind, false); local
1710 arena_t *arena; local
1745 arena_t *arena; local
1780 arena_t *arena; local
[all...]
H A Dhuge.c43 huge_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) argument
48 return (huge_palloc(tsdn, arena, usize, chunksize, zero));
52 huge_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument
62 assert(!tsdn_null(tsdn) || arena != NULL);
71 CACHELINE, false, NULL, true, arena_ichoose(tsdn, arena));
81 arena = arena_choose(tsdn_tsd(tsdn), arena);
82 if (unlikely(arena == NULL) || (ret = arena_chunk_alloc_huge(tsdn,
83 arena, usize, alignment, &is_zeroed)) == NULL) {
88 extent_node_init(node, arena, re
141 arena_t *arena; local
203 arena_t *arena; local
255 arena_t *arena; local
352 huge_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument
362 huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
396 arena_t *arena; local
426 arena_t *arena; local
442 arena_t *arena; local
457 arena_t *arena; local
[all...]
H A Dtcache.c74 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, argument
79 arena_tcache_fill_small(tsdn, arena, tbin, binind, config_prof ?
92 arena_t *arena; local
100 arena = arena_choose(tsd, NULL);
101 assert(arena != NULL);
103 /* Lock the arena bin associated with the first object. */
109 if (config_prof && bin_arena == arena) {
110 if (arena_prof_accum(tsd_tsdn(tsd), arena,
117 if (config_stats && bin_arena == arena) {
139 * arena bi
174 arena_t *arena; local
256 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument
269 tcache_arena_dissociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument
304 arena_t *arena; local
318 tcache_create(tsdn_t *tsdn, arena_t *arena) argument
360 arena_t *arena; local
364 tcache_arena_dissociate(tsd_tsdn(tsd), tcache, arena); local
420 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument
450 arena_t *arena; local
[all...]
/freebsd-11-stable/crypto/openssl/crypto/threads/
H A Dmttest.c860 static usptr_t *arena; variable
875 arena = usinit(filename);
880 lock_cs[i] = usnewsema(arena, 1);
897 usfreesema(lock_cs[i], arena);
H A Dth-lock.c254 static usptr_t *arena; variable
275 arena = usinit(filename);
279 lock_cs[i] = usnewsema(arena, 1);
296 usfreesema(lock_cs[i], arena);
/freebsd-11-stable/contrib/jemalloc/include/jemalloc/internal/
H A Darena.h17 * The minimum ratio of active:dirty pages per arena is computed as:
186 * A pointer to the arena that owns the chunk is stored within the node.
204 * than one per arena), but mainly to avoid false cacheline sharing.
263 * locked. Run allocation/deallocation are protected by the arena lock,
289 /* This arena's index within the arenas array. */
293 * Number of threads currently assigned to this arena, synchronized via
297 * created via the arenas.extend mallctl, because the arena.<i>.reset
299 * arena.
307 * There are three classes of arena operations from a locking
317 * List of tcaches for extant threads associated with this arena
[all...]
H A Dchunk.h51 chunk_hooks_t chunk_hooks_get(tsdn_t *tsdn, arena_t *arena);
52 chunk_hooks_t chunk_hooks_set(tsdn_t *tsdn, arena_t *arena,
59 void *chunk_alloc_cache(tsdn_t *tsdn, arena_t *arena,
62 void *chunk_alloc_wrapper(tsdn_t *tsdn, arena_t *arena,
65 void chunk_dalloc_cache(tsdn_t *tsdn, arena_t *arena,
67 void chunk_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena,
70 bool chunk_purge_wrapper(tsdn_t *tsdn, arena_t *arena,
H A Dchunk_dss.h26 void *chunk_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr,
H A Dextent.h36 * don't return arena chunks.
43 /* Linkage for arena's runs_dirty and chunks_cache rings. */
51 /* Linkage for arena's achunks, huge, and node_cache lists. */
80 void extent_node_arena_set(extent_node_t *node, arena_t *arena);
87 void extent_node_init(extent_node_t *node, arena_t *arena, void *addr,
147 extent_node_arena_set(extent_node_t *node, arena_t *arena) argument
150 node->en_arena = arena;
196 extent_node_init(extent_node_t *node, arena_t *arena, void *addr, size_t size, argument
200 extent_node_arena_set(node, arena);
H A Dhuge.h12 void *huge_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
13 void *huge_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize,
17 void *huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize,
H A Djemalloc_internal.h168 * jemalloc can conceptually be broken into components (arena, tcache, etc.),
191 * a: arena
376 #include "jemalloc/internal/arena.h"
408 #include "jemalloc/internal/arena.h"
412 #include "jemalloc/internal/arena.h"
504 #include "jemalloc/internal/arena.h"
553 arena_t *arena_choose_impl(tsd_t *tsd, arena_t *arena, bool internal);
554 arena_t *arena_choose(tsd_t *tsd, arena_t *arena);
555 arena_t *arena_ichoose(tsdn_t *tsdn, arena_t *arena);
791 /* Choose an arena base
[all...]
H A Dtcache.h135 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
144 tcache_t *tcache_create(tsdn_t *tsdn, arena_t *arena);
147 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
164 void *tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
166 void *tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
280 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, argument
294 arena = arena_choose(tsd, arena);
295 if (unlikely(arena == NULL))
298 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcach
[all...]
H A Dtsd.h570 O(arena, arena_t *) \
/freebsd-11-stable/sys/contrib/octeon-sdk/cvmx-malloc/
H A Darena.c24 /* $Id: arena.c 30481 2007-12-05 21:46:59Z rfranz $ */
28 #define HEAP_MIN_SIZE (4096) /* Must leave room for struct malloc_state, arena ptrs, etc., totals about 2400 bytes */
79 /* find the heap and corresponding arena for a given ptr */
82 #define set_arena_for_chunk(ptr, arena) (ptr)->arena_ptr = (arena)
99 /* Magic value for the thread-specific arena pointer when
218 /* We must zero out the arena as the malloc code assumes this. */
255 debug_printf("Adding arena at addr: %p, size %d\n", ptr, size);
H A Dmalloc.c948 arena: current total non-mmapped bytes allocated from system
1322 the main arena is obtained via mmap.
1340 * When sbrk is called to extend the top of the arena to satisfy
1348 so that the end of the arena is always a system page boundary.
1653 mstate arena_ptr; /* ptr to arena chunk belongs to */
1783 ** when in use. NOTE - last SIZE_SZ of arena must be left
2196 /* Memory allocated from the system in this arena. */
2281 #include "arena.c"
2487 chunk, or the base of its memory arena. This is ensured
2644 debug_printf("New arena
[all...]
H A Dmalloc.h128 int arena; /* non-mmapped space allocated from system */ member in struct:mallinfo

Completed in 163 milliseconds