/freebsd-11-stable/contrib/jemalloc/include/jemalloc/internal/ |
H A D | base.h | 12 void *base_alloc(tsdn_t *tsdn, size_t size); 13 void base_stats_get(tsdn_t *tsdn, size_t *allocated, size_t *resident, 16 void base_prefork(tsdn_t *tsdn); 17 void base_postfork_parent(tsdn_t *tsdn); 18 void base_postfork_child(tsdn_t *tsdn);
|
H A D | chunk_dss.h | 24 dss_prec_t chunk_dss_prec_get(tsdn_t *tsdn); 25 bool chunk_dss_prec_set(tsdn_t *tsdn, dss_prec_t dss_prec); 26 void *chunk_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, 28 bool chunk_in_dss(tsdn_t *tsdn, void *chunk); 30 void chunk_dss_prefork(tsdn_t *tsdn); 31 void chunk_dss_postfork_parent(tsdn_t *tsdn); 32 void chunk_dss_postfork_child(tsdn_t *tsdn);
|
H A D | huge.h | 12 void *huge_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 13 void *huge_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, 15 bool huge_ralloc_no_move(tsdn_t *tsdn, void *ptr, size_t oldsize, 23 void huge_dalloc(tsdn_t *tsdn, void *ptr); 25 size_t huge_salloc(tsdn_t *tsdn, const void *ptr); 26 prof_tctx_t *huge_prof_tctx_get(tsdn_t *tsdn, const void *ptr); 27 void huge_prof_tctx_set(tsdn_t *tsdn, const void *ptr, prof_tctx_t *tctx); 28 void huge_prof_tctx_reset(tsdn_t *tsdn, const void *ptr);
|
H A D | mutex.h | 59 void malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex); 60 void malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex); 61 void malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex); 70 void malloc_mutex_lock(tsdn_t *tsdn, malloc_mutex_t *mutex); 71 void malloc_mutex_unlock(tsdn_t *tsdn, malloc_mutex_t *mutex); 72 void malloc_mutex_assert_owner(tsdn_t *tsdn, malloc_mutex_t *mutex); 73 void malloc_mutex_assert_not_owner(tsdn_t *tsdn, malloc_mutex_t *mutex); 78 malloc_mutex_lock(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 82 witness_assert_not_owner(tsdn, &mutex->witness); 94 witness_lock(tsdn, 99 malloc_mutex_unlock(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 119 malloc_mutex_assert_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 127 malloc_mutex_assert_not_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) argument [all...] |
H A D | chunk.h | 51 chunk_hooks_t chunk_hooks_get(tsdn_t *tsdn, arena_t *arena); 52 chunk_hooks_t chunk_hooks_set(tsdn_t *tsdn, arena_t *arena, 55 bool chunk_register(tsdn_t *tsdn, const void *chunk, 59 void *chunk_alloc_cache(tsdn_t *tsdn, arena_t *arena, 62 void *chunk_alloc_wrapper(tsdn_t *tsdn, arena_t *arena, 65 void chunk_dalloc_cache(tsdn_t *tsdn, arena_t *arena, 67 void chunk_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, 70 bool chunk_purge_wrapper(tsdn_t *tsdn, arena_t *arena, 74 void chunk_prefork(tsdn_t *tsdn); 75 void chunk_postfork_parent(tsdn_t *tsdn); [all...] |
H A D | witness.h | 111 void witness_assert_owner(tsdn_t *tsdn, const witness_t *witness); 112 void witness_assert_not_owner(tsdn_t *tsdn, const witness_t *witness); 113 void witness_assert_lockless(tsdn_t *tsdn); 114 void witness_lock(tsdn_t *tsdn, witness_t *witness); 115 void witness_unlock(tsdn_t *tsdn, witness_t *witness); 120 witness_assert_owner(tsdn_t *tsdn, const witness_t *witness) argument 129 if (tsdn_null(tsdn)) 131 tsd = tsdn_tsd(tsdn); 144 witness_assert_not_owner(tsdn_t *tsdn, const witness_t *witness) argument 153 if (tsdn_null(tsdn)) 167 witness_assert_lockless(tsdn_t *tsdn) argument 187 witness_lock(tsdn_t *tsdn, witness_t *witness) argument 227 witness_unlock(tsdn_t *tsdn, witness_t *witness) argument [all...] |
H A D | prof.h | 284 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, 299 void prof_idump(tsdn_t *tsdn); 301 void prof_gdump(tsdn_t *tsdn); 302 prof_tdata_t *prof_tdata_init(tsdn_t *tsdn); 304 void prof_reset(tsdn_t *tsdn, size_t lg_sample); 306 bool prof_active_get(tsdn_t *tsdn); 307 bool prof_active_set(tsdn_t *tsdn, bool active); 312 bool prof_thread_active_init_get(tsdn_t *tsdn); 313 bool prof_thread_active_init_set(tsdn_t *tsdn, bool active_init); 314 bool prof_gdump_get(tsdn_t *tsdn); 401 prof_tctx_get(tsdn_t *tsdn, const void *ptr) argument 411 prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, prof_tctx_t *tctx) argument 421 prof_tctx_reset(tsdn_t *tsdn, const void *ptr, size_t usize, const void *old_ptr, prof_tctx_t *old_tctx) argument 483 prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, prof_tctx_t *tctx) argument [all...] |
H A D | ckh.h | 67 bool ckh_new(tsdn_t *tsdn, ckh_t *ckh, size_t minitems, ckh_hash_t *hash, 69 void ckh_delete(tsdn_t *tsdn, ckh_t *ckh); 72 bool ckh_insert(tsdn_t *tsdn, ckh_t *ckh, const void *key, const void *data); 73 bool ckh_remove(tsdn_t *tsdn, ckh_t *ckh, const void *searchkey, void **key,
|
H A D | arena.h | 521 extent_node_t *arena_node_alloc(tsdn_t *tsdn, arena_t *arena); 522 void arena_node_dalloc(tsdn_t *tsdn, arena_t *arena, extent_node_t *node); 523 void *arena_chunk_alloc_huge(tsdn_t *tsdn, arena_t *arena, size_t usize, 525 void arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, 527 void arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, 529 void arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, 531 bool arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, 533 ssize_t arena_lg_dirty_mult_get(tsdn_t *tsdn, arena_t *arena); 534 bool arena_lg_dirty_mult_set(tsdn_t *tsdn, arena_t *arena, 536 ssize_t arena_decay_time_get(tsdn_t *tsdn, arena_ [all...] |
H A D | valgrind.h | 33 #define JEMALLOC_VALGRIND_MALLOC(cond, tsdn, ptr, usize, zero) do { \ 35 VALGRIND_MALLOCLIKE_BLOCK(ptr, usize, p2rz(tsdn, ptr), \ 39 #define JEMALLOC_VALGRIND_REALLOC(maybe_moved, tsdn, ptr, usize, \ 43 size_t rzsize = p2rz(tsdn, ptr); \ 86 #define JEMALLOC_VALGRIND_MALLOC(cond, tsdn, ptr, usize, zero) do {} while (0) 87 #define JEMALLOC_VALGRIND_REALLOC(maybe_moved, tsdn, ptr, usize, \
|
H A D | ctl.h | 75 int ctl_nametomib(tsdn_t *tsdn, const char *name, size_t *mibp, 81 void ctl_prefork(tsdn_t *tsdn); 82 void ctl_postfork_parent(tsdn_t *tsdn); 83 void ctl_postfork_child(tsdn_t *tsdn);
|
H A D | jemalloc_internal.h | 473 arena_t *arena_init(tsdn_t *tsdn, unsigned ind); 555 arena_t *arena_ichoose(tsdn_t *tsdn, arena_t *arena); 558 arena_t *arena_get(tsdn_t *tsdn, unsigned ind, bool init_if_missing); 815 arena_ichoose(tsdn_t *tsdn, arena_t *arena) 818 assert(!tsdn_null(tsdn) || arena != NULL); 820 if (!tsdn_null(tsdn)) 821 return (arena_choose_impl(tsdn_tsd(tsdn), NULL, true)); 851 arena_get(tsdn_t *tsdn, unsigned ind, bool init_if_missing) 861 ret = arena_init(tsdn, ind); 895 size_t isalloc(tsdn_t *tsdn, cons [all...] |
H A D | tcache.h | 133 size_t tcache_salloc(tsdn_t *tsdn, const void *ptr); 135 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 141 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, 144 tcache_t *tcache_create(tsdn_t *tsdn, arena_t *arena); 147 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 148 bool tcaches_create(tsdn_t *tsdn, unsigned *r_ind); 151 bool tcache_boot(tsdn_t *tsdn);
|
/freebsd-11-stable/contrib/jemalloc/src/ |
H A D | huge.c | 18 huge_node_set(tsdn_t *tsdn, const void *ptr, extent_node_t *node) argument 23 return (chunk_register(tsdn, ptr, node)); 27 huge_node_reset(tsdn_t *tsdn, const void *ptr, extent_node_t *node) argument 31 err = huge_node_set(tsdn, ptr, node); 43 huge_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) argument 48 return (huge_palloc(tsdn, arena, usize, chunksize, zero)); 52 huge_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument 62 assert(!tsdn_null(tsdn) || arena != NULL); 70 node = ipallocztm(tsdn, CACHELINE_CEILING(sizeof(extent_node_t)), 71 CACHELINE, false, NULL, true, arena_ichoose(tsdn, aren 117 huge_dalloc_junk(tsdn_t *tsdn, void *ptr, size_t usize) argument 136 huge_ralloc_no_move_similar(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument 199 huge_ralloc_no_move_shrink(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize) argument 252 huge_ralloc_no_move_expand(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize, bool zero) argument 300 huge_ralloc_no_move(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument 352 huge_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 393 huge_dalloc(tsdn_t *tsdn, void *ptr) argument 422 huge_salloc(tsdn_t *tsdn, const void *ptr) argument 438 huge_prof_tctx_get(tsdn_t *tsdn, const void *ptr) argument 454 huge_prof_tctx_set(tsdn_t *tsdn, const void *ptr, prof_tctx_t *tctx) argument 467 huge_prof_tctx_reset(tsdn_t *tsdn, const void *ptr) argument [all...] |
H A D | base.c | 17 base_node_try_alloc(tsdn_t *tsdn) argument 21 malloc_mutex_assert_owner(tsdn, &base_mtx); 32 base_node_dalloc(tsdn_t *tsdn, extent_node_t *node) argument 35 malloc_mutex_assert_owner(tsdn, &base_mtx); 43 base_chunk_alloc(tsdn_t *tsdn, size_t minsize) argument 49 malloc_mutex_assert_owner(tsdn, &base_mtx); 51 node = base_node_try_alloc(tsdn); 58 base_node_dalloc(tsdn, node); 81 base_alloc(tsdn_t *tsdn, size_t size) argument 96 malloc_mutex_lock(tsdn, 133 base_stats_get(tsdn_t *tsdn, size_t *allocated, size_t *resident, size_t *mapped) argument 159 base_prefork(tsdn_t *tsdn) argument 166 base_postfork_parent(tsdn_t *tsdn) argument 173 base_postfork_child(tsdn_t *tsdn) argument [all...] |
H A D | chunk_dss.c | 44 chunk_dss_prec_get(tsdn_t *tsdn) argument 50 malloc_mutex_lock(tsdn, &dss_mtx); 52 malloc_mutex_unlock(tsdn, &dss_mtx); 57 chunk_dss_prec_set(tsdn_t *tsdn, dss_prec_t dss_prec) argument 62 malloc_mutex_lock(tsdn, &dss_mtx); 64 malloc_mutex_unlock(tsdn, &dss_mtx); 69 chunk_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, argument 83 malloc_mutex_lock(tsdn, &dss_mtx); 125 malloc_mutex_unlock(tsdn, &dss_mtx); 133 malloc_mutex_unlock(tsdn, 158 chunk_in_dss(tsdn_t *tsdn, void *chunk) argument 191 chunk_dss_prefork(tsdn_t *tsdn) argument 199 chunk_dss_postfork_parent(tsdn_t *tsdn) argument 207 chunk_dss_postfork_child(tsdn_t *tsdn) argument [all...] |
H A D | prof.c | 124 static bool prof_tctx_should_destroy(tsdn_t *tsdn, prof_tctx_t *tctx); 126 static bool prof_tdata_should_destroy(tsdn_t *tsdn, prof_tdata_t *tdata, 128 static void prof_tdata_destroy(tsdn_t *tsdn, prof_tdata_t *tdata, 130 static char *prof_thread_name_alloc(tsdn_t *tsdn, const char *thread_name); 226 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, argument 230 prof_tctx_set(tsdn, ptr, usize, tctx); 232 malloc_mutex_lock(tsdn, tctx->tdata->lock); 240 malloc_mutex_unlock(tsdn, tctx->tdata->lock); 550 prof_gctx_create(tsdn_t *tsdn, prof_bt_t *bt) argument 556 prof_gctx_t *gctx = (prof_gctx_t *)iallocztm(tsdn, siz 612 prof_tctx_should_destroy(tsdn_t *tsdn, prof_tctx_t *tctx) argument 901 tsdn_t *tsdn; local 1042 prof_tctx_merge_tdata(tsdn_t *tsdn, prof_tctx_t *tctx, prof_tdata_t *tdata) argument 1075 prof_tctx_merge_gctx(tsdn_t *tsdn, prof_tctx_t *tctx, prof_gctx_t *gctx) argument 1091 tsdn_t *tsdn = (tsdn_t *)arg; local 1111 tsdn_t *tsdn; member in struct:prof_tctx_dump_iter_arg_s 1146 tsdn_t *tsdn = (tsdn_t *)arg; local 1171 prof_dump_gctx_prep(tsdn_t *tsdn, prof_gctx_t *gctx, prof_gctx_tree_t *gctxs) argument 1192 tsdn_t *tsdn; member in struct:prof_gctx_merge_iter_arg_s 1258 tsdn_t *tsdn; member in struct:prof_tdata_merge_iter_arg_s 1320 prof_dump_header(tsdn_t *tsdn, bool propagate_err, const prof_cnt_t *cnt_all) argument 1344 prof_dump_gctx(tsdn_t *tsdn, bool propagate_err, prof_gctx_t *gctx, const prof_bt_t *bt, prof_gctx_tree_t *gctxs) argument 1520 tsdn_t *tsdn; member in struct:prof_gctx_dump_iter_arg_s 1677 prof_idump(tsdn_t *tsdn) argument 1729 prof_gdump(tsdn_t *tsdn) argument 1781 prof_thr_uid_alloc(tsdn_t *tsdn) argument 1794 prof_tdata_init_impl(tsdn_t *tsdn, uint64_t thr_uid, uint64_t thr_discrim, char *thread_name, bool active) argument 1840 prof_tdata_init(tsdn_t *tsdn) argument 1859 prof_tdata_should_destroy(tsdn_t *tsdn, prof_tdata_t *tdata, bool even_if_attached) argument 1869 prof_tdata_destroy_locked(tsdn_t *tsdn, prof_tdata_t *tdata, bool even_if_attached) argument 1888 prof_tdata_destroy(tsdn_t *tsdn, prof_tdata_t *tdata, bool even_if_attached) argument 1934 prof_tdata_expire(tsdn_t *tsdn, prof_tdata_t *tdata) argument 1953 tsdn_t *tsdn = (tsdn_t *)arg; local 1959 prof_reset(tsdn_t *tsdn, size_t lg_sample) argument 1999 prof_active_get(tsdn_t *tsdn) argument 2010 prof_active_set(tsdn_t *tsdn, bool active) argument 2033 prof_thread_name_alloc(tsdn_t *tsdn, const char *thread_name) argument 2110 prof_thread_active_init_get(tsdn_t *tsdn) argument 2121 prof_thread_active_init_set(tsdn_t *tsdn, bool active_init) argument 2133 prof_gdump_get(tsdn_t *tsdn) argument 2144 prof_gdump_set(tsdn_t *tsdn, bool gdump) argument 2192 prof_boot2(tsdn_t *tsdn) argument 2284 prof_prefork0(tsdn_t *tsdn) argument 2301 prof_prefork1(tsdn_t *tsdn) argument 2314 prof_postfork_parent(tsdn_t *tsdn) argument 2337 prof_postfork_child(tsdn_t *tsdn) argument [all...] |
H A D | chunk.c | 52 static void chunk_record(tsdn_t *tsdn, arena_t *arena, 67 chunk_hooks_get(tsdn_t *tsdn, arena_t *arena) argument 71 malloc_mutex_lock(tsdn, &arena->chunks_mtx); 73 malloc_mutex_unlock(tsdn, &arena->chunks_mtx); 79 chunk_hooks_set(tsdn_t *tsdn, arena_t *arena, const chunk_hooks_t *chunk_hooks) argument 83 malloc_mutex_lock(tsdn, &arena->chunks_mtx); 108 malloc_mutex_unlock(tsdn, &arena->chunks_mtx); 114 chunk_hooks_assure_initialized_impl(tsdn_t *tsdn, arena_t *arena, argument 123 chunk_hooks_get(tsdn, arena); 128 chunk_hooks_assure_initialized_locked(tsdn_t *tsdn, arena_ argument 136 chunk_hooks_assure_initialized(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks) argument 144 chunk_register(tsdn_t *tsdn, const void *chunk, const extent_node_t *node) argument 202 chunk_recycle(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, bool dalloc_node) argument 334 chunk_alloc_core(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, dss_prec_t dss_prec) argument 386 chunk_alloc_cache(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool dalloc_node) argument 410 chunk_arena_get(tsdn_t *tsdn, unsigned arena_ind) argument 424 chunk_alloc_default_impl(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument 443 tsdn_t *tsdn; local 454 chunk_alloc_retained(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument 475 chunk_alloc_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument 506 chunk_record(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *chunk, size_t size, bool zeroed, bool committed) argument 598 chunk_dalloc_cache(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed) argument 613 chunk_dalloc_default_impl(tsdn_t *tsdn, void *chunk, size_t size) argument 625 tsdn_t *tsdn; local 633 chunk_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed, bool committed) argument 701 chunk_purge_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset, size_t length) argument 720 chunk_merge_default_impl(tsdn_t *tsdn, void *chunk_a, void *chunk_b) argument 736 tsdn_t *tsdn; local 794 chunk_prefork(tsdn_t *tsdn) argument 801 chunk_postfork_parent(tsdn_t *tsdn) argument 808 chunk_postfork_child(tsdn_t *tsdn) argument [all...] |
H A D | arena.c | 40 static void arena_purge_to_limit(tsdn_t *tsdn, arena_t *arena, 42 static void arena_run_dalloc(tsdn_t *tsdn, arena_t *arena, arena_run_t *run, 44 static void arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena, 595 arena_chunk_register(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, argument 607 return (chunk_register(tsdn, chunk, &chunk->node)); 611 arena_chunk_alloc_internal_hard(tsdn_t *tsdn, arena_t *arena, argument 616 malloc_mutex_unlock(tsdn, &arena->lock); 618 chunk = (arena_chunk_t *)chunk_alloc_wrapper(tsdn, arena, chunk_hooks, 624 chunk_dalloc_wrapper(tsdn, arena, chunk_hooks, 629 if (chunk != NULL && arena_chunk_register(tsdn, aren 645 arena_chunk_alloc_internal(tsdn_t *tsdn, arena_t *arena, bool *zero, bool *commit) argument 675 arena_chunk_init_hard(tsdn_t *tsdn, arena_t *arena) argument 730 arena_chunk_alloc(tsdn_t *tsdn, arena_t *arena) argument 750 arena_chunk_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument 780 arena_spare_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *spare) argument 794 arena_chunk_dalloc(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument 900 arena_node_alloc(tsdn_t *tsdn, arena_t *arena) argument 916 arena_node_dalloc(tsdn_t *tsdn, arena_t *arena, extent_node_t *node) argument 926 arena_chunk_alloc_huge_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, size_t usize, size_t alignment, bool *zero, size_t csize) argument 950 arena_chunk_alloc_huge(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool *zero) argument 978 arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t usize) argument 996 arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1014 arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1040 arena_chunk_ralloc_huge_expand_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument 1069 arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument 1137 arena_run_alloc_large(tsdn_t *tsdn, arena_t *arena, size_t size, bool zero) argument 1181 arena_run_alloc_small(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t binind) argument 1223 arena_lg_dirty_mult_get(tsdn_t *tsdn, arena_t *arena) argument 1235 arena_lg_dirty_mult_set(tsdn_t *tsdn, arena_t *arena, ssize_t lg_dirty_mult) argument 1401 arena_decay_time_get(tsdn_t *tsdn, arena_t *arena) argument 1413 arena_decay_time_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_time) argument 1436 arena_maybe_purge_ratio(tsdn_t *tsdn, arena_t *arena) argument 1464 arena_maybe_purge_decay(tsdn_t *tsdn, arena_t *arena) argument 1499 arena_maybe_purge(tsdn_t *tsdn, arena_t *arena) argument 1547 arena_stash_dirty(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, size_t ndirty_limit, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument 1637 arena_purge_stashed(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument 1730 arena_unstash_purged(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument 1780 arena_purge_to_limit(tsdn_t *tsdn, arena_t *arena, size_t ndirty_limit) argument 1821 arena_purge(tsdn_t *tsdn, arena_t *arena, bool all) argument 2090 arena_run_dalloc(tsdn_t *tsdn, arena_t *arena, arena_run_t *run, bool dirty, bool cleaned, bool decommitted) argument 2166 arena_run_trim_head(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument 2207 arena_run_trim_tail(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument 2274 arena_bin_nonfull_run_get(tsdn_t *tsdn, arena_t *arena, arena_bin_t *bin) argument 2325 arena_bin_malloc_hard(tsdn_t *tsdn, arena_t *arena, arena_bin_t *bin) argument 2375 arena_tcache_fill_small(tsdn_t *tsdn, arena_t *arena, tcache_bin_t *tbin, szind_t binind, uint64_t prof_accumbytes) argument 2535 arena_malloc_small(tsdn_t *tsdn, arena_t *arena, szind_t binind, bool zero) argument 2589 arena_malloc_large(tsdn_t *tsdn, arena_t *arena, szind_t binind, bool zero) argument 2651 arena_malloc_hard(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero) argument 2671 arena_palloc_large(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 2759 arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 2794 arena_prof_promoted(tsdn_t *tsdn, const void *ptr, size_t size) argument 2845 arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 2884 arena_dalloc_bin_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm, bool junked) argument 2917 arena_dalloc_bin_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm) argument 2925 arena_dalloc_bin(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_bits_t *bitselm) argument 2941 arena_dalloc_small(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument 2975 arena_dalloc_large_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, bool junked) argument 3003 arena_dalloc_large_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 3011 arena_dalloc_large(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 3022 arena_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument 3059 arena_ralloc_large_grow(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument 3183 arena_ralloc_large(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument 3221 arena_ralloc_no_move(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, size_t extra, bool zero) argument 3266 arena_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 3325 arena_dss_prec_get(tsdn_t *tsdn, arena_t *arena) argument 3336 arena_dss_prec_set(tsdn_t *tsdn, arena_t *arena, dss_prec_t dss_prec) argument 3400 arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty) argument 3412 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty, arena_stats_t *astats, malloc_bin_stats_t *bstats, malloc_large_stats_t *lstats, malloc_huge_stats_t *hstats) argument 3496 arena_new(tsdn_t *tsdn, unsigned ind) argument 3837 arena_prefork0(tsdn_t *tsdn, arena_t *arena) argument 3844 arena_prefork1(tsdn_t *tsdn, arena_t *arena) argument 3851 arena_prefork2(tsdn_t *tsdn, arena_t *arena) argument 3858 arena_prefork3(tsdn_t *tsdn, arena_t *arena) argument 3868 arena_postfork_parent(tsdn_t *tsdn, arena_t *arena) argument 3881 arena_postfork_child(tsdn_t *tsdn, arena_t *arena) argument [all...] |
H A D | mutex.c | 123 malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 126 malloc_mutex_lock(tsdn, mutex); 130 malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 133 malloc_mutex_unlock(tsdn, mutex); 137 malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 141 malloc_mutex_unlock(tsdn, mutex);
|
H A D | ckh.c | 43 static bool ckh_grow(tsdn_t *tsdn, ckh_t *ckh); 44 static void ckh_shrink(tsdn_t *tsdn, ckh_t *ckh); 247 ckh_grow(tsdn_t *tsdn, ckh_t *ckh) argument 273 tab = (ckhc_t *)ipallocztm(tsdn, usize, CACHELINE, true, NULL, 274 true, arena_ichoose(tsdn, NULL)); 286 idalloctm(tsdn, tab, NULL, true, true); 291 idalloctm(tsdn, ckh->tab, NULL, true, true); 302 ckh_shrink(tsdn_t *tsdn, ckh_t *ckh) argument 317 tab = (ckhc_t *)ipallocztm(tsdn, usize, CACHELINE, true, NULL, true, 318 arena_ichoose(tsdn, NUL 350 ckh_new(tsdn_t *tsdn, ckh_t *ckh, size_t minitems, ckh_hash_t *hash, ckh_keycomp_t *keycomp) argument 407 ckh_delete(tsdn_t *tsdn, ckh_t *ckh) argument 459 ckh_insert(tsdn_t *tsdn, ckh_t *ckh, const void *key, const void *data) argument 483 ckh_remove(tsdn_t *tsdn, ckh_t *ckh, const void *searchkey, void **key, void **data) argument [all...] |
H A D | tcache.c | 27 tcache_salloc(tsdn_t *tsdn, const void *ptr) argument 30 return (arena_salloc(tsdn, ptr, false)); 74 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, argument 79 arena_tcache_fill_small(tsdn, arena, tbin, binind, config_prof ? 256 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument 261 malloc_mutex_lock(tsdn, &arena->lock); 264 malloc_mutex_unlock(tsdn, &arena->lock); 269 tcache_arena_dissociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument 274 malloc_mutex_lock(tsdn, &arena->lock); 287 tcache_stats_merge(tsdn, tcach 293 tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *oldarena, arena_t *newarena) argument 318 tcache_create(tsdn_t *tsdn, arena_t *arena) argument 420 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument 448 tcaches_create(tsdn_t *tsdn, unsigned *r_ind) argument 512 tcache_boot(tsdn_t *tsdn) argument [all...] |
H A D | jemalloc.c | 420 arena_init_locked(tsdn_t *tsdn, unsigned ind) argument 434 arena = arena_get(tsdn, ind, false); 441 arena = arena_new(tsdn, ind); 447 arena_init(tsdn_t *tsdn, unsigned ind) argument 451 malloc_mutex_lock(tsdn, &arenas_lock); 452 arena = arena_init_locked(tsdn, ind); 453 malloc_mutex_unlock(tsdn, &arenas_lock); 732 tsdn_t *tsdn; local 735 tsdn = tsdn_fetch(); 745 arena_t *arena = arena_get(tsdn, 1356 malloc_init_hard_finish(tsdn_t *tsdn) argument 1496 ialloc_body(size_t size, bool zero, tsdn_t **tsdn, size_t *usize, bool slow_path) argument 1528 ialloc_post_check(void *ret, tsdn_t *tsdn, size_t usize, const char *func, bool update_errno, bool slow_path) argument 1556 tsdn_t *tsdn; local 1720 tsdn_t *tsdn; local 2035 imallocx_flags(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument 2049 imallocx_prof_sample(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument 2123 imallocx_body(size_t size, int flags, tsdn_t **tsdn, size_t *usize, bool slow_path) argument 2166 tsdn_t *tsdn; local 2330 ixallocx_helper(tsdn_t *tsdn, void *ptr, size_t old_usize, size_t size, size_t extra, size_t alignment, bool zero) argument 2343 ixallocx_prof_sample(tsdn_t *tsdn, void *ptr, size_t old_usize, size_t size, size_t extra, size_t alignment, bool zero, prof_tctx_t *tctx) argument 2471 tsdn_t *tsdn; local 2516 inallocx(tsdn_t *tsdn, size_t size, int flags) argument 2565 tsdn_t *tsdn; local 2604 tsdn_t *tsdn; local 2637 tsdn_t *tsdn; local 2649 tsdn_t *tsdn; local [all...] |
H A D | quarantine.c | 16 static void quarantine_drain_one(tsdn_t *tsdn, quarantine_t *quarantine); 17 static void quarantine_drain(tsdn_t *tsdn, quarantine_t *quarantine, 23 quarantine_init(tsdn_t *tsdn, size_t lg_maxobjs) argument 30 quarantine = (quarantine_t *)iallocztm(tsdn, size, size2index(size), 97 quarantine_drain_one(tsdn_t *tsdn, quarantine_t *quarantine) argument 100 assert(obj->usize == isalloc(tsdn, obj->ptr, config_prof)); 101 idalloctm(tsdn, obj->ptr, NULL, false, true); 109 quarantine_drain(tsdn_t *tsdn, quarantine_t *quarantine, size_t upper_bound) argument 113 quarantine_drain_one(tsdn, quarantine);
|
H A D | ctl.c | 49 static const ctl_named_node_t *n##_index(tsdn_t *tsdn, \ 54 static void ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_stats_t *cstats, 58 static void ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, unsigned i); 59 static bool ctl_grow(tsdn_t *tsdn); 60 static void ctl_refresh(tsdn_t *tsdn); 61 static bool ctl_init(tsdn_t *tsdn); 62 static int ctl_lookup(tsdn_t *tsdn, const char *name, 120 static void arena_i_purge(tsdn_t *tsdn, unsigned arena_ind, bool all); 563 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_stats_t *cstats, arena_t *arena) argument 568 arena_stats_merge(tsdn, aren 659 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, unsigned i) argument 671 ctl_grow(tsdn_t *tsdn) argument 711 ctl_refresh(tsdn_t *tsdn) argument 761 ctl_init(tsdn_t *tsdn) argument 818 ctl_lookup(tsdn_t *tsdn, const char *name, ctl_node_t const **nodesp, size_t *mibp, size_t *depthp) argument 947 ctl_nametomib(tsdn_t *tsdn, const char *name, size_t *mibp, size_t *miblenp) argument 1024 ctl_prefork(tsdn_t *tsdn) argument 1031 ctl_postfork_parent(tsdn_t *tsdn) argument 1038 ctl_postfork_child(tsdn_t *tsdn) argument 1544 arena_i_purge(tsdn_t *tsdn, unsigned arena_ind, bool all) argument 1807 arena_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i) argument 1934 arenas_bin_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i) argument 1945 arenas_lrun_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i) argument 1957 arenas_hchunk_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i) argument 2191 stats_arenas_i_bins_j_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t j) argument 2210 stats_arenas_i_lruns_j_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t j) argument 2230 stats_arenas_i_hchunks_j_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t j) argument 2240 stats_arenas_i_index(tsdn_t *tsdn, const size_t *mib, size_t miblen, size_t i) argument [all...] |