/freebsd-current/contrib/jemalloc/include/jemalloc/internal/ |
H A D | jemalloc_internal_inlines_b.h | 79 iealloc(tsdn_t *tsdn, const void *ptr) { argument
|
H A D | prof_inlines_a.h | 7 prof_accum_add(tsdn_t *tsdn, prof_accum_t *prof_accum, argument 46 prof_accum_cancel(tsdn_t *tsdn, prof_accum_t *prof_accum, argument
|
H A D | jemalloc_internal_inlines_a.h | 85 arena_get(tsdn_t *tsdn, unsigned ind, bool init_if_missing) { argument
|
H A D | arena_inlines_a.h | 25 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes) { argument
|
H A D | background_thread_inlines.h | 10 background_thread_enabled_set(tsdn_t *tsdn, bool state) { argument 35 background_thread_wakeup_time_set(tsdn_t *tsdn, background_thread_info_t *info, argument 49 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, argument
|
H A D | mutex_pool.h | 27 mutex_pool_assert_not_held(tsdn_t *tsdn, mutex_pool_t *pool) { argument 41 mutex_pool_lock(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) { argument 49 mutex_pool_unlock(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) { argument 57 mutex_pool_lock2(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key1, argument 75 mutex_pool_unlock2(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key1, uintptr_t key2) argument 90 mutex_pool_assert_owner(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) argument [all...] |
H A D | bin.h | 107 bin_stats_merge(tsdn_t *tsdn, bin_stats_t *dst_bin_stats, bin_t *bin) { argument
|
H A D | prof_inlines_b.h | 41 prof_tctx_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) { argument 49 prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, argument 58 prof_tctx_reset(tsdn_t *tsdn, const void *ptr, prof_tctx_t *tctx) { argument 66 prof_alloc_time_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) { argument 74 prof_alloc_time_set(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx, argument 166 prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, alloc_ctx_t *alloc_ctx, prof_tctx_t *tctx) argument [all...] |
H A D | arena_inlines_b.h | 38 arena_prof_tctx_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) { argument 57 arena_prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, argument 76 arena_prof_tctx_reset(tsdn_t *tsdn, const void *ptr, prof_tctx_t *tctx) argument 87 arena_prof_alloc_time_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) argument 102 arena_prof_alloc_time_set(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx, nstime_t t) argument 113 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) argument 131 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) argument 140 arena_decay_extent(tsdn_t *tsdn,arena_t *arena, extent_hooks_t **r_extent_hooks, extent_t *extent) argument 159 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, tcache_t *tcache, bool slow_path) argument 180 arena_aalloc(tsdn_t *tsdn, const void *ptr) argument 185 arena_salloc(tsdn_t *tsdn, const void *ptr) argument 199 arena_vsalloc(tsdn_t *tsdn, const void *ptr) argument 232 arena_dalloc_large_no_tcache(tsdn_t *tsdn, void *ptr, szind_t szind) argument 242 arena_dalloc_no_tcache(tsdn_t *tsdn, void *ptr) argument [all...] |
H A D | extent_inlines.h | 13 extent_lock(tsdn_t *tsdn, extent_t *extent) { argument 19 extent_unlock(tsdn_t *tsdn, extent_t *extent) { argument 25 extent_lock2(tsdn_t *tsdn, extent_t *extent1, extent_t *extent2) { argument 32 extent_unlock2(tsdn_t *tsdn, extent_t *extent1, extent_t *extent2) { argument [all...] |
H A D | mutex.h | 163 mutex_owner_stats_update(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument 176 malloc_mutex_trylock(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument 215 malloc_mutex_lock(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 228 malloc_mutex_unlock(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 237 malloc_mutex_assert_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 242 malloc_mutex_assert_not_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) argument 248 malloc_mutex_prof_read(tsdn_t *tsdn, mutex_prof_data_t *data, malloc_mutex_t *mutex) argument 265 malloc_mutex_prof_accum(tsdn_t *tsdn, mutex_prof_data_t *data, malloc_mutex_t *mutex) argument [all...] |
H A D | jemalloc_internal_inlines_c.h | 28 iaalloc(tsdn_t *tsdn, const void *ptr) { argument 35 isalloc(tsdn_t *tsdn, const void *ptr) { argument 42 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, argument 67 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, bool is_internal, arena_t *arena) argument 87 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument 99 ivsalloc(tsdn_t *tsdn, const void *ptr) argument 104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx, bool is_internal, bool slow_path) argument 127 isdalloct(tsdn_t *tsdn, void *ptr, size_t size, tcache_t *tcache, alloc_ctx_t *alloc_ctx, bool slow_path) argument 135 iralloct_realign(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, hook_ralloc_args_t *hook_args) argument 174 iralloct(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, hook_ralloc_args_t *hook_args) argument 204 ixalloc(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, size_t extra, size_t alignment, bool zero, size_t *newsize) argument [all...] |
H A D | tsd.h | 230 tsdn_null(const tsdn_t *tsdn) { argument 235 tsdn_tsd(tsdn_t *tsdn) { argument 404 tsdn_rtree_ctx(tsdn_t *tsdn, rtree_ctx_t *fallback) argument [all...] |
H A D | arena_stats.h | 131 arena_stats_init(tsdn_t *tsdn, arena_stats_t *arena_stats) { argument 148 arena_stats_lock(tsdn_t *tsdn, arena_stats_t *arena_stats) { argument 155 arena_stats_unlock(tsdn_t *tsdn, arena_stats_t *arena_stats) { argument 162 arena_stats_read_u64(tsdn_t *tsdn, arena_stats_t *arena_stats, argument 173 arena_stats_add_u64(tsdn_t *tsdn, arena_stats_t *arena_stats, argument 184 arena_stats_sub_u64(tsdn_t *tsdn, arena_stats_ argument 212 arena_stats_read_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p) argument 223 arena_stats_add_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p, size_t x) argument 235 arena_stats_sub_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p, size_t x) argument 255 arena_stats_large_flush_nrequests_add(tsdn_t *tsdn, arena_stats_t *arena_stats, szind_t szind, uint64_t nrequests) argument 265 arena_stats_mapped_add(tsdn_t *tsdn, arena_stats_t *arena_stats, size_t size) argument [all...] |
H A D | rtree.h | 173 rtree_leaf_elm_bits_read(tsdn_t *tsdn, rtree_t *rtree, argument 211 rtree_leaf_elm_extent_read(tsdn_t *tsdn, rtree_t *rtree, argument 224 rtree_leaf_elm_szind_read(tsdn_t *tsdn, rtree_t *rtree, argument 236 rtree_leaf_elm_slab_read(tsdn_t *tsdn, rtree_t *rtree, argument 248 rtree_leaf_elm_extent_write(tsdn_t *tsdn, rtree_ argument 262 rtree_leaf_elm_szind_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm, szind_t szind) argument 280 rtree_leaf_elm_slab_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm, bool slab) argument 295 rtree_leaf_elm_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm, extent_t *extent, szind_t szind, bool slab) argument 314 rtree_leaf_elm_szind_slab_update(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm, szind_t szind, bool slab) argument 327 rtree_leaf_elm_lookup(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, bool init_missing) argument 386 rtree_write(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, extent_t *extent, szind_t szind, bool slab) argument 404 rtree_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent) argument 416 rtree_extent_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent) argument 427 rtree_szind_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent) argument 443 rtree_extent_szind_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, extent_t **r_extent, szind_t *r_szind) argument 463 rtree_szind_slab_read_fast(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, szind_t *r_szind, bool *r_slab) argument 492 rtree_szind_slab_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, szind_t *r_szind, bool *r_slab) argument 511 rtree_szind_slab_update(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, szind_t szind, bool slab) argument 520 rtree_clear(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key) argument [all...] |
/freebsd-current/contrib/jemalloc/src/ |
H A D | background_thread.c | 895 background_thread_boot1(tsdn_t *tsdn) argument [all...] |
H A D | bin.c | 83 bin_prefork(tsdn_t *tsdn, bin_t *bin) { argument 88 bin_postfork_parent(tsdn_t *tsdn, bin_t *bin) { argument 93 bin_postfork_child(tsdn_t *tsdn, bin_t *bin) { argument
|
H A D | hook.c | 48 hook_install(tsdn_t *tsdn, hooks_t *to_install) { argument 73 hook_remove(tsdn_t *tsdn, void *opaque) { argument 132 tsdn_t *tsdn = tsdn_fetch(); local [all...] |
H A D | mutex.c | 124 malloc_mutex_prof_data_reset(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument 206 malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument 211 malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument 216 malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex) { argument
|
H A D | extent_dss.c | 110 extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, argument
|
H A D | rtree.c | 31 rtree_node_alloc_impl(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { argument 38 rtree_node_dalloc_impl(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *node) { argument 46 rtree_leaf_alloc_impl(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { argument 53 rtree_leaf_dalloc_impl(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *leaf) { argument 63 rtree_delete_subtree(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *subtree, argument 94 rtree_delete(tsdn_t *tsdn, rtree_t *rtree) argument 102 rtree_node_init(tsdn_t *tsdn, rtree_t *rtree, unsigned level, atomic_p_t *elmp) argument 129 rtree_leaf_init(tsdn_t *tsdn, rtree_t *rtree, atomic_p_t *elmp) argument 181 rtree_child_node_read(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *elm, unsigned level, bool dependent) argument 210 rtree_child_leaf_read(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *elm, unsigned level, bool dependent) argument 223 rtree_leaf_elm_lookup_hard(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, bool init_missing) argument [all...] |
H A D | large.c | 14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { argument 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument 96 large_ralloc_no_move_shrink(tsdn_t *tsdn, extent_t *extent, size_t usize) argument 131 large_ralloc_no_move_expand(tsdn_t *tsdn, extent_t *extent, size_t usize, bool zero) argument 219 large_ralloc_no_move(tsdn_t *tsdn, extent_t *extent, size_t usize_min, size_t usize_max, bool zero) argument 265 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 274 large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize, size_t alignment, bool zero, tcache_t *tcache, hook_ralloc_args_t *hook_args) argument 322 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent, bool junked_locked) argument 344 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent) argument 350 large_dalloc_prep_junked_locked(tsdn_t *tsdn, extent_t *extent) argument 355 large_dalloc_finish(tsdn_t *tsdn, extent_t *extent) argument 360 large_dalloc(tsdn_t *tsdn, extent_t *extent) argument 368 large_salloc(tsdn_t *tsdn, const extent_t *extent) argument 373 large_prof_tctx_get(tsdn_t *tsdn, const extent_t *extent) argument 378 large_prof_tctx_set(tsdn_t *tsdn, extent_t *extent, prof_tctx_t *tctx) argument 383 large_prof_tctx_reset(tsdn_t *tsdn, extent_t *extent) argument [all...] |
H A D | tsd.c | 107 tsd_force_recompute(tsdn_t *tsdn) { argument 125 tsd_global_slow_inc(tsdn_t *tsdn) { argument 138 void tsd_global_slow_dec(tsdn_t *tsdn) { argument
|
H A D | tcache.c | 36 tcache_salloc(tsdn_t *tsdn, const void *ptr) { argument 89 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, argument 106 tbin_extents_lookup_size_check(tsdn_t *tsdn, cache_bin_t *tbin, szind_t binind, argument 341 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { argument 362 tcache_arena_dissociate(tsdn_t *tsdn, tcache_t *tcache) argument 389 tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument 601 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) argument 733 tcache_boot(tsdn_t *tsdn) argument 780 tcache_prefork(tsdn_t *tsdn) argument 787 tcache_postfork_parent(tsdn_t *tsdn) argument 794 tcache_postfork_child(tsdn_t *tsdn) argument [all...] |
H A D | base.c | 32 base_map(tsdn_t *tsdn, extent_hooks_t *extent_hooks, unsigned ind, size_t size) { argument 55 base_unmap(tsdn_t *tsdn, extent_hooks_t *extent_hooks, unsigned ind, void *addr, argument 144 base_auto_thp_switch(tsdn_t *tsdn, base_t *base) { argument 248 base_block_alloc(tsdn_t *tsdn, base_t *base, extent_hooks_t *extent_hooks, argument 308 base_extent_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) argument 350 base_new(tsdn_t *tsdn, unsigned ind, extent_hooks_t *extent_hooks) argument 396 base_delete(tsdn_t *tsdn, base_t *base) argument 421 base_alloc_impl(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment, size_t *esn) argument 464 base_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) argument 469 base_alloc_extent(tsdn_t *tsdn, base_t *base) argument 481 base_stats_get(tsdn_t *tsdn, base_t *base, size_t *allocated, size_t *resident, size_t *mapped, size_t *n_thp) argument 496 base_prefork(tsdn_t *tsdn, base_t *base) argument 501 base_postfork_parent(tsdn_t *tsdn, base_t *base) argument 506 base_postfork_child(tsdn_t *tsdn, base_t *base) argument 511 base_boot(tsdn_t *tsdn) argument [all...] |