Lines Matching defs:arena

391 arena_set(unsigned ind, arena_t *arena)
394 atomic_write_p((void **)&arenas[ind], arena);
418 /* Create a new arena and insert it into the arenas array at index ind. */
422 arena_t *arena;
432 * auto arena.
434 arena = arena_get(tsdn, ind, false);
435 if (arena != NULL) {
437 return (arena);
440 /* Actually initialize the arena. */
441 arena = arena_new(tsdn, ind);
442 arena_set(ind, arena);
443 return (arena);
449 arena_t *arena;
452 arena = arena_init_locked(tsdn, ind);
454 return (arena);
460 arena_t *arena;
462 arena = arena_get(tsd_tsdn(tsd), ind, false);
463 arena_nthreads_inc(arena, internal);
467 tsd_iarena_set(tsd, arena);
469 tsd_arena_set(tsd, arena);
488 arena_t *arena;
490 arena = arena_get(tsd_tsdn(tsd), ind, false);
491 arena_nthreads_dec(arena, internal);
600 * Choose the first arena that has the lowest
614 * arena, in case all extant arenas are in use.
619 * "thread.arena" mallctl.
630 * Use an unloaded arena, or the least loaded
631 * arena if all arenas are already initialized.
638 arena_t *arena;
640 /* Initialize a new arena. */
642 arena = arena_init_locked(tsd_tsdn(tsd),
644 if (arena == NULL) {
650 ret = arena;
691 arena_t *arena;
693 arena = tsd_arena_get(tsd);
694 if (arena != NULL)
695 arena_unbind(tsd, arena->ind, false);
745 arena_t *arena = arena_get(tsdn, i, false);
746 if (arena != NULL) {
751 * code is introduced that acquires both arena
755 malloc_mutex_lock(tsdn, &arena->lock);
756 ql_foreach(tcache, &arena->tcache_ql, link) {
757 tcache_stats_merge(tsdn, tcache, arena);
759 malloc_mutex_unlock(tsdn, &arena->lock);
1309 * Initialize one arena here. The rest are lazily created in
1364 * For SMP systems, create more than one arena per CPU by
1388 /* Copy the pointer to the one arena that was already initialized. */
2013 size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena)
2035 *arena = arena_get(tsd_tsdn(tsd), arena_ind, true);
2036 if (unlikely(*arena == NULL))
2039 *arena = NULL;
2045 tcache_t *tcache, arena_t *arena, bool slow_path)
2050 return (ipalloct(tsdn, usize, alignment, zero, tcache, arena));
2053 return (iallocztm(tsdn, usize, ind, zero, tcache, false, arena,
2059 tcache_t *tcache, arena_t *arena, bool slow_path)
2067 tcache, arena, slow_path);
2072 p = imallocx_flags(tsdn, usize, alignment, zero, tcache, arena,
2086 arena_t *arena;
2090 &zero, &tcache, &arena)))
2095 tcache, arena, slow_path);
2098 tcache, arena, slow_path);
2119 arena_t *arena;
2122 &zero, &tcache, &arena)))
2125 arena, slow_path);
2197 size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena,
2206 zero, tcache, arena);
2212 tcache, arena);
2221 arena_t *arena)
2232 alignment, zero, tcache, arena, tctx);
2235 tcache, arena);
2271 arena_t *arena;
2283 arena = arena_get(tsd_tsdn(tsd), arena_ind, true);
2284 if (unlikely(arena == NULL))
2287 arena = NULL;
2306 zero, tcache, arena);
2311 tcache, arena);
2818 arena_t *arena;
2837 if ((arena = arena_get(tsd_tsdn(tsd), j, false)) !=
2841 arena_prefork0(tsd_tsdn(tsd), arena);
2844 arena_prefork1(tsd_tsdn(tsd), arena);
2847 arena_prefork2(tsd_tsdn(tsd), arena);
2857 if ((arena = arena_get(tsd_tsdn(tsd), i, false)) != NULL)
2858 arena_prefork3(tsd_tsdn(tsd), arena);
2887 arena_t *arena;
2889 if ((arena = arena_get(tsd_tsdn(tsd), i, false)) != NULL)
2890 arena_postfork_parent(tsd_tsdn(tsd), arena);
2912 arena_t *arena;
2914 if ((arena = arena_get(tsd_tsdn(tsd), i, false)) != NULL)
2915 arena_postfork_child(tsd_tsdn(tsd), arena);