/freebsd-current/contrib/jemalloc/include/jemalloc/internal/ |
H A D | safety_check.h | 9 safety_check_set_redzone(void *ptr, size_t usize, size_t bumped_usize) { argument 10 assert(usize < bumped_usize); 11 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { 17 safety_check_verify_redzone(const void *ptr, size_t usize, size_t bumped_usize) 19 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) {
|
H A D | prof_inlines_b.h | 49 prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, argument 54 arena_prof_tctx_set(tsdn, ptr, usize, alloc_ctx, tctx); 83 prof_sample_check(tsd_t *tsd, size_t usize, bool update) { argument 84 ssize_t check = update ? 0 : usize; 88 bytes_until_sample -= usize; 101 prof_sample_accum_update(tsd_t *tsd, size_t usize, bool update, argument 108 if (likely(prof_sample_check(tsd, usize, update))) { 131 if (!booted && prof_sample_check(tsd, usize, update)) { 146 prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, bool update) { argument 151 assert(usize 166 prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, alloc_ctx_t *alloc_ctx, prof_tctx_t *tctx) argument 181 prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx, bool prof_active, bool updated, const void *old_ptr, size_t old_usize, prof_tctx_t *old_tctx) argument 209 prof_malloc_sample_object(tsd_tsdn(tsd), ptr, usize, tctx); local 211 prof_tctx_set(tsd_tsdn(tsd), ptr, usize, NULL, local 239 prof_free(tsd_t *tsd, const void *ptr, size_t usize, alloc_ctx_t *alloc_ctx) argument [all...] |
H A D | prof_inlines_a.h | 47 size_t usize) { 60 a1 = (a0 >= SC_LARGE_MINCLASS - usize) 61 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; 67 a1 = (a0 >= SC_LARGE_MINCLASS - usize) 68 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; 46 prof_accum_cancel(tsdn_t *tsdn, prof_accum_t *prof_accum, size_t usize) argument
|
H A D | sz.h | 116 size_t usize = (psz + delta_mask) & ~delta_mask; local 117 return usize; 192 size_t usize = grp_size + mod_size; local 193 return usize; 233 size_t usize = (size + delta_mask) & ~delta_mask; local 234 return usize; 264 size_t usize; local 284 usize = sz_s2u(ALIGNMENT_CEILING(size, alignment)); 285 if (usize < SC_LARGE_MINCLASS) { 286 return usize; [all...] |
H A D | jemalloc_internal_inlines_c.h | 67 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, argument 71 assert(usize != 0); 72 assert(usize == sz_sa2u(usize, alignment)); 78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); 87 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, argument 89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); 93 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero) { argument 94 return ipallocztm(tsd_tsdn(tsd), usize, alignment, zero, 141 size_t usize, copysiz local [all...] |
H A D | tcache_inlines.h | 47 size_t usize JEMALLOC_CC_SILENCE_INIT(0); 69 * Only compute usize if required. The checks in the following if 73 usize = sz_index2size(binind); 74 assert(tcache_salloc(tsd_tsdn(tsd), ret) == usize); 83 memset(ret, 0, usize); 90 memset(ret, 0, usize); 97 tcache->prof_accumbytes += usize; 129 size_t usize JEMALLOC_CC_SILENCE_INIT(0); 131 /* Only compute usize on demand */ 134 usize [all...] |
H A D | large_externs.h | 6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
|
H A D | prof_externs.h | 47 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, 49 void prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize,
|
H A D | arena_externs.h | 36 size_t usize, size_t alignment, bool *zero); 61 void *arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, 63 void arena_prof_promote(tsdn_t *tsdn, void *ptr, size_t usize);
|
H A D | arena_inlines_b.h | 57 arena_prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, argument
|
/freebsd-current/contrib/jemalloc/src/ |
H A D | large.c | 14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { argument 15 assert(usize == sz_s2u(usize)); 17 return large_palloc(tsdn, arena, usize, CACHELINE, zero); 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument 30 ausize = sz_sa2u(usize, alignment); 45 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize); 48 arena, usize, alignment, &is_zeroed)) == NULL) { 59 if (config_prof && arena_prof_accum(tsdn, arena, usize)) { 96 large_ralloc_no_move_shrink(tsdn_t *tsdn, extent_t *extent, size_t usize) { argument 131 large_ralloc_no_move_expand(tsdn_t *tsdn, extent_t *extent, size_t usize, bool zero) argument 265 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 274 large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize, size_t alignment, bool zero, tcache_t *tcache, hook_ralloc_args_t *hook_args) argument [all...] |
H A D | ckh.c | 274 size_t usize; local 277 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 278 if (unlikely(usize == 0 279 || usize > SC_LARGE_MAXCLASS)) { 283 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, 314 size_t usize; local 323 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 324 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { 327 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELIN 363 size_t mincells, usize; local [all...] |
H A D | jemalloc.c | 1864 bool usize; member in struct:static_opts_s 1878 static_opts->usize = false; 1893 size_t usize; member in struct:dynamic_opts_s 1905 dynamic_opts->usize = 0; 1917 size_t size, size_t usize, szind_t ind) { 1949 return ipalloct(tsd_tsdn(tsd), usize, dopts->alignment, 1959 size_t usize, szind_t ind) { 1963 * For small allocations, sampling bumps the usize. If so, we allocate 1967 size_t bumped_usize = usize; 1969 if (usize < 1916 imalloc_no_sample(static_opts_t *sopts, dynamic_opts_t *dopts, tsd_t *tsd, size_t size, size_t usize, szind_t ind) argument 1958 imalloc_sample(static_opts_t *sopts, dynamic_opts_t *dopts, tsd_t *tsd, size_t usize, szind_t ind) argument 1981 arena_prof_promote(tsd_tsdn(tsd), ret, usize); local 2047 size_t usize = 0; local 2145 prof_malloc(tsd_tsdn(tsd), allocation, usize, &alloc_ctx, tctx); local 2346 size_t usize; local 2510 irealloc_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, prof_tctx_t *tctx, hook_ralloc_args_t *hook_args) argument 2523 arena_prof_promote(tsd_tsdn(tsd), p, usize); local 2533 irealloc_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, alloc_ctx_t *alloc_ctx, hook_ralloc_args_t *hook_args) argument 2578 size_t usize; local 2599 isfree(tsd_t *tsd, void *ptr, size_t usize, tcache_t *tcache, bool slow_path) argument 2648 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, ctx, false); local 2650 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, ctx, true); local 2840 size_t usize = sz_index2size(alloc_ctx.szind); local 3134 irallocx_prof_sample(tsdn_t *tsdn, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx, hook_ralloc_args_t *hook_args) argument 3159 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena, alloc_ctx_t *alloc_ctx, hook_ralloc_args_t *hook_args) argument 3204 size_t usize; local 3312 size_t usize; local 3326 size_t usize_max, usize; local 3377 size_t usize, old_usize; local 3447 size_t usize; local 3522 size_t usize; local 3539 size_t usize = inallocx(tsd_tsdn(tsd), size, flags); local 3605 size_t usize; local 3777 size_t usize = je_xallocx(*ptr, size, extra, flags); local 3812 size_t usize = je_nallocx(size, flags); local [all...] |
H A D | arena.c | 384 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { argument 389 if (usize < SC_LARGE_MINCLASS) { 390 usize = SC_LARGE_MINCLASS; 392 index = sz_size2index(usize); 400 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { argument 405 if (usize < SC_LARGE_MINCLASS) { 406 usize = SC_LARGE_MINCLASS; 408 index = sz_size2index(usize); 417 size_t usize) { 419 arena_large_malloc_stats_update(tsdn, arena, usize); 416 arena_large_ralloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t oldusize, size_t usize) argument 428 arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool *zero) argument 493 size_t usize = extent_usize_get(extent); local 507 size_t usize = extent_usize_get(extent); local 1117 size_t usize; local 1463 size_t usize; local 1532 arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 1553 arena_prof_promote(tsdn_t *tsdn, void *ptr, size_t usize) argument 1603 size_t usize = extent_usize_get(extent); local 1786 arena_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 1803 size_t usize = sz_s2u(size); local [all...] |
H A D | base.c | 252 size_t usize = ALIGNMENT_CEILING(size, alignment); local 264 + usize)); 424 size_t usize = ALIGNMENT_CEILING(size, alignment); local 425 size_t asize = usize + alignment - QUANTUM; 438 extent = base_extent_alloc(tsdn, base, usize, alignment); 446 ret = base_extent_bump_alloc(base, extent, usize, alignment);
|
H A D | prof.c | 146 size_t usize; member in struct:prof_alloc_node_s 343 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, argument 345 prof_tctx_set(tsdn, ptr, usize, NULL, tctx); 355 tctx->cnts.curbytes += usize; 358 tctx->cnts.accumbytes += usize; 444 prof_try_log(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx) { argument 506 new_node->usize = usize; 521 prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize, argument 526 assert(tctx->cnts.curbytes >= usize); [all...] |
/freebsd-current/bin/ed/ |
H A D | undo.c | 34 static long usize = 0; /* stack size variable */ variable 45 (ustack = (undo_t *) malloc((usize = USIZE) * sizeof(undo_t))) == NULL) { 52 if (u_p < usize || 53 (t = (undo_t *) realloc(ustack, (usize += USIZE) * sizeof(undo_t))) != NULL) { 66 usize = 0;
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/mount/ |
H A D | umount_unlinked_drain.ksh | 52 usize=$(($nunlinks - $nunlinked)) 53 if [[ $iters == $MAX_ITERS && $usize == $1 ]]; then 56 if [[ $usize == $last_usize ]]; then 61 last_usize=$usize
|
/freebsd-current/contrib/libarchive/libarchive/ |
H A D | archive_write_add_filter_lzop.c | 321 lzo_uint usize, csize; local 331 usize = (lzo_uint) 337 r = lzo1x_1_compress(data->uncompressed, usize, 342 r = lzo1x_1_15_compress(data->uncompressed, usize, 347 r = lzo1x_999_compress_level(data->uncompressed, usize, 359 archive_be32enc(p + header_bytes, (uint32_t)usize); 361 checksum = lzo_adler32(1, data->uncompressed, usize); 364 if (csize < usize) { 374 archive_be32enc(p + header_bytes + 4, (uint32_t)usize); 380 usize); [all...] |
/freebsd-current/usr.bin/gzip/ |
H A D | gzip.c | 1750 off_t usize, gsize; local 1793 usize = cat_fd(fourbytes, sizeof fourbytes, &gsize, STDIN_FILENO); 1796 usize = gz_uncompress(STDIN_FILENO, STDOUT_FILENO, 1801 usize = unbzip2(STDIN_FILENO, STDOUT_FILENO, 1812 usize = zuncompress(in, stdout, (char *)fourbytes, 1819 usize = unpack(STDIN_FILENO, STDOUT_FILENO, 1825 usize = unxz(STDIN_FILENO, STDOUT_FILENO, 1831 usize = unlz(STDIN_FILENO, STDOUT_FILENO, 1837 usize = unzstd(STDIN_FILENO, STDOUT_FILENO, 1843 if (vflag && !tflag && usize ! 1856 off_t usize; local 1953 off_t usize, gsize; local 2054 print_verbage(const char *file, const char *nfile, off_t usize, off_t gsize) argument 2110 uint32_t usize; local [all...] |
/freebsd-current/sys/compat/linux/ |
H A D | linux_fork.c | 437 if (args->usize > PAGE_SIZE) 439 if (args->usize < LINUX_CLONE_ARGS_SIZE_VER0) 443 * usize can be less than size of struct clone_args, to avoid using 447 size = max(args->usize, sizeof(*uca)); 449 error = copyin(args->uargs, uca, args->usize);
|
/freebsd-current/sys/contrib/openzfs/cmd/ |
H A D | dbufstat.in | 40 "usize", "meta", "state", "dbholds", "dbc", "list", "atype", "flags", 50 dincompat = ["level", "blkid", "offset", "dbsize", "usize", "meta", "state", 58 tincompat = ["object", "level", "blkid", "offset", "dbsize", "usize", "meta", 73 "usize": [7, 1024, "size of attached user data"], 403 usize = int(line[labels['usize']]) 421 d[pool][objset][key]['cached'] += dbsize + usize
|
/freebsd-current/sys/dev/iwi/ |
H A D | if_iwireg.h | 165 uint32_t usize; /* size of ucode image */ member in struct:iwi_firmware_hdr
|
/freebsd-current/sys/dev/qat_c2xxx/ |
H A D | qat_ae.c | 1420 int ii, vali, fixup, usize = 0; local 1432 usize = nitems(ae_inst_1b); 1436 usize = nitems(ae_inst_2b); 1440 usize = nitems(ae_inst_3b); 1444 usize = nitems(ae_inst_4b); 1449 for (ii = 0; ii < usize; ii++)
|
/freebsd-current/sys/dev/qat/qat_common/ |
H A D | qat_hal.c | 1579 int usize = 0; local 1587 usize = ARRAY_SIZE(inst_4b); 1588 for (i = 0; i < usize; i++)
|