Searched refs:chunk (Results 1 - 25 of 160) sorted by path

1234567

/freebsd-11-stable/contrib/binutils/include/
H A Dobstack.h68 The way we do this is to take a large chunk, allocating memory from
69 low addresses. When you want to build a symbol in the chunk you just
70 add chars above the current "high water mark" in the chunk. When you
73 Mostly the chars will not burst over the highest address of the chunk,
74 because you would typically expect a chunk to be (say) 100 times as
84 When the chars burst over a chunk boundary, we allocate a larger
85 chunk, and then copy the partly formed object from the end of the old
86 chunk to the beginning of the new larger chunk. We then carry on
95 We carve out one object at a time from the current chunk
167 struct _obstack_chunk *chunk; /* address of current struct obstack_chunk */ member in struct:obstack
[all...]
/freebsd-11-stable/contrib/binutils/libiberty/
H A Dobjalloc.c56 /* This structure appears at the start of each chunk. */
60 /* Next chunk. */
62 /* If this chunk contains large objects, this is the value of
63 current_ptr when this chunk was allocated. If this chunk
74 /* We ask for this much memory each time we create a chunk which is to
89 struct objalloc_chunk *chunk; local
102 chunk = (struct objalloc_chunk *) ret->chunks;
103 chunk->next = NULL;
104 chunk
134 struct objalloc_chunk *chunk; local
150 struct objalloc_chunk *chunk; local
[all...]
H A Dobstack.c133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default).
146 register struct _obstack_chunk *chunk; /* points to new chunk */
173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size);
174 if (!chunk)
176 h->next_free = h->object_base = chunk->contents;
177 h->chunk_limit = chunk->limit
178 = (char *) chunk + h->chunk_size;
179 chunk
145 register struct _obstack_chunk *chunk; /* points to new chunk */ local
190 register struct _obstack_chunk *chunk; /* points to new chunk */ local
[all...]
/freebsd-11-stable/contrib/gcc/cp/
H A Dmangle.c1237 unsigned HOST_WIDE_INT chunk; local
1246 chunk = 1000000000;
1253 chunk *= chunk;
1257 base = build_int_cstu (type, chunk);
/freebsd-11-stable/contrib/gcc/
H A Ddbxout.c856 char *chunk;
864 chunk = str = XOBFINISH (&stabstr_ob, char *);
871 chunklen = strlen (chunk);
872 fwrite (chunk, 1, chunklen, asm_out_file);
876 chunk += chunklen + 1;
853 char *chunk; local
H A Dggc-zone.c110 a bitmap which marks the start position of each chunk (whether
112 lists and contain a pointer to the next free chunk. Chunks in
135 initialized if the chunk is in the "other" sized free list. Large
383 /* The most recent free chunk is saved here, instead of in the linked
974 struct alloc_chunk *chunk = (struct alloc_chunk *) ptr; local
982 VALGRIND_DISCARD (VALGRIND_MAKE_WRITABLE (chunk, sizeof (struct alloc_chunk)));
983 chunk->size = size;
984 chunk->next_free = zone->free_chunks[bin];
990 VALGRIND_DISCARD (VALGRIND_MAKE_WRITABLE (chunk, sizeof (struct alloc_chunk *)));
991 chunk
1012 struct alloc_chunk *chunk, **pp; local
2023 struct alloc_chunk *chunk = zone->free_chunks[i]; local
[all...]
H A Dpretty-print.c215 unsigned int curarg = 0, chunk = 0, argno; local
220 /* Allocate a new chunk structure. */
277 /* Handled in phase 2. Terminate the plain chunk here. */
279 gcc_assert (chunk < PP_NL_ARGMAX * 2);
280 args[chunk++] = XOBFINISH (&buffer->chunk_obstack, const char *);
303 formatters[argno] = &args[chunk];
358 gcc_assert (chunk < PP_NL_ARGMAX * 2);
359 args[chunk++] = XOBFINISH (&buffer->chunk_obstack, const char *);
363 gcc_assert (chunk < PP_NL_ARGMAX * 2);
364 args[chunk
557 unsigned int chunk; local
[all...]
/freebsd-11-stable/contrib/gcclibs/include/
H A Dobstack.h68 The way we do this is to take a large chunk, allocating memory from
69 low addresses. When you want to build a symbol in the chunk you just
70 add chars above the current "high water mark" in the chunk. When you
73 Mostly the chars will not burst over the highest address of the chunk,
74 because you would typically expect a chunk to be (say) 100 times as
84 When the chars burst over a chunk boundary, we allocate a larger
85 chunk, and then copy the partly formed object from the end of the old
86 chunk to the beginning of the new larger chunk. We then carry on
95 We carve out one object at a time from the current chunk
167 struct _obstack_chunk *chunk; /* address of current struct obstack_chunk */ member in struct:obstack
[all...]
/freebsd-11-stable/contrib/gcclibs/libgomp/
H A Diter.c115 /* Initial guess is a C sized chunk positioned nthreads iterations
151 long start, end, chunk, left; local
157 chunk = ws->chunk_size * ws->incr;
161 if (chunk < left)
162 chunk = left;
166 if (chunk > left)
167 chunk = left;
169 end = start + chunk;
187 long start, end, nend, chunk, incr; local
192 chunk
[all...]
/freebsd-11-stable/contrib/gcclibs/libiberty/
H A Dobjalloc.c56 /* This structure appears at the start of each chunk. */
60 /* Next chunk. */
62 /* If this chunk contains large objects, this is the value of
63 current_ptr when this chunk was allocated. If this chunk
74 /* We ask for this much memory each time we create a chunk which is to
89 struct objalloc_chunk *chunk; local
102 chunk = (struct objalloc_chunk *) ret->chunks;
103 chunk->next = NULL;
104 chunk
141 struct objalloc_chunk *chunk; local
157 struct objalloc_chunk *chunk; local
[all...]
H A Dobstack.c133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default).
146 register struct _obstack_chunk *chunk; /* points to new chunk */
173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size);
174 if (!chunk)
176 h->next_free = h->object_base = chunk->contents;
177 h->chunk_limit = chunk->limit
178 = (char *) chunk + h->chunk_size;
179 chunk
145 register struct _obstack_chunk *chunk; /* points to new chunk */ local
190 register struct _obstack_chunk *chunk; /* points to new chunk */ local
[all...]
/freebsd-11-stable/contrib/gdb/gdb/
H A Ddve3900-rom.c756 int i, chunk; local
763 for (i = chunk = 0, seq = 0; i < buflen; i += chunk, seq++)
766 chunk = buflen - i;
767 if (chunk > MAXPSIZE)
768 chunk = MAXPSIZE;
771 write_long_le (len, chunk);
775 debug_write (&buf[i], chunk);
H A Dvalues.c53 by print commands during this session. Each chunk
54 records 60 consecutive values. The first chunk on
296 struct value_history_chunk *chunk;
319 chunk = value_history_chain;
322 chunk = chunk->next;
324 return value_copy (chunk->values[absnum % VALUE_HISTORY_CHUNK]);
294 struct value_history_chunk *chunk; local
/freebsd-11-stable/contrib/gdb/include/
H A Dobstack.h68 The way we do this is to take a large chunk, allocating memory from
69 low addresses. When you want to build a symbol in the chunk you just
70 add chars above the current "high water mark" in the chunk. When you
73 Mostly the chars will not burst over the highest address of the chunk,
74 because you would typically expect a chunk to be (say) 100 times as
84 When the chars burst over a chunk boundary, we allocate a larger
85 chunk, and then copy the partly formed object from the end of the old
86 chunk to the beginning of the new larger chunk. We then carry on
95 We carve out one object at a time from the current chunk
171 struct _obstack_chunk *chunk; /* address of current struct obstack_chunk */ member in struct:obstack
[all...]
/freebsd-11-stable/contrib/jemalloc/include/jemalloc/internal/
H A Darena.h65 /* Each element of the chunk map corresponds to one page within the chunk. */
151 * Each arena_chunk_map_misc_t corresponds to one page within the chunk, just
153 * chunk header in order to improve cache locality.
183 /* Arena chunk header. */
186 * A pointer to the arena that owns the chunk is stored within the node.
193 * Map of pages within chunk that keeps track of free/large/small. The
194 * first map_bias entries are omitted, since the chunk header does not
196 * for common chunk sizes (e.g. 4 MiB).
338 * In order to avoid rapid chunk allocatio
691 arena_bitselm_get_mutable(arena_chunk_t *chunk, size_t pageind) argument
701 arena_bitselm_get_const(const arena_chunk_t *chunk, size_t pageind) argument
708 arena_miscelm_get_mutable(arena_chunk_t *chunk, size_t pageind) argument
[all...]
H A Dchunk.h10 /* Return the chunk address for allocation address a. */
14 /* Return the chunk offset of address a. */
18 /* Return the smallest chunk multiple that is >= s. */
55 bool chunk_register(tsdn_t *tsdn, const void *chunk,
57 void chunk_deregister(const void *chunk, const extent_node_t *node);
66 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed);
68 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed,
71 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset,
83 extent_node_t *chunk_lookup(const void *chunk, bool dependent);
H A Dchunk_dss.h28 bool chunk_in_dss(tsdn_t *tsdn, void *chunk);
H A Dchunk_mmap.h14 bool chunk_dalloc_mmap(void *chunk, size_t size);
H A Dtcache.h373 arena_chunk_t *chunk =
375 size_t pageind = (((uintptr_t)ret - (uintptr_t)chunk) >>
377 arena_mapbits_large_binind_set(chunk, pageind,
/freebsd-11-stable/contrib/jemalloc/src/
H A Darena.c45 arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin);
46 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk,
54 arena_chunk_t *chunk; local
57 chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(miscelm);
59 mapbits = arena_mapbits_get(chunk, pageind);
214 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument
218 arena_miscelm_get_const(chunk, pageind))));
219 assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >>
222 arena_miscelm_get_mutable(chunk, pageind));
226 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_ argument
238 arena_run_dirty_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
256 arena_run_dirty_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
326 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local
350 arena_run_zero(arena_chunk_t *chunk, size_t run_ind, size_t npages) argument
360 arena_run_page_mark_zeroed(arena_chunk_t *chunk, size_t run_ind) argument
368 arena_run_page_validate_zeroed(arena_chunk_t *chunk, size_t run_ind) argument
406 arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind, size_t flag_dirty, size_t flag_decommitted, size_t need_pages) argument
451 arena_chunk_t *chunk; local
538 arena_chunk_t *chunk; local
575 arena_chunk_t *chunk; local
595 arena_chunk_register(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, bool zero) argument
614 arena_chunk_t *chunk; local
648 arena_chunk_t *chunk; local
677 arena_chunk_t *chunk; local
732 arena_chunk_t *chunk; local
750 arena_chunk_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument
794 arena_chunk_dalloc(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument
978 arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t usize) argument
996 arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
1014 arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
1040 arena_chunk_ralloc_huge_expand_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument
1069 arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument
1139 arena_chunk_t *chunk; local
1183 arena_chunk_t *chunk; local
1528 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE( local
1565 UNUSED void *chunk; local
1590 arena_chunk_t *chunk = local
1670 arena_chunk_t *chunk = local
1755 arena_chunk_t *chunk = local
1833 arena_achunk_prof_reset(tsd_t *tsd, arena_t *arena, arena_chunk_t *chunk) argument
1977 arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size, size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty, size_t flag_decommitted) argument
2069 arena_run_size_get(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t run_ind) argument
2093 arena_chunk_t *chunk; local
2166 arena_run_trim_head(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument
2207 arena_run_trim_tail(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument
2344 arena_chunk_t *chunk; local
2677 arena_chunk_t *chunk; local
2796 arena_chunk_t *chunk; local
2818 arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2845 arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2863 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2884 arena_dalloc_bin_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm, bool junked) argument
2917 arena_dalloc_bin_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm) argument
2925 arena_dalloc_bin(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_bits_t *bitselm) argument
2941 arena_dalloc_small(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument
2975 arena_dalloc_large_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, bool junked) argument
3003 arena_dalloc_large_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
3011 arena_dalloc_large(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
3022 arena_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument
3059 arena_ralloc_large_grow(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument
3186 arena_chunk_t *chunk; local
3235 arena_chunk_t *chunk; local
[all...]
H A Dchunk.c16 /* Various chunk-related settings. */
23 static bool chunk_dalloc_default(void *chunk, size_t size, bool committed,
25 static bool chunk_commit_default(void *chunk, size_t size, size_t offset,
27 static bool chunk_decommit_default(void *chunk, size_t size, size_t offset,
29 static bool chunk_purge_default(void *chunk, size_t size, size_t offset,
31 static bool chunk_split_default(void *chunk, size_t size, size_t size_a,
54 extent_tree_t *chunks_ad, bool cache, void *chunk, size_t size, bool zeroed,
144 chunk_register(tsdn_t *tsdn, const void *chunk, const extent_node_t *node) argument
147 assert(extent_node_addr_get(node) == chunk);
149 if (rtree_set(&chunks_rtree, (uintptr_t)chunk, nod
171 chunk_deregister(const void *chunk, const extent_node_t *node) argument
506 chunk_record(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *chunk, size_t size, bool zeroed, bool committed) argument
598 chunk_dalloc_cache(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed) argument
613 chunk_dalloc_default_impl(tsdn_t *tsdn, void *chunk, size_t size) argument
622 chunk_dalloc_default(void *chunk, size_t size, bool committed, unsigned arena_ind) argument
633 chunk_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed, bool committed) argument
668 chunk_commit_default(void *chunk, size_t size, size_t offset, size_t length, unsigned arena_ind) argument
677 chunk_decommit_default(void *chunk, size_t size, size_t offset, size_t length, unsigned arena_ind) argument
686 chunk_purge_default(void *chunk, size_t size, size_t offset, size_t length, unsigned arena_ind) argument
701 chunk_purge_wrapper(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset, size_t length) argument
710 chunk_split_default(void *chunk, size_t size, size_t size_a, size_t size_b, bool committed, unsigned arena_ind) argument
[all...]
H A Dchunk_dss.c108 * chunk-align the end of the DSS.
113 * Compute how much chunk-aligned pad space (if any) is
158 chunk_in_dss(tsdn_t *tsdn, void *chunk) argument
165 if ((uintptr_t)chunk >= (uintptr_t)dss_base
166 && (uintptr_t)chunk < (uintptr_t)dss_max)
H A Dchunk_mmap.c71 chunk_dalloc_mmap(void *chunk, size_t size) argument
75 pages_unmap(chunk, size);
H A Dtcache.c104 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE( local
106 arena_t *bin_arena = extent_node_arena_get(&chunk->node);
128 chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(ptr);
129 if (extent_node_arena_get(&chunk->node) == bin_arena) {
131 (uintptr_t)chunk) >> LG_PAGE;
133 arena_bitselm_get_mutable(chunk, pageind);
135 bin_arena, chunk, ptr, bitselm); local
186 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE( local
188 arena_t *locked_arena = extent_node_arena_get(&chunk->node);
213 chunk
217 locked_arena, chunk, ptr); local
[all...]
/freebsd-11-stable/contrib/libucl/src/
H A Ducl_parser.c45 * @return new position in chunk
47 #define ucl_chunk_skipc(chunk, p) do{ \
49 (chunk)->line ++; \
50 (chunk)->column = 0; \
52 else (chunk)->column ++; \
54 (chunk)->pos ++; \
55 (chunk)->remain --; \
62 struct ucl_chunk *chunk = parser->chunks; local
71 if (chunk->pos < chunk
133 struct ucl_chunk *chunk = parser->chunks; local
938 ucl_lex_number(struct ucl_parser *parser, struct ucl_chunk *chunk, ucl_object_t *obj) argument
971 ucl_lex_json_string(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *need_unescape, bool *ucl_escape, bool *var_expand) argument
1188 ucl_parse_key(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *next_key, bool *end_of_object) argument
1408 ucl_parse_string_value(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *var_expand, bool *need_unescape) argument
1479 ucl_parse_multiline_string(struct ucl_parser *parser, struct ucl_chunk *chunk, const unsigned char *term, int term_len, unsigned char const **beg, bool *var_expand) argument
1564 ucl_parse_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument
1760 ucl_parse_after_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument
1851 ucl_skip_macro_as_comment(struct ucl_parser *parser, struct ucl_chunk *chunk) argument
1963 ucl_parse_macro_value(struct ucl_parser *parser, struct ucl_chunk *chunk, struct ucl_macro *macro, unsigned char const **macro_start, size_t *macro_len) argument
2040 ucl_parse_macro_arguments(struct ucl_parser *parser, struct ucl_chunk *chunk) argument
2157 struct ucl_chunk *chunk = parser->chunks; local
2604 struct ucl_chunk *chunk; local
[all...]

Completed in 231 milliseconds

1234567