/freebsd-11-stable/crypto/openssl/crypto/md2/ |
H A D | md2_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD2_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-11-stable/crypto/openssl/crypto/md4/ |
H A D | md4_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD4_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-11-stable/crypto/openssl/crypto/md5/ |
H A D | md5_one.c | 82 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD5_Update(&c, temp, chunk); 88 n -= chunk; 89 d += chunk;
|
/freebsd-11-stable/contrib/jemalloc/include/jemalloc/internal/ |
H A D | chunk_mmap.h | 14 bool chunk_dalloc_mmap(void *chunk, size_t size);
|
H A D | arena.h | 65 /* Each element of the chunk map corresponds to one page within the chunk. */ 151 * Each arena_chunk_map_misc_t corresponds to one page within the chunk, just 153 * chunk header in order to improve cache locality. 183 /* Arena chunk header. */ 186 * A pointer to the arena that owns the chunk is stored within the node. 193 * Map of pages within chunk that keeps track of free/large/small. The 194 * first map_bias entries are omitted, since the chunk header does not 196 * for common chunk sizes (e.g. 4 MiB). 338 * In order to avoid rapid chunk allocatio 691 arena_bitselm_get_mutable(arena_chunk_t *chunk, size_t pageind) argument 701 arena_bitselm_get_const(const arena_chunk_t *chunk, size_t pageind) argument 708 arena_miscelm_get_mutable(arena_chunk_t *chunk, size_t pageind) argument [all...] |
H A D | chunk.h | 10 /* Return the chunk address for allocation address a. */ 14 /* Return the chunk offset of address a. */ 18 /* Return the smallest chunk multiple that is >= s. */ 55 bool chunk_register(tsdn_t *tsdn, const void *chunk, 57 void chunk_deregister(const void *chunk, const extent_node_t *node); 66 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed); 68 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed, 71 chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset, 83 extent_node_t *chunk_lookup(const void *chunk, bool dependent);
|
/freebsd-11-stable/contrib/libucl/src/ |
H A D | ucl_parser.c | 45 * @return new position in chunk 47 #define ucl_chunk_skipc(chunk, p) do{ \ 49 (chunk)->line ++; \ 50 (chunk)->column = 0; \ 52 else (chunk)->column ++; \ 54 (chunk)->pos ++; \ 55 (chunk)->remain --; \ 62 struct ucl_chunk *chunk = parser->chunks; local 71 if (chunk->pos < chunk 133 struct ucl_chunk *chunk = parser->chunks; local 938 ucl_lex_number(struct ucl_parser *parser, struct ucl_chunk *chunk, ucl_object_t *obj) argument 971 ucl_lex_json_string(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *need_unescape, bool *ucl_escape, bool *var_expand) argument 1188 ucl_parse_key(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *next_key, bool *end_of_object) argument 1408 ucl_parse_string_value(struct ucl_parser *parser, struct ucl_chunk *chunk, bool *var_expand, bool *need_unescape) argument 1479 ucl_parse_multiline_string(struct ucl_parser *parser, struct ucl_chunk *chunk, const unsigned char *term, int term_len, unsigned char const **beg, bool *var_expand) argument 1564 ucl_parse_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 1760 ucl_parse_after_value(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 1851 ucl_skip_macro_as_comment(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 1963 ucl_parse_macro_value(struct ucl_parser *parser, struct ucl_chunk *chunk, struct ucl_macro *macro, unsigned char const **macro_start, size_t *macro_len) argument 2040 ucl_parse_macro_arguments(struct ucl_parser *parser, struct ucl_chunk *chunk) argument 2157 struct ucl_chunk *chunk = parser->chunks; local 2604 struct ucl_chunk *chunk; local [all...] |
/freebsd-11-stable/sys/dev/mlx4/mlx4_core/ |
H A D | mlx4_icm.c | 48 * per chunk. 55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) argument 59 if (chunk->nsg > 0) 60 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, 63 for (i = 0; i < chunk->npages; ++i) 64 __free_pages(sg_page(&chunk->mem[i]), 65 get_order(chunk->mem[i].length)); 68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) argument 72 for (i = 0; i < chunk 81 struct mlx4_icm_chunk *chunk, *tmp; local 132 struct mlx4_icm_chunk *chunk = NULL; local 312 struct mlx4_icm_chunk *chunk; local [all...] |
H A D | icm.h | 65 struct mlx4_icm_chunk *chunk; member in struct:mlx4_icm_iter 92 iter->chunk = list_empty(&icm->chunk_list) ? 100 return !iter->chunk; 105 if (++iter->page_idx >= iter->chunk->nsg) { 106 if (iter->chunk->list.next == &iter->icm->chunk_list) { 107 iter->chunk = NULL; 111 iter->chunk = list_entry(iter->chunk->list.next, 119 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); 124 return sg_dma_len(&iter->chunk [all...] |
/freebsd-11-stable/contrib/gcclibs/libiberty/ |
H A D | objalloc.c | 56 /* This structure appears at the start of each chunk. */ 60 /* Next chunk. */ 62 /* If this chunk contains large objects, this is the value of 63 current_ptr when this chunk was allocated. If this chunk 74 /* We ask for this much memory each time we create a chunk which is to 89 struct objalloc_chunk *chunk; local 102 chunk = (struct objalloc_chunk *) ret->chunks; 103 chunk->next = NULL; 104 chunk 141 struct objalloc_chunk *chunk; local 157 struct objalloc_chunk *chunk; local [all...] |
H A D | obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 146 register struct _obstack_chunk *chunk; /* points to new chunk */ 173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 174 if (!chunk) 176 h->next_free = h->object_base = chunk->contents; 177 h->chunk_limit = chunk->limit 178 = (char *) chunk + h->chunk_size; 179 chunk 145 register struct _obstack_chunk *chunk; /* points to new chunk */ local 190 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |
/freebsd-11-stable/contrib/binutils/libiberty/ |
H A D | objalloc.c | 56 /* This structure appears at the start of each chunk. */ 60 /* Next chunk. */ 62 /* If this chunk contains large objects, this is the value of 63 current_ptr when this chunk was allocated. If this chunk 74 /* We ask for this much memory each time we create a chunk which is to 89 struct objalloc_chunk *chunk; local 102 chunk = (struct objalloc_chunk *) ret->chunks; 103 chunk->next = NULL; 104 chunk 134 struct objalloc_chunk *chunk; local 150 struct objalloc_chunk *chunk; local [all...] |
H A D | obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 146 register struct _obstack_chunk *chunk; /* points to new chunk */ 173 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 174 if (!chunk) 176 h->next_free = h->object_base = chunk->contents; 177 h->chunk_limit = chunk->limit 178 = (char *) chunk + h->chunk_size; 179 chunk 145 register struct _obstack_chunk *chunk; /* points to new chunk */ local 190 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |
/freebsd-11-stable/contrib/vis/ |
H A D | foldit.c | 45 foldit(const char *chunk, int col, int max, int flags) argument 51 * if this chunk puts us over the limit. 54 cp = chunk;
|
/freebsd-11-stable/contrib/jemalloc/src/ |
H A D | arena.c | 45 arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin); 46 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 54 arena_chunk_t *chunk; local 57 chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(miscelm); 59 mapbits = arena_mapbits_get(chunk, pageind); 214 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument 218 arena_miscelm_get_const(chunk, pageind)))); 219 assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >> 222 arena_miscelm_get_mutable(chunk, pageind)); 226 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_ argument 238 arena_run_dirty_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument 256 arena_run_dirty_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument 326 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 350 arena_run_zero(arena_chunk_t *chunk, size_t run_ind, size_t npages) argument 360 arena_run_page_mark_zeroed(arena_chunk_t *chunk, size_t run_ind) argument 368 arena_run_page_validate_zeroed(arena_chunk_t *chunk, size_t run_ind) argument 406 arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind, size_t flag_dirty, size_t flag_decommitted, size_t need_pages) argument 451 arena_chunk_t *chunk; local 538 arena_chunk_t *chunk; local 575 arena_chunk_t *chunk; local 595 arena_chunk_register(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, bool zero) argument 614 arena_chunk_t *chunk; local 648 arena_chunk_t *chunk; local 677 arena_chunk_t *chunk; local 732 arena_chunk_t *chunk; local 750 arena_chunk_discard(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument 794 arena_chunk_dalloc(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk) argument 978 arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t usize) argument 996 arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1014 arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1040 arena_chunk_ralloc_huge_expand_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument 1069 arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument 1139 arena_chunk_t *chunk; local 1183 arena_chunk_t *chunk; local 1528 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE( local 1565 UNUSED void *chunk; local 1590 arena_chunk_t *chunk = local 1670 arena_chunk_t *chunk = local 1755 arena_chunk_t *chunk = local 1833 arena_achunk_prof_reset(tsd_t *tsd, arena_t *arena, arena_chunk_t *chunk) argument 1977 arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size, size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty, size_t flag_decommitted) argument 2069 arena_run_size_get(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t run_ind) argument 2093 arena_chunk_t *chunk; local 2166 arena_run_trim_head(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument 2207 arena_run_trim_tail(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument 2344 arena_chunk_t *chunk; local 2677 arena_chunk_t *chunk; local 2796 arena_chunk_t *chunk; local 2818 arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 2845 arena_dalloc_bin_run(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 2863 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 2884 arena_dalloc_bin_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm, bool junked) argument 2917 arena_dalloc_bin_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm) argument 2925 arena_dalloc_bin(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_bits_t *bitselm) argument 2941 arena_dalloc_small(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument 2975 arena_dalloc_large_locked_impl(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, bool junked) argument 3003 arena_dalloc_large_junked_locked(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 3011 arena_dalloc_large(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 3022 arena_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument 3059 arena_ralloc_large_grow(tsdn_t *tsdn, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument 3186 arena_chunk_t *chunk; local 3235 arena_chunk_t *chunk; local [all...] |
/freebsd-11-stable/contrib/subversion/subversion/libsvn_subr/ |
H A D | eol.c | 44 apr_uintptr_t chunk = *(const apr_uintptr_t *)buf; 48 apr_uintptr_t r_test = chunk ^ SVN__R_MASK; 49 apr_uintptr_t n_test = chunk ^ SVN__N_MASK; 43 apr_uintptr_t chunk = *(const apr_uintptr_t *)buf; local
|
/freebsd-11-stable/sys/x86/x86/ |
H A D | dump_machdep.c | 45 dumpsys_map_chunk(vm_paddr_t pa, size_t chunk, void **va) argument 50 for (i = 0; i < chunk; i++) {
|
/freebsd-11-stable/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/ |
H A D | zap_leaf.c | 30 * chunk number n means l_chunk[n], even though the header precedes it. 46 #define CHAIN_END 0xffff /* end of the chunk chain */ 190 int chunk = zap_leaf_phys(l)->l_hdr.lh_freelist; 191 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNKS(l)); 192 ASSERT3U(ZAP_LEAF_CHUNK(l, chunk).l_free.lf_type, ==, ZAP_CHUNK_FREE); 195 ZAP_LEAF_CHUNK(l, chunk).l_free.lf_next; 199 return (chunk); 203 zap_leaf_chunk_free(zap_leaf_t *l, uint16_t chunk) argument 205 struct zap_leaf_free *zlf = &ZAP_LEAF_CHUNK(l, chunk).l_free; 207 ASSERT3U(chunk, <, ZAP_LEAF_NUMCHUNK 236 uint16_t chunk = zap_leaf_chunk_alloc(l); local 264 uint16_t chunk = *chunkp; local 279 zap_leaf_array_read(zap_leaf_t *l, uint16_t chunk, int array_int_len, int array_len, int buf_int_len, uint64_t buf_len, void *buf) argument 337 zap_leaf_array_match(zap_leaf_t *l, zap_name_t *zn, int chunk, int array_numints) argument 398 uint16_t chunk = *chunkp; local 560 uint16_t chunk; local 715 zap_leaf_transfer_array(zap_leaf_t *l, uint16_t chunk, zap_leaf_t *nl) argument 748 uint16_t chunk = zap_leaf_chunk_alloc(nl); local 829 int chunk = zap_leaf_phys(l)->l_hash[i]; local [all...] |
/freebsd-11-stable/sys/dev/mthca/ |
H A D | mthca_memfree.h | 76 struct mthca_icm_chunk *chunk; member in struct:mthca_icm_iter 103 iter->chunk = list_empty(&icm->chunk_list) ? 111 return !iter->chunk; 116 if (++iter->page_idx >= iter->chunk->nsg) { 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { 118 iter->chunk = NULL; 122 iter->chunk = list_entry(iter->chunk->list.next, 130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); 135 return sg_dma_len(&iter->chunk [all...] |
H A D | mthca_memfree.c | 47 * per chunk. 63 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) argument 67 if (chunk->nsg > 0) 68 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, 71 for (i = 0; i < chunk->npages; ++i) 72 __free_pages(sg_page(&chunk->mem[i]), 73 get_order(chunk->mem[i].length)); 76 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) argument 80 for (i = 0; i < chunk 89 struct mthca_icm_chunk *chunk, *tmp; local 140 struct mthca_icm_chunk *chunk = NULL; local 280 struct mthca_icm_chunk *chunk; local [all...] |
/freebsd-11-stable/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_poisoning.cpp | 54 u8 *chunk; member in struct:__asan::ShadowSegmentEndpoint 56 s8 value; // = *chunk; 59 chunk = (u8*)MemToShadow(address); 61 value = *chunk; 66 // Since asan's mapping is compacting, the shadow chunk may be 115 if (beg.chunk == end.chunk) { 123 *beg.chunk = Min(value, beg.offset); 125 *beg.chunk = kAsanUserPoisonedMemoryMagic; 130 CHECK_LT(beg.chunk, en [all...] |
/freebsd-11-stable/crypto/openssl/engines/ |
H A D | e_padlock.c | 926 size_t chunk, allocated = 0; local 931 if ((chunk = ctx->num)) { /* borrow chunk variable */ 936 if (chunk >= AES_BLOCK_SIZE) 940 while (chunk < AES_BLOCK_SIZE && nbytes != 0) { 941 ivp[chunk] = *(out_arg++) = *(in_arg++) ^ ivp[chunk]; 942 chunk++, nbytes--; 944 while (chunk < AES_BLOCK_SIZE && nbytes != 0) { 946 *(out_arg++) = c ^ ivp[chunk]; [all...] |
/freebsd-11-stable/contrib/serf/buckets/ |
H A D | chunk_buckets.c | 37 serf_bucket_t *chunk; member in struct:__anon59 51 ctx->chunk = serf_bucket_aggregate_create(allocator); 65 struct iovec vecs[66]; /* 64 + chunk trailer + EOF trailer = 66 */ 90 /* Inserting a 0 byte chunk indicates a terminator, which already happens 92 * EOF chunk to be interpreted by the server as a new request. So, 96 /* Build the chunk header. */ 101 /* Create a copy of the chunk header so we can have multiple chunks 106 serf_bucket_aggregate_append(ctx->chunk, simple_bkt); 108 /* Insert the chunk footer. */ 115 /* Insert the chunk foote [all...] |
/freebsd-11-stable/contrib/gcclibs/libgomp/ |
H A D | iter.c | 115 /* Initial guess is a C sized chunk positioned nthreads iterations 151 long start, end, chunk, left; local 157 chunk = ws->chunk_size * ws->incr; 161 if (chunk < left) 162 chunk = left; 166 if (chunk > left) 167 chunk = left; 169 end = start + chunk; 187 long start, end, nend, chunk, incr; local 192 chunk [all...] |
/freebsd-11-stable/gnu/usr.bin/grep/ |
H A D | obstack.c | 143 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 164 register struct _obstack_chunk *chunk; /* points to new chunk */ 196 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 197 if (!chunk) 199 h->next_free = h->object_base = chunk->contents; 200 h->chunk_limit = chunk->limit 201 = (char *) chunk + h->chunk_size; 202 chunk 163 register struct _obstack_chunk *chunk; /* points to new chunk */ local 222 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |