/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | id_table.c | 44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) argument 49 spin_lock_irqsave(&alloc->lock, flags); 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); 52 if (obj >= alloc->max) 53 obj = find_first_zero_bit(alloc->table, alloc->max); 55 if (obj < alloc->max) { 56 if (alloc 71 c4iw_id_free(struct c4iw_id_table *alloc, u32 obj) argument 82 c4iw_id_table_alloc(struct c4iw_id_table *alloc, u32 start, u32 num, u32 reserved, u32 flags) argument 103 c4iw_id_table_free(struct c4iw_id_table *alloc) argument [all...] |
/linux-master/drivers/android/ |
H A D | binder_alloc.c | 60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, argument 63 if (list_is_last(&buffer->entry, &alloc->buffers)) 64 return alloc->buffer + alloc->buffer_size - buffer->user_data; 68 static void binder_insert_free_buffer(struct binder_alloc *alloc, argument 71 struct rb_node **p = &alloc->free_buffers.rb_node; 79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); 83 alloc->pid, new_buffer_size, new_buffer); 90 buffer_size = binder_alloc_buffer_size(alloc, buffer); 98 rb_insert_color(&new_buffer->rb_node, &alloc 101 binder_insert_allocated_buffer_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer) argument 126 binder_alloc_prepare_to_free_locked( struct binder_alloc *alloc, unsigned long user_ptr) argument 167 binder_alloc_prepare_to_free(struct binder_alloc *alloc, unsigned long user_ptr) argument 193 binder_lru_freelist_add(struct binder_alloc *alloc, unsigned long start, unsigned long end) argument 220 binder_install_single_page(struct binder_alloc *alloc, struct binder_lru_page *lru_page, unsigned long addr) argument 268 binder_install_buffer_pages(struct binder_alloc *alloc, struct binder_buffer *buffer, size_t size) argument 302 binder_lru_freelist_del(struct binder_alloc *alloc, unsigned long start, unsigned long end) argument 332 binder_alloc_set_vma(struct binder_alloc *alloc, struct vm_area_struct *vma) argument 339 binder_alloc_get_vma( struct binder_alloc *alloc) argument 346 debug_no_space_locked(struct binder_alloc *alloc) argument 383 debug_low_async_space_locked(struct binder_alloc *alloc) argument 436 binder_alloc_new_buf_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer, size_t size, int is_async) argument 568 binder_alloc_new_buf(struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async) argument 632 binder_delete_free_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer) argument 659 binder_free_buf_locked(struct binder_alloc *alloc, struct binder_buffer *buffer) argument 731 binder_alloc_get_page(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, pgoff_t *pgoffp) argument 754 binder_alloc_clear_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) argument 781 binder_alloc_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) argument 814 binder_alloc_mmap_handler(struct binder_alloc *alloc, struct vm_area_struct *vma) argument 889 binder_alloc_deferred_release(struct binder_alloc *alloc) argument 961 binder_alloc_print_allocated(struct seq_file *m, struct binder_alloc *alloc) argument 985 binder_alloc_print_pages(struct seq_file *m, struct binder_alloc *alloc) argument 1021 binder_alloc_get_allocated_count(struct binder_alloc *alloc) argument 1042 binder_alloc_vma_close(struct binder_alloc *alloc) argument 1063 struct binder_alloc *alloc = page->alloc; local 1145 binder_alloc_init(struct binder_alloc *alloc) argument 1201 check_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t offset, size_t bytes) argument 1227 binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, const void __user *from, size_t bytes) argument 1258 binder_alloc_do_buffer_copy(struct binder_alloc *alloc, bool to_buffer, struct binder_buffer *buffer, binder_size_t buffer_offset, void *ptr, size_t bytes) argument 1289 binder_alloc_copy_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, void *src, size_t bytes) argument 1299 binder_alloc_copy_from_buffer(struct binder_alloc *alloc, void *dest, struct binder_buffer *buffer, binder_size_t buffer_offset, size_t bytes) argument [all...] |
H A D | binder_alloc.h | 23 * @entry: entry alloc->buffers 65 * @alloc: binder_alloc for a proc 70 struct binder_alloc *alloc; member in struct:binder_lru_page 115 void binder_selftest_alloc(struct binder_alloc *alloc); 117 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} argument 122 struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc, 127 void binder_alloc_init(struct binder_alloc *alloc); 130 void binder_alloc_vma_close(struct binder_alloc *alloc); 132 binder_alloc_prepare_to_free(struct binder_alloc *alloc, 134 void binder_alloc_free_buf(struct binder_alloc *alloc, 152 binder_alloc_get_free_async_space(struct binder_alloc *alloc) argument [all...] |
H A D | binder_alloc_selftest.c | 86 pr_err("alloc sizes: "); 96 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, argument 107 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; 108 if (!alloc->pages[page_index].page_ptr || 109 !list_empty(&alloc->pages[page_index].lru)) { 110 pr_err("expect alloc but is %s at page index %d\n", 111 alloc->pages[page_index].page_ptr ? 119 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, argument 126 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0); 128 !check_buffer_pages_allocated(alloc, buffer 136 binder_selftest_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffers[], size_t *sizes, int *seq, size_t end) argument 160 binder_selftest_free_page(struct binder_alloc *alloc) argument 180 binder_selftest_alloc_free(struct binder_alloc *alloc, size_t *sizes, int *seq, size_t end) argument 209 binder_selftest_free_seq(struct binder_alloc *alloc, size_t *sizes, int *seq, int index, size_t end) argument 227 binder_selftest_alloc_size(struct binder_alloc *alloc, size_t *end_offset) argument 253 binder_selftest_alloc_offset(struct binder_alloc *alloc, size_t *end_offset, int index) argument 287 binder_selftest_alloc(struct binder_alloc *alloc) argument [all...] |
H A D | binder_trace.h | 319 TP_PROTO(struct binder_alloc *alloc, bool allocate, 321 TP_ARGS(alloc, allocate, start, end), 329 __entry->proc = alloc->pid; 331 __entry->offset = start - alloc->buffer; 340 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 341 TP_ARGS(alloc, page_index), 347 __entry->proc = alloc->pid; 355 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 356 TP_ARGS(alloc, page_index)); 359 TP_PROTO(const struct binder_alloc *alloc, size_ [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 40 u32 mthca_alloc(struct mthca_alloc *alloc) argument 45 spin_lock_irqsave(&alloc->lock, flags); 47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); 48 if (obj >= alloc->max) { 49 alloc->top = (alloc->top + alloc->max) & alloc 64 mthca_free(struct mthca_alloc *alloc, u32 obj) argument 79 mthca_alloc_init(struct mthca_alloc *alloc, u32 num, u32 mask, u32 reserved) argument 100 mthca_alloc_cleanup(struct mthca_alloc *alloc) argument [all...] |
H A D | mthca_pd.c | 46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); 56 mthca_free(&dev->pd_table.alloc, pd->pd_num); 66 mthca_free(&dev->pd_table.alloc, pd->pd_num); 71 return mthca_alloc_init(&dev->pd_table.alloc, 80 mthca_alloc_cleanup(&dev->pd_table.alloc);
|
H A D | mthca_uar.c | 40 uar->index = mthca_alloc(&dev->uar_table.alloc); 51 mthca_free(&dev->uar_table.alloc, uar->index); 58 ret = mthca_alloc_init(&dev->uar_table.alloc, 67 mthca_alloc_cleanup(&dev->uar_table.alloc); 77 mthca_alloc_cleanup(&dev->uar_table.alloc);
|
/linux-master/sound/isa/gus/ |
H A D | gus_mem.c | 18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) argument 21 mutex_lock(&alloc->memory_mutex); 23 mutex_unlock(&alloc->memory_mutex); 28 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, argument 43 pblock = alloc->first; 49 if (pblock == alloc->first) 50 alloc->first = nblock; 53 mutex_unlock(&alloc->memory_mutex); 59 if (alloc->last == NULL) { 61 alloc 70 snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) argument 100 snd_gf1_mem_look(struct snd_gf1_mem * alloc, unsigned int address) argument 113 snd_gf1_mem_share(struct snd_gf1_mem * alloc, unsigned int *share_id) argument 128 snd_gf1_mem_find(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block, unsigned int size, int w_16, int align) argument 180 snd_gf1_mem_alloc(struct snd_gf1_mem * alloc, int owner, char *name, int size, int w_16, int align, unsigned int *share_id) argument 213 snd_gf1_mem_free(struct snd_gf1_mem * alloc, unsigned int address) argument 231 struct snd_gf1_mem *alloc; local 260 struct snd_gf1_mem *alloc; local 278 struct snd_gf1_mem *alloc; local [all...] |
/linux-master/tools/lib/subcmd/ |
H A D | subcmd-util.h | 34 * is 'alloc', using the standard growing factor alloc_nr() macro. 36 * DO NOT USE any expression with side-effect for 'x' or 'alloc'. 38 #define ALLOC_GROW(x, nr, alloc) \ 40 if ((nr) > alloc) { \ 41 if (alloc_nr(alloc) < (nr)) \ 42 alloc = (nr); \ 44 alloc = alloc_nr(alloc); \ 45 x = xrealloc((x), alloc * sizeof(*(x))); \
|
/linux-master/tools/perf/util/ |
H A D | strbuf.h | 37 * XXX: do _not_ assume that the area that is yours is of size ->alloc - 1 51 size_t alloc; member in struct:strbuf 65 return sb->alloc ? sb->alloc - sb->len - 1 : 0; 71 if (!sb->alloc) { 76 assert(len < sb->alloc);
|
H A D | strbuf.c | 22 sb->alloc = sb->len = 0; 31 if (sb->alloc) { 39 char *res = sb->alloc ? sb->buf : NULL; 51 if (nr < sb->alloc) 57 if (alloc_nr(sb->alloc) > nr) 58 nr = alloc_nr(sb->alloc); 61 * Note that sb->buf == strbuf_slopbuf if sb->alloc == 0, and it is 64 buf = realloc(sb->alloc ? sb->buf : NULL, nr * sizeof(*buf)); 69 sb->alloc = nr; 106 len = vsnprintf(sb->buf + sb->len, sb->alloc [all...] |
/linux-master/fs/ocfs2/ |
H A D | localalloc.c | 20 #include "alloc.h" 35 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc); 38 struct ocfs2_dinode *alloc, 42 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc); 46 struct ocfs2_dinode *alloc, 64 * the local alloc. 66 * Generally, we'd like to pick as large a local alloc as 72 * Some things work against us when trying to choose a large local alloc: 87 * alloc maximums at various cluster sizes (4k blocksize) 142 * local alloc siz 271 struct ocfs2_dinode *alloc = NULL; local 375 struct ocfs2_dinode *alloc = NULL; local 483 struct ocfs2_dinode *alloc; local 546 ocfs2_complete_local_alloc_recovery(struct ocfs2_super *osb, struct ocfs2_dinode *alloc) argument 619 struct ocfs2_dinode *alloc; local 723 struct ocfs2_dinode *alloc; local 779 struct ocfs2_dinode *alloc; local 811 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc) argument 822 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) argument 910 ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc) argument 947 ocfs2_sync_local_to_main(struct ocfs2_super *osb, handle_t *handle, struct ocfs2_dinode *alloc, struct inode *main_bm_inode, struct buffer_head *main_bm_bh) argument 1152 struct ocfs2_dinode *alloc = NULL; local 1241 struct ocfs2_dinode *alloc; local [all...] |
/linux-master/arch/s390/kernel/ |
H A D | abs_lowcore.c | 8 int abs_lowcore_map(int cpu, struct lowcore *lc, bool alloc) argument 15 rc = __vmem_map_4k_page(addr, phys, PAGE_KERNEL, alloc); 23 if (alloc) {
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_alloc_btree.c | 144 len = rec->alloc.ar_blockcount; 147 if (be32_to_cpu(rec->alloc.ar_blockcount) <= 150 len = rec->alloc.ar_blockcount; 200 key->alloc.ar_startblock = rec->alloc.ar_startblock; 201 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; 211 x = be32_to_cpu(rec->alloc.ar_startblock); 212 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; 213 key->alloc [all...] |
/linux-master/fs/hpfs/ |
H A D | Makefile | 8 hpfs-objs := alloc.o anode.o buffer.o dentry.o dir.o dnode.o ea.o file.o \
|
/linux-master/arch/s390/include/asm/ |
H A D | abs_lowcore.h | 11 int abs_lowcore_map(int cpu, struct lowcore *lc, bool alloc);
|
/linux-master/lib/zstd/compress/ |
H A D | zstd_cwksp.h | 201 * Returns an adjusted alloc size that is the nearest larger multiple of 64 bytes. 247 void* const alloc = (BYTE*)ws->allocStart - bytes; local 250 alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes); 252 assert(alloc >= bottom); 253 if (alloc < bottom) { 254 DEBUGLOG(4, "cwksp: alloc failed!"); 260 if (alloc < ws->tableValidEnd) { 261 ws->tableValidEnd = alloc; 263 ws->allocStart = alloc; 264 return alloc; 295 void* const alloc = ws->objectEnd; local 326 void* alloc; local 365 void* alloc; local 401 void* alloc = ws->objectEnd; local [all...] |
/linux-master/arch/xtensa/variants/csp/include/variant/ |
H A D | tie-asm.h | 72 * alloc Select what category(ies) of registers to allocate; if any 76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 138 * alloc Select what category(ies) of registers to allocate; if any 142 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 150 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 162 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 182 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) [all...] |
/linux-master/arch/xtensa/variants/dc233c/include/variant/ |
H A D | tie-asm.h | 73 * alloc Select what category(ies) of registers to allocate; if any 77 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 85 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 97 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 115 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 137 * alloc Select what category(ies) of registers to allocate; if any 141 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 149 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 161 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 179 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) [all...] |
/linux-master/sound/core/ |
H A D | memalloc_local.h | 6 void *(*alloc)(struct snd_dma_buffer *dmab, size_t size); member in struct:snd_malloc_ops
|
/linux-master/drivers/md/bcache/ |
H A D | Makefile | 5 bcache-y := alloc.o bset.o btree.o debug.o extents.o io.o\
|
/linux-master/drivers/iommu/ |
H A D | io-pgtable.c | 41 if (!cfg->alloc && !cfg->free) 44 /* When passing a custom allocator, both the alloc and free 47 if (!cfg->alloc || !cfg->free) 74 iop = fns->alloc(cfg, cookie);
|
/linux-master/arch/s390/mm/ |
H A D | pgalloc.c | 294 unsigned long end, int alloc) 298 if (!alloc) 310 unsigned long end, int alloc) 320 if (!alloc) 328 rc = base_page_walk(table, addr, next, alloc); 331 if (!alloc) 339 unsigned long end, int alloc) 349 if (!alloc) 357 rc = base_segment_walk(table, addr, next, alloc); 360 if (!alloc) 293 base_page_walk(unsigned long *origin, unsigned long addr, unsigned long end, int alloc) argument 309 base_segment_walk(unsigned long *origin, unsigned long addr, unsigned long end, int alloc) argument 338 base_region3_walk(unsigned long *origin, unsigned long addr, unsigned long end, int alloc) argument 366 base_region2_walk(unsigned long *origin, unsigned long addr, unsigned long end, int alloc) argument 394 base_region1_walk(unsigned long *origin, unsigned long addr, unsigned long end, int alloc) argument [all...] |
/linux-master/arch/xtensa/variants/de212/include/variant/ |
H A D | tie-asm.h | 72 * alloc Select what category(ies) of registers to allocate; if any 76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 86 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 104 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 126 * alloc Select what category(ies) of registers to allocate; if any 130 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 140 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 158 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
|