/barrelfish-2018-10-04/lib/devif/ |
H A D | region_pool.c | 23 // number of regions in pool 33 struct region** pool; member in struct:region_pool 37 * @brief initialized a pool of regions 39 * @param pool Return pointer to the region pool 43 errval_t region_pool_init(struct region_pool** pool) argument 45 // Allocate pool struct itself including pointers to region 46 (*pool) = calloc(1, sizeof(struct region_pool)); 47 if (*pool == NULL) { 48 DQI_DEBUG_REGION("Allocationg inital pool faile 79 region_pool_destroy(struct region_pool* pool) argument 117 region_pool_grow(struct region_pool* pool) argument 160 region_pool_add_region(struct region_pool* pool, struct capref cap, regionid_t* region_id) argument 246 region_pool_add_region_with_id(struct region_pool* pool, struct capref cap, regionid_t region_id) argument 285 region_pool_remove_region(struct region_pool* pool, regionid_t region_id, struct capref* cap) argument 347 region_pool_buffer_check_bounds(struct region_pool* pool, regionid_t region_id, genoffset_t offset, genoffset_t length, genoffset_t valid_data, genoffset_t valid_length) argument [all...] |
H A D | region_pool.h | 19 * @brief initialized a pool of regions 21 * @param pool Return pointer to the region pool 25 errval_t region_pool_init(struct region_pool** pool); 29 * @brief freeing region pool 31 * @param pool The region pool to free 35 errval_t region_pool_destroy(struct region_pool* pool); 38 * @brief add a memory region to the region pool 40 * @param pool Th [all...] |
H A D | queue_interface_internal.c | 36 err = region_pool_init(&(q->pool)); 46 err = region_pool_add_region_with_id(q->pool, cap, rid); 55 err = region_pool_remove_region(q->pool, rid, &cap);
|
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/infiniband/core/ |
H A D | fmr_pool.c | 56 * its pool's free_list (if the FMR can be mapped again; that is, 57 * remap_count < pool->max_remaps) or its pool's dirty_list (if the 94 void (*flush_function)(struct ib_fmr_pool *pool, 113 static inline struct ib_pool_fmr *ib_fmr_cache_lookup(struct ib_fmr_pool *pool, argument 122 if (!pool->cache_bucket) 125 bucket = pool->cache_bucket + ib_fmr_hash(*page_list); 137 static void ib_fmr_batch_release(struct ib_fmr_pool *pool) argument 144 spin_lock_irq(&pool->pool_lock); 146 list_for_each_entry(fmr, &pool 179 struct ib_fmr_pool *pool = pool_ptr; local 214 struct ib_fmr_pool *pool; local 358 ib_destroy_fmr_pool(struct ib_fmr_pool *pool) argument 396 ib_flush_fmr_pool(struct ib_fmr_pool *pool) argument 439 struct ib_fmr_pool *pool = pool_handle; local 514 struct ib_fmr_pool *pool; local [all...] |
/barrelfish-2018-10-04/lib/bulk_transfer/ |
H A D | bulk_pool.h | 16 * internal representation of a bulk pool 20 struct bulk_pool pool; ///< pointer to the public interface member in struct:bulk_pool_internal 22 struct vregion *vregion; ///< pointer to the vregion of the pool 24 uint32_t refcnt; ///< ref count for pool to channel assignments 30 * returns a pointer to the pool with the given id 32 * @param id the id of the pool to look up 34 * @return NULL if the pool is not present in the domain 39 * inserts a pool into the domain global bulk pool list 41 * @param pool th [all...] |
H A D | bulk_pool.c | 44 * returns a pointer to the pool with the given id 46 * @param id the id of the pool to look up 48 * @return NULL if the pool is not present in the domain 54 if (bulk_pool_cmp_id(&list->pool->id, id) == 0) { 55 return list->pool; 63 * inserts a pool into the domain global bulk pool list 65 * @param pool the pool to insert 67 errval_t bulk_pool_domain_list_insert(struct bulk_pool *pool) argument 123 bulk_pool_domain_list_remove(struct bulk_pool *pool) argument 205 bulk_pool_is_assigned(struct bulk_pool *pool, struct bulk_channel *channel) argument 258 bulk_pool_assign(struct bulk_pool *pool, struct bulk_channel *channel) argument 317 bulk_pool_remove(struct bulk_pool *pool, struct bulk_channel *channel) argument 357 bulk_pool_unmap(struct bulk_pool *pool) argument 420 bulk_pool_map(struct bulk_pool *pool) argument 583 bulk_pool_init_bufs(struct bulk_pool *pool) argument 618 bulk_pool_alloc_with_id(struct bulk_pool **pool, size_t buffer_count, size_t buffer_size, struct bulk_pool_id id) argument 659 bulk_pool_alloc(struct bulk_pool **pool, size_t buffer_count, size_t buffer_size) argument 673 bulk_pool_dealloc(struct bulk_pool *pool) argument [all...] |
H A D | bulk_channel.c | 125 * Assign a pool to a channel. 128 * @param pool Pool to assign (must not be assigned to this channel yet) 131 struct bulk_pool *pool, 135 assert(pool); 141 if (pool->trust == BULK_TRUST_UNINITIALIZED) { 142 /* this pool has never been assigned to a channel */ 143 pool->trust = channel->trust; 147 if (channel->trust != pool->trust) { 151 if (bulk_pool_is_assigned(pool, channel)) { 152 debug_printf("bulk_channel_assign_pool: pool alread 130 bulk_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument 176 bulk_channel_remove_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument [all...] |
H A D | bulk_allocator.c | 26 * initializes a new bulk allocator with a pool and allocates memory for it. 31 * @param constraints memory requirements for this pool or NULL if none 52 err = bulk_pool_alloc(&alloc->pool, buffer_count, buffer_size); 57 /* reserve a virtual memory range for the pool */ 69 err = frame_alloc(&alloc->pool->pool_cap, pool_size, &ret_size); 81 alloc->pool->trust = constraints->trust; 89 err = bulk_pool_map(alloc->pool); 97 struct bulk_buffer *buf = alloc->pool->buffers[i]; 115 * @param frame capability for backing the bulk pool 126 * Frees up the bulk allocator and it's pool [all...] |
H A D | bulk_buffer.c | 38 if (buffer->pool->trust == BULK_TRUST_FULL) { 48 pool_int = (struct bulk_pool_internal *) buffer->pool; 53 size_t size = buffer->pool->buffer_size; 57 buffer->pool->base_address + offset); 64 (void *) buffer->pool->base_address); 103 if (buffer->pool->trust == BULK_TRUST_FULL) { 107 pool_int = (struct bulk_pool_internal *) buffer->pool; 112 size_t offset = (lvaddr_t) buffer->address - buffer->pool->base_address; 148 if (buffer->pool->trust == BULK_TRUST_FULL) { 153 ->pool; [all...] |
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/ |
H A D | dmapool.h | 47 struct dma_pool *pool; local 49 pool = kmalloc(sizeof(*pool), GFP_KERNEL); 55 pool->pool_zone = uma_zcreate(name, size, NULL, NULL, NULL, NULL, 58 return (pool); 62 dma_pool_destroy(struct dma_pool *pool) argument 64 uma_zdestroy(pool->pool_zone); 65 kfree(pool); 69 dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, dma_addr_t *handle) argument 73 vaddr = uma_zalloc(pool 80 dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr) argument [all...] |
/barrelfish-2018-10-04/lib/ahci/ |
H A D | ahci_dma_pool.c | 19 // meta-data structure for the dma pool 21 // size of the pool in bytes 60 static struct dma_pool pool; variable in typeref:struct:dma_pool 64 * This function grows the address arrays in the pool metadata 71 if (pool.addr_entries == 0) { // initial alloc 72 pool.virt_addrs = calloc(8, sizeof(void *)); 73 if (pool.virt_addrs == NULL) { 76 pool.phys_addrs = calloc(8, sizeof(genpaddr_t)); 77 if (pool.phys_addrs == NULL) { 80 pool [all...] |
/barrelfish-2018-10-04/include/lwip2/lwip/priv/ |
H A D | tcpip_priv.h | 58 #define API_VAR_ALLOC(type, pool, name, errorval) do { \ 59 name = (type *)memp_malloc(pool); \ 64 #define API_VAR_ALLOC_POOL(type, pool, name, errorval) do { \ 65 name = (type *)LWIP_MEMPOOL_ALLOC(pool); \ 70 #define API_VAR_FREE(pool, name) memp_free(pool, name) 71 #define API_VAR_FREE_POOL(pool, name) LWIP_MEMPOOL_FREE(pool, name) 84 #define API_VAR_ALLOC(type, pool, name, errorval) 85 #define API_VAR_ALLOC_POOL(type, pool, nam [all...] |
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/rdma/ |
H A D | ib_fmr_pool.h | 42 * struct ib_fmr_pool_param - Parameters for creating FMR pool 45 * @access:Access flags for FMRs in pool. 46 * @pool_size:Number of FMRs to allocate for pool. 61 void (*flush_function)(struct ib_fmr_pool *pool, 69 struct ib_fmr_pool *pool; member in struct:ib_pool_fmr 82 void ib_destroy_fmr_pool(struct ib_fmr_pool *pool); 84 int ib_flush_fmr_pool(struct ib_fmr_pool *pool);
|
/barrelfish-2018-10-04/lib/bulk_transfer/backends/local/ |
H A D | control_channel.c | 62 struct bulk_pool *pool; member in struct:local_event::__anon526::__anon529 104 struct bulk_pool *pool, 302 /* -------------------------- event pool assign ---------------------------- */ 305 * Gets called when the pool assignment on the other side is completed 318 err = bulk_pool_assign(lev->params.pool_assigned.pool, 328 lev->cont.handler(lev->params.pool_assigned.pool, result, lev->channel); 335 * Gets called when a pool is assigned to the channel 347 struct bulk_pool *pool = lev->params.pool_assigned.pool; local 351 if (bulk_pool_is_assigned(pool, cha 441 impl_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument 470 struct bulk_pool *pool = bulk_pool_get(&lev->params.move_received.pool_id, local 553 struct bulk_pool *pool = bulk_pool_get(&lev->params.buffer_received.pool_id, local 633 struct bulk_pool *pool = bulk_pool_get(&lev->params.copy_received.pool_id, local 704 struct bulk_pool *pool = bulk_pool_get(&lev->params.copy_released.pool_id, local [all...] |
/barrelfish-2018-10-04/usr/bench/bulk_transfer/ |
H A D | micro_throughput.c | 50 struct bulk_pool *pool) 52 debug_printf("pool_assigned: %p [%d,%d,%d]\n", pool, pool->id.machine, 53 pool->id.dom, pool->id.local); 63 buffer->pool); 71 if (memcmp(&txalloc.pool->id, &buffer->pool->id, sizeof(txalloc.pool->id)) 85 buffer->pool); 49 cb_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument [all...] |
H A D | micro_echo.c | 49 struct bulk_pool *pool) 52 debug_printf("pool_assigned: RX %p [%d,%d,%d]\n", pool, 53 pool->id.machine, pool->id.dom, pool->id.local); 60 expect_success(bulk_channel_assign_pool(&txc, pool, wait_cont)); 62 //but in that case, we don't expect to get a pool over this channel anyway 64 event_dispatch(txc.waitset); //wait until pool is assigned 67 debug_printf("pool_assigned: TX %p [%d,%d,%d]\n", pool, 68 pool 48 cb_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument [all...] |
/barrelfish-2018-10-04/lib/bulk_transfer/backends/sm/ |
H A D | pool.c | 26 struct bulk_pool *pool; ///< pool to be added to channel member in struct:c_assign_pool_data 28 bulk_ctrl_pool_t flounder_pool; ///< flounder representation of pool 41 * Reply handler for pool_assign RPC call. On success, pool is added to channel. 51 // if accepted, add pool to channel 54 errval_t err = bulk_pool_assign(cdata->pool, channel); 92 * confirmation of pool and replies to the original side. 96 bulk_ctrl_pool_t pool, 113 // Create library representation from pool information recevied by flounder. 116 //first look up if we already know this pool fro 94 bulk_sm_assign_pool_rx_call( struct bulk_ctrl_binding *b, bulk_ctrl_pool_t pool, uint64_t id) argument 197 bulk_sm_assign_pool( struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument 222 bulk_sm_remove_pool( struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument [all...] |
H A D | flounder_helpers.c | 154 errval_t create_pool_from_flounder(struct bulk_pool **pool, argument 157 assert(pool); 161 // allocate pool 185 *pool = p; 189 void generate_pool_for_flounder(const struct bulk_pool *pool, argument 192 assert(pool); 195 f_pool->pool_id_machine = pool->id.machine; 196 f_pool->pool_id_dom = pool->id.dom; 197 f_pool->pool_id_local = pool->id.local; 198 f_pool->trust = bulk2flounder_trust(pool [all...] |
/barrelfish-2018-10-04/lib/bulk_transfer/backends/net/ |
H A D | bulk_net_no_copy.c | 47 struct bulk_pool *pool; member in struct:bulk_net_nocopy 168 struct bulk_pool *pool; member in struct:pending_pool_request 182 static inline struct bulk_net_pool_data *get_pool_data(struct bulk_pool *pool) argument 184 return ((struct bulk_pool_internal*) pool)->impl_data; 187 static inline struct bulk_net_nocopy *get_net_nocopy(struct bulk_pool *pool) argument 189 struct bulk_net_pool_data *pd = get_pool_data(pool); 200 struct bulk_pool *pool = bulk_pool_get(pool_id, chan); local 201 assert(pool); 202 assert(buffer_id < pool->num_buffers); 203 return pool 215 get_remote_bufid(struct bulk_pool *pool, uint32_t local_buf_id) argument 223 set_remote_bufid(struct bulk_pool *pool, uint32_t local_buf_id, uint32_t remote_buf_id) argument 438 send_pool_assign_response(struct bulk_net_nocopy *p, errval_t err, struct bulk_pool *pool, uint16_t l_port) argument 510 struct bulk_pool *pool = ppr->pool; local 550 send_pool_assign_request(struct bulk_net_nocopy *p, struct bulk_pool *pool, uint16_t l_port) argument 610 struct bulk_pool *pool = bulk_pool_domain_list_get(&id); local 1357 impl_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument 1466 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local 1484 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local 1524 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local 1538 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local [all...] |
H A D | bulk_net_transparent.c | 43 struct bulk_pool *pool; member in struct:pending_pool_request 179 struct bulk_pool *pool = bulk_pool_get(pool_id, chan); local 180 assert(pool); 181 assert(buffer_id < pool->num_buffers); 182 return pool->buffers[buffer_id]; 353 /* -------------------------- pool assignment -------------------------------*/ 357 struct bulk_pool *pool) 370 t->pool_domain_id = pool->id.dom; 371 t->pool_machine_id = pool->id.machine; 372 t->pool_local_id = pool 355 send_pool_assign_response(struct bulk_net_transp *p, errval_t err, struct bulk_pool *pool) argument 419 struct bulk_pool *pool = ppr->pool; local 439 send_pool_assign_request(struct bulk_net_transp *p, struct bulk_pool *pool) argument 485 struct bulk_pool *pool = bulk_pool_domain_list_get(&id); local 1066 impl_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument [all...] |
/barrelfish-2018-10-04/lib/openssl-1.0.0d/crypto/bn/ |
H A D | bn_ctx.c | 81 /* How many bignums are in each "pool item"; */ 142 BN_POOL pool; member in struct:bignum_ctx 159 BN_POOL_ITEM *item = ctx->pool.head; 206 BN_POOL_reset(&ctx->pool); 223 BN_POOL_init(&ret->pool); 237 BN_POOL_ITEM *pool = ctx->pool.head; local 238 fprintf(stderr,"BN_CTX_free, stack-size=%d, pool-bignums=%d\n", 239 ctx->stack.size, ctx->pool.size); 241 while(pool) { [all...] |
/barrelfish-2018-10-04/usr/tests/bulk_transfer/ |
H A D | bulk_netproxy.c | 35 struct bulk_pool *pool); 51 struct bulk_pool *pool) 67 for (int i = 0; i < buffer->pool->buffer_size; ++i) { 81 for (int i = 0; i < buffer->pool->buffer_size; ++i) { 101 assert(arg == rx_allocator.pool); 128 rx_allocator.pool }; 130 err = bulk_channel_assign_pool(&rx_channel, rx_allocator.pool, cont); 170 struct bulk_pool *pool); 193 assert(tx_allocator.pool == arg); 202 struct bulk_pool *pool) 50 cb_rx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument 201 cb_tx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument [all...] |
H A D | bulk_mini.c | 28 struct bulk_pool *pool); 42 struct bulk_pool *pool) 58 for (int i = 0; i < buffer->pool->buffer_size; ++i) { 72 for (int i = 0; i < buffer->pool->buffer_size; ++i) { 113 struct bulk_pool *pool); 128 assert(tx_allocator.pool == arg); 142 err = bulk_channel_assign_pool(channel, tx_allocator.pool, cont); 147 struct bulk_pool *pool) 206 memset(buffer->address, 123, buffer->pool->buffer_size); 220 memset(buffer->address, 101, buffer->pool 41 cb_rx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument 146 cb_tx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument [all...] |
/barrelfish-2018-10-04/lib/barrelfish/ |
H A D | bulk_transfer.c | 67 current_pbuf->pool = bt; 82 current_pbuf->pool = bt; 113 void *pool; local 114 r = vspace_map_one_frame(&pool, allocated_size, *shared_mem, NULL, NULL); 119 assert(pool != NULL); 121 return bulk_init(pool, size, block_size, bt);
|
/barrelfish-2018-10-04/include/vm/ |
H A D | vm_phys.h | 76 vm_page_t vm_phys_alloc_freelist_pages(int freelist, int pool, int order); 77 vm_page_t vm_phys_alloc_pages(int pool, int order); 89 void vm_phys_set_pool(int pool, vm_page_t m, int order);
|