Searched refs:pool (Results 1 - 25 of 65) sorted by relevance

123

/barrelfish-2018-10-04/lib/devif/
H A Dregion_pool.c23 // number of regions in pool
33 struct region** pool; member in struct:region_pool
37 * @brief initialized a pool of regions
39 * @param pool Return pointer to the region pool
43 errval_t region_pool_init(struct region_pool** pool) argument
45 // Allocate pool struct itself including pointers to region
46 (*pool) = calloc(1, sizeof(struct region_pool));
47 if (*pool == NULL) {
48 DQI_DEBUG_REGION("Allocationg inital pool faile
79 region_pool_destroy(struct region_pool* pool) argument
117 region_pool_grow(struct region_pool* pool) argument
160 region_pool_add_region(struct region_pool* pool, struct capref cap, regionid_t* region_id) argument
246 region_pool_add_region_with_id(struct region_pool* pool, struct capref cap, regionid_t region_id) argument
285 region_pool_remove_region(struct region_pool* pool, regionid_t region_id, struct capref* cap) argument
347 region_pool_buffer_check_bounds(struct region_pool* pool, regionid_t region_id, genoffset_t offset, genoffset_t length, genoffset_t valid_data, genoffset_t valid_length) argument
[all...]
H A Dregion_pool.h19 * @brief initialized a pool of regions
21 * @param pool Return pointer to the region pool
25 errval_t region_pool_init(struct region_pool** pool);
29 * @brief freeing region pool
31 * @param pool The region pool to free
35 errval_t region_pool_destroy(struct region_pool* pool);
38 * @brief add a memory region to the region pool
40 * @param pool Th
[all...]
H A Dqueue_interface_internal.c36 err = region_pool_init(&(q->pool));
46 err = region_pool_add_region_with_id(q->pool, cap, rid);
55 err = region_pool_remove_region(q->pool, rid, &cap);
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/infiniband/core/
H A Dfmr_pool.c56 * its pool's free_list (if the FMR can be mapped again; that is,
57 * remap_count < pool->max_remaps) or its pool's dirty_list (if the
94 void (*flush_function)(struct ib_fmr_pool *pool,
113 static inline struct ib_pool_fmr *ib_fmr_cache_lookup(struct ib_fmr_pool *pool, argument
122 if (!pool->cache_bucket)
125 bucket = pool->cache_bucket + ib_fmr_hash(*page_list);
137 static void ib_fmr_batch_release(struct ib_fmr_pool *pool) argument
144 spin_lock_irq(&pool->pool_lock);
146 list_for_each_entry(fmr, &pool
179 struct ib_fmr_pool *pool = pool_ptr; local
214 struct ib_fmr_pool *pool; local
358 ib_destroy_fmr_pool(struct ib_fmr_pool *pool) argument
396 ib_flush_fmr_pool(struct ib_fmr_pool *pool) argument
439 struct ib_fmr_pool *pool = pool_handle; local
514 struct ib_fmr_pool *pool; local
[all...]
/barrelfish-2018-10-04/lib/bulk_transfer/
H A Dbulk_pool.h16 * internal representation of a bulk pool
20 struct bulk_pool pool; ///< pointer to the public interface member in struct:bulk_pool_internal
22 struct vregion *vregion; ///< pointer to the vregion of the pool
24 uint32_t refcnt; ///< ref count for pool to channel assignments
30 * returns a pointer to the pool with the given id
32 * @param id the id of the pool to look up
34 * @return NULL if the pool is not present in the domain
39 * inserts a pool into the domain global bulk pool list
41 * @param pool th
[all...]
H A Dbulk_pool.c44 * returns a pointer to the pool with the given id
46 * @param id the id of the pool to look up
48 * @return NULL if the pool is not present in the domain
54 if (bulk_pool_cmp_id(&list->pool->id, id) == 0) {
55 return list->pool;
63 * inserts a pool into the domain global bulk pool list
65 * @param pool the pool to insert
67 errval_t bulk_pool_domain_list_insert(struct bulk_pool *pool) argument
123 bulk_pool_domain_list_remove(struct bulk_pool *pool) argument
205 bulk_pool_is_assigned(struct bulk_pool *pool, struct bulk_channel *channel) argument
258 bulk_pool_assign(struct bulk_pool *pool, struct bulk_channel *channel) argument
317 bulk_pool_remove(struct bulk_pool *pool, struct bulk_channel *channel) argument
357 bulk_pool_unmap(struct bulk_pool *pool) argument
420 bulk_pool_map(struct bulk_pool *pool) argument
583 bulk_pool_init_bufs(struct bulk_pool *pool) argument
618 bulk_pool_alloc_with_id(struct bulk_pool **pool, size_t buffer_count, size_t buffer_size, struct bulk_pool_id id) argument
659 bulk_pool_alloc(struct bulk_pool **pool, size_t buffer_count, size_t buffer_size) argument
673 bulk_pool_dealloc(struct bulk_pool *pool) argument
[all...]
H A Dbulk_channel.c125 * Assign a pool to a channel.
128 * @param pool Pool to assign (must not be assigned to this channel yet)
131 struct bulk_pool *pool,
135 assert(pool);
141 if (pool->trust == BULK_TRUST_UNINITIALIZED) {
142 /* this pool has never been assigned to a channel */
143 pool->trust = channel->trust;
147 if (channel->trust != pool->trust) {
151 if (bulk_pool_is_assigned(pool, channel)) {
152 debug_printf("bulk_channel_assign_pool: pool alread
130 bulk_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
176 bulk_channel_remove_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
[all...]
H A Dbulk_allocator.c26 * initializes a new bulk allocator with a pool and allocates memory for it.
31 * @param constraints memory requirements for this pool or NULL if none
52 err = bulk_pool_alloc(&alloc->pool, buffer_count, buffer_size);
57 /* reserve a virtual memory range for the pool */
69 err = frame_alloc(&alloc->pool->pool_cap, pool_size, &ret_size);
81 alloc->pool->trust = constraints->trust;
89 err = bulk_pool_map(alloc->pool);
97 struct bulk_buffer *buf = alloc->pool->buffers[i];
115 * @param frame capability for backing the bulk pool
126 * Frees up the bulk allocator and it's pool
[all...]
H A Dbulk_buffer.c38 if (buffer->pool->trust == BULK_TRUST_FULL) {
48 pool_int = (struct bulk_pool_internal *) buffer->pool;
53 size_t size = buffer->pool->buffer_size;
57 buffer->pool->base_address + offset);
64 (void *) buffer->pool->base_address);
103 if (buffer->pool->trust == BULK_TRUST_FULL) {
107 pool_int = (struct bulk_pool_internal *) buffer->pool;
112 size_t offset = (lvaddr_t) buffer->address - buffer->pool->base_address;
148 if (buffer->pool->trust == BULK_TRUST_FULL) {
153 ->pool;
[all...]
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/
H A Ddmapool.h47 struct dma_pool *pool; local
49 pool = kmalloc(sizeof(*pool), GFP_KERNEL);
55 pool->pool_zone = uma_zcreate(name, size, NULL, NULL, NULL, NULL,
58 return (pool);
62 dma_pool_destroy(struct dma_pool *pool) argument
64 uma_zdestroy(pool->pool_zone);
65 kfree(pool);
69 dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, dma_addr_t *handle) argument
73 vaddr = uma_zalloc(pool
80 dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t addr) argument
[all...]
/barrelfish-2018-10-04/lib/ahci/
H A Dahci_dma_pool.c19 // meta-data structure for the dma pool
21 // size of the pool in bytes
60 static struct dma_pool pool; variable in typeref:struct:dma_pool
64 * This function grows the address arrays in the pool metadata
71 if (pool.addr_entries == 0) { // initial alloc
72 pool.virt_addrs = calloc(8, sizeof(void *));
73 if (pool.virt_addrs == NULL) {
76 pool.phys_addrs = calloc(8, sizeof(genpaddr_t));
77 if (pool.phys_addrs == NULL) {
80 pool
[all...]
/barrelfish-2018-10-04/include/lwip2/lwip/priv/
H A Dtcpip_priv.h58 #define API_VAR_ALLOC(type, pool, name, errorval) do { \
59 name = (type *)memp_malloc(pool); \
64 #define API_VAR_ALLOC_POOL(type, pool, name, errorval) do { \
65 name = (type *)LWIP_MEMPOOL_ALLOC(pool); \
70 #define API_VAR_FREE(pool, name) memp_free(pool, name)
71 #define API_VAR_FREE_POOL(pool, name) LWIP_MEMPOOL_FREE(pool, name)
84 #define API_VAR_ALLOC(type, pool, name, errorval)
85 #define API_VAR_ALLOC_POOL(type, pool, nam
[all...]
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/rdma/
H A Dib_fmr_pool.h42 * struct ib_fmr_pool_param - Parameters for creating FMR pool
45 * @access:Access flags for FMRs in pool.
46 * @pool_size:Number of FMRs to allocate for pool.
61 void (*flush_function)(struct ib_fmr_pool *pool,
69 struct ib_fmr_pool *pool; member in struct:ib_pool_fmr
82 void ib_destroy_fmr_pool(struct ib_fmr_pool *pool);
84 int ib_flush_fmr_pool(struct ib_fmr_pool *pool);
/barrelfish-2018-10-04/lib/bulk_transfer/backends/local/
H A Dcontrol_channel.c62 struct bulk_pool *pool; member in struct:local_event::__anon526::__anon529
104 struct bulk_pool *pool,
302 /* -------------------------- event pool assign ---------------------------- */
305 * Gets called when the pool assignment on the other side is completed
318 err = bulk_pool_assign(lev->params.pool_assigned.pool,
328 lev->cont.handler(lev->params.pool_assigned.pool, result, lev->channel);
335 * Gets called when a pool is assigned to the channel
347 struct bulk_pool *pool = lev->params.pool_assigned.pool; local
351 if (bulk_pool_is_assigned(pool, cha
441 impl_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
470 struct bulk_pool *pool = bulk_pool_get(&lev->params.move_received.pool_id, local
553 struct bulk_pool *pool = bulk_pool_get(&lev->params.buffer_received.pool_id, local
633 struct bulk_pool *pool = bulk_pool_get(&lev->params.copy_received.pool_id, local
704 struct bulk_pool *pool = bulk_pool_get(&lev->params.copy_released.pool_id, local
[all...]
/barrelfish-2018-10-04/usr/bench/bulk_transfer/
H A Dmicro_throughput.c50 struct bulk_pool *pool)
52 debug_printf("pool_assigned: %p [%d,%d,%d]\n", pool, pool->id.machine,
53 pool->id.dom, pool->id.local);
63 buffer->pool);
71 if (memcmp(&txalloc.pool->id, &buffer->pool->id, sizeof(txalloc.pool->id))
85 buffer->pool);
49 cb_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
[all...]
H A Dmicro_echo.c49 struct bulk_pool *pool)
52 debug_printf("pool_assigned: RX %p [%d,%d,%d]\n", pool,
53 pool->id.machine, pool->id.dom, pool->id.local);
60 expect_success(bulk_channel_assign_pool(&txc, pool, wait_cont));
62 //but in that case, we don't expect to get a pool over this channel anyway
64 event_dispatch(txc.waitset); //wait until pool is assigned
67 debug_printf("pool_assigned: TX %p [%d,%d,%d]\n", pool,
68 pool
48 cb_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
[all...]
/barrelfish-2018-10-04/lib/bulk_transfer/backends/sm/
H A Dpool.c26 struct bulk_pool *pool; ///< pool to be added to channel member in struct:c_assign_pool_data
28 bulk_ctrl_pool_t flounder_pool; ///< flounder representation of pool
41 * Reply handler for pool_assign RPC call. On success, pool is added to channel.
51 // if accepted, add pool to channel
54 errval_t err = bulk_pool_assign(cdata->pool, channel);
92 * confirmation of pool and replies to the original side.
96 bulk_ctrl_pool_t pool,
113 // Create library representation from pool information recevied by flounder.
116 //first look up if we already know this pool fro
94 bulk_sm_assign_pool_rx_call( struct bulk_ctrl_binding *b, bulk_ctrl_pool_t pool, uint64_t id) argument
197 bulk_sm_assign_pool( struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
222 bulk_sm_remove_pool( struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
[all...]
H A Dflounder_helpers.c154 errval_t create_pool_from_flounder(struct bulk_pool **pool, argument
157 assert(pool);
161 // allocate pool
185 *pool = p;
189 void generate_pool_for_flounder(const struct bulk_pool *pool, argument
192 assert(pool);
195 f_pool->pool_id_machine = pool->id.machine;
196 f_pool->pool_id_dom = pool->id.dom;
197 f_pool->pool_id_local = pool->id.local;
198 f_pool->trust = bulk2flounder_trust(pool
[all...]
/barrelfish-2018-10-04/lib/bulk_transfer/backends/net/
H A Dbulk_net_no_copy.c47 struct bulk_pool *pool; member in struct:bulk_net_nocopy
168 struct bulk_pool *pool; member in struct:pending_pool_request
182 static inline struct bulk_net_pool_data *get_pool_data(struct bulk_pool *pool) argument
184 return ((struct bulk_pool_internal*) pool)->impl_data;
187 static inline struct bulk_net_nocopy *get_net_nocopy(struct bulk_pool *pool) argument
189 struct bulk_net_pool_data *pd = get_pool_data(pool);
200 struct bulk_pool *pool = bulk_pool_get(pool_id, chan); local
201 assert(pool);
202 assert(buffer_id < pool->num_buffers);
203 return pool
215 get_remote_bufid(struct bulk_pool *pool, uint32_t local_buf_id) argument
223 set_remote_bufid(struct bulk_pool *pool, uint32_t local_buf_id, uint32_t remote_buf_id) argument
438 send_pool_assign_response(struct bulk_net_nocopy *p, errval_t err, struct bulk_pool *pool, uint16_t l_port) argument
510 struct bulk_pool *pool = ppr->pool; local
550 send_pool_assign_request(struct bulk_net_nocopy *p, struct bulk_pool *pool, uint16_t l_port) argument
610 struct bulk_pool *pool = bulk_pool_domain_list_get(&id); local
1357 impl_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
1466 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local
1484 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local
1524 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local
1538 struct bulk_pool_internal *pool = (struct bulk_pool_internal *) buffer->pool; local
[all...]
H A Dbulk_net_transparent.c43 struct bulk_pool *pool; member in struct:pending_pool_request
179 struct bulk_pool *pool = bulk_pool_get(pool_id, chan); local
180 assert(pool);
181 assert(buffer_id < pool->num_buffers);
182 return pool->buffers[buffer_id];
353 /* -------------------------- pool assignment -------------------------------*/
357 struct bulk_pool *pool)
370 t->pool_domain_id = pool->id.dom;
371 t->pool_machine_id = pool->id.machine;
372 t->pool_local_id = pool
355 send_pool_assign_response(struct bulk_net_transp *p, errval_t err, struct bulk_pool *pool) argument
419 struct bulk_pool *pool = ppr->pool; local
439 send_pool_assign_request(struct bulk_net_transp *p, struct bulk_pool *pool) argument
485 struct bulk_pool *pool = bulk_pool_domain_list_get(&id); local
1066 impl_channel_assign_pool(struct bulk_channel *channel, struct bulk_pool *pool, struct bulk_continuation cont) argument
[all...]
/barrelfish-2018-10-04/lib/openssl-1.0.0d/crypto/bn/
H A Dbn_ctx.c81 /* How many bignums are in each "pool item"; */
142 BN_POOL pool; member in struct:bignum_ctx
159 BN_POOL_ITEM *item = ctx->pool.head;
206 BN_POOL_reset(&ctx->pool);
223 BN_POOL_init(&ret->pool);
237 BN_POOL_ITEM *pool = ctx->pool.head; local
238 fprintf(stderr,"BN_CTX_free, stack-size=%d, pool-bignums=%d\n",
239 ctx->stack.size, ctx->pool.size);
241 while(pool) {
[all...]
/barrelfish-2018-10-04/usr/tests/bulk_transfer/
H A Dbulk_netproxy.c35 struct bulk_pool *pool);
51 struct bulk_pool *pool)
67 for (int i = 0; i < buffer->pool->buffer_size; ++i) {
81 for (int i = 0; i < buffer->pool->buffer_size; ++i) {
101 assert(arg == rx_allocator.pool);
128 rx_allocator.pool };
130 err = bulk_channel_assign_pool(&rx_channel, rx_allocator.pool, cont);
170 struct bulk_pool *pool);
193 assert(tx_allocator.pool == arg);
202 struct bulk_pool *pool)
50 cb_rx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
201 cb_tx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
[all...]
H A Dbulk_mini.c28 struct bulk_pool *pool);
42 struct bulk_pool *pool)
58 for (int i = 0; i < buffer->pool->buffer_size; ++i) {
72 for (int i = 0; i < buffer->pool->buffer_size; ++i) {
113 struct bulk_pool *pool);
128 assert(tx_allocator.pool == arg);
142 err = bulk_channel_assign_pool(channel, tx_allocator.pool, cont);
147 struct bulk_pool *pool)
206 memset(buffer->address, 123, buffer->pool->buffer_size);
220 memset(buffer->address, 101, buffer->pool
41 cb_rx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
146 cb_tx_pool_assigned(struct bulk_channel *channel, struct bulk_pool *pool) argument
[all...]
/barrelfish-2018-10-04/lib/barrelfish/
H A Dbulk_transfer.c67 current_pbuf->pool = bt;
82 current_pbuf->pool = bt;
113 void *pool; local
114 r = vspace_map_one_frame(&pool, allocated_size, *shared_mem, NULL, NULL);
119 assert(pool != NULL);
121 return bulk_init(pool, size, block_size, bt);
/barrelfish-2018-10-04/include/vm/
H A Dvm_phys.h76 vm_page_t vm_phys_alloc_freelist_pages(int freelist, int pool, int order);
77 vm_page_t vm_phys_alloc_pages(int pool, int order);
89 void vm_phys_set_pool(int pool, vm_page_t m, int order);

Completed in 371 milliseconds

123