Searched refs:pool_size (Results 1 - 25 of 70) sorted by last modified time

123

/linux-master/kernel/dma/
H A Dswiotlb.c690 size_t pool_size; local
698 pool_size = sizeof(*pool) + array_size(sizeof(*pool->areas), nareas);
699 pool = kzalloc(pool_size, gfp);
/linux-master/mm/
H A Dzswap.c283 u64 pool_size = 0; local
287 pool_size += zpool_get_total_size(pool->zpools[i]);
289 return pool_size;
/linux-master/drivers/net/ethernet/wangxun/libwx/
H A Dwx_lib.c2401 .pool_size = rx_ring->size,
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt.c3558 pp.pool_size = bp->rx_agg_ring_size;
3560 pp.pool_size += bp->rx_ring_size;
/linux-master/net/core/
H A Ddev.c11719 .pool_size = SYSTEM_PERCPU_PAGE_POOL_SIZE,
H A Dpage_pool.c190 if (pool->p.pool_size)
191 ring_qsize = pool->p.pool_size;
/linux-master/drivers/net/ethernet/stmicro/stmmac/
H A Dstmmac_main.c2022 pp_params.pool_size = dma_conf->dma_rx_size;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c740 u32 *pool_size,
770 *pool_size += (rq->mpwqe.shampo->hd_per_wqe * wq_size) /
803 u32 pool_size; local
817 pool_size = 1 << params->log_rq_mtu_frames;
848 pool_size = rq->mpwqe.pages_per_wqe <<
852 pool_size *= 2; /* additional page per packet for the linear part */
869 err = mlx5_rq_shampo_alloc(mdev, params, rqp, rq, &pool_size, node);
902 pp_params.pool_size = pool_size;
736 mlx5_rq_shampo_alloc(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_rq_param *rqp, struct mlx5e_rq *rq, u32 *pool_size, int node) argument
/linux-master/drivers/gpu/drm/amd/pm/swsmu/inc/
H A Damdgpu_smu.h496 uint64_t pool_size; member in struct:smu_context
/linux-master/drivers/gpu/drm/amd/pm/swsmu/
H A Damdgpu_smu.c982 uint64_t pool_size = smu->pool_size; local
985 if (pool_size == SMU_MEMORY_POOL_SIZE_ZERO)
988 memory_pool->size = pool_size;
992 switch (pool_size) {
1212 smu->pool_size = adev->pm.smu_prv_buffer_size;
/linux-master/drivers/net/
H A Dxen-netfront.c2166 .pool_size = NET_RX_RING_SIZE,
H A Dveth.c968 .pool_size = VETH_RING_SIZE,
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c2138 pprm.pool_size = RX_BUFFERS_PER_QUEUE;
/linux-master/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c591 .pool_size = rxbufq->desc_count,
/linux-master/drivers/net/ethernet/intel/i40e/
H A Di40e_virtchnl_pf.c2708 unsigned int i, cur_queues, more, pool_size; local
2738 pool_size = 0;
2741 pool_size = 0;
2744 if (needed <= ++pool_size)
/linux-master/drivers/net/ethernet/freescale/
H A Dfec_main.c473 .pool_size = size,
/linux-master/drivers/net/ethernet/mediatek/
H A Dmtk_eth_soc.c1698 .pool_size = size,
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_common.c1413 pp_params.pool_size = min(OTX2_PAGE_POOL_SZ, numptrs);
/linux-master/drivers/mtd/ubi/
H A Dfastmap.c393 * @pool_size: size of the pool (number of entries in @pebs)
401 __be32 *pebs, int pool_size, unsigned long long *max_sqnum,
422 dbg_bld("scanning fastmap pool: size = %i", pool_size);
428 for (i = 0; i < pool_size; i++) {
561 int ret, i, j, pool_size, wl_pool_size; local
607 pool_size = be16_to_cpu(fmpl->size);
612 if (pool_size > UBI_FM_MAX_POOL_SIZE || pool_size < 0) {
613 ubi_err(ubi, "bad pool size: %i", pool_size);
765 ret = scan_pool(ubi, ai, fmpl->pebs, pool_size,
400 scan_pool(struct ubi_device *ubi, struct ubi_attach_info *ai, __be32 *pebs, int pool_size, unsigned long long *max_sqnum, struct list_head *free) argument
[all...]
/linux-master/arch/powerpc/platforms/pseries/
H A Dlparcfg.c281 if (mpp_data.pool_size != -1)
283 mpp_data.pool_size);
/linux-master/mm/kasan/
H A Dkasan_test.c1066 int pool_size = 4; local
1071 ret = mempool_init_kmalloc_pool(pool, pool_size, size);
1089 int pool_size = 4; local
1096 ret = mempool_init_slab_pool(pool, pool_size, cache);
1109 int pool_size = 4; local
1114 ret = mempool_init_page_pool(pool, pool_size, order);
/linux-master/drivers/md/dm-vdo/
H A Dvio.c302 * @pool_size: The number of vios in the pool.
311 int make_vio_pool(struct vdo *vdo, size_t pool_size, thread_id_t thread_id, argument
319 result = vdo_allocate_extended(struct vio_pool, pool_size, struct pooled_vio,
328 result = vdo_allocate(pool_size * VDO_BLOCK_SIZE, char,
336 for (pool->size = 0; pool->size < pool_size; pool->size++, ptr += VDO_BLOCK_SIZE) {
H A Dvio.h191 int __must_check make_vio_pool(struct vdo *vdo, size_t pool_size, thread_id_t thread_id,
H A Ddata-vio.h331 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size,
H A Ddata-vio.c837 * @pool_size: The number of data_vios in the pool.
841 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size, argument
848 result = vdo_allocate_extended(struct data_vio_pool, pool_size, struct data_vio,
853 VDO_ASSERT_LOG_ONLY((discard_limit <= pool_size),
858 initialize_limiter(&pool->limiter, pool, assign_data_vio_to_waiter, pool_size);
874 for (i = 0; i < pool_size; i++) {

Completed in 556 milliseconds

123