/linux-master/kernel/dma/ |
H A D | swiotlb.c | 690 size_t pool_size; local 698 pool_size = sizeof(*pool) + array_size(sizeof(*pool->areas), nareas); 699 pool = kzalloc(pool_size, gfp);
|
/linux-master/mm/ |
H A D | zswap.c | 283 u64 pool_size = 0; local 287 pool_size += zpool_get_total_size(pool->zpools[i]); 289 return pool_size;
|
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 2401 .pool_size = rx_ring->size,
|
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 3558 pp.pool_size = bp->rx_agg_ring_size; 3560 pp.pool_size += bp->rx_ring_size;
|
/linux-master/net/core/ |
H A D | dev.c | 11719 .pool_size = SYSTEM_PERCPU_PAGE_POOL_SIZE,
|
H A D | page_pool.c | 190 if (pool->p.pool_size) 191 ring_qsize = pool->p.pool_size;
|
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_main.c | 2022 pp_params.pool_size = dma_conf->dma_rx_size;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 740 u32 *pool_size, 770 *pool_size += (rq->mpwqe.shampo->hd_per_wqe * wq_size) / 803 u32 pool_size; local 817 pool_size = 1 << params->log_rq_mtu_frames; 848 pool_size = rq->mpwqe.pages_per_wqe << 852 pool_size *= 2; /* additional page per packet for the linear part */ 869 err = mlx5_rq_shampo_alloc(mdev, params, rqp, rq, &pool_size, node); 902 pp_params.pool_size = pool_size; 736 mlx5_rq_shampo_alloc(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_rq_param *rqp, struct mlx5e_rq *rq, u32 *pool_size, int node) argument
|
/linux-master/drivers/gpu/drm/amd/pm/swsmu/inc/ |
H A D | amdgpu_smu.h | 496 uint64_t pool_size; member in struct:smu_context
|
/linux-master/drivers/gpu/drm/amd/pm/swsmu/ |
H A D | amdgpu_smu.c | 982 uint64_t pool_size = smu->pool_size; local 985 if (pool_size == SMU_MEMORY_POOL_SIZE_ZERO) 988 memory_pool->size = pool_size; 992 switch (pool_size) { 1212 smu->pool_size = adev->pm.smu_prv_buffer_size;
|
/linux-master/drivers/net/ |
H A D | xen-netfront.c | 2166 .pool_size = NET_RX_RING_SIZE,
|
H A D | veth.c | 968 .pool_size = VETH_RING_SIZE,
|
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_en.c | 2138 pprm.pool_size = RX_BUFFERS_PER_QUEUE;
|
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.c | 591 .pool_size = rxbufq->desc_count,
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_virtchnl_pf.c | 2708 unsigned int i, cur_queues, more, pool_size; local 2738 pool_size = 0; 2741 pool_size = 0; 2744 if (needed <= ++pool_size)
|
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 473 .pool_size = size,
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_eth_soc.c | 1698 .pool_size = size,
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 1413 pp_params.pool_size = min(OTX2_PAGE_POOL_SZ, numptrs);
|
/linux-master/drivers/mtd/ubi/ |
H A D | fastmap.c | 393 * @pool_size: size of the pool (number of entries in @pebs) 401 __be32 *pebs, int pool_size, unsigned long long *max_sqnum, 422 dbg_bld("scanning fastmap pool: size = %i", pool_size); 428 for (i = 0; i < pool_size; i++) { 561 int ret, i, j, pool_size, wl_pool_size; local 607 pool_size = be16_to_cpu(fmpl->size); 612 if (pool_size > UBI_FM_MAX_POOL_SIZE || pool_size < 0) { 613 ubi_err(ubi, "bad pool size: %i", pool_size); 765 ret = scan_pool(ubi, ai, fmpl->pebs, pool_size, 400 scan_pool(struct ubi_device *ubi, struct ubi_attach_info *ai, __be32 *pebs, int pool_size, unsigned long long *max_sqnum, struct list_head *free) argument [all...] |
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | lparcfg.c | 281 if (mpp_data.pool_size != -1) 283 mpp_data.pool_size);
|
/linux-master/mm/kasan/ |
H A D | kasan_test.c | 1066 int pool_size = 4; local 1071 ret = mempool_init_kmalloc_pool(pool, pool_size, size); 1089 int pool_size = 4; local 1096 ret = mempool_init_slab_pool(pool, pool_size, cache); 1109 int pool_size = 4; local 1114 ret = mempool_init_page_pool(pool, pool_size, order);
|
/linux-master/drivers/md/dm-vdo/ |
H A D | vio.c | 302 * @pool_size: The number of vios in the pool. 311 int make_vio_pool(struct vdo *vdo, size_t pool_size, thread_id_t thread_id, argument 319 result = vdo_allocate_extended(struct vio_pool, pool_size, struct pooled_vio, 328 result = vdo_allocate(pool_size * VDO_BLOCK_SIZE, char, 336 for (pool->size = 0; pool->size < pool_size; pool->size++, ptr += VDO_BLOCK_SIZE) {
|
H A D | vio.h | 191 int __must_check make_vio_pool(struct vdo *vdo, size_t pool_size, thread_id_t thread_id,
|
H A D | data-vio.h | 331 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size,
|
H A D | data-vio.c | 837 * @pool_size: The number of data_vios in the pool. 841 int make_data_vio_pool(struct vdo *vdo, data_vio_count_t pool_size, argument 848 result = vdo_allocate_extended(struct data_vio_pool, pool_size, struct data_vio, 853 VDO_ASSERT_LOG_ONLY((discard_limit <= pool_size), 858 initialize_limiter(&pool->limiter, pool, assign_data_vio_to_waiter, pool_size); 874 for (i = 0; i < pool_size; i++) {
|