/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | pool.h | 12 if (!xsk || !xsk->pools) 18 return xsk->pools[ix];
|
H A D | pool.c | 25 if (!xsk->pools) { 26 xsk->pools = kcalloc(MLX5E_MAX_NUM_CHANNELS, 27 sizeof(*xsk->pools), GFP_KERNEL); 28 if (unlikely(!xsk->pools)) 41 kfree(xsk->pools); 42 xsk->pools = NULL; 54 xsk->pools[ix] = pool; 60 xsk->pools[ix] = NULL;
|
/linux-master/tools/testing/selftests/drivers/net/mlxsw/ |
H A D | sharedbuffer_configuration.py | 28 def __init__(self, pools): 30 for pool in pools: 72 # The threshold type of pools 4, 8, 9 and 10 cannot be changed 110 # The threshold type of pools 4, 8, 9 and 10 cannot be changed 187 pools = PoolList() 190 pools.append(Pool(pooldict)) 191 return pools 194 def do_check_pools(dlname, pools, vp): 195 for pool in pools: 219 def check_pools(dlname, pools) [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_arg.c | 28 struct dr_arg_pool *pools[DR_ARG_CHUNK_SIZE_MAX]; member in struct:mlx5dr_arg_mgr 201 arg_obj = dr_arg_pool_get_arg_obj(mgr->pools[size]); 226 dr_arg_pool_put_arg_obj(mgr->pools[arg_obj->log_chunk_size], arg_obj); 245 pool_mgr->pools[i] = dr_arg_pool_create(dmn, i); 246 if (!pool_mgr->pools[i]) 254 dr_arg_pool_destroy(pool_mgr->pools[i]); 262 struct dr_arg_pool **pools; local 268 pools = mgr->pools; 270 dr_arg_pool_destroy(pools[ [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | iommu-common.c | 82 spin_lock_init(&(iommu->pools[i].lock)); 83 iommu->pools[i].start = start; 84 iommu->pools[i].hint = start; 86 iommu->pools[i].end = start - 1; 131 pool = &(iommu->pools[pool_nr]); 161 pool = &(iommu->pools[0]); 193 pool = &(iommu->pools[pool_nr]); 237 p = &tbl->pools[pool_nr];
|
/linux-master/tools/net/ynl/samples/ |
H A D | page-pool.c | 60 struct netdev_page_pool_get_list *pools; local 76 pools = netdev_page_pool_get_dump(ys); 77 if (!pools) 80 ynl_dump_foreach(pools, pp) { 87 netdev_page_pool_get_list_free(pools); 124 printf("page pools: %u (zombies: %u)\n",
|
/linux-master/arch/sparc/include/asm/ |
H A D | iommu-common.h | 26 struct iommu_pool pools[IOMMU_NR_POOLS]; member in struct:iommu_map_table
|
/linux-master/drivers/net/ethernet/chelsio/libcxgb/ |
H A D | libcxgb_ppm.c | 348 struct cxgbi_ppm_pool *pools; local 350 unsigned int max = (PCPU_MIN_UNIT_SIZE - sizeof(*pools)) << 3; 367 alloc_sz = sizeof(*pools) + sizeof(unsigned long) * bmap; 368 pools = __alloc_percpu(alloc_sz, __alignof__(struct cxgbi_ppm_pool)); 370 if (!pools) 374 struct cxgbi_ppm_pool *ppool = per_cpu_ptr(pools, cpu); 384 return pools;
|
/linux-master/mm/ |
H A D | dmapool.c | 60 struct list_head pools; member in struct:dma_pool 80 list_for_each_entry(pool, &dev->dma_pools, pools) { 92 static DEVICE_ATTR_RO(pools); 210 * Given one of these pools, dma_pool_alloc() 267 INIT_LIST_HEAD(&retval->pools); 280 list_add(&retval->pools, &dev->dma_pools); 288 list_del(&retval->pools); 370 list_del(&pool->pools);
|
/linux-master/drivers/soc/ti/ |
H A D | knav_qmss.h | 191 * @pools: list of descriptor pools in the region 203 struct list_head pools; member in struct:knav_region 207 * struct knav_pool: qmss pools 304 struct list_head pools; member in struct:knav_device 363 list_for_each_entry(pool, &kdev->pools, list)
|
H A D | knav_qmss_queue.c | 815 /* Region maintains a sorted (by region offset) list of pools 820 node = ®ion->pools; 821 list_for_each_entry(iter, ®ion->pools, region_inst) { 835 list_add_tail(&pool->list, &kdev->pools); 1037 list_add(&pool->region_inst, ®ion->pools); 1121 INIT_LIST_HEAD(®ion->pools); 1359 list_for_each_entry_safe(pool, tmp, ®ion->pools, region_inst) 1781 INIT_LIST_HEAD(&kdev->pools); 1827 queue_pools = of_get_child_by_name(node, "queue-pools"); 1829 dev_err(dev, "queue-pools no [all...] |
/linux-master/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-xsk.c | 162 pools_params->pools[curr_bp].priority_mask |= (1 << j); 163 if (!pools_params->pools[curr_bp].priority_mask) 166 pools_params->pools[curr_bp].dpbp_id = priv->bp[i]->bpid; 167 pools_params->pools[curr_bp].buffer_size = priv->rx_buf_size; 168 pools_params->pools[curr_bp++].backup_pool = 0;
|
H A D | dpsw.c | 1138 * dpsw_ctrl_if_set_pools() - Set control interface buffer pools 1142 * @cfg: Buffer pools configuration 1158 cmd_params->dpbp_id[i] = cpu_to_le32(cfg->pools[i].dpbp_id); 1160 cpu_to_le16(cfg->pools[i].buffer_size); 1162 DPSW_BACKUP_POOL(cfg->pools[i].backup_pool, i);
|
H A D | dpsw.h | 196 * struct dpsw_ctrl_if_pools_cfg - Control interface buffer pools configuration 198 * @pools: Array of buffer pools parameters; The number of valid entries 200 * @pools.dpbp_id: DPBP object ID 201 * @pools.buffer_size: Buffer size 202 * @pools.backup_pool: Backup pool 210 } pools[DPSW_MAX_DPBP]; member in struct:dpsw_ctrl_if_pools_cfg
|
H A D | dpni.c | 150 * dpni_set_pools() - Set buffer pools configuration 154 * @cfg: Buffer pools configuration 179 cpu_to_le16(cfg->pools[i].dpbp_id); 181 cfg->pools[i].priority_mask; 183 cpu_to_le16(cfg->pools[i].buffer_size); 185 DPNI_BACKUP_POOL(cfg->pools[i].backup_pool, i);
|
/linux-master/arch/powerpc/kernel/ |
H A D | iommu.c | 100 * The hash is important to spread CPUs across all the pools. For example, 102 * with 4 pools all primary threads would map to the same pool. 252 pool = &(tbl->pools[pool_nr]); 280 pool = &(tbl->pools[0]); 299 /* Now try scanning all the other pools */ 302 pool = &tbl->pools[pool_nr]; 429 p = &tbl->pools[pool_nr]; 755 p = &tbl->pools[i]; 1122 spin_lock_nest_lock(&tbl->pools[i].lock, &tbl->large_pool.lock); 1132 spin_unlock(&tbl->pools[ [all...] |
/linux-master/include/linux/ |
H A D | swiotlb.h | 109 * @nslabs: Total number of IO TLB slabs in all pools. 113 * @can_grow: %true if more pools can be allocated dynamically. 116 * @pools: List of IO TLB memory pool descriptors (if dynamic). 123 * @transient_nslabs: The total number of slots in all transient pools that 136 struct list_head pools; member in struct:io_tlb_mem 185 * dev->dma_uses_io_tlb here and mem->pools in swiotlb_find_pool().
|
/linux-master/drivers/md/ |
H A D | dm-table.c | 1024 struct dm_md_mempools *pools; local 1031 pools = kzalloc_node(sizeof(*pools), GFP_KERNEL, md->numa_node_id); 1032 if (!pools) 1053 if (bioset_init(&pools->io_bs, pool_size, io_front_pad, 1057 bioset_integrity_create(&pools->io_bs, pool_size)) 1060 if (bioset_init(&pools->bs, pool_size, front_pad, 0)) 1063 bioset_integrity_create(&pools->bs, pool_size)) 1066 t->mempools = pools; 1070 dm_free_md_mempools(pools); [all...] |
H A D | dm.h | 224 void dm_free_md_mempools(struct dm_md_mempools *pools);
|
H A D | dm.c | 3165 void dm_free_md_mempools(struct dm_md_mempools *pools) 3167 if (!pools) 3170 bioset_exit(&pools->bs); 3171 bioset_exit(&pools->io_bs); 3173 kfree(pools); 3159 dm_free_md_mempools(struct dm_md_mempools *pools) argument
|
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_hw.c | 599 * @pools: VMDq "set" or "pool" index 604 static int wx_set_rar(struct wx *wx, u32 index, u8 *addr, u64 pools, argument 620 wr32(wx, WX_PSR_MAC_SWC_VM_L, pools & 0xFFFFFFFF); 622 wr32(wx, WX_PSR_MAC_SWC_VM_H, pools >> 32); 797 wx->mac_table[i].pools, 811 wx->mac_table[0].pools = 1ULL; 814 wx->mac_table[0].pools, 830 wx->mac_table[i].pools = 0; 846 if (wx->mac_table[i].pools != (1ULL << pool)) { 848 wx->mac_table[i].pools | [all...] |
/linux-master/kernel/dma/ |
H A D | swiotlb.c | 91 .pools = LIST_HEAD_INIT(io_tlb_default_mem.pools), 310 list_add_rcu(&pool->node, &mem->pools); 618 * Allocate from the atomic pools if memory is encrypted and 781 list_for_each_entry_rcu(pool, &mem->pools, node) { 1125 * swiotlb_search_area() - search one memory area in all pools 1135 * Search one memory area in all pools for a sequence of slots that match the 1150 list_for_each_entry_rcu(pool, &mem->pools, node) { 1240 * Second, the load from mem->pools must be ordered before the same 1331 list_for_each_entry_rcu(pool, &mem->pools, nod [all...] |
/linux-master/drivers/soc/fsl/qbman/ |
H A D | qman_priv.h | 177 u32 pools; member in struct:qm_portal_config
|
H A D | qman_portal.c | 248 pcfg->pools = qm_get_pools_sdqcr();
|
/linux-master/arch/powerpc/include/asm/ |
H A D | iommu.h | 108 struct iommu_pool pools[IOMMU_NR_POOLS]; member in struct:iommu_table
|