/linux-master/fs/ceph/ |
H A D | util.c | 37 fl->pool_id = le32_to_cpu(legacy->fl_pg_pool); 38 if (fl->pool_id == 0 && fl->stripe_unit == 0 && 40 fl->pool_id = -1; 49 if (fl->pool_id >= 0) 50 legacy->fl_pg_pool = cpu_to_le32(fl->pool_id);
|
H A D | ioctl.c | 29 l.data_pool = ci->i_layout.pool_id; 99 nl.data_pool = ci->i_layout.pool_id; 212 oloc.pool = ci->i_layout.pool_id;
|
H A D | xattr.c | 53 fl->object_size > 0 || fl->pool_id >= 0 || 64 s64 pool = ci->i_layout.pool_id; 167 s64 pool = ci->i_layout.pool_id;
|
H A D | addr.c | 2189 pool = ci->i_layout.pool_id; 2217 if (pool == ci->i_layout.pool_id && 2221 pool = ci->i_layout.pool_id;
|
H A D | file.c | 2334 pool_flags = ceph_pg_pool_flags(osdc->osdmap, ci->i_layout.pool_id); 2832 src_oloc.pool = src_ci->i_layout.pool_id; 2834 dst_oloc.pool = dst_ci->i_layout.pool_id;
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 19 static void otx2_qos_aura_pool_free(struct otx2_nic *pfvf, int pool_id) argument 26 pool = &pfvf->qset.pool[pool_id]; 36 int pool_id, stack_pages, num_sqbs; local 56 pool_id = otx2_get_pool_idx(pfvf, AURA_NIX_SQ, qidx); 57 pool = &pfvf->qset.pool[pool_id]; 60 err = otx2_aura_init(pfvf, pool_id, pool_id, num_sqbs); 65 err = otx2_pool_init(pfvf, pool_id, stack_pages, 88 pfvf->hw_ops->aura_freeptr(pfvf, pool_id, bufptr); 104 otx2_aura_allocptr(pfvf, pool_id); 238 int pool_id, sq_idx, err; local 265 int pool_id, sq_idx; local [all...] |
H A D | otx2_common.c | 988 int err, pool_id, non_xdp_queues; local 1030 pool_id = ((cq->cq_type == CQ_RX) && 1032 cq->rbpool = &qset->pool[pool_id]; 1233 int pool_id, pool_start = 0, pool_end = 0, size = 0; local 1249 for (pool_id = pool_start; pool_id < pool_end; pool_id++) { 1250 iova = otx2_aura_allocptr(pfvf, pool_id); 1251 pool = &pfvf->qset.pool[pool_id]; 1258 iova = otx2_aura_allocptr(pfvf, pool_id); 1266 int pool_id; local 1282 otx2_aura_init(struct otx2_nic *pfvf, int aura_id, int pool_id, int numptrs) argument 1359 otx2_pool_init(struct otx2_nic *pfvf, u16 pool_id, int stack_pages, int numptrs, int buf_size, int type) argument 1428 int qidx, pool_id, stack_pages, num_sqbs; local 1501 int stack_pages, pool_id, rq; local [all...] |
H A D | otx2_common.h | 994 int otx2_pool_init(struct otx2_nic *pfvf, u16 pool_id, 997 int pool_id, int numptrs);
|
H A D | otx2_txrx.c | 1215 u16 pool_id; local 1224 pool_id = otx2_get_pool_idx(pfvf, AURA_NIX_RQ, qidx); 1225 pool = &pfvf->qset.pool[pool_id];
|
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mvneta_bm.c | 40 static void mvneta_bm_pool_enable(struct mvneta_bm *priv, int pool_id) argument 44 val = mvneta_bm_read(priv, MVNETA_BM_POOL_BASE_REG(pool_id)); 46 mvneta_bm_write(priv, MVNETA_BM_POOL_BASE_REG(pool_id), val); 52 static void mvneta_bm_pool_disable(struct mvneta_bm *priv, int pool_id) argument 56 val = mvneta_bm_read(priv, MVNETA_BM_POOL_BASE_REG(pool_id)); 58 mvneta_bm_write(priv, MVNETA_BM_POOL_BASE_REG(pool_id), val); 79 static void mvneta_bm_pool_target_set(struct mvneta_bm *priv, int pool_id, argument 84 val = mvneta_bm_read(priv, MVNETA_BM_XBAR_POOL_REG(pool_id)); 85 val &= ~MVNETA_BM_TARGET_ID_MASK(pool_id); 86 val &= ~MVNETA_BM_XBAR_ATTR_MASK(pool_id); 158 mvneta_bm_pool_use(struct mvneta_bm *priv, u8 pool_id, enum mvneta_bm_type type, u8 port_id, int pkt_size) argument [all...] |
H A D | mvneta_bm.h | 144 struct mvneta_bm_pool *mvneta_bm_pool_use(struct mvneta_bm *priv, u8 pool_id, 175 u8 pool_id, 174 mvneta_bm_pool_use(struct mvneta_bm *priv, u8 pool_id, enum mvneta_bm_type type, u8 port_id, int pkt_size) argument
|
H A D | mvneta.c | 1067 u8 pool_id) 1078 val = mvreg_read(pp, MVNETA_PORT_POOL_BUFFER_SZ_REG(pool_id)); 1080 mvreg_write(pp, MVNETA_PORT_POOL_BUFFER_SZ_REG(pool_id), val); 1999 u8 pool_id = MVNETA_RX_GET_BM_POOL_ID(rx_desc); local 2002 bm_pool = &pp->bm_priv->bm_pools[pool_id]; 2564 u8 pool_id; local 2571 pool_id = MVNETA_RX_GET_BM_POOL_ID(rx_desc); 2572 bm_pool = &pp->bm_priv->bm_pools[pool_id]; 1065 mvneta_bm_pool_bufsize_set(struct mvneta_port *pp, int buf_size, u8 pool_id) argument
|
/linux-master/drivers/net/wireless/ath/ath12k/ |
H A D | dp_tx.c | 81 u8 pool_id) 83 spin_lock_bh(&dp->tx_desc_lock[pool_id]); 84 list_move_tail(&tx_desc->list, &dp->tx_desc_free_list[pool_id]); 85 spin_unlock_bh(&dp->tx_desc_lock[pool_id]); 89 u8 pool_id) 93 spin_lock_bh(&dp->tx_desc_lock[pool_id]); 94 desc = list_first_entry_or_null(&dp->tx_desc_free_list[pool_id], 98 spin_unlock_bh(&dp->tx_desc_lock[pool_id]); 103 list_move_tail(&desc->list, &dp->tx_desc_used_list[pool_id]); 104 spin_unlock_bh(&dp->tx_desc_lock[pool_id]); 79 ath12k_dp_tx_release_txbuf(struct ath12k_dp *dp, struct ath12k_tx_desc_info *tx_desc, u8 pool_id) argument 88 ath12k_dp_tx_assign_buffer(struct ath12k_dp *dp, u8 pool_id) argument 142 u8 pool_id; local [all...] |
H A D | dp.c | 1154 u32 pool_id, tx_spt_page; local 1204 for (pool_id = 0; pool_id < ATH12K_HW_MAX_QUEUES; pool_id++) { 1205 spin_lock_bh(&dp->tx_desc_lock[pool_id]); 1208 tx_spt_page = i + pool_id * ATH12K_TX_SPT_PAGES_PER_POOL; 1216 spin_unlock_bh(&dp->tx_desc_lock[pool_id]); 1377 u32 i, j, pool_id, tx_spt_page; local 1407 for (pool_id = 0; pool_id < ATH12K_HW_MAX_QUEUE [all...] |
H A D | dp.h | 292 u8 pool_id; member in struct:ath12k_tx_desc_info
|
/linux-master/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_cle.c | 701 u16 pool_id; local 705 pool_id = pdata->rx_ring[idx]->buf_pool->id; 706 fpsel = xgene_enet_get_fpsel(pool_id); 710 pool_id = pdata->rx_ring[idx]->page_pool->id; 711 nfpsel = xgene_enet_get_fpsel(pool_id); 764 u32 def_qid, def_fpsel, def_nxtfpsel, pool_id; local 783 pool_id = pdata->rx_ring[0]->buf_pool->id; 784 def_fpsel = xgene_enet_get_fpsel(pool_id); 787 pool_id = pdata->rx_ring[0]->page_pool->id; 788 def_nxtfpsel = xgene_enet_get_fpsel(pool_id); [all...] |
/linux-master/include/xen/interface/ |
H A D | xen.h | 744 int32_t pool_id; member in struct:tmem_op
|
/linux-master/tools/workqueue/ |
H A D | wq_dump.py | 185 pool_id = per_cpu_ptr(wq.cpu_pwq, cpu)[0].pool.id.value_() variable 187 print(f' {pool_id:{field_len}}', end='')
|
/linux-master/drivers/block/ |
H A D | rbd.c | 165 * The tuple (pool_id, image_id, snap_id) is sufficient to uniquely 188 u64 pool_id; member in struct:rbd_spec 992 rbd_dev->layout.pool_id = rbd_dev->header.data_pool_id == CEPH_NOPOOL ? 993 rbd_dev->spec->pool_id : rbd_dev->header.data_pool_id; 1450 req->r_base_oloc.pool = rbd_dev->layout.pool_id; 5113 (unsigned long long) rbd_dev->spec->pool_id); 5183 "pool_id %llu\npool_name %s\n" 5189 spec->pool_id, spec->pool_name, 5226 static DEVICE_ATTR(pool_id, 0444, rbd_pool_id_show, NULL); 5295 spec->pool_id 5595 u64 pool_id; member in struct:parent_image_info [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 653 int32_t pool_id; member in struct:amdgpu_ttm_tt 1084 gtt->pool_id = KFD_XCP_MEM_ID(adev, abo->xcp_id); 1086 gtt->pool_id = abo->xcp_id; 1128 if (adev->mman.ttm_pools && gtt->pool_id >= 0) 1129 pool = &adev->mman.ttm_pools[gtt->pool_id]; 1173 if (adev->mman.ttm_pools && gtt->pool_id >= 0) 1174 pool = &adev->mman.ttm_pools[gtt->pool_id];
|
/linux-master/net/ipv4/ |
H A D | tcp_ao.c | 1528 int err, pool_id; local 1541 pool_id = tcp_sigpool_alloc_ahash(algo, 60); 1542 if (pool_id < 0) 1543 return ERR_PTR(pool_id); 1545 err = tcp_sigpool_start(pool_id, &hp); 1560 key->tcp_sigpool_id = pool_id; 1565 tcp_sigpool_release(pool_id);
|
/linux-master/drivers/net/wireless/ath/ath11k/ |
H A D | dp_tx.c | 94 u8 pool_id; local 108 pool_id = skb_get_queue_mapping(skb) & (ATH11K_HW_MAX_QUEUES - 1); 141 FIELD_PREP(DP_TX_DESC_ID_POOL_ID, pool_id);
|
/linux-master/kernel/ |
H A D | workqueue.c | 799 int pool_id, unsigned long flags) 801 set_work_data(work, ((unsigned long)pool_id << WORK_OFFQ_POOL_SHIFT) | 806 int pool_id, unsigned long flags) 815 set_work_data(work, ((unsigned long)pool_id << WORK_OFFQ_POOL_SHIFT) | 881 int pool_id; local 888 pool_id = data >> WORK_OFFQ_POOL_SHIFT; 889 if (pool_id == WORK_OFFQ_POOL_NONE) 892 return idr_find(&worker_pool_idr, pool_id); 914 unsigned long pool_id = get_work_pool_id(work); local 916 pool_id << 798 set_work_pool_and_keep_pending(struct work_struct *work, int pool_id, unsigned long flags) argument 805 set_work_pool_and_clear_pending(struct work_struct *work, int pool_id, unsigned long flags) argument [all...] |
/linux-master/drivers/s390/block/ |
H A D | dasd_eckd.h | 430 __u16 pool_id; member in struct:dasd_ext_pool_sum
|
/linux-master/include/linux/ceph/ |
H A D | ceph_fs.h | 66 s64 pool_id; /* rados pool id */ member in struct:ceph_file_layout
|