/linux-master/drivers/net/ethernet/ti/ |
H A D | cpsw_priv.c | 21 #include <net/page_pool/helpers.h> 1106 struct page_pool *pool; 1113 pool = cpsw->page_pool[ch]; 1147 static struct page_pool *cpsw_create_page_pool(struct cpsw_common *cpsw, 1151 struct page_pool *pool; 1169 struct page_pool *pool; 1177 cpsw->page_pool[ch] = pool; 1186 struct page_pool *pool; 1189 pool = cpsw->page_pool[ch]; 1227 page_pool_destroy(cpsw->page_pool[c [all...] |
H A D | cpsw_priv.h | 360 struct page_pool *page_pool[CPSW_MAX_QUEUES]; member in struct:cpsw_common
|
/linux-master/drivers/net/ethernet/socionext/ |
H A D | netsec.c | 18 #include <net/page_pool/helpers.h> 286 struct page_pool *page_pool; member in struct:netsec_desc_ring 742 page = page_pool_dev_alloc_pages(dring->page_pool); 747 * page_pool API will map the whole page, skip what's needed for 857 /* This is the device Rx buffer from page_pool. No need to remap 863 page_pool_get_dma_dir(rx_ring->page_pool); 922 page_pool_put_page(dring->page_pool, page, sync, true); 932 page_pool_put_page(dring->page_pool, page, sync, true); 944 page_pool_put_page(dring->page_pool, pag [all...] |
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.h | 7 #include <net/page_pool/helpers.h> 709 struct page_pool *pp; 926 * @pool: page_pool to allocate from 932 static inline dma_addr_t idpf_alloc_page(struct page_pool *pool, 970 struct page_pool *pp = page->pp;
|
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 31 #include <net/page_pool/helpers.h> 953 page_pool_put_full_page(rx->page_pool, entry->page, 961 if (rx->page_pool) 962 page_pool_destroy(rx->page_pool); 1011 rx->page_pool = page_pool_create(&pp_params); 1012 if (IS_ERR(rx->page_pool)) { 1013 retval = PTR_ERR(rx->page_pool); 1014 rx->page_pool = NULL; 1079 page_pool_put_full_page(rx->page_pool, *page, false); 1093 rx->page_buffer[i] = page_pool_dev_alloc_pages(rx->page_pool); [all...] |
/linux-master/drivers/net/netdevsim/ |
H A D | netdevsim.h | 96 struct page_pool *page_pool; member in struct:nsim_rq
|
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 6 #include <net/page_pool/helpers.h> 177 page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, false); 272 page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, true); 313 page = page_pool_dev_alloc_pages(rx_ring->page_pool); 2192 page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, false); 2207 page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, false); 2256 if (rx_ring->page_pool) { 2257 page_pool_destroy(rx_ring->page_pool); 2258 rx_ring->page_pool = NULL; 2409 rx_ring->page_pool [all...] |
/linux-master/drivers/net/ethernet/marvell/mvpp2/ |
H A D | mvpp2_main.c | 38 #include <net/page_pool/helpers.h> 104 static struct page_pool * 109 /* internal DMA mapping in page_pool */ 359 /* Returns a struct page if page_pool is set, otherwise a buffer */ 361 struct page_pool *page_pool) 363 if (page_pool) 364 return page_pool_dev_alloc_pages(page_pool); 373 struct page_pool *page_pool, voi 360 mvpp2_frag_alloc(const struct mvpp2_bm_pool *pool, struct page_pool *page_pool) argument 372 mvpp2_frag_free(const struct mvpp2_bm_pool *pool, struct page_pool *page_pool, void *data) argument 761 mvpp2_buf_alloc(struct mvpp2_port *port, struct mvpp2_bm_pool *bm_pool, struct page_pool *page_pool, dma_addr_t *buf_dma_addr, phys_addr_t *buf_phys_addr, gfp_t gfp_mask) argument 3602 mvpp2_rx_refill(struct mvpp2_port *port, struct mvpp2_bm_pool *bm_pool, struct page_pool *page_pool, int pool) argument [all...] |
/linux-master/drivers/net/ethernet/hisilicon/hns3/ |
H A D | hns3_enet.h | 9 #include <net/page_pool/types.h> 472 struct page_pool *page_pool; member in struct:hns3_enet_ring
|
H A D | hns3_debugfs.c | 1132 READ_ONCE(ring->page_pool->pages_state_hold_cnt)); 1134 atomic_read(&ring->page_pool->pages_state_release_cnt)); 1135 sprintf(result[j++], "%u", ring->page_pool->p.pool_size); 1136 sprintf(result[j++], "%u", ring->page_pool->p.order); 1137 sprintf(result[j++], "%d", ring->page_pool->p.nid); 1138 sprintf(result[j++], "%uK", ring->page_pool->p.max_len / 1024); 1157 if (!priv->ring[h->kinfo.num_tqps].page_pool) {
|
H A D | hns3_enet.c | 21 #include <net/page_pool/helpers.h> 3369 if (ring->page_pool) { 3370 p = page_pool_dev_alloc_frag(ring->page_pool, 3410 page_pool_put_full_page(ring->page_pool, cb->priv, 3500 if (ret || ring->page_pool) 3759 if (ring->page_pool) { 4057 page_pool_put_full_page(ring->page_pool, desc_cb->priv, 4067 if (ring->page_pool) 4107 if (ring->page_pool) 4950 ring->page_pool [all...] |
/linux-master/net/core/ |
H A D | xdp.c | 17 #include <net/page_pool/helpers.h> 132 page_pool_destroy(xa->page_pool); 289 return ERR_PTR(-EINVAL); /* Setup time check page_pool req */ 389 * as mem->type knows this a page_pool page 466 page_pool_put_page_bulk(xa->page_pool, bq->q, bq->count);
|
/linux-master/drivers/net/ethernet/microchip/lan966x/ |
H A D | lan966x_main.h | 13 #include <net/page_pool/types.h> 257 struct page_pool *page_pool; member in struct:lan966x_rx
|
/linux-master/drivers/net/ |
H A D | xen-netfront.c | 48 #include <net/page_pool/types.h> 156 struct page_pool *page_pool; member in struct:netfront_queue 281 page = page_pool_alloc_pages(queue->page_pool, 1847 page_pool_destroy(queue->page_pool); 2173 queue->page_pool = page_pool_create(&pp_params); 2174 if (IS_ERR(queue->page_pool)) { 2175 err = PTR_ERR(queue->page_pool); 2176 queue->page_pool = NULL; 2188 MEM_TYPE_PAGE_POOL, queue->page_pool); [all...] |
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 40 #include <net/page_pool/helpers.h> 718 /* page_pool */ 719 struct page_pool *page_pool; member in struct:mvneta_rx_queue 1943 page = page_pool_alloc_pages(rxq->page_pool, 2016 page_pool_put_full_page(rxq->page_pool, data, false); 2020 page_pool_destroy(rxq->page_pool); 2021 rxq->page_pool = NULL; 2082 page_pool_put_full_page(rxq->page_pool, 2086 page_pool_put_page(rxq->page_pool, virt_to_head_pag [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_common.c | 10 #include <net/page_pool/helpers.h> 533 page = page_pool_alloc_frag(pool->page_pool, &offset, sz, GFP_ATOMIC); 546 if (pool->page_pool) 1220 if (pool->page_pool) { 1221 page_pool_put_full_page(pool->page_pool, page, true); 1275 page_pool_destroy(pool->page_pool); 1276 pool->page_pool = NULL; 1407 pool->page_pool = NULL; 1417 pool->page_pool = page_pool_create(&pp_params); 1418 if (IS_ERR(pool->page_pool)) { [all...] |
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 41 #include <net/page_pool/helpers.h> 482 rxq->page_pool = page_pool_create(&pp_params); 483 if (IS_ERR(rxq->page_pool)) { 484 err = PTR_ERR(rxq->page_pool); 485 rxq->page_pool = NULL; 494 rxq->page_pool); 503 page_pool_destroy(rxq->page_pool); 504 rxq->page_pool = NULL; 1583 new_page = page_pool_dev_alloc_pages(rxq->page_pool); 1648 page_pool_put_page(rxq->page_pool, pag [all...] |
H A D | fec.h | 576 /* page_pool */ 577 struct page_pool *page_pool; member in struct:fec_enet_priv_rx_q
|
/linux-master/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_cle.c | 709 if (pdata->rx_ring[idx]->page_pool) { 710 pool_id = pdata->rx_ring[idx]->page_pool->id; 786 if (pdata->rx_ring[0]->page_pool) { 787 pool_id = pdata->rx_ring[0]->page_pool->id;
|
H A D | xgene_enet_main.h | 120 struct xgene_enet_desc_ring *page_pool; member in struct:xgene_enet_desc_ring
|
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | wed.c | 56 dir = page_pool_get_dma_dir(q->page_pool);
|
H A D | dma.c | 455 page_pool_get_dma_dir(q->page_pool)); 467 page_pool_get_dma_dir(q->page_pool)); 660 dir = page_pool_get_dma_dir(q->page_pool); 1011 page_pool_destroy(q->page_pool);
|
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_txrx.h | 239 struct page_pool *pp; /* Used on Rx for buffer management */
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_eth_soc.c | 29 #include <net/page_pool/helpers.h> 1691 static struct page_pool *mtk_create_page_pool(struct mtk_eth *eth, 1704 struct page_pool *pp; 1732 static void *mtk_page_pool_get_buff(struct page_pool *pp, dma_addr_t *dma_addr, 1747 if (ring->page_pool) 1748 page_pool_put_full_page(ring->page_pool, 1986 page_pool_put_full_page(ring->page_pool, 2062 if (ring->page_pool) { 2067 new_data = mtk_page_pool_get_buff(ring->page_pool, 2077 pktlen, page_pool_get_dma_dir(ring->page_pool)); [all...] |
/linux-master/include/net/mana/ |
H A D | mana.h | 336 struct page_pool *page_pool; member in struct:mana_rxq
|