/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 1755 /* keep pages to prevent allocation failures when xsk is 2284 return tsnep_xdp_setup_pool(adapter, bpf->xsk.pool, 2285 bpf->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/freescale/dpaa2/ |
H A D | Makefile | 10 fsl-dpaa2-eth-objs := dpaa2-eth.o dpaa2-ethtool.o dpni.o dpaa2-mac.o dpmac.o dpaa2-eth-devlink.o dpaa2-xsk.o
|
H A D | dpaa2-eth.c | 300 xdp_buff = swa->xsk.xdp_buff; 1191 dma_unmap_single(dev, fd_addr, swa->xsk.sgt_size, 1728 swa->xsk.xdp_buff = xdp_buffs[i]; 2799 return dpaa2_xsk_setup_pool(dev, xdp->xsk.pool, xdp->xsk.queue_id);
|
H A D | dpaa2-eth.h | 165 } xsk; member in union:dpaa2_eth_swa::__anon1
|
H A D | dpaa2-xsk.c | 50 xdp_buff = swa->xsk.xdp_buff; 365 swa->xsk.sgt_size = sgt_buf_size;
|
/linux-master/drivers/net/ethernet/google/gve/ |
H A D | gve_main.c | 1557 dev_err(&priv->pdev->dev, "xsk pool invalid qid %d", qid); 1562 dev_err(&priv->pdev->dev, "xsk pool frame_len too small"); 1724 if (xdp->xsk.pool) 1725 return gve_xsk_pool_enable(dev, xdp->xsk.pool, xdp->xsk.queue_id); 1727 return gve_xsk_pool_disable(dev, xdp->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 13623 return i40e_xsk_pool_setup(vsi, xdp->xsk.pool, 13624 xdp->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_main.c | 3074 return ice_xsk_pool_setup(vsi, xdp->xsk.pool, 3075 xdp->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_main.c | 6562 return igc_xdp_setup_pool(adapter, bpf->xsk.pool, 6563 bpf->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_main.c | 10346 return ixgbe_xsk_pool_setup(adapter, xdp->xsk.pool, 10347 xdp->xsk.queue_id);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | Makefile | 29 en/reporter_tx.o en/reporter_rx.o en/params.o en/xsk/pool.o \ 30 en/xsk/setup.o en/xsk/rx.o en/xsk/tx.o en/devlink.o en/ptp.o \
|
H A D | en.h | 327 struct mlx5e_xsk *xsk; member in struct:mlx5e_params 352 MLX5E_RQ_STATE_XSK, /* set to indicate an xsk rq */ 613 int mlx5e_rq_set_handlers(struct mlx5e_rq *rq, struct mlx5e_params *params, bool xsk); 949 struct mlx5e_xsk xsk; member in struct:mlx5e_priv 1041 struct mlx5e_xsk_param *xsk, int node, u16 q_counter, 1237 void mlx5e_build_nic_params(struct mlx5e_priv *priv, struct mlx5e_xsk *xsk, u16 mtu);
|
H A D | en_main.c | 63 #include "en/xsk/pool.h" 64 #include "en/xsk/setup.h" 65 #include "en/xsk/rx.h" 66 #include "en/xsk/tx.h" 798 struct mlx5e_xsk_param *xsk, 818 rq->buff.headroom = mlx5e_get_rq_headroom(mdev, params, xsk); 838 rq->mpwqe.page_shift = mlx5e_mpwrq_page_shift(mdev, xsk); 839 rq->mpwqe.umr_mode = mlx5e_mpwrq_umr_mode(mdev, xsk); 851 mlx5e_mpwqe_get_log_rq_size(mdev, params, xsk); 853 if (!mlx5e_rx_mpwqe_is_linear_skb(mdev, params, xsk) 797 mlx5e_alloc_rq(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_rq_param *rqp, int node, struct mlx5e_rq *rq) argument 1270 mlx5e_open_rq(struct mlx5e_params *params, struct mlx5e_rq_param *param, struct mlx5e_xsk_param *xsk, int node, u16 q_counter, struct mlx5e_rq *rq) argument 2637 struct mlx5e_xsk_param xsk; local 4403 struct mlx5e_xsk_param xsk; local 5160 mlx5e_build_nic_params(struct mlx5e_priv *priv, struct mlx5e_xsk *xsk, u16 mtu) argument [all...] |
H A D | en_rx.c | 57 #include "en/xsk/rx.h" 2683 int mlx5e_rq_set_handlers(struct mlx5e_rq *rq, struct mlx5e_params *params, bool xsk) argument 2691 rq->mpwqe.skb_from_cqe_mpwrq = xsk ? 2715 rq->wqe.skb_from_cqe = xsk ?
|
H A D | en_stats.c | 2435 (NUM_XSKRQ_STATS * max_nch * priv->xsk.ever_used) + 2436 (NUM_XSKSQ_STATS * max_nch * priv->xsk.ever_used); 2441 bool is_xsk = priv->xsk.ever_used; 2474 bool is_xsk = priv->xsk.ever_used;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.c | 20 u8 mlx5e_mpwrq_page_shift(struct mlx5_core_dev *mdev, struct mlx5e_xsk_param *xsk) argument 22 u8 req_page_shift = xsk ? order_base_2(xsk->chunk_size) : PAGE_SHIFT; 26 if (WARN_ON_ONCE(!xsk && req_page_shift < min_page_shift)) 33 mlx5e_mpwrq_umr_mode(struct mlx5_core_dev *mdev, struct mlx5e_xsk_param *xsk) argument 44 u8 page_shift = mlx5e_mpwrq_page_shift(mdev, xsk); 45 bool unaligned = xsk ? xsk->unaligned : false; 48 if (xsk) { 49 oversized = xsk 219 mlx5e_get_linear_rq_headroom(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 236 mlx5e_rx_get_linear_sz_xsk(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 256 mlx5e_rx_get_linear_stride_sz(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, bool mpwqe) argument 284 mlx5e_mpwqe_log_pkts_per_wqe(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 296 mlx5e_rx_is_linear_skb(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 342 mlx5e_verify_params_rx_mpwqe_strides(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 356 mlx5e_rx_mpwqe_is_linear_skb(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 382 mlx5e_mpwqe_get_log_rq_size(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 431 mlx5e_mpwqe_get_log_stride_size(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 445 mlx5e_mpwqe_get_log_num_strides(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 467 mlx5e_get_rq_headroom(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 544 mlx5e_mpwrq_validate_xsk(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 679 mlx5e_build_rq_frags_info(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_rq_frags_info *info, u32 *xdp_frag_size) argument 826 mlx5e_shampo_get_log_cq_size(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 843 mlx5e_build_rx_cq_param(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_cq_param *param) argument 889 mlx5e_build_rq_param(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_rq_param *param) argument 1090 mlx5e_mpwrq_total_umr_wqebbs(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 1133 struct mlx5e_xsk_param xsk = { local 1218 mlx5e_build_xdpsq_param(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_sq_param *param) argument [all...] |
H A D | params.h | 59 u8 mlx5e_mpwrq_page_shift(struct mlx5_core_dev *mdev, struct mlx5e_xsk_param *xsk); 61 mlx5e_mpwrq_umr_mode(struct mlx5_core_dev *mdev, struct mlx5e_xsk_param *xsk); 83 struct mlx5e_xsk_param *xsk); 89 struct mlx5e_xsk_param *xsk); 92 struct mlx5e_xsk_param *xsk); 95 struct mlx5e_xsk_param *xsk); 98 struct mlx5e_xsk_param *xsk); 113 struct mlx5e_xsk_param *xsk); 116 struct mlx5e_xsk_param *xsk); 120 struct mlx5e_xsk_param *xsk); [all...] |
H A D | rx_res.c | 615 unsigned int ix, bool xsk) 619 if (xsk) 614 mlx5e_rx_res_xsk_update(struct mlx5e_rx_res *res, struct mlx5e_channels *chs, unsigned int ix, bool xsk) argument
|
H A D | rx_res.h | 42 unsigned int ix, bool xsk);
|
H A D | xdp.c | 40 int mlx5e_xdp_max_mtu(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 42 int hr = mlx5e_get_linear_rq_headroom(params, xsk);
|
H A D | xdp.h | 105 int mlx5e_xdp_max_mtu(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | pool.c | 23 static int mlx5e_xsk_get_pools(struct mlx5e_xsk *xsk) argument 25 if (!xsk->pools) { 26 xsk->pools = kcalloc(MLX5E_MAX_NUM_CHANNELS, 27 sizeof(*xsk->pools), GFP_KERNEL); 28 if (unlikely(!xsk->pools)) 32 xsk->refcnt++; 33 xsk->ever_used = true; 38 static void mlx5e_xsk_put_pools(struct mlx5e_xsk *xsk) argument 40 if (!--xsk->refcnt) { 41 kfree(xsk 46 mlx5e_xsk_add_pool(struct mlx5e_xsk *xsk, struct xsk_buff_pool *pool, u16 ix) argument 58 mlx5e_xsk_remove_pool(struct mlx5e_xsk *xsk, u16 ix) argument 71 mlx5e_build_xsk_param(struct xsk_buff_pool *pool, struct mlx5e_xsk_param *xsk) argument 82 struct mlx5e_xsk_param xsk; local [all...] |
H A D | pool.h | 10 struct mlx5e_xsk *xsk, u16 ix) 12 if (!xsk || !xsk->pools) 18 return xsk->pools[ix]; 22 void mlx5e_build_xsk_param(struct xsk_buff_pool *pool, struct mlx5e_xsk_param *xsk); 9 mlx5e_xsk_get_pool(struct mlx5e_params *params, struct mlx5e_xsk *xsk, u16 ix) argument
|
H A D | setup.c | 12 struct mlx5e_xsk_param *xsk) 14 if (!mlx5e_rx_is_linear_skb(mdev, params, xsk)) { 28 struct mlx5e_xsk_param *xsk, 32 * and xsk->chunk_size is limited to 65535 bytes. 34 if ((size_t)xsk->chunk_size > PAGE_SIZE || xsk->chunk_size < MLX5E_MIN_XSK_CHUNK_SIZE) { 35 mlx5_core_err(mdev, "XSK chunk size %u out of bounds [%u, %lu]\n", xsk->chunk_size, 45 return !mlx5e_mpwrq_validate_xsk(mdev, params, xsk); 47 return !mlx5e_legacy_rq_validate_xsk(mdev, params, xsk); 53 struct mlx5e_xsk_param *xsk, 10 mlx5e_legacy_rq_validate_xsk(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk) argument 27 mlx5e_validate_xsk_param(struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5_core_dev *mdev) argument 51 mlx5e_build_xsk_cparam(struct mlx5_core_dev *mdev, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct mlx5e_channel_param *cparam) argument 60 mlx5e_init_xsk_rq(struct mlx5e_channel *c, struct mlx5e_params *params, struct xsk_buff_pool *pool, struct mlx5e_xsk_param *xsk, struct mlx5e_rq *rq) argument 93 mlx5e_open_xsk_rq(struct mlx5e_channel *c, struct mlx5e_params *params, struct mlx5e_rq_param *rq_params, struct xsk_buff_pool *pool, struct mlx5e_xsk_param *xsk) argument 113 mlx5e_open_xsk(struct mlx5e_priv *priv, struct mlx5e_params *params, struct mlx5e_xsk_param *xsk, struct xsk_buff_pool *pool, struct mlx5e_channel *c) argument [all...] |
H A D | setup.h | 12 struct mlx5e_xsk_param *xsk, 15 struct mlx5e_xsk_param *xsk, struct xsk_buff_pool *pool,
|