Lines Matching refs:mxbuf
56 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(xsk_buffs[i]);
62 mxbuf->rq = rq;
66 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(xsk_buffs[i]);
73 mxbuf->rq = rq;
79 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(xsk_buffs[i]);
98 mxbuf->rq = rq;
106 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(xsk_buffs[i]);
119 mxbuf->rq = rq;
254 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(wi->alloc_units.xsk_buffs[page_idx]);
270 /* mxbuf->rq is set on allocation, but cqe is per-packet so set it here */
271 mxbuf->cqe = cqe;
272 xsk_buff_set_size(&mxbuf->xdp, cqe_bcnt);
273 xsk_buff_dma_sync_for_cpu(&mxbuf->xdp);
274 net_prefetch(mxbuf->xdp.data);
292 if (likely(prog && mlx5e_xdp_handle(rq, prog, mxbuf))) {
301 return mlx5e_xsk_construct_skb(rq, &mxbuf->xdp);
309 struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(*wi->xskp);
319 /* mxbuf->rq is set on allocation, but cqe is per-packet so set it here */
320 mxbuf->cqe = cqe;
321 xsk_buff_set_size(&mxbuf->xdp, cqe_bcnt);
322 xsk_buff_dma_sync_for_cpu(&mxbuf->xdp);
323 net_prefetch(mxbuf->xdp.data);
326 if (likely(prog && mlx5e_xdp_handle(rq, prog, mxbuf))) {
336 return mlx5e_xsk_construct_skb(rq, &mxbuf->xdp);