Lines Matching defs:nn

57 	struct nfp_net *nn = r_vec->nfp_net;
64 tx_ring->qcidx = tx_ring->idx * nn->stride_tx;
66 tx_ring->qcp_q = nn->tx_bar + NFP_QCP_QUEUE_OFF(tx_ring->qcidx);
79 struct nfp_net *nn = r_vec->nfp_net;
85 rx_ring->fl_qcidx = rx_ring->idx * nn->stride_rx;
86 rx_ring->qcp_fl = nn->rx_bar + NFP_QCP_QUEUE_OFF(rx_ring->fl_qcidx);
183 int nfp_net_tx_rings_prepare(struct nfp_net *nn, struct nfp_net_dp *dp)
207 &nn->r_vecs[r - bias], r, bias);
339 int nfp_net_rx_rings_prepare(struct nfp_net *nn, struct nfp_net_dp *dp)
349 nfp_net_rx_ring_init(&dp->rx_rings[r], &nn->r_vecs[r], r);
383 nfp_net_rx_ring_hw_cfg_write(struct nfp_net *nn,
387 nn_writeq(nn, NFP_NET_CFG_RXR_ADDR(idx), rx_ring->dma);
388 nn_writeb(nn, NFP_NET_CFG_RXR_SZ(idx), ilog2(rx_ring->cnt));
389 nn_writeb(nn, NFP_NET_CFG_RXR_VEC(idx), rx_ring->r_vec->irq_entry);
393 nfp_net_tx_ring_hw_cfg_write(struct nfp_net *nn,
396 nn_writeq(nn, NFP_NET_CFG_TXR_ADDR(idx), tx_ring->dma);
399 nn_writeq(nn, NFP_NET_CFG_TXR_WB_ADDR(idx),
400 nn->dp.txrwb_dma + idx * sizeof(u64));
402 nn_writeb(nn, NFP_NET_CFG_TXR_SZ(idx), ilog2(tx_ring->cnt));
403 nn_writeb(nn, NFP_NET_CFG_TXR_VEC(idx), tx_ring->r_vec->irq_entry);
406 void nfp_net_vec_clear_ring_data(struct nfp_net *nn, unsigned int idx)
408 nn_writeq(nn, NFP_NET_CFG_RXR_ADDR(idx), 0);
409 nn_writeb(nn, NFP_NET_CFG_RXR_SZ(idx), 0);
410 nn_writeb(nn, NFP_NET_CFG_RXR_VEC(idx), 0);
412 nn_writeq(nn, NFP_NET_CFG_TXR_ADDR(idx), 0);
413 nn_writeq(nn, NFP_NET_CFG_TXR_WB_ADDR(idx), 0);
414 nn_writeb(nn, NFP_NET_CFG_TXR_SZ(idx), 0);
415 nn_writeb(nn, NFP_NET_CFG_TXR_VEC(idx), 0);
420 struct nfp_net *nn = netdev_priv(netdev);
422 return nn->dp.ops->xmit(skb, netdev);
425 bool __nfp_ctrl_tx(struct nfp_net *nn, struct sk_buff *skb)
427 struct nfp_net_r_vector *r_vec = &nn->r_vecs[0];
429 return nn->dp.ops->ctrl_tx_one(nn, r_vec, skb, false);
432 bool nfp_ctrl_tx(struct nfp_net *nn, struct sk_buff *skb)
434 struct nfp_net_r_vector *r_vec = &nn->r_vecs[0];
438 ret = nn->dp.ops->ctrl_tx_one(nn, r_vec, skb, false);