Lines Matching refs:dp
631 ring->tx_max_pending = qc_max / nn->dp.ops->tx_min_desc_per_pkt;
632 ring->rx_pending = nn->dp.rxd_cnt;
633 ring->tx_pending = nn->dp.txd_cnt;
639 struct nfp_net_dp *dp;
641 dp = nfp_net_clone_dp(nn);
642 if (!dp)
645 dp->rxd_cnt = rxd_cnt;
646 dp->txd_cnt = txd_cnt;
648 return nfp_net_ring_reconfig(nn, dp, extack);
665 tx_dpp = nn->dp.ops->tx_min_desc_per_pkt;
680 if (nn->dp.rxd_cnt == rxd_cnt && nn->dp.txd_cnt == txd_cnt)
684 nn->dp.rxd_cnt, rxd_cnt, nn->dp.txd_cnt, txd_cnt);
1000 mem = nn->dp.ctrl_bar + nn->tlv_caps.vnic_stats_off;
1032 mem = nn->dp.ctrl_bar + nn->tlv_caps.vnic_stats_off;
1037 mem = nn->dp.ctrl_bar;
1120 data = nfp_vnic_get_hw_stats(data, nn->dp.ctrl_bar,
1444 cmd->data = nn->dp.num_rx_rings;
1502 writel(new_rss_cfg, nn->dp.ctrl_bar + NFP_NET_CFG_RSS_CTRL);
1630 fs->ring_cookie >= nn->dp.num_rx_rings)
1865 regs_buf[i] = readl(nn->dp.ctrl_bar + (i * sizeof(u32)));
2210 num_tx_rings = nn->dp.num_tx_rings;
2211 if (nn->dp.xdp_prog)
2212 num_tx_rings -= nn->dp.num_rx_rings;
2218 channel->combined_count = min(nn->dp.num_rx_rings, num_tx_rings);
2219 channel->rx_count = nn->dp.num_rx_rings - channel->combined_count;
2227 struct nfp_net_dp *dp;
2229 dp = nfp_net_clone_dp(nn);
2230 if (!dp)
2233 dp->num_rx_rings = total_rx;
2234 dp->num_tx_rings = total_tx;
2236 if (dp->xdp_prog)
2237 dp->num_tx_rings += total_rx;
2239 return nfp_net_ring_reconfig(nn, dp, NULL);