// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2015 Cavium, Inc. */ /* ETHTOOL Support for VNIC_VF Device*/ #include #include #include #include "nic_reg.h" #include "nic.h" #include "nicvf_queues.h" #include "q_struct.h" #include "thunder_bgx.h" #include "../common/cavium_ptp.h" #define DRV_NAME "nicvf" struct nicvf_stat { char name[ETH_GSTRING_LEN]; unsigned int index; }; #define NICVF_HW_STAT(stat) { \ .name = #stat, \ .index = offsetof(struct nicvf_hw_stats, stat) / sizeof(u64), \ } #define NICVF_DRV_STAT(stat) { \ .name = #stat, \ .index = offsetof(struct nicvf_drv_stats, stat) / sizeof(u64), \ } static const struct nicvf_stat nicvf_hw_stats[] = { NICVF_HW_STAT(rx_bytes), NICVF_HW_STAT(rx_frames), NICVF_HW_STAT(rx_ucast_frames), NICVF_HW_STAT(rx_bcast_frames), NICVF_HW_STAT(rx_mcast_frames), NICVF_HW_STAT(rx_drops), NICVF_HW_STAT(rx_drop_red), NICVF_HW_STAT(rx_drop_red_bytes), NICVF_HW_STAT(rx_drop_overrun), NICVF_HW_STAT(rx_drop_overrun_bytes), NICVF_HW_STAT(rx_drop_bcast), NICVF_HW_STAT(rx_drop_mcast), NICVF_HW_STAT(rx_drop_l3_bcast), NICVF_HW_STAT(rx_drop_l3_mcast), NICVF_HW_STAT(rx_fcs_errors), NICVF_HW_STAT(rx_l2_errors), NICVF_HW_STAT(tx_bytes), NICVF_HW_STAT(tx_frames), NICVF_HW_STAT(tx_ucast_frames), NICVF_HW_STAT(tx_bcast_frames), NICVF_HW_STAT(tx_mcast_frames), NICVF_HW_STAT(tx_drops), }; static const struct nicvf_stat nicvf_drv_stats[] = { NICVF_DRV_STAT(rx_bgx_truncated_pkts), NICVF_DRV_STAT(rx_jabber_errs), NICVF_DRV_STAT(rx_fcs_errs), NICVF_DRV_STAT(rx_bgx_errs), NICVF_DRV_STAT(rx_prel2_errs), NICVF_DRV_STAT(rx_l2_hdr_malformed), NICVF_DRV_STAT(rx_oversize), NICVF_DRV_STAT(rx_undersize), NICVF_DRV_STAT(rx_l2_len_mismatch), NICVF_DRV_STAT(rx_l2_pclp), NICVF_DRV_STAT(rx_ip_ver_errs), NICVF_DRV_STAT(rx_ip_csum_errs), NICVF_DRV_STAT(rx_ip_hdr_malformed), NICVF_DRV_STAT(rx_ip_payload_malformed), NICVF_DRV_STAT(rx_ip_ttl_errs), NICVF_DRV_STAT(rx_l3_pclp), NICVF_DRV_STAT(rx_l4_malformed), NICVF_DRV_STAT(rx_l4_csum_errs), NICVF_DRV_STAT(rx_udp_len_errs), NICVF_DRV_STAT(rx_l4_port_errs), NICVF_DRV_STAT(rx_tcp_flag_errs), NICVF_DRV_STAT(rx_tcp_offset_errs), NICVF_DRV_STAT(rx_l4_pclp), NICVF_DRV_STAT(rx_truncated_pkts), NICVF_DRV_STAT(tx_desc_fault), NICVF_DRV_STAT(tx_hdr_cons_err), NICVF_DRV_STAT(tx_subdesc_err), NICVF_DRV_STAT(tx_max_size_exceeded), NICVF_DRV_STAT(tx_imm_size_oflow), NICVF_DRV_STAT(tx_data_seq_err), NICVF_DRV_STAT(tx_mem_seq_err), NICVF_DRV_STAT(tx_lock_viol), NICVF_DRV_STAT(tx_data_fault), NICVF_DRV_STAT(tx_tstmp_conflict), NICVF_DRV_STAT(tx_tstmp_timeout), NICVF_DRV_STAT(tx_mem_fault), NICVF_DRV_STAT(tx_csum_overlap), NICVF_DRV_STAT(tx_csum_overflow), NICVF_DRV_STAT(tx_tso), NICVF_DRV_STAT(tx_timeout), NICVF_DRV_STAT(txq_stop), NICVF_DRV_STAT(txq_wake), NICVF_DRV_STAT(rcv_buffer_alloc_failures), NICVF_DRV_STAT(page_alloc), }; static const struct nicvf_stat nicvf_queue_stats[] = { { "bytes", 0 }, { "frames", 1 }, }; static const unsigned int nicvf_n_hw_stats = ARRAY_SIZE(nicvf_hw_stats); static const unsigned int nicvf_n_drv_stats = ARRAY_SIZE(nicvf_drv_stats); static const unsigned int nicvf_n_queue_stats = ARRAY_SIZE(nicvf_queue_stats); static int nicvf_get_link_ksettings(struct net_device *netdev, struct ethtool_link_ksettings *cmd) { struct nicvf *nic = netdev_priv(netdev); u32 supported, advertising; supported = 0; advertising = 0; if (!nic->link_up) { cmd->base.duplex = DUPLEX_UNKNOWN; cmd->base.speed = SPEED_UNKNOWN; return 0; } switch (nic->speed) { case SPEED_1000: cmd->base.port = PORT_MII | PORT_TP; cmd->base.autoneg = AUTONEG_ENABLE; supported |= SUPPORTED_MII | SUPPORTED_TP; supported |= SUPPORTED_1000baseT_Full | SUPPORTED_1000baseT_Half | SUPPORTED_100baseT_Full | SUPPORTED_100baseT_Half | SUPPORTED_10baseT_Full | SUPPORTED_10baseT_Half; supported |= SUPPORTED_Autoneg; advertising |= ADVERTISED_1000baseT_Full | ADVERTISED_1000baseT_Half | ADVERTISED_100baseT_Full | ADVERTISED_100baseT_Half | ADVERTISED_10baseT_Full | ADVERTISED_10baseT_Half; break; case SPEED_10000: if (nic->mac_type == BGX_MODE_RXAUI) { cmd->base.port = PORT_TP; supported |= SUPPORTED_TP; } else { cmd->base.port = PORT_FIBRE; supported |= SUPPORTED_FIBRE; } cmd->base.autoneg = AUTONEG_DISABLE; supported |= SUPPORTED_10000baseT_Full; break; case SPEED_40000: cmd->base.port = PORT_FIBRE; cmd->base.autoneg = AUTONEG_DISABLE; supported |= SUPPORTED_FIBRE; supported |= SUPPORTED_40000baseCR4_Full; break; } cmd->base.duplex = nic->duplex; cmd->base.speed = nic->speed; ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, supported); ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, advertising); return 0; } static u32 nicvf_get_link(struct net_device *netdev) { struct nicvf *nic = netdev_priv(netdev); return nic->link_up; } static void nicvf_get_drvinfo(struct net_device *netdev, struct ethtool_drvinfo *info) { struct nicvf *nic = netdev_priv(netdev); strscpy(info->driver, DRV_NAME, sizeof(info->driver)); strscpy(info->bus_info, pci_name(nic->pdev), sizeof(info->bus_info)); } static u32 nicvf_get_msglevel(struct net_device *netdev) { struct nicvf *nic = netdev_priv(netdev); return nic->msg_enable; } static void nicvf_set_msglevel(struct net_device *netdev, u32 lvl) { struct nicvf *nic = netdev_priv(netdev); nic->msg_enable = lvl; } static void nicvf_get_qset_strings(struct nicvf *nic, u8 **data, int qset) { int stats, qidx; int start_qidx = qset * MAX_RCV_QUEUES_PER_QS; for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { for (stats = 0; stats < nicvf_n_queue_stats; stats++) { sprintf(*data, "rxq%d: %s", qidx + start_qidx, nicvf_queue_stats[stats].name); *data += ETH_GSTRING_LEN; } } for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { for (stats = 0; stats < nicvf_n_queue_stats; stats++) { sprintf(*data, "txq%d: %s", qidx + start_qidx, nicvf_queue_stats[stats].name); *data += ETH_GSTRING_LEN; } } } static void nicvf_get_strings(struct net_device *netdev, u32 sset, u8 *data) { struct nicvf *nic = netdev_priv(netdev); int stats; int sqs; if (sset != ETH_SS_STATS) return; for (stats = 0; stats < nicvf_n_hw_stats; stats++) { memcpy(data, nicvf_hw_stats[stats].name, ETH_GSTRING_LEN); data += ETH_GSTRING_LEN; } for (stats = 0; stats < nicvf_n_drv_stats; stats++) { memcpy(data, nicvf_drv_stats[stats].name, ETH_GSTRING_LEN); data += ETH_GSTRING_LEN; } nicvf_get_qset_strings(nic, &data, 0); for (sqs = 0; sqs < nic->sqs_count; sqs++) { if (!nic->snicvf[sqs]) continue; nicvf_get_qset_strings(nic->snicvf[sqs], &data, sqs + 1); } for (stats = 0; stats < BGX_RX_STATS_COUNT; stats++) { sprintf(data, "bgx_rxstat%d: ", stats); data += ETH_GSTRING_LEN; } for (stats = 0; stats < BGX_TX_STATS_COUNT; stats++) { sprintf(data, "bgx_txstat%d: ", stats); data += ETH_GSTRING_LEN; } } static int nicvf_get_sset_count(struct net_device *netdev, int sset) { struct nicvf *nic = netdev_priv(netdev); int qstats_count; int sqs; if (sset != ETH_SS_STATS) return -EINVAL; qstats_count = nicvf_n_queue_stats * (nic->qs->rq_cnt + nic->qs->sq_cnt); for (sqs = 0; sqs < nic->sqs_count; sqs++) { struct nicvf *snic; snic = nic->snicvf[sqs]; if (!snic) continue; qstats_count += nicvf_n_queue_stats * (snic->qs->rq_cnt + snic->qs->sq_cnt); } return nicvf_n_hw_stats + nicvf_n_drv_stats + qstats_count + BGX_RX_STATS_COUNT + BGX_TX_STATS_COUNT; } static void nicvf_get_qset_stats(struct nicvf *nic, struct ethtool_stats *stats, u64 **data) { int stat, qidx; if (!nic) return; for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { nicvf_update_rq_stats(nic, qidx); for (stat = 0; stat < nicvf_n_queue_stats; stat++) *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) [nicvf_queue_stats[stat].index]; } for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { nicvf_update_sq_stats(nic, qidx); for (stat = 0; stat < nicvf_n_queue_stats; stat++) *((*data)++) = ((u64 *)&nic->qs->sq[qidx].stats) [nicvf_queue_stats[stat].index]; } } static void nicvf_get_ethtool_stats(struct net_device *netdev, struct ethtool_stats *stats, u64 *data) { struct nicvf *nic = netdev_priv(netdev); int stat, tmp_stats; int sqs, cpu; nicvf_update_stats(nic); /* Update LMAC stats */ nicvf_update_lmac_stats(nic); for (stat = 0; stat < nicvf_n_hw_stats; stat++) *(data++) = ((u64 *)&nic->hw_stats) [nicvf_hw_stats[stat].index]; for (stat = 0; stat < nicvf_n_drv_stats; stat++) { tmp_stats = 0; for_each_possible_cpu(cpu) tmp_stats += ((u64 *)per_cpu_ptr(nic->drv_stats, cpu)) [nicvf_drv_stats[stat].index]; *(data++) = tmp_stats; } nicvf_get_qset_stats(nic, stats, &data); for (sqs = 0; sqs < nic->sqs_count; sqs++) { if (!nic->snicvf[sqs]) continue; nicvf_get_qset_stats(nic->snicvf[sqs], stats, &data); } for (stat = 0; stat < BGX_RX_STATS_COUNT; stat++) *(data++) = nic->bgx_stats.rx_stats[stat]; for (stat = 0; stat < BGX_TX_STATS_COUNT; stat++) *(data++) = nic->bgx_stats.tx_stats[stat]; } static int nicvf_get_regs_len(struct net_device *dev) { return sizeof(u64) * NIC_VF_REG_COUNT; } static void nicvf_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *reg) { struct nicvf *nic = netdev_priv(dev); u64 *p = (u64 *)reg; u64 reg_offset; int mbox, key, stat, q; int i = 0; regs->version = 0; memset(p, 0, NIC_VF_REG_COUNT); p[i++] = nicvf_reg_read(nic, NIC_VNIC_CFG); /* Mailbox registers */ for (mbox = 0; mbox < NIC_PF_VF_MAILBOX_SIZE; mbox++) p[i++] = nicvf_reg_read(nic, NIC_VF_PF_MAILBOX_0_1 | (mbox << 3)); p[i++] = nicvf_reg_read(nic, NIC_VF_INT); p[i++] = nicvf_reg_read(nic, NIC_VF_INT_W1S); p[i++] = nicvf_reg_read(nic, NIC_VF_ENA_W1C); p[i++] = nicvf_reg_read(nic, NIC_VF_ENA_W1S); p[i++] = nicvf_reg_read(nic, NIC_VNIC_RSS_CFG); for (key = 0; key < RSS_HASH_KEY_SIZE; key++) p[i++] = nicvf_reg_read(nic, NIC_VNIC_RSS_KEY_0_4 | (key << 3)); /* Tx/Rx statistics */ for (stat = 0; stat < TX_STATS_ENUM_LAST; stat++) p[i++] = nicvf_reg_read(nic, NIC_VNIC_TX_STAT_0_4 | (stat << 3)); for (i = 0; i < RX_STATS_ENUM_LAST; i++) p[i++] = nicvf_reg_read(nic, NIC_VNIC_RX_STAT_0_13 | (stat << 3)); p[i++] = nicvf_reg_read(nic, NIC_QSET_RQ_GEN_CFG); /* All completion queue's registers */ for (q = 0; q < MAX_CMP_QUEUES_PER_QS; q++) { p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_CFG, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_CFG2, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_THRESH, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_BASE, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_TAIL, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_DOOR, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_STATUS, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_STATUS2, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_DEBUG, q); } /* All receive queue's registers */ for (q = 0; q < MAX_RCV_QUEUES_PER_QS; q++) { p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RQ_0_7_CFG, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RQ_0_7_STAT_0_1, q); reg_offset = NIC_QSET_RQ_0_7_STAT_0_1 | (1 << 3); p[i++] = nicvf_queue_reg_read(nic, reg_offset, q); } for (q = 0; q < MAX_SND_QUEUES_PER_QS; q++) { p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_CFG, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_THRESH, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_BASE, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_HEAD, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_TAIL, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_DOOR, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_STATUS, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_DEBUG, q); /* Padding, was NIC_QSET_SQ_0_7_CNM_CHG, which * produces bus errors when read */ p[i++] = 0; p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_SQ_0_7_STAT_0_1, q); reg_offset = NIC_QSET_SQ_0_7_STAT_0_1 | (1 << 3); p[i++] = nicvf_queue_reg_read(nic, reg_offset, q); } for (q = 0; q < MAX_RCV_BUF_DESC_RINGS_PER_QS; q++) { p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_CFG, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_THRESH, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_BASE, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_HEAD, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_TAIL, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_DOOR, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS0, q); p[i++] = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS1, q); reg_offset = NIC_QSET_RBDR_0_1_PREFETCH_STATUS; p[i++] = nicvf_queue_reg_read(nic, reg_offset, q); } } static int nicvf_get_coalesce(struct net_device *netdev, struct ethtool_coalesce *cmd, struct kernel_ethtool_coalesce *kernel_coal, struct netlink_ext_ack *extack) { struct nicvf *nic = netdev_priv(netdev); cmd->rx_coalesce_usecs = nic->cq_coalesce_usecs; return 0; } static void nicvf_get_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) { struct nicvf *nic = netdev_priv(netdev); struct queue_set *qs = nic->qs; ring->rx_max_pending = MAX_CMP_QUEUE_LEN; ring->rx_pending = qs->cq_len; ring->tx_max_pending = MAX_SND_QUEUE_LEN; ring->tx_pending = qs->sq_len; } static int nicvf_set_ringparam(struct net_device *netdev, struct ethtool_ringparam *ring, struct kernel_ethtool_ringparam *kernel_ring, struct netlink_ext_ack *extack) { struct nicvf *nic = netdev_priv(netdev); struct queue_set *qs = nic->qs; u32 rx_count, tx_count; /* Due to HW errata this is not supported on T88 pass 1.x silicon */ if (pass1_silicon(nic->pdev)) return -EINVAL; if ((ring->rx_mini_pending) || (ring->rx_jumbo_pending)) return -EINVAL; tx_count = clamp_t(u32, ring->tx_pending, MIN_SND_QUEUE_LEN, MAX_SND_QUEUE_LEN); rx_count = clamp_t(u32, ring->rx_pending, MIN_CMP_QUEUE_LEN, MAX_CMP_QUEUE_LEN); if ((tx_count == qs->sq_len) && (rx_count == qs->cq_len)) return 0; /* Permitted lengths are 1K, 2K, 4K, 8K, 16K, 32K, 64K */ qs->sq_len = rounddown_pow_of_two(tx_count); qs->cq_len = rounddown_pow_of_two(rx_count); if (netif_running(netdev)) { nicvf_stop(netdev); nicvf_open(netdev); } return 0; } static int nicvf_get_rss_hash_opts(struct nicvf *nic, struct ethtool_rxnfc *info) { info->data = 0; switch (info->flow_type) { case TCP_V4_FLOW: case TCP_V6_FLOW: case UDP_V4_FLOW: case UDP_V6_FLOW: case SCTP_V4_FLOW: case SCTP_V6_FLOW: info->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; fallthrough; case IPV4_FLOW: case IPV6_FLOW: info->data |= RXH_IP_SRC | RXH_IP_DST; break; default: return -EINVAL; } return 0; } static int nicvf_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *info, u32 *rules) { struct nicvf *nic = netdev_priv(dev); int ret = -EOPNOTSUPP; switch (info->cmd) { case ETHTOOL_GRXRINGS: info->data = nic->rx_queues; ret = 0; break; case ETHTOOL_GRXFH: return nicvf_get_rss_hash_opts(nic, info); default: break; } return ret; } static int nicvf_set_rss_hash_opts(struct nicvf *nic, struct ethtool_rxnfc *info) { struct nicvf_rss_info *rss = &nic->rss_info; u64 rss_cfg = nicvf_reg_read(nic, NIC_VNIC_RSS_CFG); if (!rss->enable) netdev_err(nic->netdev, "RSS is disabled, hash cannot be set\n"); netdev_info(nic->netdev, "Set RSS flow type = %d, data = %lld\n", info->flow_type, info->data); if (!(info->data & RXH_IP_SRC) || !(info->data & RXH_IP_DST)) return -EINVAL; switch (info->flow_type) { case TCP_V4_FLOW: case TCP_V6_FLOW: switch (info->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) { case 0: rss_cfg &= ~(1ULL << RSS_HASH_TCP); break; case (RXH_L4_B_0_1 | RXH_L4_B_2_3): rss_cfg |= (1ULL << RSS_HASH_TCP); break; default: return -EINVAL; } break; case UDP_V4_FLOW: case UDP_V6_FLOW: switch (info->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) { case 0: rss_cfg &= ~(1ULL << RSS_HASH_UDP); break; case (RXH_L4_B_0_1 | RXH_L4_B_2_3): rss_cfg |= (1ULL << RSS_HASH_UDP); break; default: return -EINVAL; } break; case SCTP_V4_FLOW: case SCTP_V6_FLOW: switch (info->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) { case 0: rss_cfg &= ~(1ULL << RSS_HASH_L4ETC); break; case (RXH_L4_B_0_1 | RXH_L4_B_2_3): rss_cfg |= (1ULL << RSS_HASH_L4ETC); break; default: return -EINVAL; } break; case IPV4_FLOW: case IPV6_FLOW: rss_cfg = RSS_HASH_IP; break; default: return -EINVAL; } nicvf_reg_write(nic, NIC_VNIC_RSS_CFG, rss_cfg); return 0; } static int nicvf_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *info) { struct nicvf *nic = netdev_priv(dev); switch (info->cmd) { case ETHTOOL_SRXFH: return nicvf_set_rss_hash_opts(nic, info); default: break; } return -EOPNOTSUPP; } static u32 nicvf_get_rxfh_key_size(struct net_device *netdev) { return RSS_HASH_KEY_SIZE * sizeof(u64); } static u32 nicvf_get_rxfh_indir_size(struct net_device *dev) { struct nicvf *nic = netdev_priv(dev); return nic->rss_info.rss_size; } static int nicvf_get_rxfh(struct net_device *dev, struct ethtool_rxfh_param *rxfh) { struct nicvf *nic = netdev_priv(dev); struct nicvf_rss_info *rss = &nic->rss_info; int idx; if (rxfh->indir) { for (idx = 0; idx < rss->rss_size; idx++) rxfh->indir[idx] = rss->ind_tbl[idx]; } if (rxfh->key) memcpy(rxfh->key, rss->key, RSS_HASH_KEY_SIZE * sizeof(u64)); rxfh->hfunc = ETH_RSS_HASH_TOP; return 0; } static int nicvf_set_rxfh(struct net_device *dev, struct ethtool_rxfh_param *rxfh, struct netlink_ext_ack *extack) { struct nicvf *nic = netdev_priv(dev); struct nicvf_rss_info *rss = &nic->rss_info; int idx; if (rxfh->hfunc != ETH_RSS_HASH_NO_CHANGE && rxfh->hfunc != ETH_RSS_HASH_TOP) return -EOPNOTSUPP; if (!rss->enable) { netdev_err(nic->netdev, "RSS is disabled, cannot change settings\n"); return -EIO; } if (rxfh->indir) { for (idx = 0; idx < rss->rss_size; idx++) rss->ind_tbl[idx] = rxfh->indir[idx]; } if (rxfh->key) { memcpy(rss->key, rxfh->key, RSS_HASH_KEY_SIZE * sizeof(u64)); nicvf_set_rss_key(nic); } nicvf_config_rss(nic); return 0; } /* Get no of queues device supports and current queue count */ static void nicvf_get_channels(struct net_device *dev, struct ethtool_channels *channel) { struct nicvf *nic = netdev_priv(dev); memset(channel, 0, sizeof(*channel)); channel->max_rx = nic->max_queues; channel->max_tx = nic->max_queues; channel->rx_count = nic->rx_queues; channel->tx_count = nic->tx_queues; } /* Set no of Tx, Rx queues to be used */ static int nicvf_set_channels(struct net_device *dev, struct ethtool_channels *channel) { struct nicvf *nic = netdev_priv(dev); int err = 0; bool if_up = netif_running(dev); u8 cqcount, txq_count; if (!channel->rx_count || !channel->tx_count) return -EINVAL; if (channel->rx_count > nic->max_queues) return -EINVAL; if (channel->tx_count > nic->max_queues) return -EINVAL; if (channel->tx_count + channel->rx_count > nic->max_queues) { if (nic->xdp_prog) { netdev_err(nic->netdev, "XDP mode, RXQs + TXQs > Max %d\n", nic->max_queues); return -EINVAL; } xdp_clear_features_flag(nic->netdev); } else if (!pass1_silicon(nic->pdev)) { xdp_set_features_flag(dev, NETDEV_XDP_ACT_BASIC); } if (if_up) nicvf_stop(dev); nic->rx_queues = channel->rx_count; nic->tx_queues = channel->tx_count; if (!nic->xdp_prog) nic->xdp_tx_queues = 0; else nic->xdp_tx_queues = channel->rx_count; txq_count = nic->xdp_tx_queues + nic->tx_queues; cqcount = max(nic->rx_queues, txq_count); if (cqcount > MAX_CMP_QUEUES_PER_QS) { nic->sqs_count = roundup(cqcount, MAX_CMP_QUEUES_PER_QS); nic->sqs_count = (nic->sqs_count / MAX_CMP_QUEUES_PER_QS) - 1; } else { nic->sqs_count = 0; } nic->qs->rq_cnt = min_t(u8, nic->rx_queues, MAX_RCV_QUEUES_PER_QS); nic->qs->sq_cnt = min_t(u8, txq_count, MAX_SND_QUEUES_PER_QS); nic->qs->cq_cnt = max(nic->qs->rq_cnt, nic->qs->sq_cnt); err = nicvf_set_real_num_queues(dev, nic->tx_queues, nic->rx_queues); if (err) return err; if (if_up) nicvf_open(dev); netdev_info(dev, "Setting num Tx rings to %d, Rx rings to %d success\n", nic->tx_queues, nic->rx_queues); return err; } static void nicvf_get_pauseparam(struct net_device *dev, struct ethtool_pauseparam *pause) { struct nicvf *nic = netdev_priv(dev); union nic_mbx mbx = {}; /* Supported only for 10G/40G interfaces */ if ((nic->mac_type == BGX_MODE_SGMII) || (nic->mac_type == BGX_MODE_QSGMII) || (nic->mac_type == BGX_MODE_RGMII)) return; mbx.pfc.msg = NIC_MBOX_MSG_PFC; mbx.pfc.get = 1; if (!nicvf_send_msg_to_pf(nic, &mbx)) { pause->autoneg = nic->pfc.autoneg; pause->rx_pause = nic->pfc.fc_rx; pause->tx_pause = nic->pfc.fc_tx; } } static int nicvf_set_pauseparam(struct net_device *dev, struct ethtool_pauseparam *pause) { struct nicvf *nic = netdev_priv(dev); union nic_mbx mbx = {}; /* Supported only for 10G/40G interfaces */ if ((nic->mac_type == BGX_MODE_SGMII) || (nic->mac_type == BGX_MODE_QSGMII) || (nic->mac_type == BGX_MODE_RGMII)) return -EOPNOTSUPP; if (pause->autoneg) return -EOPNOTSUPP; mbx.pfc.msg = NIC_MBOX_MSG_PFC; mbx.pfc.get = 0; mbx.pfc.fc_rx = pause->rx_pause; mbx.pfc.fc_tx = pause->tx_pause; if (nicvf_send_msg_to_pf(nic, &mbx)) return -EAGAIN; nic->pfc.fc_rx = pause->rx_pause; nic->pfc.fc_tx = pause->tx_pause; return 0; } static int nicvf_get_ts_info(struct net_device *netdev, struct ethtool_ts_info *info) { struct nicvf *nic = netdev_priv(netdev); if (!nic->ptp_clock) return ethtool_op_get_ts_info(netdev, info); info->so_timestamping = SOF_TIMESTAMPING_TX_SOFTWARE | SOF_TIMESTAMPING_RX_SOFTWARE | SOF_TIMESTAMPING_SOFTWARE | SOF_TIMESTAMPING_TX_HARDWARE | SOF_TIMESTAMPING_RX_HARDWARE | SOF_TIMESTAMPING_RAW_HARDWARE; info->phc_index = cavium_ptp_clock_index(nic->ptp_clock); info->tx_types = (1 << HWTSTAMP_TX_OFF) | (1 << HWTSTAMP_TX_ON); info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) | (1 << HWTSTAMP_FILTER_ALL); return 0; } static const struct ethtool_ops nicvf_ethtool_ops = { .get_link = nicvf_get_link, .get_drvinfo = nicvf_get_drvinfo, .get_msglevel = nicvf_get_msglevel, .set_msglevel = nicvf_set_msglevel, .get_strings = nicvf_get_strings, .get_sset_count = nicvf_get_sset_count, .get_ethtool_stats = nicvf_get_ethtool_stats, .get_regs_len = nicvf_get_regs_len, .get_regs = nicvf_get_regs, .get_coalesce = nicvf_get_coalesce, .get_ringparam = nicvf_get_ringparam, .set_ringparam = nicvf_set_ringparam, .get_rxnfc = nicvf_get_rxnfc, .set_rxnfc = nicvf_set_rxnfc, .get_rxfh_key_size = nicvf_get_rxfh_key_size, .get_rxfh_indir_size = nicvf_get_rxfh_indir_size, .get_rxfh = nicvf_get_rxfh, .set_rxfh = nicvf_set_rxfh, .get_channels = nicvf_get_channels, .set_channels = nicvf_set_channels, .get_pauseparam = nicvf_get_pauseparam, .set_pauseparam = nicvf_set_pauseparam, .get_ts_info = nicvf_get_ts_info, .get_link_ksettings = nicvf_get_link_ksettings, }; void nicvf_set_ethtool_ops(struct net_device *netdev) { netdev->ethtool_ops = &nicvf_ethtool_ops; }