// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) /* * Copyright(c) 2020 Intel Corporation. * */ /* * This file contains HFI1 support for netdev RX functionality */ #include "sdma.h" #include "verbs.h" #include "netdev.h" #include "hfi.h" #include #include #include static int hfi1_netdev_setup_ctxt(struct hfi1_netdev_rx *rx, struct hfi1_ctxtdata *uctxt) { unsigned int rcvctrl_ops; struct hfi1_devdata *dd = rx->dd; int ret; uctxt->rhf_rcv_function_map = netdev_rhf_rcv_functions; uctxt->do_interrupt = &handle_receive_interrupt_napi_sp; /* Now allocate the RcvHdr queue and eager buffers. */ ret = hfi1_create_rcvhdrq(dd, uctxt); if (ret) goto done; ret = hfi1_setup_eagerbufs(uctxt); if (ret) goto done; clear_rcvhdrtail(uctxt); rcvctrl_ops = HFI1_RCVCTRL_CTXT_DIS; rcvctrl_ops |= HFI1_RCVCTRL_INTRAVAIL_DIS; if (!HFI1_CAP_KGET_MASK(uctxt->flags, MULTI_PKT_EGR)) rcvctrl_ops |= HFI1_RCVCTRL_ONE_PKT_EGR_ENB; if (HFI1_CAP_KGET_MASK(uctxt->flags, NODROP_EGR_FULL)) rcvctrl_ops |= HFI1_RCVCTRL_NO_EGR_DROP_ENB; if (HFI1_CAP_KGET_MASK(uctxt->flags, NODROP_RHQ_FULL)) rcvctrl_ops |= HFI1_RCVCTRL_NO_RHQ_DROP_ENB; if (HFI1_CAP_KGET_MASK(uctxt->flags, DMA_RTAIL)) rcvctrl_ops |= HFI1_RCVCTRL_TAILUPD_ENB; hfi1_rcvctrl(uctxt->dd, rcvctrl_ops, uctxt); done: return ret; } static int hfi1_netdev_allocate_ctxt(struct hfi1_devdata *dd, struct hfi1_ctxtdata **ctxt) { struct hfi1_ctxtdata *uctxt; int ret; if (dd->flags & HFI1_FROZEN) return -EIO; ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt); if (ret < 0) { dd_dev_err(dd, "Unable to create ctxtdata, failing open\n"); return -ENOMEM; } uctxt->flags = HFI1_CAP_KGET(MULTI_PKT_EGR) | HFI1_CAP_KGET(NODROP_RHQ_FULL) | HFI1_CAP_KGET(NODROP_EGR_FULL) | HFI1_CAP_KGET(DMA_RTAIL); /* Netdev contexts are always NO_RDMA_RTAIL */ uctxt->fast_handler = handle_receive_interrupt_napi_fp; uctxt->slow_handler = handle_receive_interrupt_napi_sp; hfi1_set_seq_cnt(uctxt, 1); uctxt->is_vnic = true; hfi1_stats.sps_ctxts++; dd_dev_info(dd, "created netdev context %d\n", uctxt->ctxt); *ctxt = uctxt; return 0; } static void hfi1_netdev_deallocate_ctxt(struct hfi1_devdata *dd, struct hfi1_ctxtdata *uctxt) { flush_wc(); /* * Disable receive context and interrupt available, reset all * RcvCtxtCtrl bits to default values. */ hfi1_rcvctrl(dd, HFI1_RCVCTRL_CTXT_DIS | HFI1_RCVCTRL_TIDFLOW_DIS | HFI1_RCVCTRL_INTRAVAIL_DIS | HFI1_RCVCTRL_ONE_PKT_EGR_DIS | HFI1_RCVCTRL_NO_RHQ_DROP_DIS | HFI1_RCVCTRL_NO_EGR_DROP_DIS, uctxt); if (uctxt->msix_intr != CCE_NUM_MSIX_VECTORS) msix_free_irq(dd, uctxt->msix_intr); uctxt->msix_intr = CCE_NUM_MSIX_VECTORS; uctxt->event_flags = 0; hfi1_clear_tids(uctxt); hfi1_clear_ctxt_pkey(dd, uctxt); hfi1_stats.sps_ctxts--; hfi1_free_ctxt(uctxt); } static int hfi1_netdev_allot_ctxt(struct hfi1_netdev_rx *rx, struct hfi1_ctxtdata **ctxt) { int rc; struct hfi1_devdata *dd = rx->dd; rc = hfi1_netdev_allocate_ctxt(dd, ctxt); if (rc) { dd_dev_err(dd, "netdev ctxt alloc failed %d\n", rc); return rc; } rc = hfi1_netdev_setup_ctxt(rx, *ctxt); if (rc) { dd_dev_err(dd, "netdev ctxt setup failed %d\n", rc); hfi1_netdev_deallocate_ctxt(dd, *ctxt); *ctxt = NULL; } return rc; } /** * hfi1_num_netdev_contexts - Count of netdev recv contexts to use. * @dd: device on which to allocate netdev contexts * @available_contexts: count of available receive contexts * @cpu_mask: mask of possible cpus to include for contexts * * Return: count of physical cores on a node or the remaining available recv * contexts for netdev recv context usage up to the maximum of * HFI1_MAX_NETDEV_CTXTS. * A value of 0 can be returned when acceleration is explicitly turned off, * a memory allocation error occurs or when there are no available contexts. * */ u32 hfi1_num_netdev_contexts(struct hfi1_devdata *dd, u32 available_contexts, struct cpumask *cpu_mask) { cpumask_var_t node_cpu_mask; unsigned int available_cpus; if (!HFI1_CAP_IS_KSET(AIP)) return 0; /* Always give user contexts priority over netdev contexts */ if (available_contexts == 0) { dd_dev_info(dd, "No receive contexts available for netdevs.\n"); return 0; } if (!zalloc_cpumask_var(&node_cpu_mask, GFP_KERNEL)) { dd_dev_err(dd, "Unable to allocate cpu_mask for netdevs.\n"); return 0; } cpumask_and(node_cpu_mask, cpu_mask, cpumask_of_node(dd->node)); available_cpus = cpumask_weight(node_cpu_mask); free_cpumask_var(node_cpu_mask); return min3(available_cpus, available_contexts, (u32)HFI1_MAX_NETDEV_CTXTS); } static int hfi1_netdev_rxq_init(struct hfi1_netdev_rx *rx) { int i; int rc; struct hfi1_devdata *dd = rx->dd; struct net_device *dev = rx->rx_napi; rx->num_rx_q = dd->num_netdev_contexts; rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), GFP_KERNEL, dd->node); if (!rx->rxq) { dd_dev_err(dd, "Unable to allocate netdev queue data\n"); return (-ENOMEM); } for (i = 0; i < rx->num_rx_q; i++) { struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); if (rc) goto bail_context_irq_failure; hfi1_rcd_get(rxq->rcd); rxq->rx = rx; rxq->rcd->napi = &rxq->napi; dd_dev_info(dd, "Setting rcv queue %d napi to context %d\n", i, rxq->rcd->ctxt); /* * Disable BUSY_POLL on this NAPI as this is not supported * right now. */ set_bit(NAPI_STATE_NO_BUSY_POLL, &rxq->napi.state); netif_napi_add(dev, &rxq->napi, hfi1_netdev_rx_napi); rc = msix_netdev_request_rcd_irq(rxq->rcd); if (rc) goto bail_context_irq_failure; } return 0; bail_context_irq_failure: dd_dev_err(dd, "Unable to allot receive context\n"); for (; i >= 0; i--) { struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; if (rxq->rcd) { hfi1_netdev_deallocate_ctxt(dd, rxq->rcd); hfi1_rcd_put(rxq->rcd); rxq->rcd = NULL; } } kfree(rx->rxq); rx->rxq = NULL; return rc; } static void hfi1_netdev_rxq_deinit(struct hfi1_netdev_rx *rx) { int i; struct hfi1_devdata *dd = rx->dd; for (i = 0; i < rx->num_rx_q; i++) { struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; netif_napi_del(&rxq->napi); hfi1_netdev_deallocate_ctxt(dd, rxq->rcd); hfi1_rcd_put(rxq->rcd); rxq->rcd = NULL; } kfree(rx->rxq); rx->rxq = NULL; rx->num_rx_q = 0; } static void enable_queues(struct hfi1_netdev_rx *rx) { int i; for (i = 0; i < rx->num_rx_q; i++) { struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; dd_dev_info(rx->dd, "enabling queue %d on context %d\n", i, rxq->rcd->ctxt); napi_enable(&rxq->napi); hfi1_rcvctrl(rx->dd, HFI1_RCVCTRL_CTXT_ENB | HFI1_RCVCTRL_INTRAVAIL_ENB, rxq->rcd); } } static void disable_queues(struct hfi1_netdev_rx *rx) { int i; msix_netdev_synchronize_irq(rx->dd); for (i = 0; i < rx->num_rx_q; i++) { struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; dd_dev_info(rx->dd, "disabling queue %d on context %d\n", i, rxq->rcd->ctxt); /* wait for napi if it was scheduled */ hfi1_rcvctrl(rx->dd, HFI1_RCVCTRL_CTXT_DIS | HFI1_RCVCTRL_INTRAVAIL_DIS, rxq->rcd); napi_synchronize(&rxq->napi); napi_disable(&rxq->napi); } } /** * hfi1_netdev_rx_init - Incrememnts netdevs counter. When called first time, * it allocates receive queue data and calls netif_napi_add * for each queue. * * @dd: hfi1 dev data */ int hfi1_netdev_rx_init(struct hfi1_devdata *dd) { struct hfi1_netdev_rx *rx = dd->netdev_rx; int res; if (atomic_fetch_inc(&rx->netdevs)) return 0; mutex_lock(&hfi1_mutex); res = hfi1_netdev_rxq_init(rx); mutex_unlock(&hfi1_mutex); return res; } /** * hfi1_netdev_rx_destroy - Decrements netdevs counter, when it reaches 0 * napi is deleted and receive queses memory is freed. * * @dd: hfi1 dev data */ int hfi1_netdev_rx_destroy(struct hfi1_devdata *dd) { struct hfi1_netdev_rx *rx = dd->netdev_rx; /* destroy the RX queues only if it is the last netdev going away */ if (atomic_fetch_add_unless(&rx->netdevs, -1, 0) == 1) { mutex_lock(&hfi1_mutex); hfi1_netdev_rxq_deinit(rx); mutex_unlock(&hfi1_mutex); } return 0; } /** * hfi1_alloc_rx - Allocates the rx support structure * @dd: hfi1 dev data * * Allocate the rx structure to support gathering the receive * resources and the dummy netdev. * * Updates dd struct pointer upon success. * * Return: 0 (success) -error on failure * */ int hfi1_alloc_rx(struct hfi1_devdata *dd) { struct hfi1_netdev_rx *rx; dd_dev_info(dd, "allocating rx size %ld\n", sizeof(*rx)); rx = kzalloc_node(sizeof(*rx), GFP_KERNEL, dd->node); if (!rx) return -ENOMEM; rx->dd = dd; rx->rx_napi = alloc_netdev_dummy(0); if (!rx->rx_napi) { kfree(rx); return -ENOMEM; } xa_init(&rx->dev_tbl); atomic_set(&rx->enabled, 0); atomic_set(&rx->netdevs, 0); dd->netdev_rx = rx; return 0; } void hfi1_free_rx(struct hfi1_devdata *dd) { if (dd->netdev_rx) { dd_dev_info(dd, "hfi1 rx freed\n"); free_netdev(dd->netdev_rx->rx_napi); kfree(dd->netdev_rx); dd->netdev_rx = NULL; } } /** * hfi1_netdev_enable_queues - This is napi enable function. * It enables napi objects associated with queues. * When at least one device has called it it increments atomic counter. * Disable function decrements counter and when it is 0, * calls napi_disable for every queue. * * @dd: hfi1 dev data */ void hfi1_netdev_enable_queues(struct hfi1_devdata *dd) { struct hfi1_netdev_rx *rx; if (!dd->netdev_rx) return; rx = dd->netdev_rx; if (atomic_fetch_inc(&rx->enabled)) return; mutex_lock(&hfi1_mutex); enable_queues(rx); mutex_unlock(&hfi1_mutex); } void hfi1_netdev_disable_queues(struct hfi1_devdata *dd) { struct hfi1_netdev_rx *rx; if (!dd->netdev_rx) return; rx = dd->netdev_rx; if (atomic_dec_if_positive(&rx->enabled)) return; mutex_lock(&hfi1_mutex); disable_queues(rx); mutex_unlock(&hfi1_mutex); } /** * hfi1_netdev_add_data - Registers data with unique identifier * to be requested later this is needed for VNIC and IPoIB VLANs * implementations. * This call is protected by mutex idr_lock. * * @dd: hfi1 dev data * @id: requested integer id up to INT_MAX * @data: data to be associated with index */ int hfi1_netdev_add_data(struct hfi1_devdata *dd, int id, void *data) { struct hfi1_netdev_rx *rx = dd->netdev_rx; return xa_insert(&rx->dev_tbl, id, data, GFP_NOWAIT); } /** * hfi1_netdev_remove_data - Removes data with previously given id. * Returns the reference to removed entry. * * @dd: hfi1 dev data * @id: requested integer id up to INT_MAX */ void *hfi1_netdev_remove_data(struct hfi1_devdata *dd, int id) { struct hfi1_netdev_rx *rx = dd->netdev_rx; return xa_erase(&rx->dev_tbl, id); } /** * hfi1_netdev_get_data - Gets data with given id * * @dd: hfi1 dev data * @id: requested integer id up to INT_MAX */ void *hfi1_netdev_get_data(struct hfi1_devdata *dd, int id) { struct hfi1_netdev_rx *rx = dd->netdev_rx; return xa_load(&rx->dev_tbl, id); } /** * hfi1_netdev_get_first_data - Gets first entry with greater or equal id. * * @dd: hfi1 dev data * @start_id: requested integer id up to INT_MAX */ void *hfi1_netdev_get_first_data(struct hfi1_devdata *dd, int *start_id) { struct hfi1_netdev_rx *rx = dd->netdev_rx; unsigned long index = *start_id; void *ret; ret = xa_find(&rx->dev_tbl, &index, UINT_MAX, XA_PRESENT); *start_id = (int)index; return ret; }