/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | vnic.h | 40 * @q_idx - vnic Tx queue index 49 u8 q_idx; member in struct:hfi1_vnic_sdma 113 u8 q_idx); 122 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx,
|
H A D | vnic_main.c | 124 u8 q_idx, struct sk_buff *skb, int err) 127 struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; 153 u8 q_idx, struct sk_buff *skb, int err) 156 struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; 205 u8 q_idx) 207 netif_stop_subqueue(vinfo->netdev, q_idx); 208 if (!hfi1_vnic_sdma_write_avail(vinfo, q_idx)) 211 netif_start_subqueue(vinfo->netdev, q_idx); 218 u8 pad_len, q_idx = skb->queue_mapping; local 225 v_dbg("xmit: queue %d skb len %d\n", q_idx, sk 123 hfi1_vnic_update_tx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument 152 hfi1_vnic_update_rx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument 204 hfi1_vnic_maybe_stop_tx(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument 339 u8 q_idx; local [all...] |
H A D | vnic_sdma.c | 126 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, argument 130 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; 224 if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx)) 225 netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx); 229 u8 q_idx) 231 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; 249 vnic_sdma->q_idx = i; 228 hfi1_vnic_sdma_write_avail(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument
|
H A D | ipoib.h | 102 * @q_idx: ipoib Tx queue index 114 u8 q_idx; member in struct:hfi1_ipoib_txq
|
H A D | ipoib_tx.c | 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); 64 netif_wake_subqueue(txq->priv->netdev, txq->q_idx); 125 le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx, 723 txq->q_idx = i; 789 txq->q_idx, 848 __netif_subqueue_stopped(dev, txq->q_idx),
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_base.h | 9 int ice_vsi_cfg_single_rxq(struct ice_vsi *vsi, u16 q_idx); 19 u16 q_idx);
|
H A D | ice_xsk.c | 23 * @q_idx: ring index in array 25 static void ice_qp_reset_stats(struct ice_vsi *vsi, u16 q_idx) argument 38 memset(&vsi_stat->rx_ring_stats[q_idx]->rx_stats, 0, 39 sizeof(vsi_stat->rx_ring_stats[q_idx]->rx_stats)); 40 memset(&vsi_stat->tx_ring_stats[q_idx]->stats, 0, 41 sizeof(vsi_stat->tx_ring_stats[q_idx]->stats)); 43 memset(&vsi->xdp_rings[q_idx]->ring_stats->stats, 0, 44 sizeof(vsi->xdp_rings[q_idx]->ring_stats->stats)); 50 * @q_idx: ring index in array 52 static void ice_qp_clean_rings(struct ice_vsi *vsi, u16 q_idx) argument 160 ice_qp_dis(struct ice_vsi *vsi, u16 q_idx) argument 219 ice_qp_ena(struct ice_vsi *vsi, u16 q_idx) argument [all...] |
H A D | ice_base.c | 634 int ice_vsi_cfg_single_rxq(struct ice_vsi *vsi, u16 q_idx) argument 636 if (q_idx >= vsi->num_rxq) 639 return ice_vsi_cfg_rxq(vsi->rx_rings[q_idx]); 957 u16 q_idx) 961 if (q_idx >= vsi->alloc_txq || !tx_rings || !tx_rings[q_idx]) 966 return ice_vsi_cfg_txq(vsi, tx_rings[q_idx], qg_buf); 983 u16 q_idx; local 987 for (q_idx = 0; q_idx < coun 956 ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings, u16 q_idx) argument [all...] |
/linux-master/drivers/net/hyperv/ |
H A D | netvsc_bpf.c | 230 struct xdp_frame *frame, u16 q_idx) 240 skb_record_rx_queue(skb, q_idx); 256 u16 q_idx; local 276 q_idx = smp_processor_id() % ndev->real_num_tx_queues; 279 if (netvsc_ndoxdp_xmit_fm(ndev, frames[i], q_idx)) 285 tx_stats = &nvsc_dev->chan_table[q_idx].tx_stats; 229 netvsc_ndoxdp_xmit_fm(struct net_device *ndev, struct xdp_frame *frame, u16 q_idx) argument
|
H A D | netvsc.c | 321 int netvsc_alloc_recv_comp_ring(struct netvsc_device *net_device, u32 q_idx) argument 323 struct netvsc_channel *nvchan = &net_device->chan_table[q_idx]; 760 u16 q_idx = 0; local 781 q_idx = packet->q_idx; 783 tx_stats = &net_device->chan_table[q_idx].tx_stats; 795 atomic_dec_return(&net_device->chan_table[q_idx].queue_sends); 801 struct netdev_queue *txq = netdev_get_tx_queue(ndev, q_idx); 1063 &net_device->chan_table[packet->q_idx]; 1067 struct netdev_queue *txq = netdev_get_tx_queue(ndev, packet->q_idx); 1362 enq_receive_complete(struct net_device *ndev, struct netvsc_device *nvdev, u16 q_idx, u64 tid, u32 status) argument 1403 u16 q_idx = channel->offermsg.offer.sub_channel_index; local [all...] |
H A D | netvsc_drv.c | 254 int q_idx; local 256 q_idx = ndc->tx_table[netvsc_get_hash(skb, ndc) & 260 if (q_idx != old_idx && 262 sk_tx_queue_set(sk, q_idx); 264 return q_idx; 280 int q_idx = sk_tx_queue_get(skb->sk); local 282 if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) { 287 q_idx = skb_get_rx_queue(skb); 289 q_idx 880 u16 q_idx = channel->offermsg.offer.sub_channel_index; local [all...] |
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_bpf.c | 36 u16 q_idx) 44 skb_set_queue_mapping(skb, q_idx); 57 u16 q_idx; local 62 q_idx = smp_processor_id() % ndev->real_num_tx_queues; 65 if (mana_xdp_xmit_fm(ndev, frames[i], q_idx)) 71 tx_stats = &apc->tx_qp[q_idx].txq.stats; 35 mana_xdp_xmit_fm(struct net_device *ndev, struct xdp_frame *frame, u16 q_idx) argument
|
/linux-master/drivers/accel/habanalabs/common/ |
H A D | hw_queue.c | 420 u32 q_idx; local 423 q_idx = job->hw_queue_id; 424 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 431 "generate signal CB, sob_id: %d, sob val: %u, q_idx: %d, seq: %llu\n", 432 cs_cmpl->hw_sob->sob_id, cs_cmpl->sob_val, q_idx, 441 rc = hl_cs_signal_sob_wraparound_handler(hdev, q_idx, &hw_sob, 1, 480 u32 q_idx; local 482 q_idx = job->hw_queue_id; 483 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 497 cs->encaps_sig_hdl->q_idx, 913 sync_stream_queue_init(struct hl_device *hdev, u32 q_idx) argument 972 sync_stream_queue_reset(struct hl_device *hdev, u32 q_idx) argument [all...] |
H A D | command_submission.c | 155 "SOB release shouldn't be called here, q_idx: %d, sob_id: %d\n", 156 hw_sob->q_idx, hw_sob->sob_id); 1793 * @q_idx: stream queue index 1800 int hl_cs_signal_sob_wraparound_handler(struct hl_device *hdev, u32 q_idx, argument 1808 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 1830 dev_err(hdev->dev, "error: Cannot switch SOBs q_idx: %d\n", 1831 q_idx); 1871 dev_dbg(hdev->dev, "switched to SOB %d, q_idx: %d\n", 1872 prop->curr_sob_offset, q_idx); 1940 enum hl_queue_type q_type, u32 q_idx, u3 1938 cs_ioctl_signal_wait_create_jobs(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_cs *cs, enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) argument 2003 cs_ioctl_reserve_signals(struct hl_fpriv *hpriv, u32 q_idx, u32 count, u32 *handle_id, u32 *sob_addr, u32 *signals_count) argument 2132 u32 q_idx, sob_addr; local 2212 u32 q_idx, collective_engine_id = 0; local [all...] |
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.h | 354 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx); 355 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx); 356 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx); 357 int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx);
|
H A D | nicvf_queues.c | 1723 static u64 nicvf_int_type_to_mask(int int_type, int q_idx) argument 1729 reg_val = ((1ULL << q_idx) << NICVF_INTR_CQ_SHIFT); 1732 reg_val = ((1ULL << q_idx) << NICVF_INTR_SQ_SHIFT); 1735 reg_val = ((1ULL << q_idx) << NICVF_INTR_RBDR_SHIFT); 1757 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx) argument 1759 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1771 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx) argument 1773 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1785 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx) argument 1787 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1799 nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx) argument [all...] |
/linux-master/net/core/ |
H A D | netdev-genl.c | 293 u32 q_idx, u32 q_type, const struct genl_info *info) 303 if (nla_put_u32(rsp, NETDEV_A_QUEUE_ID, q_idx) || 310 rxq = __netif_get_rx_queue(netdev, q_idx); 316 txq = netdev_get_tx_queue(netdev, q_idx); 347 netdev_nl_queue_fill(struct sk_buff *rsp, struct net_device *netdev, u32 q_idx, argument 355 err = netdev_nl_queue_validate(netdev, q_idx, q_type); 359 return netdev_nl_queue_fill_one(rsp, netdev, q_idx, q_type, info); 292 netdev_nl_queue_fill_one(struct sk_buff *rsp, struct net_device *netdev, u32 q_idx, u32 q_type, const struct genl_info *info) argument
|
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_hw.h | 293 u8 q_idx; member in struct:qed_qm_pq_params::__anon1336
|
/linux-master/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_pf.c | 503 u16 vsi, queue, pc, q_idx; local 520 q_idx = dglort->queue_b; 524 for (queue = 0; queue < queue_count; queue++, q_idx++) { 525 if (q_idx >= FM10K_MAX_QUEUES) 528 fm10k_write_reg(hw, FM10K_TX_SGLORT(q_idx), glort); 529 fm10k_write_reg(hw, FM10K_RX_SGLORT(q_idx), glort); 539 q_idx = pc + dglort->queue_b; 541 if (q_idx >= FM10K_MAX_QUEUES) 544 txqctl = fm10k_read_reg(hw, FM10K_TXQCTL(q_idx)); 547 fm10k_write_reg(hw, FM10K_TXQCTL(q_idx), txqct [all...] |
H A D | fm10k_pci.c | 1175 int q_idx; local 1177 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { 1178 q_vector = interface->q_vector[q_idx]; 1872 int q_idx; local 1874 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { 1875 q_vector = interface->q_vector[q_idx]; [all...] |
/linux-master/drivers/net/wireless/realtek/rtlwifi/rtl8192ee/ |
H A D | trx.c | 534 u16 rtl92ee_get_available_desc(struct ieee80211_hw *hw, u8 q_idx) argument 542 get_desc_addr_fr_q_idx(q_idx)); 889 u8 q_idx = *val; local 900 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx]; 903 if (q_idx == BEACON_QUEUE) { 914 get_desc_addr_fr_q_idx(q_idx),
|
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_main.c | 1201 int q_idx; local 1205 for (q_idx = 0; q_idx < q_vectors; q_idx++) { 1208 q_vector = &adapter->q_vectors[q_idx]; 1220 int q_idx; local 1224 for (q_idx = 0; q_idx < q_vectors; q_idx++) { 1225 q_vector = &adapter->q_vectors[q_idx]; 1815 int q_idx = 0, num_q_vectors; local 1848 int q_idx, num_q_vectors; local [all...] |
/linux-master/net/sched/ |
H A D | sch_api.c | 1802 int ret = 0, q_idx = *q_idx_p; local 1810 if (q_idx < s_q_idx) { 1811 q_idx++; 1818 q_idx++; 1831 if (q_idx < s_q_idx) { 1832 q_idx++; 1840 q_idx++; 1844 *q_idx_p = q_idx; 1854 int idx, q_idx; local 1862 s_q_idx = q_idx [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | mt76_connac_mac.c | 478 u8 p_fmt, q_idx, omac_idx = 0, wmm_idx = 0, band_idx = 0; local 500 q_idx = MT_LMAC_ALTX0; 503 q_idx = MT_LMAC_BCN0; 506 q_idx = MT_LMAC_ALTX0; 509 q_idx = wmm_idx * MT76_CONNAC_MAX_WMM_SETS + 519 FIELD_PREP(MT_TXD0_Q_IDX, q_idx);
|
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_defines.h | 643 #define IGC_VLANPQF_QSEL(_n, q_idx) ((q_idx) << ((_n) * 4))
|