/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_tx.c | 54 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; local 58 if (!txq->que_started) 61 old_sw_rd_idx = txq->drb_rd_idx; 71 drb_cnt = txq->drb_size_cnt - old_sw_rd_idx + new_hw_rd_idx; 73 spin_lock_irqsave(&txq->tx_lock, flags); 74 txq->drb_rd_idx = new_hw_rd_idx; 75 spin_unlock_irqrestore(&txq->tx_lock, flags); 83 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_nu local 137 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; local 153 t7xx_dpmaif_drb_ring_not_empty(struct dpmaif_tx_queue *txq) argument 160 struct dpmaif_tx_queue *txq = container_of(work, struct dpmaif_tx_queue, dpmaif_tx_work); local 248 struct dpmaif_tx_queue *txq; local 349 struct dpmaif_tx_queue *txq; local 358 t7xx_txq_drb_wr_available(struct dpmaif_tx_queue *txq) argument 370 t7xx_txq_burst_send_skb(struct dpmaif_tx_queue *txq) argument 416 struct dpmaif_tx_queue *txq; local 510 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[txq_number]; local 537 t7xx_dpmaif_tx_drb_buf_init(struct dpmaif_tx_queue *txq) argument 563 t7xx_dpmaif_tx_free_drb_skb(struct dpmaif_tx_queue *txq) argument 587 t7xx_dpmaif_tx_drb_buf_rel(struct dpmaif_tx_queue *txq) argument 607 t7xx_dpmaif_txq_init(struct dpmaif_tx_queue *txq) argument 633 t7xx_dpmaif_txq_free(struct dpmaif_tx_queue *txq) argument 647 struct dpmaif_tx_queue *txq; local 665 t7xx_dpmaif_txq_flush_rel(struct dpmaif_tx_queue *txq) argument [all...] |
/linux-master/include/net/ |
H A D | netdev_queues.h | 90 #define netif_txq_try_stop(txq, get_desc, start_thrs) \ 94 netif_tx_stop_queue(txq); \ 106 netif_tx_start_queue(txq); \ 114 * @txq: struct netdev_queue to stop/start 131 #define netif_txq_maybe_stop(txq, get_desc, stop_thrs, start_thrs) \ 137 _res = netif_txq_try_stop(txq, get_desc, start_thrs); \ 156 * @txq: struct netdev_queue to stop/start 174 #define __netif_txq_completed_wake(txq, pkts, bytes, \ 184 netdev_txq_completed_mb(txq, pkts, bytes); \ 189 if (unlikely(netif_tx_queue_stopped(txq)) [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | selq.h | 30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) argument 32 while (unlikely(txq >= num_channels)) 33 txq -= num_channels; 34 return txq; 37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) argument 39 if (unlikely(txq >= num_channels)) { 40 if (unlikely(txq >= num_channels << 3)) 41 txq %= num_channels; 44 txq -= num_channels; 45 while (txq > [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 29 struct hfi1_ipoib_txq *txq; member in struct:ipoib_txparms 47 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) argument 49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs, 50 txq->tx_ring.complete_txreqs); 53 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) argument 55 trace_hfi1_txq_stop(txq); 56 if (atomic_inc_return(&txq->tx_ring.stops) == 1) 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); 60 static void hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) argument 67 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) argument 73 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) argument 79 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) argument 89 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) argument 134 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq) argument 156 struct hfi1_ipoib_txq *txq = local 353 struct hfi1_ipoib_txq *txq = txp->txq; local 404 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) argument 422 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) argument 439 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) argument 460 struct hfi1_ipoib_txq *txq = txp->txq; local 511 struct hfi1_ipoib_txq *txq = txp->txq; local 619 struct hfi1_ipoib_txq *txq = local 660 struct hfi1_ipoib_txq *txq = local 673 struct hfi1_ipoib_txq *txq = local 705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local 756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local 770 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) argument 799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local 822 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local 834 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local 844 struct hfi1_ipoib_txq *txq = &priv->txqs[q]; local [all...] |
H A D | trace_tx.h | 897 TP_PROTO(struct hfi1_ipoib_txq *txq), 898 TP_ARGS(txq), 900 DD_DEV_ENTRY(txq->priv->dd) 901 __field(struct hfi1_ipoib_txq *, txq) 913 DD_DEV_ASSIGN(txq->priv->dd); 914 __entry->txq = txq; 915 __entry->sde = txq->sde; 916 __entry->head = txq->tx_ring.head; 917 __entry->tail = txq [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/queue/ |
H A D | tx.c | 23 struct iwl_txq *txq, u16 byte_cnt, 26 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); 31 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) 47 struct iwl_gen3_bc_tbl_entry *scd_bc_tbl_gen3 = txq->bc_tbl.addr; 55 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; 69 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument 71 lockdep_assert_held(&txq->lock); 73 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq 22 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) argument 162 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument 465 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) argument 552 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) argument 647 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) argument 727 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 822 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 852 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) argument 884 struct iwl_txq *txq; local 939 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) argument 991 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) argument 1023 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); local 1048 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) argument 1130 struct iwl_txq *txq; local 1173 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) argument 1230 struct iwl_txq *txq; local 1392 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) argument 1438 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) argument 1482 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) argument 1516 iwl_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument 1553 iwl_txq_progress(struct iwl_txq *txq) argument 1581 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 1707 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 1723 struct iwl_txq *txq = trans->txqs.txq[queue]; local 1775 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local [all...] |
H A D | tx.h | 16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) argument 18 return txq->first_tb_dma + 30 struct iwl_txq *txq) 32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { 33 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); 34 iwl_op_mode_queue_not_full(trans->op_mode, txq->id); 39 struct iwl_txq *txq, int idx) 42 idx = iwl_txq_get_cmd_index(txq, idx); 44 return (u8 *)txq->tfds + trans->txqs.tfd.size * idx; 47 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in 29 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument 38 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) argument 61 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) argument [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | tx.c | 73 struct iwl_txq *txq) 76 int txq_id = txq->id; 78 lockdep_assert_held(&txq->lock); 101 txq->need_update = true; 110 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); 111 if (!txq->block) 113 txq->write_ptr | (txq_id << 8)); 121 struct iwl_txq *txq = trans->txqs.txq[i]; local 126 spin_lock_bh(&txq 72 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument 135 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) argument 189 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 239 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 350 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 653 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 723 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 833 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local 881 struct iwl_txq *txq = trans->txqs.txq[i]; local 915 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local 1199 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local 1257 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) argument 1309 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) argument 1440 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) argument 1459 struct iwl_txq *txq; local [all...] |
H A D | tx-gen2.c | 31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local 115 spin_lock_irqsave(&txq->lock, flags); 117 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); 118 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); 121 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { 122 spin_unlock_irqrestore(&txq->lock, flags); 130 out_cmd = txq [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_tx.c | 75 * @txq: Logical Tx Queue 77 static void hinic_txq_clean_stats(struct hinic_txq *txq) argument 79 struct hinic_txq_stats *txq_stats = &txq->txq_stats; 93 * @txq: Logical Tx Queue 96 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) argument 98 struct hinic_txq_stats *txq_stats = &txq->txq_stats; 114 * @txq: Logical Tx Queue 116 static void txq_stats_init(struct hinic_txq *txq) argument 118 struct hinic_txq_stats *txq_stats = &txq->txq_stats; 121 hinic_txq_clean_stats(txq); 499 struct hinic_txq *txq; local 560 struct hinic_txq *txq; local 665 free_all_tx_skbs(struct hinic_txq *txq) argument 699 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); local 779 struct hinic_txq *txq = data; local 796 tx_request_irq(struct hinic_txq *txq) argument 842 tx_free_irq(struct hinic_txq *txq) argument 858 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) argument 921 hinic_clean_txq(struct hinic_txq *txq) argument [all...] |
H A D | hinic_tx.h | 43 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats); 49 int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, 52 void hinic_clean_txq(struct hinic_txq *txq);
|
/linux-master/drivers/net/wireless/mediatek/mt76/mt7925/ |
H A D | pci_mcu.c | 12 enum mt76_mcuq_id txq = MT_MCUQ_WM; local 22 txq = MT_MCUQ_FWDL; 24 return mt76_tx_queue_skb_raw(dev, mdev->q_mcu[txq], skb, 0);
|
/linux-master/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_trace.h | 15 TP_PROTO(const struct funeth_txq *txq, 20 TP_ARGS(txq, len, sqe_idx, ngle), 27 __string(devname, txq->netdev->name) 31 __entry->qidx = txq->qidx; 35 __assign_str(devname, txq->netdev->name); 45 TP_PROTO(const struct funeth_txq *txq, 50 TP_ARGS(txq, sqe_idx, num_sqes, hw_head), 57 __string(devname, txq->netdev->name) 61 __entry->qidx = txq->qidx; 65 __assign_str(devname, txq [all...] |
/linux-master/include/trace/events/ |
H A D | qdisc.h | 16 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, 19 TP_ARGS(qdisc, txq, packets, skb), 23 __field(const struct netdev_queue *, txq ) 35 __entry->txq = txq; 38 __entry->ifindex = txq->dev ? txq->dev->ifindex : 0; 41 __entry->txq_state = txq->state; 51 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, struct sk_buff *skb), 53 TP_ARGS(qdisc, txq, sk [all...] |
/linux-master/drivers/net/wireless/ath/ath9k/ |
H A D | xmit.c | 56 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 59 int tx_flags, struct ath_txq *txq, 62 struct ath_txq *txq, struct list_head *bf_q, 65 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 73 struct ath_txq *txq, 107 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq) 108 __releases(&txq->axq_lock) 115 skb_queue_splice_init(&txq->complete_q, &q); 116 spin_unlock_bh(&txq->axq_lock); 135 struct ath_txq *txq local 211 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) argument 236 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); local 281 struct ath_txq *txq = tid->txq; local 363 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument 390 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) argument 487 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) argument 729 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) argument 942 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) argument 1031 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) argument 1407 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) argument 1500 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) argument 1533 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument 1580 struct ath_txq *txq; local 1620 struct ath_txq *txq = txtid->txq; local 1652 struct ath_txq *txq; local 1696 struct ath_txq *txq = sc->tx.uapsdq; local 1808 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; local 1870 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) argument 1903 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) argument 1931 struct ath_txq *txq; local 1969 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) argument 1978 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) argument 2019 struct ath_txq *txq; local 2039 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) argument 2106 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument 2211 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument 2340 struct ath_txq *txq = txctl->txq; local 2464 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) argument 2516 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) argument 2629 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) argument 2723 struct ath_txq *txq; local 2890 struct ath_txq *txq; local [all...] |
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 76 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) argument 78 u16 idx = txq->sw_tx_cons; 79 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; 84 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; 89 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n", 90 idx, txq->sw_tx_cons, txq->sw_tx_prod); 96 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq 131 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) argument 243 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) argument 288 qede_update_tx_producer(struct qede_tx_queue *txq) argument 304 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) argument 387 qede_txq_has_work(struct qede_tx_queue *txq) argument 400 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) argument 432 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) argument 1482 struct qede_tx_queue *txq; local [all...] |
H A D | qede_main.c | 544 struct qede_fastpath *fp, struct qede_tx_queue *txq) 546 struct qed_chain *p_chain = &txq->tx_pbl; 548 /* Dump txq/fp/sb ids etc. other metadata */ 551 fp->id, fp->sb_info->igu_sb_id, txq->index, txq->ndev_txq_id, txq->cos, 557 le16_to_cpu(*txq->hw_cons_ptr), txq->sw_tx_prod, txq->sw_tx_cons, 562 qede_tx_log_print(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq) argument 543 qede_txq_fp_log_metadata(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq) argument 599 struct qede_tx_queue *txq; local 1649 qede_free_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) argument 1662 qede_alloc_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) argument 1787 qede_empty_tx_queue(struct qede_dev *edev, struct qede_tx_queue *txq) argument 1887 struct qede_tx_queue *txq = &fp->txq[cos]; local 2057 qede_drain_txq(struct qede_dev *edev, struct qede_tx_queue *txq, bool allow_drain) argument 2090 qede_stop_txq(struct qede_dev *edev, struct qede_tx_queue *txq, int rss_id) argument 2188 qede_start_txq(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq, u8 rss_id, u16 sb_idx) argument 2807 qede_is_txq_full(struct qede_dev *edev, struct qede_tx_queue *txq) argument 2876 struct qede_tx_queue *txq = QEDE_FP_TC0_TXQ(fp); local [all...] |
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 179 #define IS_TSO_HEADER(txq, addr) \ 180 ((addr >= txq->tso_hdrs_dma) && \ 181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 408 struct tx_queue txq[8]; member in struct:mv643xx_eth_private 440 /* rxq/txq helper functions *************************************************/ 446 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) argument 448 return container_of(txq, struct mv643xx_eth_private, txq[txq 467 txq_reset_hw_ptr(struct tx_queue *txq) argument 477 txq_enable(struct tx_queue *txq) argument 483 txq_disable(struct tx_queue *txq) argument 493 txq_maybe_wake(struct tx_queue *txq) argument 727 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) argument 775 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) argument 818 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument 891 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) argument 929 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument 997 struct tx_queue *txq; local 1032 txq_kick(struct tx_queue *txq) argument 1057 txq_reclaim(struct tx_queue *txq, int budget, int force) argument 1165 txq_set_rate(struct tx_queue *txq, int rate, int burst) argument 1183 txq_set_fixed_prio_mode(struct tx_queue *txq) argument 1262 struct tx_queue *txq = mp->txq + i; local 2036 struct tx_queue *txq = mp->txq + index; local 2122 txq_deinit(struct tx_queue *txq) argument 2212 struct tx_queue *txq = mp->txq + i; local 2355 struct tx_queue *txq = mp->txq + i; local [all...] |
H A D | mvneta.c | 135 #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8) 773 /* Increment txq get counter */ 774 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) argument 776 txq->txq_get_index++; 777 if (txq->txq_get_index == txq->size) 778 txq->txq_get_index = 0; 781 /* Increment txq put counter */ 782 static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq) argument 963 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) argument 982 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) argument 993 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) argument 1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 1498 int rxq, txq; local 1708 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) argument 1734 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) argument 1752 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1768 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) argument 1913 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 2091 mvneta_xdp_submit_frame(struct mvneta_port *pp, struct mvneta_tx_queue *txq, struct xdp_frame *xdpf, int *nxmit_byte, bool dma_map) argument 2177 struct mvneta_tx_queue *txq; local 2219 struct mvneta_tx_queue *txq; local 2670 mvneta_free_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 2686 mvneta_alloc_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 2706 mvneta_get_tso_hdr(struct mvneta_tx_queue *txq, dma_addr_t *dma) argument 2718 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_tx_queue *txq, struct tso_t *tso, int size, bool is_last) argument 2742 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) argument 2775 mvneta_release_descs(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int first, int num) argument 2803 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) argument 2862 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) argument 2915 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; local 3005 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3025 struct mvneta_tx_queue *txq; local 3524 mvneta_txq_sw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3569 mvneta_txq_hw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3584 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3599 mvneta_txq_sw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3621 mvneta_txq_hw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 3633 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument 4691 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 5222 int rxq, txq, tc, ret; local 5357 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 5794 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local 5847 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/mt7921/ |
H A D | pci_mcu.c | 26 enum mt76_mcuq_id txq = MT_MCUQ_WM; local 36 txq = MT_MCUQ_FWDL; 38 return mt76_tx_queue_skb_raw(dev, mdev->q_mcu[txq], skb, 0);
|
/linux-master/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 53 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) argument 55 struct alx_buffer *txb = &txq->bufs[entry]; 58 dma_unmap_single(txq->dev, 149 return alx->qnapi[r_idx]->txq; 152 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) argument 154 return netdev_get_tx_queue(txq->netdev, txq->queue_idx); 157 static inline int alx_tpd_avail(struct alx_tx_queue *txq) argument 159 if (txq->write_idx >= txq 164 alx_clean_tx_irq(struct alx_tx_queue *txq) argument 488 alx_free_txring_buf(struct alx_tx_queue *txq) argument 616 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) argument 742 struct alx_tx_queue *txq; local 1460 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) argument 1531 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) argument [all...] |
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.h | 124 #define IDPF_BASE_TX_DESC(txq, i) \ 125 (&(((struct idpf_base_tx_desc *)((txq)->desc_ring))[i])) 126 #define IDPF_BASE_TX_CTX_DESC(txq, i) \ 127 (&(((struct idpf_base_tx_ctx_desc *)((txq)->desc_ring))[i])) 131 #define IDPF_FLEX_TX_DESC(txq, i) \ 132 (&(((union idpf_tx_flex_desc *)((txq)->desc_ring))[i])) 133 #define IDPF_FLEX_TX_CTX_DESC(txq, i) \ 134 (&(((struct idpf_flex_tx_ctx_desc *)((txq)->desc_ring))[i])) 136 #define IDPF_DESC_UNUSED(txq) \ 137 ((((txq) [all...] |
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 307 #define IS_TSO_HEADER(txq, addr) \ 308 ((addr >= txq->tso_hdrs_dma) && \ 309 (addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE)) 333 static int fec_enet_get_free_txdesc_num(struct fec_enet_priv_tx_q *txq) argument 337 entries = (((const char *)txq->dirty_tx - 338 (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1; 340 return entries >= 0 ? entries : entries + txq->bd.ring_size; 356 struct fec_enet_priv_tx_q *txq; local 509 fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 597 fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 723 fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index, char *data, int size, bool last_tcp, bool is_last) argument 783 fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index) argument 836 fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument 926 struct fec_enet_priv_tx_q *txq; local 953 struct fec_enet_priv_tx_q *txq; local 1041 struct fec_enet_priv_tx_q *txq; local 1419 struct fec_enet_priv_tx_q *txq; local 3265 struct fec_enet_priv_tx_q *txq; local 3314 struct fec_enet_priv_tx_q *txq; local 3335 struct fec_enet_priv_tx_q *txq; local 3433 struct fec_enet_priv_tx_q *txq; local 3819 fec_enet_txq_xmit_frame(struct fec_enet_private *fep, struct fec_enet_priv_tx_q *txq, void *frame, u32 dma_sync_len, bool ndo_xmit) argument 3914 struct fec_enet_priv_tx_q *txq; local 3939 struct fec_enet_priv_tx_q *txq; local 4097 struct fec_enet_priv_tx_q *txq = fep->tx_queue[i]; local [all...] |
/linux-master/drivers/net/wireless/ath/ath10k/ |
H A D | txrx.c | 50 struct ieee80211_txq *txq; local 76 txq = skb_cb->txq; 78 if (txq) { 79 artxq = (void *)txq->drv_priv; 89 if (txq && txq->sta && skb_cb->airtime_est) 90 ieee80211_sta_register_airtime(txq->sta, txq->tid, 138 if (txq) [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 1133 static void txq_stop(struct sge_eth_txq *txq) argument 1135 netif_tx_stop_queue(txq->txq); 1136 txq->q.stops++; 1164 struct sge_eth_txq *txq; local 1195 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; 1205 reclaim_completed_tx(adapter, &txq->q, true); 1214 credits = txq_avail(&txq->q) - ndesc; 1223 txq_stop(txq); 1237 txq 2131 struct sge_eth_txq *txq = &s->ethtxq[i]; local 2410 t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, struct net_device *dev, struct netdev_queue *devq, unsigned int iqid) argument 2565 struct sge_eth_txq *txq = s->ethtxq; local [all...] |