Searched refs:txq (Results 1 - 25 of 340) sorted by relevance

1234567891011>>

/linux-master/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_tx.c54 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; local
58 if (!txq->que_started)
61 old_sw_rd_idx = txq->drb_rd_idx;
71 drb_cnt = txq->drb_size_cnt - old_sw_rd_idx + new_hw_rd_idx;
73 spin_lock_irqsave(&txq->tx_lock, flags);
74 txq->drb_rd_idx = new_hw_rd_idx;
75 spin_unlock_irqrestore(&txq->tx_lock, flags);
83 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_nu local
137 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; local
153 t7xx_dpmaif_drb_ring_not_empty(struct dpmaif_tx_queue *txq) argument
160 struct dpmaif_tx_queue *txq = container_of(work, struct dpmaif_tx_queue, dpmaif_tx_work); local
248 struct dpmaif_tx_queue *txq; local
349 struct dpmaif_tx_queue *txq; local
358 t7xx_txq_drb_wr_available(struct dpmaif_tx_queue *txq) argument
370 t7xx_txq_burst_send_skb(struct dpmaif_tx_queue *txq) argument
416 struct dpmaif_tx_queue *txq; local
510 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[txq_number]; local
537 t7xx_dpmaif_tx_drb_buf_init(struct dpmaif_tx_queue *txq) argument
563 t7xx_dpmaif_tx_free_drb_skb(struct dpmaif_tx_queue *txq) argument
587 t7xx_dpmaif_tx_drb_buf_rel(struct dpmaif_tx_queue *txq) argument
607 t7xx_dpmaif_txq_init(struct dpmaif_tx_queue *txq) argument
633 t7xx_dpmaif_txq_free(struct dpmaif_tx_queue *txq) argument
647 struct dpmaif_tx_queue *txq; local
665 t7xx_dpmaif_txq_flush_rel(struct dpmaif_tx_queue *txq) argument
[all...]
/linux-master/include/net/
H A Dnetdev_queues.h90 #define netif_txq_try_stop(txq, get_desc, start_thrs) \
94 netif_tx_stop_queue(txq); \
106 netif_tx_start_queue(txq); \
114 * @txq: struct netdev_queue to stop/start
131 #define netif_txq_maybe_stop(txq, get_desc, stop_thrs, start_thrs) \
137 _res = netif_txq_try_stop(txq, get_desc, start_thrs); \
156 * @txq: struct netdev_queue to stop/start
174 #define __netif_txq_completed_wake(txq, pkts, bytes, \
184 netdev_txq_completed_mb(txq, pkts, bytes); \
189 if (unlikely(netif_tx_queue_stopped(txq))
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dselq.h30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) argument
32 while (unlikely(txq >= num_channels))
33 txq -= num_channels;
34 return txq;
37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) argument
39 if (unlikely(txq >= num_channels)) {
40 if (unlikely(txq >= num_channels << 3))
41 txq %= num_channels;
44 txq -= num_channels;
45 while (txq >
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c29 struct hfi1_ipoib_txq *txq; member in struct:ipoib_txparms
47 static u64 hfi1_ipoib_used(struct hfi1_ipoib_txq *txq) argument
49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs,
50 txq->tx_ring.complete_txreqs);
53 static void hfi1_ipoib_stop_txq(struct hfi1_ipoib_txq *txq) argument
55 trace_hfi1_txq_stop(txq);
56 if (atomic_inc_return(&txq->tx_ring.stops) == 1)
57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx);
60 static void hfi1_ipoib_wake_txq(struct hfi1_ipoib_txq *txq) argument
67 hfi1_ipoib_ring_hwat(struct hfi1_ipoib_txq *txq) argument
73 hfi1_ipoib_ring_lwat(struct hfi1_ipoib_txq *txq) argument
79 hfi1_ipoib_check_queue_depth(struct hfi1_ipoib_txq *txq) argument
89 hfi1_ipoib_check_queue_stopped(struct hfi1_ipoib_txq *txq) argument
134 hfi1_ipoib_drain_tx_ring(struct hfi1_ipoib_txq *txq) argument
156 struct hfi1_ipoib_txq *txq = local
353 struct hfi1_ipoib_txq *txq = txp->txq; local
404 hfi1_ipoib_submit_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) argument
422 hfi1_ipoib_flush_tx_list(struct net_device *dev, struct hfi1_ipoib_txq *txq) argument
439 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) argument
460 struct hfi1_ipoib_txq *txq = txp->txq; local
511 struct hfi1_ipoib_txq *txq = txp->txq; local
619 struct hfi1_ipoib_txq *txq = local
660 struct hfi1_ipoib_txq *txq = local
673 struct hfi1_ipoib_txq *txq = local
705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local
756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local
770 hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) argument
799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local
822 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local
834 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; local
844 struct hfi1_ipoib_txq *txq = &priv->txqs[q]; local
[all...]
H A Dtrace_tx.h897 TP_PROTO(struct hfi1_ipoib_txq *txq),
898 TP_ARGS(txq),
900 DD_DEV_ENTRY(txq->priv->dd)
901 __field(struct hfi1_ipoib_txq *, txq)
913 DD_DEV_ASSIGN(txq->priv->dd);
914 __entry->txq = txq;
915 __entry->sde = txq->sde;
916 __entry->head = txq->tx_ring.head;
917 __entry->tail = txq
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/queue/
H A Dtx.c23 struct iwl_txq *txq, u16 byte_cnt,
26 int idx = iwl_txq_get_cmd_index(txq, txq->write_ptr);
31 if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window))
47 struct iwl_gen3_bc_tbl_entry *scd_bc_tbl_gen3 = txq->bc_tbl.addr;
55 struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr;
69 void iwl_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument
71 lockdep_assert_held(&txq->lock);
73 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq
22 iwl_pcie_gen2_update_byte_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) argument
162 iwl_txq_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument
465 iwl_txq_gen2_build_tx_amsdu(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len) argument
552 iwl_txq_gen2_build_tx(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta, int hdr_len, int tx_cmd_len, bool pad) argument
647 iwl_txq_gen2_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_device_tx_cmd *dev_cmd, struct sk_buff *skb, struct iwl_cmd_meta *out_meta) argument
727 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
822 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
852 iwl_txq_gen2_free_memory(struct iwl_trans *trans, struct iwl_txq *txq) argument
884 struct iwl_txq *txq; local
939 iwl_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) argument
991 iwl_txq_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) argument
1023 struct iwl_txq *txq = from_timer(txq, t, stuck_timer); local
1048 iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, bool cmd_queue) argument
1130 struct iwl_txq *txq; local
1173 iwl_txq_alloc_response(struct iwl_trans *trans, struct iwl_txq *txq, struct iwl_host_cmd *hcmd) argument
1230 struct iwl_txq *txq; local
1392 iwl_txq_gen1_tfd_unmap(struct iwl_trans *trans, struct iwl_cmd_meta *meta, struct iwl_txq *txq, int index) argument
1438 iwl_txq_gen1_update_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq, u16 byte_cnt, int num_tbs) argument
1482 iwl_txq_gen1_inval_byte_cnt_tbl(struct iwl_trans *trans, struct iwl_txq *txq) argument
1516 iwl_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) argument
1553 iwl_txq_progress(struct iwl_txq *txq) argument
1581 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
1707 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
1723 struct iwl_txq *txq = trans->txqs.txq[queue]; local
1775 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local
[all...]
H A Dtx.h16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) argument
18 return txq->first_tb_dma +
30 struct iwl_txq *txq)
32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) {
33 IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id);
34 iwl_op_mode_queue_not_full(trans->op_mode, txq->id);
39 struct iwl_txq *txq, int idx)
42 idx = iwl_txq_get_cmd_index(txq, idx);
44 return (u8 *)txq->tfds + trans->txqs.tfd.size * idx;
47 int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, in
29 iwl_wake_queue(struct iwl_trans *trans, struct iwl_txq *txq) argument
38 iwl_txq_get_tfd(struct iwl_trans *trans, struct iwl_txq *txq, int idx) argument
61 iwl_txq_stop(struct iwl_trans *trans, struct iwl_txq *txq) argument
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/pcie/
H A Dtx.c73 struct iwl_txq *txq)
76 int txq_id = txq->id;
78 lockdep_assert_held(&txq->lock);
101 txq->need_update = true;
110 IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr);
111 if (!txq->block)
113 txq->write_ptr | (txq_id << 8));
121 struct iwl_txq *txq = trans->txqs.txq[i]; local
126 spin_lock_bh(&txq
72 iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_txq *txq) argument
135 iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, dma_addr_t addr, u16 len, bool reset) argument
189 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
239 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
350 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
653 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
723 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
833 struct iwl_txq *txq = trans->txqs.txq[txq_id]; local
881 struct iwl_txq *txq = trans->txqs.txq[i]; local
915 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local
1199 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local
1257 iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta) argument
1309 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) argument
1440 iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb, struct iwl_txq *txq, u8 hdr_len, struct iwl_cmd_meta *out_meta, struct iwl_device_tx_cmd *dev_cmd, u16 tb1_len) argument
1459 struct iwl_txq *txq; local
[all...]
H A Dtx-gen2.c31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; local
115 spin_lock_irqsave(&txq->lock, flags);
117 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr);
118 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr);
121 if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) {
122 spin_unlock_irqrestore(&txq->lock, flags);
130 out_cmd = txq
[all...]
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c75 * @txq: Logical Tx Queue
77 static void hinic_txq_clean_stats(struct hinic_txq *txq) argument
79 struct hinic_txq_stats *txq_stats = &txq->txq_stats;
93 * @txq: Logical Tx Queue
96 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) argument
98 struct hinic_txq_stats *txq_stats = &txq->txq_stats;
114 * @txq: Logical Tx Queue
116 static void txq_stats_init(struct hinic_txq *txq) argument
118 struct hinic_txq_stats *txq_stats = &txq->txq_stats;
121 hinic_txq_clean_stats(txq);
499 struct hinic_txq *txq; local
560 struct hinic_txq *txq; local
665 free_all_tx_skbs(struct hinic_txq *txq) argument
699 struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); local
779 struct hinic_txq *txq = data; local
796 tx_request_irq(struct hinic_txq *txq) argument
842 tx_free_irq(struct hinic_txq *txq) argument
858 hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, struct net_device *netdev) argument
921 hinic_clean_txq(struct hinic_txq *txq) argument
[all...]
H A Dhinic_tx.h43 void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats);
49 int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq,
52 void hinic_clean_txq(struct hinic_txq *txq);
/linux-master/drivers/net/wireless/mediatek/mt76/mt7925/
H A Dpci_mcu.c12 enum mt76_mcuq_id txq = MT_MCUQ_WM; local
22 txq = MT_MCUQ_FWDL;
24 return mt76_tx_queue_skb_raw(dev, mdev->q_mcu[txq], skb, 0);
/linux-master/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_trace.h15 TP_PROTO(const struct funeth_txq *txq,
20 TP_ARGS(txq, len, sqe_idx, ngle),
27 __string(devname, txq->netdev->name)
31 __entry->qidx = txq->qidx;
35 __assign_str(devname, txq->netdev->name);
45 TP_PROTO(const struct funeth_txq *txq,
50 TP_ARGS(txq, sqe_idx, num_sqes, hw_head),
57 __string(devname, txq->netdev->name)
61 __entry->qidx = txq->qidx;
65 __assign_str(devname, txq
[all...]
/linux-master/include/trace/events/
H A Dqdisc.h16 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq,
19 TP_ARGS(qdisc, txq, packets, skb),
23 __field(const struct netdev_queue *, txq )
35 __entry->txq = txq;
38 __entry->ifindex = txq->dev ? txq->dev->ifindex : 0;
41 __entry->txq_state = txq->state;
51 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, struct sk_buff *skb),
53 TP_ARGS(qdisc, txq, sk
[all...]
/linux-master/drivers/net/wireless/ath/ath9k/
H A Dxmit.c56 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
59 int tx_flags, struct ath_txq *txq,
62 struct ath_txq *txq, struct list_head *bf_q,
65 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
73 struct ath_txq *txq,
107 void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq)
108 __releases(&txq->axq_lock)
115 skb_queue_splice_init(&txq->complete_q, &q);
116 spin_unlock_bh(&txq->axq_lock);
135 struct ath_txq *txq local
211 ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb) argument
236 struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); local
281 struct ath_txq *txq = tid->txq; local
363 ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument
390 ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, struct sk_buff *skb, int count) argument
487 ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_buf *bf, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_atx_tid *tid, struct ath_tx_status *ts, int txok) argument
729 ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_tx_status *ts, struct ath_buf *bf, struct list_head *bf_head) argument
942 ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct ath_buf **buf) argument
1031 ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) argument
1407 ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, int len) argument
1500 ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct list_head *bf_q, struct ath_buf *bf_first) argument
1533 ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid) argument
1580 struct ath_txq *txq; local
1620 struct ath_txq *txq = txtid->txq; local
1652 struct ath_txq *txq; local
1696 struct ath_txq *txq = sc->tx.uapsdq; local
1808 struct ath_txq *txq = &sc->tx.txq[axq_qnum]; local
1870 ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, struct list_head *list) argument
1903 ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) argument
1931 struct ath_txq *txq; local
1969 ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) argument
1978 ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) argument
2019 struct ath_txq *txq; local
2039 ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, struct list_head *head, bool internal) argument
2106 ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument
2211 ath_tx_setup_buffer(struct ath_softc *sc, struct ath_txq *txq, struct ath_atx_tid *tid, struct sk_buff *skb) argument
2340 struct ath_txq *txq = txctl->txq; local
2464 ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb, int tx_flags, struct ath_txq *txq, struct ieee80211_sta *sta) argument
2516 ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf, struct ath_txq *txq, struct list_head *bf_q, struct ieee80211_sta *sta, struct ath_tx_status *ts, int txok) argument
2629 ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) argument
2723 struct ath_txq *txq; local
2890 struct ath_txq *txq; local
[all...]
/linux-master/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c76 int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) argument
78 u16 idx = txq->sw_tx_cons;
79 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb;
84 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD;
89 "skb is null for txq idx=%d txq->sw_tx_cons=%d txq->sw_tx_prod=%d\n",
90 idx, txq->sw_tx_cons, txq->sw_tx_prod);
96 first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq
131 qede_free_failed_tx_pkt(struct qede_tx_queue *txq, struct eth_tx_1st_bd *first_bd, int nbd, bool data_split) argument
243 map_frag_to_bd(struct qede_tx_queue *txq, skb_frag_t *frag, struct eth_tx_bd *bd) argument
288 qede_update_tx_producer(struct qede_tx_queue *txq) argument
304 qede_xdp_xmit(struct qede_tx_queue *txq, dma_addr_t dma, u16 pad, u16 len, struct page *page, struct xdp_frame *xdpf) argument
387 qede_txq_has_work(struct qede_tx_queue *txq) argument
400 qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) argument
432 qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) argument
1482 struct qede_tx_queue *txq; local
[all...]
H A Dqede_main.c544 struct qede_fastpath *fp, struct qede_tx_queue *txq)
546 struct qed_chain *p_chain = &txq->tx_pbl;
548 /* Dump txq/fp/sb ids etc. other metadata */
551 fp->id, fp->sb_info->igu_sb_id, txq->index, txq->ndev_txq_id, txq->cos,
557 le16_to_cpu(*txq->hw_cons_ptr), txq->sw_tx_prod, txq->sw_tx_cons,
562 qede_tx_log_print(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq) argument
543 qede_txq_fp_log_metadata(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq) argument
599 struct qede_tx_queue *txq; local
1649 qede_free_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) argument
1662 qede_alloc_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) argument
1787 qede_empty_tx_queue(struct qede_dev *edev, struct qede_tx_queue *txq) argument
1887 struct qede_tx_queue *txq = &fp->txq[cos]; local
2057 qede_drain_txq(struct qede_dev *edev, struct qede_tx_queue *txq, bool allow_drain) argument
2090 qede_stop_txq(struct qede_dev *edev, struct qede_tx_queue *txq, int rss_id) argument
2188 qede_start_txq(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_tx_queue *txq, u8 rss_id, u16 sb_idx) argument
2807 qede_is_txq_full(struct qede_dev *edev, struct qede_tx_queue *txq) argument
2876 struct qede_tx_queue *txq = QEDE_FP_TC0_TXQ(fp); local
[all...]
/linux-master/drivers/net/ethernet/marvell/
H A Dmv643xx_eth.c179 #define IS_TSO_HEADER(txq, addr) \
180 ((addr >= txq->tso_hdrs_dma) && \
181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
408 struct tx_queue txq[8]; member in struct:mv643xx_eth_private
440 /* rxq/txq helper functions *************************************************/
446 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) argument
448 return container_of(txq, struct mv643xx_eth_private, txq[txq
467 txq_reset_hw_ptr(struct tx_queue *txq) argument
477 txq_enable(struct tx_queue *txq) argument
483 txq_disable(struct tx_queue *txq) argument
493 txq_maybe_wake(struct tx_queue *txq) argument
727 txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, struct sk_buff *skb, char *data, int length, bool last_tcp, bool is_last) argument
775 txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, u32 *first_cmd_sts, bool first_desc) argument
818 txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument
891 txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) argument
929 txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, struct net_device *dev) argument
997 struct tx_queue *txq; local
1032 txq_kick(struct tx_queue *txq) argument
1057 txq_reclaim(struct tx_queue *txq, int budget, int force) argument
1165 txq_set_rate(struct tx_queue *txq, int rate, int burst) argument
1183 txq_set_fixed_prio_mode(struct tx_queue *txq) argument
1262 struct tx_queue *txq = mp->txq + i; local
2036 struct tx_queue *txq = mp->txq + index; local
2122 txq_deinit(struct tx_queue *txq) argument
2212 struct tx_queue *txq = mp->txq + i; local
2355 struct tx_queue *txq = mp->txq + i; local
[all...]
H A Dmvneta.c135 #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8)
773 /* Increment txq get counter */
774 static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) argument
776 txq->txq_get_index++;
777 if (txq->txq_get_index == txq->size)
778 txq->txq_get_index = 0;
781 /* Increment txq put counter */
782 static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq) argument
963 mvneta_txq_pend_desc_add(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int pend_desc) argument
982 mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) argument
993 mvneta_txq_desc_put(struct mvneta_tx_queue *txq) argument
1260 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
1498 int rxq, txq; local
1708 mvneta_tx_done_pkts_coal_set(struct mvneta_port *pp, struct mvneta_tx_queue *txq, u32 value) argument
1734 mvneta_txq_sent_desc_dec(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int sent_desc) argument
1752 mvneta_txq_sent_desc_num_get(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1768 mvneta_txq_sent_desc_proc(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) argument
1913 mvneta_txq_done(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
2091 mvneta_xdp_submit_frame(struct mvneta_port *pp, struct mvneta_tx_queue *txq, struct xdp_frame *xdpf, int *nxmit_byte, bool dma_map) argument
2177 struct mvneta_tx_queue *txq; local
2219 struct mvneta_tx_queue *txq; local
2670 mvneta_free_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
2686 mvneta_alloc_tso_hdrs(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
2706 mvneta_get_tso_hdr(struct mvneta_tx_queue *txq, dma_addr_t *dma) argument
2718 mvneta_tso_put_hdr(struct sk_buff *skb, struct mvneta_tx_queue *txq, struct tso_t *tso, int size, bool is_last) argument
2742 mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, struct sk_buff *skb, char *data, int size, bool last_tcp, bool is_last) argument
2775 mvneta_release_descs(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int first, int num) argument
2803 mvneta_tx_tso(struct sk_buff *skb, struct net_device *dev, struct mvneta_tx_queue *txq) argument
2862 mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb, struct mvneta_tx_queue *txq) argument
2915 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; local
3005 mvneta_txq_done_force(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3025 struct mvneta_tx_queue *txq; local
3524 mvneta_txq_sw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3569 mvneta_txq_hw_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3584 mvneta_txq_init(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3599 mvneta_txq_sw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3621 mvneta_txq_hw_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
3633 mvneta_txq_deinit(struct mvneta_port *pp, struct mvneta_tx_queue *txq) argument
4691 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
5222 int rxq, txq, tc, ret; local
5357 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
5794 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
5847 struct mvneta_tx_queue *txq = &pp->txqs[queue]; local
[all...]
/linux-master/drivers/net/wireless/mediatek/mt76/mt7921/
H A Dpci_mcu.c26 enum mt76_mcuq_id txq = MT_MCUQ_WM; local
36 txq = MT_MCUQ_FWDL;
38 return mt76_tx_queue_skb_raw(dev, mdev->q_mcu[txq], skb, 0);
/linux-master/drivers/net/ethernet/atheros/alx/
H A Dmain.c53 static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) argument
55 struct alx_buffer *txb = &txq->bufs[entry];
58 dma_unmap_single(txq->dev,
149 return alx->qnapi[r_idx]->txq;
152 static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) argument
154 return netdev_get_tx_queue(txq->netdev, txq->queue_idx);
157 static inline int alx_tpd_avail(struct alx_tx_queue *txq) argument
159 if (txq->write_idx >= txq
164 alx_clean_tx_irq(struct alx_tx_queue *txq) argument
488 alx_free_txring_buf(struct alx_tx_queue *txq) argument
616 alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, int offset) argument
742 struct alx_tx_queue *txq; local
1460 alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) argument
1531 alx_start_xmit_ring(struct sk_buff *skb, struct alx_tx_queue *txq) argument
[all...]
/linux-master/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.h124 #define IDPF_BASE_TX_DESC(txq, i) \
125 (&(((struct idpf_base_tx_desc *)((txq)->desc_ring))[i]))
126 #define IDPF_BASE_TX_CTX_DESC(txq, i) \
127 (&(((struct idpf_base_tx_ctx_desc *)((txq)->desc_ring))[i]))
131 #define IDPF_FLEX_TX_DESC(txq, i) \
132 (&(((union idpf_tx_flex_desc *)((txq)->desc_ring))[i]))
133 #define IDPF_FLEX_TX_CTX_DESC(txq, i) \
134 (&(((struct idpf_flex_tx_ctx_desc *)((txq)->desc_ring))[i]))
136 #define IDPF_DESC_UNUSED(txq) \
137 ((((txq)
[all...]
/linux-master/drivers/net/ethernet/freescale/
H A Dfec_main.c307 #define IS_TSO_HEADER(txq, addr) \
308 ((addr >= txq->tso_hdrs_dma) && \
309 (addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE))
333 static int fec_enet_get_free_txdesc_num(struct fec_enet_priv_tx_q *txq) argument
337 entries = (((const char *)txq->dirty_tx -
338 (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1;
340 return entries >= 0 ? entries : entries + txq->bd.ring_size;
356 struct fec_enet_priv_tx_q *txq; local
509 fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
597 fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
723 fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index, char *data, int size, bool last_tcp, bool is_last) argument
783 fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev, struct bufdesc *bdp, int index) argument
836 fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, struct net_device *ndev) argument
926 struct fec_enet_priv_tx_q *txq; local
953 struct fec_enet_priv_tx_q *txq; local
1041 struct fec_enet_priv_tx_q *txq; local
1419 struct fec_enet_priv_tx_q *txq; local
3265 struct fec_enet_priv_tx_q *txq; local
3314 struct fec_enet_priv_tx_q *txq; local
3335 struct fec_enet_priv_tx_q *txq; local
3433 struct fec_enet_priv_tx_q *txq; local
3819 fec_enet_txq_xmit_frame(struct fec_enet_private *fep, struct fec_enet_priv_tx_q *txq, void *frame, u32 dma_sync_len, bool ndo_xmit) argument
3914 struct fec_enet_priv_tx_q *txq; local
3939 struct fec_enet_priv_tx_q *txq; local
4097 struct fec_enet_priv_tx_q *txq = fep->tx_queue[i]; local
[all...]
/linux-master/drivers/net/wireless/ath/ath10k/
H A Dtxrx.c50 struct ieee80211_txq *txq; local
76 txq = skb_cb->txq;
78 if (txq) {
79 artxq = (void *)txq->drv_priv;
89 if (txq && txq->sta && skb_cb->airtime_est)
90 ieee80211_sta_register_airtime(txq->sta, txq->tid,
138 if (txq)
[all...]
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c1133 static void txq_stop(struct sge_eth_txq *txq) argument
1135 netif_tx_stop_queue(txq->txq);
1136 txq->q.stops++;
1164 struct sge_eth_txq *txq; local
1195 txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
1205 reclaim_completed_tx(adapter, &txq->q, true);
1214 credits = txq_avail(&txq->q) - ndesc;
1223 txq_stop(txq);
1237 txq
2131 struct sge_eth_txq *txq = &s->ethtxq[i]; local
2410 t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, struct net_device *dev, struct netdev_queue *devq, unsigned int iqid) argument
2565 struct sge_eth_txq *txq = s->ethtxq; local
[all...]

Completed in 245 milliseconds

1234567891011>>