/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | ring_mode.c | 17 static int jumbo_frm(struct stmmac_tx_queue *tx_q, struct sk_buff *skb, argument 21 struct stmmac_priv *priv = tx_q->priv_data; 22 unsigned int entry = tx_q->cur_tx; 27 desc = (struct dma_desc *)(tx_q->dma_etx + entry); 29 desc = tx_q->dma_tx + entry; 46 tx_q->tx_skbuff_dma[entry].buf = des2; 47 tx_q->tx_skbuff_dma[entry].len = bmax; 48 tx_q->tx_skbuff_dma[entry].is_jumbo = true; 53 tx_q->tx_skbuff[entry] = NULL; 57 desc = (struct dma_desc *)(tx_q 119 clean_desc3(struct stmmac_tx_queue *tx_q, struct dma_desc *p) argument [all...] |
H A D | chain_mode.c | 17 static int jumbo_frm(struct stmmac_tx_queue *tx_q, struct sk_buff *skb, argument 21 struct stmmac_priv *priv = tx_q->priv_data; 22 unsigned int entry = tx_q->cur_tx; 27 desc = tx_q->dma_tx + entry; 41 tx_q->tx_skbuff_dma[entry].buf = des2; 42 tx_q->tx_skbuff_dma[entry].len = bmax; 48 tx_q->tx_skbuff[entry] = NULL; 50 desc = tx_q->dma_tx + entry; 59 tx_q->tx_skbuff_dma[entry].buf = des2; 60 tx_q 143 clean_desc3(struct stmmac_tx_queue *tx_q, struct dma_desc *p) argument [all...] |
H A D | stmmac_main.c | 364 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 367 if (tx_q->dirty_tx > tx_q->cur_tx) 368 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; 370 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; 416 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 418 if (tx_q->dirty_tx != tx_q 1288 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 1377 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 1502 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 1773 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 1871 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 1959 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 2116 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; local 2462 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 2605 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 2799 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; local 2867 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; local 3000 struct stmmac_tx_queue *tx_q; local 3062 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 3093 struct stmmac_tx_queue *tx_q = container_of(t, struct stmmac_tx_queue, txtimer); local 3128 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; local 3490 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; local 3512 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; local 3893 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[chan]; local 4102 stmmac_vlan_insert(struct stmmac_priv *priv, struct sk_buff *skb, struct stmmac_tx_queue *tx_q) argument 4147 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 4186 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 4242 struct stmmac_tx_queue *tx_q; local 4511 struct stmmac_tx_queue *tx_q; local 4903 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 6133 struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data; local 6358 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 6851 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local 6929 struct stmmac_tx_queue *tx_q; local 7035 struct stmmac_tx_queue *tx_q; local 7930 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; local [all...] |
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_singleq_txrx.c | 181 * @tx_q: queue to send buffer on 189 static void idpf_tx_singleq_map(struct idpf_queue *tx_q, argument 199 u16 i = tx_q->next_to_use; 208 tx_desc = IDPF_BASE_TX_DESC(tx_q, i); 210 dma = dma_map_single(tx_q->dev, skb->data, size, DMA_TO_DEVICE); 213 if (tx_q->vport->crc_enable) 219 if (dma_mapping_error(tx_q->dev, dma)) 220 return idpf_tx_dma_map_error(tx_q, skb, first, i); 241 if (i == tx_q->desc_count) { 242 tx_desc = IDPF_BASE_TX_DESC(tx_q, 354 idpf_tx_singleq_frame(struct sk_buff *skb, struct idpf_queue *tx_q) argument 422 struct idpf_queue *tx_q; local 445 idpf_tx_singleq_clean(struct idpf_queue *tx_q, int napi_budget, int *cleaned) argument [all...] |
H A D | idpf_txrx.c | 60 * @tx_q: the queue that owns the buffer 63 static void idpf_tx_buf_rel(struct idpf_queue *tx_q, struct idpf_tx_buf *tx_buf) argument 67 dma_unmap_single(tx_q->dev, 73 dma_unmap_page(tx_q->dev, 162 * @tx_q: queue for which the buffers are allocated 166 static int idpf_tx_buf_alloc_all(struct idpf_queue *tx_q) argument 174 buf_size = sizeof(struct idpf_tx_buf) * tx_q->desc_count; 175 tx_q->tx_buf = kzalloc(buf_size, GFP_KERNEL); 176 if (!tx_q->tx_buf) 180 for (i = 0; i < tx_q 212 idpf_tx_desc_alloc(struct idpf_queue *tx_q, bool bufq) argument 1448 idpf_tx_handle_sw_marker(struct idpf_queue *tx_q) argument 1477 idpf_tx_splitq_clean_hdr(struct idpf_queue *tx_q, struct idpf_tx_buf *tx_buf, struct idpf_cleaned_stats *cleaned, int napi_budget) argument 1610 idpf_tx_splitq_clean(struct idpf_queue *tx_q, u16 end, int napi_budget, struct idpf_cleaned_stats *cleaned, bool descs_only) argument 1822 struct idpf_queue *tx_q; local 1901 struct idpf_queue *tx_q = complq->txq_grp->txqs[i]; local 1979 idpf_tx_maybe_stop_common(struct idpf_queue *tx_q, unsigned int size) argument 2002 idpf_tx_maybe_stop_splitq(struct idpf_queue *tx_q, unsigned int descs_needed) argument 2043 idpf_tx_buf_hw_update(struct idpf_queue *tx_q, u32 val, bool xmit_more) argument 2184 idpf_tx_splitq_map(struct idpf_queue *tx_q, struct idpf_tx_splitq_params *params, struct idpf_tx_buf *first) argument 2567 idpf_tx_drop_skb(struct idpf_queue *tx_q, struct sk_buff *skb) argument 2587 idpf_tx_splitq_frame(struct sk_buff *skb, struct idpf_queue *tx_q) argument 2685 struct idpf_queue *tx_q; local [all...] |
H A D | idpf_txrx.h | 1003 void idpf_tx_buf_hw_update(struct idpf_queue *tx_q, u32 val, 1006 netdev_tx_t idpf_tx_drop_skb(struct idpf_queue *tx_q, struct sk_buff *skb); 1013 int idpf_tx_maybe_stop_common(struct idpf_queue *tx_q, unsigned int size);
|
/linux-master/drivers/net/ethernet/qualcomm/emac/ |
H A D | emac-mac.c | 305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), 308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), 311 writel(adpt->tx_q.tpd.count & TPD_RING_SIZE_BMSK, 587 struct emac_tx_queue *tx_q = &adpt->tx_q; local 592 if (!tx_q->tpd.tpbuff) 595 for (i = 0; i < tx_q->tpd.count; i++) { 596 struct emac_buffer *tpbuf = GET_TPD_BUFFER(tx_q, i); 610 size = sizeof(struct emac_buffer) * tx_q->tpd.count; 611 memset(tx_q 662 struct emac_tx_queue *tx_q = &adpt->tx_q; local 674 emac_tx_q_desc_alloc(struct emac_adapter *adpt, struct emac_tx_queue *tx_q) argument 1025 emac_tx_tpd_create(struct emac_adapter *adpt, struct emac_tx_queue *tx_q, struct emac_tpd *tpd) argument 1043 emac_tx_tpd_mark_last(struct emac_adapter *adpt, struct emac_tx_queue *tx_q) argument 1168 emac_tpd_num_free_descs(struct emac_tx_queue *tx_q) argument 1179 emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q) argument 1245 emac_tso_csum(struct emac_adapter *adpt, struct emac_tx_queue *tx_q, struct sk_buff *skb, struct emac_tpd *tpd) argument 1330 emac_tx_fill_tpd(struct emac_adapter *adpt, struct emac_tx_queue *tx_q, struct sk_buff *skb, struct emac_tpd *tpd) argument 1437 emac_mac_tx_buf_send(struct emac_adapter *adpt, struct emac_tx_queue *tx_q, struct sk_buff *skb) argument [all...] |
H A D | emac-mac.h | 231 struct emac_tx_queue *tx_q, 233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
|
H A D | emac.c | 122 return emac_mac_tx_buf_send(adpt, &adpt->tx_q, skb); 160 emac_mac_tx_process(adpt, &adpt->tx_q);
|
H A D | emac.h | 341 struct emac_tx_queue tx_q; member in struct:emac_adapter
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_dcb_lib.h | 40 * @tx_q: hung_queue/tx_queue 42 * finds if queue 'tx_q' falls between the two offsets of any given TC 44 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) argument 46 return (tx_q >= low) && (tx_q < high);
|
/linux-master/net/nfc/nci/ |
H A D | data.c | 146 spin_lock_irqsave(&ndev->tx_q.lock, flags); 149 __skb_queue_tail(&ndev->tx_q, skb_frag); 151 spin_unlock_irqrestore(&ndev->tx_q.lock, flags); 185 skb_queue_tail(&ndev->tx_q, skb);
|
H A D | uart.c | 44 skb = skb_dequeue(&nu->tx_q); 56 return skb_queue_empty(&nu->tx_q); 123 skb_queue_head_init(&nu->tx_q); 186 skb_queue_purge(&nu->tx_q); 363 skb_queue_tail(&nu->tx_q, skb);
|
H A D | ntf.c | 78 if (!skb_queue_empty(&ndev->tx_q)) 709 skb_queue_purge(&ndev->tx_q);
|
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif.c | 211 struct dpmaif_tx_queue *tx_q; local 236 tx_q = &dpmaif_ctrl->txq[tx_idx]; 237 tx_q->index = tx_idx; 238 tx_q->dpmaif_ctrl = dpmaif_ctrl; 239 ret = t7xx_dpmaif_txq_init(tx_q); 261 tx_q = &dpmaif_ctrl->txq[i]; 262 t7xx_dpmaif_txq_free(tx_q); 282 struct dpmaif_tx_queue *tx_q; local 289 tx_q = &dpmaif_ctrl->txq[i]; 290 t7xx_dpmaif_txq_free(tx_q); [all...] |
/linux-master/drivers/net/fddi/skfp/h/ |
H A D | hwmtm.h | 242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used 261 (smc)->hw.fp.tx_q[queue].tx_curr_put
|
H A D | fplustm.h | 191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member in struct:s_smt_fp
|
/linux-master/drivers/net/wireless/mediatek/mt7601u/ |
H A D | dma.c | 312 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; 477 if (!dev->tx_q) 481 mt7601u_free_tx_queue(&dev->tx_q[i]); 505 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, 506 sizeof(*dev->tx_q), GFP_KERNEL); 507 if (!dev->tx_q) 511 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i]))
|
H A D | mt7601u.h | 140 * @tx_lock: protects @tx_q and changes of MT7601U_STATE_*_STATS 201 struct mt7601u_tx_queue *tx_q; member in struct:mt7601u_dev
|
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | mt76_connac_mac.c | 89 if (wcid && pm->tx_q[i].wcid != wcid) 92 dev_kfree_skb(pm->tx_q[i].skb); 93 pm->tx_q[i].skb = NULL; 108 if (!pm->tx_q[qid].skb) { 110 pm->tx_q[qid].wcid = wcid; 111 pm->tx_q[qid].skb = skb; 127 struct mt76_wcid *wcid = pm->tx_q[i].wcid; 130 if (!pm->tx_q[i].skb) 137 mt76_tx(phy, sta, wcid, pm->tx_q[i].skb); 138 pm->tx_q[ [all...] |
/linux-master/drivers/media/radio/wl128x/ |
H A D | fmdrv.h | 205 struct sk_buff_head tx_q; /* TX queue */ member in struct:fmdev
|
H A D | fmdrv_common.c | 328 if (!skb_queue_empty(&fmdev->tx_q)) 352 skb = skb_dequeue(&fmdev->tx_q); 442 skb_queue_tail(&fmdev->tx_q, skb); 1541 skb_queue_head_init(&fmdev->tx_q); 1595 skb_queue_purge(&fmdev->tx_q);
|
/linux-master/include/net/nfc/ |
H A D | nci_core.h | 227 struct sk_buff_head tx_q; member in struct:nci_dev 454 struct sk_buff_head tx_q; member in struct:nci_uart
|
/linux-master/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member in struct:sge_qset
|
/linux-master/net/bluetooth/ |
H A D | l2cap_core.c | 450 skb_queue_head_init(&chan->tx_q); 549 skb_queue_head_init(&chan->tx_q); 659 skb_queue_purge(&chan->tx_q); 674 skb_queue_purge(&chan->tx_q); 1904 skb_queue_splice_tail_init(skbs, &chan->tx_q); 1906 while (!skb_queue_empty(&chan->tx_q)) { 1908 skb = skb_dequeue(&chan->tx_q); 1984 if (skb_queue_is_last(&chan->tx_q, skb)) 1987 chan->tx_send_head = skb_queue_next(&chan->tx_q, skb); 1994 chan->unacked_frames, skb_queue_len(&chan->tx_q)); [all...] |