/linux-master/drivers/slimbus/ |
H A D | qcom-ngd-ctrl.c | 177 int tx_head; member in struct:qcom_slim_ngd_ctrl 536 if ((ctrl->tx_tail + 1) % QCOM_SLIM_NGD_DESC_NUM == ctrl->tx_head) { 563 ctrl->tx_head = (ctrl->tx_head + 1) % QCOM_SLIM_NGD_DESC_NUM; 736 ctrl->tx_head = 0;
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.c | 929 struct i40e_tx_desc *tx_head; local 938 tx_head = I40E_TX_DESC(tx_ring, i40e_get_head(tx_ring)); 952 if (tx_head == tx_desc) 3822 struct i40e_tx_buffer *tx_head = &xdp_ring->tx_bi[index]; local 3823 struct i40e_tx_buffer *tx_bi = tx_head; 3833 tx_head->bytecount = xdp_get_frame_len(xdpf); 3834 tx_head->gso_segs = 1; 3835 tx_head->xdpf = xdpf; 3876 tx_head->next_to_watch = tx_desc; 3890 if (tx_bi == tx_head) [all...] |
/linux-master/drivers/net/ethernet/qualcomm/ |
H A D | qca_uart.c | 43 unsigned char *tx_head; /* pointer to next XMIT byte */ member in struct:qcauart 136 written = serdev_device_write_buf(qca->serdev, qca->tx_head, 140 qca->tx_head += written; 225 qca->tx_head = qca->tx_buffer + written;
|
/linux-master/arch/sparc/kernel/ |
H A D | ldc.c | 123 unsigned long tx_head; member in struct:ldc_channel 227 if (t == lp->tx_head) 247 return lp->tx_head; 529 &lp->tx_head, 926 &lp->tx_head, 931 orig_state, lp->chan_state, lp->tx_head, lp->tx_tail); 1339 &lp->tx_head, 1346 lp->tx_acked = lp->tx_head; 1495 lp->tx_head, lp->tx_tail, lp->tx_num_entries, 1506 hv_err = sun4v_ldc_tx_get_state(lp->id, &lp->tx_head, [all...] |
/linux-master/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_main.c | 8665 struct ixgbe_tx_buffer *tx_head = &ring->tx_buffer_info[index]; local 8666 struct ixgbe_tx_buffer *tx_buff = tx_head; 8674 tx_head->bytecount = xdp_get_frame_len(xdpf); 8675 tx_head->gso_segs = 1; 8676 tx_head->xdpf = xdpf; 8679 cpu_to_le32(tx_head->bytecount << IXGBE_ADVTXD_PAYLEN_SHIFT); 8717 tx_head->next_to_watch = tx_desc; 8730 if (tx_buff == tx_head)
|
/linux-master/drivers/net/ethernet/intel/igb/ |
H A D | igb_main.c | 6359 struct igb_tx_buffer *tx_head = &tx_ring->tx_buffer_info[index]; local 6360 struct igb_tx_buffer *tx_buffer = tx_head; 6374 tx_head->bytecount = xdp_get_frame_len(xdpf); 6375 tx_head->type = IGB_TYPE_XDP; 6376 tx_head->gso_segs = 1; 6377 tx_head->xdpf = xdpf; 6379 olinfo_status = tx_head->bytecount << E1000_ADVTXD_PAYLEN_SHIFT; 6421 netdev_tx_sent_queue(txring_txq(tx_ring), tx_head->bytecount); 6423 tx_head->time_stamp = jiffies; 6429 tx_head [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_txrx_lib.c | 379 struct ice_tx_buf *tx_head; local 402 tx_head = &xdp_ring->tx_buf[ntu]; 403 tx_buf = tx_head; 442 tx_head->bytecount = xdp_get_buff_len(xdp); 443 tx_head->nr_frags = nr_frags; 446 tx_head->type = ICE_TX_BUF_XDP_XMIT; 447 tx_head->xdpf = xdp->data_hard_start; 465 if (tx_buf == tx_head)
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_cmd.h | 579 __le16 tx_head; member in struct:hclge_cfg_tx_queue_pointer_cmd
|
/linux-master/drivers/net/ethernet/google/gve/ |
H A D | gve_tx_dqo.c | 1244 u16 tx_head = le16_to_cpu(compl_desc->tx_head); local 1246 atomic_set_release(&tx->dqo_compl.hw_tx_head, tx_head);
|
/linux-master/drivers/net/can/ |
H A D | xilinx_can.c | 216 * @tx_head: Tx CAN packets ready to send on the queue 242 unsigned int tx_head; member in struct:xcan_priv 467 priv->tx_head = 0; 695 can_put_echo_skb(skb, ndev, priv->tx_head % priv->tx_max, 0); 699 priv->tx_head++; 761 if ((priv->tx_head - priv->tx_tail) == priv->tx_max) 1397 frames_in_fifo = priv->tx_head - priv->tx_tail;
|
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_main.c | 1199 "queue stopped %d, tx_head %u, tx_tail %u, global_tx_head %u, global_tx_tail %u\n", 1200 netif_queue_stopped(dev), priv->tx_head, priv->tx_tail, 1747 /* priv->tx_head, tx_tail and global_tx_tail/head are already 0 */
|
H A D | ipoib_ib.c | 650 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; 662 /* increase the tx_head after send success, but use it for queue state */ 677 rc = post_send(priv, priv->tx_head & (ipoib_sendq_size - 1), 690 rc = priv->tx_head; 691 ++priv->tx_head; 820 while (priv->tx_head != priv->tx_tail || recvs_pending(dev)) { 824 priv->tx_head - priv->tx_tail, 831 while ((int)priv->tx_tail - (int)priv->tx_head < 0) {
|
H A D | ipoib.h | 252 unsigned int tx_head; member in struct:ipoib_cm_tx 383 unsigned int tx_head; member in struct:ipoib_dev_priv
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | sdma.c | 530 return sde->tx_ring[sde->tx_head & sde->sdma_mask]; 555 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; 1810 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; 2195 sde->tx_head,
|
H A D | vnic_sdma.c | 259 list_add_tail(&vnic_sdma->stx.list, &work->tx_head);
|
H A D | verbs.c | 638 &ps->wait->tx_head); 901 &ps->wait->tx_head);
|
H A D | trace_tx.h | 371 __field(u16, tx_head) 378 __entry->tx_head = sde->tx_head; 384 "[%s] SDE(%u) sn %llu hwhead %u swhead %u next_descq_idx %u tx_head %u tx_tail %u", 391 __entry->tx_head, 407 __field(u16, tx_head) 414 __entry->tx_head = sde->tx_head; 419 "[%s] SDE(%u) hwhead %u swhead %u next_descq_idx %u tx_head %u tx_tail %u", 425 __entry->tx_head, [all...] |
H A D | sdma.h | 333 u16 tx_head; member in struct:sdma_engine
|
H A D | qp.c | 126 flush_list_head(&iowait_get_ib_work(&priv->s_iowait)->tx_head); 127 flush_list_head(&iowait_get_tid_work(&priv->s_iowait)->tx_head); 460 list_add_tail(&stx->list, &wait->tx_head);
|
H A D | iowait.h | 39 * @tx_head: list of prebuilt packets 52 struct list_head tx_head; member in struct:iowait_work 58 * @tx_head: overflow list of sdma_txreq's 68 * @count: total number of descriptors in tx_head'ed list 70 * @tx_count: number of tx entry's in tx_head'ed list 289 if (!list_empty(&wait->tx_head)) { 291 &wait->tx_head, 304 if (!list_empty(&w->tx_head)) { 305 tx = list_first_entry(&w->tx_head, struct sdma_txreq, 327 if (!list_empty(&w->tx_head)) { [all...] |
/linux-master/drivers/net/ethernet/cadence/ |
H A D | macb_main.c | 1098 queue->tx_tail, queue->tx_head); 1125 for (tail = queue->tx_tail; tail != queue->tx_head; tail++) { 1184 queue->tx_head = 0; 1241 head = queue->tx_head; 1296 CIRC_CNT(queue->tx_head, queue->tx_tail, 1718 if (queue->tx_head == queue->tx_tail) 1723 head_idx = macb_adj_dma_desc_idx(bp, macb_tx_ring_wrap(bp, queue->tx_head)); 1741 if (queue->tx_head != queue->tx_tail) { 2035 unsigned int len, entry, i, tx_head = queue->tx_head; local [all...] |
H A D | macb.h | 1218 /* Lock to protect tx_head and tx_tail */ 1220 unsigned int tx_head, tx_tail; member in struct:macb_queue
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_send.c | 585 buff_offset = (send_ring->tx_head & 595 send_ring->tx_head++;
|
H A D | dr_types.h | 1459 u32 tx_head; member in struct:mlx5dr_send_ring
|
/linux-master/drivers/net/ethernet/moxa/ |
H A D | moxart_ether.c | 131 priv->tx_head = 0; 282 return CIRC_SPACE(priv->tx_head, priv->tx_tail, TX_DESC_NUM); 288 unsigned int tx_head = priv->tx_head; local 291 while (tx_tail != tx_head) { 335 unsigned int tx_head; local 341 tx_head = priv->tx_head; 342 desc = priv->tx_desc_base + (TX_REG_DESC_SIZE * tx_head); 356 priv->tx_mapping[tx_head] [all...] |