/linux-master/drivers/net/ethernet/netronome/nfp/nfd3/ |
H A D | rings.c | 11 static void nfp_nfd3_xsk_tx_bufs_free(struct nfp_net_tx_ring *tx_ring) argument 16 while (tx_ring->rd_p != tx_ring->wr_p) { 17 idx = D_IDX(tx_ring, tx_ring->rd_p); 18 txbuf = &tx_ring->txbufs[idx]; 22 tx_ring->qcp_rd_p++; 23 tx_ring->rd_p++; 25 if (tx_ring->r_vec->xsk_pool) { 29 xsk_tx_completed(tx_ring 42 nfp_nfd3_tx_ring_reset(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 101 nfp_nfd3_tx_ring_free(struct nfp_net_tx_ring *tx_ring) argument 127 nfp_nfd3_tx_ring_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 160 nfp_nfd3_tx_ring_bufs_free(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 178 nfp_nfd3_tx_ring_bufs_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 199 nfp_nfd3_print_tx_descs(struct seq_file *file, struct nfp_net_r_vector *r_vec, struct nfp_net_tx_ring *tx_ring, u32 d_rd_p, u32 d_wr_p) argument [all...] |
H A D | xsk.c | 17 struct nfp_net_tx_ring *tx_ring, 26 if (nfp_net_tx_space(tx_ring) < 1) 32 wr_idx = D_IDX(tx_ring, tx_ring->wr_p); 34 txbuf = &tx_ring->txbufs[wr_idx]; 40 txd = &tx_ring->txds[wr_idx]; 50 tx_ring->wr_ptr_add++; 51 tx_ring->wr_p++; 120 struct nfp_net_tx_ring *tx_ring; local 126 tx_ring 15 nfp_nfd3_xsk_tx_xdp(const struct nfp_net_dp *dp, struct nfp_net_r_vector *r_vec, struct nfp_net_rx_ring *rx_ring, struct nfp_net_tx_ring *tx_ring, struct nfp_net_xsk_rx_buf *xrxbuf, unsigned int pkt_len, int pkt_off) argument 275 nfp_nfd3_xsk_complete(struct nfp_net_tx_ring *tx_ring) argument 333 nfp_nfd3_xsk_tx(struct nfp_net_tx_ring *tx_ring) argument [all...] |
H A D | dp.c | 31 static int nfp_nfd3_tx_ring_should_wake(struct nfp_net_tx_ring *tx_ring) argument 33 return !nfp_net_tx_full(tx_ring, MAX_SKB_FRAGS * 4); 36 static int nfp_nfd3_tx_ring_should_stop(struct nfp_net_tx_ring *tx_ring) argument 38 return nfp_net_tx_full(tx_ring, MAX_SKB_FRAGS + 1); 44 * @tx_ring: driver tx queue structure 52 struct nfp_net_tx_ring *tx_ring) 58 if (unlikely(nfp_nfd3_tx_ring_should_wake(tx_ring))) 259 struct nfp_net_tx_ring *tx_ring; local 274 tx_ring = &dp->tx_rings[qidx]; 275 r_vec = tx_ring 51 nfp_nfd3_tx_ring_stop(struct netdev_queue *nd_q, struct nfp_net_tx_ring *tx_ring) argument 421 nfp_nfd3_tx_complete(struct nfp_net_tx_ring *tx_ring, int budget) argument 506 nfp_nfd3_xdp_complete(struct nfp_net_tx_ring *tx_ring) argument 836 nfp_nfd3_tx_xdp_buf(struct nfp_net_dp *dp, struct nfp_net_rx_ring *rx_ring, struct nfp_net_tx_ring *tx_ring, struct nfp_net_rx_buf *rxbuf, unsigned int dma_off, unsigned int pkt_len, bool *completed) argument 910 struct nfp_net_tx_ring *tx_ring; local 1211 struct nfp_net_tx_ring *tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/nfdk/ |
H A D | rings.c | 11 nfp_nfdk_tx_ring_reset(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 16 while (!tx_ring->is_xdp && tx_ring->rd_p != tx_ring->wr_p) { 23 rd_idx = D_IDX(tx_ring, tx_ring->rd_p); 24 txbuf = &tx_ring->ktxbufs[rd_idx]; 28 n_descs = D_BLOCK_CPL(tx_ring->rd_p); 57 tx_ring->rd_p += n_descs; 60 memset(tx_ring 74 nfp_nfdk_tx_ring_free(struct nfp_net_tx_ring *tx_ring) argument 93 nfp_nfdk_tx_ring_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 125 nfp_nfdk_tx_ring_bufs_free(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 131 nfp_nfdk_tx_ring_bufs_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 138 nfp_nfdk_print_tx_descs(struct seq_file *file, struct nfp_net_r_vector *r_vec, struct nfp_net_tx_ring *tx_ring, u32 d_rd_p, u32 d_wr_p) argument [all...] |
H A D | dp.c | 18 static int nfp_nfdk_tx_ring_should_wake(struct nfp_net_tx_ring *tx_ring) argument 20 return !nfp_net_tx_full(tx_ring, NFDK_TX_DESC_STOP_CNT * 2); 23 static int nfp_nfdk_tx_ring_should_stop(struct nfp_net_tx_ring *tx_ring) argument 25 return nfp_net_tx_full(tx_ring, NFDK_TX_DESC_STOP_CNT); 29 struct nfp_net_tx_ring *tx_ring) 35 if (unlikely(nfp_nfdk_tx_ring_should_wake(tx_ring))) 119 nfp_nfdk_tx_maybe_close_block(struct nfp_net_tx_ring *tx_ring, argument 151 if (round_down(tx_ring->wr_p, NFDK_TX_DESC_BLOCK_CNT) != 152 round_down(tx_ring->wr_p + n_descs, NFDK_TX_DESC_BLOCK_CNT)) 155 if ((u32)tx_ring 28 nfp_nfdk_tx_ring_stop(struct netdev_queue *nd_q, struct nfp_net_tx_ring *tx_ring) argument 258 struct nfp_net_tx_ring *tx_ring; local 471 nfp_nfdk_tx_complete(struct nfp_net_tx_ring *tx_ring, int budget) argument 835 nfp_nfdk_xdp_complete(struct nfp_net_tx_ring *tx_ring) argument 899 nfp_nfdk_tx_xdp_buf(struct nfp_net_dp *dp, struct nfp_net_rx_ring *rx_ring, struct nfp_net_tx_ring *tx_ring, struct nfp_net_rx_buf *rxbuf, unsigned int dma_off, unsigned int pkt_len, bool *completed) argument 1031 struct nfp_net_tx_ring *tx_ring; local 1321 struct nfp_net_tx_ring *tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx_common.h | 42 * @tx_ring: Tx ring to update 46 static inline void i40e_update_tx_stats(struct i40e_ring *tx_ring, argument 50 u64_stats_update_begin(&tx_ring->syncp); 51 tx_ring->stats.bytes += total_bytes; 52 tx_ring->stats.packets += total_packets; 53 u64_stats_update_end(&tx_ring->syncp); 54 tx_ring->q_vector->tx.total_bytes += total_bytes; 55 tx_ring->q_vector->tx.total_packets += total_packets; 62 * @tx_ring: Tx ring to update 66 static inline void i40e_arm_wb(struct i40e_ring *tx_ring, argument [all...] |
H A D | i40e_xsk.c | 631 * @tx_ring: XDP Tx ring 634 static void i40e_clean_xdp_tx_buffer(struct i40e_ring *tx_ring, argument 638 tx_ring->xdp_tx_active--; 639 dma_unmap_single(tx_ring->dev, 648 * @tx_ring: XDP Tx ring 652 bool i40e_clean_xdp_tx_irq(struct i40e_vsi *vsi, struct i40e_ring *tx_ring) argument 654 struct xsk_buff_pool *bp = tx_ring->xsk_pool; 656 u32 head_idx = i40e_get_head(tx_ring); 660 if (head_idx < tx_ring->next_to_clean) 661 head_idx += tx_ring 768 i40e_xsk_clean_tx_ring(struct i40e_ring *tx_ring) argument [all...] |
/linux-master/drivers/net/ethernet/amazon/ena/ |
H A D | ena_xdp.c | 8 static int validate_xdp_req_id(struct ena_ring *tx_ring, u16 req_id) argument 12 tx_info = &tx_ring->tx_buffer_info[req_id]; 16 return handle_invalid_req_id(tx_ring, req_id, tx_info, true); 19 static int ena_xdp_tx_map_frame(struct ena_ring *tx_ring, argument 24 struct ena_adapter *adapter = tx_ring->adapter; 35 if (tx_ring->tx_mem_queue_type == ENA_ADMIN_PLACEMENT_POLICY_DEV) { 37 push_len = min_t(u32, size, tx_ring->tx_max_header_size); 48 dma = dma_map_single(tx_ring->dev, 52 if (unlikely(dma_mapping_error(tx_ring->dev, dma))) 68 ena_increase_stat(&tx_ring 75 ena_xdp_xmit_frame(struct ena_ring *tx_ring, struct ena_adapter *adapter, struct xdp_frame *xdpf, int flags) argument 125 struct ena_ring *tx_ring; local 367 ena_clean_xdp_irq(struct ena_ring *tx_ring, u32 budget) argument 431 struct ena_ring *tx_ring; local [all...] |
H A D | ena_netdev.c | 53 struct ena_ring *tx_ring; local 62 tx_ring = &adapter->tx_ring[txqueue]; 64 time_since_last_napi = jiffies_to_usecs(jiffies - tx_ring->tx_stats.last_napi_jiffies); 65 napi_scheduled = !!(tx_ring->napi->state & NAPIF_STATE_SCHED); 216 txr = &adapter->tx_ring[i]; 246 rxr->xdp_ring = &adapter->tx_ring[i + adapter->num_io_queues]; 259 struct ena_ring *tx_ring = &adapter->tx_ring[qid]; local 263 if (tx_ring 326 struct ena_ring *tx_ring = &adapter->tx_ring[qid]; local 685 ena_unmap_tx_buff(struct ena_ring *tx_ring, struct ena_tx_buffer *tx_info) argument 718 ena_free_tx_bufs(struct ena_ring *tx_ring) argument 758 struct ena_ring *tx_ring; local 819 validate_tx_req_id(struct ena_ring *tx_ring, u16 req_id) argument 830 ena_clean_tx_irq(struct ena_ring *tx_ring, u32 budget) argument 1391 ena_unmask_interrupt(struct ena_ring *tx_ring, struct ena_ring *rx_ring) argument 1424 ena_update_ring_numa_node(struct ena_ring *tx_ring, struct ena_ring *rx_ring) argument 1463 struct ena_ring *tx_ring, *rx_ring; local 1789 struct ena_ring *rx_ring, *tx_ring; local 1889 struct ena_ring *tx_ring; local 2468 ena_check_and_linearize_skb(struct ena_ring *tx_ring, struct sk_buff *skb) argument 2494 ena_tx_map_skb(struct ena_ring *tx_ring, struct ena_tx_buffer *tx_info, struct sk_buff *skb, void **push_hdr, u16 *header_len) argument 2604 struct ena_ring *tx_ring; local 2818 struct ena_ring *rx_ring, *tx_ring; local 3415 check_missing_comp_in_tx_queue(struct ena_adapter *adapter, struct ena_ring *tx_ring) argument 3507 struct ena_ring *tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_dp.h | 42 * @tx_ring: TX ring to check 51 static inline int nfp_net_tx_full(struct nfp_net_tx_ring *tx_ring, int dcnt) argument 53 return (tx_ring->wr_p - tx_ring->rd_p) >= (tx_ring->cnt - dcnt); 56 static inline void nfp_net_tx_xmit_more_flush(struct nfp_net_tx_ring *tx_ring) argument 59 nfp_qcp_wr_ptr_add(tx_ring->qcp_q, tx_ring->wr_ptr_add); 60 tx_ring->wr_ptr_add = 0; 64 nfp_net_read_tx_cmpl(struct nfp_net_tx_ring *tx_ring, struc argument 167 nfp_net_tx_ring_reset(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 180 nfp_net_tx_ring_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 186 nfp_net_tx_ring_free(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 192 nfp_net_tx_ring_bufs_alloc(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 199 nfp_net_tx_ring_bufs_free(struct nfp_net_dp *dp, struct nfp_net_tx_ring *tx_ring) argument 206 nfp_net_debugfs_print_tx_descs(struct seq_file *file, struct nfp_net_dp *dp, struct nfp_net_r_vector *r_vec, struct nfp_net_tx_ring *tx_ring, u32 d_rd_p, u32 d_wr_p) argument [all...] |
H A D | nfp_net_xsk.h | 23 static inline int nfp_net_tx_space(struct nfp_net_tx_ring *tx_ring) argument 25 return tx_ring->cnt - tx_ring->wr_p + tx_ring->rd_p - 1;
|
H A D | nfp_net_debugfs.c | 83 struct nfp_net_tx_ring *tx_ring; local 90 tx_ring = r_vec->tx_ring; 92 tx_ring = r_vec->xdp_ring; 93 if (!r_vec->nfp_net || !tx_ring) 99 d_rd_p = nfp_qcp_rd_ptr_read(tx_ring->qcp_q); 100 d_wr_p = nfp_qcp_wr_ptr_read(tx_ring->qcp_q); 103 tx_ring->idx, tx_ring->qcidx, 104 tx_ring [all...] |
/linux-master/drivers/net/can/spi/mcp251xfd/ |
H A D | mcp251xfd-tx.c | 21 mcp251xfd_tx_obj *mcp251xfd_get_tx_obj_next(struct mcp251xfd_tx_ring *tx_ring) argument 25 tx_head = mcp251xfd_get_tx_head(tx_ring); 27 return &tx_ring->obj[tx_head]; 141 struct mcp251xfd_tx_ring *tx_ring) 143 if (mcp251xfd_get_tx_free(tx_ring) > 0) 151 if (mcp251xfd_get_tx_free(tx_ring) == 0) { 154 tx_ring->head, tx_ring->tail, 155 tx_ring->head - tx_ring 140 mcp251xfd_tx_busy(const struct mcp251xfd_priv *priv, struct mcp251xfd_tx_ring *tx_ring) argument 169 struct mcp251xfd_tx_ring *tx_ring = priv->tx; local [all...] |
H A D | mcp251xfd-tef.c | 61 const struct mcp251xfd_tx_ring *tx_ring = priv->tx; local 80 seq, priv->tef->tail, priv->tef->head, tx_ring->head); 125 const struct mcp251xfd_tx_ring *tx_ring = priv->tx; local 137 new_head = round_down(priv->tef->head, tx_ring->obj_num) + chip_tx_tail; 139 new_head += tx_ring->obj_num; 142 priv->tef->head = min(new_head, tx_ring->head); 152 const struct mcp251xfd_tx_ring *tx_ring = priv->tx; local 156 (offset > tx_ring->obj_num || 157 len > tx_ring->obj_num || 158 offset + len > tx_ring 223 struct mcp251xfd_tx_ring *tx_ring = priv->tx; local [all...] |
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_algs_send.c | 15 ret = adf_send_message(req->tx_ring, req->fw_req); 30 if (adf_send_message(req->tx_ring, req->fw_req)) { 46 struct adf_etr_ring_data *tx_ring = req->tx_ring; local 54 if (adf_ring_nearly_full(tx_ring)) 58 if (adf_send_message(tx_ring, fw_req))
|
H A D | qat_algs_send.h | 16 struct adf_etr_ring_data *tx_ring; member in struct:qat_alg_req
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_txrx.c | 39 struct ice_tx_ring *tx_ring; local 48 tx_ring = vsi->tx_rings[0]; 49 if (!tx_ring || !tx_ring->desc) 51 dev = tx_ring->dev; 54 for (i = ICE_FDIR_CLEAN_DELAY; ICE_DESC_UNUSED(tx_ring) < 2; i--) { 67 i = tx_ring->next_to_use; 68 first = &tx_ring->tx_buf[i]; 69 f_desc = ICE_TX_FDIRDESC(tx_ring, i); 73 i = (i < tx_ring 151 ice_clean_tx_ring(struct ice_tx_ring *tx_ring) argument 193 ice_free_tx_ring(struct ice_tx_ring *tx_ring) argument 217 ice_clean_tx_irq(struct ice_tx_ring *tx_ring, int napi_budget) argument 341 ice_setup_tx_ring(struct ice_tx_ring *tx_ring) argument 1339 struct ice_tx_ring *tx_ring; local 1514 struct ice_tx_ring *tx_ring; local 1597 __ice_maybe_stop_tx(struct ice_tx_ring *tx_ring, unsigned int size) argument 1620 ice_maybe_stop_tx(struct ice_tx_ring *tx_ring, unsigned int size) argument 1639 ice_tx_map(struct ice_tx_ring *tx_ring, struct ice_tx_buf *first, struct ice_tx_offload_params *off) argument 1984 ice_tx_prepare_vlan_flags(struct ice_tx_ring *tx_ring, struct ice_tx_buf *first) argument 2297 ice_tstamp(struct ice_tx_ring *tx_ring, struct sk_buff *skb, struct ice_tx_buf *first, struct ice_tx_offload_params *off) argument 2331 ice_xmit_frame_ring(struct sk_buff *skb, struct ice_tx_ring *tx_ring) argument 2447 struct ice_tx_ring *tx_ring; local 2497 ice_clean_ctrl_tx_irq(struct ice_tx_ring *tx_ring) argument [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 49 return hfi1_ipoib_txreqs(txq->tx_ring.sent_txreqs, 50 txq->tx_ring.complete_txreqs); 56 if (atomic_inc_return(&txq->tx_ring.stops) == 1) 63 if (atomic_dec_and_test(&txq->tx_ring.stops)) 70 txq->tx_ring.max_items - 1); 76 txq->tx_ring.max_items) >> 1; 81 ++txq->tx_ring.sent_txreqs; 83 !atomic_xchg(&txq->tx_ring.ring_full, 1)) { 108 atomic_xchg(&txq->tx_ring.ring_full, 0)) { 136 struct hfi1_ipoib_circ_buf *tx_ring local 158 struct hfi1_ipoib_circ_buf *tx_ring = &txq->tx_ring; local 355 struct hfi1_ipoib_circ_buf *tx_ring = &txq->tx_ring; local 461 struct hfi1_ipoib_circ_buf *tx_ring; local 512 struct hfi1_ipoib_circ_buf *tx_ring; local 687 struct hfi1_ipoib_circ_buf *tx_ring; local 800 struct hfi1_ipoib_circ_buf *tx_ring = &txq->tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_main.c | 743 static int fm10k_tso(struct fm10k_ring *tx_ring, argument 776 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); 783 tx_ring->netdev->features &= ~NETIF_F_GSO_UDP_TUNNEL; 785 netdev_err(tx_ring->netdev, 790 static void fm10k_tx_csum(struct fm10k_ring *tx_ring, argument 812 dev_warn(tx_ring->dev, 814 tx_ring->tx_stats.csum_err++; 855 dev_warn(tx_ring->dev, 860 tx_ring 892 fm10k_tx_desc_push(struct fm10k_ring *tx_ring, struct fm10k_tx_desc *tx_desc, u16 i, dma_addr_t dma, unsigned int size, u8 desc_flags) argument 909 __fm10k_maybe_stop_tx(struct fm10k_ring *tx_ring, u16 size) argument 926 fm10k_maybe_stop_tx(struct fm10k_ring *tx_ring, u16 size) argument 933 fm10k_tx_map(struct fm10k_ring *tx_ring, struct fm10k_tx_buffer *first) argument 1052 fm10k_xmit_frame_ring(struct sk_buff *skb, struct fm10k_ring *tx_ring) argument 1130 fm10k_check_tx_hang(struct fm10k_ring *tx_ring) argument 1179 fm10k_clean_tx_irq(struct fm10k_q_vector *q_vector, struct fm10k_ring *tx_ring, int napi_budget) argument [all...] |
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_txrx.c | 57 * @tx_ring: ring to be cleaned 59 static void iavf_clean_tx_ring(struct iavf_ring *tx_ring) argument 65 if (!tx_ring->tx_bi) 69 for (i = 0; i < tx_ring->count; i++) 70 iavf_unmap_and_free_tx_resource(tx_ring, &tx_ring->tx_bi[i]); 72 bi_size = sizeof(struct iavf_tx_buffer) * tx_ring->count; 73 memset(tx_ring->tx_bi, 0, bi_size); 76 memset(tx_ring->desc, 0, tx_ring 94 iavf_free_tx_resources(struct iavf_ring *tx_ring) argument 159 struct iavf_ring *tx_ring = NULL; local 213 iavf_clean_tx_irq(struct iavf_vsi *vsi, struct iavf_ring *tx_ring, int napi_budget) argument 650 iavf_setup_tx_descriptors(struct iavf_ring *tx_ring) argument 1462 iavf_tx_prepare_vlan_flags(struct sk_buff *skb, struct iavf_ring *tx_ring, u32 *flags) argument 1611 iavf_tx_enable_csum(struct sk_buff *skb, u32 *tx_flags, u32 *td_cmd, u32 *td_offset, struct iavf_ring *tx_ring, u32 *cd_tunneling) argument 1775 iavf_create_tx_ctx(struct iavf_ring *tx_ring, const u64 cd_type_cmd_tso_mss, const u32 cd_tunneling, const u32 cd_l2tag2) argument 1890 __iavf_maybe_stop_tx(struct iavf_ring *tx_ring, int size) argument 1916 iavf_tx_map(struct iavf_ring *tx_ring, struct sk_buff *skb, struct iavf_tx_buffer *first, u32 tx_flags, const u8 hdr_len, u32 td_cmd, u32 td_offset) argument 2057 iavf_xmit_frame_ring(struct sk_buff *skb, struct iavf_ring *tx_ring) argument 2160 struct iavf_ring *tx_ring = &adapter->tx_rings[skb->queue_mapping]; local [all...] |
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_xdp.c | 48 struct igc_ring *rx_ring, *tx_ring; local 77 tx_ring = adapter->tx_ring[queue_id]; 83 igc_disable_tx_ring(tx_ring); 88 set_bit(IGC_RING_FLAG_AF_XDP_ZC, &tx_ring->flags); 93 igc_enable_tx_ring(tx_ring); 107 struct igc_ring *rx_ring, *tx_ring; local 123 tx_ring = adapter->tx_ring[queue_id]; 129 igc_disable_tx_ring(tx_ring); [all...] |
H A D | igc_dump.c | 118 struct igc_ring *tx_ring; local 138 tx_ring = adapter->tx_ring[n]; 139 buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; 142 n, tx_ring->next_to_use, tx_ring->next_to_clean, 167 tx_ring = adapter->tx_ring[n]; 170 tx_ring [all...] |
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 702 * @tx_ring: tx ring to clean 706 struct wx_ring *tx_ring, int napi_budget) 710 unsigned int i = tx_ring->next_to_clean; 714 if (!netif_carrier_ok(tx_ring->netdev)) 717 tx_buffer = &tx_ring->tx_buffer_info[i]; 718 tx_desc = WX_TX_DESC(tx_ring, i); 719 i -= tx_ring->count; 746 dma_unmap_single(tx_ring->dev, 760 i -= tx_ring->count; 761 tx_buffer = tx_ring 705 wx_clean_tx_irq(struct wx_q_vector *q_vector, struct wx_ring *tx_ring, int napi_budget) argument 877 wx_maybe_stop_tx(struct wx_ring *tx_ring, u16 size) argument 937 wx_tx_map(struct wx_ring *tx_ring, struct wx_tx_buffer *first, const u8 hdr_len) argument 1069 wx_tx_ctxtdesc(struct wx_ring *tx_ring, u32 vlan_macip_lens, u32 fcoe_sof_eof, u32 type_tucmd, u32 mss_l4len_idx) argument 1209 wx_tso(struct wx_ring *tx_ring, struct wx_tx_buffer *first, u8 *hdr_len, u8 ptype) argument 1330 wx_tx_csum(struct wx_ring *tx_ring, struct wx_tx_buffer *first, u8 ptype) argument 1453 wx_xmit_frame_ring(struct sk_buff *skb, struct wx_ring *tx_ring) argument 1516 struct wx_ring *tx_ring; local 2280 wx_clean_tx_ring(struct wx_ring *tx_ring) argument 2357 wx_free_tx_resources(struct wx_ring *tx_ring) argument 2515 wx_setup_tx_resources(struct wx_ring *tx_ring) argument [all...] |
/linux-master/drivers/net/ethernet/intel/igbvf/ |
H A D | netdev.c | 412 * @tx_ring: ring being initialized 417 struct igbvf_ring *tx_ring) 422 size = sizeof(struct igbvf_buffer) * tx_ring->count; 423 tx_ring->buffer_info = vzalloc(size); 424 if (!tx_ring->buffer_info) 428 tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc); 429 tx_ring->size = ALIGN(tx_ring->size, 4096); 431 tx_ring 416 igbvf_setup_tx_resources(struct igbvf_adapter *adapter, struct igbvf_ring *tx_ring) argument 496 igbvf_clean_tx_ring(struct igbvf_ring *tx_ring) argument 531 igbvf_free_tx_resources(struct igbvf_ring *tx_ring) argument 772 igbvf_clean_tx_irq(struct igbvf_ring *tx_ring) argument 874 struct igbvf_ring *tx_ring = adapter->tx_ring; local 978 struct igbvf_ring *tx_ring = adapter->tx_ring; local 1291 struct igbvf_ring *tx_ring = adapter->tx_ring; local 1919 struct igbvf_ring *tx_ring = adapter->tx_ring; local 1978 igbvf_tx_ctxtdesc(struct igbvf_ring *tx_ring, u32 vlan_macip_lens, u32 type_tucmd, u32 mss_l4len_idx) argument 2003 igbvf_tso(struct igbvf_ring *tx_ring, struct sk_buff *skb, u32 tx_flags, u8 *hdr_len) argument 2077 igbvf_tx_csum(struct igbvf_ring *tx_ring, struct sk_buff *skb, u32 tx_flags, __be16 protocol) argument 2147 igbvf_tx_map_adv(struct igbvf_adapter *adapter, struct igbvf_ring *tx_ring, struct sk_buff *skb) argument 2219 igbvf_tx_queue_adv(struct igbvf_adapter *adapter, struct igbvf_ring *tx_ring, int tx_flags, int count, unsigned int first, u32 paylen, u8 hdr_len) argument 2278 igbvf_xmit_frame_ring_adv(struct sk_buff *skb, struct net_device *netdev, struct igbvf_ring *tx_ring) argument 2356 struct igbvf_ring *tx_ring; local [all...] |
/linux-master/drivers/net/ethernet/freescale/enetc/ |
H A D | enetc.c | 46 struct enetc_bdr *tx_ring) 48 int index = &priv->tx_ring[tx_ring->index] - priv->xdp_tx_ring; 70 static void enetc_unmap_tx_buff(struct enetc_bdr *tx_ring, argument 78 dma_unmap_page(tx_ring->dev, tx_swbd->dma, 82 dma_unmap_single(tx_ring->dev, tx_swbd->dma, 87 static void enetc_free_tx_frame(struct enetc_bdr *tx_ring, argument 94 enetc_unmap_tx_buff(tx_ring, tx_swbd); 106 static void enetc_update_tx_ring_tail(struct enetc_bdr *tx_ring) argument 109 enetc_wr_reg_hot(tx_ring 45 enetc_rx_ring_from_xdp_tx_ring(struct enetc_ndev_priv *priv, struct enetc_bdr *tx_ring) argument 145 enetc_map_tx_buffs(struct enetc_bdr *tx_ring, struct sk_buff *skb) argument 339 enetc_map_tx_tso_hdr(struct enetc_bdr *tx_ring, struct sk_buff *skb, struct enetc_tx_swbd *tx_swbd, union enetc_tx_bd *txbd, int *i, int hdr_len, int data_len) argument 392 enetc_map_tx_tso_data(struct enetc_bdr *tx_ring, struct sk_buff *skb, struct enetc_tx_swbd *tx_swbd, union enetc_tx_bd *txbd, char *data, int size, bool last_bd) argument 458 enetc_tso_complete_csum(struct enetc_bdr *tx_ring, struct tso_t *tso, struct sk_buff *skb, char *hdr, int len, __wsum sum) argument 488 enetc_map_tx_tso_buffs(struct enetc_bdr *tx_ring, struct sk_buff *skb) argument 592 struct enetc_bdr *tx_ring; local 724 enetc_bd_ready_count(struct enetc_bdr *tx_ring, int ci) argument 775 enetc_recycle_xdp_tx_buff(struct enetc_bdr *tx_ring, struct enetc_tx_swbd *tx_swbd) argument 814 enetc_clean_tx_ring(struct enetc_bdr *tx_ring, int napi_budget) argument 1252 enetc_xdp_map_tx_buff(struct enetc_bdr *tx_ring, int i, struct enetc_tx_swbd *tx_swbd, int frm_len) argument 1271 enetc_xdp_tx(struct enetc_bdr *tx_ring, struct enetc_tx_swbd *xdp_tx_arr, int num_tx_swbd) argument 1307 enetc_xdp_frame_to_xdp_tx_swbd(struct enetc_bdr *tx_ring, struct enetc_tx_swbd *xdp_tx_arr, struct xdp_frame *xdp_frame) argument 1379 struct enetc_bdr *tx_ring; local 1538 struct enetc_bdr *tx_ring; local 1838 struct enetc_bdr *tx_ring = priv->tx_ring[i]; local 1940 enetc_assign_tx_resource(struct enetc_bdr *tx_ring, const struct enetc_bdr_resource *res) argument 1990 enetc_free_tx_ring(struct enetc_bdr *tx_ring) argument 2111 enetc_setup_txbdr(struct enetc_hw *hw, struct enetc_bdr *tx_ring) argument 2209 enetc_enable_txbdr(struct enetc_hw *hw, struct enetc_bdr *tx_ring) argument 2269 enetc_wait_txbdr(struct enetc_hw *hw, struct enetc_bdr *tx_ring) argument 2645 struct enetc_bdr *tx_ring; local 2675 struct enetc_bdr *tx_ring; local [all...] |