/linux-master/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_dp.c | 71 * @rx_ring: RX ring structure 76 nfp_net_rx_ring_init(struct nfp_net_rx_ring *rx_ring, argument 81 rx_ring->idx = idx; 82 rx_ring->r_vec = r_vec; 83 u64_stats_init(&rx_ring->r_vec->rx_sync); 85 rx_ring->fl_qcidx = rx_ring->idx * nn->stride_rx; 86 rx_ring->qcp_fl = nn->rx_bar + NFP_QCP_QUEUE_OFF(rx_ring->fl_qcidx); 91 * @rx_ring 95 nfp_net_rx_ring_reset(struct nfp_net_rx_ring *rx_ring) argument 132 nfp_net_rx_ring_bufs_free(struct nfp_net_dp *dp, struct nfp_net_rx_ring *rx_ring) argument 161 nfp_net_rx_ring_bufs_alloc(struct nfp_net_dp *dp, struct nfp_net_rx_ring *rx_ring) argument 251 nfp_net_rx_ring_free(struct nfp_net_rx_ring *rx_ring) argument 284 nfp_net_rx_ring_alloc(struct nfp_net_dp *dp, struct nfp_net_rx_ring *rx_ring) argument 383 nfp_net_rx_ring_hw_cfg_write(struct nfp_net *nn, struct nfp_net_rx_ring *rx_ring, unsigned int idx) argument [all...] |
H A D | nfp_net_xsk.h | 18 static inline int nfp_net_rx_space(struct nfp_net_rx_ring *rx_ring) argument 20 return rx_ring->cnt - rx_ring->wr_p + rx_ring->rd_p - 1; 35 void nfp_net_xsk_rx_bufs_free(struct nfp_net_rx_ring *rx_ring); 37 void nfp_net_xsk_rx_ring_fill_freelist(struct nfp_net_rx_ring *rx_ring);
|
H A D | nfp_net_xsk.c | 17 nfp_net_xsk_rx_bufs_stash(struct nfp_net_rx_ring *rx_ring, unsigned int idx, argument 22 headroom = xsk_pool_get_headroom(rx_ring->r_vec->xsk_pool); 24 rx_ring->rxds[idx].fld.reserved = 0; 25 rx_ring->rxds[idx].fld.meta_len_dd = 0; 27 rx_ring->xsk_rxbufs[idx].xdp = xdp; 28 rx_ring->xsk_rxbufs[idx].dma_addr = 46 void nfp_net_xsk_rx_bufs_free(struct nfp_net_rx_ring *rx_ring) argument 50 if (!rx_ring->cnt) 53 for (i = 0; i < rx_ring->cnt - 1; i++) 54 nfp_net_xsk_rx_free(&rx_ring 57 nfp_net_xsk_rx_ring_fill_freelist(struct nfp_net_rx_ring *rx_ring) argument [all...] |
H A D | nfp_net_debugfs.c | 15 struct nfp_net_rx_ring *rx_ring; local 24 if (!r_vec->nfp_net || !r_vec->rx_ring) 27 rx_ring = r_vec->rx_ring; 31 rxd_cnt = rx_ring->cnt; 33 fl_rd_p = nfp_qcp_rd_ptr_read(rx_ring->qcp_fl); 34 fl_wr_p = nfp_qcp_wr_ptr_read(rx_ring->qcp_fl); 37 rx_ring->idx, rx_ring->fl_qcidx, 38 rx_ring [all...] |
/linux-master/drivers/net/ethernet/amazon/ena/ |
H A D | ena_xdp.h | 45 int ena_xdp_register_rxq_info(struct ena_ring *rx_ring); 46 void ena_xdp_unregister_rxq_info(struct ena_ring *rx_ring); 82 static inline int ena_xdp_execute(struct ena_ring *rx_ring, struct xdp_buff *xdp) argument 90 xdp_prog = READ_ONCE(rx_ring->xdp_bpf_prog); 98 trace_xdp_exception(rx_ring->netdev, xdp_prog, verdict); 99 xdp_stat = &rx_ring->rx_stats.xdp_aborted; 105 xdp_ring = rx_ring->xdp_ring; 110 if (ena_xdp_xmit_frame(xdp_ring, rx_ring->adapter, xdpf, 115 xdp_stat = &rx_ring->rx_stats.xdp_tx; 119 if (likely(!xdp_do_redirect(rx_ring [all...] |
H A D | ena_netdev.c | 95 adapter->rx_ring[i].mtu = mtu; 217 rxr = &adapter->rx_ring[i]; 393 struct ena_ring *rx_ring = &adapter->rx_ring[qid]; local 397 if (rx_ring->rx_buffer_info) { 406 size = sizeof(struct ena_rx_buffer) * (rx_ring->ring_size + 1); 409 rx_ring->rx_buffer_info = vzalloc_node(size, node); 410 if (!rx_ring->rx_buffer_info) { 411 rx_ring->rx_buffer_info = vzalloc(size); 412 if (!rx_ring 451 struct ena_ring *rx_ring = &adapter->rx_ring[qid]; local 501 ena_alloc_map_page(struct ena_ring *rx_ring, dma_addr_t *dma) argument 530 ena_alloc_rx_buffer(struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info) argument 566 ena_unmap_rx_buff_attrs(struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info, unsigned long attrs) argument 574 ena_free_rx_page(struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info) argument 591 ena_refill_rx_bufs(struct ena_ring *rx_ring, u32 num) argument 646 struct ena_ring *rx_ring = &adapter->rx_ring[qid]; local 662 struct ena_ring *rx_ring; local 911 ena_alloc_skb(struct ena_ring *rx_ring, void *first_frag, u16 len) argument 951 ena_rx_skb(struct ena_ring *rx_ring, struct ena_com_rx_buf_info *ena_bufs, u32 descs, u16 *next_to_clean) argument 1084 ena_rx_checksum(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx, struct sk_buff *skb) argument 1141 ena_set_rx_hash(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx, struct sk_buff *skb) argument 1163 ena_xdp_handle_buff(struct ena_ring *rx_ring, struct xdp_buff *xdp, u16 num_descs) argument 1199 ena_clean_rx_irq(struct ena_ring *rx_ring, struct napi_struct *napi, u32 budget) argument 1364 struct ena_ring *rx_ring = ena_napi->rx_ring; local 1381 ena_unmask_interrupt(struct ena_ring *tx_ring, struct ena_ring *rx_ring) argument 1414 ena_update_ring_numa_node(struct ena_ring *tx_ring, struct ena_ring *rx_ring) argument 1453 struct ena_ring *tx_ring, *rx_ring; local 1779 struct ena_ring *rx_ring, *tx_ring; local 1947 struct ena_ring *rx_ring; local 2351 struct ena_ring *rx_ring; local 2799 struct ena_ring *rx_ring, *tx_ring; local 3365 check_for_rx_interrupt_queue(struct ena_adapter *adapter, struct ena_ring *rx_ring) argument 3483 struct ena_ring *rx_ring; local 3541 struct ena_ring *rx_ring; local [all...] |
H A D | ena_xdp.c | 133 if (!READ_ONCE(adapter->rx_ring->xdp_bpf_prog)) 194 int ena_xdp_register_rxq_info(struct ena_ring *rx_ring) argument 198 rc = xdp_rxq_info_reg(&rx_ring->xdp_rxq, rx_ring->netdev, rx_ring->qid, 0); 200 netif_dbg(rx_ring->adapter, ifup, rx_ring->netdev, "Registering RX info for queue %d", 201 rx_ring->qid); 203 netif_err(rx_ring->adapter, ifup, rx_ring 222 ena_xdp_unregister_rxq_info(struct ena_ring *rx_ring) argument 236 struct ena_ring *rx_ring; local [all...] |
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_txrx_lib.h | 11 * @rx_ring: Rx ring struct 18 ice_set_rx_bufs_act(struct xdp_buff *xdp, const struct ice_rx_ring *rx_ring, argument 22 u32 nr_frags = rx_ring->nr_frags + 1; 23 u32 idx = rx_ring->first_desc; 24 u32 cnt = rx_ring->count; 28 buf = &rx_ring->rx_buf[idx]; 36 if (sinfo_frags < rx_ring->nr_frags && act == ICE_XDP_CONSUMED) { 37 u32 delta = rx_ring->nr_frags - sinfo_frags; 44 buf = &rx_ring->rx_buf[idx]; 69 * @rx_ring 76 ice_is_non_eop(const struct ice_rx_ring *rx_ring, const union ice_32b_rx_flex_desc *rx_desc) argument [all...] |
H A D | ice_xsk.h | 23 int ice_clean_rx_irq_zc(struct ice_rx_ring *rx_ring, int budget); 25 bool ice_alloc_rx_bufs_zc(struct ice_rx_ring *rx_ring, u16 count); 27 void ice_xsk_clean_rx_ring(struct ice_rx_ring *rx_ring); 46 ice_clean_rx_irq_zc(struct ice_rx_ring __always_unused *rx_ring, argument 53 ice_alloc_rx_bufs_zc(struct ice_rx_ring __always_unused *rx_ring, argument 71 static inline void ice_xsk_clean_rx_ring(struct ice_rx_ring *rx_ring) { } argument
|
H A D | ice_txrx.c | 381 * @rx_ring: ring to be cleaned 383 void ice_clean_rx_ring(struct ice_rx_ring *rx_ring) argument 385 struct xdp_buff *xdp = &rx_ring->xdp; 386 struct device *dev = rx_ring->dev; 391 if (!rx_ring->rx_buf) 394 if (rx_ring->xsk_pool) { 395 ice_xsk_clean_rx_ring(rx_ring); 405 for (i = 0; i < rx_ring->count; i++) { 406 struct ice_rx_buf *rx_buf = &rx_ring->rx_buf[i]; 416 rx_ring 451 ice_free_rx_ring(struct ice_rx_ring *rx_ring) argument 483 ice_setup_rx_ring(struct ice_rx_ring *rx_ring) argument 533 ice_rx_frame_truesize(struct ice_rx_ring *rx_ring, const unsigned int size) argument 560 ice_run_xdp(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, struct bpf_prog *xdp_prog, struct ice_tx_ring *xdp_ring, struct ice_rx_buf *rx_buf, union ice_32b_rx_flex_desc *eop_desc) argument 696 ice_alloc_mapped_page(struct ice_rx_ring *rx_ring, struct ice_rx_buf *bi) argument 747 ice_alloc_rx_bufs(struct ice_rx_ring *rx_ring, unsigned int cleaned_count) argument 874 ice_add_xdp_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, struct ice_rx_buf *rx_buf, const unsigned int size) argument 915 ice_reuse_rx_page(struct ice_rx_ring *rx_ring, struct ice_rx_buf *old_buf) argument 946 ice_get_rx_buf(struct ice_rx_ring *rx_ring, const unsigned int size, const unsigned int ntc) argument 983 ice_build_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) argument 1036 ice_construct_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) argument 1117 ice_put_rx_buf(struct ice_rx_ring *rx_ring, struct ice_rx_buf *rx_buf) argument 1149 ice_clean_rx_irq(struct ice_rx_ring *rx_ring, int budget) argument 1352 struct ice_rx_ring *rx_ring; local 1516 struct ice_rx_ring *rx_ring; local [all...] |
H A D | ice_xsk.c | 15 static struct xdp_buff **ice_xdp_buf(struct ice_rx_ring *rx_ring, u32 idx) argument 17 return &rx_ring->xdp_buf[idx]; 84 * @rx_ring: Rx ring that will have its IRQ disabled 88 ice_qvec_dis_irq(struct ice_vsi *vsi, struct ice_rx_ring *rx_ring, argument 99 reg = rx_ring->reg_idx; 123 struct ice_rx_ring *rx_ring; local 131 ice_for_each_rx_ring(rx_ring, q_vector->rx) 132 ice_cfg_rxq_interrupt(vsi, rx_ring->reg_idx, reg_idx, 165 struct ice_rx_ring *rx_ring; local 173 rx_ring 318 ice_realloc_rx_xdp_bufs(struct ice_rx_ring *rx_ring, bool pool_present) argument 351 struct ice_rx_ring *rx_ring; local 386 struct ice_rx_ring *rx_ring = vsi->rx_rings[qid]; local 471 __ice_alloc_rx_bufs_zc(struct ice_rx_ring *rx_ring, u16 count) argument 520 ice_alloc_rx_bufs_zc(struct ice_rx_ring *rx_ring, u16 count) argument 544 ice_construct_skb_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) argument 751 ice_run_xdp_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, struct bpf_prog *xdp_prog, struct ice_tx_ring *xdp_ring) argument 795 ice_add_xsk_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *first, struct xdp_buff *xdp, const unsigned int size) argument 830 ice_clean_rx_irq_zc(struct ice_rx_ring *rx_ring, int budget) argument 1145 ice_xsk_clean_rx_ring(struct ice_rx_ring *rx_ring) argument [all...] |
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_xsk.c | 9 void i40e_clear_rx_bi_zc(struct i40e_ring *rx_ring) argument 11 memset(rx_ring->rx_bi_zc, 0, 12 sizeof(*rx_ring->rx_bi_zc) * rx_ring->count); 15 static struct xdp_buff **i40e_rx_bi(struct i40e_ring *rx_ring, u32 idx) argument 17 return &rx_ring->rx_bi_zc[idx]; 22 * @rx_ring: Current rx ring 29 static int i40e_realloc_rx_xdp_bi(struct i40e_ring *rx_ring, bool pool_present) argument 31 size_t elem_size = pool_present ? sizeof(*rx_ring->rx_bi_zc) : 32 sizeof(*rx_ring 61 struct i40e_ring *rx_ring; local 197 i40e_run_xdp_zc(struct i40e_ring *rx_ring, struct xdp_buff *xdp, struct bpf_prog *xdp_prog) argument 240 i40e_alloc_rx_buffers_zc(struct i40e_ring *rx_ring, u16 count) argument 288 i40e_construct_skb_zc(struct i40e_ring *rx_ring, struct xdp_buff *xdp) argument 344 i40e_handle_xdp_result_zc(struct i40e_ring *rx_ring, struct xdp_buff *xdp_buff, union i40e_rx_desc *rx_desc, unsigned int *rx_packets, unsigned int *rx_bytes, unsigned int xdp_res, bool *failure) argument 400 i40e_add_xsk_frag(struct i40e_ring *rx_ring, struct xdp_buff *first, struct xdp_buff *xdp, const unsigned int size) argument 432 i40e_clean_rx_irq_zc(struct i40e_ring *rx_ring, int budget) argument 750 i40e_xsk_clean_rx_ring(struct i40e_ring *rx_ring) argument [all...] |
H A D | i40e_xsk.h | 34 bool i40e_alloc_rx_buffers_zc(struct i40e_ring *rx_ring, u16 cleaned_count); 35 int i40e_clean_rx_irq_zc(struct i40e_ring *rx_ring, int budget); 40 void i40e_clear_rx_bi_zc(struct i40e_ring *rx_ring);
|
H A D | i40e_txrx_common.h | 10 void i40e_clean_programming_status(struct i40e_ring *rx_ring, u64 qword0_raw, 12 void i40e_process_skb_fields(struct i40e_ring *rx_ring, 15 void i40e_update_rx_stats(struct i40e_ring *rx_ring, 18 void i40e_finalize_xdp_rx(struct i40e_ring *rx_ring, unsigned int xdp_res); 19 void i40e_release_rx_desc(struct i40e_ring *rx_ring, u32 val); 105 void i40e_xsk_clean_rx_ring(struct i40e_ring *rx_ring);
|
/linux-master/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_txrx_common.h | 18 bool ixgbe_cleanup_headers(struct ixgbe_ring *rx_ring, 21 void ixgbe_process_skb_fields(struct ixgbe_ring *rx_ring, 39 bool ixgbe_alloc_rx_buffers_zc(struct ixgbe_ring *rx_ring, u16 cleaned_count); 41 struct ixgbe_ring *rx_ring, 43 void ixgbe_xsk_clean_rx_ring(struct ixgbe_ring *rx_ring); 52 void ixgbe_update_rx_ring_stats(struct ixgbe_ring *rx_ring,
|
H A D | ixgbe_xsk.c | 98 struct ixgbe_ring *rx_ring, 107 xdp_prog = READ_ONCE(rx_ring->xdp_prog); 111 err = xdp_do_redirect(rx_ring->netdev, xdp, xdp_prog); 114 if (xsk_uses_need_wakeup(rx_ring->xsk_pool) && err == -ENOBUFS) 141 bpf_warn_invalid_xdp_action(rx_ring->netdev, xdp_prog, act); 146 trace_xdp_exception(rx_ring->netdev, xdp_prog, act); 151 bool ixgbe_alloc_rx_buffers_zc(struct ixgbe_ring *rx_ring, u16 count) argument 155 u16 i = rx_ring->next_to_use; 163 rx_desc = IXGBE_RX_DESC(rx_ring, i); 164 bi = &rx_ring 97 ixgbe_run_xdp_zc(struct ixgbe_adapter *adapter, struct ixgbe_ring *rx_ring, struct xdp_buff *xdp) argument 213 ixgbe_construct_skb_zc(struct ixgbe_ring *rx_ring, const struct xdp_buff *xdp) argument 239 ixgbe_inc_ntc(struct ixgbe_ring *rx_ring) argument 248 ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector, struct ixgbe_ring *rx_ring, const int budget) argument 376 ixgbe_xsk_clean_rx_ring(struct ixgbe_ring *rx_ring) argument [all...] |
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_xdp.c | 48 struct igc_ring *rx_ring, *tx_ring; local 76 rx_ring = adapter->rx_ring[queue_id]; 79 napi = &rx_ring->q_vector->napi; 82 igc_disable_rx_ring(rx_ring); 87 set_bit(IGC_RING_FLAG_AF_XDP_ZC, &rx_ring->flags); 92 igc_enable_rx_ring(rx_ring); 107 struct igc_ring *rx_ring, *tx_ring; local 122 rx_ring = adapter->rx_ring[queue_i [all...] |
H A D | igc_dump.c | 119 struct igc_ring *rx_ring; local 214 rx_ring = adapter->rx_ring[n]; 215 netdev_info(netdev, "%5d %5X %5X\n", n, rx_ring->next_to_use, 216 rx_ring->next_to_clean); 247 rx_ring = adapter->rx_ring[n]; 250 rx_ring->queue_index); 255 for (i = 0; i < rx_ring->count; i++) { 259 buffer_info = &rx_ring [all...] |
/linux-master/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_txrx.c | 688 * @rx_ring: ring to be cleaned 690 static void iavf_clean_rx_ring(struct iavf_ring *rx_ring) argument 696 if (!rx_ring->rx_bi) 699 if (rx_ring->skb) { 700 dev_kfree_skb(rx_ring->skb); 701 rx_ring->skb = NULL; 705 for (i = 0; i < rx_ring->count; i++) { 706 struct iavf_rx_buffer *rx_bi = &rx_ring->rx_bi[i]; 714 dma_sync_single_range_for_cpu(rx_ring->dev, 717 rx_ring 749 iavf_free_rx_resources(struct iavf_ring *rx_ring) argument 768 iavf_setup_rx_descriptors(struct iavf_ring *rx_ring) argument 810 iavf_release_rx_desc(struct iavf_ring *rx_ring, u32 val) argument 832 iavf_rx_offset(struct iavf_ring *rx_ring) argument 845 iavf_alloc_mapped_page(struct iavf_ring *rx_ring, struct iavf_rx_buffer *bi) argument 895 iavf_receive_skb(struct iavf_ring *rx_ring, struct sk_buff *skb, u16 vlan_tag) argument 917 iavf_alloc_rx_buffers(struct iavf_ring *rx_ring, u16 cleaned_count) argument 1118 iavf_process_skb_fields(struct iavf_ring *rx_ring, union iavf_rx_desc *rx_desc, struct sk_buff *skb, u8 rx_ptype) argument 1145 iavf_cleanup_headers(struct iavf_ring *rx_ring, struct sk_buff *skb) argument 1161 iavf_reuse_rx_page(struct iavf_ring *rx_ring, struct iavf_rx_buffer *old_buff) argument 1251 iavf_add_rx_frag(struct iavf_ring *rx_ring, struct iavf_rx_buffer *rx_buffer, struct sk_buff *skb, unsigned int size) argument 1284 iavf_get_rx_buffer(struct iavf_ring *rx_ring, const unsigned int size) argument 1317 iavf_construct_skb(struct iavf_ring *rx_ring, struct iavf_rx_buffer *rx_buffer, unsigned int size) argument 1381 iavf_build_skb(struct iavf_ring *rx_ring, struct iavf_rx_buffer *rx_buffer, unsigned int size) argument 1427 iavf_put_rx_buffer(struct iavf_ring *rx_ring, struct iavf_rx_buffer *rx_buffer) argument 1461 iavf_is_non_eop(struct iavf_ring *rx_ring, union iavf_rx_desc *rx_desc, struct sk_buff *skb) argument 1495 iavf_clean_rx_irq(struct iavf_ring *rx_ring, int budget) argument [all...] |
/linux-master/drivers/net/can/spi/mcp251xfd/ |
H A D | mcp251xfd-ring.c | 201 struct mcp251xfd_rx_ring *rx_ring; local 208 mcp251xfd_for_each_rx_ring(priv, rx_ring, i) { 209 rx_ring->head = 0; 210 rx_ring->tail = 0; 211 rx_ring->base = *base; 212 rx_ring->nr = i; 213 rx_ring->fifo_nr = *fifo_nr; 215 *base = mcp251xfd_get_rx_obj_addr(rx_ring, rx_ring->obj_num); 219 addr = MCP251XFD_REG_FIFOCON(rx_ring 289 const struct mcp251xfd_rx_ring *rx_ring; local 465 struct mcp251xfd_rx_ring *rx_ring; local [all...] |
/linux-master/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_ring.c | 52 static int aq_alloc_rxpages(struct aq_rxpage *rxpage, struct aq_ring_s *rx_ring) argument 54 struct device *dev = aq_nic_get_dev(rx_ring->aq_nic); 55 unsigned int order = rx_ring->page_order; 73 rxpage->pg_off = rx_ring->page_offset; 407 struct aq_ring_s *rx_ring, 417 u64_stats_update_begin(&rx_ring->stats.rx.syncp); 418 ++rx_ring->stats.rx.packets; 419 rx_ring->stats.rx.bytes += xdp_get_buff_len(xdp); 420 u64_stats_update_end(&rx_ring->stats.rx.syncp); 422 prog = READ_ONCE(rx_ring 405 aq_xdp_run_prog(struct aq_nic_s *aq_nic, struct xdp_buff *xdp, struct aq_ring_s *rx_ring, struct aq_ring_buff_s *buff) argument 698 __aq_ring_xdp_clean(struct aq_ring_s *rx_ring, struct napi_struct *napi, int *work_done, int budget) argument [all...] |
/linux-master/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_main.c | 75 static bool fm10k_alloc_mapped_page(struct fm10k_ring *rx_ring, argument 88 rx_ring->rx_stats.alloc_failed++; 93 dma = dma_map_page(rx_ring->dev, page, 0, PAGE_SIZE, DMA_FROM_DEVICE); 98 if (dma_mapping_error(rx_ring->dev, dma)) { 101 rx_ring->rx_stats.alloc_failed++; 114 * @rx_ring: ring to place buffers on 117 void fm10k_alloc_rx_buffers(struct fm10k_ring *rx_ring, u16 cleaned_count) argument 121 u16 i = rx_ring->next_to_use; 127 rx_desc = FM10K_RX_DESC(rx_ring, i); 128 bi = &rx_ring 183 fm10k_reuse_rx_page(struct fm10k_ring *rx_ring, struct fm10k_rx_buffer *old_buff) argument 299 fm10k_fetch_rx_buffer(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 411 fm10k_type_trans(struct fm10k_ring *rx_ring, union fm10k_rx_desc __maybe_unused *rx_desc, struct sk_buff *skb) argument 449 fm10k_process_skb_fields(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 490 fm10k_is_non_eop(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc) argument 521 fm10k_cleanup_headers(struct fm10k_ring *rx_ring, union fm10k_rx_desc *rx_desc, struct sk_buff *skb) argument 562 fm10k_clean_rx_irq(struct fm10k_q_vector *q_vector, struct fm10k_ring *rx_ring, int budget) argument [all...] |
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | bcm4908_enet.c | 86 struct bcm4908_enet_dma_ring rx_ring; member in struct:bcm4908_enet 199 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 203 size = rx_ring->length * sizeof(struct bcm4908_enet_dma_ring_bd); 204 if (rx_ring->cpu_addr) 205 dma_free_coherent(dev, size, rx_ring->cpu_addr, rx_ring->dma_addr); 206 kfree(rx_ring->slots); 217 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 321 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 338 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 452 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 502 struct bcm4908_enet_dma_ring *rx_ring = &enet->rx_ring; local 583 struct bcm4908_enet_dma_ring *rx_ring = container_of(napi, struct bcm4908_enet_dma_ring, napi); local [all...] |
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 163 static void wx_dma_sync_frag(struct wx_ring *rx_ring, argument 169 dma_sync_single_range_for_cpu(rx_ring->dev, 177 page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, false); 180 static struct wx_rx_buffer *wx_get_rx_buffer(struct wx_ring *rx_ring, argument 188 rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; 209 wx_dma_sync_frag(rx_ring, rx_buffer); 213 dma_sync_single_range_for_cpu(rx_ring->dev, 222 static void wx_put_rx_buffer(struct wx_ring *rx_ring, argument 236 static struct sk_buff *wx_build_skb(struct wx_ring *rx_ring, argument 303 wx_alloc_mapped_page(struct wx_ring *rx_ring, struct wx_rx_buffer *bi) argument 329 wx_alloc_rx_buffers(struct wx_ring *rx_ring, u16 cleaned_count) argument 407 wx_is_non_eop(struct wx_ring *rx_ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument 474 wx_cleanup_headers(struct wx_ring *rx_ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument 598 wx_process_skb_fields(struct wx_ring *rx_ring, union wx_rx_desc *rx_desc, struct sk_buff *skb) argument 622 wx_clean_rx_irq(struct wx_q_vector *q_vector, struct wx_ring *rx_ring, int budget) argument 2179 wx_clean_rx_ring(struct wx_ring *rx_ring) argument 2241 wx_free_rx_resources(struct wx_ring *rx_ring) argument 2394 wx_alloc_page_pool(struct wx_ring *rx_ring) argument 2424 wx_setup_rx_resources(struct wx_ring *rx_ring) argument [all...] |
/linux-master/drivers/net/ethernet/mscc/ |
H A D | ocelot_fdma.c | 58 struct ocelot_fdma_rx_ring *rx_ring = &fdma->rx_ring; local 60 if (rx_ring->next_to_use >= rx_ring->next_to_clean) 62 (rx_ring->next_to_use - rx_ring->next_to_clean) - 1; 64 return rx_ring->next_to_clean - rx_ring->next_to_use - 1; 149 struct ocelot_fdma_rx_ring *rx_ring; local 156 rx_ring 223 ocelot_fdma_rx_set_llp(struct ocelot_fdma_rx_ring *rx_ring) argument 237 struct ocelot_fdma_rx_ring *rx_ring; local 298 struct ocelot_fdma_rx_ring *rx_ring = &ocelot->fdma->rx_ring; local 392 struct ocelot_fdma_rx_ring *rx_ring; local 702 struct ocelot_fdma_rx_ring *rx_ring; local [all...] |