Searched refs:rxq (Results 1 - 25 of 171) sorted by relevance

1234567

/linux-master/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_rx.c83 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; local
84 struct dpmaif_bat_request *bat_req = rxq->bat_req;
87 if (!rxq->que_started) {
88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index);
235 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, argument
238 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info;
242 if (!rxq->que_started)
245 if (rel_entry_num >= rxq->pit_size_cnt) {
246 dev_err(rxq
275 t7xx_frag_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) argument
392 t7xx_dpmaif_set_frag_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct sk_buff *skb) argument
424 t7xx_dpmaif_get_frag(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, const struct dpmaif_cur_rx_skb_info *skb_info) argument
445 t7xx_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) argument
461 t7xx_dpmaif_check_pit_seq(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pit) argument
499 t7xx_dpmaif_release_bat_entry(const struct dpmaif_rx_queue *rxq, const unsigned int rel_entry_num, const enum bat_type buf_type) argument
555 t7xx_dpmaif_pit_release_and_add(struct dpmaif_rx_queue *rxq) argument
570 t7xx_dpmaif_bat_release_and_add(const struct dpmaif_rx_queue *rxq) argument
592 t7xx_dpmaif_frag_bat_release_and_add(const struct dpmaif_rx_queue *rxq) argument
610 t7xx_dpmaif_parse_msg_pit(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *msg_pit, struct dpmaif_cur_rx_skb_info *skb_info) argument
622 t7xx_dpmaif_set_data_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) argument
658 t7xx_dpmaif_get_rx_pkt(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) argument
679 t7xx_dpmaifq_rx_notify_hw(struct dpmaif_rx_queue *rxq) argument
693 t7xx_dpmaif_rx_skb(struct dpmaif_rx_queue *rxq, struct dpmaif_cur_rx_skb_info *skb_info) argument
717 t7xx_dpmaif_rx_start(struct dpmaif_rx_queue *rxq, const unsigned int pit_cnt, const unsigned int budget, int *once_more) argument
798 t7xx_dpmaifq_poll_pit(struct dpmaif_rx_queue *rxq) argument
816 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; local
833 struct dpmaif_rx_queue *rxq = container_of(napi, struct dpmaif_rx_queue, napi); local
891 struct dpmaif_rx_queue *rxq; local
1006 t7xx_dpmaif_rx_alloc(struct dpmaif_rx_queue *rxq) argument
1030 t7xx_dpmaif_rx_buf_free(const struct dpmaif_rx_queue *rxq) argument
1063 struct dpmaif_rx_queue *rxq; local
1116 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[i]; local
1130 t7xx_dpmaif_stop_rxq(struct dpmaif_rx_queue *rxq) argument
[all...]
H A Dt7xx_hif_dpmaif.c227 rx_q = &dpmaif_ctrl->rxq[rx_idx];
267 rx_q = &dpmaif_ctrl->rxq[i];
294 rx_q = &dpmaif_ctrl->rxq[i];
303 struct dpmaif_rx_queue *rxq; local
314 rxq = &dpmaif_ctrl->rxq[i];
315 rxq->que_started = true;
316 rxq->index = i;
317 rxq->budget = rxq
432 struct dpmaif_rx_queue *rxq; local
[all...]
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_rx.c51 * @rxq: Logical Rx Queue
53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) argument
55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats;
68 * @rxq: Logical Rx Queue
71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) argument
73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats;
89 * @rxq: Logical Rx Queue
91 static void rxq_stats_init(struct hinic_rxq *rxq) argument
93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats;
96 hinic_rxq_clean_stats(rxq);
99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) argument
127 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) argument
163 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) argument
180 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) argument
193 rx_alloc_pkts(struct hinic_rxq *rxq) argument
243 free_all_rx_skbs(struct hinic_rxq *rxq) argument
271 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) argument
354 rxq_recv(struct hinic_rxq *rxq, int budget) argument
459 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); local
478 rx_add_napi(struct hinic_rxq *rxq) argument
487 rx_del_napi(struct hinic_rxq *rxq) argument
495 struct hinic_rxq *rxq = (struct hinic_rxq *)data; local
513 rx_request_irq(struct hinic_rxq *rxq) argument
563 rx_free_irq(struct hinic_rxq *rxq) argument
580 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) argument
623 hinic_clean_rxq(struct hinic_rxq *rxq) argument
[all...]
H A Dhinic_rx.h44 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats);
46 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq,
49 void hinic_clean_rxq(struct hinic_rxq *rxq);
/linux-master/drivers/net/wireless/intel/iwlwifi/pcie/
H A Drx.c55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free.
58 * or reused - added immediately to the iwl->rxq->rx_free list.
66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them
76 * detached from the iwl->rxq. The driver 'processed' index is updated.
77 * + If there are no allocated buffers in iwl->rxq->rx_free,
104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue
108 * rxq
119 iwl_rxq_space(const struct iwl_rxq *rxq) argument
166 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
220 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) argument
244 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
291 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
349 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
429 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) argument
613 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
678 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
713 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
757 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
807 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
835 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) argument
991 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) argument
1005 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); local
1032 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); local
1071 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
1108 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
1229 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local
1246 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) argument
1260 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) argument
1292 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) argument
1441 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) argument
1499 struct iwl_rxq *rxq; local
1642 struct iwl_rxq *rxq; local
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dnetdev_rx.c194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq),
197 if (!rx->rxq) {
203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local
205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd);
209 hfi1_rcd_get(rxq->rcd);
210 rxq->rx = rx;
211 rxq->rcd->napi = &rxq
230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local
250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local
268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local
286 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local
[all...]
H A Dnetdev.h41 * @rxq: pointer to dummy netdev receive queues.
54 struct hfi1_netdev_rxq *rxq; member in struct:hfi1_netdev_rx
73 return dd->netdev_rx->rxq[ctxt].rcd;
H A Dvnic_main.c292 static inline int hfi1_vnic_decap_skb(struct hfi1_vnic_rx_queue *rxq, argument
295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo;
303 vinfo->stats[rxq->idx].rx_oversize++;
305 vinfo->stats[rxq->idx].rx_runt++;
336 struct hfi1_vnic_rx_queue *rxq; local
370 rxq = &vinfo->rxq[q_idx];
389 rc = hfi1_vnic_decap_skb(rxq, skb);
392 hfi1_vnic_update_rx_counters(vinfo, rxq->idx, skb, rc);
399 skb->protocol = eth_type_trans(skb, rxq
599 struct hfi1_vnic_rx_queue *rxq = &vinfo->rxq[i]; local
[all...]
H A Dipoib_rx.c46 struct sk_buff *hfi1_ipoib_prepare_skb(struct hfi1_netdev_rxq *rxq, argument
49 struct napi_struct *napi = &rxq->napi;
/linux-master/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c27 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) argument
38 if (allow_lazy && likely(rxq->filled_buffers > 12)) {
39 rxq->filled_buffers--;
50 mapping = dma_map_page(rxq->dev, data, 0,
51 PAGE_SIZE, rxq->data_direction);
52 if (unlikely(dma_mapping_error(rxq->dev, mapping))) {
57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX];
63 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring);
67 rxq
502 qede_has_rx_work(struct qede_rx_queue *rxq) argument
515 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) argument
524 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) argument
547 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) argument
558 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) argument
588 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) argument
637 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) argument
664 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) argument
740 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) argument
759 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) argument
790 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) argument
831 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) argument
957 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) argument
976 struct qede_rx_queue *rxq = fp->rxq; local
1079 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) argument
1169 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) argument
1223 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) argument
1243 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) argument
1341 struct qede_rx_queue *rxq = fp->rxq; local
[all...]
H A Dqede_main.c964 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq))
965 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq);
966 kfree(fp->rxq);
1035 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL);
1036 if (!fp->rxq)
1528 struct qede_rx_queue *rxq)
1532 for (i = rxq->sw_rx_cons; i != rxq
1527 qede_free_rx_buffers(struct qede_dev *edev, struct qede_rx_queue *rxq) argument
1547 qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) argument
1560 qede_set_tpa_param(struct qede_rx_queue *rxq) argument
1572 qede_alloc_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) argument
2287 struct qede_rx_queue *rxq = fp->rxq; local
[all...]
/linux-master/include/net/
H A Dnetdev_rx_queue.h41 __netif_get_rx_queue(struct net_device *dev, unsigned int rxq) argument
43 return dev->_rx + rxq;
/linux-master/tools/testing/selftests/bpf/progs/
H A Dtest_xdp_bpf2bpf.c29 struct xdp_rxq_info *rxq; member in struct:xdp_buff
49 meta.ifindex = xdp->rxq->dev->ifindex;
56 test_result_fentry = xdp->rxq->dev->ifindex;
/linux-master/drivers/net/ethernet/marvell/
H A Dmvneta.c109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3))
134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq)
857 struct mvneta_rx_queue *rxq,
864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id),
870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id),
876 struct mvneta_rx_queue *rxq)
880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id));
888 struct mvneta_rx_queue *rxq,
856 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) argument
875 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
887 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) argument
922 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) argument
945 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) argument
1002 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) argument
1017 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
1028 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
1039 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
1052 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local
1498 int rxq, txq; local
1685 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) argument
1695 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) argument
1722 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) argument
1935 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) argument
1986 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
2041 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
2072 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len) argument
2257 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) argument
2314 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) argument
2349 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) argument
2425 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) argument
2538 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) argument
3346 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) argument
3390 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) argument
3439 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
3456 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
3489 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
3505 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument
4343 int rxq; local
4683 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local
5145 mvneta_map_vlan_prio_to_rxq(struct mvneta_port *pp, u8 pri, u8 rxq) argument
5222 int rxq, txq, tc, ret; local
5369 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local
5788 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local
5840 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local
[all...]
H A Dmv643xx_eth.c399 struct rx_queue rxq[8]; member in struct:mv643xx_eth_private
440 /* rxq/txq helper functions *************************************************/
441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) argument
443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]);
451 static void rxq_enable(struct rx_queue *rxq) argument
453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index);
457 static void rxq_disable(struct rx_queue *rxq) argument
506 rxq_process(struct rx_queue *rxq, int budget) argument
598 rxq_refill(struct rx_queue *rxq, int budget) argument
1940 struct rx_queue *rxq = mp->rxq + index; local
2005 rxq_deinit(struct rx_queue *rxq) argument
2383 struct rx_queue *rxq = mp->rxq + i; local
[all...]
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c548 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) argument
550 struct net_device *ndev = rxq->ndev;
562 if (mpc->rxbpre_datasize != rxq->datasize) {
564 mpc->rxbpre_datasize, rxq->datasize);
568 if (mpc->rxbpre_alloc_size != rxq->alloc_size) {
570 mpc->rxbpre_alloc_size, rxq->alloc_size);
574 if (mpc->rxbpre_headroom != rxq->headroom) {
576 mpc->rxbpre_headroom, rxq->headroom);
1259 static int mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) argument
1265 init_completion(&rxq
1301 struct mana_rxq *rxq; local
1462 mana_post_pkt_rxq(struct mana_rxq *rxq) argument
1482 mana_build_skb(struct mana_rxq *rxq, void *buf_va, uint pkt_len, struct xdp_buff *xdp) argument
1502 mana_rx_skb(void *buf_va, bool from_pool, struct mana_rxcomp_oob *cqe, struct mana_rxq *rxq) argument
1603 mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev, dma_addr_t *da, bool *from_pool, bool is_napi) argument
1654 mana_refill_rx_oob(struct device *dev, struct mana_rxq *rxq, struct mana_recv_buf_oob *rxoob, void **old_buf, bool *old_fp) argument
1676 mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq, struct gdma_comp *cqe) argument
1746 struct mana_rxq *rxq = cq->rxq; local
1987 mana_destroy_rxq(struct mana_port_context *apc, struct mana_rxq *rxq, bool validate_state) argument
2046 mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key, struct mana_rxq *rxq, struct device *dev) argument
2075 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) argument
2115 mana_push_wqe(struct mana_rxq *rxq) argument
2133 mana_create_page_pool(struct mana_rxq *rxq, struct gdma_context *gc) argument
2165 struct mana_rxq *rxq; local
2285 struct mana_rxq *rxq; local
2309 struct mana_rxq *rxq; local
[all...]
H A Dmana_bpf.c80 u32 mana_run_xdp(struct net_device *ndev, struct mana_rxq *rxq, argument
88 prog = rcu_dereference(rxq->bpf_prog);
93 xdp_init_buff(xdp, PAGE_SIZE, &rxq->xdp_rxq);
98 rx_stats = &rxq->stats;
107 rxq->xdp_rc = xdp_do_redirect(ndev, xdp, prog);
108 if (!rxq->xdp_rc) {
109 rxq->xdp_flush = true;
/linux-master/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_trace.h75 TP_PROTO(const struct funeth_rxq *rxq,
81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec),
90 __string(devname, rxq->netdev->name)
94 __entry->qidx = rxq->qidx;
95 __entry->cq_head = rxq->cq_head;
100 __assign_str(devname, rxq->netdev->name);
/linux-master/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c332 * @rxq: the queue that owns the buffer
335 static void idpf_rx_page_rel(struct idpf_queue *rxq, struct idpf_rx_buf *rx_buf) argument
340 page_pool_put_full_page(rxq->pp, rx_buf->page, false);
348 * @rxq: queue to use
350 static void idpf_rx_hdr_buf_rel_all(struct idpf_queue *rxq) argument
352 struct idpf_adapter *adapter = rxq->vport->adapter;
355 rxq->desc_count * IDPF_HDR_BUF_SIZE,
356 rxq->rx_buf.hdr_buf_va,
357 rxq->rx_buf.hdr_buf_pa);
358 rxq
365 idpf_rx_buf_rel_all(struct idpf_queue *rxq) argument
395 idpf_rx_desc_rel(struct idpf_queue *rxq, bool bufq, s32 q_model) argument
466 idpf_rx_buf_hw_update(struct idpf_queue *rxq, u32 val) argument
483 idpf_rx_hdr_buf_alloc_all(struct idpf_queue *rxq) argument
714 idpf_rx_desc_alloc(struct idpf_queue *rxq, bool bufq, s32 q_model) argument
2738 idpf_rx_hash(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc, struct idpf_rx_ptype_decoded *decoded) argument
2763 idpf_rx_csum(struct idpf_queue *rxq, struct sk_buff *skb, struct idpf_rx_csum_decoded *csum_bits, struct idpf_rx_ptype_decoded *decoded) argument
2863 idpf_rx_rsc(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc, struct idpf_rx_ptype_decoded *decoded) argument
2933 idpf_rx_process_skb_fields(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc) argument
2995 idpf_rx_construct_skb(struct idpf_queue *rxq, struct idpf_rx_buf *rx_buf, unsigned int size) argument
3056 idpf_rx_hdr_construct_skb(struct idpf_queue *rxq, const void *va, unsigned int size) argument
3119 idpf_rx_splitq_clean(struct idpf_queue *rxq, int budget) argument
3602 struct idpf_queue *rxq = q_vector->rx[i]; local
3886 struct idpf_queue *rxq = q_vec->rx[i]; local
[all...]
H A Didpf_txrx.h78 #define IDPF_RX_BUFQ_WORKING_SET(rxq) ((rxq)->desc_count - 1)
80 #define IDPF_RX_BUMP_NTC(rxq, ntc) \
82 if (unlikely(++(ntc) == (rxq)->desc_count)) { \
84 change_bit(__IDPF_Q_GEN_CHK, (rxq)->flags); \
118 #define IDPF_SINGLEQ_RX_BUF_DESC(rxq, i) \
119 (&(((struct virtchnl2_singleq_rx_buf_desc *)((rxq)->desc_ring))[i]))
120 #define IDPF_SPLITQ_RX_BUF_DESC(rxq, i) \
121 (&(((struct virtchnl2_splitq_rx_buf_desc *)((rxq)->desc_ring))[i]))
122 #define IDPF_SPLITQ_RX_BI_DESC(rxq,
785 struct idpf_queue rxq; member in struct:idpf_rxq_set
[all...]
/linux-master/drivers/net/ethernet/hisilicon/
H A Dhisi_femac.c121 struct hisi_femac_queue rxq; member in struct:hisi_femac_priv
212 struct hisi_femac_queue *rxq = &priv->rxq; local
218 pos = rxq->head;
220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num))
222 if (unlikely(rxq->skb[pos])) {
224 pos, rxq->skb[pos]);
237 rxq->dma_phys[pos] = addr;
238 rxq
248 struct hisi_femac_queue *rxq = &priv->rxq; local
388 struct hisi_femac_queue *rxq = &priv->rxq; local
[all...]
/linux-master/tools/testing/selftests/bpf/
H A Dxdp_hw_metadata.c63 int rxq; variable
405 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) argument
408 struct pollfd fds[rxq + 1];
415 for (i = 0; i < rxq; i++) {
421 fds[rxq].fd = server_fd;
422 fds[rxq].events = POLLIN;
423 fds[rxq].revents = 0;
428 for (i = 0; i < rxq; i++) {
434 ret = poll(fds, rxq + 1, 1000);
443 if (fds[rxq]
[all...]
/linux-master/drivers/net/ethernet/atheros/alx/
H A Dmain.c73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; local
79 next = cur = rxq->write_idx;
82 cur_buf = &rxq->bufs[cur];
84 while (!cur_buf->skb && next != rxq->read_idx) {
85 struct alx_rfd *rfd = &rxq->rfd[cur];
127 cur_buf = &rxq->bufs[cur];
134 rxq->write_idx = cur;
216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) argument
225 alx = netdev_priv(rxq
506 alx_free_rxring_buf(struct alx_rx_queue *rxq) argument
630 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) argument
741 struct alx_rx_queue *rxq; local
[all...]
/linux-master/drivers/vhost/
H A Dnet.c127 struct vhost_net_buf rxq; member in struct:vhost_net_virtqueue
150 static void *vhost_net_buf_get_ptr(struct vhost_net_buf *rxq) argument
152 if (rxq->tail != rxq->head)
153 return rxq->queue[rxq->head];
158 static int vhost_net_buf_get_size(struct vhost_net_buf *rxq) argument
160 return rxq->tail - rxq->head;
163 static int vhost_net_buf_is_empty(struct vhost_net_buf *rxq) argument
168 vhost_net_buf_consume(struct vhost_net_buf *rxq) argument
177 struct vhost_net_buf *rxq = &nvq->rxq; local
187 struct vhost_net_buf *rxq = &nvq->rxq; local
210 struct vhost_net_buf *rxq = &nvq->rxq; local
222 vhost_net_buf_init(struct vhost_net_buf *rxq) argument
[all...]
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dcxgb4vf_main.c383 int rxq, msi, err; local
397 for_each_ethrxq(s, rxq) {
401 &s->ethrxq[rxq].rspq);
409 while (--rxq >= 0)
410 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq);
421 int rxq, msi; local
425 for_each_ethrxq(s, rxq)
427 &s->ethrxq[rxq].rspq);
452 int rxq; local
455 for_each_ethrxq(s, rxq)
477 int rxq; local
632 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local
663 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local
705 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local
1236 struct sge_eth_rxq *rxq; local
1810 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local
2062 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local
2202 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local
2729 struct sge_eth_rxq *rxq = &s->ethrxq[qs]; local
[all...]

Completed in 246 milliseconds

1234567