/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_rx.c | 83 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; local 84 struct dpmaif_bat_request *bat_req = rxq->bat_req; 87 if (!rxq->que_started) { 88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); 235 static int t7xx_dpmaifq_release_pit_entry(struct dpmaif_rx_queue *rxq, argument 238 struct dpmaif_hw_info *hw_info = &rxq->dpmaif_ctrl->hw_info; 242 if (!rxq->que_started) 245 if (rel_entry_num >= rxq->pit_size_cnt) { 246 dev_err(rxq 275 t7xx_frag_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) argument 392 t7xx_dpmaif_set_frag_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct sk_buff *skb) argument 424 t7xx_dpmaif_get_frag(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, const struct dpmaif_cur_rx_skb_info *skb_info) argument 445 t7xx_bat_cur_bid_check(struct dpmaif_rx_queue *rxq, const unsigned int cur_bid) argument 461 t7xx_dpmaif_check_pit_seq(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pit) argument 499 t7xx_dpmaif_release_bat_entry(const struct dpmaif_rx_queue *rxq, const unsigned int rel_entry_num, const enum bat_type buf_type) argument 555 t7xx_dpmaif_pit_release_and_add(struct dpmaif_rx_queue *rxq) argument 570 t7xx_dpmaif_bat_release_and_add(const struct dpmaif_rx_queue *rxq) argument 592 t7xx_dpmaif_frag_bat_release_and_add(const struct dpmaif_rx_queue *rxq) argument 610 t7xx_dpmaif_parse_msg_pit(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *msg_pit, struct dpmaif_cur_rx_skb_info *skb_info) argument 622 t7xx_dpmaif_set_data_to_skb(const struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) argument 658 t7xx_dpmaif_get_rx_pkt(struct dpmaif_rx_queue *rxq, const struct dpmaif_pit *pkt_info, struct dpmaif_cur_rx_skb_info *skb_info) argument 679 t7xx_dpmaifq_rx_notify_hw(struct dpmaif_rx_queue *rxq) argument 693 t7xx_dpmaif_rx_skb(struct dpmaif_rx_queue *rxq, struct dpmaif_cur_rx_skb_info *skb_info) argument 717 t7xx_dpmaif_rx_start(struct dpmaif_rx_queue *rxq, const unsigned int pit_cnt, const unsigned int budget, int *once_more) argument 798 t7xx_dpmaifq_poll_pit(struct dpmaif_rx_queue *rxq) argument 816 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; local 833 struct dpmaif_rx_queue *rxq = container_of(napi, struct dpmaif_rx_queue, napi); local 891 struct dpmaif_rx_queue *rxq; local 1006 t7xx_dpmaif_rx_alloc(struct dpmaif_rx_queue *rxq) argument 1030 t7xx_dpmaif_rx_buf_free(const struct dpmaif_rx_queue *rxq) argument 1063 struct dpmaif_rx_queue *rxq; local 1116 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[i]; local 1130 t7xx_dpmaif_stop_rxq(struct dpmaif_rx_queue *rxq) argument [all...] |
H A D | t7xx_hif_dpmaif.c | 227 rx_q = &dpmaif_ctrl->rxq[rx_idx]; 267 rx_q = &dpmaif_ctrl->rxq[i]; 294 rx_q = &dpmaif_ctrl->rxq[i]; 303 struct dpmaif_rx_queue *rxq; local 314 rxq = &dpmaif_ctrl->rxq[i]; 315 rxq->que_started = true; 316 rxq->index = i; 317 rxq->budget = rxq 432 struct dpmaif_rx_queue *rxq; local [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_rx.c | 51 * @rxq: Logical Rx Queue 53 static void hinic_rxq_clean_stats(struct hinic_rxq *rxq) argument 55 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; 68 * @rxq: Logical Rx Queue 71 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats) argument 73 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; 89 * @rxq: Logical Rx Queue 91 static void rxq_stats_init(struct hinic_rxq *rxq) argument 93 struct hinic_rxq_stats *rxq_stats = &rxq->rxq_stats; 96 hinic_rxq_clean_stats(rxq); 99 rx_csum(struct hinic_rxq *rxq, u32 status, struct sk_buff *skb) argument 127 rx_alloc_skb(struct hinic_rxq *rxq, dma_addr_t *dma_addr) argument 163 rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr) argument 180 rx_free_skb(struct hinic_rxq *rxq, struct sk_buff *skb, dma_addr_t dma_addr) argument 193 rx_alloc_pkts(struct hinic_rxq *rxq) argument 243 free_all_rx_skbs(struct hinic_rxq *rxq) argument 271 rx_recv_jumbo_pkt(struct hinic_rxq *rxq, struct sk_buff *head_skb, unsigned int left_pkt_len, u16 ci) argument 354 rxq_recv(struct hinic_rxq *rxq, int budget) argument 459 struct hinic_rxq *rxq = container_of(napi, struct hinic_rxq, napi); local 478 rx_add_napi(struct hinic_rxq *rxq) argument 487 rx_del_napi(struct hinic_rxq *rxq) argument 495 struct hinic_rxq *rxq = (struct hinic_rxq *)data; local 513 rx_request_irq(struct hinic_rxq *rxq) argument 563 rx_free_irq(struct hinic_rxq *rxq) argument 580 hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, struct net_device *netdev) argument 623 hinic_clean_rxq(struct hinic_rxq *rxq) argument [all...] |
H A D | hinic_rx.h | 44 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats); 46 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, 49 void hinic_clean_rxq(struct hinic_rxq *rxq);
|
/linux-master/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | rx.c | 55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free. 58 * or reused - added immediately to the iwl->rxq->rx_free list. 66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them 76 * detached from the iwl->rxq. The driver 'processed' index is updated. 77 * + If there are no allocated buffers in iwl->rxq->rx_free, 104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue 108 * rxq 119 iwl_rxq_space(const struct iwl_rxq *rxq) argument 166 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 220 iwl_pcie_restock_bd(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb) argument 244 iwl_pcie_rxmq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 291 iwl_pcie_rxsq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 349 iwl_pcie_rxq_restock(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 429 iwl_pcie_rxq_alloc_rbs(struct iwl_trans *trans, gfp_t priority, struct iwl_rxq *rxq) argument 613 iwl_pcie_rx_allocator_get(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 678 iwl_pcie_free_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 713 iwl_pcie_alloc_rxq_dma(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 757 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 807 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 835 iwl_pcie_rx_hw_init(struct iwl_trans *trans, struct iwl_rxq *rxq) argument 991 iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq) argument 1005 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); local 1032 struct iwl_rxq *rxq = container_of(napi, struct iwl_rxq, napi); local 1071 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 1108 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 1229 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; local 1246 iwl_pcie_rx_move_to_allocator(struct iwl_rxq *rxq, struct iwl_rb_allocator *rba) argument 1260 iwl_pcie_rx_reuse_rbd(struct iwl_trans *trans, struct iwl_rx_mem_buffer *rxb, struct iwl_rxq *rxq, bool emergency) argument 1292 iwl_pcie_rx_handle_rb(struct iwl_trans *trans, struct iwl_rxq *rxq, struct iwl_rx_mem_buffer *rxb, bool emergency, int i) argument 1441 iwl_pcie_get_rxb(struct iwl_trans *trans, struct iwl_rxq *rxq, int i, bool *join) argument 1499 struct iwl_rxq *rxq; local 1642 struct iwl_rxq *rxq; local [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | netdev_rx.c | 194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), 197 if (!rx->rxq) { 203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local 205 rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd); 209 hfi1_rcd_get(rxq->rcd); 210 rxq->rx = rx; 211 rxq->rcd->napi = &rxq 230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local 250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local 268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local 286 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; local [all...] |
H A D | netdev.h | 41 * @rxq: pointer to dummy netdev receive queues. 54 struct hfi1_netdev_rxq *rxq; member in struct:hfi1_netdev_rx 73 return dd->netdev_rx->rxq[ctxt].rcd;
|
H A D | vnic_main.c | 292 static inline int hfi1_vnic_decap_skb(struct hfi1_vnic_rx_queue *rxq, argument 295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; 303 vinfo->stats[rxq->idx].rx_oversize++; 305 vinfo->stats[rxq->idx].rx_runt++; 336 struct hfi1_vnic_rx_queue *rxq; local 370 rxq = &vinfo->rxq[q_idx]; 389 rc = hfi1_vnic_decap_skb(rxq, skb); 392 hfi1_vnic_update_rx_counters(vinfo, rxq->idx, skb, rc); 399 skb->protocol = eth_type_trans(skb, rxq 599 struct hfi1_vnic_rx_queue *rxq = &vinfo->rxq[i]; local [all...] |
H A D | ipoib_rx.c | 46 struct sk_buff *hfi1_ipoib_prepare_skb(struct hfi1_netdev_rxq *rxq, argument 49 struct napi_struct *napi = &rxq->napi;
|
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 27 int qede_alloc_rx_buffer(struct qede_rx_queue *rxq, bool allow_lazy) argument 38 if (allow_lazy && likely(rxq->filled_buffers > 12)) { 39 rxq->filled_buffers--; 50 mapping = dma_map_page(rxq->dev, data, 0, 51 PAGE_SIZE, rxq->data_direction); 52 if (unlikely(dma_mapping_error(rxq->dev, mapping))) { 57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; 63 rx_bd = (struct eth_rx_bd *)qed_chain_produce(&rxq->rx_bd_ring); 67 rxq 502 qede_has_rx_work(struct qede_rx_queue *rxq) argument 515 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) argument 524 qede_reuse_page(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) argument 547 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, u8 count) argument 558 qede_realloc_rx_buffer(struct qede_rx_queue *rxq, struct sw_rx_data *curr_cons) argument 588 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) argument 637 qede_skb_receive(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct sk_buff *skb, u16 vlan_tag) argument 664 qede_fill_frag_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, u8 tpa_agg_index, u16 len_on_bd) argument 740 qede_build_skb(struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) argument 759 qede_tpa_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad, bool alloc_skb) argument 790 qede_rx_build_skb(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sw_rx_data *bd, u16 len, u16 pad) argument 831 qede_tpa_start(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_start_cqe *cqe) argument 957 qede_tpa_cont(struct qede_dev *edev, struct qede_rx_queue *rxq, struct eth_fast_path_rx_tpa_cont_cqe *cqe) argument 976 struct qede_rx_queue *rxq = fp->rxq; local 1079 qede_rx_xdp(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, struct bpf_prog *prog, struct sw_rx_data *bd, struct eth_fast_path_rx_reg_cqe *cqe, u16 *data_offset, u16 *len) argument 1169 qede_rx_build_jumbo(struct qede_dev *edev, struct qede_rx_queue *rxq, struct sk_buff *skb, struct eth_fast_path_rx_reg_cqe *cqe, u16 first_bd_len) argument 1223 qede_rx_process_tpa_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq, union eth_rx_cqe *cqe, enum eth_rx_cqe_type type) argument 1243 qede_rx_process_cqe(struct qede_dev *edev, struct qede_fastpath *fp, struct qede_rx_queue *rxq) argument 1341 struct qede_rx_queue *rxq = fp->rxq; local [all...] |
H A D | qede_main.c | 964 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) 965 xdp_rxq_info_unreg(&fp->rxq->xdp_rxq); 966 kfree(fp->rxq); 1035 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); 1036 if (!fp->rxq) 1528 struct qede_rx_queue *rxq) 1532 for (i = rxq->sw_rx_cons; i != rxq 1527 qede_free_rx_buffers(struct qede_dev *edev, struct qede_rx_queue *rxq) argument 1547 qede_free_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) argument 1560 qede_set_tpa_param(struct qede_rx_queue *rxq) argument 1572 qede_alloc_mem_rxq(struct qede_dev *edev, struct qede_rx_queue *rxq) argument 2287 struct qede_rx_queue *rxq = fp->rxq; local [all...] |
/linux-master/include/net/ |
H A D | netdev_rx_queue.h | 41 __netif_get_rx_queue(struct net_device *dev, unsigned int rxq) argument 43 return dev->_rx + rxq;
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_xdp_bpf2bpf.c | 29 struct xdp_rxq_info *rxq; member in struct:xdp_buff 49 meta.ifindex = xdp->rxq->dev->ifindex; 56 test_result_fentry = xdp->rxq->dev->ifindex;
|
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 109 #define MVNETA_VLAN_PRIO_RXQ_MAP(prio, rxq) ((rxq) << ((prio) * 3)) 134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) 857 struct mvneta_rx_queue *rxq, 864 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), 870 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), 876 struct mvneta_rx_queue *rxq) 880 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); 888 struct mvneta_rx_queue *rxq, 856 mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int ndescs) argument 875 mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 887 mvneta_rxq_desc_num_update(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int rx_done, int rx_filled) argument 922 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) argument 945 mvneta_rxq_offset_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int offset) argument 1002 mvneta_rxq_buf_size_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int buf_size) argument 1017 mvneta_rxq_bm_disable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 1028 mvneta_rxq_bm_enable(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 1039 mvneta_rxq_long_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 1052 mvneta_rxq_short_pool_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local 1498 int rxq, txq; local 1685 mvneta_rx_pkts_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) argument 1695 mvneta_rx_time_coal_set(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, u32 value) argument 1722 mvneta_rx_desc_fill(struct mvneta_rx_desc *rx_desc, u32 phys_addr, void *virt_addr, struct mvneta_rx_queue *rxq) argument 1935 mvneta_rx_refill(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, gfp_t gfp_mask) argument 1986 mvneta_rxq_drop_pkts(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 2041 mvneta_rx_refill_queue(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 2072 mvneta_xdp_put_buff(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int sync_len) argument 2257 mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, struct bpf_prog *prog, struct xdp_buff *xdp, u32 frame_sz, struct mvneta_stats *stats) argument 2314 mvneta_swbm_rx_frame(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) argument 2349 mvneta_swbm_add_rx_fragment(struct mvneta_port *pp, struct mvneta_rx_desc *rx_desc, struct mvneta_rx_queue *rxq, struct xdp_buff *xdp, int *size, struct page *page) argument 2425 mvneta_rx_swbm(struct napi_struct *napi, struct mvneta_port *pp, int budget, struct mvneta_rx_queue *rxq) argument 2538 mvneta_rx_hwbm(struct napi_struct *napi, struct mvneta_port *pp, int rx_todo, struct mvneta_rx_queue *rxq) argument 3346 mvneta_create_page_pool(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int size) argument 3390 mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq, int num) argument 3439 mvneta_rxq_sw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 3456 mvneta_rxq_hw_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 3489 mvneta_rxq_init(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 3505 mvneta_rxq_deinit(struct mvneta_port *pp, struct mvneta_rx_queue *rxq) argument 4343 int rxq; local 4683 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local 5145 mvneta_map_vlan_prio_to_rxq(struct mvneta_port *pp, u8 pri, u8 rxq) argument 5222 int rxq, txq, tc, ret; local 5369 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local 5788 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local 5840 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; local [all...] |
H A D | mv643xx_eth.c | 399 struct rx_queue rxq[8]; member in struct:mv643xx_eth_private 440 /* rxq/txq helper functions *************************************************/ 441 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) argument 443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); 451 static void rxq_enable(struct rx_queue *rxq) argument 453 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); 454 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); 457 static void rxq_disable(struct rx_queue *rxq) argument 506 rxq_process(struct rx_queue *rxq, int budget) argument 598 rxq_refill(struct rx_queue *rxq, int budget) argument 1940 struct rx_queue *rxq = mp->rxq + index; local 2005 rxq_deinit(struct rx_queue *rxq) argument 2383 struct rx_queue *rxq = mp->rxq + i; local [all...] |
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_en.c | 548 static void *mana_get_rxbuf_pre(struct mana_rxq *rxq, dma_addr_t *da) argument 550 struct net_device *ndev = rxq->ndev; 562 if (mpc->rxbpre_datasize != rxq->datasize) { 564 mpc->rxbpre_datasize, rxq->datasize); 568 if (mpc->rxbpre_alloc_size != rxq->alloc_size) { 570 mpc->rxbpre_alloc_size, rxq->alloc_size); 574 if (mpc->rxbpre_headroom != rxq->headroom) { 576 mpc->rxbpre_headroom, rxq->headroom); 1259 static int mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) argument 1265 init_completion(&rxq 1301 struct mana_rxq *rxq; local 1462 mana_post_pkt_rxq(struct mana_rxq *rxq) argument 1482 mana_build_skb(struct mana_rxq *rxq, void *buf_va, uint pkt_len, struct xdp_buff *xdp) argument 1502 mana_rx_skb(void *buf_va, bool from_pool, struct mana_rxcomp_oob *cqe, struct mana_rxq *rxq) argument 1603 mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev, dma_addr_t *da, bool *from_pool, bool is_napi) argument 1654 mana_refill_rx_oob(struct device *dev, struct mana_rxq *rxq, struct mana_recv_buf_oob *rxoob, void **old_buf, bool *old_fp) argument 1676 mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq, struct gdma_comp *cqe) argument 1746 struct mana_rxq *rxq = cq->rxq; local 1987 mana_destroy_rxq(struct mana_port_context *apc, struct mana_rxq *rxq, bool validate_state) argument 2046 mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key, struct mana_rxq *rxq, struct device *dev) argument 2075 mana_alloc_rx_wqe(struct mana_port_context *apc, struct mana_rxq *rxq, u32 *rxq_size, u32 *cq_size) argument 2115 mana_push_wqe(struct mana_rxq *rxq) argument 2133 mana_create_page_pool(struct mana_rxq *rxq, struct gdma_context *gc) argument 2165 struct mana_rxq *rxq; local 2285 struct mana_rxq *rxq; local 2309 struct mana_rxq *rxq; local [all...] |
H A D | mana_bpf.c | 80 u32 mana_run_xdp(struct net_device *ndev, struct mana_rxq *rxq, argument 88 prog = rcu_dereference(rxq->bpf_prog); 93 xdp_init_buff(xdp, PAGE_SIZE, &rxq->xdp_rxq); 98 rx_stats = &rxq->stats; 107 rxq->xdp_rc = xdp_do_redirect(ndev, xdp, prog); 108 if (!rxq->xdp_rc) { 109 rxq->xdp_flush = true;
|
/linux-master/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_trace.h | 75 TP_PROTO(const struct funeth_rxq *rxq, 81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec), 90 __string(devname, rxq->netdev->name) 94 __entry->qidx = rxq->qidx; 95 __entry->cq_head = rxq->cq_head; 100 __assign_str(devname, rxq->netdev->name);
|
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.c | 332 * @rxq: the queue that owns the buffer 335 static void idpf_rx_page_rel(struct idpf_queue *rxq, struct idpf_rx_buf *rx_buf) argument 340 page_pool_put_full_page(rxq->pp, rx_buf->page, false); 348 * @rxq: queue to use 350 static void idpf_rx_hdr_buf_rel_all(struct idpf_queue *rxq) argument 352 struct idpf_adapter *adapter = rxq->vport->adapter; 355 rxq->desc_count * IDPF_HDR_BUF_SIZE, 356 rxq->rx_buf.hdr_buf_va, 357 rxq->rx_buf.hdr_buf_pa); 358 rxq 365 idpf_rx_buf_rel_all(struct idpf_queue *rxq) argument 395 idpf_rx_desc_rel(struct idpf_queue *rxq, bool bufq, s32 q_model) argument 466 idpf_rx_buf_hw_update(struct idpf_queue *rxq, u32 val) argument 483 idpf_rx_hdr_buf_alloc_all(struct idpf_queue *rxq) argument 714 idpf_rx_desc_alloc(struct idpf_queue *rxq, bool bufq, s32 q_model) argument 2738 idpf_rx_hash(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc, struct idpf_rx_ptype_decoded *decoded) argument 2763 idpf_rx_csum(struct idpf_queue *rxq, struct sk_buff *skb, struct idpf_rx_csum_decoded *csum_bits, struct idpf_rx_ptype_decoded *decoded) argument 2863 idpf_rx_rsc(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc, struct idpf_rx_ptype_decoded *decoded) argument 2933 idpf_rx_process_skb_fields(struct idpf_queue *rxq, struct sk_buff *skb, struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc) argument 2995 idpf_rx_construct_skb(struct idpf_queue *rxq, struct idpf_rx_buf *rx_buf, unsigned int size) argument 3056 idpf_rx_hdr_construct_skb(struct idpf_queue *rxq, const void *va, unsigned int size) argument 3119 idpf_rx_splitq_clean(struct idpf_queue *rxq, int budget) argument 3602 struct idpf_queue *rxq = q_vector->rx[i]; local 3886 struct idpf_queue *rxq = q_vec->rx[i]; local [all...] |
H A D | idpf_txrx.h | 78 #define IDPF_RX_BUFQ_WORKING_SET(rxq) ((rxq)->desc_count - 1) 80 #define IDPF_RX_BUMP_NTC(rxq, ntc) \ 82 if (unlikely(++(ntc) == (rxq)->desc_count)) { \ 84 change_bit(__IDPF_Q_GEN_CHK, (rxq)->flags); \ 118 #define IDPF_SINGLEQ_RX_BUF_DESC(rxq, i) \ 119 (&(((struct virtchnl2_singleq_rx_buf_desc *)((rxq)->desc_ring))[i])) 120 #define IDPF_SPLITQ_RX_BUF_DESC(rxq, i) \ 121 (&(((struct virtchnl2_splitq_rx_buf_desc *)((rxq)->desc_ring))[i])) 122 #define IDPF_SPLITQ_RX_BI_DESC(rxq, 785 struct idpf_queue rxq; member in struct:idpf_rxq_set [all...] |
/linux-master/drivers/net/ethernet/hisilicon/ |
H A D | hisi_femac.c | 121 struct hisi_femac_queue rxq; member in struct:hisi_femac_priv 212 struct hisi_femac_queue *rxq = &priv->rxq; local 218 pos = rxq->head; 220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) 222 if (unlikely(rxq->skb[pos])) { 224 pos, rxq->skb[pos]); 237 rxq->dma_phys[pos] = addr; 238 rxq 248 struct hisi_femac_queue *rxq = &priv->rxq; local 388 struct hisi_femac_queue *rxq = &priv->rxq; local [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | xdp_hw_metadata.c | 63 int rxq; variable 405 static int verify_metadata(struct xsk *rx_xsk, int rxq, int server_fd, clockid_t clock_id) argument 408 struct pollfd fds[rxq + 1]; 415 for (i = 0; i < rxq; i++) { 421 fds[rxq].fd = server_fd; 422 fds[rxq].events = POLLIN; 423 fds[rxq].revents = 0; 428 for (i = 0; i < rxq; i++) { 434 ret = poll(fds, rxq + 1, 1000); 443 if (fds[rxq] [all...] |
/linux-master/drivers/net/ethernet/atheros/alx/ |
H A D | main.c | 73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; local 79 next = cur = rxq->write_idx; 82 cur_buf = &rxq->bufs[cur]; 84 while (!cur_buf->skb && next != rxq->read_idx) { 85 struct alx_rfd *rfd = &rxq->rfd[cur]; 127 cur_buf = &rxq->bufs[cur]; 134 rxq->write_idx = cur; 216 static int alx_clean_rx_irq(struct alx_rx_queue *rxq, int budget) argument 225 alx = netdev_priv(rxq 506 alx_free_rxring_buf(struct alx_rx_queue *rxq) argument 630 alx_alloc_rx_ring(struct alx_priv *alx, struct alx_rx_queue *rxq, int offset) argument 741 struct alx_rx_queue *rxq; local [all...] |
/linux-master/drivers/vhost/ |
H A D | net.c | 127 struct vhost_net_buf rxq; member in struct:vhost_net_virtqueue 150 static void *vhost_net_buf_get_ptr(struct vhost_net_buf *rxq) argument 152 if (rxq->tail != rxq->head) 153 return rxq->queue[rxq->head]; 158 static int vhost_net_buf_get_size(struct vhost_net_buf *rxq) argument 160 return rxq->tail - rxq->head; 163 static int vhost_net_buf_is_empty(struct vhost_net_buf *rxq) argument 168 vhost_net_buf_consume(struct vhost_net_buf *rxq) argument 177 struct vhost_net_buf *rxq = &nvq->rxq; local 187 struct vhost_net_buf *rxq = &nvq->rxq; local 210 struct vhost_net_buf *rxq = &nvq->rxq; local 222 vhost_net_buf_init(struct vhost_net_buf *rxq) argument [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | cxgb4vf_main.c | 383 int rxq, msi, err; local 397 for_each_ethrxq(s, rxq) { 401 &s->ethrxq[rxq].rspq); 409 while (--rxq >= 0) 410 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); 421 int rxq, msi; local 425 for_each_ethrxq(s, rxq) 427 &s->ethrxq[rxq].rspq); 452 int rxq; local 455 for_each_ethrxq(s, rxq) 477 int rxq; local 632 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local 663 struct sge_eth_rxq *rxq = &s->ethrxq[pi->first_qset]; local 705 struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local 1236 struct sge_eth_rxq *rxq; local 1810 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset]; local 2062 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local 2202 const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL]; local 2729 struct sge_eth_rxq *rxq = &s->ethrxq[qs]; local [all...] |