/linux-master/net/core/ |
H A D | netdev-genl.c | 20 unsigned int txq_idx; member in struct:netdev_nl_dump_ctx 420 for (i = ctx->txq_idx; i < netdev->real_num_tx_queues;) { 425 ctx->txq_idx = i++; 456 ctx->txq_idx = 0; 572 i = ctx->txq_idx; 578 ctx->txq_idx = i++; 582 ctx->txq_idx = 0;
|
/linux-master/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_bpf.c | 15 u16 txq_idx = skb_get_queue_mapping(skb); local 21 ndevtxq = netdev_get_tx_queue(ndev, txq_idx);
|
H A D | mana_en.c | 225 u16 txq_idx = skb_get_queue_mapping(skb); local 243 txq = &apc->tx_qp[txq_idx].txq; 245 cq = &apc->tx_qp[txq_idx].tx_cq; 385 net_txq = netdev_get_tx_queue(ndev, txq_idx);
|
/linux-master/drivers/target/iscsi/cxgbit/ |
H A D | cxgbit.h | 231 u16 txq_idx; member in struct:cxgbit_sock
|
H A D | cxgbit_cm.c | 631 cxgb_mk_close_con_req(skb, len, csk->tid, csk->txq_idx, 672 cxgb_mk_abort_req(skb, len, csk->tid, csk->txq_idx, 943 csk->txq_idx = cxgb4_port_idx(ndev) * step; 979 csk->txq_idx = (port_id * step) + 1483 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx); 1771 cxgb_mk_abort_rpl(rpl_skb, len, csk->tid, csk->txq_idx);
|
H A D | cxgbit_target.c | 244 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx);
|
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_main.c | 858 unsigned int tx_pkts = 0, tx_bytes = 0, txq_idx; local 921 txq_idx = nicvf_netdev_qidx(nic, cq_idx); 924 if (txq_idx < nic->pnicvf->xdp_tx_queues) { 929 txq_idx -= nic->pnicvf->xdp_tx_queues; 936 txq = netdev_get_tx_queue(netdev, txq_idx); 947 "Transmit queue wakeup SQ%d\n", txq_idx);
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | cm.c | 642 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 655 cxgb_mk_close_con_req(skb, wrlen, ep->hwtid, ep->txq_idx, 696 cxgb_mk_abort_req(req_skb, wrlen, ep->hwtid, ep->txq_idx, 960 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 1066 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 1118 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 1146 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 2113 ep->txq_idx = cxgb4_port_idx(pdev) * step; 2132 ep->txq_idx = cxgb4_port_idx(pdev) * step; 2234 pr_debug("txq_idx [all...] |
H A D | iw_cxgb4.h | 891 u16 txq_idx; member in struct:c4iw_ep
|
H A D | qp.c | 1574 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx); 1705 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); 1769 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx);
|
/linux-master/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_cm.c | 237 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); 1233 csk->txq_idx = (rxq_idx < cdev->lldi->ntxq) ? rxq_idx : 2061 queue = csk->txq_idx; 2094 int queue = csk->txq_idx; 2112 int queue = csk->txq_idx;
|
H A D | chtls_io.c | 101 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); 128 csk->txq_idx, 661 skb_set_queue_mapping(skb, (csk->txq_idx << 1) |
|
H A D | chtls_hw.c | 81 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA);
|
H A D | chtls.h | 297 u32 txq_idx; member in struct:chtls_sock
|
/linux-master/drivers/scsi/cxgbi/ |
H A D | libcxgbi.h | 120 unsigned short txq_idx; member in struct:cxgbi_sock
|
/linux-master/drivers/net/wireless/intel/iwlwifi/pcie/ |
H A D | trans.c | 2398 static int iwl_trans_pcie_wait_txq_empty(struct iwl_trans *trans, int txq_idx) argument 2409 if (!test_bit(txq_idx, trans->txqs.queue_used)) 2412 IWL_DEBUG_TX_QUEUES(trans, "Emptying queue %d...\n", txq_idx); 2413 txq = trans->txqs.txq[txq_idx]; 2449 "fail to flush all tx fifo queues Q %d\n", txq_idx); 2454 IWL_DEBUG_TX_QUEUES(trans, "Queue %d is now empty.\n", txq_idx);
|
/linux-master/drivers/scsi/cxgbi/cxgb4i/ |
H A D | cxgb4i.c | 1864 csk->txq_idx = cxgb4_port_idx(ndev) * step; 1893 csk->flags, csk->tx_chan, csk->txq_idx, csk->rss_qid,
|