Lines Matching refs:tx_queue

181 			   "XDP_TX and XDP_REDIRECT will work with reduced performance (%d cpus/tx_queue)\n",
533 struct efx_tx_queue *tx_queue;
546 tx_queue = &channel->tx_queue[j];
547 tx_queue->efx = efx;
548 tx_queue->queue = -1;
549 tx_queue->label = j;
550 tx_queue->channel = channel;
603 struct efx_tx_queue *tx_queue;
620 tx_queue = &channel->tx_queue[j];
621 if (tx_queue->channel)
622 tx_queue->channel = channel;
623 tx_queue->buffer = NULL;
624 tx_queue->cb_page = NULL;
625 memset(&tx_queue->txd, 0, sizeof(tx_queue->txd));
641 struct efx_tx_queue *tx_queue;
656 efx_for_each_channel_tx_queue(tx_queue, channel) {
657 rc = efx_probe_tx_queue(tx_queue);
741 struct efx_tx_queue *tx_queue;
749 efx_for_each_channel_tx_queue(tx_queue, channel)
750 efx_remove_tx_queue(tx_queue);
766 struct efx_tx_queue *tx_queue)
773 tx_queue->channel->channel, tx_queue->label,
774 xdp_queue_number, tx_queue->queue);
775 efx->xdp_tx_queues[xdp_queue_number] = tx_queue;
781 struct efx_tx_queue *tx_queue;
796 efx_for_each_channel_tx_queue(tx_queue, channel) {
797 tx_queue->queue = next_queue++;
799 tx_queue);
804 efx_for_each_channel_tx_queue(tx_queue, channel) {
805 tx_queue->queue = next_queue++;
808 channel->channel, tx_queue->label,
809 tx_queue->queue);
819 tx_queue = &channel->tx_queue[0];
821 tx_queue);
837 tx_queue = efx->xdp_tx_queues[next_queue++];
838 rc = efx_set_xdp_tx_queue(efx, xdp_queue_number, tx_queue);
1091 struct efx_tx_queue *tx_queue;
1098 efx_for_each_channel_tx_queue(tx_queue, channel) {
1099 efx_init_tx_queue(tx_queue);
1117 struct efx_tx_queue *tx_queue;
1159 efx_for_each_channel_tx_queue(tx_queue, channel)
1160 efx_fini_tx_queue(tx_queue);
1179 struct efx_tx_queue *tx_queue;
1191 efx_for_each_channel_tx_queue(tx_queue, channel) {
1192 tx_queue->pkts_compl = 0;
1193 tx_queue->bytes_compl = 0;
1206 efx_for_each_channel_tx_queue(tx_queue, channel) {
1207 if (tx_queue->bytes_compl) {
1208 netdev_tx_completed_queue(tx_queue->core_txq,
1209 tx_queue->pkts_compl,
1210 tx_queue->bytes_compl);