Lines Matching defs:qos

75 	int qos, queues_per_port;
82 for (qos = 0; qos < queues_per_port; qos++) {
83 if (skb_queue_len(&priv->tx_free_list[qos]) == 0)
85 skb_to_free = cvmx_fau_fetch_and_add32(priv->fau + qos * 4,
88 priv->fau + qos * 4);
92 spin_lock_irqsave(&priv->tx_free_list[qos].lock, flags);
96 t = __skb_dequeue(&priv->tx_free_list[qos]);
101 spin_unlock_irqrestore(&priv->tx_free_list[qos].lock,
111 total_remaining += skb_queue_len(&priv->tx_free_list[qos]);
132 int qos;
153 * completely remove "qos" in the event neither interface
158 qos = GET_SKBUFF_QOS(skb);
159 if (qos <= 0)
160 qos = 0;
161 else if (qos >= cvmx_pko_get_num_queues(priv->port))
162 qos = 0;
164 qos = 0;
181 priv->fau + qos * 4,
207 qos * 4,
212 qos * 4);
213 spin_lock_irqsave(&priv->tx_free_list[qos].lock, flags);
381 skb_to_free = cvmx_fau_fetch_and_add32(priv->fau + qos * 4,
388 priv->fau + qos * 4);
399 pko_command.s.reg0 = priv->fau + qos * 4;
407 spin_lock_irqsave(&priv->tx_free_list[qos].lock, flags);
410 if (unlikely(skb_queue_len(&priv->tx_free_list[qos]) >=
414 spin_unlock_irqrestore(&priv->tx_free_list[qos].lock,
417 spin_lock_irqsave(&priv->tx_free_list[qos].lock,
426 cvmx_pko_send_packet_prepare(priv->port, priv->queue + qos,
431 priv->queue + qos,
451 __skb_queue_tail(&priv->tx_free_list[qos], skb);
458 struct sk_buff *t = __skb_dequeue(&priv->tx_free_list[qos]);
465 spin_unlock_irqrestore(&priv->tx_free_list[qos].lock, flags);
662 int qos;
664 for (qos = 0; qos < 16; qos++) {
665 spin_lock_irqsave(&priv->tx_free_list[qos].lock, flags);
666 while (skb_queue_len(&priv->tx_free_list[qos]))
668 (&priv->tx_free_list[qos]));
669 spin_unlock_irqrestore(&priv->tx_free_list[qos].lock, flags);