Lines Matching refs:netdev_queue

634 struct netdev_queue {
711 static inline int netdev_queue_numa_node_read(const struct netdev_queue *q)
720 static inline void netdev_queue_numa_node_write(struct netdev_queue *q, int node)
2044 struct netdev_queue *_tx;
2267 struct netdev_queue __rcu *ingress_queue;
2482 struct netdev_queue *netdev_get_tx_queue(const struct net_device *dev,
2489 static inline struct netdev_queue *skb_get_tx_queue(const struct net_device *dev,
2497 struct netdev_queue *,
2524 struct netdev_queue *netdev_core_pick_tx(struct net_device *dev,
3265 void netif_schedule_queue(struct netdev_queue *txq);
3275 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue)
3296 struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
3301 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
3320 struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
3325 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue)
3345 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue)
3361 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue)
3367 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue)
3373 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue)
3388 static inline void netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue,
3396 static inline int netdev_queue_dql_avail(const struct netdev_queue *txq)
3413 static inline void netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue)
3427 static inline void netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue)
3444 static inline void netdev_tx_sent_queue(struct netdev_queue *dev_queue,
3474 static inline bool __netdev_tx_sent_queue(struct netdev_queue *dev_queue,
3520 static inline void netdev_tx_completed_queue(struct netdev_queue *dev_queue,
3560 static inline void netdev_tx_reset_queue(struct netdev_queue *q)
3627 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
3641 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
3655 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
3682 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
3940 struct netdev_queue *txq, int *ret);
4345 static inline void __netif_tx_lock(struct netdev_queue *txq, int cpu)
4352 static inline bool __netif_tx_acquire(struct netdev_queue *txq)
4358 static inline void __netif_tx_release(struct netdev_queue *txq)
4363 static inline void __netif_tx_lock_bh(struct netdev_queue *txq)
4370 static inline bool __netif_tx_trylock(struct netdev_queue *txq)
4381 static inline void __netif_tx_unlock(struct netdev_queue *txq)
4388 static inline void __netif_tx_unlock_bh(struct netdev_queue *txq)
4398 static inline void txq_trans_update(struct netdev_queue *txq)
4404 static inline void txq_trans_cond_update(struct netdev_queue *txq)
4415 struct netdev_queue *txq = netdev_get_tx_queue(dev, 0);
4472 struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
4891 struct netdev_queue *txq, bool more)