Searched refs:dev_queue (Results 1 - 25 of 32) sorted by path

12

/linux-master/drivers/net/ethernet/hisilicon/hns/
H A Dhns_enet.c333 struct netdev_queue *dev_queue; local
386 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping);
387 netdev_tx_sent_queue(dev_queue, skb->len);
976 struct netdev_queue *dev_queue; local
1005 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index);
1006 netdev_tx_completed_queue(dev_queue, pkts, bytes);
1017 if (netif_tx_queue_stopped(dev_queue) &&
1019 netif_tx_wake_queue(dev_queue);
1060 struct netdev_queue *dev_queue; local
1070 dev_queue
[all...]
H A Dhns_ethtool.c376 struct netdev_queue *dev_queue; local
407 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index);
408 netdev_tx_reset_queue(dev_queue);
/linux-master/drivers/net/ethernet/hisilicon/hns3/
H A Dhns3_enet.c813 struct netdev_queue *dev_queue; local
817 dev_queue = netdev_get_tx_queue(ndev,
819 netdev_tx_reset_queue(dev_queue);
2299 struct netdev_queue *dev_queue; local
2343 dev_queue = netdev_get_tx_queue(netdev, ring->queue_index);
2344 doorbell = __netdev_tx_sent_queue(dev_queue, desc_cb->send_bytes,
3630 struct netdev_queue *dev_queue; local
3647 dev_queue = netdev_get_tx_queue(netdev, ring->tqp->tqp_index);
3648 netdev_tx_completed_queue(dev_queue, pkts, bytes);
3656 if (netif_tx_queue_stopped(dev_queue)
[all...]
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dqos.c1502 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); local
1503 struct Qdisc *qdisc = rtnl_dereference(dev_queue->qdisc_sleeping);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dqos.c367 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); local
368 struct Qdisc *qdisc = dev_queue->qdisc_sleeping;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c4937 struct netdev_queue *dev_queue = local
4941 if (!netif_xmit_stopped(dev_queue))
/linux-master/drivers/net/ethernet/myricom/myri10ge/
H A Dmyri10ge.c1381 struct netdev_queue *dev_queue; local
1416 dev_queue = netdev_get_tx_queue(ss->dev, ss - ss->mgp->ss);
1427 __netif_tx_trylock(dev_queue)) {
1433 __netif_tx_unlock(dev_queue);
1437 if (netif_tx_queue_stopped(dev_queue) &&
1441 netif_tx_wake_queue(dev_queue);
/linux-master/drivers/net/xen-netback/
H A Dxenbus.c31 struct netdev_queue *dev_queue; local
92 dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id);
97 netif_tx_queue_stopped(dev_queue) ? "stopped" : "running");
/linux-master/drivers/net/
H A Dxen-netfront.c261 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, queue->id); local
263 if (unlikely(netif_tx_queue_stopped(dev_queue)) &&
/linux-master/include/linux/
H A Dnetdevice.h3275 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) argument
3277 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);
3301 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
3325 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) argument
3328 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);
3345 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) argument
3347 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state);
3361 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) argument
3363 return dev_queue->state & QUEUE_STATE_ANY_XOFF;
3367 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) argument
3373 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) argument
3388 netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue, unsigned int min_limit) argument
3413 netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) argument
3427 netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) argument
3444 netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes) argument
3474 __netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes, bool xmit_more) argument
3520 netdev_tx_completed_queue(struct netdev_queue *dev_queue, unsigned int pkts, unsigned int bytes) argument
3575 netdev_reset_queue(struct net_device *dev_queue) argument
[all...]
/linux-master/include/net/
H A Dnetdev_queues.h206 netdev_txq_completed_mb(struct netdev_queue *dev_queue, argument
210 netdev_tx_completed_queue(dev_queue, pkts, bytes);
H A Dpkt_sched.h144 return dev_net(q->dev_queue->dev);
H A Dsch_generic.h84 * q->dev_queue : It can test
105 struct netdev_queue *dev_queue; member in struct:Qdisc
550 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc);
557 return rcu_dereference_bh(qdisc->dev_queue->qdisc);
562 return rcu_dereference_rtnl(qdisc->dev_queue->qdisc_sleeping);
575 return qdisc->dev_queue->dev;
693 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue,
727 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue,
731 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
/linux-master/net/core/
H A Ddev.c3159 void netif_tx_wake_queue(struct netdev_queue *dev_queue) argument
3161 if (test_and_clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state)) {
3165 q = rcu_dereference(dev_queue->qdisc);
/linux-master/net/sched/
H A Dcls_api.c866 struct net_device *dev = q->dev_queue->dev;
905 struct net_device *dev = q->dev_queue->dev;
H A Dsch_api.c1111 struct netdev_queue *dev_queue; local
1118 dev_queue = dev_ingress_queue(dev);
1119 if (!dev_queue) {
1124 q = rtnl_dereference(dev_queue->qdisc_sleeping);
1146 dev_queue = netdev_get_tx_queue(dev, i);
1147 old = dev_graft_qdisc(dev_queue, new);
1154 old = dev_graft_qdisc(dev_queue, NULL);
1163 dev_graft_qdisc(dev_queue, new);
1253 struct netdev_queue *dev_queue,
1299 sch = qdisc_alloc(dev_queue, op
1252 qdisc_create(struct net_device *dev, struct netdev_queue *dev_queue, u32 parent, u32 handle, struct nlattr **tca, int *errp, struct netlink_ext_ack *extack) argument
1768 struct netdev_queue *dev_queue; local
1874 struct netdev_queue *dev_queue; local
2342 struct netdev_queue *dev_queue; local
[all...]
H A Dsch_cbs.c412 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
423 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
498 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
H A Dsch_drr.c110 cl->qdisc = qdisc_create_dflt(sch->dev_queue,
211 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
H A Dsch_etf.c379 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
H A Dsch_ets.c254 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
H A Dsch_fifo.c260 q = qdisc_create_dflt(sch->dev_queue, ops, TC_H_MAKE(sch->handle, 1),
H A Dsch_fq.c879 netdev_queue_numa_node_read(sch->dev_queue));
H A Dsch_generic.c75 const struct netdev_queue *txq = q->dev_queue;
232 const struct netdev_queue *txq = q->dev_queue;
665 .dev_queue = &noop_netdev_queue,
928 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, argument
937 if (!dev_queue) {
943 dev = dev_queue->dev;
944 sch = kzalloc_node(size, GFP_KERNEL, netdev_queue_numa_node_read(dev_queue));
981 sch->dev_queue = dev_queue;
994 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue, argument
1124 dev_graft_qdisc(struct netdev_queue *dev_queue, struct Qdisc *qdisc) argument
1145 shutdown_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) argument
1160 attach_one_default_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) argument
1222 transition_one_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_need_watchdog) argument
1275 dev_deactivate_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) argument
1289 dev_reset_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) argument
1321 struct netdev_queue *dev_queue; local
1400 qdisc_change_tx_queue_len(struct net_device *dev, struct netdev_queue *dev_queue) argument
1466 dev_init_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc) argument
[all...]
H A Dsch_hfsc.c1061 cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
1194 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
1420 q->root.qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
H A Dsch_htb.c1119 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local
1122 qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops,
1161 old = dev_graft_qdisc(qdisc->dev_queue, qdisc);
1166 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local
1167 struct Qdisc *old = dev_graft_qdisc(dev_queue, NULL);
1183 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local
1184 struct Qdisc *old = dev_graft_qdisc(dev_queue, sch);
1369 return sch->dev_queue;
1382 htb_graft_helper(struct netdev_queue *dev_queue, struct Qdisc *new_q) argument
1384 struct net_device *dev = dev_queue
1444 struct netdev_queue *dev_queue = sch->dev_queue; local
1525 htb_parent_to_leaf_offload(struct Qdisc *sch, struct netdev_queue *dev_queue, struct Qdisc *new_q) argument
1543 struct netdev_queue *dev_queue; local
1719 struct netdev_queue *dev_queue = sch->dev_queue; local
1768 struct netdev_queue *dev_queue; local
[all...]

Completed in 544 milliseconds

12