/linux-master/drivers/net/ethernet/hisilicon/hns/ |
H A D | hns_enet.c | 333 struct netdev_queue *dev_queue; local 386 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping); 387 netdev_tx_sent_queue(dev_queue, skb->len); 976 struct netdev_queue *dev_queue; local 1005 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); 1006 netdev_tx_completed_queue(dev_queue, pkts, bytes); 1017 if (netif_tx_queue_stopped(dev_queue) && 1019 netif_tx_wake_queue(dev_queue); 1060 struct netdev_queue *dev_queue; local 1070 dev_queue [all...] |
H A D | hns_ethtool.c | 376 struct netdev_queue *dev_queue; local 407 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); 408 netdev_tx_reset_queue(dev_queue);
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/ |
H A D | hns3_enet.c | 813 struct netdev_queue *dev_queue; local 817 dev_queue = netdev_get_tx_queue(ndev, 819 netdev_tx_reset_queue(dev_queue); 2299 struct netdev_queue *dev_queue; local 2343 dev_queue = netdev_get_tx_queue(netdev, ring->queue_index); 2344 doorbell = __netdev_tx_sent_queue(dev_queue, desc_cb->send_bytes, 3630 struct netdev_queue *dev_queue; local 3647 dev_queue = netdev_get_tx_queue(netdev, ring->tqp->tqp_index); 3648 netdev_tx_completed_queue(dev_queue, pkts, bytes); 3656 if (netif_tx_queue_stopped(dev_queue) [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos.c | 1502 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); local 1503 struct Qdisc *qdisc = rtnl_dereference(dev_queue->qdisc_sleeping);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | qos.c | 367 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); local 368 struct Qdisc *qdisc = dev_queue->qdisc_sleeping;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 4937 struct netdev_queue *dev_queue = local 4941 if (!netif_xmit_stopped(dev_queue))
|
/linux-master/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge.c | 1381 struct netdev_queue *dev_queue; local 1416 dev_queue = netdev_get_tx_queue(ss->dev, ss - ss->mgp->ss); 1427 __netif_tx_trylock(dev_queue)) { 1433 __netif_tx_unlock(dev_queue); 1437 if (netif_tx_queue_stopped(dev_queue) && 1441 netif_tx_wake_queue(dev_queue);
|
/linux-master/drivers/net/xen-netback/ |
H A D | xenbus.c | 31 struct netdev_queue *dev_queue; local 92 dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id); 97 netif_tx_queue_stopped(dev_queue) ? "stopped" : "running");
|
/linux-master/drivers/net/ |
H A D | xen-netfront.c | 261 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, queue->id); local 263 if (unlikely(netif_tx_queue_stopped(dev_queue)) &&
|
/linux-master/include/linux/ |
H A D | netdevice.h | 3275 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) argument 3277 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); 3301 void netif_tx_wake_queue(struct netdev_queue *dev_queue); 3325 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) argument 3328 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); 3345 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) argument 3347 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); 3361 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) argument 3363 return dev_queue->state & QUEUE_STATE_ANY_XOFF; 3367 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) argument 3373 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) argument 3388 netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue, unsigned int min_limit) argument 3413 netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) argument 3427 netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) argument 3444 netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes) argument 3474 __netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes, bool xmit_more) argument 3520 netdev_tx_completed_queue(struct netdev_queue *dev_queue, unsigned int pkts, unsigned int bytes) argument 3575 netdev_reset_queue(struct net_device *dev_queue) argument [all...] |
/linux-master/include/net/ |
H A D | netdev_queues.h | 206 netdev_txq_completed_mb(struct netdev_queue *dev_queue, argument 210 netdev_tx_completed_queue(dev_queue, pkts, bytes);
|
H A D | pkt_sched.h | 144 return dev_net(q->dev_queue->dev);
|
H A D | sch_generic.h | 84 * q->dev_queue : It can test 105 struct netdev_queue *dev_queue; member in struct:Qdisc 550 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc); 557 return rcu_dereference_bh(qdisc->dev_queue->qdisc); 562 return rcu_dereference_rtnl(qdisc->dev_queue->qdisc_sleeping); 575 return qdisc->dev_queue->dev; 693 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue, 727 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, 731 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
|
/linux-master/net/core/ |
H A D | dev.c | 3159 void netif_tx_wake_queue(struct netdev_queue *dev_queue) argument 3161 if (test_and_clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state)) { 3165 q = rcu_dereference(dev_queue->qdisc);
|
/linux-master/net/sched/ |
H A D | cls_api.c | 866 struct net_device *dev = q->dev_queue->dev; 905 struct net_device *dev = q->dev_queue->dev;
|
H A D | sch_api.c | 1111 struct netdev_queue *dev_queue; local 1118 dev_queue = dev_ingress_queue(dev); 1119 if (!dev_queue) { 1124 q = rtnl_dereference(dev_queue->qdisc_sleeping); 1146 dev_queue = netdev_get_tx_queue(dev, i); 1147 old = dev_graft_qdisc(dev_queue, new); 1154 old = dev_graft_qdisc(dev_queue, NULL); 1163 dev_graft_qdisc(dev_queue, new); 1253 struct netdev_queue *dev_queue, 1299 sch = qdisc_alloc(dev_queue, op 1252 qdisc_create(struct net_device *dev, struct netdev_queue *dev_queue, u32 parent, u32 handle, struct nlattr **tca, int *errp, struct netlink_ext_ack *extack) argument 1768 struct netdev_queue *dev_queue; local 1874 struct netdev_queue *dev_queue; local 2342 struct netdev_queue *dev_queue; local [all...] |
H A D | sch_cbs.c | 412 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 423 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); 498 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
|
H A D | sch_drr.c | 110 cl->qdisc = qdisc_create_dflt(sch->dev_queue, 211 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
|
H A D | sch_etf.c | 379 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
|
H A D | sch_ets.c | 254 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
|
H A D | sch_fifo.c | 260 q = qdisc_create_dflt(sch->dev_queue, ops, TC_H_MAKE(sch->handle, 1),
|
H A D | sch_fq.c | 879 netdev_queue_numa_node_read(sch->dev_queue));
|
H A D | sch_generic.c | 75 const struct netdev_queue *txq = q->dev_queue; 232 const struct netdev_queue *txq = q->dev_queue; 665 .dev_queue = &noop_netdev_queue, 928 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, argument 937 if (!dev_queue) { 943 dev = dev_queue->dev; 944 sch = kzalloc_node(size, GFP_KERNEL, netdev_queue_numa_node_read(dev_queue)); 981 sch->dev_queue = dev_queue; 994 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue, argument 1124 dev_graft_qdisc(struct netdev_queue *dev_queue, struct Qdisc *qdisc) argument 1145 shutdown_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) argument 1160 attach_one_default_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) argument 1222 transition_one_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_need_watchdog) argument 1275 dev_deactivate_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) argument 1289 dev_reset_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) argument 1321 struct netdev_queue *dev_queue; local 1400 qdisc_change_tx_queue_len(struct net_device *dev, struct netdev_queue *dev_queue) argument 1466 dev_init_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc) argument [all...] |
H A D | sch_hfsc.c | 1061 cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 1194 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 1420 q->root.qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
|
H A D | sch_htb.c | 1119 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local 1122 qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops, 1161 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); 1166 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local 1167 struct Qdisc *old = dev_graft_qdisc(dev_queue, NULL); 1183 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); local 1184 struct Qdisc *old = dev_graft_qdisc(dev_queue, sch); 1369 return sch->dev_queue; 1382 htb_graft_helper(struct netdev_queue *dev_queue, struct Qdisc *new_q) argument 1384 struct net_device *dev = dev_queue 1444 struct netdev_queue *dev_queue = sch->dev_queue; local 1525 htb_parent_to_leaf_offload(struct Qdisc *sch, struct netdev_queue *dev_queue, struct Qdisc *new_q) argument 1543 struct netdev_queue *dev_queue; local 1719 struct netdev_queue *dev_queue = sch->dev_queue; local 1768 struct netdev_queue *dev_queue; local [all...] |