/linux-master/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-xsk.c | 431 err = priv->enqueue(priv, fq, &fds[total_enqueued], 0,
|
/linux-master/kernel/sched/ |
H A D | rt.c | 876 int enqueue = 0; local 907 enqueue = 1; 925 enqueue = 1; 930 if (enqueue) 1382 * Don't enqueue the group if its throttled, or when empty. 1507 * dequeue followed by enqueue.
|
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gsc.c | 608 bool enqueue) 610 bool masked = !enqueue; 925 bool enqueue) 927 bool masked = !enqueue; 942 if (enqueue && gsc_dst_get_buf_seq(ctx) >= GSC_BUF_START) 946 if (!enqueue && gsc_dst_get_buf_seq(ctx) <= GSC_BUF_STOP) 607 gsc_src_set_buf_seq(struct gsc_context *ctx, u32 buf_id, bool enqueue) argument 924 gsc_dst_set_buf_seq(struct gsc_context *ctx, u32 buf_id, bool enqueue) argument
|
H A D | exynos_drm_fimc.c | 905 bool enqueue) 911 DRM_DEV_DEBUG_KMS(ctx->dev, "buf_id[%d]enqueu[%d]\n", buf_id, enqueue); 917 if (enqueue) 926 if (enqueue && buf_num >= FIMC_BUF_START) 928 else if (!enqueue && buf_num <= FIMC_BUF_STOP) 904 fimc_dst_set_buf_seq(struct fimc_context *ctx, u32 buf_id, bool enqueue) argument
|
/linux-master/net/sched/ |
H A D | sch_teql.c | 447 ops->enqueue = teql_enqueue;
|
H A D | sch_taprio.c | 1920 /* Protects against enqueue()/dequeue() */ 2524 .enqueue = taprio_enqueue,
|
H A D | sch_sfq.c | 376 goto enqueue; 441 enqueue: 923 .enqueue = sfq_enqueue,
|
H A D | sch_pie.c | 89 bool enqueue = false; local 98 enqueue = true; 105 enqueue = true; 108 /* we can enqueue the packet */ 109 if (enqueue) { 110 /* Set enqueue time only when dq_rate_estimator is disabled. */ 551 .enqueue = pie_qdisc_enqueue,
|
H A D | sch_skbprio.c | 287 .enqueue = skbprio_enqueue,
|
H A D | sch_htb.c | 214 * node). If we end up with classid MAJOR:0 we enqueue the skb into special 627 /* enqueue to helper queue */ 888 * qdisc drops packets in enqueue routine or if someone used 2143 .enqueue = htb_enqueue,
|
H A D | sch_hhf.c | 701 .enqueue = hhf_enqueue,
|
H A D | sch_hfsc.c | 1674 .enqueue = hfsc_enqueue,
|
H A D | sch_generic.c | 66 * - enqueue, dequeue are serialized via qdisc root lock 648 .enqueue = noop_enqueue, 660 .enqueue = noop_enqueue, 688 qdisc->enqueue = NULL; 696 .enqueue = noop_enqueue, 913 .enqueue = pfifo_fast_enqueue, 979 sch->enqueue = ops->enqueue;
|
H A D | sch_fq_pie.c | 29 * - Drops during enqueue only. 137 u8 enqueue = false; local 166 enqueue = true; 174 enqueue = true; 176 if (enqueue) { 177 /* Set enqueue time only when dq_rate_estimator is disabled. */ 563 .enqueue = fq_pie_qdisc_enqueue,
|
H A D | sch_fq.c | 17 * Transport (eg TCP) can set in sk->sk_pacing_rate a rate, enqueue a 21 * enqueue() : 1286 .enqueue = fq_enqueue,
|
H A D | sch_fq_codel.c | 150 * in fast path (packet queue/enqueue) with many cache misses. 722 .enqueue = fq_codel_enqueue,
|
H A D | sch_ets.c | 808 .enqueue = ets_qdisc_enqueue,
|
H A D | sch_fifo.c | 187 .enqueue = pfifo_enqueue, 202 .enqueue = bfifo_enqueue, 217 .enqueue = pfifo_tail_enqueue,
|
H A D | sch_choke.c | 491 .enqueue = choke_enqueue,
|
H A D | sch_codel.c | 257 .enqueue = codel_qdisc_enqueue,
|
H A D | sch_etf.c | 494 .enqueue = etf_enqueue_timesortedlist,
|
H A D | sch_cake.c | 3111 .enqueue = cake_enqueue,
|
H A D | sch_cbs.c | 81 int (*enqueue)(struct sk_buff *skb, struct Qdisc *sch, member in struct:cbs_sched_data 95 err = child->ops->enqueue(skb, child, to_free); 136 return q->enqueue(skb, sch, to_free); 260 q->enqueue = cbs_enqueue_soft; 303 q->enqueue = cbs_enqueue_offload; 425 q->enqueue = cbs_enqueue_soft; 539 .enqueue = cbs_enqueue,
|
H A D | sch_api.c | 49 which is able to enqueue packets and to dequeue them (when 71 Every discipline has two major routines: enqueue and dequeue. 82 ---enqueue 84 enqueue returns 0, if packet was enqueued successfully. 141 if (qops->enqueue == NULL) 142 qops->enqueue = noop_qdisc_ops.enqueue;
|
/linux-master/net/core/ |
H A D | dev.c | 3773 rc = q->enqueue(skb, q, to_free) & NET_XMIT_MASK; 3835 * and then other tasks will only enqueue packets. The packets will be 4358 if (q->enqueue) {
|