Searched refs:backlog (Results 1 - 25 of 181) sorted by relevance

12345678

/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dqat_algs_send.c24 void qat_alg_send_backlog(struct qat_instance_backlog *backlog) argument
28 spin_lock_bh(&backlog->lock);
29 list_for_each_entry_safe(req, tmp, &backlog->list, list) {
40 spin_unlock_bh(&backlog->lock);
45 struct qat_instance_backlog *backlog = req->backlog; local
50 if (!list_empty(&backlog->list))
67 struct qat_instance_backlog *backlog = req->backlog; local
73 spin_lock_bh(&backlog
[all...]
H A Dqat_algs_send.h11 spinlock_t lock; /* protects backlog list */
19 struct qat_instance_backlog *backlog; member in struct:qat_alg_req
23 void qat_alg_send_backlog(struct qat_instance_backlog *backlog);
H A Dqat_compression.h21 struct qat_instance_backlog backlog; member in struct:qat_compression_instance
H A Dqat_crypto.h24 struct qat_instance_backlog backlog; member in struct:qat_crypto_instance
/linux-master/tools/perf/ui/gtk/
H A Dhelpline.c32 static int backlog; local
34 ret = vscnprintf(ui_helpline__current + backlog,
35 sizeof(ui_helpline__current) - backlog, fmt, ap);
36 backlog += ret;
40 if (ptr && (ptr - ui_helpline__current) <= backlog) {
43 backlog = 0;
/linux-master/tools/perf/ui/tui/
H A Dhelpline.c33 static int backlog; local
36 ret = vscnprintf(ui_helpline__last_msg + backlog,
37 sizeof(ui_helpline__last_msg) - backlog, format, ap);
38 backlog += ret;
42 if (ui_helpline__last_msg[backlog - 1] == '\n') {
45 backlog = 0;
/linux-master/drivers/crypto/ccp/
H A Dccp-crypto-main.c55 struct list_head *backlog; member in struct:ccp_crypto_queue
90 struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog)
95 *backlog = NULL;
110 /* Process the backlog:
112 * special precautions have to be taken when handling the backlog.
114 if (req_queue.backlog != &req_queue.cmds) {
115 /* Skip over this cmd if it is the next backlog cmd */
116 if (req_queue.backlog == &crypto_cmd->entry)
117 req_queue.backlog = crypto_cmd->entry.next;
119 *backlog
89 ccp_crypto_cmd_complete( struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) argument
140 struct ccp_crypto_cmd *held, *next, *backlog; local
[all...]
H A Dccp-dev.c280 * backlogged) or advancement out of the backlog. If the cmd has
281 * advanced out of the backlog the "err" value of the callback
315 list_add_tail(&cmd->entry, &ccp->backlog);
378 struct ccp_cmd *backlog = NULL; local
403 if (!list_empty(&ccp->backlog)) {
404 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd,
406 list_del(&backlog->entry);
411 if (backlog) {
412 INIT_WORK(&backlog
[all...]
/linux-master/drivers/infiniband/hw/qedr/
H A Dqedr_iw_cm.h37 int qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog);
/linux-master/tools/testing/selftests/net/forwarding/
H A Dsch_red.sh9 # to go to backlog.
164 qdisc_stats_get $swp3 11: .backlog
190 # This sends traffic in an attempt to build a backlog of $size. Returns 0 on
192 # backlog size to stdout.
258 local backlog
261 # Build the below-the-limit backlog using UDP. We could use TCP just
266 backlog=$(build_backlog $((2 * limit / 3)) udp)
267 check_err $? "Could not build the requested backlog"
269 check_err $? "backlog $backlog /
[all...]
/linux-master/include/uapi/linux/
H A Dgen_stats.h54 * @backlog: backlog size of queue
61 __u32 backlog; member in struct:gnet_stats_queue
/linux-master/tools/testing/selftests/bpf/progs/
H A Dtest_skc_to_unix_sock.c13 int BPF_PROG(unix_listen, struct socket *sock, int backlog) argument
/linux-master/include/net/
H A Dfq.h27 * @backlog: number of bytes pending in the queue. The number of packets can be
35 u32 backlog; member in struct:fq_flow
66 * @backlog: number of packets queued across all flows
79 u32 backlog; member in struct:fq
H A Dpie.h49 * @backlog_old: queue backlog during previous qdelay calculation
127 struct pie_vars *vars, u32 backlog, u32 packet_size);
130 struct pie_vars *vars, u32 backlog);
133 u32 backlog);
H A Dcodel_impl.h111 u32 *backlog,
129 *backlog <= params->mtu) {
147 u32 *backlog,
166 skb_len_func, skb_time_func, backlog, now);
176 * A large backlog might result in drop rates so high
202 backlog, now)) {
227 skb_time_func, backlog, now);
104 codel_should_drop(const struct sk_buff *skb, void *ctx, struct codel_vars *vars, struct codel_params *params, struct codel_stats *stats, codel_skb_len_t skb_len_func, codel_skb_time_t skb_time_func, u32 *backlog, codel_time_t now) argument
146 codel_dequeue(void *ctx, u32 *backlog, struct codel_params *params, struct codel_vars *vars, struct codel_stats *stats, codel_skb_len_t skb_len_func, codel_skb_time_t skb_time_func, codel_skb_drop_t drop_func, codel_skb_dequeue_t dequeue_func) argument
H A Dfq_impl.h24 flow->backlog -= bytes;
25 fq->backlog -= packets;
28 if (flow->backlog)
176 cur_len = cur->backlog;
185 unsigned int cur_len = tin->default_flow.backlog;
210 if (!flow->backlog) {
220 flow->backlog += skb->len;
224 fq->backlog++;
235 while (fq->backlog > fq->limit || oom) {
308 WARN_ON_ONCE(flow->backlog);
[all...]
H A Dinet_common.h42 int inet_listen(struct socket *sock, int backlog);
43 int __inet_listen_sk(struct sock *sk, int backlog);
H A Dred.h339 unsigned int backlog)
346 * qavg = qavg*(1-W) + backlog*W;
350 return v->qavg + (backlog - (v->qavg >> p->Wlog));
355 unsigned int backlog)
358 return red_calc_qavg_no_idle_time(p, v, backlog);
337 red_calc_qavg_no_idle_time(const struct red_parms *p, const struct red_vars *v, unsigned int backlog) argument
353 red_calc_qavg(const struct red_parms *p, const struct red_vars *v, unsigned int backlog) argument
/linux-master/drivers/xen/
H A Dpvcalls-front.h12 int pvcalls_front_listen(struct socket *sock, int backlog);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.h75 struct list_head backlog; member in struct:mlx5_fpga_conn::__anon1228::__anon1229
/linux-master/tools/testing/selftests/drivers/net/mlxsw/
H A Dsch_red_core.sh8 # to the backlog. Any extra packets sent should almost 1:1 go to backlog. That
9 # is what H2 is used for--it sends the extra traffic to create backlog.
315 qdisc_stats_get $swp3 $(get_qdisc_handle $vlan) .backlog
360 # This sends traffic in an attempt to build a backlog of $size. Returns 0 on
362 # backlog size to stdout.
418 local backlog
421 # Build the below-the-limit backlog using UDP. We could use TCP just
426 backlog=$(build_backlog $vlan $((2 * limit / 3)) udp)
427 check_err $? "Could not build the requested backlog"
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlxsw/
H A Dspectrum_qdisc.c79 u64 backlog; member in struct:mlxsw_sp_qdisc::mlxsw_sp_qdisc_stats
200 tmp->stats_base.backlog -= mlxsw_sp_qdisc->stats_base.backlog;
503 return xstats->backlog[tclass_num] +
504 xstats->backlog[tclass_num + 8];
561 u64 drops, u64 backlog,
569 backlog -= stats_base->backlog;
573 stats_ptr->qstats->backlog += mlxsw_sp_cells_bytes(mlxsw_sp, backlog);
558 mlxsw_sp_qdisc_update_stats(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u64 tx_bytes, u64 tx_packets, u64 drops, u64 backlog, struct tc_qopt_offload_stats *stats_ptr) argument
588 u64 backlog = 0; local
751 u64 backlog; local
1298 u64 backlog; local
1422 u64 backlog; local
1448 u64 backlog = 0; local
[all...]
/linux-master/net/sched/
H A Dsch_pie.c34 struct pie_vars *vars, u32 backlog, u32 packet_size)
54 if (backlog < 2 * mtu)
96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog,
206 struct pie_vars *vars, u32 backlog)
222 if (backlog == 0)
235 if (backlog >= QUEUE_THRESHOLD && vars->dq_count == DQCOUNT_INVALID) {
274 if (backlog < QUEUE_THRESHOLD) {
298 u32 backlog)
313 qdelay = (backlog << PIE_SCALE) / vars->avg_dq_rate;
321 /* If qdelay is zero and backlog i
33 pie_drop_early(struct Qdisc *sch, struct pie_params *params, struct pie_vars *vars, u32 backlog, u32 packet_size) argument
205 pie_process_dequeue(struct sk_buff *skb, struct pie_params *params, struct pie_vars *vars, u32 backlog) argument
297 pie_calculate_probability(struct pie_params *params, struct pie_vars *vars, u32 backlog) argument
[all...]
H A Dsch_gred.c39 u32 backlog; /* bytes on the virtualQ */ member in struct:gred_sched_data
118 return sch->qstats.backlog;
120 return q->backlog;
182 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <=
248 q->backlog += qdisc_pkt_len(skb);
273 net_warn_ratelimited("GRED: Unable to relocate VQ 0x%x after dequeue, screwing up backlog\n",
276 q->backlog -= qdisc_pkt_len(skb);
279 if (!sch->qstats.backlog)
282 if (!q->backlog)
307 q->backlog
[all...]
/linux-master/include/rdma/
H A Diw_cm.h122 * @backlog: The maximum number of outstanding un-accepted inbound listen
128 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog);

Completed in 524 milliseconds

12345678