/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_algs_send.c | 24 void qat_alg_send_backlog(struct qat_instance_backlog *backlog) argument 28 spin_lock_bh(&backlog->lock); 29 list_for_each_entry_safe(req, tmp, &backlog->list, list) { 40 spin_unlock_bh(&backlog->lock); 45 struct qat_instance_backlog *backlog = req->backlog; local 50 if (!list_empty(&backlog->list)) 67 struct qat_instance_backlog *backlog = req->backlog; local 73 spin_lock_bh(&backlog [all...] |
H A D | qat_algs_send.h | 11 spinlock_t lock; /* protects backlog list */ 19 struct qat_instance_backlog *backlog; member in struct:qat_alg_req 23 void qat_alg_send_backlog(struct qat_instance_backlog *backlog);
|
H A D | qat_compression.h | 21 struct qat_instance_backlog backlog; member in struct:qat_compression_instance
|
H A D | qat_crypto.h | 24 struct qat_instance_backlog backlog; member in struct:qat_crypto_instance
|
/linux-master/tools/perf/ui/gtk/ |
H A D | helpline.c | 32 static int backlog; local 34 ret = vscnprintf(ui_helpline__current + backlog, 35 sizeof(ui_helpline__current) - backlog, fmt, ap); 36 backlog += ret; 40 if (ptr && (ptr - ui_helpline__current) <= backlog) { 43 backlog = 0;
|
/linux-master/tools/perf/ui/tui/ |
H A D | helpline.c | 33 static int backlog; local 36 ret = vscnprintf(ui_helpline__last_msg + backlog, 37 sizeof(ui_helpline__last_msg) - backlog, format, ap); 38 backlog += ret; 42 if (ui_helpline__last_msg[backlog - 1] == '\n') { 45 backlog = 0;
|
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-crypto-main.c | 55 struct list_head *backlog; member in struct:ccp_crypto_queue 90 struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) 95 *backlog = NULL; 110 /* Process the backlog: 112 * special precautions have to be taken when handling the backlog. 114 if (req_queue.backlog != &req_queue.cmds) { 115 /* Skip over this cmd if it is the next backlog cmd */ 116 if (req_queue.backlog == &crypto_cmd->entry) 117 req_queue.backlog = crypto_cmd->entry.next; 119 *backlog 89 ccp_crypto_cmd_complete( struct ccp_crypto_cmd *crypto_cmd, struct ccp_crypto_cmd **backlog) argument 140 struct ccp_crypto_cmd *held, *next, *backlog; local [all...] |
H A D | ccp-dev.c | 280 * backlogged) or advancement out of the backlog. If the cmd has 281 * advanced out of the backlog the "err" value of the callback 315 list_add_tail(&cmd->entry, &ccp->backlog); 378 struct ccp_cmd *backlog = NULL; local 403 if (!list_empty(&ccp->backlog)) { 404 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd, 406 list_del(&backlog->entry); 411 if (backlog) { 412 INIT_WORK(&backlog [all...] |
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | qedr_iw_cm.h | 37 int qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog);
|
/linux-master/tools/testing/selftests/net/forwarding/ |
H A D | sch_red.sh | 9 # to go to backlog. 164 qdisc_stats_get $swp3 11: .backlog 190 # This sends traffic in an attempt to build a backlog of $size. Returns 0 on 192 # backlog size to stdout. 258 local backlog 261 # Build the below-the-limit backlog using UDP. We could use TCP just 266 backlog=$(build_backlog $((2 * limit / 3)) udp) 267 check_err $? "Could not build the requested backlog" 269 check_err $? "backlog $backlog / [all...] |
/linux-master/include/uapi/linux/ |
H A D | gen_stats.h | 54 * @backlog: backlog size of queue 61 __u32 backlog; member in struct:gnet_stats_queue
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_skc_to_unix_sock.c | 13 int BPF_PROG(unix_listen, struct socket *sock, int backlog) argument
|
/linux-master/include/net/ |
H A D | fq.h | 27 * @backlog: number of bytes pending in the queue. The number of packets can be 35 u32 backlog; member in struct:fq_flow 66 * @backlog: number of packets queued across all flows 79 u32 backlog; member in struct:fq
|
H A D | pie.h | 49 * @backlog_old: queue backlog during previous qdelay calculation 127 struct pie_vars *vars, u32 backlog, u32 packet_size); 130 struct pie_vars *vars, u32 backlog); 133 u32 backlog);
|
H A D | codel_impl.h | 111 u32 *backlog, 129 *backlog <= params->mtu) { 147 u32 *backlog, 166 skb_len_func, skb_time_func, backlog, now); 176 * A large backlog might result in drop rates so high 202 backlog, now)) { 227 skb_time_func, backlog, now); 104 codel_should_drop(const struct sk_buff *skb, void *ctx, struct codel_vars *vars, struct codel_params *params, struct codel_stats *stats, codel_skb_len_t skb_len_func, codel_skb_time_t skb_time_func, u32 *backlog, codel_time_t now) argument 146 codel_dequeue(void *ctx, u32 *backlog, struct codel_params *params, struct codel_vars *vars, struct codel_stats *stats, codel_skb_len_t skb_len_func, codel_skb_time_t skb_time_func, codel_skb_drop_t drop_func, codel_skb_dequeue_t dequeue_func) argument
|
H A D | fq_impl.h | 24 flow->backlog -= bytes; 25 fq->backlog -= packets; 28 if (flow->backlog) 176 cur_len = cur->backlog; 185 unsigned int cur_len = tin->default_flow.backlog; 210 if (!flow->backlog) { 220 flow->backlog += skb->len; 224 fq->backlog++; 235 while (fq->backlog > fq->limit || oom) { 308 WARN_ON_ONCE(flow->backlog); [all...] |
H A D | inet_common.h | 42 int inet_listen(struct socket *sock, int backlog); 43 int __inet_listen_sk(struct sock *sk, int backlog);
|
H A D | red.h | 339 unsigned int backlog) 346 * qavg = qavg*(1-W) + backlog*W; 350 return v->qavg + (backlog - (v->qavg >> p->Wlog)); 355 unsigned int backlog) 358 return red_calc_qavg_no_idle_time(p, v, backlog); 337 red_calc_qavg_no_idle_time(const struct red_parms *p, const struct red_vars *v, unsigned int backlog) argument 353 red_calc_qavg(const struct red_parms *p, const struct red_vars *v, unsigned int backlog) argument
|
/linux-master/drivers/xen/ |
H A D | pvcalls-front.h | 12 int pvcalls_front_listen(struct socket *sock, int backlog);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.h | 75 struct list_head backlog; member in struct:mlx5_fpga_conn::__anon1228::__anon1229
|
/linux-master/tools/testing/selftests/drivers/net/mlxsw/ |
H A D | sch_red_core.sh | 8 # to the backlog. Any extra packets sent should almost 1:1 go to backlog. That 9 # is what H2 is used for--it sends the extra traffic to create backlog. 315 qdisc_stats_get $swp3 $(get_qdisc_handle $vlan) .backlog 360 # This sends traffic in an attempt to build a backlog of $size. Returns 0 on 362 # backlog size to stdout. 418 local backlog 421 # Build the below-the-limit backlog using UDP. We could use TCP just 426 backlog=$(build_backlog $vlan $((2 * limit / 3)) udp) 427 check_err $? "Could not build the requested backlog" [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_qdisc.c | 79 u64 backlog; member in struct:mlxsw_sp_qdisc::mlxsw_sp_qdisc_stats 200 tmp->stats_base.backlog -= mlxsw_sp_qdisc->stats_base.backlog; 503 return xstats->backlog[tclass_num] + 504 xstats->backlog[tclass_num + 8]; 561 u64 drops, u64 backlog, 569 backlog -= stats_base->backlog; 573 stats_ptr->qstats->backlog += mlxsw_sp_cells_bytes(mlxsw_sp, backlog); 558 mlxsw_sp_qdisc_update_stats(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u64 tx_bytes, u64 tx_packets, u64 drops, u64 backlog, struct tc_qopt_offload_stats *stats_ptr) argument 588 u64 backlog = 0; local 751 u64 backlog; local 1298 u64 backlog; local 1422 u64 backlog; local 1448 u64 backlog = 0; local [all...] |
/linux-master/net/sched/ |
H A D | sch_pie.c | 34 struct pie_vars *vars, u32 backlog, u32 packet_size) 54 if (backlog < 2 * mtu) 96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, 206 struct pie_vars *vars, u32 backlog) 222 if (backlog == 0) 235 if (backlog >= QUEUE_THRESHOLD && vars->dq_count == DQCOUNT_INVALID) { 274 if (backlog < QUEUE_THRESHOLD) { 298 u32 backlog) 313 qdelay = (backlog << PIE_SCALE) / vars->avg_dq_rate; 321 /* If qdelay is zero and backlog i 33 pie_drop_early(struct Qdisc *sch, struct pie_params *params, struct pie_vars *vars, u32 backlog, u32 packet_size) argument 205 pie_process_dequeue(struct sk_buff *skb, struct pie_params *params, struct pie_vars *vars, u32 backlog) argument 297 pie_calculate_probability(struct pie_params *params, struct pie_vars *vars, u32 backlog) argument [all...] |
H A D | sch_gred.c | 39 u32 backlog; /* bytes on the virtualQ */ member in struct:gred_sched_data 118 return sch->qstats.backlog; 120 return q->backlog; 182 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= 248 q->backlog += qdisc_pkt_len(skb); 273 net_warn_ratelimited("GRED: Unable to relocate VQ 0x%x after dequeue, screwing up backlog\n", 276 q->backlog -= qdisc_pkt_len(skb); 279 if (!sch->qstats.backlog) 282 if (!q->backlog) 307 q->backlog [all...] |
/linux-master/include/rdma/ |
H A D | iw_cm.h | 122 * @backlog: The maximum number of outstanding un-accepted inbound listen 128 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog);
|