Searched refs:fq (Results 1 - 25 of 59) sorted by relevance

123

/linux-master/include/net/
H A Dfq_impl.h10 #include <net/fq.h>
16 __fq_adjust_removal(struct fq *fq, struct fq_flow *flow, unsigned int packets, argument
25 fq->backlog -= packets;
26 fq->memory_usage -= truesize;
36 idx = flow - fq->flows;
37 __clear_bit(idx, fq->flows_bitmap);
40 static void fq_adjust_removal(struct fq *fq, argument
44 __fq_adjust_removal(fq, flo
47 fq_flow_dequeue(struct fq *fq, struct fq_flow *flow) argument
63 fq_flow_drop(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) argument
90 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) argument
137 fq_flow_idx(struct fq *fq, struct sk_buff *skb) argument
144 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb) argument
165 fq_find_fattest_flow(struct fq *fq) argument
197 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func) argument
252 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) argument
273 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) argument
289 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) argument
311 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) argument
349 fq_init(struct fq *fq, int flows_cnt) argument
378 fq_reset(struct fq *fq, fq_skb_free_t free_func) argument
[all...]
H A Dipv6_frag.h34 struct frag_queue *fq = container_of(q, struct frag_queue, q); local
38 fq->ecn = 0;
49 const struct inet_frag_queue *fq = data; local
51 return jhash2((const u32 *)&fq->key.v6,
59 const struct inet_frag_queue *fq = ptr; local
61 return !!memcmp(&fq->key, key, sizeof(*key));
65 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) argument
72 if (READ_ONCE(fq->q.fqdir->dead))
74 spin_lock(&fq->q.lock);
76 if (fq
[all...]
H A Dfq.h63 * struct fq - main container for fair queuing purposes
68 struct fq { struct
85 typedef struct sk_buff *fq_tin_dequeue_t(struct fq *,
89 typedef void fq_skb_free_t(struct fq *,
95 typedef bool fq_skb_filter_t(struct fq *,
/linux-master/block/
H A Dblk-flush.c30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a
95 struct blk_flush_queue *fq, blk_opf_t flags);
100 return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq;
153 * @fq: flush queue
161 * spin_lock_irq(fq->mq_flush_lock)
164 struct blk_flush_queue *fq,
168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx];
185 fq
163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) argument
223 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); local
289 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, blk_opf_t flags) argument
358 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); local
401 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); local
481 struct blk_flush_queue *fq; local
506 blk_free_flush_queue(struct blk_flush_queue *fq) argument
[all...]
/linux-master/drivers/soc/fsl/qbman/
H A Dqman_test_api.c106 static int do_enqueues(struct qman_fq *fq) argument
112 if (qman_enqueue(fq, &fd)) {
126 struct qman_fq *fq = &fq_base; local
133 err = qman_create_fq(0, FQ_FLAGS, fq);
138 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL);
144 err = do_enqueues(fq);
149 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt);
154 err = do_enqueues(fq);
159 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt);
167 err = qman_volatile_dequeue(fq, VDQCR_FLAG
210 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument
227 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) argument
234 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) argument
[all...]
H A Dqman.c268 struct qm_mcc_fq fq; member in union:qm_mc_command
957 static inline void fq_set(struct qman_fq *fq, u32 mask) argument
959 fq->flags |= mask;
962 static inline void fq_clear(struct qman_fq *fq, u32 mask) argument
964 fq->flags &= ~mask;
967 static inline int fq_isset(struct qman_fq *fq, u32 mask) argument
969 return fq->flags & mask;
972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) argument
974 return !(fq->flags & mask);
1114 pr_debug("Allocated fq looku
1121 struct qman_fq *fq; local
1151 fq_to_tag(struct qman_fq *fq) argument
1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) argument
1491 struct qman_fq *fq; local
1574 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) argument
1611 struct qman_fq *fq; local
1798 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) argument
1828 qman_destroy_fq(struct qman_fq *fq) argument
1850 qman_fq_fqid(struct qman_fq *fq) argument
1856 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) argument
1969 qman_schedule_fq(struct qman_fq *fq) argument
2010 qman_retire_fq(struct qman_fq *fq, u32 *flags) argument
2092 qman_oos_fq(struct qman_fq *fq) argument
2130 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) argument
2155 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) argument
2226 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) argument
2246 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) argument
2256 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) argument
2269 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) argument
2315 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) argument
2952 struct qman_fq fq = { local
3010 struct qman_fq fq = { local
[all...]
/linux-master/drivers/net/ethernet/freescale/dpaa/
H A Ddpaa_eth_trace.h32 struct qman_fq *fq,
36 TP_ARGS(netdev, fq, fd),
54 __entry->fqid = fq->fqid;
78 struct qman_fq *fq,
81 TP_ARGS(netdev, fq, fd)
88 struct qman_fq *fq,
91 TP_ARGS(netdev, fq, fd)
98 struct qman_fq *fq,
101 TP_ARGS(netdev, fq, fd)
H A Ddpaa_eth_sysfs.c33 struct dpaa_fq *fq; local
40 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) {
41 switch (fq->fq_type) {
67 if (prev && (abs(fq->fqid - prev->fqid) != 1 ||
78 if (prev && abs(fq->fqid - prev->fqid) == 1 &&
80 last_fqid = fq->fqid;
82 first_fqid = fq->fqid;
83 last_fqid = fq->fqid;
86 prev = fq;
H A Ddpaa_eth.c636 static inline void dpaa_assign_wq(struct dpaa_fq *fq, int idx) argument
638 switch (fq->fq_type) {
641 fq->wq = 1;
645 fq->wq = 5;
649 fq->wq = 6;
655 fq->wq = 6;
659 fq->wq = 2;
663 fq->wq = 1;
667 fq->wq = 0;
676 fq
907 dpaa_setup_ingress(const struct dpaa_priv *priv, struct dpaa_fq *fq, const struct qman_fq *template) argument
918 dpaa_setup_egress(const struct dpaa_priv *priv, struct dpaa_fq *fq, struct fman_port *port, const struct qman_fq *template) argument
941 struct dpaa_fq *fq; local
1020 struct qman_fq *fq; local
1176 dpaa_fq_free_entry(struct device *dev, struct qman_fq *fq) argument
2466 rx_error_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument
2656 rx_default_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument
2815 conf_error_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument
2837 conf_dflt_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument
2862 egress_ern(struct qman_portal *portal, struct qman_fq *fq, const union qm_mr_entry *msg) argument
[all...]
/linux-master/net/ieee802154/6lowpan/
H A Dreassembly.c33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb,
47 struct frag_queue *fq; local
49 fq = container_of(frag, struct frag_queue, q);
51 spin_lock(&fq->q.lock);
53 if (fq->q.flags & INET_FRAG_COMPLETE)
56 inet_frag_kill(&fq->q);
58 spin_unlock(&fq->q.lock);
59 inet_frag_put(&fq->q);
84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, argument
97 if (fq
164 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) argument
279 struct lowpan_frag_queue *fq; local
498 const struct inet_frag_queue *fq = data; local
507 const struct inet_frag_queue *fq = ptr; local
[all...]
/linux-master/net/ipv6/
H A Dreassembly.c70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb,
76 struct frag_queue *fq; local
78 fq = container_of(frag, struct frag_queue, q);
80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq);
106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, argument
119 if (fq->q.flags & INET_FRAG_COMPLETE) {
151 if (end < fq->q.len ||
152 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq
256 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) argument
330 struct frag_queue *fq; local
[all...]
/linux-master/net/ipv6/netfilter/
H A Dnf_conntrack_reasm.c126 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb,
137 struct frag_queue *fq; local
139 fq = container_of(frag, struct frag_queue, q);
141 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq);
166 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, argument
175 if (fq->q.flags & INET_FRAG_COMPLETE) {
205 if (end < fq->q.len ||
206 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq
314 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) argument
445 struct frag_queue *fq; local
[all...]
/linux-master/drivers/crypto/caam/
H A Dqi.c142 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, argument
225 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) argument
229 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT |
234 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid);
243 } while (fq->flags & QMAN_FQ_STATE_NE);
248 static int kill_fq(struct device *qidev, struct qman_fq *fq) argument
253 ret = qman_retire_fq(fq, &flags);
267 } while (fq->state != qman_fq_state_retired);
269 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS);
270 WARN_ON(fq
293 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) argument
621 struct qman_fq *fq; local
[all...]
/linux-master/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-eth-debugfs.c47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) argument
49 switch (fq->type) {
62 struct dpaa2_eth_fq *fq; local
71 fq = &priv->fq[i];
72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt);
77 if (!fq->stats.frames && !fcnt)
81 fq->fqid,
82 fq->target_cpu,
83 fq
[all...]
H A Ddpaa2-xsk.c17 struct dpaa2_eth_fq *fq; local
21 fq = &priv->fq[i];
23 if (fq->type != type)
25 if (fq->channel != ch)
28 fq->consume = consume;
106 struct dpaa2_eth_fq *fq)
127 xdp_act = dpaa2_xsk_run_xdp(priv, ch, fq, (struct dpaa2_fd *)fd, vaddr);
143 dpaa2_eth_receive_skb(priv, ch, fd, vaddr, fq, percpu_stats, skb);
399 struct dpaa2_eth_fq *fq; local
103 dpaa2_xsk_rx(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument
[all...]
H A Ddpaa2-eth.c335 struct dpaa2_eth_fq *fq,
350 err = priv->enqueue(priv, fq, &fds[total_enqueued],
365 struct dpaa2_eth_fq *fq)
374 enqueued = dpaa2_eth_xdp_flush(priv, fq, &fq->xdp_tx_fds);
378 fds = fq->xdp_tx_fds.fds;
383 for (i = enqueued; i < fq->xdp_tx_fds.num; i++) {
388 fq->xdp_tx_fds.num = 0;
398 struct dpaa2_eth_fq *fq; local
415 fq
334 dpaa2_eth_xdp_flush(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_eth_xdp_fds *xdp_fds) argument
363 dpaa2_eth_xdp_tx_flush(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, struct dpaa2_eth_fq *fq) argument
535 dpaa2_eth_receive_skb(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, void *vaddr, struct dpaa2_eth_fq *fq, struct rtnl_link_stats64 *percpu_stats, struct sk_buff *skb) argument
578 dpaa2_eth_rx(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument
714 struct dpaa2_eth_fq *fq = NULL; local
1119 dpaa2_eth_free_tx_fd(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, struct dpaa2_eth_fq *fq, const struct dpaa2_fd *fd, bool in_napi) argument
1404 struct dpaa2_eth_fq *fq; local
1579 dpaa2_eth_tx_conf(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument
1947 struct dpaa2_eth_fq *fq, *txc_fq = NULL; local
2074 struct dpaa2_eth_fq *fq; local
2231 struct dpaa2_eth_fq *fq; local
2861 struct dpaa2_eth_fq *fq; local
2900 struct dpaa2_eth_fq *fq; local
3296 struct dpaa2_eth_fq *fq; local
3562 dpaa2_eth_enqueue_qd(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_fd *fd, u8 prio, u32 num_frames __always_unused, int *frames_enqueued) argument
3578 dpaa2_eth_enqueue_fq_multiple(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_fd *fd, u8 prio, u32 num_frames, int *frames_enqueued) argument
3637 struct dpaa2_eth_fq *fq; local
3881 dpaa2_eth_setup_rx_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument
3933 dpaa2_eth_setup_tx_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument
3981 setup_rx_err_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument
[all...]
/linux-master/net/ipv4/
H A Dinet_fragment.c133 struct inet_frag_queue *fq = ptr; local
136 count = del_timer_sync(&fq->timer) ? 1 : 0;
138 spin_lock_bh(&fq->lock);
139 fq->flags |= INET_FRAG_DROP;
140 if (!(fq->flags & INET_FRAG_COMPLETE)) {
141 fq->flags |= INET_FRAG_COMPLETE;
143 } else if (fq->flags & INET_FRAG_HASH_DEAD) {
146 spin_unlock_bh(&fq->lock);
148 if (refcount_sub_and_test(count, &fq->refcnt))
149 inet_frag_destroy(fq);
229 inet_frag_kill(struct inet_frag_queue *fq) argument
366 struct inet_frag_queue *fq = NULL, *prev; local
[all...]
/linux-master/drivers/iommu/
H A Ddma-iommu.c127 #define fq_ring_for_each(i, fq) \
128 for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) & (fq)->mod_mask)
130 static inline bool fq_full(struct iova_fq *fq) argument
132 assert_spin_locked(&fq->lock);
133 return (((fq->tail + 1) & fq->mod_mask) == fq->head);
136 static inline unsigned int fq_ring_add(struct iova_fq *fq) argument
147 fq_ring_free_locked(struct iommu_dma_cookie *cookie, struct iova_fq *fq) argument
168 fq_ring_free(struct iommu_dma_cookie *cookie, struct iova_fq *fq) argument
204 struct iova_fq *fq; local
252 iommu_dma_free_fq_single(struct iova_fq *fq) argument
267 struct iova_fq *fq = per_cpu_ptr(percpu_fq, cpu); local
288 iommu_dma_init_one_fq(struct iova_fq *fq, size_t fq_size) argument
[all...]
/linux-master/net/xdp/
H A Dxsk_buff_pool.c97 pool->fq = xs->fq_tmp;
249 if (!pool->fq || !pool->cq)
279 if (pool->fq) {
280 xskq_destroy(pool->fq);
281 pool->fq = NULL;
513 if (!xskq_cons_peek_addr_unchecked(pool->fq, &addr)) {
514 pool->fq->queue_empty_descs++;
521 pool->fq->invalid_descs++;
522 xskq_cons_release(pool->fq);
537 xskq_cons_release(pool->fq);
[all...]
H A Dxsk_diag.c70 if (!err && pool && pool->fq)
71 err = xsk_diag_put_ring(pool->fq,
86 du.n_fill_ring_empty = xs->pool ? xskq_nb_queue_empty_descs(xs->pool->fq) : 0;
/linux-master/net/mac80211/
H A Ddebugfs.c81 struct fq *fq = &local->fq; local
85 spin_lock_bh(&local->fq.lock);
99 fq->flows_cnt,
100 fq->backlog,
101 fq->overmemory,
102 fq->overlimit,
103 fq->collisions,
104 fq
[all...]
H A Dtx.c1358 struct fq *fq; local
1363 fq = &local->fq;
1368 flow = &fq->flows[cvars - local->cvars];
1370 return fq_flow_dequeue(fq, flow);
1387 static struct sk_buff *fq_tin_dequeue_func(struct fq *fq, argument
1397 local = container_of(fq, struct ieee80211_local, fq);
1425 fq_skb_free_func(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow, struct sk_buff *skb) argument
1440 struct fq *fq = &local->fq; local
1464 fq_vlan_filter_func(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow, struct sk_buff *skb, void *data) argument
1476 struct fq *fq = &local->fq; local
1542 struct fq *fq = &local->fq; local
1575 struct fq *fq = &local->fq; local
1626 struct fq *fq = &local->fq; local
3401 struct fq *fq = &local->fq; local
3809 struct fq *fq = &local->fq; local
[all...]
/linux-master/include/soc/fsl/
H A Dqman.h300 } __packed fq; /* FQRN/FQRNI/FQRL/FQPN */ member in union:qm_mr_entry
691 struct qman_fq *fq,
699 typedef void (*qman_cb_mr)(struct qman_portal *qm, struct qman_fq *fq,
743 * struct qman_fq *fq = qman_create_fq(fqid, flags, &myfq->base);
745 * // in a dequeue callback, access extra fields from 'fq' via a cast;
746 * struct my_fq *myfq = (struct my_fq *)fq;
953 * @fq: memory for storing the 'fq', with callbacks filled in
958 * created, the caller should not touch the memory at 'fq' except as extended to
969 int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq);
[all...]
/linux-master/tools/testing/selftests/net/
H A Dfq_band_pktlimit.sh30 tc qdisc replace dev dummy0 root handle 1: fq quantum 1514 initial_quantum 1514 limit 10
/linux-master/tools/testing/selftests/netfilter/
H A Dnft_trans_stress.sh84 ip netns exec "$testns" taskset $mask ping -4 127.0.0.1 -fq > /dev/null &
85 ip netns exec "$testns" taskset $mask ping -6 ::1 -fq > /dev/null &

Completed in 359 milliseconds

123