/linux-master/include/net/ |
H A D | fq_impl.h | 10 #include <net/fq.h> 16 __fq_adjust_removal(struct fq *fq, struct fq_flow *flow, unsigned int packets, argument 25 fq->backlog -= packets; 26 fq->memory_usage -= truesize; 36 idx = flow - fq->flows; 37 __clear_bit(idx, fq->flows_bitmap); 40 static void fq_adjust_removal(struct fq *fq, argument 44 __fq_adjust_removal(fq, flo 47 fq_flow_dequeue(struct fq *fq, struct fq_flow *flow) argument 63 fq_flow_drop(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) argument 90 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) argument 137 fq_flow_idx(struct fq *fq, struct sk_buff *skb) argument 144 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb) argument 165 fq_find_fattest_flow(struct fq *fq) argument 197 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func) argument 252 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) argument 273 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) argument 289 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) argument 311 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) argument 349 fq_init(struct fq *fq, int flows_cnt) argument 378 fq_reset(struct fq *fq, fq_skb_free_t free_func) argument [all...] |
H A D | ipv6_frag.h | 34 struct frag_queue *fq = container_of(q, struct frag_queue, q); local 38 fq->ecn = 0; 49 const struct inet_frag_queue *fq = data; local 51 return jhash2((const u32 *)&fq->key.v6, 59 const struct inet_frag_queue *fq = ptr; local 61 return !!memcmp(&fq->key, key, sizeof(*key)); 65 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) argument 72 if (READ_ONCE(fq->q.fqdir->dead)) 74 spin_lock(&fq->q.lock); 76 if (fq [all...] |
H A D | fq.h | 63 * struct fq - main container for fair queuing purposes 68 struct fq { struct 85 typedef struct sk_buff *fq_tin_dequeue_t(struct fq *, 89 typedef void fq_skb_free_t(struct fq *, 95 typedef bool fq_skb_filter_t(struct fq *,
|
/linux-master/block/ |
H A D | blk-flush.c | 30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a 95 struct blk_flush_queue *fq, blk_opf_t flags); 100 return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq; 153 * @fq: flush queue 161 * spin_lock_irq(fq->mq_flush_lock) 164 struct blk_flush_queue *fq, 168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; 185 fq 163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) argument 223 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); local 289 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, blk_opf_t flags) argument 358 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); local 401 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); local 481 struct blk_flush_queue *fq; local 506 blk_free_flush_queue(struct blk_flush_queue *fq) argument [all...] |
/linux-master/drivers/soc/fsl/qbman/ |
H A D | qman_test_api.c | 106 static int do_enqueues(struct qman_fq *fq) argument 112 if (qman_enqueue(fq, &fd)) { 126 struct qman_fq *fq = &fq_base; local 133 err = qman_create_fq(0, FQ_FLAGS, fq); 138 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL); 144 err = do_enqueues(fq); 149 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); 154 err = do_enqueues(fq); 159 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); 167 err = qman_volatile_dequeue(fq, VDQCR_FLAG 210 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument 227 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) argument 234 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) argument [all...] |
H A D | qman.c | 268 struct qm_mcc_fq fq; member in union:qm_mc_command 957 static inline void fq_set(struct qman_fq *fq, u32 mask) argument 959 fq->flags |= mask; 962 static inline void fq_clear(struct qman_fq *fq, u32 mask) argument 964 fq->flags &= ~mask; 967 static inline int fq_isset(struct qman_fq *fq, u32 mask) argument 969 return fq->flags & mask; 972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) argument 974 return !(fq->flags & mask); 1114 pr_debug("Allocated fq looku 1121 struct qman_fq *fq; local 1151 fq_to_tag(struct qman_fq *fq) argument 1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) argument 1491 struct qman_fq *fq; local 1574 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) argument 1611 struct qman_fq *fq; local 1798 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) argument 1828 qman_destroy_fq(struct qman_fq *fq) argument 1850 qman_fq_fqid(struct qman_fq *fq) argument 1856 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) argument 1969 qman_schedule_fq(struct qman_fq *fq) argument 2010 qman_retire_fq(struct qman_fq *fq, u32 *flags) argument 2092 qman_oos_fq(struct qman_fq *fq) argument 2130 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) argument 2155 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) argument 2226 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) argument 2246 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) argument 2256 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) argument 2269 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) argument 2315 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) argument 2952 struct qman_fq fq = { local 3010 struct qman_fq fq = { local [all...] |
/linux-master/drivers/net/ethernet/freescale/dpaa/ |
H A D | dpaa_eth_trace.h | 32 struct qman_fq *fq, 36 TP_ARGS(netdev, fq, fd), 54 __entry->fqid = fq->fqid; 78 struct qman_fq *fq, 81 TP_ARGS(netdev, fq, fd) 88 struct qman_fq *fq, 91 TP_ARGS(netdev, fq, fd) 98 struct qman_fq *fq, 101 TP_ARGS(netdev, fq, fd)
|
H A D | dpaa_eth_sysfs.c | 33 struct dpaa_fq *fq; local 40 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) { 41 switch (fq->fq_type) { 67 if (prev && (abs(fq->fqid - prev->fqid) != 1 || 78 if (prev && abs(fq->fqid - prev->fqid) == 1 && 80 last_fqid = fq->fqid; 82 first_fqid = fq->fqid; 83 last_fqid = fq->fqid; 86 prev = fq;
|
H A D | dpaa_eth.c | 636 static inline void dpaa_assign_wq(struct dpaa_fq *fq, int idx) argument 638 switch (fq->fq_type) { 641 fq->wq = 1; 645 fq->wq = 5; 649 fq->wq = 6; 655 fq->wq = 6; 659 fq->wq = 2; 663 fq->wq = 1; 667 fq->wq = 0; 676 fq 907 dpaa_setup_ingress(const struct dpaa_priv *priv, struct dpaa_fq *fq, const struct qman_fq *template) argument 918 dpaa_setup_egress(const struct dpaa_priv *priv, struct dpaa_fq *fq, struct fman_port *port, const struct qman_fq *template) argument 941 struct dpaa_fq *fq; local 1020 struct qman_fq *fq; local 1176 dpaa_fq_free_entry(struct device *dev, struct qman_fq *fq) argument 2466 rx_error_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument 2656 rx_default_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument 2815 conf_error_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument 2837 conf_dflt_dqrr(struct qman_portal *portal, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) argument 2862 egress_ern(struct qman_portal *portal, struct qman_fq *fq, const union qm_mr_entry *msg) argument [all...] |
/linux-master/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, 47 struct frag_queue *fq; local 49 fq = container_of(frag, struct frag_queue, q); 51 spin_lock(&fq->q.lock); 53 if (fq->q.flags & INET_FRAG_COMPLETE) 56 inet_frag_kill(&fq->q); 58 spin_unlock(&fq->q.lock); 59 inet_frag_put(&fq->q); 84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, argument 97 if (fq 164 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) argument 279 struct lowpan_frag_queue *fq; local 498 const struct inet_frag_queue *fq = data; local 507 const struct inet_frag_queue *fq = ptr; local [all...] |
/linux-master/net/ipv6/ |
H A D | reassembly.c | 70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, 76 struct frag_queue *fq; local 78 fq = container_of(frag, struct frag_queue, q); 80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); 106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, argument 119 if (fq->q.flags & INET_FRAG_COMPLETE) { 151 if (end < fq->q.len || 152 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq 256 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) argument 330 struct frag_queue *fq; local [all...] |
/linux-master/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 126 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, 137 struct frag_queue *fq; local 139 fq = container_of(frag, struct frag_queue, q); 141 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); 166 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, argument 175 if (fq->q.flags & INET_FRAG_COMPLETE) { 205 if (end < fq->q.len || 206 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq 314 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) argument 445 struct frag_queue *fq; local [all...] |
/linux-master/drivers/crypto/caam/ |
H A D | qi.c | 142 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, argument 225 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) argument 229 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT | 234 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid); 243 } while (fq->flags & QMAN_FQ_STATE_NE); 248 static int kill_fq(struct device *qidev, struct qman_fq *fq) argument 253 ret = qman_retire_fq(fq, &flags); 267 } while (fq->state != qman_fq_state_retired); 269 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS); 270 WARN_ON(fq 293 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) argument 621 struct qman_fq *fq; local [all...] |
/linux-master/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-eth-debugfs.c | 47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) argument 49 switch (fq->type) { 62 struct dpaa2_eth_fq *fq; local 71 fq = &priv->fq[i]; 72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt); 77 if (!fq->stats.frames && !fcnt) 81 fq->fqid, 82 fq->target_cpu, 83 fq [all...] |
H A D | dpaa2-xsk.c | 17 struct dpaa2_eth_fq *fq; local 21 fq = &priv->fq[i]; 23 if (fq->type != type) 25 if (fq->channel != ch) 28 fq->consume = consume; 106 struct dpaa2_eth_fq *fq) 127 xdp_act = dpaa2_xsk_run_xdp(priv, ch, fq, (struct dpaa2_fd *)fd, vaddr); 143 dpaa2_eth_receive_skb(priv, ch, fd, vaddr, fq, percpu_stats, skb); 399 struct dpaa2_eth_fq *fq; local 103 dpaa2_xsk_rx(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument [all...] |
H A D | dpaa2-eth.c | 335 struct dpaa2_eth_fq *fq, 350 err = priv->enqueue(priv, fq, &fds[total_enqueued], 365 struct dpaa2_eth_fq *fq) 374 enqueued = dpaa2_eth_xdp_flush(priv, fq, &fq->xdp_tx_fds); 378 fds = fq->xdp_tx_fds.fds; 383 for (i = enqueued; i < fq->xdp_tx_fds.num; i++) { 388 fq->xdp_tx_fds.num = 0; 398 struct dpaa2_eth_fq *fq; local 415 fq 334 dpaa2_eth_xdp_flush(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_eth_xdp_fds *xdp_fds) argument 363 dpaa2_eth_xdp_tx_flush(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, struct dpaa2_eth_fq *fq) argument 535 dpaa2_eth_receive_skb(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, void *vaddr, struct dpaa2_eth_fq *fq, struct rtnl_link_stats64 *percpu_stats, struct sk_buff *skb) argument 578 dpaa2_eth_rx(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument 714 struct dpaa2_eth_fq *fq = NULL; local 1119 dpaa2_eth_free_tx_fd(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, struct dpaa2_eth_fq *fq, const struct dpaa2_fd *fd, bool in_napi) argument 1404 struct dpaa2_eth_fq *fq; local 1579 dpaa2_eth_tx_conf(struct dpaa2_eth_priv *priv, struct dpaa2_eth_channel *ch, const struct dpaa2_fd *fd, struct dpaa2_eth_fq *fq) argument 1947 struct dpaa2_eth_fq *fq, *txc_fq = NULL; local 2074 struct dpaa2_eth_fq *fq; local 2231 struct dpaa2_eth_fq *fq; local 2861 struct dpaa2_eth_fq *fq; local 2900 struct dpaa2_eth_fq *fq; local 3296 struct dpaa2_eth_fq *fq; local 3562 dpaa2_eth_enqueue_qd(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_fd *fd, u8 prio, u32 num_frames __always_unused, int *frames_enqueued) argument 3578 dpaa2_eth_enqueue_fq_multiple(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq, struct dpaa2_fd *fd, u8 prio, u32 num_frames, int *frames_enqueued) argument 3637 struct dpaa2_eth_fq *fq; local 3881 dpaa2_eth_setup_rx_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument 3933 dpaa2_eth_setup_tx_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument 3981 setup_rx_err_flow(struct dpaa2_eth_priv *priv, struct dpaa2_eth_fq *fq) argument [all...] |
/linux-master/net/ipv4/ |
H A D | inet_fragment.c | 133 struct inet_frag_queue *fq = ptr; local 136 count = del_timer_sync(&fq->timer) ? 1 : 0; 138 spin_lock_bh(&fq->lock); 139 fq->flags |= INET_FRAG_DROP; 140 if (!(fq->flags & INET_FRAG_COMPLETE)) { 141 fq->flags |= INET_FRAG_COMPLETE; 143 } else if (fq->flags & INET_FRAG_HASH_DEAD) { 146 spin_unlock_bh(&fq->lock); 148 if (refcount_sub_and_test(count, &fq->refcnt)) 149 inet_frag_destroy(fq); 229 inet_frag_kill(struct inet_frag_queue *fq) argument 366 struct inet_frag_queue *fq = NULL, *prev; local [all...] |
/linux-master/drivers/iommu/ |
H A D | dma-iommu.c | 127 #define fq_ring_for_each(i, fq) \ 128 for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) & (fq)->mod_mask) 130 static inline bool fq_full(struct iova_fq *fq) argument 132 assert_spin_locked(&fq->lock); 133 return (((fq->tail + 1) & fq->mod_mask) == fq->head); 136 static inline unsigned int fq_ring_add(struct iova_fq *fq) argument 147 fq_ring_free_locked(struct iommu_dma_cookie *cookie, struct iova_fq *fq) argument 168 fq_ring_free(struct iommu_dma_cookie *cookie, struct iova_fq *fq) argument 204 struct iova_fq *fq; local 252 iommu_dma_free_fq_single(struct iova_fq *fq) argument 267 struct iova_fq *fq = per_cpu_ptr(percpu_fq, cpu); local 288 iommu_dma_init_one_fq(struct iova_fq *fq, size_t fq_size) argument [all...] |
/linux-master/net/xdp/ |
H A D | xsk_buff_pool.c | 97 pool->fq = xs->fq_tmp; 249 if (!pool->fq || !pool->cq) 279 if (pool->fq) { 280 xskq_destroy(pool->fq); 281 pool->fq = NULL; 513 if (!xskq_cons_peek_addr_unchecked(pool->fq, &addr)) { 514 pool->fq->queue_empty_descs++; 521 pool->fq->invalid_descs++; 522 xskq_cons_release(pool->fq); 537 xskq_cons_release(pool->fq); [all...] |
H A D | xsk_diag.c | 70 if (!err && pool && pool->fq) 71 err = xsk_diag_put_ring(pool->fq, 86 du.n_fill_ring_empty = xs->pool ? xskq_nb_queue_empty_descs(xs->pool->fq) : 0;
|
/linux-master/net/mac80211/ |
H A D | debugfs.c | 81 struct fq *fq = &local->fq; local 85 spin_lock_bh(&local->fq.lock); 99 fq->flows_cnt, 100 fq->backlog, 101 fq->overmemory, 102 fq->overlimit, 103 fq->collisions, 104 fq [all...] |
H A D | tx.c | 1358 struct fq *fq; local 1363 fq = &local->fq; 1368 flow = &fq->flows[cvars - local->cvars]; 1370 return fq_flow_dequeue(fq, flow); 1387 static struct sk_buff *fq_tin_dequeue_func(struct fq *fq, argument 1397 local = container_of(fq, struct ieee80211_local, fq); 1425 fq_skb_free_func(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow, struct sk_buff *skb) argument 1440 struct fq *fq = &local->fq; local 1464 fq_vlan_filter_func(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow, struct sk_buff *skb, void *data) argument 1476 struct fq *fq = &local->fq; local 1542 struct fq *fq = &local->fq; local 1575 struct fq *fq = &local->fq; local 1626 struct fq *fq = &local->fq; local 3401 struct fq *fq = &local->fq; local 3809 struct fq *fq = &local->fq; local [all...] |
/linux-master/include/soc/fsl/ |
H A D | qman.h | 300 } __packed fq; /* FQRN/FQRNI/FQRL/FQPN */ member in union:qm_mr_entry 691 struct qman_fq *fq, 699 typedef void (*qman_cb_mr)(struct qman_portal *qm, struct qman_fq *fq, 743 * struct qman_fq *fq = qman_create_fq(fqid, flags, &myfq->base); 745 * // in a dequeue callback, access extra fields from 'fq' via a cast; 746 * struct my_fq *myfq = (struct my_fq *)fq; 953 * @fq: memory for storing the 'fq', with callbacks filled in 958 * created, the caller should not touch the memory at 'fq' except as extended to 969 int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq); [all...] |
/linux-master/tools/testing/selftests/net/ |
H A D | fq_band_pktlimit.sh | 30 tc qdisc replace dev dummy0 root handle 1: fq quantum 1514 initial_quantum 1514 limit 10
|
/linux-master/tools/testing/selftests/netfilter/ |
H A D | nft_trans_stress.sh | 84 ip netns exec "$testns" taskset $mask ping -4 127.0.0.1 -fq > /dev/null & 85 ip netns exec "$testns" taskset $mask ping -6 ::1 -fq > /dev/null &
|