Searched refs:wq (Results 76 - 100 of 710) sorted by last modified time

1234567891011>>

/linux-master/net/smc/
H A Daf_smc.c65 static struct workqueue_struct *smc_tcp_ls_wq; /* wq for tcp listen work */
66 struct workqueue_struct *smc_hs_wq; /* wq for handshake work */
67 struct workqueue_struct *smc_close_wq; /* wq for close work */
785 struct socket_wq *wq; local
788 wq = rcu_dereference(smc->sk.sk_wq);
789 if (!skwq_has_sleeper(wq))
795 wake_up_interruptible_all(&wq->wait);
800 wake_up_interruptible_sync_poll(&wq->wait, flags);
803 wake_up_interruptible_poll(&wq->wait, flags);
822 struct socket_wq *wq; local
[all...]
/linux-master/net/sctp/
H A Dsocket.c9081 struct socket_wq *wq; local
9084 wq = rcu_dereference(sk->sk_wq);
9085 if (wq) {
9086 if (waitqueue_active(&wq->wait))
9087 wake_up_interruptible(&wq->wait);
9094 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT);
9271 struct socket_wq *wq; local
9276 wq = rcu_dereference(sk->sk_wq);
9277 if (skwq_has_sleeper(wq))
9278 wake_up_interruptible_sync_poll(&wq
[all...]
/linux-master/net/rxrpc/
H A Daf_rxrpc.c64 struct socket_wq *wq = rcu_dereference(sk->sk_wq); local
66 if (skwq_has_sleeper(wq))
67 wake_up_interruptible(&wq->wait);
/linux-master/net/mptcp/
H A Dsubflow.c1772 struct socket_wq *wq; local
1775 wq = rcu_dereference(sk->sk_wq);
1776 if (skwq_has_sleeper(wq))
1777 wake_up_interruptible_all(&wq->wait);
H A Dprotocol.c3572 rcu_assign_pointer(sk->sk_wq, &parent->wq);
/linux-master/net/iucv/
H A Daf_iucv.c181 struct socket_wq *wq; local
184 wq = rcu_dereference(sk->sk_wq);
185 if (skwq_has_sleeper(wq))
186 wake_up_interruptible_all(&wq->wait);
/linux-master/net/dccp/
H A Doutput.c199 struct socket_wq *wq; local
202 wq = rcu_dereference(sk->sk_wq);
203 if (skwq_has_sleeper(wq))
204 wake_up_interruptible(&wq->wait);
/linux-master/net/core/
H A Dsock.c2956 prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
2964 finish_wait(&sk->sk_lock.wq, &wait);
3323 struct socket_wq *wq; local
3326 wq = rcu_dereference(sk->sk_wq);
3327 if (skwq_has_sleeper(wq))
3328 wake_up_interruptible_all(&wq->wait);
3334 struct socket_wq *wq; local
3337 wq = rcu_dereference(sk->sk_wq);
3338 if (skwq_has_sleeper(wq))
3339 wake_up_interruptible_poll(&wq
3346 struct socket_wq *wq; local
3361 struct socket_wq *wq; local
3391 struct socket_wq *wq = rcu_dereference(sk->sk_wq); local
[all...]
H A Dpage_pool.c965 static void page_pool_release_retry(struct work_struct *wq) argument
967 struct delayed_work *dwq = to_delayed_work(wq);
/linux-master/net/atm/
H A Dcommon.c91 struct socket_wq *wq; local
94 wq = rcu_dereference(sk->sk_wq);
95 if (skwq_has_sleeper(wq))
96 wake_up(&wq->wait);
110 struct socket_wq *wq; local
115 wq = rcu_dereference(sk->sk_wq);
116 if (skwq_has_sleeper(wq))
117 wake_up_interruptible(&wq->wait);
/linux-master/net/9p/
H A Dclient.c310 init_waitqueue_head(&req->wq);
453 wake_up(&req->wq);
714 err = wait_event_killable(req->wq,
/linux-master/kernel/bpf/
H A Dhelpers.c2692 __bpf_kfunc int bpf_wq_init(struct bpf_wq *wq, void *p__map, unsigned int flags) argument
2694 struct bpf_async_kern *async = (struct bpf_async_kern *)wq;
2706 __bpf_kfunc int bpf_wq_start(struct bpf_wq *wq, unsigned int flags) argument
2708 struct bpf_async_kern *async = (struct bpf_async_kern *)wq;
2723 __bpf_kfunc int bpf_wq_set_callback_impl(struct bpf_wq *wq, argument
2724 int (callback_fn)(void *map, int *key, struct bpf_wq *wq),
2729 struct bpf_async_kern *async = (struct bpf_async_kern *)wq;
/linux-master/io_uring/
H A Dio_uring.h11 #include "io-wq.h"
39 struct wait_queue_entry wq; member in struct:io_wait_queue
H A Dio_uring.c80 #include "io-wq.h"
517 * canceled. That will make io-wq go through the usual work cancel
918 * All execution paths but io-wq use the deferred completions by
943 * io-wq only, which holds a reference, so it cannot be the last put.
1807 /* one will be dropped by ->io_wq_free_work() after returning to io-wq */
1815 /* either cancelled or io-wq is dying, so don't touch tctx->iowq */
1832 * Don't allow any multishot execution from io-wq. It's more restrictive
2359 struct io_wait_queue *iowq = container_of(curr, struct io_wait_queue, wq);
2449 init_waitqueue_func_entry(&iowq.wq, io_wake_function);
2450 iowq.wq
[all...]
H A Dio-wq.c21 #include "io-wq.h"
35 IO_WQ_BIT_EXIT = 0, /* wq exiting */
43 * One for each thread in a wq pool
52 struct io_wq *wq; member in struct:io_worker
135 static bool create_io_worker(struct io_wq *wq, int index);
137 static bool io_acct_cancel_pending_work(struct io_wq *wq,
141 static void io_wq_cancel_tw_create(struct io_wq *wq);
154 static inline struct io_wq_acct *io_get_acct(struct io_wq *wq, bool bound) argument
156 return &wq->acct[bound ? IO_WQ_ACCT_BOUND : IO_WQ_ACCT_UNBOUND];
159 static inline struct io_wq_acct *io_work_get_acct(struct io_wq *wq, argument
170 io_worker_ref_put(struct io_wq *wq) argument
189 struct io_wq *wq = worker->wq; local
212 struct io_wq *wq = worker->wq; local
305 io_wq_create_worker(struct io_wq *wq, struct io_wq_acct *acct) argument
336 struct io_wq *wq; local
365 struct io_wq *wq = worker->wq; local
410 struct io_wq *wq = worker->wq; local
430 __io_worker_busy(struct io_wq *wq, struct io_worker *worker) argument
457 io_wait_on_hash(struct io_wq *wq, unsigned int hash) argument
481 struct io_wq *wq = worker->wq; variable in typeref:struct:io_wq
551 struct io_wq *wq = worker->wq; variable in typeref:struct:io_wq
631 struct io_wq *wq = worker->wq; local
727 io_init_new_worker(struct io_wq *wq, struct io_worker *worker, struct task_struct *tsk) argument
771 struct io_wq *wq; local
818 create_io_worker(struct io_wq *wq, int index) argument
863 io_wq_for_each_worker(struct io_wq *wq, bool (*func)(struct io_worker *, void *), void *data) argument
891 io_run_cancel(struct io_wq_work *work, struct io_wq *wq) argument
900 io_wq_insert_work(struct io_wq *wq, struct io_wq_work *work) argument
926 io_wq_enqueue(struct io_wq *wq, struct io_wq_work *work) argument
1017 io_wq_remove_pending(struct io_wq *wq, struct io_wq_work *work, struct io_wq_work_node *prev) argument
1036 io_acct_cancel_pending_work(struct io_wq *wq, struct io_wq_acct *acct, struct io_cb_cancel_data *match) argument
1060 io_wq_cancel_pending_work(struct io_wq *wq, struct io_cb_cancel_data *match) argument
1076 io_wq_cancel_running_work(struct io_wq *wq, struct io_cb_cancel_data *match) argument
1084 io_wq_cancel_cb(struct io_wq *wq, work_cancel_fn *cancel, void *data, bool cancel_all) argument
1126 struct io_wq *wq = container_of(wait, struct io_wq, wait); local
1145 struct io_wq *wq; local
1209 io_wq_exit_start(struct io_wq *wq) argument
1214 io_wq_cancel_tw_create(struct io_wq *wq) argument
1232 io_wq_exit_workers(struct io_wq *wq) argument
1253 io_wq_destroy(struct io_wq *wq) argument
1267 io_wq_put_and_exit(struct io_wq *wq) argument
1291 __io_wq_cpu_online(struct io_wq *wq, unsigned int cpu, bool online) argument
1306 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node); local
1313 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node); local
1337 io_wq_max_workers(struct io_wq *wq, int *new_count) argument
[all...]
H A DMakefile13 obj-$(CONFIG_IO_WQ) += io-wq.o
/linux-master/include/trace/events/
H A Dbtrfs.h1543 __field( const void *, wq )
1551 __entry->wq = work->wq;
1557 TP_printk_btrfs("work=%p (normal_work=%p) wq=%p func=%ps ordered_func=%p",
1558 __entry->work, __entry->normal_work, __entry->wq,
1614 TP_PROTO(const struct btrfs_workqueue *wq, const char *name),
1616 TP_ARGS(wq, name),
1619 __field( const void *, wq )
1623 TP_fast_assign_btrfs(btrfs_workqueue_owner(wq),
1624 __entry->wq
[all...]
/linux-master/include/net/
H A Dtls.h138 struct wait_queue_head wq; member in struct:tls_sw_context_rx
H A Dsock.h86 wait_queue_head_t wq; member in struct:__anon2823
1576 init_waitqueue_head(&sk->sk_lock.wq); \
1998 rcu_assign_pointer(sk->sk_wq, &parent->wq);
2234 * @wq: struct socket_wq
2251 * wq = rcu_dereference(sk->sk_wq);
2252 * if (wq && waitqueue_active(&wq->wait))
2253 * wake_up_interruptible(&wq->wait)
2263 static inline bool skwq_has_sleeper(struct socket_wq *wq) argument
2265 return wq
[all...]
/linux-master/include/net/9p/
H A Dclient.h72 * @wq: wait_queue for the client to block on for this request
81 wait_queue_head_t wq; member in struct:p9_req_t
/linux-master/include/linux/mmc/
H A Dhost.h473 wait_queue_head_t wq; member in struct:mmc_host
/linux-master/include/linux/mlx5/
H A Dmlx5_ifc.h3899 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_sqc_bits
4017 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_rqc_bits
4035 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_rmpc_bits
4518 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_xrqc_bits
H A Ddriver.h321 struct workqueue_struct *wq; member in struct:mlx5_cmd
440 struct workqueue_struct *wq; member in struct:mlx5_core_health
493 struct workqueue_struct *wq; member in struct:mlx5_fc_stats
/linux-master/fs/
H A Duserfaultfd.c69 wait_queue_entry_t wq; member in struct:userfaultfd_wait_queue
122 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, argument
130 uwq = container_of(wq, struct userfaultfd_wait_queue, wq);
143 ret = wake_up_state(wq->private, mode);
153 * wq->private is read before calling the extern function
156 list_del_init(&wq->entry);
485 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function);
486 uwq.wq.private = current;
508 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq);
939 wait_queue_entry_t *wq; local
2150 wait_queue_entry_t *wq; local
[all...]
H A Dnamei.c1671 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq);
1677 dentry = d_alloc_parallel(dir, name, &wq);
3421 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq);
3430 dentry = d_alloc_parallel(dir, &nd->last, &wq);

Completed in 530 milliseconds

1234567891011>>