/linux-master/net/smc/ |
H A D | af_smc.c | 65 static struct workqueue_struct *smc_tcp_ls_wq; /* wq for tcp listen work */ 66 struct workqueue_struct *smc_hs_wq; /* wq for handshake work */ 67 struct workqueue_struct *smc_close_wq; /* wq for close work */ 785 struct socket_wq *wq; local 788 wq = rcu_dereference(smc->sk.sk_wq); 789 if (!skwq_has_sleeper(wq)) 795 wake_up_interruptible_all(&wq->wait); 800 wake_up_interruptible_sync_poll(&wq->wait, flags); 803 wake_up_interruptible_poll(&wq->wait, flags); 822 struct socket_wq *wq; local [all...] |
/linux-master/net/sctp/ |
H A D | socket.c | 9081 struct socket_wq *wq; local 9084 wq = rcu_dereference(sk->sk_wq); 9085 if (wq) { 9086 if (waitqueue_active(&wq->wait)) 9087 wake_up_interruptible(&wq->wait); 9094 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); 9271 struct socket_wq *wq; local 9276 wq = rcu_dereference(sk->sk_wq); 9277 if (skwq_has_sleeper(wq)) 9278 wake_up_interruptible_sync_poll(&wq [all...] |
/linux-master/net/rxrpc/ |
H A D | af_rxrpc.c | 64 struct socket_wq *wq = rcu_dereference(sk->sk_wq); local 66 if (skwq_has_sleeper(wq)) 67 wake_up_interruptible(&wq->wait);
|
/linux-master/net/mptcp/ |
H A D | subflow.c | 1772 struct socket_wq *wq; local 1775 wq = rcu_dereference(sk->sk_wq); 1776 if (skwq_has_sleeper(wq)) 1777 wake_up_interruptible_all(&wq->wait);
|
H A D | protocol.c | 3572 rcu_assign_pointer(sk->sk_wq, &parent->wq);
|
/linux-master/net/iucv/ |
H A D | af_iucv.c | 181 struct socket_wq *wq; local 184 wq = rcu_dereference(sk->sk_wq); 185 if (skwq_has_sleeper(wq)) 186 wake_up_interruptible_all(&wq->wait);
|
/linux-master/net/dccp/ |
H A D | output.c | 199 struct socket_wq *wq; local 202 wq = rcu_dereference(sk->sk_wq); 203 if (skwq_has_sleeper(wq)) 204 wake_up_interruptible(&wq->wait);
|
/linux-master/net/core/ |
H A D | sock.c | 2956 prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait, 2964 finish_wait(&sk->sk_lock.wq, &wait); 3323 struct socket_wq *wq; local 3326 wq = rcu_dereference(sk->sk_wq); 3327 if (skwq_has_sleeper(wq)) 3328 wake_up_interruptible_all(&wq->wait); 3334 struct socket_wq *wq; local 3337 wq = rcu_dereference(sk->sk_wq); 3338 if (skwq_has_sleeper(wq)) 3339 wake_up_interruptible_poll(&wq 3346 struct socket_wq *wq; local 3361 struct socket_wq *wq; local 3391 struct socket_wq *wq = rcu_dereference(sk->sk_wq); local [all...] |
H A D | page_pool.c | 965 static void page_pool_release_retry(struct work_struct *wq) argument 967 struct delayed_work *dwq = to_delayed_work(wq);
|
/linux-master/net/atm/ |
H A D | common.c | 91 struct socket_wq *wq; local 94 wq = rcu_dereference(sk->sk_wq); 95 if (skwq_has_sleeper(wq)) 96 wake_up(&wq->wait); 110 struct socket_wq *wq; local 115 wq = rcu_dereference(sk->sk_wq); 116 if (skwq_has_sleeper(wq)) 117 wake_up_interruptible(&wq->wait);
|
/linux-master/net/9p/ |
H A D | client.c | 310 init_waitqueue_head(&req->wq); 453 wake_up(&req->wq); 714 err = wait_event_killable(req->wq,
|
/linux-master/kernel/bpf/ |
H A D | helpers.c | 2692 __bpf_kfunc int bpf_wq_init(struct bpf_wq *wq, void *p__map, unsigned int flags) argument 2694 struct bpf_async_kern *async = (struct bpf_async_kern *)wq; 2706 __bpf_kfunc int bpf_wq_start(struct bpf_wq *wq, unsigned int flags) argument 2708 struct bpf_async_kern *async = (struct bpf_async_kern *)wq; 2723 __bpf_kfunc int bpf_wq_set_callback_impl(struct bpf_wq *wq, argument 2724 int (callback_fn)(void *map, int *key, struct bpf_wq *wq), 2729 struct bpf_async_kern *async = (struct bpf_async_kern *)wq;
|
/linux-master/io_uring/ |
H A D | io_uring.h | 11 #include "io-wq.h" 39 struct wait_queue_entry wq; member in struct:io_wait_queue
|
H A D | io_uring.c | 80 #include "io-wq.h" 517 * canceled. That will make io-wq go through the usual work cancel 918 * All execution paths but io-wq use the deferred completions by 943 * io-wq only, which holds a reference, so it cannot be the last put. 1807 /* one will be dropped by ->io_wq_free_work() after returning to io-wq */ 1815 /* either cancelled or io-wq is dying, so don't touch tctx->iowq */ 1832 * Don't allow any multishot execution from io-wq. It's more restrictive 2359 struct io_wait_queue *iowq = container_of(curr, struct io_wait_queue, wq); 2449 init_waitqueue_func_entry(&iowq.wq, io_wake_function); 2450 iowq.wq [all...] |
H A D | io-wq.c | 21 #include "io-wq.h" 35 IO_WQ_BIT_EXIT = 0, /* wq exiting */ 43 * One for each thread in a wq pool 52 struct io_wq *wq; member in struct:io_worker 135 static bool create_io_worker(struct io_wq *wq, int index); 137 static bool io_acct_cancel_pending_work(struct io_wq *wq, 141 static void io_wq_cancel_tw_create(struct io_wq *wq); 154 static inline struct io_wq_acct *io_get_acct(struct io_wq *wq, bool bound) argument 156 return &wq->acct[bound ? IO_WQ_ACCT_BOUND : IO_WQ_ACCT_UNBOUND]; 159 static inline struct io_wq_acct *io_work_get_acct(struct io_wq *wq, argument 170 io_worker_ref_put(struct io_wq *wq) argument 189 struct io_wq *wq = worker->wq; local 212 struct io_wq *wq = worker->wq; local 305 io_wq_create_worker(struct io_wq *wq, struct io_wq_acct *acct) argument 336 struct io_wq *wq; local 365 struct io_wq *wq = worker->wq; local 410 struct io_wq *wq = worker->wq; local 430 __io_worker_busy(struct io_wq *wq, struct io_worker *worker) argument 457 io_wait_on_hash(struct io_wq *wq, unsigned int hash) argument 481 struct io_wq *wq = worker->wq; variable in typeref:struct:io_wq 551 struct io_wq *wq = worker->wq; variable in typeref:struct:io_wq 631 struct io_wq *wq = worker->wq; local 727 io_init_new_worker(struct io_wq *wq, struct io_worker *worker, struct task_struct *tsk) argument 771 struct io_wq *wq; local 818 create_io_worker(struct io_wq *wq, int index) argument 863 io_wq_for_each_worker(struct io_wq *wq, bool (*func)(struct io_worker *, void *), void *data) argument 891 io_run_cancel(struct io_wq_work *work, struct io_wq *wq) argument 900 io_wq_insert_work(struct io_wq *wq, struct io_wq_work *work) argument 926 io_wq_enqueue(struct io_wq *wq, struct io_wq_work *work) argument 1017 io_wq_remove_pending(struct io_wq *wq, struct io_wq_work *work, struct io_wq_work_node *prev) argument 1036 io_acct_cancel_pending_work(struct io_wq *wq, struct io_wq_acct *acct, struct io_cb_cancel_data *match) argument 1060 io_wq_cancel_pending_work(struct io_wq *wq, struct io_cb_cancel_data *match) argument 1076 io_wq_cancel_running_work(struct io_wq *wq, struct io_cb_cancel_data *match) argument 1084 io_wq_cancel_cb(struct io_wq *wq, work_cancel_fn *cancel, void *data, bool cancel_all) argument 1126 struct io_wq *wq = container_of(wait, struct io_wq, wait); local 1145 struct io_wq *wq; local 1209 io_wq_exit_start(struct io_wq *wq) argument 1214 io_wq_cancel_tw_create(struct io_wq *wq) argument 1232 io_wq_exit_workers(struct io_wq *wq) argument 1253 io_wq_destroy(struct io_wq *wq) argument 1267 io_wq_put_and_exit(struct io_wq *wq) argument 1291 __io_wq_cpu_online(struct io_wq *wq, unsigned int cpu, bool online) argument 1306 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node); local 1313 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node); local 1337 io_wq_max_workers(struct io_wq *wq, int *new_count) argument [all...] |
H A D | Makefile | 13 obj-$(CONFIG_IO_WQ) += io-wq.o
|
/linux-master/include/trace/events/ |
H A D | btrfs.h | 1543 __field( const void *, wq ) 1551 __entry->wq = work->wq; 1557 TP_printk_btrfs("work=%p (normal_work=%p) wq=%p func=%ps ordered_func=%p", 1558 __entry->work, __entry->normal_work, __entry->wq, 1614 TP_PROTO(const struct btrfs_workqueue *wq, const char *name), 1616 TP_ARGS(wq, name), 1619 __field( const void *, wq ) 1623 TP_fast_assign_btrfs(btrfs_workqueue_owner(wq), 1624 __entry->wq [all...] |
/linux-master/include/net/ |
H A D | tls.h | 138 struct wait_queue_head wq; member in struct:tls_sw_context_rx
|
H A D | sock.h | 86 wait_queue_head_t wq; member in struct:__anon2823 1576 init_waitqueue_head(&sk->sk_lock.wq); \ 1998 rcu_assign_pointer(sk->sk_wq, &parent->wq); 2234 * @wq: struct socket_wq 2251 * wq = rcu_dereference(sk->sk_wq); 2252 * if (wq && waitqueue_active(&wq->wait)) 2253 * wake_up_interruptible(&wq->wait) 2263 static inline bool skwq_has_sleeper(struct socket_wq *wq) argument 2265 return wq [all...] |
/linux-master/include/net/9p/ |
H A D | client.h | 72 * @wq: wait_queue for the client to block on for this request 81 wait_queue_head_t wq; member in struct:p9_req_t
|
/linux-master/include/linux/mmc/ |
H A D | host.h | 473 wait_queue_head_t wq; member in struct:mmc_host
|
/linux-master/include/linux/mlx5/ |
H A D | mlx5_ifc.h | 3899 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_sqc_bits 4017 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_rqc_bits 4035 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_rmpc_bits 4518 struct mlx5_ifc_wq_bits wq; member in struct:mlx5_ifc_xrqc_bits
|
H A D | driver.h | 321 struct workqueue_struct *wq; member in struct:mlx5_cmd 440 struct workqueue_struct *wq; member in struct:mlx5_core_health 493 struct workqueue_struct *wq; member in struct:mlx5_fc_stats
|
/linux-master/fs/ |
H A D | userfaultfd.c | 69 wait_queue_entry_t wq; member in struct:userfaultfd_wait_queue 122 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, argument 130 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); 143 ret = wake_up_state(wq->private, mode); 153 * wq->private is read before calling the extern function 156 list_del_init(&wq->entry); 485 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); 486 uwq.wq.private = current; 508 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); 939 wait_queue_entry_t *wq; local 2150 wait_queue_entry_t *wq; local [all...] |
H A D | namei.c | 1671 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 1677 dentry = d_alloc_parallel(dir, name, &wq); 3421 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); 3430 dentry = d_alloc_parallel(dir, &nd->last, &wq);
|