/linux-master/fs/ |
H A D | eventfd.c | 32 wait_queue_head_t wqh; member in struct:eventfd_ctx 36 * wakeup is performed on "wqh". If EFD_SEMAPHORE flag was not 71 spin_lock_irqsave(&ctx->wqh.lock, flags); 75 if (waitqueue_active(&ctx->wqh)) 76 wake_up_locked_poll(&ctx->wqh, EPOLLIN | mask); 78 spin_unlock_irqrestore(&ctx->wqh.lock, flags); 113 wake_up_poll(&ctx->wqh, EPOLLHUP); 124 poll_wait(file, &ctx->wqh, wait); 127 * All writes to ctx->count occur within ctx->wqh.lock. This read 128 * can be done outside ctx->wqh [all...] |
H A D | timerfd.c | 38 wait_queue_head_t wqh; member in struct:timerfd_ctx 67 spin_lock_irqsave(&ctx->wqh.lock, flags); 70 wake_up_locked_poll(&ctx->wqh, EPOLLIN); 71 spin_unlock_irqrestore(&ctx->wqh.lock, flags); 107 spin_lock_irqsave(&ctx->wqh.lock, flags); 111 wake_up_locked_poll(&ctx->wqh, EPOLLIN); 113 spin_unlock_irqrestore(&ctx->wqh.lock, flags); 255 poll_wait(file, &ctx->wqh, wait); 257 spin_lock_irqsave(&ctx->wqh.lock, flags); 260 spin_unlock_irqrestore(&ctx->wqh [all...] |
H A D | fs-writeback.c | 1514 wait_queue_head_t *wqh; variable 1516 wqh = bit_waitqueue(&inode->i_state, __I_SYNC); 1519 __wait_on_bit(wqh, &wq, bit_wait, 1544 wait_queue_head_t *wqh = bit_waitqueue(&inode->i_state, __I_SYNC); variable 1547 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); 1552 finish_wait(wqh, &wait);
|
H A D | userfaultfd.c | 913 * After no new page faults can wait on this fault_*wqh, flush 915 * the fault_*wqh. 932 wait_queue_head_t *wqh) 937 lockdep_assert_held(&wqh->lock); 940 if (!waitqueue_active(wqh)) 943 wq = list_last_entry(&wqh->head, typeof(*wq), entry); 931 find_userfault_in( wait_queue_head_t *wqh) argument
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | rtas-work-area.c | 44 struct wait_queue_head wqh; member in struct:__anon2 49 .wqh = __WAIT_QUEUE_HEAD_INITIALIZER(rwa_state.wqh), 103 wait_event(rwa_state.wqh, 123 wake_up(&rwa_state.wqh);
|
/linux-master/kernel/locking/ |
H A D | rtmutex_api.c | 163 * @wqh: The wake queue head from which to get the next lock waiter 166 struct rt_wake_q_head *wqh) 183 mark_wakeup_next_waiter(wqh, lock); 190 DEFINE_RT_WAKE_Q(wqh); 195 postunlock = __rt_mutex_futex_unlock(lock, &wqh); 199 rt_mutex_postunlock(&wqh); 479 void __sched rt_mutex_postunlock(struct rt_wake_q_head *wqh) argument 481 rt_mutex_wake_up_q(wqh); 165 __rt_mutex_futex_unlock(struct rt_mutex_base *lock, struct rt_wake_q_head *wqh) argument
|
H A D | rwbase_rt.c | 149 DEFINE_RT_WAKE_Q(wqh); 160 rt_mutex_wake_q_add_task(&wqh, owner, state); 165 rt_mutex_wake_up_q(&wqh);
|
H A D | rtmutex.c | 540 static __always_inline void rt_mutex_wake_q_add_task(struct rt_wake_q_head *wqh, argument 546 WARN_ON_ONCE(wqh->rtlock_task); 548 wqh->rtlock_task = task; 550 wake_q_add(&wqh->head, task); 554 static __always_inline void rt_mutex_wake_q_add(struct rt_wake_q_head *wqh, argument 557 rt_mutex_wake_q_add_task(wqh, w->task, w->wake_state); 560 static __always_inline void rt_mutex_wake_up_q(struct rt_wake_q_head *wqh) argument 562 if (IS_ENABLED(CONFIG_PREEMPT_RT) && wqh->rtlock_task) { 563 wake_up_state(wqh->rtlock_task, TASK_RTLOCK_WAIT); 564 put_task_struct(wqh 1308 mark_wakeup_next_waiter(struct rt_wake_q_head *wqh, struct rt_mutex_base *lock) argument [all...] |
H A D | rtmutex_common.h | 101 struct rt_wake_q_head *wqh); 103 extern void rt_mutex_postunlock(struct rt_wake_q_head *wqh);
|
/linux-master/drivers/vfio/ |
H A D | virqfd.c | 79 wait_queue_head_t *wqh, poll_table *pt) 82 add_wait_queue(wqh, &virqfd->wait); 78 virqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, poll_table *pt) argument
|
/linux-master/drivers/virt/acrn/ |
H A D | irqfd.c | 75 /* Called with wqh->lock held and interrupts disabled */ 90 /* Do shutdown work in thread to hold wqh->lock */ 96 static void hsm_irqfd_poll_func(struct file *file, wait_queue_head_t *wqh, argument 102 add_wait_queue(wqh, &irqfd->wait);
|
/linux-master/kernel/ |
H A D | seccomp.c | 214 * @wqh: A wait queue for poll if a notifier is in use. 236 wait_queue_head_t wqh; member in struct:seccomp_filter 534 if (waitqueue_active(&orig->wqh)) 535 wake_up_poll(&orig->wqh, EPOLLHUP); 689 init_waitqueue_head(&sfilter->wqh); 1130 wake_up_poll_on_current_cpu(&match->wqh, EPOLLIN | EPOLLRDNORM); 1132 wake_up_poll(&match->wqh, EPOLLIN | EPOLLRDNORM); 1483 ret = prepare_to_wait_event(&filter->wqh, &wait, TASK_INTERRUPTIBLE); 1493 finish_wait(&filter->wqh, &wait); 1540 wake_up_poll(&filter->wqh, EPOLLOU [all...] |
/linux-master/kernel/futex/ |
H A D | pi.c | 620 DEFINE_RT_WAKE_Q(wqh); 659 postunlock = __rt_mutex_futex_unlock(&pi_state->pi_mutex, &wqh); 666 rt_mutex_postunlock(&wqh);
|
/linux-master/mm/ |
H A D | internal.h | 192 wait_queue_head_t *wqh; local 194 wqh = &pgdat->reclaim_wait[VMSCAN_THROTTLE_ISOLATED]; 195 if (waitqueue_active(wqh)) 196 wake_up(wqh);
|
H A D | vmscan.c | 501 wait_queue_head_t *wqh = &pgdat->reclaim_wait[reason]; local 556 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); 558 finish_wait(wqh, &wait); 6042 wait_queue_head_t *wqh; local 6044 wqh = &pgdat->reclaim_wait[VMSCAN_THROTTLE_NOPROGRESS]; 6045 if (waitqueue_active(wqh)) 6046 wake_up(wqh);
|
H A D | memcontrol.c | 168 wait_queue_head_t *wqh; member in struct:mem_cgroup_event 4989 remove_wait_queue(event->wqh, &event->wait); 5004 * Called with wqh->lock held and interrupts disabled. 5021 * side will require wqh->lock via remove_wait_queue(), 5040 wait_queue_head_t *wqh, poll_table *pt) 5045 event->wqh = wqh; 5046 add_wait_queue(wqh, &event->wait); 5039 memcg_event_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, poll_table *pt) argument
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 154 static void vhost_poll_func(struct file *file, wait_queue_head_t *wqh, argument 160 poll->wqh = wqh; 161 add_wait_queue(wqh, &poll->wait); 197 poll->wqh = NULL; 210 if (poll->wqh) 229 if (poll->wqh) { 230 remove_wait_queue(poll->wqh, &poll->wait); 231 poll->wqh = NULL;
|
H A D | vhost.h | 43 wait_queue_head_t *wqh; member in struct:vhost_poll
|
/linux-master/virt/kvm/ |
H A D | eventfd.c | 191 * Called with wqh->lock held and interrupts disabled 236 * other side is required to acquire wqh->lock, which we hold 248 irqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, argument 253 add_wait_queue_priority(wqh, &irqfd->wait);
|
/linux-master/drivers/xen/ |
H A D | privcmd.c | 921 irqfd_poll_func(struct file *file, wait_queue_head_t *wqh, poll_table *pt) argument 926 add_wait_queue_priority(wqh, &kirqfd->wait);
|
/linux-master/kernel/rcu/ |
H A D | tree.c | 1018 struct swait_queue_head *wqh = arg; local 1020 swake_up_one(wqh); 1023 static void swake_up_one_online(struct swait_queue_head *wqh) argument 1040 wqh, 0); 1044 swake_up_one(wqh);
|