Lines Matching refs:wqh
32 wait_queue_head_t wqh;
36 * wakeup is performed on "wqh". If EFD_SEMAPHORE flag was not
71 spin_lock_irqsave(&ctx->wqh.lock, flags);
75 if (waitqueue_active(&ctx->wqh))
76 wake_up_locked_poll(&ctx->wqh, EPOLLIN | mask);
78 spin_unlock_irqrestore(&ctx->wqh.lock, flags);
113 wake_up_poll(&ctx->wqh, EPOLLHUP);
124 poll_wait(file, &ctx->wqh, wait);
127 * All writes to ctx->count occur within ctx->wqh.lock. This read
128 * can be done outside ctx->wqh.lock because we know that poll_wait
138 * lock ctx->wqh.lock (in poll_wait)
141 * unlock ctx->wqh.lock
159 * lock ctx->wqh.lock (in poll_wait)
161 * unlock ctx->wqh.lock
178 lockdep_assert_held(&ctx->wqh.lock);
203 spin_lock_irqsave(&ctx->wqh.lock, flags);
205 __remove_wait_queue(&ctx->wqh, wait);
206 if (*cnt != 0 && waitqueue_active(&ctx->wqh))
207 wake_up_locked_poll(&ctx->wqh, EPOLLOUT);
208 spin_unlock_irqrestore(&ctx->wqh.lock, flags);
222 spin_lock_irq(&ctx->wqh.lock);
226 spin_unlock_irq(&ctx->wqh.lock);
230 if (wait_event_interruptible_locked_irq(ctx->wqh, ctx->count)) {
231 spin_unlock_irq(&ctx->wqh.lock);
237 if (waitqueue_active(&ctx->wqh))
238 wake_up_locked_poll(&ctx->wqh, EPOLLOUT);
240 spin_unlock_irq(&ctx->wqh.lock);
260 spin_lock_irq(&ctx->wqh.lock);
265 res = wait_event_interruptible_locked_irq(ctx->wqh,
273 if (waitqueue_active(&ctx->wqh))
274 wake_up_locked_poll(&ctx->wqh, EPOLLIN);
277 spin_unlock_irq(&ctx->wqh.lock);
288 spin_lock_irq(&ctx->wqh.lock);
290 spin_unlock_irq(&ctx->wqh.lock);
401 init_waitqueue_head(&ctx->wqh);