Lines Matching refs:waiter

163  * @wqh:	The wake queue head from which to get the next lock waiter
276 * @waiter: the pre-initialized rt_mutex_waiter
279 * Starts the rt_mutex acquire; it enqueues the @waiter and does deadlock
282 * NOTE: does _NOT_ remove the @waiter on failure; must either call
293 struct rt_mutex_waiter *waiter,
304 ret = task_blocks_on_rt_mutex(lock, waiter, task, NULL,
312 * pi chain. Let the waiter sort it out.
323 * @waiter: the pre-initialized rt_mutex_waiter
326 * Starts the rt_mutex acquire; it enqueues the @waiter and does deadlock
329 * NOTE: unlike __rt_mutex_start_proxy_lock this _DOES_ remove the @waiter
340 struct rt_mutex_waiter *waiter,
346 ret = __rt_mutex_start_proxy_lock(lock, waiter, task);
348 remove_waiter(lock, waiter);
359 * @waiter: the pre-initialized rt_mutex_waiter
373 struct rt_mutex_waiter *waiter)
380 ret = rt_mutex_slowlock_block(lock, NULL, TASK_INTERRUPTIBLE, to, waiter);
382 * try_to_take_rt_mutex() sets the waiter bit unconditionally. We might
394 * @waiter: the pre-initialized rt_mutex_waiter
412 struct rt_mutex_waiter *waiter)
424 * we will own the lock and it will have removed the waiter. If we
428 try_to_take_rt_mutex(lock, current, waiter);
434 remove_waiter(lock, waiter);
438 * try_to_take_rt_mutex() sets the waiter bit unconditionally. We might
455 struct rt_mutex_waiter *waiter;
461 waiter = task->pi_blocked_on;
462 if (!waiter || rt_waiter_node_equal(&waiter->tree, task_to_waiter_node(task))) {
466 next_lock = waiter->lock;
477 * Performs the wakeup of the top-waiter and re-enables preemption.