Lines Matching refs:rq
270 * Our polling bit is clear if we're not scheduled (i.e. if rq->curr !=
271 * rq->idle). This means that, if rq->idle has the polling bit set,
445 balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
454 static void wakeup_preempt_idle(struct rq *rq, struct task_struct *p, int flags)
456 resched_curr(rq);
459 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev)
463 static void set_next_task_idle(struct rq *rq, struct task_struct *next, bool first)
465 update_idle_core(rq);
466 schedstat_inc(rq->sched_goidle);
470 static struct task_struct *pick_task_idle(struct rq *rq)
472 return rq->idle;
476 struct task_struct *pick_next_task_idle(struct rq *rq)
478 struct task_struct *next = rq->idle;
480 set_next_task_idle(rq, next, true);
490 dequeue_task_idle(struct rq *rq, struct task_struct *p, int flags)
492 raw_spin_rq_unlock_irq(rq);
495 raw_spin_rq_lock_irq(rq);
503 * and everything must be accessed through the @rq and @curr passed in
506 static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued)
510 static void switched_to_idle(struct rq *rq, struct task_struct *p)
516 prio_changed_idle(struct rq *rq, struct task_struct *p, int oldprio)
521 static void update_curr_idle(struct rq *rq)