• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-WNDR4500-V1.0.1.40_1.0.68/src/linux/linux-2.6/kernel/

Lines Matching refs:this_cpu

1301 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
1317 local_group = cpu_isset(this_cpu, group->cpumask);
1356 find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
1369 if (load < min_load || (load == min_load && i == this_cpu)) {
1509 int cpu, this_cpu, success = 0;
1528 this_cpu = smp_processor_id();
1537 if (cpu == this_cpu) {
1542 for_each_domain(this_cpu, sd) {
1550 if (unlikely(!cpu_isset(this_cpu, p->cpus_allowed)))
1563 this_load = target_load(this_cpu, idx);
1565 new_cpu = this_cpu; /* Wake to this CPU if we can */
1571 tl_per_task = cpu_avg_load_per_task(this_cpu);
1606 new_cpu = cpu; /* Could not wake to this_cpu. Wake to cpu instead */
1620 this_cpu = smp_processor_id();
1644 activate_task(p, rq, cpu == this_cpu);
1653 if (!sync || cpu != this_cpu) {
1757 int this_cpu, cpu;
1761 this_cpu = smp_processor_id();
1775 if (likely(cpu == this_cpu)) {
1797 * We skip the following code due to cpu == this_cpu
1804 this_rq = cpu_rq(this_cpu);
2176 int new_cpu, this_cpu = get_cpu();
2177 new_cpu = sched_balance_self(this_cpu, SD_BALANCE_EXEC);
2179 if (new_cpu != this_cpu)
2189 struct prio_array *this_array, int this_cpu)
2193 set_task_cpu(p, this_cpu);
2207 * can_migrate_task - may task p from runqueue rq be migrated to this_cpu?
2210 int can_migrate_task(struct task_struct *p, struct rq *rq, int this_cpu,
2220 if (!cpu_isset(this_cpu, p->cpus_allowed))
2255 static int move_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
2330 !can_migrate_task(tmp, busiest, this_cpu, sd, idle, &pinned)) {
2339 pull_task(busiest, array, tmp, this_rq, dst_array, this_cpu);
2374 find_busiest_group(struct sched_domain *sd, int this_cpu,
2408 local_group = cpu_isset(this_cpu, group->cpumask);
2448 if (local_group && balance_cpu != this_cpu && balance) {
2697 * Check this_cpu to ensure it is balanced within domain. Attempt to move
2700 static int load_balance(int this_cpu, struct rq *this_rq,
2724 group = find_busiest_group(sd, this_cpu, &imbalance, idle, &sd_idle,
2755 nr_moved = move_tasks(this_rq, this_cpu, busiest,
2764 if (nr_moved && this_cpu != smp_processor_id())
2765 resched_cpu(this_cpu);
2785 * task on busiest cpu can't be moved to this_cpu
2787 if (!cpu_isset(this_cpu, busiest->curr->cpus_allowed)) {
2795 busiest->push_cpu = this_cpu;
2848 * Check this_cpu to ensure it is balanced within domain. Attempt to move
2855 load_balance_newidle(int this_cpu, struct rq *this_rq, struct sched_domain *sd)
2876 group = find_busiest_group(sd, this_cpu, &imbalance, NEWLY_IDLE,
2898 nr_moved = move_tasks(this_rq, this_cpu, busiest,
2931 * idle_balance is called by schedule() if this_cpu is about to become
2934 static void idle_balance(int this_cpu, struct rq *this_rq)
2940 for_each_domain(this_cpu, sd) {
2948 pulled_task = load_balance_newidle(this_cpu,
5351 int this_cpu = smp_processor_id();
5352 struct rq *rq = cpu_rq(this_cpu);
5357 BUG_ON(cpu_online(this_cpu));