Lines Matching defs:src_cpu

1976 	int src_cpu, src_nid;
2207 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr))
2420 .src_cpu = task_cpu(p),
2445 sd = rcu_dereference(per_cpu(sd_numa, env.src_cpu));
2528 trace_sched_stick_numa(p, env.src_cpu, NULL, -1);
2537 trace_sched_stick_numa(p, env.src_cpu, NULL, env.best_cpu);
2541 ret = migrate_swap(p, env.best_task, env.best_cpu, env.src_cpu);
2545 trace_sched_stick_numa(p, env.src_cpu, env.best_task, env.best_cpu);
5692 * Ensure that neither of the group entities corresponding to src_cpu or
5697 int src_cpu, int dest_cpu)
5701 src_cfs_rq = tg->cfs_rq[src_cpu];
6596 int src_cpu, int dest_cpu)
8792 int src_cpu;
8877 src_nid = cpu_to_node(env->src_cpu);
8936 if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu))
8953 * meet load balance goals by pulling other tasks on src_cpu.
9151 if (task_fits_cpu(p, env->src_cpu))
9768 static inline bool sched_asym(struct sched_domain *sd, int dst_cpu, int src_cpu)
9772 * if it has higher priority than @src_cpu.
9775 sched_asym_prefer(dst_cpu, src_cpu);
11132 * If @env::src_cpu is an SMT core with busy siblings, let
11137 (sched_asym_prefer(env->dst_cpu, env->src_cpu) ||
11138 !sched_use_asym_prio(env->sd, env->src_cpu));
11169 * The dst_cpu is idle and the src_cpu CPU has only 1 CFS task.
11170 * It's worth migrating the task if the src_cpu's capacity is reduced
11177 (capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100))
11307 env.src_cpu = busiest->cpu;
11357 * Revisit (affine) tasks on src_cpu that couldn't be moved to
11360 * iterate on same src_cpu is dependent on number of CPUs in our
11388 * need to continue with same src_cpu.
11617 .src_cpu = busiest_rq->cpu,