Lines Matching refs:cpumask
1311 if (!cpus_intersects(group->cpumask, p->cpus_allowed))
1314 local_group = cpu_isset(this_cpu, group->cpumask);
1319 for_each_cpu_mask(i, group->cpumask) {
1359 cpus_and(tmp, group->cpumask, p->cpus_allowed);
2435 local_group = cpu_isset(this_cpu, group->cpumask);
2438 balance_cpu = first_cpu(group->cpumask);
2445 for_each_cpu_mask(i, group->cpumask) {
2547 first_cpu(group->cpumask) <
2548 first_cpu(group_min->cpumask))) {
2563 first_cpu(group->cpumask) >
2564 first_cpu(group_leader->cpumask))) {
2709 for_each_cpu_mask(i, group->cpumask) {
5708 if (!cpu_isset(cpu, group->cpumask)) {
5728 if (!cpus_weight(group->cpumask)) {
5734 if (cpus_intersects(groupmask, group->cpumask)) {
5740 cpus_or(groupmask, groupmask, group->cpumask);
5742 cpumask_scnprintf(str, NR_CPUS, group->cpumask);
5888 * init_sched_build_groups takes the cpumask we wish to span, and a pointer
5894 * covered by the given span, and will set each group's ->cpumask correctly,
5914 sg->cpumask = CPU_MASK_NONE;
5922 cpu_set(j, sg->cpumask);
5978 * sched_domain_node_span - get a cpumask for a node's sched_domain
5979 * @node: node whose cpumask we're constructing
5982 * Given a node, construct a good cpumask for its sched_domain to span. It
6115 for_each_cpu_mask(j, sg->cpumask) {
6119 if (j != first_cpu(sd->groups->cpumask)) {
6196 if (cpu != first_cpu(sd->groups->cpumask))
6385 sg->cpumask = nodemask;
6413 sg->cpumask = tmp;
6538 * not be load balanced. If the same cpumask appears both in the