/linux-master/kernel/ |
H A D | watchdog_buddy.c | 16 if (next_cpu >= nr_cpu_ids) 20 return nr_cpu_ids; 51 if (next_cpu < nr_cpu_ids) 75 if (next_cpu < nr_cpu_ids) 102 if (next_cpu >= nr_cpu_ids)
|
H A D | scftorture.c | 342 cpu = torture_random(trsp) % nr_cpu_ids; 349 cpu = torture_random(trsp) % nr_cpu_ids; 372 cpu = torture_random(trsp) % nr_cpu_ids; 455 cpu = scfp->cpu % nr_cpu_ids; 465 WARN_ONCE(curcpu != scfp->cpu % nr_cpu_ids, 466 "%s: Wanted CPU %d, running on %d, nr_cpu_ids = %d\n", 467 __func__, scfp->cpu, curcpu, nr_cpu_ids); 564 weight_resched1 = weight_resched == 0 ? 0 : 2 * nr_cpu_ids; 565 weight_single1 = weight_single == 0 ? 0 : 2 * nr_cpu_ids; 566 weight_single_rpc1 = weight_single_rpc == 0 ? 0 : 2 * nr_cpu_ids; [all...] |
H A D | smp.c | 244 if (WARN_ONCE(cpu < 0 || cpu >= nr_cpu_ids, "%s: cpu = %d\n", __func__, cpu)) 410 if ((unsigned)cpu >= nr_cpu_ids || !cpu_online(cpu)) { 732 for (cpu = cpumask_first_and(nodemask, mask); cpu < nr_cpu_ids; 738 /* Any online will do: smp_call_function_single handles nr_cpu_ids. */ 796 if (cpu < nr_cpu_ids) 939 if (get_option(&str, &nr_cpus) && nr_cpus > 0 && nr_cpus < nr_cpu_ids) 960 unsigned int nr_cpu_ids __read_mostly = NR_CPUS; 961 EXPORT_SYMBOL(nr_cpu_ids); variable 964 /* An arch may set nr_cpu_ids earlier if needed, so this would be redundant */ 1117 if (cpu >= nr_cpu_ids || !cpu_onlin [all...] |
/linux-master/lib/ |
H A D | cpumask_kunit.c | 44 for_each_cpu_wrap(cpu, m, nr_cpu_ids / 2) \ 68 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids, cpumask_weight(cpu_possible_mask), 70 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids, cpumask_weight(&mask_all), MASK_MSG(&mask_all)); 75 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_first(&mask_empty), MASK_MSG(&mask_empty)); 79 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_first_zero(cpu_possible_mask), 87 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids - 1, cpumask_last(cpu_possible_mask), 94 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next_zero(-1, cpu_possible_mask), 97 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next(-1, &mask_empty),
|
H A D | cpumask.c | 17 * Return: >= nr_cpu_ids on completion 139 * enumeration is O(sched_domains_numa_levels * nr_cpu_ids), while 141 * O(sched_domains_numa_levels * nr_cpu_ids * log(nr_cpu_ids)). 152 WARN_ON(cpu >= nr_cpu_ids); 167 * Return: >= nr_cpu_ids if the intersection is empty. 179 if (next < nr_cpu_ids) 190 * Return: >= nr_cpu_ids if the intersection is empty. 199 if (next < nr_cpu_ids)
|
H A D | flex_proportions.c | 89 #define PROP_BATCH (8*(1+ilog2(nr_cpu_ids))) 127 if (val < (nr_cpu_ids * PROP_BATCH))
|
/linux-master/arch/arm64/include/asm/ |
H A D | smp_plat.h | 38 for (cpu = 0; cpu < nr_cpu_ids; cpu++)
|
/linux-master/arch/arm/mach-spear/ |
H A D | platsmp.c | 102 if (ncores > nr_cpu_ids) { 104 ncores, nr_cpu_ids); 105 ncores = nr_cpu_ids;
|
/linux-master/arch/arm/mach-bcm/ |
H A D | bcm63xx_smp.c | 64 if (ncores > nr_cpu_ids) { 66 ncores, nr_cpu_ids); 67 ncores = nr_cpu_ids;
|
/linux-master/arch/x86/kernel/apic/ |
H A D | apic_common.c | 23 if (mps_cpu < nr_cpu_ids && cpu_present(mps_cpu))
|
H A D | probe_32.c | 98 if (nr_cpu_ids <= 8 || xen_pv_domain()) 121 if (nr_cpu_ids > 8 && !xen_pv_domain())
|
/linux-master/kernel/irq/ |
H A D | migration.c | 29 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) >= nr_cpu_ids) { 77 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
|
H A D | cpuhotplug.c | 39 if (cpumask_any_but(m, cpu) < nr_cpu_ids && 40 cpumask_any_and(m, cpu_online_mask) >= nr_cpu_ids) { 113 if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) {
|
H A D | ipi.c | 70 if (next < nr_cpu_ids) 72 if (next < nr_cpu_ids) { 168 if (!data || cpu >= nr_cpu_ids) 199 if (cpu >= nr_cpu_ids)
|
/linux-master/net/netfilter/ |
H A D | nf_flow_table_procfs.c | 14 for (cpu = *pos - 1; cpu < nr_cpu_ids; ++cpu) { 29 for (cpu = *pos; cpu < nr_cpu_ids; ++cpu) {
|
/linux-master/arch/x86/kernel/cpu/ |
H A D | topology.c | 241 if (apic_id != topo_info.boot_cpu_apic_id && topo_info.nr_assigned_cpus >= nr_cpu_ids) { 242 pr_warn_once("CPU limit of %d reached. Ignoring further CPUs\n", nr_cpu_ids); 382 unsigned int possible = nr_cpu_ids; 391 if (possible < nr_cpu_ids) { 433 if (WARN_ON_ONCE(assigned > nr_cpu_ids)) { 434 disabled += assigned - nr_cpu_ids; 435 assigned = nr_cpu_ids; 437 allowed = min_t(unsigned int, total, nr_cpu_ids);
|
/linux-master/arch/powerpc/kernel/ |
H A D | paca.c | 62 size_t shared_lppaca_total_size = PAGE_ALIGN(nr_cpu_ids * LPPACA_SIZE); 245 paca_nr_cpu_ids = nr_cpu_ids; 247 paca_ptrs_size = sizeof(struct paca_struct *) * nr_cpu_ids; 291 new_ptrs_size = sizeof(struct paca_struct *) * nr_cpu_ids; 296 paca_nr_cpu_ids = nr_cpu_ids; 309 paca_ptrs_size + paca_struct_size, nr_cpu_ids);
|
H A D | setup-common.c | 325 if (cpumask_next(cpu_id, cpu_online_mask) >= nr_cpu_ids) 337 if ((*pos) < nr_cpu_ids) 418 for (int i = 0; i < nthreads && cpu < nr_cpu_ids; i++) { 440 * of things like irqstacks to nr_cpu_ids rather than NR_CPUS. 460 cpu_to_phys_id = memblock_alloc(nr_cpu_ids * sizeof(u32), 464 __func__, nr_cpu_ids * sizeof(u32), __alignof__(u32)); 509 } else if (cpu >= nr_cpu_ids) { 514 if (cpu < nr_cpu_ids) 548 if (maxcpus > nr_cpu_ids) { 552 maxcpus, nr_cpu_ids); [all...] |
/linux-master/arch/arm/kernel/ |
H A D | devtree.c | 129 if (WARN(cpuidx > nr_cpu_ids, "DT /cpu %u nodes greater than " 131 cpuidx, nr_cpu_ids)) { 132 cpuidx = nr_cpu_ids;
|
/linux-master/include/linux/ |
H A D | cpumask.h | 8 * only nr_cpu_ids (<= NR_CPUS) bits are valid. 26 * You should only assume nr_cpu_ids bits of this mask are valid. This is 37 #define cpumask_pr_args(maskp) nr_cpu_ids, cpumask_bits(maskp) 40 #define nr_cpu_ids ((unsigned int)NR_CPUS) macro 42 extern unsigned int nr_cpu_ids; 48 WARN_ON(nr != nr_cpu_ids); 50 nr_cpu_ids = nr; 73 * least as big as 'nr_cpu_ids' itself is, and all our cpumask 82 #define small_cpumask_bits nr_cpu_ids 85 #define small_cpumask_bits nr_cpu_ids [all...] |
/linux-master/arch/riscv/kernel/ |
H A D | smpboot.c | 155 if (cpuid > nr_cpu_ids) 157 cpuid, nr_cpu_ids); 171 for (cpuid = 1; cpuid < nr_cpu_ids; cpuid++)
|
/linux-master/arch/arm/include/asm/ |
H A D | smp_plat.h | 83 for (cpu = 0; cpu < nr_cpu_ids; cpu++)
|
/linux-master/arch/powerpc/include/asm/ |
H A D | cputhreads.h | 37 return nr_cpu_ids >> threads_shift;
|
/linux-master/scripts/gdb/linux/ |
H A D | timerlist.py | 151 nr_cpu_ids = 1 153 nr_cpu_ids = gdb.parse_and_eval("nr_cpu_ids") 157 num_bytes = (nr_cpu_ids + 7) / 8 173 extra = nr_cpu_ids % 8
|
/linux-master/kernel/sched/ |
H A D | cpupri.c | 99 if (cpumask_any_and(&p->cpus_mask, vec->mask) >= nr_cpu_ids) 290 cp->cpu_to_pri = kcalloc(nr_cpu_ids, sizeof(int), GFP_KERNEL);
|