Searched refs:cpu_present_mask (Results 1 - 25 of 33) sorted by relevance

12

/linux-master/arch/powerpc/platforms/cell/
H A Dsmp.c150 cpumask_copy(&of_spin_map, cpu_present_mask);
/linux-master/drivers/thermal/intel/
H A Dintel_powerclamp.c148 if (cpumask_equal(cpu_present_mask, mask) && idle > MAX_ALL_CPU_IDLE)
238 ret = allocate_copy_idle_injection_mask(cpu_present_mask);
588 if (cpumask_equal(cpu_present_mask, idle_injection_cpu_mask)) {
773 retval = allocate_copy_idle_injection_mask(cpu_present_mask);
/linux-master/arch/arm64/kernel/
H A Dtopology.c103 if ((cpu >= nr_cpu_ids) || !cpumask_test_cpu(cpu, cpu_present_mask))
316 if ((cpu >= nr_cpu_ids) || !cpumask_test_cpu(cpu, cpu_present_mask))
/linux-master/arch/powerpc/platforms/pseries/
H A Dsmp.c276 cpumask_copy(of_spin_mask, cpu_present_mask);
H A Dhotplug-cpu.c167 BUG_ON(!cpumask_subset(cpu_present_mask, cpu_possible_mask));
170 cpumask_xor(candidate_mask, cpu_possible_mask, cpu_present_mask);
205 * Update cpu_present_mask and paca(s) for a new cpu node. The wrinkle
/linux-master/arch/x86/kernel/apic/
H A Dipi.c44 !cpumask_equal(cpu_present_mask, &cpus_booted_once_mask)) {
/linux-master/include/linux/
H A Dcpumask.h95 * cpu_present_mask - has bit 'cpu' set iff cpu is populated
103 * life of that system boot. The cpu_present_mask is dynamic(*),
105 * cpu_online_mask is the dynamic subset of cpu_present_mask,
108 * If HOTPLUG is enabled, then cpu_present_mask varies dynamically,
110 * cpu_present_mask is just a copy of cpu_possible_mask.
112 * (*) Well, cpu_present_mask is dynamic in the hotplug case. If not
133 #define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask) macro
1012 #define for_each_present_cpu(cpu) for_each_cpu((cpu), cpu_present_mask)
1115 #define num_present_cpus() cpumask_weight(cpu_present_mask)
1130 return cpumask_test_cpu(cpu, cpu_present_mask);
[all...]
/linux-master/arch/x86/kernel/cpu/microcode/
H A Dcore.c568 for_each_cpu_and(cpu, cpu_present_mask, &cpus_booted_once_mask) {
648 for_each_cpu_and(cpu, cpu_present_mask, &cpus_booted_once_mask) {
/linux-master/arch/alpha/kernel/
H A Dsys_titan.c72 cpumask_copy(&cpm, cpu_present_mask);
H A Dprocess.c129 while (!cpumask_empty(cpu_present_mask))
H A Dsmp.c439 printk(KERN_INFO "SMP: %d CPUs probed -- cpu_present_mask = %lx\n",
440 smp_num_probed, cpumask_bits(cpu_present_mask)[0]);
/linux-master/drivers/irqchip/
H A Dirq-csky-mpintc.c148 * equal to cpu_present_mask.
151 if (cpumask_equal(mask_val, cpu_present_mask))
H A Dirq-sifive-plic.c259 for_each_cpu(cpu, cpu_present_mask) {
291 for_each_cpu(cpu, cpu_present_mask) {
/linux-master/kernel/sched/
H A Disolation.c132 if (!cpumask_intersects(cpu_present_mask, housekeeping_staging)) {
/linux-master/drivers/cpuidle/
H A Dcpuidle-powernv.c225 drv->cpumask = (struct cpumask *)cpu_present_mask;
/linux-master/arch/x86/hyperv/
H A Dmmu.c97 if (cpumask_equal(cpus, cpu_present_mask)) {
H A Dhv_apic.c132 if (!cpumask_equal(mask, cpu_present_mask) || exclude_self) {
/linux-master/lib/
H A Dgroup_cpus.c372 * Make a local cache of 'cpu_present_mask', so the two stages
373 * spread can observe consistent 'cpu_present_mask' without holding
377 * Here CPU hotplug may happen when reading `cpu_present_mask`, and
383 cpumask_copy(npresmsk, data_race(cpu_present_mask));
/linux-master/arch/powerpc/platforms/powernv/
H A Dsubcore.c304 cpumask_andnot(cpu_offline_mask, cpu_present_mask,
/linux-master/arch/loongarch/kernel/
H A Dacpi.c76 cpu = cpumask_next_zero(-1, cpu_present_mask);
H A Dsmp.c267 cpu = cpumask_next_zero(-1, cpu_present_mask);
/linux-master/arch/powerpc/sysdev/xics/
H A Dxics-common.c300 if (cpumask_equal(cpu_online_mask, cpu_present_mask))
/linux-master/arch/s390/kernel/
H A Dsmp.c737 if (pcpu_find_address(cpu_present_mask, address + i))
768 cpumask_xor(&avail, cpu_possible_mask, cpu_present_mask);
H A Dtopology.c83 cpumask_copy(&mask, cpu_present_mask);
/linux-master/drivers/hwtracing/coresight/
H A Dcoresight-etm-perf.c260 cpumask_copy(mask, cpu_present_mask);

Completed in 221 milliseconds

12