/linux-master/include/asm-generic/ |
H A D | topology.h | 49 #define cpumask_of_node(node) ((node) == 0 ? cpu_online_mask : cpu_none_mask) 51 #define cpumask_of_node(node) ((void)(node), cpu_online_mask)
|
/linux-master/arch/mips/include/asm/mach-loongson64/ |
H A D | topology.h | 15 #define cpumask_of_pcibus(bus) (cpu_online_mask)
|
/linux-master/arch/sparc/include/asm/ |
H A D | smp_32.h | 68 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, 0, 0, 0, 0); 73 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, arg1, 0, 0, 0); 77 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, arg1, arg2, 0, 0); 83 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, 90 sparc32_ipi_ops->cross_call(func, *cpu_online_mask,
|
/linux-master/arch/sh/include/asm/ |
H A D | topology.h | 9 #define cpumask_of_node(node) ((void)node, cpu_online_mask)
|
/linux-master/arch/x86/include/asm/ |
H A D | cpumask.h | 16 return arch_test_bit(cpu, cpumask_bits(cpu_online_mask));
|
/linux-master/kernel/power/ |
H A D | poweroff.c | 29 schedule_work_on(cpumask_first(cpu_online_mask), &poweroff_work);
|
/linux-master/kernel/irq/ |
H A D | migration.c | 29 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) >= nr_cpu_ids) { 77 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
|
H A D | cpuhotplug.c | 40 cpumask_any_and(m, cpu_online_mask) >= nr_cpu_ids) { 113 if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) { 123 affinity = cpu_online_mask;
|
/linux-master/arch/x86/kernel/apic/ |
H A D | bigsmp_32.c | 23 default_send_IPI_mask_allbutself_phys(cpu_online_mask, vector); 28 default_send_IPI_mask_sequence_phys(cpu_online_mask, vector);
|
H A D | ipi.c | 59 __apic_send_IPI_mask_allbutself(cpu_online_mask, vector); 86 if (!cpumask_or_equal(mask, cpumask_of(cpu), cpu_online_mask)) 286 WARN_ON(mask & ~cpumask_bits(cpu_online_mask)[0]);
|
/linux-master/arch/x86/xen/ |
H A D | suspend_pv.c | 42 cpumask_copy(xen_cpu_initialized_map, cpu_online_mask);
|
H A D | smp.c | 150 for_each_cpu_and(cpu, mask, cpu_online_mask) 221 __xen_send_IPI_mask(cpu_online_mask, xen_vector); 242 for_each_cpu_and(cpu, mask, cpu_online_mask) { 252 xen_send_IPI_mask_allbutself(cpu_online_mask, vector);
|
/linux-master/arch/loongarch/include/asm/ |
H A D | topology.h | 19 #define cpumask_of_pcibus(bus) (cpu_online_mask)
|
/linux-master/arch/riscv/kernel/ |
H A D | unaligned_access_speed.c | 151 cpumask_and(&fast_except_me, &fast_misaligned_access, cpu_online_mask); 164 * cpu_online_mask and calling num_online_cpus, cpus_read_lock must be 170 cpumask_and(&fast_and_online, &fast_misaligned_access, cpu_online_mask); 232 for_each_cpu(cpu, cpu_online_mask) { 254 for_each_cpu(cpu, cpu_online_mask) {
|
H A D | sys_hwprobe.c | 251 cpumask_copy(&cpus, cpu_online_mask); 264 cpumask_and(&cpus, &cpus, cpu_online_mask); 312 cpumask_copy(&cpus, cpu_online_mask); 314 cpumask_and(&cpus, &cpus, cpu_online_mask); 390 hwprobe_one_pair(&pair, cpu_online_mask);
|
/linux-master/include/linux/ |
H A D | nmi.h | 160 arch_trigger_cpumask_backtrace(cpu_online_mask, -1); 166 arch_trigger_cpumask_backtrace(cpu_online_mask, exclude_cpu);
|
H A D | topology.h | 254 return cpumask_nth_and(cpu, cpus, cpu_online_mask); 272 * Yields cpu_online_mask for @node == NUMA_NO_NODE. 278 cpu_online_mask, \
|
H A D | smp.h | 71 on_each_cpu_cond_mask(NULL, func, info, wait, cpu_online_mask); 105 on_each_cpu_cond_mask(cond_func, func, info, wait, cpu_online_mask);
|
/linux-master/drivers/net/wireguard/ |
H A D | queueing.h | 110 !cpumask_test_cpu(cpu, cpu_online_mask))) { 111 cpu_index = id % cpumask_weight(cpu_online_mask); 112 cpu = cpumask_first(cpu_online_mask); 114 cpu = cpumask_next(cpu, cpu_online_mask); 127 int cpu = cpumask_next(*last_cpu, cpu_online_mask); 129 cpu = cpumask_first(cpu_online_mask);
|
/linux-master/arch/riscv/mm/ |
H A D | tlbflush.c | 112 if (cmask != cpu_online_mask) { 138 if (cmask != cpu_online_mask) 203 __flush_tlb_range(cpu_online_mask, FLUSH_TLB_NO_ASID,
|
/linux-master/drivers/cpufreq/ |
H A D | speedstep-ich.c | 263 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); 298 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | irq_affinity.c | 25 for_each_cpu_and(cpu, req_mask, cpu_online_mask) { 40 best_cpu = cpumask_first(cpu_online_mask);
|
/linux-master/arch/xtensa/kernel/ |
H A D | jump_label.c | 64 &patch, cpu_online_mask);
|
/linux-master/arch/riscv/kvm/ |
H A D | vmid.c | 109 on_each_cpu_mask(cpu_online_mask, __local_hfence_gvma_all,
|
/linux-master/lib/ |
H A D | cpumask.c | 150 cpu = sched_numa_find_nth_cpu(cpu_online_mask, i, node);
|