/linux-master/include/linux/ |
H A D | stop_machine.h | 6 #include <linux/cpumask.h> 38 void stop_machine_yield(const struct cpumask *cpumask); 114 int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 125 int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 144 const struct cpumask *cpus); 148 const struct cpumask *cpus) 159 stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus) 166 const struct cpumask *cpus)
|
H A D | arch_topology.h | 44 void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq, 60 void topology_set_scale_freq_source(struct scale_freq_data *data, const struct cpumask *cpus); 61 void topology_clear_scale_freq_source(enum scale_freq_source source, const struct cpumask *cpus); 70 void topology_update_thermal_pressure(const struct cpumask *cpus, 96 const struct cpumask *cpu_coregroup_mask(int cpu); 97 const struct cpumask *cpu_clustergroup_mask(int cpu);
|
H A D | pm_opp.h | 194 int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask); 195 int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); 197 void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask); 425 static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask) argument 430 static inline int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) argument 439 dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) argument 508 dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask) argument 513 dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask) argument 517 dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) argument [all...] |
/linux-master/drivers/cpuidle/ |
H A D | cpuidle-big_little.c | 144 struct cpumask *cpumask; local 147 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); 148 if (!cpumask) 153 cpumask_set_cpu(cpu, cpumask); 155 drv->cpumask = cpumask; 227 kfree(bl_idle_big_driver.cpumask); 229 kfree(bl_idle_little_driver.cpumask);
|
H A D | dt_idle_states.c | 12 #include <linux/cpumask.h> 94 * cpumask 97 const cpumask_t *cpumask) 105 * CPUidle driver cpumask. Start from next logical cpu following 106 * cpumask_first(cpumask) since that's the CPU state_node was 110 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); 111 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { 156 const cpumask_t *cpumask; local 163 * driver mask (or cpu_possible_mask if the driver cpumask i 96 idle_state_valid(struct device_node *state_node, unsigned int idx, const cpumask_t *cpumask) argument [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | smp.h | 5 #include <linux/cpumask.h> 7 #include <asm/cpumask.h> 39 void (*send_call_func_ipi)(const struct cpumask *mask); 96 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask) 118 void native_send_call_func_ipi(const struct cpumask *mask); 144 static inline struct cpumask *cpu_llc_shared_mask(int cpu) 149 static inline struct cpumask *cpu_l2c_shared_mask(int cpu) 162 static inline struct cpumask *cpu_llc_shared_mask(int cpu) 164 return (struct cpumask *)cpumask_of(0);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | affinity.h | 28 struct cpumask mask; 29 struct cpumask used; 64 struct cpumask general_intr_mask; 65 struct cpumask comp_vect_mask; 71 struct cpumask real_cpu_mask;
|
/linux-master/arch/x86/kernel/apic/ |
H A D | apic_noop.c | 16 #include <linux/cpumask.h> 24 static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { } argument 25 static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { } argument
|
H A D | local.h | 58 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector); 59 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, int vector); 65 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector); 66 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, int vector); 67 void default_send_IPI_mask_logical(const struct cpumask *mask, int vector);
|
H A D | apic_flat_64.c | 11 #include <linux/cpumask.h> 40 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) argument 42 unsigned long mask = cpumask_bits(cpumask)[0]; 48 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) argument 50 unsigned long mask = cpumask_bits(cpumask)[0];
|
/linux-master/arch/mips/include/asm/ |
H A D | smp-ops.h | 18 #include <linux/cpumask.h> 24 void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action); 51 extern void mips_smp_send_ipi_mask(const struct cpumask *mask,
|
H A D | smp.h | 17 #include <linux/cpumask.h> 117 int mips_smp_ipi_allocate(const struct cpumask *mask); 124 int mips_smp_ipi_free(const struct cpumask *mask); 133 static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask)
|
/linux-master/drivers/clocksource/ |
H A D | dummy_timer.c | 12 #include <linux/cpumask.h> 25 evt->cpumask = cpumask_of(cpu);
|
/linux-master/kernel/sched/ |
H A D | cpupri.h | 22 struct cpumask *lowest_mask); 24 struct cpumask *lowest_mask,
|
/linux-master/arch/parisc/include/asm/ |
H A D | irq.h | 11 #include <linux/cpumask.h> 48 extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest);
|
H A D | smp.h | 18 #include <linux/cpumask.h> 33 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/linux-master/arch/arm/include/asm/ |
H A D | irq.h | 32 #include <linux/cpumask.h>
|
H A D | topology.h | 7 #include <linux/cpumask.h>
|
/linux-master/arch/powerpc/include/asm/ |
H A D | mmzone.h | 12 #include <linux/cpumask.h>
|
H A D | smp.h | 17 #include <linux/cpumask.h> 114 static inline struct cpumask *cpu_sibling_mask(int cpu) 119 static inline struct cpumask *cpu_core_mask(int cpu) 124 static inline struct cpumask *cpu_l2_cache_mask(int cpu) 129 static inline struct cpumask *cpu_smallcore_mask(int cpu) 142 static inline const struct cpumask *cpu_smt_mask(int cpu) 194 static inline const struct cpumask *cpu_sibling_mask(int cpu) 199 static inline const struct cpumask *cpu_smallcore_mask(int cpu) 204 static inline const struct cpumask *cpu_l2_cache_mask(int cpu) 256 extern void arch_send_call_function_ipi_mask(const struct cpumask *mas [all...] |
/linux-master/arch/loongarch/kernel/ |
H A D | topology.c | 4 #include <linux/cpumask.h>
|
/linux-master/arch/arm64/include/asm/ |
H A D | irq.h | 7 #include <linux/cpumask.h>
|
/linux-master/lib/ |
H A D | cpumask.c | 5 #include <linux/cpumask.h> 13 * @mask: the cpumask pointer 22 unsigned int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool wrap) 45 * alloc_cpumask_var_node - allocate a struct cpumask on a given node 46 * @mask: pointer to cpumask_var_t where the cpumask is returned 51 * a nop returning a constant 1 (in <linux/cpumask.h>). 76 * alloc_bootmem_cpumask_var - allocate a struct cpumask from the bootmem arena. 77 * @mask: pointer to cpumask_var_t where the cpumask is returned 80 * a nop (in <linux/cpumask.h>). 81 * Either returns an allocated (zero-filled) cpumask, o [all...] |
/linux-master/drivers/powercap/ |
H A D | idle_inject.c | 15 * Next, the users of the idle injection framework provide a cpumask via 17 * this cpumask. 29 * all of the CPUs in the cpumask provided by the user. 68 * @cpumask: mask of CPUs affected by idle injection 89 unsigned long cpumask[]; member in struct:idle_inject_device 107 for_each_cpu_and(cpu, to_cpumask(ii_dev->cpumask), cpu_online_mask) { 230 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); 260 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); 266 * associated with the given cpumask to be parked and stay that way, so 279 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { 331 idle_inject_register_full(struct cpumask *cpumask, bool (*update)(void)) argument 383 idle_inject_register(struct cpumask *cpumask) argument [all...] |
/linux-master/kernel/ |
H A D | padata.c | 16 #include <linux/cpumask.h> 54 target_cpu = cpumask_first(pd->cpumask.pcpu); 56 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); 67 int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); 165 * run on. If it's not in the serial cpumask of @pinst 166 * (i.e. cpumask.cbcpu), this function selects a fallback CPU and if 191 if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) { 192 if (cpumask_empty(pd->cpumask.cbcpu)) 196 cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu); 198 cpu = cpumask_first(pd->cpumask 689 padata_validate_cpumask(struct padata_instance *pinst, const struct cpumask *cpumask) argument 740 padata_set_cpumask(struct padata_instance *pinst, int cpumask_type, cpumask_var_t cpumask) argument 881 struct cpumask *cpumask; local [all...] |