Searched refs:cpu_mask (Results 1 - 25 of 74) sorted by relevance

123

/linux-master/arch/arm/include/asm/
H A Dprocinfo.h28 unsigned int cpu_mask; member in struct:proc_info_list
/linux-master/arch/riscv/kernel/
H A Dsbi.c23 static int (*__sbi_rfence)(int fid, const struct cpumask *cpu_mask,
73 static unsigned long __sbi_v01_cpumask_to_hartmask(const struct cpumask *cpu_mask) argument
85 for_each_cpu(cpuid, cpu_mask) {
159 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, argument
166 if (!cpu_mask || cpumask_empty(cpu_mask))
167 cpu_mask = cpu_online_mask;
168 hart_mask = __sbi_v01_cpumask_to_hartmask(cpu_mask);
211 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, argument
302 static int __sbi_rfence_v02(int fid, const struct cpumask *cpu_mask, argument
374 sbi_remote_fence_i(const struct cpumask *cpu_mask) argument
393 sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask, unsigned long start, unsigned long size, unsigned long asid) argument
416 sbi_remote_hfence_gvma(const struct cpumask *cpu_mask, unsigned long start, unsigned long size) argument
436 sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask, unsigned long start, unsigned long size, unsigned long vmid) argument
455 sbi_remote_hfence_vvma(const struct cpumask *cpu_mask, unsigned long start, unsigned long size) argument
476 sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask, unsigned long start, unsigned long size, unsigned long asid) argument
[all...]
/linux-master/scripts/gdb/linux/
H A Dcpus.py56 cpu_mask = {} variable
60 global cpu_mask
61 cpu_mask = {}
68 global cpu_mask
70 if mask_name in cpu_mask:
71 mask = cpu_mask[mask_name]
75 cpu_mask[mask_name] = mask
/linux-master/tools/testing/selftests/bpf/prog_tests/
H A Dcpu_mask.c33 const char *cpu_mask; member in struct:__anon2393
65 err = parse_cpu_mask_str(test_cases[i].cpu_mask, &mask, &n);
/linux-master/lib/
H A Dgroup_cpus.c127 * Active CPUs means the CPUs in '@cpu_mask AND @node_to_cpumask[]'
132 const struct cpumask *cpu_mask,
147 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]);
251 const struct cpumask *cpu_mask,
260 if (cpumask_empty(cpu_mask))
263 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk);
272 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]);
287 alloc_nodes_groups(numgrps, node_to_cpumask, cpu_mask,
297 cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]);
130 alloc_nodes_groups(unsigned int numgrps, cpumask_var_t *node_to_cpumask, const struct cpumask *cpu_mask, const nodemask_t nodemsk, struct cpumask *nmsk, struct node_groups *node_groups) argument
249 __group_cpus_evenly(unsigned int startgrp, unsigned int numgrps, cpumask_var_t *node_to_cpumask, const struct cpumask *cpu_mask, struct cpumask *nmsk, struct cpumask *masks) argument
/linux-master/arch/riscv/include/asm/
H A Dsbi.h313 int sbi_remote_fence_i(const struct cpumask *cpu_mask);
315 int sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask,
319 int sbi_remote_hfence_gvma(const struct cpumask *cpu_mask,
322 int sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask,
326 int sbi_remote_hfence_vvma(const struct cpumask *cpu_mask,
329 int sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask,
369 static inline int sbi_remote_fence_i(const struct cpumask *cpu_mask) { return -1; } argument
/linux-master/tools/power/x86/amd_pstate_tracer/
H A Damd_pstate_trace.py155 def store_csv(cpu_int, time_pre_dec, time_post_dec, min_perf, des_perf, max_perf, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask):
160 if cpu_mask[cpu_int] == 0:
185 def read_trace_data(file_name, cpu_mask):
234 store_csv(cpu_int, time_pre_dec, time_post_dec, min_perf, des_perf, max_perf, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask)
239 ipt.split_csv(current_max_cpu, cpu_mask)
263 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable
297 cpu_mask[int(p)] = 1
300 cpu_mask[i] = 1
330 read_trace_data(file_name, cpu_mask)
/linux-master/arch/powerpc/platforms/pseries/
H A Dhotplug-cpu.c149 * @cpu_mask: the returned CPU mask.
154 cpumask_var_t *cpu_mask)
163 cpumask_clear(*cpu_mask);
165 cpumask_set_cpu(cpu, *cpu_mask);
189 while (!cpumask_empty(*cpu_mask)) {
190 if (cpumask_subset(*cpu_mask, candidate_mask))
193 cpumask_shift_left(*cpu_mask, *cpu_mask, nthreads);
196 if (!cpumask_empty(*cpu_mask))
215 cpumask_var_t cpu_mask; local
153 find_cpu_id_range(unsigned int nthreads, int assigned_node, cpumask_var_t *cpu_mask) argument
[all...]
/linux-master/kernel/trace/
H A Dpreemptirq_delay_test.c122 struct cpumask cpu_mask; local
125 cpumask_clear(&cpu_mask);
126 cpumask_set_cpu(cpu_affinity, &cpu_mask);
127 if (set_cpus_allowed_ptr(current, &cpu_mask))
/linux-master/arch/arm/mach-omap2/
H A Dclkt2xxx_virt_prcm_set.c42 static u16 cpu_mask; variable
82 if (!(ptr->flags & cpu_mask))
106 if (!(prcm->flags & cpu_mask))
179 if (!(prcm->flags & cpu_mask))
/linux-master/arch/arm/mm/
H A Dproc-arm9tdmi.S71 .macro arm9tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req
75 .long \cpu_mask
H A Dproc-arm7tdmi.S77 .macro arm7tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req, \
82 .long \cpu_mask
/linux-master/arch/sparc/mm/
H A Dsrmmu.c1659 cpumask_t cpu_mask; local
1660 cpumask_copy(&cpu_mask, mm_cpumask(mm));
1661 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1662 if (!cpumask_empty(&cpu_mask))
1671 cpumask_t cpu_mask; local
1672 cpumask_copy(&cpu_mask, mm_cpumask(mm));
1673 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1674 if (!cpumask_empty(&cpu_mask)) {
1691 cpumask_t cpu_mask; local
1692 cpumask_copy(&cpu_mask, mm_cpumas
1693 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1708 cpumask_t cpu_mask; local
1710 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1723 cpumask_t cpu_mask; local
1725 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1737 cpumask_t cpu_mask; local
1739 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
1762 cpumask_t cpu_mask; local
1764 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); local
[all...]
/linux-master/arch/x86/events/amd/
H A Dpower.c38 * MSR_F15H_CU_PWR_ACCUMULATOR. cpu_mask represents CPU bit map of all cores
41 static cpumask_t cpu_mask; variable
149 return cpumap_print_to_pagebuf(true, buf, &cpu_mask);
223 if (!cpumask_test_and_clear_cpu(cpu, &cpu_mask))
233 cpumask_set_cpu(target, &cpu_mask);
244 * 1) If any CPU is set at cpu_mask in the same compute unit, do
246 * 2) If no CPU is set at cpu_mask in the same compute unit,
250 * sibling mask, then it is also in cpu_mask.
254 cpumask_set_cpu(cpu, &cpu_mask);
/linux-master/arch/x86/kernel/cpu/resctrl/
H A Dctrlmondata.c277 cpumask_var_t cpu_mask)
282 cpumask_set_cpu(cpumask_any(&dom->cpu_mask), cpu_mask);
299 if (!cpumask_test_cpu(smp_processor_id(), &d->cpu_mask))
318 cpumask_var_t cpu_mask; local
325 if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL))
337 if (!apply_config(hw_dom, cfg, idx, cpu_mask))
351 if (cpumask_empty(cpu_mask))
355 on_each_cpu_mask(cpu_mask, rdt_ctrl_update, &msr_param, 1);
358 free_cpumask_var(cpu_mask);
275 apply_config(struct rdt_hw_domain *hw_dom, struct resctrl_staged_config *cfg, u32 idx, cpumask_var_t cpu_mask) argument
[all...]
H A Drdtgroup.c320 mask = &rdtgrp->plr->d->cpu_mask;
327 cpumask_pr_args(&rdtgrp->cpu_mask));
361 * Update the PGR_ASSOC MSR on all cpus in @cpu_mask,
366 update_closid_rmid(const struct cpumask *cpu_mask, struct rdtgroup *r) argument
368 on_each_cpu_mask(cpu_mask, update_cpu_closid_rmid, r, 1);
378 cpumask_andnot(tmpmask, newmask, &prgrp->cpu_mask);
385 cpumask_andnot(tmpmask, &rdtgrp->cpu_mask, newmask);
388 cpumask_or(&prgrp->cpu_mask, &prgrp->cpu_mask, tmpmask);
396 cpumask_andnot(tmpmask, newmask, &rdtgrp->cpu_mask);
2262 cpumask_var_t cpu_mask; local
2816 cpumask_var_t cpu_mask; local
[all...]
/linux-master/tools/power/x86/intel_pstate_tracer/
H A Dintel_pstate_tracer.py326 def store_csv(cpu_int, time_pre_dec, time_post_dec, core_busy, scaled, _from, _to, mperf, aperf, tsc, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask):
331 if cpu_mask[cpu_int] == 0:
345 def split_csv(current_max_cpu, cpu_mask):
350 if cpu_mask[int(index)] != 0:
420 def read_trace_data(filename, cpu_mask):
480 store_csv(cpu_int, time_pre_dec, time_post_dec, core_busy, scaled, _from, _to, mperf, aperf, tsc, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask)
486 split_csv(current_max_cpu, cpu_mask)
511 cpu_mask = zeros((MAX_CPUS,), dtype=int)
544 cpu_mask[int(p)] = 1
547 cpu_mask[
[all...]
/linux-master/drivers/irqchip/
H A Dirq-vf610-mscm-ir.c43 u16 cpu_mask; member in struct:vf610_mscm_ir_chip_data
95 WARN_ON(irsprc & ~chip_data->cpu_mask);
97 writew_relaxed(chip_data->cpu_mask,
210 mscm_ir_data->cpu_mask = 0x1 << cpuid;
H A Dirq-hip04.c272 unsigned int cpu_mask, cpu = smp_processor_id(); local
279 cpu_mask = hip04_get_cpumask(intc);
280 hip04_cpu_map[cpu] = cpu_mask;
288 hip04_cpu_map[i] &= ~cpu_mask;
/linux-master/drivers/infiniband/hw/hfi1/
H A Dnetdev.h89 struct cpumask *cpu_mask);
/linux-master/tools/perf/tests/
H A Dperf-record.c54 cpu_set_t cpu_mask; local
55 size_t cpu_mask_size = sizeof(cpu_mask);
114 err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask);
126 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) {
/linux-master/arch/x86/hyperv/
H A Dhv_apic.c201 ipi_arg.cpu_mask = 0;
217 __set_bit(vcpu, (unsigned long *)&ipi_arg.cpu_mask);
221 ipi_arg.cpu_mask);
/linux-master/drivers/soc/ti/
H A Dknav_qmss_acc.c200 struct cpumask *cpu_mask; local
207 cpu_mask = range->irqs[0].cpu_mask;
211 cpu_mask = range->irqs[queue].cpu_mask;
234 if (!ret && cpu_mask) {
235 ret = irq_set_affinity_hint(irq, cpu_mask);
/linux-master/arch/powerpc/include/asm/
H A Dfadump-internal.h53 struct cpumask cpu_mask; member in struct:fadump_crash_info_header
/linux-master/drivers/virtio/
H A Dvirtio_pci_common.h140 int vp_set_vq_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask);

Completed in 450 milliseconds

123