/linux-master/arch/arm64/kernel/ |
H A D | cpuidle.c | 24 struct acpi_processor *pr = per_cpu(processors, cpu);
|
/linux-master/kernel/ |
H A D | smpboot.c | 32 struct task_struct *tsk = per_cpu(idle_threads, cpu); 41 per_cpu(idle_threads, smp_processor_id()) = current; 52 struct task_struct *tsk = per_cpu(idle_threads, cpu); 59 per_cpu(idle_threads, cpu) = tsk; 282 * smpboot_register_percpu_thread - Register a per_cpu thread related 312 * smpboot_unregister_percpu_thread - Unregister a per_cpu thread related to hotplug
|
H A D | watchdog.c | 127 per_cpu(watchdog_hardlockup_touched, cpu) = true; 132 int hrint = atomic_read(&per_cpu(hrtimer_interrupts, cpu)); 134 if (per_cpu(hrtimer_interrupts_saved, cpu) == hrint) 142 per_cpu(hrtimer_interrupts_saved, cpu) = hrint; 157 if (per_cpu(watchdog_hardlockup_touched, cpu)) { 158 per_cpu(watchdog_hardlockup_touched, cpu) = false; 173 if (per_cpu(watchdog_hardlockup_warned, cpu)) 218 per_cpu(watchdog_hardlockup_warned, cpu) = true; 220 per_cpu(watchdog_hardlockup_warned, cpu) = false; 619 per_cpu(watchdog_report_t [all...] |
H A D | smp.c | 55 struct call_function_data *cfd = &per_cpu(cfd_data, cpu); 77 struct call_function_data *cfd = &per_cpu(cfd_data, cpu); 106 init_llist_head(&per_cpu(call_single_queue, i)); 248 cpu_cur_csd = smp_load_acquire(&per_cpu(cur_csd, cpux)); /* Before func and info. */ 262 *bug_id, READ_ONCE(per_cpu(cur_csd_func, cpux)), 263 READ_ONCE(per_cpu(cur_csd_info, cpux))); 269 if (atomic_cmpxchg_acquire(&per_cpu(trigger_backtrace, cpu), 1, 0)) 381 if (llist_add(node, &per_cpu(call_single_queue, cpu))) 824 if (llist_add(&csd->node.llist, &per_cpu(call_single_queue, cpu))) {
|
/linux-master/drivers/acpi/ |
H A D | processor_thermal.c | 45 per_cpu(cpufreq_thermal_reduction_step, phys_package_first_cpu(cpu)) 118 pr = per_cpu(processors, i); 165 struct acpi_processor *pr = per_cpu(processors, cpu); 189 struct acpi_processor *pr = per_cpu(processors, cpu);
|
/linux-master/arch/x86/include/asm/ |
H A D | kmsan.h | 46 return &per_cpu(metadata_array[off], cpu);
|
/linux-master/arch/x86/kernel/ |
H A D | kvmclock.c | 249 if (per_cpu(hv_clock_per_cpu, 0) && kvmclock_vsyscall) { 266 struct pvclock_vsyscall_time_info *p = per_cpu(hv_clock_per_cpu, cpu); 273 if (!cpu || (p && p != per_cpu(hv_clock_per_cpu, 0))) 284 per_cpu(hv_clock_per_cpu, cpu) = p;
|
/linux-master/drivers/cpuidle/ |
H A D | driver.c | 38 return per_cpu(cpuidle_drivers, cpu); 58 per_cpu(cpuidle_drivers, cpu) = NULL; 82 per_cpu(cpuidle_drivers, cpu) = drv; 377 struct cpuidle_device *dev = per_cpu(cpuidle_devices, cpu);
|
H A D | coupled.c | 334 call_single_data_t *csd = &per_cpu(cpuidle_coupled_poke_cb, cpu); 646 other_dev = per_cpu(cpuidle_devices, cpu); 669 csd = &per_cpu(cpuidle_coupled_poke_cb, dev->cpu); 742 dev = per_cpu(cpuidle_devices, cpu); 758 dev = per_cpu(cpuidle_devices, cpu);
|
/linux-master/arch/sparc/kernel/ |
H A D | nmi.c | 61 if (per_cpu(nmi_touch, cpu) != 1) 62 per_cpu(nmi_touch, cpu) = 1; 154 per_cpu(wd_enabled, cpu) = 0; 192 if (!per_cpu(wd_enabled, cpu))
|
/linux-master/drivers/acpi/riscv/ |
H A D | cpuidle.c | 26 struct acpi_processor *pr = per_cpu(processors, cpu);
|
/linux-master/arch/arc/include/asm/ |
H A D | mmu_context.h | 53 #define asid_cpu(cpu) per_cpu(asid_cache, cpu)
|
/linux-master/arch/x86/kernel/cpu/microcode/ |
H A D | core.c | 376 ret = per_cpu(ucode_ctrl.result, ctrl_cpu); 414 per_cpu(ucode_ctrl.ctrl, sibling) = ctrl; 424 per_cpu(ucode_ctrl.nmi_enabled, cpu) = true; 443 per_cpu(ucode_ctrl.ctrl, cpu) = SCTRL_DONE; 564 switch (per_cpu(ucode_ctrl.result, cpu)) { 663 per_cpu(ucode_ctrl, cpu) = ctrl; 672 per_cpu(ucode_ctrl, cpu) = ctrl;
|
/linux-master/arch/parisc/kernel/ |
H A D | processor.c | 161 p = &per_cpu(cpu_data, cpuid); 341 per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision; 342 per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model; 385 strscpy(cpu_name, per_cpu(cpu_data, 0).dev->name, sizeof(cpu_name)); 392 const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
|
H A D | time.c | 70 struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); 162 per_cpu(cpu_data, cpu).it_value = next_tick;
|
/linux-master/arch/mips/kernel/ |
H A D | mips-cm.c | 292 spin_lock_init(&per_cpu(cm_core_lock, cpu)); 340 spin_lock_irqsave(&per_cpu(cm_core_lock, curr_core), 341 per_cpu(cm_core_lock_flags, curr_core)); 361 spin_unlock_irqrestore(&per_cpu(cm_core_lock, curr_core), 362 per_cpu(cm_core_lock_flags, curr_core));
|
H A D | cevt-bcm1480.c | 100 struct clock_event_device *cd = &per_cpu(sibyte_hpt_clockevent, cpu); 101 unsigned char *name = per_cpu(sibyte_hpt_name, cpu);
|
H A D | cevt-sb1250.c | 99 struct clock_event_device *cd = &per_cpu(sibyte_hpt_clockevent, cpu); 100 unsigned char *name = per_cpu(sibyte_hpt_name, cpu);
|
/linux-master/kernel/trace/ |
H A D | trace_irqsoff.c | 113 if (likely(!per_cpu(tracing_cpu, cpu))) 170 per_cpu(tracing_cpu, cpu) = 0; 383 if (per_cpu(tracing_cpu, cpu)) 399 per_cpu(tracing_cpu, cpu) = 1; 414 if (unlikely(per_cpu(tracing_cpu, cpu))) 415 per_cpu(tracing_cpu, cpu) = 0;
|
/linux-master/drivers/cpufreq/ |
H A D | cpufreq_governor.c | 103 struct cpu_dbs_info *j_cdbs = &per_cpu(cpu_dbs, j); 138 struct cpu_dbs_info *j_cdbs = &per_cpu(cpu_dbs, j); 332 struct cpu_dbs_info *cdbs = &per_cpu(cpu_dbs, cpu); 368 struct cpu_dbs_info *j_cdbs = &per_cpu(cpu_dbs, j); 383 struct cpu_dbs_info *j_cdbs = &per_cpu(cpu_dbs, j); 532 struct cpu_dbs_info *j_cdbs = &per_cpu(cpu_dbs, j);
|
/linux-master/tools/perf/scripts/python/ |
H A D | parallel-perf.py | 437 def ProcessCommandOutputLines(cmd, per_cpu, fn, *x): 446 if per_cpu: 489 def SplitTimeRangesByTraceDataDensity(time_ranges, cpus, nr, cmd, file_name, per_cpu, min_size, min_interval, verbosity): 496 nr_cpus = cpus[-1] + 1 if per_cpu else 1 497 if per_cpu: 510 ProcessCommandOutputLines(cnts_cmd, per_cpu, CountSamplesByCPU, cpu_time_ranges) 573 ProcessCommandOutputLines(times_cmd, per_cpu, CalcTimeRangesByCPU, cpu_time_ranges, max_time) 730 if self.per_cpu: 752 if self.per_cpu: 772 self.orig_cmd, self.file_name, self.per_cpu, [all...] |
/linux-master/arch/microblaze/kernel/ |
H A D | setup.c | 172 per_cpu(KM, 0) = 0x1; /* We start in kernel mode */ 173 per_cpu(CURRENT_SAVE, 0) = (unsigned long)current;
|
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | interrupt.c | 180 pd = &per_cpu(ps3_private, cpu); 684 struct ps3_private *pd = &per_cpu(ps3_private, cpu); 694 struct ps3_private *pd = &per_cpu(ps3_private, cpu); 719 dump_bmp(&per_cpu(ps3_private, 0)); 720 dump_bmp(&per_cpu(ps3_private, 1)); 726 dump_bmp(&per_cpu(ps3_private, 0)); 727 dump_bmp(&per_cpu(ps3_private, 1)); 750 struct ps3_private *pd = &per_cpu(ps3_private, cpu);
|
/linux-master/kernel/locking/ |
H A D | lockdep_internals.h | 230 __cpu_lockdep_stats = &per_cpu(lockdep_stats, __cpu); \ 251 ops += per_cpu(lockdep_stats.lock_class_ops[idx], cpu);
|
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | rng.c | 110 if (per_cpu(pnv_rng, cpu) == NULL || 112 per_cpu(pnv_rng, cpu) = rng;
|