/linux-master/arch/x86/include/asm/ |
H A D | processor.h | 188 #define cpu_data(cpu) per_cpu(cpu_info, cpu) macro
|
/linux-master/arch/x86/kernel/cpu/ |
H A D | amd.c | 1199 struct cpuinfo_x86 *c = &cpu_data(smp_processor_id());
|
/linux-master/drivers/acpi/ |
H A D | cppc_acpi.c | 471 * @cpu_data: Pointer to CPU specific CPPC data including PSD info. 475 int acpi_get_psd_map(unsigned int cpu, struct cppc_cpudata *cpu_data) argument 491 cpumask_set_cpu(cpu, cpu_data->shared_cpu_map); 498 cpu_data->shared_type = CPUFREQ_SHARED_TYPE_ALL; 500 cpu_data->shared_type = CPUFREQ_SHARED_TYPE_HW; 502 cpu_data->shared_type = CPUFREQ_SHARED_TYPE_ANY; 523 cpumask_set_cpu(i, cpu_data->shared_cpu_map); 530 cpumask_clear(cpu_data->shared_cpu_map); 531 cpumask_set_cpu(cpu, cpu_data->shared_cpu_map); 532 cpu_data [all...] |
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 9673 struct cpuinfo_x86 *c = &cpu_data(cpu);
|
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 523 struct cpuinfo_x86 *c = &cpu_data(cpu);
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_device.c | 1632 struct cpuinfo_x86 *c = &cpu_data(0);
|
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | core.c | 129 m->socketid = cpu_data(m->extcpu).topo.pkg_id; 130 m->apicid = cpu_data(m->extcpu).topo.initial_apicid; 132 m->ppin = cpu_data(m->extcpu).ppin;
|
/linux-master/arch/x86/kernel/ |
H A D | smpboot.c | 215 * Calibrate the delay loop and update loops_per_jiffy in cpu_data. 224 cpu_data(smp_processor_id()).loops_per_jiffy = loops_per_jiffy; 322 struct cpuinfo_x86 *c = &cpu_data(id); 556 struct cpuinfo_x86 *c = &cpu_data(cpu); 573 o = &cpu_data(i); 596 cpu_data(i).smt_active = threads > 1; 603 o = &cpu_data(i); 624 cpu_data(i).booted_cores++; 626 c->booted_cores = cpu_data(i).booted_cores; 652 bogosum += cpu_data(cp [all...] |
/linux-master/arch/loongarch/kernel/ |
H A D | traps.c | 1129 if (!cpu_data[cpu].asid_cache) 1130 cpu_data[cpu].asid_cache = asid_first_version(cpu);
|
H A D | setup.c | 62 struct cpuinfo_loongarch cpu_data[NR_CPUS] __read_mostly; 64 EXPORT_SYMBOL(cpu_data); variable
|
/linux-master/tools/tracing/rtla/src/ |
H A D | timerlat_top.c | 74 struct timerlat_top_cpu *cpu_data; member in struct:timerlat_top_data 84 free(data->cpu_data); 103 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); 104 if (!data->cpu_data) 109 data->cpu_data[cpu].min_irq = ~0; 110 data->cpu_data[cpu].min_thread = ~0; 111 data->cpu_data[cpu].min_user = ~0; 130 struct timerlat_top_cpu *cpu_data = &data->cpu_data[cp local 223 struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu]; local [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | setup_32.c | 406 cpu_data(0).udelay_val = loops_per_jiffy;
|
H A D | nmi.c | 132 return cpu_data(cpu).__nmi_count;
|
H A D | leon_smp.c | 240 prev = &cpu_data(i).next;
|
H A D | irq_64.c | 309 seq_printf(p, "%10u ", cpu_data(j).__nmi_count);
|
H A D | irq_32.c | 204 seq_printf(p, "%10u ", cpu_data(j).irq_resched_count); 208 seq_printf(p, "%10u ", cpu_data(j).irq_call_count); 213 seq_printf(p, "%10u ", cpu_data(j).counter);
|
/linux-master/arch/parisc/kernel/ |
H A D | irq.c | 318 return per_cpu(cpu_data, cpu).txn_addr; 330 (!per_cpu(cpu_data, next_cpu).txn_addr || 523 per_cpu(cpu_data, cpu).hpa);
|
/linux-master/arch/mips/mm/ |
H A D | tlbex.c | 314 unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask;
|
/linux-master/arch/mips/kvm/ |
H A D | entry.c | 71 unsigned int kscratch_mask = cpu_data[0].kscratch_mask; 326 UASM_i_LA_mostly(&p, GPR_AT, (long)&cpu_data[0].asid_mask); 328 UASM_i_LW(&p, GPR_T2, uasm_rel_lo((long)&cpu_data[0].asid_mask), GPR_AT);
|
/linux-master/arch/mips/kernel/ |
H A D | traps.c | 2264 cpu_data[cpu].asid_cache = 0; 2265 else if (!cpu_data[cpu].asid_cache) 2266 cpu_data[cpu].asid_cache = asid_first_version(cpu);
|
H A D | smp-cps.c | 168 cpu_set_cluster(&cpu_data[nvpes + v], cl); 169 cpu_set_core(&cpu_data[nvpes + v], c); 170 cpu_set_vpe_id(&cpu_data[nvpes + v], v); 182 set_cpu_possible(v, cpu_cluster(&cpu_data[v]) == 0); 183 set_cpu_present(v, cpu_cluster(&cpu_data[v]) == 0); 384 unsigned core = cpu_core(&cpu_data[cpu]); 385 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); 392 if (cpu_cluster(&cpu_data[cpu]) != cpu_cluster(&raw_current_cpu_data)) 399 atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask); 500 core = cpu_core(&cpu_data[cp [all...] |
H A D | pm-cps.c | 441 vpe_id = cpu_vpe_id(&cpu_data[cpu]); 462 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache, 466 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache, 479 uasm_i_addiu(&p, GPR_T0, GPR_ZERO, 1 << cpu_core(&cpu_data[cpu])); 493 err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu], 633 unsigned core = cpu_core(&cpu_data[cpu]);
|
/linux-master/arch/mips/include/asm/ |
H A D | mips-cm.h | 411 unsigned int core = cpu_core(&cpu_data[cpu]); 412 unsigned int vp = cpu_vpe_id(&cpu_data[cpu]); 467 struct cpuinfo_mips *d = &cpu_data[cpu];
|
/linux-master/arch/arm64/kernel/ |
H A D | cpufeature.c | 3571 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
|
H A D | cpuinfo.c | 33 DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data); 199 struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, i); 341 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); 363 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); 381 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); 485 struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data); 492 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0);
|