/linux-master/drivers/cpufreq/ |
H A D | tegra194-cpufreq.c | 79 struct tegra_cpu_data *cpu_data; member in struct:tegra194_cpufreq_data 128 *ndiv = readl(data->cpu_data[cpu].freq_core_reg) & NDIV_MASK; 139 writel(ndiv, data->cpu_data[cpu].freq_core_reg); 157 actmon_reg = CORE_ACTMON_CNTR_REG(data, data->cpu_data[c->cpu].clusterid, 158 data->cpu_data[c->cpu].cpuid); 388 u32 clusterid = data->cpu_data[cpu].clusterid; 398 ret = data->soc->ops->get_cpu_ndiv(cpu, data->cpu_data[cpu].cpuid, clusterid, &ndiv); 504 u32 clusterid = data->cpu_data[policy->cpu].clusterid; 703 data->cpu_data[cpu].cpuid = cpuid; 704 data->cpu_data[cp [all...] |
H A D | amd_freq_sensitivity.c | 37 static DEFINE_PER_CPU(struct cpu_data_t, cpu_data); 46 struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu);
|
H A D | intel_pstate.c | 614 static s16 intel_pstate_get_epb(struct cpudata *cpu_data) argument 622 ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); 629 static s16 intel_pstate_get_epp(struct cpudata *cpu_data, u64 hwp_req_data) argument 639 epp = rdmsrl_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST, 647 epp = intel_pstate_get_epb(cpu_data); 707 static int intel_pstate_get_energy_pref_index(struct cpudata *cpu_data, int *raw_epp) argument 713 epp = intel_pstate_get_epp(cpu_data, 0); 771 static int intel_pstate_set_energy_pref_index(struct cpudata *cpu_data, argument 779 epp = cpu_data->epp_default; 792 if (epp > 0 && cpu_data 894 struct cpudata *cpu_data = all_cpu_data[policy->cpu]; local 968 struct cpudata *cpu_data = all_cpu_data[cpu]; local 2517 struct cpudata *cpu_data = all_cpu_data[cpu]; local [all...] |
H A D | sc520_freq.c | 74 struct cpuinfo_x86 *c = &cpu_data(0);
|
H A D | elanfreq.c | 146 struct cpuinfo_x86 *c = &cpu_data(0);
|
/linux-master/arch/s390/pci/ |
H A D | pci_irq.c | 194 struct cpu_irq_data *cpu_data; local 211 cpu_data = &per_cpu(irq_data, cpu); 212 if (atomic_inc_return(&cpu_data->scheduled) > 1) 215 INIT_CSD(&cpu_data->csd, zpci_handle_remote_irq, &cpu_data->scheduled); 216 smp_call_function_single_async(cpu, &cpu_data->csd);
|
/linux-master/arch/x86/kernel/ |
H A D | smpboot.c | 215 * Calibrate the delay loop and update loops_per_jiffy in cpu_data. 224 cpu_data(smp_processor_id()).loops_per_jiffy = loops_per_jiffy; 322 struct cpuinfo_x86 *c = &cpu_data(id); 556 struct cpuinfo_x86 *c = &cpu_data(cpu); 573 o = &cpu_data(i); 596 cpu_data(i).smt_active = threads > 1; 603 o = &cpu_data(i); 624 cpu_data(i).booted_cores++; 626 c->booted_cores = cpu_data(i).booted_cores; 652 bogosum += cpu_data(cp [all...] |
/linux-master/arch/sh/kernel/cpu/ |
H A D | proc.c | 82 unsigned int cpu = c - cpu_data; 135 return *pos < NR_CPUS ? cpu_data + *pos : NULL;
|
/linux-master/arch/parisc/kernel/ |
H A D | time.c | 59 * CR16 to interrupt every @clocktick cycles. The it_value in cpu_data 70 struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); 162 per_cpu(cpu_data, cpu).it_value = next_tick;
|
H A D | smp.c | 123 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); 199 struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu); 335 const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid); 356 ** (offset into init_tasks[],cpu_data[])
|
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | apei.c | 106 if (cpu_data(cpu).topo.initial_apicid == lapic_id) { 108 m.socketid = cpu_data(m.extcpu).topo.pkg_id;
|
/linux-master/arch/x86/kernel/acpi/ |
H A D | cstate.c | 32 struct cpuinfo_x86 *c = &cpu_data(cpu); 172 struct cpuinfo_x86 *c = &cpu_data(cpu);
|
/linux-master/arch/sparc/kernel/ |
H A D | devices.c | 130 cpu_data(0).clock_tick = prom_getintdefault(cpu_node,
|
H A D | cpu.c | 348 , cpu_data(0).udelay_val/(500000/HZ), 349 (cpu_data(0).udelay_val/(5000/HZ)) % 100, 350 cpu_data(0).clock_tick 398 , cpu_data(0).clock_tick
|
H A D | smp_64.c | 97 i, cpu_data(i).clock_tick); 1241 if (cpu_data(i).core_id == 0) { 1247 if (cpu_data(i).core_id == 1248 cpu_data(j).core_id) 1257 if (cpu_data(i).max_cache_id == 1258 cpu_data(j).max_cache_id) 1261 if (cpu_data(i).sock_id == cpu_data(j).sock_id) 1270 if (cpu_data(i).proc_id == -1) { 1276 if (cpu_data( [all...] |
H A D | irq_32.c | 204 seq_printf(p, "%10u ", cpu_data(j).irq_resched_count); 208 seq_printf(p, "%10u ", cpu_data(j).irq_call_count); 213 seq_printf(p, "%10u ", cpu_data(j).counter);
|
/linux-master/arch/alpha/kernel/ |
H A D | irq.c | 76 seq_printf(p, "%10lu ", cpu_data[j].ipi_count);
|
/linux-master/arch/mips/include/asm/ |
H A D | cpu-type.h | 198 const int cpu_type = cpu_data[0].cputype;
|
H A D | mips-cm.h | 411 unsigned int core = cpu_core(&cpu_data[cpu]); 412 unsigned int vp = cpu_vpe_id(&cpu_data[cpu]); 467 struct cpuinfo_mips *d = &cpu_data[cpu];
|
/linux-master/arch/csky/kernel/ |
H A D | process.c | 17 struct cpuinfo_csky cpu_data[NR_CPUS]; variable in typeref:struct:cpuinfo_csky
|
/linux-master/arch/sh/mm/ |
H A D | cache-sh4.c | 74 cpu_data->icache.entry_mask); 78 for (i = 0; i < cpu_data->icache.ways; i++) { 81 icacheaddr += cpu_data->icache.way_incr;
|
/linux-master/tools/perf/util/ |
H A D | bpf_lock_contention.c | 251 struct tstamp_data ts_data, *cpu_data; local 279 cpu_data = calloc(total_cpus, sizeof(*cpu_data)); 280 if (cpu_data == NULL) 285 if (bpf_map_lookup_elem(ts_fd, &key, cpu_data) < 0) 290 &cpu_data[i]); 296 free(cpu_data);
|
/linux-master/arch/mips/kernel/ |
H A D | pm-cps.c | 441 vpe_id = cpu_vpe_id(&cpu_data[cpu]); 462 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache, 466 cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache, 479 uasm_i_addiu(&p, GPR_T0, GPR_ZERO, 1 << cpu_core(&cpu_data[cpu])); 493 err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu], 633 unsigned core = cpu_core(&cpu_data[cpu]);
|
/linux-master/arch/sh/include/asm/ |
H A D | mmu_context.h | 37 #define asid_cache(cpu) (cpu_data[cpu].asid_cache)
|
/linux-master/drivers/char/hw_random/ |
H A D | via-rng.c | 131 struct cpuinfo_x86 *c = &cpu_data(0);
|