/linux-master/arch/x86/xen/ |
H A D | smp_hvm.c | 39 int cpu; local 48 for_each_possible_cpu(cpu) { 49 if (cpu == 0) 52 /* Set default vcpu_id to make sure that we don't use cpu-0's */ 53 per_cpu(xen_vcpu_id, cpu) = XEN_VCPU_ID_INVALID; 58 static void xen_hvm_cleanup_dead_cpu(unsigned int cpu) argument 61 xen_smp_intr_free(cpu); 62 xen_uninit_lock_cpu(cpu); 63 xen_teardown_timer(cpu); 67 static void xen_hvm_cleanup_dead_cpu(unsigned int cpu) argument [all...] |
H A D | smp.c | 3 #include <linux/cpu.h> 33 void xen_smp_intr_free(unsigned int cpu) argument 35 kfree(per_cpu(xen_resched_irq, cpu).name); 36 per_cpu(xen_resched_irq, cpu).name = NULL; 37 if (per_cpu(xen_resched_irq, cpu).irq >= 0) { 38 unbind_from_irqhandler(per_cpu(xen_resched_irq, cpu).irq, NULL); 39 per_cpu(xen_resched_irq, cpu).irq = -1; 41 kfree(per_cpu(xen_callfunc_irq, cpu).name); 42 per_cpu(xen_callfunc_irq, cpu).name = NULL; 43 if (per_cpu(xen_callfunc_irq, cpu) 62 xen_smp_intr_init(unsigned int cpu) argument 140 xen_smp_send_reschedule(int cpu) argument 148 unsigned cpu; local 156 int cpu; local 169 xen_smp_send_call_function_single_ipi(int cpu) argument 235 unsigned cpu; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | topology.h | 41 /* Mappings between logical cpu number and node number */ 48 extern int __cpu_to_node(int cpu); 51 extern int early_cpu_to_node(int cpu); 56 static inline int early_cpu_to_node(int cpu) argument 58 return early_per_cpu(x86_cpu_to_node_map, cpu); 94 static inline int early_cpu_to_node(int cpu) argument 134 extern const struct cpumask *cpu_coregroup_mask(int cpu); 135 extern const struct cpumask *cpu_clustergroup_mask(int cpu); 137 #define topology_logical_package_id(cpu) (cpu_data(cpu) 218 topology_is_primary_thread(unsigned int cpu) argument 226 topology_is_primary_thread(unsigned int cpu) argument 279 arch_scale_freq_capacity(int cpu) argument [all...] |
/linux-master/net/netfilter/ |
H A D | nf_flow_table_procfs.c | 9 int cpu; local 14 for (cpu = *pos - 1; cpu < nr_cpu_ids; ++cpu) { 15 if (!cpu_possible(cpu)) 17 *pos = cpu + 1; 18 return per_cpu_ptr(net->ft.stat, cpu); 27 int cpu; local 29 for (cpu = *pos; cpu < nr_cpu_id [all...] |
/linux-master/tools/perf/tests/ |
H A D | topology.c | 71 struct perf_cpu cpu; local 80 * cpu is a NULL pointer. 98 if (!session->header.env.cpu 107 * test if physical_package_id returns -1 for cpu from perf_cpu_map. 114 TEST_ASSERT_VAL("Session header CPU map not set", session->header.env.cpu); 117 cpu.cpu = i; 118 if (!perf_cpu_map__has(map, cpu)) 121 session->header.env.cpu[i].core_id, 122 session->header.env.cpu[ [all...] |
/linux-master/include/trace/events/ |
H A D | cpuhp.h | 12 TP_PROTO(unsigned int cpu, 17 TP_ARGS(cpu, target, idx, fun), 20 __field( unsigned int, cpu ) 27 __entry->cpu = cpu; 33 TP_printk("cpu: %04u target: %3d step: %3d (%ps)", 34 __entry->cpu, __entry->target, __entry->idx, __entry->fun) 39 TP_PROTO(unsigned int cpu, 45 TP_ARGS(cpu, target, idx, fun, node), 48 __field( unsigned int, cpu ) [all...] |
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | smp.c | 24 * ps3_ipi_virqs - a per cpu array of virqs for ipi use 30 static void ps3_smp_message_pass(int cpu, int msg) argument 40 virq = per_cpu(ps3_ipi_virqs, cpu)[msg]; 45 " (%d)\n", __func__, __LINE__, cpu, msg, result); 50 int cpu; local 52 for (cpu = 0; cpu < 2; cpu++) { 54 unsigned int *virqs = per_cpu(ps3_ipi_virqs, cpu); 57 DBG(" -> %s:%d: (%d)\n", __func__, __LINE__, cpu); 93 ps3_smp_cleanup_cpu(int cpu) argument [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | paravirt.h | 29 u64 pseries_paravirt_steal_clock(int cpu); 31 static inline u64 paravirt_steal_clock(int cpu) argument 33 return pseries_paravirt_steal_clock(cpu); 37 /* If bit 0 is set, the cpu has been ceded, conferred, or preempted */ 38 static inline u32 yield_count_of(int cpu) argument 40 __be32 yield_count = READ_ONCE(lppaca_of(cpu).yield_count); 60 static inline void yield_to_preempted(int cpu, u32 yield_count) argument 62 plpar_hcall_norets_notrace(H_CONFER, get_hard_smp_processor_id(cpu), yield_count); 65 static inline void prod_cpu(int cpu) argument 67 plpar_hcall_norets_notrace(H_PROD, get_hard_smp_processor_id(cpu)); 96 yield_count_of(int cpu) argument 102 yield_to_preempted(int cpu, u32 yield_count) argument 114 prod_cpu(int cpu) argument 130 vcpu_is_preempted(int cpu) argument [all...] |
/linux-master/drivers/clk/imx/ |
H A D | clk-cpu.c | 28 struct clk_cpu *cpu = to_clk_cpu(hw); local 30 return clk_get_rate(cpu->div); 36 struct clk_cpu *cpu = to_clk_cpu(hw); local 38 return clk_round_rate(cpu->pll, rate); 44 struct clk_cpu *cpu = to_clk_cpu(hw); local 48 ret = clk_set_parent(cpu->mux, cpu->step); 53 ret = clk_set_rate(cpu->pll, rate); 55 clk_set_parent(cpu->mux, cpu 77 struct clk_cpu *cpu; local [all...] |
/linux-master/include/asm-generic/ |
H A D | topology.h | 35 #define cpu_to_node(cpu) ((void)(cpu),0) 41 #define set_cpu_numa_node(cpu, node) 44 #define cpu_to_mem(cpu) ((void)(cpu),0) 72 #define set_cpu_numa_mem(cpu, node)
|
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | hyp-smp.c | 19 u64 cpu_logical_map(unsigned int cpu) argument 21 BUG_ON(cpu >= ARRAY_SIZE(hyp_cpu_logical_map)); 23 return hyp_cpu_logical_map[cpu]; 28 unsigned long __hyp_per_cpu_offset(unsigned int cpu) argument 34 BUG_ON(cpu >= ARRAY_SIZE(kvm_arm_hyp_percpu_base)); 37 this_cpu_base = kern_hyp_va(cpu_base_array[cpu]);
|
/linux-master/drivers/cpufreq/ |
H A D | intel_pstate.c | 20 #include <linux/cpu.h> 31 #include <asm/cpu.h> 98 * to account for cpu idle period 134 * Stores the per cpu model P state limits and current P state. 190 * @cpu: CPU number for this instance data 226 int cpu; member in struct:cpudata 271 * @get_cpu_scaling: Get frequency scaling factor for a given cpu 280 int (*get_max)(int cpu); 281 int (*get_max_physical)(int cpu); 282 int (*get_min)(int cpu); 351 intel_pstate_set_itmt_prio(int cpu) argument 395 intel_pstate_get_cppc_guaranteed(int cpu) argument 410 intel_pstate_cppc_get_scaling(int cpu) argument 430 intel_pstate_set_itmt_prio(int cpu) argument 437 struct cpudata *cpu; local 493 struct cpudata *cpu; local 517 intel_pstate_get_cppc_guaranteed(int cpu) argument 522 intel_pstate_cppc_get_scaling(int cpu) argument 528 intel_pstate_freq_to_hwp_rel(struct cpudata *cpu, int freq, unsigned int relation) argument 547 intel_pstate_freq_to_hwp(struct cpudata *cpu, int freq) argument 563 intel_pstate_hybrid_hwp_adjust(struct cpudata *cpu) argument 606 struct cpudata *cpu = all_cpu_data[0]; local 652 intel_pstate_set_epb(int cpu, s16 pref) argument 744 intel_pstate_set_epp(struct cpudata *cpu, u32 epp) argument 825 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 910 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 937 __intel_pstate_get_hwp_cap(struct cpudata *cpu) argument 947 intel_pstate_get_hwp_cap(struct cpudata *cpu) argument 965 intel_pstate_hwp_set(unsigned int cpu) argument 1030 intel_pstate_hwp_offline(struct cpudata *cpu) argument 1099 intel_pstate_hwp_reenable(struct cpudata *cpu) argument 1107 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 1121 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 1147 int cpu; local 1165 intel_pstate_update_limits(unsigned int cpu) argument 1179 int cpu; local 1224 struct cpudata *cpu; local 1250 struct cpudata *cpu; local 1324 struct cpudata *cpu = all_cpu_data[0]; local 1350 struct cpudata *cpu = all_cpu_data[i]; local 1833 core_get_min_pstate(int cpu) argument 1841 core_get_max_pstate_physical(int cpu) argument 1849 core_get_tdp_ratio(int cpu, u64 plat_info) argument 1882 core_get_max_pstate(int cpu) argument 1917 core_get_turbo_pstate(int cpu) argument 1946 knl_get_turbo_pstate(int cpu) argument 1966 hwp_get_cpu_scaling(int cpu) argument 1990 intel_pstate_set_pstate(struct cpudata *cpu, int pstate) argument 2003 intel_pstate_set_min_pstate(struct cpudata *cpu) argument 2008 intel_pstate_get_cpu_pstates(struct cpudata *cpu) argument 2056 intel_pstate_hwp_boost_up(struct cpudata *cpu) argument 2103 intel_pstate_hwp_boost_down(struct cpudata *cpu) argument 2119 intel_pstate_update_util_hwp_local(struct cpudata *cpu, u64 time) argument 2151 struct cpudata *cpu = container_of(data, struct cpudata, update_util); local 2159 intel_pstate_calc_avg_perf(struct cpudata *cpu) argument 2166 intel_pstate_sample(struct cpudata *cpu, u64 time) argument 2208 get_avg_frequency(struct cpudata *cpu) argument 2213 get_avg_pstate(struct cpudata *cpu) argument 2219 get_target_pstate(struct cpudata *cpu) argument 2254 intel_pstate_prepare_request(struct cpudata *cpu, int pstate) argument 2262 intel_pstate_update_pstate(struct cpudata *cpu, int pstate) argument 2271 intel_pstate_adjust_pstate(struct cpudata *cpu) argument 2297 struct cpudata *cpu = container_of(data, struct cpudata, update_util); local 2424 struct cpudata *cpu; local 2466 struct cpudata *cpu = all_cpu_data[cpu_num]; local 2483 intel_pstate_clear_update_util_hook(unsigned int cpu) argument 2495 intel_pstate_get_max_freq(struct cpudata *cpu) argument 2501 intel_pstate_update_perf_limits(struct cpudata *cpu, unsigned int policy_min, unsigned int policy_max) argument 2566 struct cpudata *cpu; local 2615 intel_pstate_adjust_policy_max(struct cpudata *cpu, struct cpufreq_policy_data *policy) argument 2627 intel_pstate_verify_cpu_policy(struct cpudata *cpu, struct cpufreq_policy_data *policy) argument 2653 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2678 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2714 struct cpudata *cpu; local 2755 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2779 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2803 intel_cpufreq_trace(struct cpudata *cpu, unsigned int trace_type, int old_pstate) argument 2825 intel_cpufreq_hwp_update(struct cpudata *cpu, u32 min, u32 max, u32 desired, bool fast_switch) argument 2849 intel_cpufreq_perf_ctl_update(struct cpudata *cpu, u32 target_pstate, bool fast_switch) argument 2863 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2889 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2911 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 2926 struct cpudata *cpu = all_cpu_data[cpunum]; local 2969 struct cpudata *cpu; local 3058 struct cpudata *cpu = all_cpu_data[policy->cpu]; local 3093 unsigned int cpu; local [all...] |
/linux-master/arch/mips/mm/ |
H A D | context.c | 16 static bool asid_versions_eq(int cpu, u64 a, u64 b) argument 18 return ((a ^ b) & asid_version_mask(cpu)) == 0; 23 unsigned int cpu; local 33 cpu = smp_processor_id(); 34 asid = asid_cache(cpu); 36 if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) { 42 set_cpu_context(cpu, mm, asid); 43 asid_cache(cpu) = asid; 49 unsigned int cpu = smp_processor_id(); local 59 if (!asid_versions_eq(cpu, cpu_contex 67 int cpu; local 102 int cpu; local 181 unsigned int cpu = smp_processor_id(); local [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | smp.c | 29 #include <linux/cpu.h> 143 * Returns 1 if the specified cpu should be brought up during boot. 277 void smp_muxed_ipi_set_message(int cpu, int msg) argument 279 struct cpu_messages *info = &per_cpu(ipi_message, cpu); 289 void smp_muxed_ipi_message_pass(int cpu, int msg) argument 291 smp_muxed_ipi_set_message(cpu, msg); 297 smp_ops->cause_ipi(cpu); 351 static inline void do_message_pass(int cpu, int msg) argument 354 smp_ops->message_pass(cpu, msg); 357 smp_muxed_ipi_message_pass(cpu, ms 361 arch_smp_send_reschedule(int cpu) argument 368 arch_send_call_function_single_ipi(int cpu) argument 375 unsigned int cpu; local 469 do_smp_send_nmi_ipi(int cpu, bool safe) argument 493 __smp_send_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us, bool safe) argument 558 smp_send_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us) argument 563 smp_send_safe_nmi_ipi(int cpu, void (*fn)(struct pt_regs *), u64 delay_us) argument 572 unsigned int cpu; local 594 int cpu; local 866 get_cpu_thread_group_start(int cpu, struct thread_groups *tg) argument 885 get_thread_groups(int cpu, int group_property, int *err) argument 920 update_mask_from_threadgroup(cpumask_var_t *mask, struct thread_groups *tg, int cpu, int cpu_group_start) argument 943 init_thread_group_cache_map(int cpu, int cache_property) argument 1030 shared_cache_mask(int cpu) argument 1036 smallcore_smt_mask(int cpu) argument 1042 cpu_coregroup_mask(int cpu) argument 1056 cpu_mc_mask(int cpu) argument 1063 int cpu; local 1094 unsigned int cpu, num_threads; local 1183 unsigned int cpu = smp_processor_id(); local 1210 generic_cpu_die(unsigned int cpu) argument 1223 generic_set_cpu_dead(unsigned int cpu) argument 1233 generic_set_cpu_up(unsigned int cpu) argument 1238 generic_check_cpu_restart(unsigned int cpu) argument 1243 is_cpu_dead(unsigned int cpu) argument 1259 cpu_idle_thread_init(unsigned int cpu, struct task_struct *idle) argument 1270 __cpu_up(unsigned int cpu, struct task_struct *tidle) argument 1359 cpu_to_core_id(int cpu) argument 1376 cpu_core_index_of_thread(int cpu) argument 1391 cpu_to_l2cache(int cpu) argument 1410 update_mask_by_l2(int cpu, cpumask_var_t *mask) argument 1481 remove_cpu_from_masks(int cpu) argument 1508 add_cpu_to_smallcore_masks(int cpu) argument 1523 update_coregroup_mask(int cpu, cpumask_var_t *mask) argument 1559 add_cpu_to_masks(int cpu) argument 1621 unsigned int cpu = raw_smp_processor_id(); local 1749 arch_asym_cpu_priority(int cpu) argument 1760 int cpu = smp_processor_id(); local 1778 __cpu_die(unsigned int cpu) argument [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | smp_plat.h | 12 #include <asm/cpu.h> 31 * smp_cpuid_part() - return part id for a given cpu 32 * @cpu: logical cpu id. 34 * Return: part id of logical cpu passed as argument. 36 static inline unsigned int smp_cpuid_part(int cpu) argument 38 struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpu); 73 #define cpu_logical_map(cpu) __cpu_logical_map[cpu] 75 * Retrieve logical cpu inde 82 int cpu; local 114 platform_can_hotplug_cpu(unsigned int cpu) argument [all...] |
/linux-master/arch/arm/mach-shmobile/ |
H A D | smp-r8a7779.c | 28 static int r8a7779_boot_secondary(unsigned int cpu, struct task_struct *idle) argument 32 cpu = cpu_logical_map(cpu); 33 if (cpu) 34 ret = rcar_sysc_power_up_cpu(cpu); 60 static int r8a7779_platform_cpu_kill(unsigned int cpu) argument 64 cpu = cpu_logical_map(cpu); 65 if (cpu) 66 ret = rcar_sysc_power_down_cpu(cpu); 71 r8a7779_cpu_kill(unsigned int cpu) argument [all...] |
/linux-master/arch/sh/kernel/cpu/sh4a/ |
H A D | smp-shx3.c | 16 #include <linux/cpu.h> 30 unsigned int cpu = hard_smp_processor_id(); local 31 unsigned int offs = 4 * cpu; 45 unsigned int cpu = 0; local 48 init_cpu_possible(cpumask_of(cpu)); 51 __raw_writel(__raw_readl(STBCR_REG(cpu)) | STBCR_LTSLP, STBCR_REG(cpu)); 84 static void shx3_start_cpu(unsigned int cpu, unsigned long entry_point) argument 87 __raw_writel(entry_point, RESET_REG(cpu)); 89 __raw_writel(virt_to_phys(entry_point), RESET_REG(cpu)); 106 shx3_send_ipi(unsigned int cpu, unsigned int message) argument 115 shx3_update_boot_vector(unsigned int cpu) argument 123 shx3_cpu_prepare(unsigned int cpu) argument [all...] |
/linux-master/drivers/base/ |
H A D | arch_topology.c | 3 * Arch specific cpu topology information 11 #include <linux/cpu.h> 65 int cpu; local 76 for_each_cpu(cpu, cpus) { 77 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); 81 rcu_assign_pointer(per_cpu(sft_data, cpu), data); 82 cpumask_set_cpu(cpu, &scale_freq_counters_mask); 96 int cpu; local 100 for_each_cpu(cpu, cpus) { 101 sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu)); 158 topology_set_cpu_scale(unsigned int cpu, unsigned long capacity) argument 184 int cpu; local 212 struct cpu *cpu = container_of(dev, struct cpu, dev); local 222 cpu_capacity_sysctl_add(unsigned int cpu) argument 234 cpu_capacity_sysctl_remove(unsigned int cpu) argument 288 int cpu; local 310 topology_parse_cpu_capacity(struct device_node *cpu_node, int cpu) argument 361 freq_inv_set_max_ratio(int cpu, u64 max_rate) argument 372 int cpu; local 432 int cpu; local 517 int cpu; local 540 int cpu; local 681 int cpu; local 727 cpu_coregroup_mask(int cpu) argument 754 cpu_clustergroup_mask(int cpu) argument 770 int cpu, ret; local 807 clear_cpu_topology(int cpu) argument 825 unsigned int cpu; local 839 remove_cpu_topology(unsigned int cpu) argument 863 int cpu, ret; local [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | cmp.h | 12 extern void cmp_boot_secondary(int cpu, struct task_struct *t); 17 extern void cmp_send_ipi(int cpu, unsigned int action);
|
/linux-master/arch/mips/power/ |
H A D | Makefile | 2 obj-$(CONFIG_HIBERNATION) += cpu.o hibernate.o hibernate_asm.o
|
/linux-master/arch/loongarch/include/asm/ |
H A D | topology.h | 32 #define topology_physical_package_id(cpu) (cpu_data[cpu].package) 33 #define topology_core_id(cpu) (cpu_data[cpu].core) 34 #define topology_core_cpumask(cpu) (&cpu_core_map[cpu]) 35 #define topology_sibling_cpumask(cpu) (&cpu_sibling_map[cpu])
|
H A D | timex.h | 12 #include <asm/cpu.h> 13 #include <asm/cpu-features.h>
|
/linux-master/tools/testing/selftests/cpu-hotplug/ |
H A D | Makefile | 4 TEST_PROGS := cpu-on-off-test.sh 9 @/bin/bash ./cpu-on-off-test.sh -a && echo "cpu-hotplug selftests: [PASS]" || echo "cpu-hotplug selftests: [FAIL]"
|
/linux-master/arch/loongarch/kernel/ |
H A D | idle.c | 7 #include <linux/cpu.h> 9 #include <asm/cpu.h>
|
/linux-master/tools/power/cpupower/lib/ |
H A D | cpupower.c | 74 int cpupower_is_cpu_online(unsigned int cpu) argument 84 snprintf(path, sizeof(path), PATH_TO_CPU "cpu%u", cpu); 93 snprintf(path, sizeof(path), PATH_TO_CPU "cpu%u/online", cpu); 117 static int sysfs_topology_read_file(unsigned int cpu, const char *fname, int *result) argument 123 snprintf(path, sizeof(path), PATH_TO_CPU "cpu%u/topology/%s", 124 cpu, fname); 145 else if (top1->cpu < top2->cpu) 161 int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF); local [all...] |