Searched refs:cpu (Results 826 - 850 of 2844) sorted by relevance

<<31323334353637383940>>

/linux-master/drivers/hwtracing/coresight/
H A Dcoresight-cpu-debug.c10 #include <linux/cpu.h>
92 int cpu; member in struct:debug_drvdata
175 __func__, drvdata->cpu);
383 int cpu; local
395 for_each_possible_cpu(cpu) {
396 drvdata = per_cpu(debug_drvdata, cpu);
400 dev_emerg(drvdata->dev, "CPU[%d]:\n", drvdata->cpu);
418 int cpu, ret = 0; local
427 for_each_possible_cpu(cpu) {
428 drvdata = per_cpu(debug_drvdata, cpu);
457 int cpu, ret, err = 0; local
[all...]
/linux-master/drivers/irqchip/
H A Dirq-sifive-plic.c6 #include <linux/cpu.h>
116 int cpu; local
118 for_each_cpu(cpu, mask) {
119 struct plic_handler *handler = per_cpu_ptr(&plic_handlers, cpu);
166 unsigned int cpu; local
170 cpu = cpumask_first_and(&priv->lmask, mask_val);
172 cpu = cpumask_first_and_and(&priv->lmask, mask_val, cpu_online_mask);
174 if (cpu >= nr_cpu_ids)
179 irq_data_update_effective_affinity(d, cpumask_of(cpu));
243 unsigned int i, cpu; local
275 unsigned int i, index, cpu; local
395 plic_dying_cpu(unsigned int cpu) argument
403 plic_starting_cpu(unsigned int cpu) argument
485 int error = 0, nr_contexts, nr_handlers = 0, cpu, i; local
[all...]
H A Dirq-bcm7038-l1.c122 struct bcm7038_l1_cpu *cpu; local
127 cpu = intc->cpus[cpu_logical_map(smp_processor_id())];
129 cpu = intc->cpus[0];
140 pending = l1_readl(cpu->map_base + reg_status(intc, idx)) &
141 ~cpu->mask_cache[idx];
228 struct bcm7038_l1_cpu *cpu; local
252 cpu = intc->cpus[idx] = kzalloc(struct_size(cpu, mask_cache, n_words),
254 if (!cpu)
257 cpu
444 struct bcm7038_l1_cpu *cpu = intc->cpus[idx]; local
[all...]
H A Dirq-mips-gic.c111 static void gic_send_ipi(struct irq_data *d, unsigned int cpu) argument
157 /* Get per-cpu bitmaps */
190 unsigned int cpu; local
195 cpu = cpumask_first(irq_data_get_effective_affinity_mask(d));
196 set_bit(intr, per_cpu_ptr(pcpu_masks, cpu));
264 unsigned int cpu; local
266 cpu = cpumask_first_and(cpumask, cpu_online_mask);
267 if (cpu >= NR_CPUS)
274 write_gic_map_vp(irq, BIT(mips_cm_vp_id(cpu)));
279 set_bit(irq, per_cpu_ptr(pcpu_masks, cpu));
354 int intr, cpu; local
372 int intr, cpu; local
431 gic_shared_irq_domain_map(struct irq_domain *d, unsigned int virq, irq_hw_number_t hw, unsigned int cpu) argument
474 int err, cpu; local
594 int cpu, ret, i; local
719 gic_cpu_startup(unsigned int cpu) argument
[all...]
/linux-master/arch/x86/kernel/cpu/mce/
H A Damd.c19 #include <linux/cpu.h>
134 enum smca_bank_types smca_get_bank_type(unsigned int cpu, unsigned int bank) argument
141 b = &per_cpu(smca_banks, cpu)[bank];
244 static void smca_set_misc_banks_map(unsigned int bank, unsigned int cpu) argument
262 per_cpu(smca_misc_banks_map, cpu) |= BIT_ULL(bank);
266 static void smca_configure(unsigned int bank, unsigned int cpu) argument
306 smca_set_misc_banks_map(bank, cpu);
389 pr_err(FW_BUG "cpu %d, failed to setup threshold interrupt "
390 "for bank %d, block %d (MSR%08X=0x%x%08x)\n", b->cpu,
404 pr_err(FW_BUG "cpu
519 smca_get_block_address(unsigned int bank, unsigned int block, unsigned int cpu) argument
531 get_block_address(u32 current_addr, u32 low, u32 high, unsigned int bank, unsigned int block, unsigned int cpu) argument
563 unsigned int cpu = smp_processor_id(); local
674 unsigned int bank, block, cpu = smp_processor_id(); local
922 unsigned int bank, cpu = smp_processor_id(); local
1090 get_name(unsigned int cpu, unsigned int bank, struct threshold_block *b) argument
1120 allocate_threshold_blocks(unsigned int cpu, struct threshold_bank *tb, unsigned int bank, unsigned int block, u32 address) argument
1221 threshold_create_bank(struct threshold_bank **bp, unsigned int cpu, unsigned int bank) argument
1364 mce_threshold_remove_device(unsigned int cpu) argument
1392 mce_threshold_create_device(unsigned int cpu) argument
[all...]
/linux-master/tools/testing/selftests/rseq/
H A Drseq-riscv-bits.h9 int RSEQ_TEMPLATE_IDENTIFIER(rseq_cmpeqv_storev)(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument
32 : [cpu_id] "r" (cpu),
63 off_t voffp, intptr_t *load, int cpu)
89 : [cpu_id] "r" (cpu),
119 int RSEQ_TEMPLATE_IDENTIFIER(rseq_addv)(intptr_t *v, intptr_t count, int cpu) argument
139 : [cpu_id] "r" (cpu),
165 intptr_t newv, int cpu)
192 : [cpu_id] "r" (cpu),
232 int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, off_t off, intptr_t inc, int cpu) argument
250 : [cpu_id] "r" (cpu),
62 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, off_t voffp, intptr_t *load, int cpu) argument
163 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument
281 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument
344 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument
[all...]
H A Drseq-x86-bits.h16 int RSEQ_TEMPLATE_IDENTIFIER(rseq_cmpeqv_storev)(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument
45 : [cpu_id] "r" (cpu),
82 long voffp, intptr_t *load, int cpu)
116 : [cpu_id] "r" (cpu),
150 int RSEQ_TEMPLATE_IDENTIFIER(rseq_addv)(intptr_t *v, intptr_t count, int cpu) argument
172 : [cpu_id] "r" (cpu),
204 int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, long off, intptr_t inc, int cpu) argument
231 : [cpu_id] "r" (cpu),
257 intptr_t newv, int cpu)
292 : [cpu_id] "r" (cpu),
81 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, long voffp, intptr_t *load, int cpu) argument
255 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument
337 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument
407 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument
528 rseq_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument
593 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, long voffp, intptr_t *load, int cpu) argument
662 rseq_addv(intptr_t *v, intptr_t count, int cpu) argument
710 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument
793 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument
870 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument
[all...]
/linux-master/lib/
H A Dpercpu_counter.c9 #include <linux/cpu.h>
62 int cpu; local
66 for_each_possible_cpu(cpu) {
67 s32 *pcount = per_cpu_ptr(fbc->counters, cpu);
126 * Add up all the per-cpu counts, return the result. This is a more accurate
129 * We use the cpu mask of (cpu_online_mask | cpu_dying_mask) to capture sums
140 int cpu; local
145 for_each_cpu_or(cpu, cpu_online_mask, cpu_dying_mask) {
146 s32 *pcount = per_cpu_ptr(fbc->counters, cpu);
224 static int compute_batch_value(unsigned int cpu) argument
232 percpu_counter_cpu_dead(unsigned int cpu) argument
335 int cpu; local
[all...]
/linux-master/arch/loongarch/kvm/
H A Dmain.c195 static void kvm_update_vpid(struct kvm_vcpu *vcpu, int cpu) argument
200 context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
219 int cpu; local
224 cpu = smp_processor_id();
229 context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
230 migrated = (vcpu->cpu != cpu);
242 kvm_update_vpid(vcpu, cpu);
244 vcpu->cpu = cpu;
315 int cpu, order; local
[all...]
/linux-master/tools/testing/selftests/kvm/lib/aarch64/
H A Dgic_v3.c53 static inline volatile void *gicr_base_cpu(uint32_t cpu) argument
56 return GICR_BASE_GVA + cpu * SZ_64K * 2;
59 static void gicv3_gicr_wait_for_rwp(uint32_t cpu) argument
63 while (readl(gicr_base_cpu(cpu) + GICR_CTLR) & GICR_CTLR_RWP) {
172 uint32_t cpu = guest_get_vcpuid(); local
193 cpu_or_dist = (intid_range == SPI_RANGE) ? DIST_BIT : cpu;
236 uint32_t cpu = guest_get_vcpuid(); local
239 gicv3_wait_for_rwp(is_spi ? DIST_BIT : cpu);
245 uint32_t cpu = guest_get_vcpuid(); local
248 gicv3_wait_for_rwp(is_spi ? DIST_BIT : cpu);
296 gicv3_cpu_init(unsigned int cpu) argument
[all...]
/linux-master/drivers/misc/sgi-gru/
H A Dgrufile.c311 static unsigned long gru_chiplet_cpu_to_mmr(int chiplet, int cpu, int *corep) argument
322 core = uv_cpu_core_number(cpu) + UV_MAX_INT_CORES * uv_cpu_socket_number(cpu);
323 if (core >= GRU_NUM_TFM || uv_cpu_ht_number(cpu))
341 irq_handler_t irq_handler, int cpu, int blade)
347 mmr = gru_chiplet_cpu_to_mmr(chiplet, cpu, &core);
351 irq = uv_setup_irq(irq_name, cpu, blade, mmr, UV_AFFINITY_CPU);
369 static void gru_chiplet_teardown_tlb_irq(int chiplet, int cpu, int blade) argument
374 mmr = gru_chiplet_cpu_to_mmr(chiplet, cpu, &core);
387 int cpu; local
340 gru_chiplet_setup_tlb_irq(int chiplet, char *irq_name, irq_handler_t irq_handler, int cpu, int blade) argument
405 int cpu; local
[all...]
/linux-master/kernel/printk/
H A Dnbcon.c23 * - The 'cpu' field denotes on which CPU the console is locked. It is used
229 unsigned int cpu = smp_processor_id(); local
253 new.cpu = cpu;
309 unsigned int cpu = smp_processor_id(); local
338 new.cpu = cpu;
392 unsigned int cpu = smp_processor_id(); local
406 if (cur->cpu == cpu)
498 unsigned int cpu = smp_processor_id(); local
544 unsigned int cpu = smp_processor_id(); local
606 unsigned int cpu = smp_processor_id(); local
659 unsigned int cpu = smp_processor_id(); local
[all...]
/linux-master/arch/parisc/kernel/
H A Dirq.c72 int cpu = smp_processor_id(); local
75 per_cpu(local_ack_eiem, cpu) &= ~mask;
78 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu));
87 int cpu = smp_processor_id(); local
90 per_cpu(local_ack_eiem, cpu) |= mask;
93 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu));
311 unsigned long txn_affinity_addr(unsigned int irq, int cpu) argument
315 irq_data_update_affinity(d, cpumask_of(cpu));
318 return per_cpu(cpu_data, cpu).txn_addr;
385 int cpu local
493 int irq, cpu = smp_processor_id(); local
518 int cpu = cpumask_first(&dest); local
[all...]
/linux-master/kernel/rcu/
H A Dtasks.h36 * @cpu: CPU number corresponding to this entry.
51 int cpu; member in struct:rcu_tasks_percpu
245 int cpu; local
265 for_each_possible_cpu(cpu) {
266 struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
269 if (cpu)
274 rtpcp->cpu = cpu;
401 int cpu; local
415 for_each_possible_cpu(cpu) {
436 int cpu; local
514 int cpu; local
604 int cpu; local
696 int cpu; local
907 int cpu; local
947 int cpu; local
1012 int cpu; local
1612 int cpu = task_cpu(t); local
1660 int cpu; local
1742 int cpu; local
1828 int cpu; local
1859 int cpu; local
1907 int cpu; local
[all...]
/linux-master/include/linux/
H A Dworkqueue.h119 int cpu; member in struct:delayed_work
371 WQ_UNBOUND = 1 << 1, /* not bound to any cpu */
375 WQ_CPU_INTENSIVE = 1 << 5, /* cpu intensive workqueue */
379 * Per-cpu workqueues are generally preferred because they tend to
380 * show better performance thanks to cache locality. Per-cpu
387 * however, for example, a per-cpu work item scheduled from an
393 * Workqueues marked with WQ_POWER_EFFICIENT are per-cpu by default
395 * specified. Per-cpu workqueues which are identified to
469 void workqueue_softirq_dead(unsigned int cpu);
478 * For a per-cpu workqueu
661 schedule_work_on(int cpu, struct work_struct *work) argument
759 schedule_delayed_work_on(int cpu, struct delayed_work *dwork, unsigned long delay) argument
780 work_on_cpu(int cpu, long (*fn)(void *), void *arg) argument
784 work_on_cpu_safe(int cpu, long (*fn)(void *), void *arg) argument
833 wq_watchdog_touch(int cpu) argument
[all...]
/linux-master/drivers/macintosh/
H A Drack-meter.c67 struct rackmeter_cpu cpu[2]; member in struct:rackmeter
81 static inline u64 get_cpu_idle_time(unsigned int cpu) argument
83 struct kernel_cpustat *kcpustat = &kcpustat_cpu(cpu);
90 retval += kcpustat_field(kcpustat, CPUTIME_NICE, cpu);
217 unsigned int cpu = smp_processor_id(); local
226 total_idle_nsecs = get_cpu_idle_time(cpu);
236 offset = cpu << 3;
246 pause = (rm->cpu[0].zero && rm->cpu[1].zero);
249 pause = (rm->cpu[
259 unsigned int cpu; local
[all...]
/linux-master/kernel/power/
H A Denergy_model.c12 #include <linux/cpu.h>
221 int i, cpu; local
227 cpu = cpumask_first(em_span_cpus(pd));
235 max_cap = (u64) arch_scale_cpu_capacity(cpu);
405 int cpu, ret, num_cpus; local
440 for_each_cpu(cpu, cpus) {
441 cpu_dev = get_cpu_device(cpu);
462 int i, cpu; local
468 cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask);
469 if (cpu >
523 em_cpu_get(int cpu) argument
564 int cpu, ret; local
751 int cpu; local
[all...]
/linux-master/arch/powerpc/kernel/
H A Dsetup-common.c26 #include <linux/cpu.h>
114 /* This keeps a track of which one is the crashing cpu. */
267 * rate of a given cpu. The rest can use ppc_proc_freq to
324 /* If this is the last cpu, print the summary */
333 if (*pos == 0) /* just in case, cpu 0 is not the first */
415 static int assign_threads(unsigned int cpu, unsigned int nthreads, bool present, argument
418 for (int i = 0; i < nthreads && cpu < nr_cpu_ids; i++) {
423 DBG(" thread %d -> cpu %d (hard id %d)\n", i, cpu, hwid);
425 set_cpu_present(cpu, presen
455 int cpu = 0; local
889 int cpu; local
[all...]
/linux-master/drivers/virt/nitro_enclaves/
H A Dne_misc_dev.c13 #include <linux/cpu.h>
36 * NE_CPUS_SIZE - Size for max 128 CPUs, for now, in a cpu-list string, comma
101 /* https://www.kernel.org/doc/html/latest/admin-guide/kernel-parameters.html#cpu-lists */
102 MODULE_PARM_DESC(ne_cpus, "<cpu-list> - CPU pool used for Nitro Enclaves");
182 unsigned int cpu = 0; local
201 cpu = cpumask_any(cpu_pool);
202 if (cpu >= nr_cpu_ids) {
214 for_each_cpu(cpu, cpu_pool)
215 if (cpu_is_offline(cpu)) {
217 ne_misc_dev.name, cpu);
388 unsigned int cpu = 0; local
480 ne_donated_cpu(struct ne_enclave *ne_enclave, unsigned int cpu) argument
530 unsigned int cpu = 0; local
576 unsigned int cpu = 0; local
1076 unsigned int cpu = 0; local
1419 unsigned int cpu = 0; local
[all...]
/linux-master/kernel/
H A Dtorture.c32 #include <linux/cpu.h>
201 bool torture_offline(int cpu, long *n_offl_attempts, long *n_offl_successes, argument
209 if (!cpu_online(cpu) || !cpu_is_hotpluggable(cpu))
217 torture_type, cpu);
220 ret = remove_cpu(cpu);
231 torture_type, cpu, s, ret);
236 torture_type, cpu);
263 bool torture_online(int cpu, long *n_onl_attempts, long *n_onl_successes, argument
271 if (cpu_online(cpu) || !cpu_is_hotpluggabl
320 int cpu; local
342 int cpu; local
[all...]
/linux-master/arch/x86/kernel/cpu/microcode/
H A Damd.c35 #include <asm/cpu.h>
432 * patch container file in initrd, traverse equivalent cpu table, look for a
436 * load_microcode_amd() to save equivalent cpu table and microcode patches in
593 static struct ucode_patch *find_patch(unsigned int cpu) argument
595 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
605 void reload_ucode_amd(unsigned int cpu) argument
611 p = find_patch(cpu);
625 static int collect_cpu_info_amd(int cpu, struct cpu_signature *csig) argument
627 struct cpuinfo_x86 *c = &cpu_data(cpu);
628 struct ucode_cpu_info *uci = ucode_cpu_info + cpu;
645 apply_microcode_amd(int cpu) argument
695 unsigned int cpu = smp_processor_id(); local
824 unsigned int nid, cpu; local
870 request_microcode_amd(int cpu, struct device *device) argument
901 microcode_fini_cpu_amd(int cpu) argument
[all...]
/linux-master/tools/tracing/rtla/src/
H A Dtimerlat_aa.c25 * Per-cpu data statistics and data.
137 * timerlat_aa_get_data - Get the per-cpu data from the timerlat context
140 *timerlat_aa_get_data(struct timerlat_aa_context *taa_ctx, int cpu) argument
142 return &taa_ctx->taa_data[cpu];
245 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
268 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
311 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
407 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
440 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
485 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu);
560 timerlat_thread_analysis(struct timerlat_aa_data *taa_data, int cpu, int irq_thresh, int thread_thresh) argument
735 int cpu; local
[all...]
/linux-master/net/core/
H A Dgen_stats.c127 struct gnet_stats_basic_sync __percpu *cpu)
133 struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i);
150 struct gnet_stats_basic_sync __percpu *cpu,
157 WARN_ON_ONCE((cpu || running) && in_hardirq());
159 if (cpu) {
160 gnet_stats_add_basic_cpu(bstats, cpu);
175 struct gnet_stats_basic_sync __percpu *cpu,
180 if (cpu) {
185 struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i);
212 struct gnet_stats_basic_sync __percpu *cpu,
126 gnet_stats_add_basic_cpu(struct gnet_stats_basic_sync *bstats, struct gnet_stats_basic_sync __percpu *cpu) argument
149 gnet_stats_add_basic(struct gnet_stats_basic_sync *bstats, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument
174 gnet_stats_read_basic(u64 *ret_bytes, u64 *ret_packets, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument
211 ___gnet_stats_copy_basic(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, int type, bool running) argument
260 gnet_stats_copy_basic(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument
287 gnet_stats_copy_basic_hw(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument
356 gnet_stats_add_queue(struct gnet_stats_queue *qstats, const struct gnet_stats_queue __percpu *cpu, const struct gnet_stats_queue *q) argument
[all...]
/linux-master/arch/x86/kernel/
H A Dirq.c5 #include <linux/cpu.h>
200 u64 arch_irq_stat_cpu(unsigned int cpu) argument
202 u64 sum = irq_stats(cpu)->__nmi_count;
205 sum += irq_stats(cpu)->apic_timer_irqs;
206 sum += irq_stats(cpu)->irq_spurious_count;
207 sum += irq_stats(cpu)->apic_perf_irqs;
208 sum += irq_stats(cpu)->apic_irq_work_irqs;
209 sum += irq_stats(cpu)->icr_read_retry_count;
211 sum += irq_stats(cpu)->x86_platform_ipis;
214 sum += irq_stats(cpu)
[all...]
/linux-master/drivers/thermal/
H A Dcpufreq_cooling.c13 #include <linux/cpu.h>
44 * @time: previous reading of the absolute time that this cpu was idle
153 * get_load() - get load for a cpu
154 * @cpufreq_cdev: struct cpufreq_cooling_device for the cpu
155 * @cpu: cpu number
156 * @cpu_idx: index of the cpu in time_in_idle array
158 * Return: The average load of cpu @cpu in percentage since this
162 static u32 get_load(struct cpufreq_cooling_device *cpufreq_cdev, int cpu, argument
170 get_load(struct cpufreq_cooling_device *cpufreq_cdev, int cpu, int cpu_idx) argument
236 int i = 0, cpu; local
[all...]

Completed in 367 milliseconds

<<31323334353637383940>>