Searched refs:cpu (Results 676 - 700 of 2844) sorted by relevance

<<21222324252627282930>>

/linux-master/arch/x86/include/asm/trace/
H A Dhyperv.h75 TP_PROTO(int cpu,
77 TP_ARGS(cpu, vector),
79 __field(int, cpu)
82 TP_fast_assign(__entry->cpu = cpu;
85 TP_printk("cpu %d vector %x",
86 __entry->cpu, __entry->vector)
/linux-master/tools/testing/selftests/rcutorture/bin/
H A Djitter.sh31 for i in /sys/devices/system/cpu/cpu[0-9]*
58 if cpus=`grep 1 /sys/devices/system/cpu/*/online 2>&1 |
H A Dkvm-assign-cpus.sh27 for j in $i/cpu*/cache/index*
37 indexlist="`ls -d $i/cpu* | grep 'cpu[0-9][0-9]*' | head -1 | sed -e 's,^.*$,ls -d &/cache/index*,' | sh | sed -e 's,^.*/,,'`"
40 for i in node*/cpu*/cache/index*/shared_cpu_list
56 f="$n/cpu*/cache/$i/shared_cpu_list"
81 cat $n/cpu*/cache/$splitindex/shared_cpu_list | sort -u -k1n |
94 print "cpu[" nodenum "][" idx "] = " j ";";
/linux-master/arch/x86/kernel/
H A Dirq_32.c17 #include <linux/cpu.h>
108 * Allocate per-cpu stacks for hardirq and softirq processing
110 int irq_init_percpu_irqstack(unsigned int cpu) argument
112 int node = cpu_to_node(cpu);
115 if (per_cpu(pcpu_hot.hardirq_stack_ptr, cpu))
127 per_cpu(pcpu_hot.hardirq_stack_ptr, cpu) = page_address(ph);
128 per_cpu(pcpu_hot.softirq_stack_ptr, cpu) = page_address(ps);
/linux-master/drivers/cpufreq/
H A Dppc_cbe_cpufreq_pmi.c34 int cbe_cpufreq_set_pmode_pmi(int cpu, unsigned int pmode) argument
42 pmi_msg.data1 = cbe_cpu_to_node(cpu);
69 int cpu, ret; local
76 cpu = cbe_node_to_cpu(node);
80 policy = cpufreq_cpu_get(cpu);
82 pr_warn("cpufreq policy not found cpu%d\n", cpu);
/linux-master/arch/openrisc/kernel/
H A Dtime.c71 unsigned int cpu = smp_processor_id(); local
73 &per_cpu(clockevent_openrisc_timer, cpu);
74 struct cpuinfo_or1k *cpuinfo = &cpuinfo_or1k[cpu];
86 evt->cpumask = cpumask_of(cpu);
116 unsigned int cpu = smp_processor_id(); local
118 &per_cpu(clockevent_openrisc_timer, cpu);
/linux-master/drivers/pinctrl/renesas/
H A DMakefile56 CFLAGS_pfc-sh7203.o += -I$(srctree)/arch/sh/include/cpu-sh2a
57 CFLAGS_pfc-sh7264.o += -I$(srctree)/arch/sh/include/cpu-sh2a
58 CFLAGS_pfc-sh7269.o += -I$(srctree)/arch/sh/include/cpu-sh2a
59 CFLAGS_pfc-sh7720.o += -I$(srctree)/arch/sh/include/cpu-sh3
60 CFLAGS_pfc-sh7722.o += -I$(srctree)/arch/sh/include/cpu-sh4
61 CFLAGS_pfc-sh7723.o += -I$(srctree)/arch/sh/include/cpu-sh4
62 CFLAGS_pfc-sh7724.o += -I$(srctree)/arch/sh/include/cpu-sh4
63 CFLAGS_pfc-sh7734.o += -I$(srctree)/arch/sh/include/cpu-sh4
64 CFLAGS_pfc-sh7757.o += -I$(srctree)/arch/sh/include/cpu-sh4
65 CFLAGS_pfc-sh7785.o += -I$(srctree)/arch/sh/include/cpu
[all...]
/linux-master/drivers/edac/
H A Docteon_edac-pc.c45 unsigned int cpu = smp_processor_id(); local
58 "CacheErr (Icache):%llx, core %d/cpu %d, cp0_errorepc == %lx\n",
59 (unsigned long long)icache_err, core, cpu,
62 edac_device_handle_ce(p->ed, cpu, 1, "icache");
66 "CacheErr (Dcache):%llx, core %d/cpu %d, cp0_errorepc == %lx\n",
67 (unsigned long long)dcache_err, core, cpu,
70 edac_device_handle_ue(p->ed, cpu, 0, "dcache");
72 edac_device_handle_ce(p->ed, cpu, 0, "dcache");
94 p->ed = edac_device_alloc_ctl_info(0, "cpu", num_possible_cpus(),
104 p->ed->mod_name = "octeon-cpu";
[all...]
/linux-master/arch/powerpc/mm/
H A Dnuma.c16 #include <linux/cpu.h>
140 unsigned int cpu; local
142 for_each_possible_cpu(cpu)
143 numa_cpu_lookup_table[cpu] = -1;
146 void map_cpu_to_node(int cpu, int node) argument
148 update_numa_cpu_lookup_table(cpu, node);
150 if (!(cpumask_test_cpu(cpu, node_to_cpumask_map[node]))) {
151 pr_debug("adding cpu %d to node %d\n", cpu, node);
152 cpumask_set_cpu(cpu, node_to_cpumask_ma
157 unmap_cpu_from_node(unsigned long cpu) argument
706 struct device_node *cpu; local
768 verify_cpu_node_mapping(int cpu, int node) argument
790 ppc_numa_cpu_prepare(unsigned int cpu) argument
799 ppc_numa_cpu_dead(unsigned int cpu) argument
935 struct device_node *cpu; local
1050 unsigned int cpu, count; local
1166 int cpu; local
1389 vphn_get_associativity(unsigned long cpu, __be32 *associativity) argument
1422 find_and_update_cpu_nid(int cpu) argument
1444 cpu_to_coregroup_id(int cpu) argument
[all...]
/linux-master/arch/powerpc/platforms/ps3/
H A Dinterrupt.c30 * struct ps3_bmp - a per cpu irq status and mask bitmap structure
68 * struct ps3_private - a per cpu data structure
160 * @cpu: enum ps3_cpu_binding indicating the cpu the interrupt should be
169 static int ps3_virq_setup(enum ps3_cpu_binding cpu, unsigned long outlet, argument
177 if (cpu == PS3_BINDING_CPU_ANY)
178 cpu = 0;
180 pd = &per_cpu(ps3_private, cpu);
191 DBG("%s:%d: outlet %lu => cpu %u, virq %u\n", __func__, __LINE__,
192 outlet, cpu, *vir
243 ps3_irq_plug_setup(enum ps3_cpu_binding cpu, unsigned long outlet, unsigned int *virq) argument
321 ps3_event_receive_port_setup(enum ps3_cpu_binding cpu, unsigned int *virq) argument
390 ps3_sb_event_receive_port_setup(struct ps3_system_bus_device *dev, enum ps3_cpu_binding cpu, unsigned int *virq) argument
466 ps3_io_irq_setup(enum ps3_cpu_binding cpu, unsigned int interrupt_id, unsigned int *virq) argument
523 ps3_vuart_irq_setup(enum ps3_cpu_binding cpu, void* virt_addr_bmp, unsigned int *virq) argument
579 ps3_spe_irq_setup(enum ps3_cpu_binding cpu, unsigned long spe_id, unsigned int class, unsigned int *virq) argument
618 _dump_64_bmp(const char *header, const u64 *p, unsigned cpu, const char* func, int line) argument
627 _dump_256_bmp(const char *header, const u64 *p, unsigned cpu, const char* func, int line) argument
682 ps3_register_ipi_debug_brk(unsigned int cpu, unsigned int virq) argument
692 ps3_register_ipi_irq(unsigned int cpu, unsigned int virq) argument
743 unsigned cpu; local
772 ps3_shutdown_IRQ(int cpu) argument
[all...]
/linux-master/arch/x86/kernel/cpu/mce/
H A Dinject.c19 #include <linux/cpu.h>
179 int cpu = smp_processor_id(); local
181 if (!cpumask_test_cpu(cpu, mce_inject_cpumask))
183 cpumask_clear_cpu(cpu, mce_inject_cpumask);
193 int cpu = smp_processor_id(); local
196 if (cpumask_test_cpu(cpu, mce_inject_cpumask) &&
198 cpumask_clear_cpu(cpu, mce_inject_cpumask);
209 int cpu = m->extcpu; local
212 pr_info("Triggering MCE exception on CPU %d\n", cpu);
228 pr_info("MCE exception done on CPU %d\n", cpu);
251 int cpu; local
317 toggle_hw_mce_inject(unsigned int cpu, bool enable) argument
502 unsigned int cpu = i_mce.extcpu; local
723 int cpu; local
[all...]
/linux-master/drivers/gpu/drm/lima/
H A Dlima_vm.c43 vm->bts[pbe].cpu[bte] = 0;
52 if (!vm->bts[pbe].cpu) {
57 vm->bts[pbe].cpu = dma_alloc_wc(
60 if (!vm->bts[pbe].cpu)
64 pd = vm->pd.cpu + (pbe << LIMA_VM_NUM_PT_PER_BT_SHIFT);
71 vm->bts[pbe].cpu[bte] = pa | LIMA_VM_FLAGS_CACHE;
212 vm->pd.cpu = dma_alloc_wc(dev->dev, LIMA_PAGE_SIZE, &vm->pd.dma,
214 if (!vm->pd.cpu)
229 dma_free_wc(dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma);
243 if (vm->bts[i].cpu)
[all...]
/linux-master/arch/powerpc/platforms/pseries/
H A Ddtl.c24 int cpu; member in struct:dtl
35 * Size of per-cpu log buffers. Firmware requires that the buffer does
43 * When CONFIG_VIRT_CPU_ACCOUNTING_NATIVE = y, the cpu accounting code controls
62 * The cpu accounting code controls the DTL ring buffer, and we get
93 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu);
104 lppaca_of(dtl->cpu).dtl_enable_mask |= dtl_event_mask;
113 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu);
121 lppaca_of(dtl->cpu).dtl_enable_mask = DTL_LOG_PREEMPT;
129 return per_cpu(dtl_rings, dtl->cpu).write_index;
143 hwcpu = get_hard_smp_processor_id(dtl->cpu);
[all...]
/linux-master/arch/sparc/kernel/
H A Dcpumap.c96 static int cpuinfo_id(int cpu, int level) argument
105 id = cpu_to_node(cpu);
108 id = cpu_data(cpu).core_id;
111 id = cpu_data(cpu).proc_id;
193 int n, id, cpu, prev_cpu, last_cpu, level; local
204 prev_cpu = cpu = cpumask_first(cpu_online_mask);
213 id = cpuinfo_id(cpu, level);
227 ? cpu : new_tree->level[level + 1].start_index;
238 while (++cpu <= last_cpu) {
239 if (!cpu_online(cpu))
[all...]
H A Dsysfs.c8 #include <linux/cpu.h>
206 static DEFINE_PER_CPU(struct cpu, cpu_devices);
208 static int register_cpu_online(unsigned int cpu) argument
210 struct cpu *c = &per_cpu(cpu_devices, cpu);
221 static int unregister_cpu_online(unsigned int cpu) argument
224 struct cpu *c = &per_cpu(cpu_devices, cpu);
249 int cpu, ret; local
253 for_each_possible_cpu(cpu) {
[all...]
H A Dsun4m_smp.c13 #include <linux/cpu.h>
94 * we need to re-initialize it for each cpu
126 /* setup cpu list for irq rotation */
139 static void sun4m_send_ipi(int cpu, int level) argument
141 sbus_writel(SUN4M_SOFT_INT(level), &sun4m_irq_percpu[cpu]->set);
144 static void sun4m_ipi_resched(int cpu) argument
146 sun4m_send_ipi(cpu, IRQ_IPI_RESCHED);
149 static void sun4m_ipi_single(int cpu) argument
151 sun4m_send_ipi(cpu, IRQ_IPI_SINGLE);
154 static void sun4m_ipi_mask_one(int cpu) argument
247 int cpu = smp_processor_id(); local
[all...]
/linux-master/arch/loongarch/kernel/
H A Dnuma.c42 * apicid, cpu, node mappings
55 static int __init pcpu_cpu_to_node(int cpu) argument
57 return early_cpu_to_node(cpu);
76 unsigned int cpu; local
104 for_each_possible_cpu(cpu)
105 __per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu];
110 * Get nodeid by logical cpu number.
111 * __cpuid_to_node maps phyical cpu id to node, so we
112 * should use cpu_logical_map(cpu) t
118 early_cpu_to_node(int cpu) argument
130 int cpu = __cpu_number_map[cpuid]; local
139 numa_add_cpu(unsigned int cpu) argument
145 numa_remove_cpu(unsigned int cpu) argument
[all...]
/linux-master/arch/alpha/kernel/
H A Dtime.c94 int cpu = smp_processor_id();
95 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu);
119 int cpu = smp_processor_id();
120 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu);
126 .cpumask = cpumask_of(cpu),
177 int cpu = smp_processor_id();
178 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu);
187 int cpu = smp_processor_id();
188 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu);
194 .cpumask = cpumask_of(cpu),
93 int cpu = smp_processor_id(); local
118 int cpu = smp_processor_id(); local
175 int cpu = smp_processor_id(); local
185 int cpu = smp_processor_id(); local
313 struct percpu_struct *cpu; local
[all...]
H A Dcore_mcpcia.c92 unsigned int stat0, value, cpu; local
94 cpu = smp_processor_id();
110 mcheck_expected(cpu) = 1;
111 mcheck_taken(cpu) = 0;
112 mcheck_extra(cpu) = mid;
120 if (mcheck_taken(cpu)) {
121 mcheck_taken(cpu) = 0;
125 mcheck_expected(cpu) = 0;
140 unsigned int stat0, cpu; local
142 cpu
256 int cpu = smp_processor_id(); local
579 unsigned int cpu = smp_processor_id(); local
[all...]
/linux-master/drivers/irqchip/
H A Dirq-csky-mpintc.c131 unsigned int cpu; local
135 cpu = cpumask_any_and(mask_val, cpu_online_mask);
137 cpu = cpumask_first(mask_val);
139 if (cpu >= nr_cpu_ids)
144 * could deliver external irq to one cpu or all cpus. So it
152 cpu = 0;
154 cpu |= BIT(31);
156 writel_relaxed(cpu, INTCG_base + INTCG_CIDSTR + offset);
158 irq_data_update_effective_affinity(d, cpumask_of(cpu));
231 unsigned int cpu, nr_ir local
[all...]
/linux-master/arch/s390/lib/
H A Dspinlock.c60 void arch_spin_lock_setup(int cpu) argument
65 node = per_cpu_ptr(&spin_wait[0], cpu);
68 node->node_id = ((cpu + 1) << _Q_TAIL_CPU_OFFSET) +
99 int ix, cpu; local
102 cpu = (lock & _Q_TAIL_CPU_MASK) >> _Q_TAIL_CPU_OFFSET;
103 return per_cpu_ptr(&spin_wait[ix], cpu - 1);
111 return 0; /* 0 -> no target cpu */
124 lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */
215 lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */
254 int cpu local
315 int cpu; local
[all...]
/linux-master/arch/powerpc/platforms/cell/
H A Dras.c27 static void dump_fir(int cpu) argument
29 struct cbe_pmd_regs __iomem *pregs = cbe_get_cpu_pmd_regs(cpu);
30 struct cbe_iic_regs __iomem *iregs = cbe_get_cpu_iic_regs(cpu);
54 int cpu = smp_processor_id(); local
56 printk(KERN_ERR "System Error Interrupt on CPU %d !\n", cpu);
57 dump_fir(cpu);
63 int cpu = smp_processor_id(); local
69 printk(KERN_ERR "Unhandled Maintenance interrupt on CPU %d !\n", cpu);
75 int cpu = smp_processor_id(); local
81 printk(KERN_ERR "Unhandled Thermal interrupt on CPU %d !\n", cpu);
87 int cpu = smp_processor_id(); local
[all...]
/linux-master/drivers/acpi/
H A Dacpi_extlog.c16 #include <asm/cpu.h>
63 #define ELOG_IDX(cpu, bank) \
64 (cpu_physical_id(cpu) * l1_percpu_entry + (bank))
72 static struct acpi_hest_generic_status *extlog_elog_entry_check(int cpu, int bank) argument
78 WARN_ON(cpu < 0);
79 idx = ELOG_IDX(cpu, bank);
95 struct acpi_hest_generic_status *estatus, int cpu)
109 printk("%s""Hardware error detected on CPU%d\n", pfx_seq, cpu);
114 struct acpi_hest_generic_status *estatus, int cpu)
127 __print_extlog_rcd(pfx, estatus, cpu);
94 __print_extlog_rcd(const char *pfx, struct acpi_hest_generic_status *estatus, int cpu) argument
113 print_extlog_rcd(const char *pfx, struct acpi_hest_generic_status *estatus, int cpu) argument
139 int cpu = mce->extcpu; local
[all...]
/linux-master/tools/perf/util/
H A Dbpf_kwork.c32 u32 cpu; member in struct:work_key
153 struct perf_cpu cpu; local
157 pr_debug("Invalid cpu filter fd\n");
168 perf_cpu_map__for_each_cpu(cpu, idx, map) {
171 if (cpu.cpu >= nr_cpus) {
173 pr_err("Requested cpu %d too large\n", cpu.cpu);
176 bpf_map_update_elem(fd, &cpu
[all...]
/linux-master/arch/x86/kernel/apic/
H A Dvector.c30 unsigned int cpu; member in struct:apic_chip_data
129 unsigned int cpu)
136 apicd->hw_irq_cfg.dest_apicid = apic->calc_dest_apicid(cpu);
137 irq_data_update_effective_affinity(irqd, cpumask_of(cpu));
138 trace_vector_config(irqd->irq, vector, cpu,
152 apicd->cpu);
171 if (cpu_online(apicd->cpu)) {
174 apicd->prev_cpu = apicd->cpu;
175 WARN_ON_ONCE(apicd->cpu == newcpu);
177 irq_matrix_free(vector_matrix, apicd->cpu, apic
128 apic_update_irq_cfg(struct irq_data *irqd, unsigned int vector, unsigned int cpu) argument
190 unsigned int cpu = cpumask_first(cpu_online_mask); local
237 unsigned int cpu = apicd->cpu; local
328 int vector, cpu; local
938 unsigned int cpu = apicd->prev_cpu; local
1010 unsigned int cpu = apicd->prev_cpu; local
1076 unsigned int cpu = smp_processor_id(); local
1171 unsigned int rsvd, avl, tomove, cpu = smp_processor_id(); local
1316 int cpu; local
[all...]

Completed in 245 milliseconds

<<21222324252627282930>>