/linux-master/arch/x86/include/asm/trace/ |
H A D | hyperv.h | 75 TP_PROTO(int cpu, 77 TP_ARGS(cpu, vector), 79 __field(int, cpu) 82 TP_fast_assign(__entry->cpu = cpu; 85 TP_printk("cpu %d vector %x", 86 __entry->cpu, __entry->vector)
|
/linux-master/tools/testing/selftests/rcutorture/bin/ |
H A D | jitter.sh | 31 for i in /sys/devices/system/cpu/cpu[0-9]* 58 if cpus=`grep 1 /sys/devices/system/cpu/*/online 2>&1 |
|
H A D | kvm-assign-cpus.sh | 27 for j in $i/cpu*/cache/index* 37 indexlist="`ls -d $i/cpu* | grep 'cpu[0-9][0-9]*' | head -1 | sed -e 's,^.*$,ls -d &/cache/index*,' | sh | sed -e 's,^.*/,,'`" 40 for i in node*/cpu*/cache/index*/shared_cpu_list 56 f="$n/cpu*/cache/$i/shared_cpu_list" 81 cat $n/cpu*/cache/$splitindex/shared_cpu_list | sort -u -k1n | 94 print "cpu[" nodenum "][" idx "] = " j ";";
|
/linux-master/arch/x86/kernel/ |
H A D | irq_32.c | 17 #include <linux/cpu.h> 108 * Allocate per-cpu stacks for hardirq and softirq processing 110 int irq_init_percpu_irqstack(unsigned int cpu) argument 112 int node = cpu_to_node(cpu); 115 if (per_cpu(pcpu_hot.hardirq_stack_ptr, cpu)) 127 per_cpu(pcpu_hot.hardirq_stack_ptr, cpu) = page_address(ph); 128 per_cpu(pcpu_hot.softirq_stack_ptr, cpu) = page_address(ps);
|
/linux-master/drivers/cpufreq/ |
H A D | ppc_cbe_cpufreq_pmi.c | 34 int cbe_cpufreq_set_pmode_pmi(int cpu, unsigned int pmode) argument 42 pmi_msg.data1 = cbe_cpu_to_node(cpu); 69 int cpu, ret; local 76 cpu = cbe_node_to_cpu(node); 80 policy = cpufreq_cpu_get(cpu); 82 pr_warn("cpufreq policy not found cpu%d\n", cpu);
|
/linux-master/arch/openrisc/kernel/ |
H A D | time.c | 71 unsigned int cpu = smp_processor_id(); local 73 &per_cpu(clockevent_openrisc_timer, cpu); 74 struct cpuinfo_or1k *cpuinfo = &cpuinfo_or1k[cpu]; 86 evt->cpumask = cpumask_of(cpu); 116 unsigned int cpu = smp_processor_id(); local 118 &per_cpu(clockevent_openrisc_timer, cpu);
|
/linux-master/drivers/pinctrl/renesas/ |
H A D | Makefile | 56 CFLAGS_pfc-sh7203.o += -I$(srctree)/arch/sh/include/cpu-sh2a 57 CFLAGS_pfc-sh7264.o += -I$(srctree)/arch/sh/include/cpu-sh2a 58 CFLAGS_pfc-sh7269.o += -I$(srctree)/arch/sh/include/cpu-sh2a 59 CFLAGS_pfc-sh7720.o += -I$(srctree)/arch/sh/include/cpu-sh3 60 CFLAGS_pfc-sh7722.o += -I$(srctree)/arch/sh/include/cpu-sh4 61 CFLAGS_pfc-sh7723.o += -I$(srctree)/arch/sh/include/cpu-sh4 62 CFLAGS_pfc-sh7724.o += -I$(srctree)/arch/sh/include/cpu-sh4 63 CFLAGS_pfc-sh7734.o += -I$(srctree)/arch/sh/include/cpu-sh4 64 CFLAGS_pfc-sh7757.o += -I$(srctree)/arch/sh/include/cpu-sh4 65 CFLAGS_pfc-sh7785.o += -I$(srctree)/arch/sh/include/cpu [all...] |
/linux-master/drivers/edac/ |
H A D | octeon_edac-pc.c | 45 unsigned int cpu = smp_processor_id(); local 58 "CacheErr (Icache):%llx, core %d/cpu %d, cp0_errorepc == %lx\n", 59 (unsigned long long)icache_err, core, cpu, 62 edac_device_handle_ce(p->ed, cpu, 1, "icache"); 66 "CacheErr (Dcache):%llx, core %d/cpu %d, cp0_errorepc == %lx\n", 67 (unsigned long long)dcache_err, core, cpu, 70 edac_device_handle_ue(p->ed, cpu, 0, "dcache"); 72 edac_device_handle_ce(p->ed, cpu, 0, "dcache"); 94 p->ed = edac_device_alloc_ctl_info(0, "cpu", num_possible_cpus(), 104 p->ed->mod_name = "octeon-cpu"; [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | numa.c | 16 #include <linux/cpu.h> 140 unsigned int cpu; local 142 for_each_possible_cpu(cpu) 143 numa_cpu_lookup_table[cpu] = -1; 146 void map_cpu_to_node(int cpu, int node) argument 148 update_numa_cpu_lookup_table(cpu, node); 150 if (!(cpumask_test_cpu(cpu, node_to_cpumask_map[node]))) { 151 pr_debug("adding cpu %d to node %d\n", cpu, node); 152 cpumask_set_cpu(cpu, node_to_cpumask_ma 157 unmap_cpu_from_node(unsigned long cpu) argument 706 struct device_node *cpu; local 768 verify_cpu_node_mapping(int cpu, int node) argument 790 ppc_numa_cpu_prepare(unsigned int cpu) argument 799 ppc_numa_cpu_dead(unsigned int cpu) argument 935 struct device_node *cpu; local 1050 unsigned int cpu, count; local 1166 int cpu; local 1389 vphn_get_associativity(unsigned long cpu, __be32 *associativity) argument 1422 find_and_update_cpu_nid(int cpu) argument 1444 cpu_to_coregroup_id(int cpu) argument [all...] |
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | interrupt.c | 30 * struct ps3_bmp - a per cpu irq status and mask bitmap structure 68 * struct ps3_private - a per cpu data structure 160 * @cpu: enum ps3_cpu_binding indicating the cpu the interrupt should be 169 static int ps3_virq_setup(enum ps3_cpu_binding cpu, unsigned long outlet, argument 177 if (cpu == PS3_BINDING_CPU_ANY) 178 cpu = 0; 180 pd = &per_cpu(ps3_private, cpu); 191 DBG("%s:%d: outlet %lu => cpu %u, virq %u\n", __func__, __LINE__, 192 outlet, cpu, *vir 243 ps3_irq_plug_setup(enum ps3_cpu_binding cpu, unsigned long outlet, unsigned int *virq) argument 321 ps3_event_receive_port_setup(enum ps3_cpu_binding cpu, unsigned int *virq) argument 390 ps3_sb_event_receive_port_setup(struct ps3_system_bus_device *dev, enum ps3_cpu_binding cpu, unsigned int *virq) argument 466 ps3_io_irq_setup(enum ps3_cpu_binding cpu, unsigned int interrupt_id, unsigned int *virq) argument 523 ps3_vuart_irq_setup(enum ps3_cpu_binding cpu, void* virt_addr_bmp, unsigned int *virq) argument 579 ps3_spe_irq_setup(enum ps3_cpu_binding cpu, unsigned long spe_id, unsigned int class, unsigned int *virq) argument 618 _dump_64_bmp(const char *header, const u64 *p, unsigned cpu, const char* func, int line) argument 627 _dump_256_bmp(const char *header, const u64 *p, unsigned cpu, const char* func, int line) argument 682 ps3_register_ipi_debug_brk(unsigned int cpu, unsigned int virq) argument 692 ps3_register_ipi_irq(unsigned int cpu, unsigned int virq) argument 743 unsigned cpu; local 772 ps3_shutdown_IRQ(int cpu) argument [all...] |
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | inject.c | 19 #include <linux/cpu.h> 179 int cpu = smp_processor_id(); local 181 if (!cpumask_test_cpu(cpu, mce_inject_cpumask)) 183 cpumask_clear_cpu(cpu, mce_inject_cpumask); 193 int cpu = smp_processor_id(); local 196 if (cpumask_test_cpu(cpu, mce_inject_cpumask) && 198 cpumask_clear_cpu(cpu, mce_inject_cpumask); 209 int cpu = m->extcpu; local 212 pr_info("Triggering MCE exception on CPU %d\n", cpu); 228 pr_info("MCE exception done on CPU %d\n", cpu); 251 int cpu; local 317 toggle_hw_mce_inject(unsigned int cpu, bool enable) argument 502 unsigned int cpu = i_mce.extcpu; local 723 int cpu; local [all...] |
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_vm.c | 43 vm->bts[pbe].cpu[bte] = 0; 52 if (!vm->bts[pbe].cpu) { 57 vm->bts[pbe].cpu = dma_alloc_wc( 60 if (!vm->bts[pbe].cpu) 64 pd = vm->pd.cpu + (pbe << LIMA_VM_NUM_PT_PER_BT_SHIFT); 71 vm->bts[pbe].cpu[bte] = pa | LIMA_VM_FLAGS_CACHE; 212 vm->pd.cpu = dma_alloc_wc(dev->dev, LIMA_PAGE_SIZE, &vm->pd.dma, 214 if (!vm->pd.cpu) 229 dma_free_wc(dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma); 243 if (vm->bts[i].cpu) [all...] |
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | dtl.c | 24 int cpu; member in struct:dtl 35 * Size of per-cpu log buffers. Firmware requires that the buffer does 43 * When CONFIG_VIRT_CPU_ACCOUNTING_NATIVE = y, the cpu accounting code controls 62 * The cpu accounting code controls the DTL ring buffer, and we get 93 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu); 104 lppaca_of(dtl->cpu).dtl_enable_mask |= dtl_event_mask; 113 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu); 121 lppaca_of(dtl->cpu).dtl_enable_mask = DTL_LOG_PREEMPT; 129 return per_cpu(dtl_rings, dtl->cpu).write_index; 143 hwcpu = get_hard_smp_processor_id(dtl->cpu); [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | cpumap.c | 96 static int cpuinfo_id(int cpu, int level) argument 105 id = cpu_to_node(cpu); 108 id = cpu_data(cpu).core_id; 111 id = cpu_data(cpu).proc_id; 193 int n, id, cpu, prev_cpu, last_cpu, level; local 204 prev_cpu = cpu = cpumask_first(cpu_online_mask); 213 id = cpuinfo_id(cpu, level); 227 ? cpu : new_tree->level[level + 1].start_index; 238 while (++cpu <= last_cpu) { 239 if (!cpu_online(cpu)) [all...] |
H A D | sysfs.c | 8 #include <linux/cpu.h> 206 static DEFINE_PER_CPU(struct cpu, cpu_devices); 208 static int register_cpu_online(unsigned int cpu) argument 210 struct cpu *c = &per_cpu(cpu_devices, cpu); 221 static int unregister_cpu_online(unsigned int cpu) argument 224 struct cpu *c = &per_cpu(cpu_devices, cpu); 249 int cpu, ret; local 253 for_each_possible_cpu(cpu) { [all...] |
H A D | sun4m_smp.c | 13 #include <linux/cpu.h> 94 * we need to re-initialize it for each cpu 126 /* setup cpu list for irq rotation */ 139 static void sun4m_send_ipi(int cpu, int level) argument 141 sbus_writel(SUN4M_SOFT_INT(level), &sun4m_irq_percpu[cpu]->set); 144 static void sun4m_ipi_resched(int cpu) argument 146 sun4m_send_ipi(cpu, IRQ_IPI_RESCHED); 149 static void sun4m_ipi_single(int cpu) argument 151 sun4m_send_ipi(cpu, IRQ_IPI_SINGLE); 154 static void sun4m_ipi_mask_one(int cpu) argument 247 int cpu = smp_processor_id(); local [all...] |
/linux-master/arch/loongarch/kernel/ |
H A D | numa.c | 42 * apicid, cpu, node mappings 55 static int __init pcpu_cpu_to_node(int cpu) argument 57 return early_cpu_to_node(cpu); 76 unsigned int cpu; local 104 for_each_possible_cpu(cpu) 105 __per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu]; 110 * Get nodeid by logical cpu number. 111 * __cpuid_to_node maps phyical cpu id to node, so we 112 * should use cpu_logical_map(cpu) t 118 early_cpu_to_node(int cpu) argument 130 int cpu = __cpu_number_map[cpuid]; local 139 numa_add_cpu(unsigned int cpu) argument 145 numa_remove_cpu(unsigned int cpu) argument [all...] |
/linux-master/arch/alpha/kernel/ |
H A D | time.c | 94 int cpu = smp_processor_id(); 95 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu); 119 int cpu = smp_processor_id(); 120 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu); 126 .cpumask = cpumask_of(cpu), 177 int cpu = smp_processor_id(); 178 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu); 187 int cpu = smp_processor_id(); 188 struct clock_event_device *ce = &per_cpu(cpu_ce, cpu); 194 .cpumask = cpumask_of(cpu), 93 int cpu = smp_processor_id(); local 118 int cpu = smp_processor_id(); local 175 int cpu = smp_processor_id(); local 185 int cpu = smp_processor_id(); local 313 struct percpu_struct *cpu; local [all...] |
H A D | core_mcpcia.c | 92 unsigned int stat0, value, cpu; local 94 cpu = smp_processor_id(); 110 mcheck_expected(cpu) = 1; 111 mcheck_taken(cpu) = 0; 112 mcheck_extra(cpu) = mid; 120 if (mcheck_taken(cpu)) { 121 mcheck_taken(cpu) = 0; 125 mcheck_expected(cpu) = 0; 140 unsigned int stat0, cpu; local 142 cpu 256 int cpu = smp_processor_id(); local 579 unsigned int cpu = smp_processor_id(); local [all...] |
/linux-master/drivers/irqchip/ |
H A D | irq-csky-mpintc.c | 131 unsigned int cpu; local 135 cpu = cpumask_any_and(mask_val, cpu_online_mask); 137 cpu = cpumask_first(mask_val); 139 if (cpu >= nr_cpu_ids) 144 * could deliver external irq to one cpu or all cpus. So it 152 cpu = 0; 154 cpu |= BIT(31); 156 writel_relaxed(cpu, INTCG_base + INTCG_CIDSTR + offset); 158 irq_data_update_effective_affinity(d, cpumask_of(cpu)); 231 unsigned int cpu, nr_ir local [all...] |
/linux-master/arch/s390/lib/ |
H A D | spinlock.c | 60 void arch_spin_lock_setup(int cpu) argument 65 node = per_cpu_ptr(&spin_wait[0], cpu); 68 node->node_id = ((cpu + 1) << _Q_TAIL_CPU_OFFSET) + 99 int ix, cpu; local 102 cpu = (lock & _Q_TAIL_CPU_MASK) >> _Q_TAIL_CPU_OFFSET; 103 return per_cpu_ptr(&spin_wait[ix], cpu - 1); 111 return 0; /* 0 -> no target cpu */ 124 lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */ 215 lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */ 254 int cpu local 315 int cpu; local [all...] |
/linux-master/arch/powerpc/platforms/cell/ |
H A D | ras.c | 27 static void dump_fir(int cpu) argument 29 struct cbe_pmd_regs __iomem *pregs = cbe_get_cpu_pmd_regs(cpu); 30 struct cbe_iic_regs __iomem *iregs = cbe_get_cpu_iic_regs(cpu); 54 int cpu = smp_processor_id(); local 56 printk(KERN_ERR "System Error Interrupt on CPU %d !\n", cpu); 57 dump_fir(cpu); 63 int cpu = smp_processor_id(); local 69 printk(KERN_ERR "Unhandled Maintenance interrupt on CPU %d !\n", cpu); 75 int cpu = smp_processor_id(); local 81 printk(KERN_ERR "Unhandled Thermal interrupt on CPU %d !\n", cpu); 87 int cpu = smp_processor_id(); local [all...] |
/linux-master/drivers/acpi/ |
H A D | acpi_extlog.c | 16 #include <asm/cpu.h> 63 #define ELOG_IDX(cpu, bank) \ 64 (cpu_physical_id(cpu) * l1_percpu_entry + (bank)) 72 static struct acpi_hest_generic_status *extlog_elog_entry_check(int cpu, int bank) argument 78 WARN_ON(cpu < 0); 79 idx = ELOG_IDX(cpu, bank); 95 struct acpi_hest_generic_status *estatus, int cpu) 109 printk("%s""Hardware error detected on CPU%d\n", pfx_seq, cpu); 114 struct acpi_hest_generic_status *estatus, int cpu) 127 __print_extlog_rcd(pfx, estatus, cpu); 94 __print_extlog_rcd(const char *pfx, struct acpi_hest_generic_status *estatus, int cpu) argument 113 print_extlog_rcd(const char *pfx, struct acpi_hest_generic_status *estatus, int cpu) argument 139 int cpu = mce->extcpu; local [all...] |
/linux-master/tools/perf/util/ |
H A D | bpf_kwork.c | 32 u32 cpu; member in struct:work_key 153 struct perf_cpu cpu; local 157 pr_debug("Invalid cpu filter fd\n"); 168 perf_cpu_map__for_each_cpu(cpu, idx, map) { 171 if (cpu.cpu >= nr_cpus) { 173 pr_err("Requested cpu %d too large\n", cpu.cpu); 176 bpf_map_update_elem(fd, &cpu [all...] |
/linux-master/arch/x86/kernel/apic/ |
H A D | vector.c | 30 unsigned int cpu; member in struct:apic_chip_data 129 unsigned int cpu) 136 apicd->hw_irq_cfg.dest_apicid = apic->calc_dest_apicid(cpu); 137 irq_data_update_effective_affinity(irqd, cpumask_of(cpu)); 138 trace_vector_config(irqd->irq, vector, cpu, 152 apicd->cpu); 171 if (cpu_online(apicd->cpu)) { 174 apicd->prev_cpu = apicd->cpu; 175 WARN_ON_ONCE(apicd->cpu == newcpu); 177 irq_matrix_free(vector_matrix, apicd->cpu, apic 128 apic_update_irq_cfg(struct irq_data *irqd, unsigned int vector, unsigned int cpu) argument 190 unsigned int cpu = cpumask_first(cpu_online_mask); local 237 unsigned int cpu = apicd->cpu; local 328 int vector, cpu; local 938 unsigned int cpu = apicd->prev_cpu; local 1010 unsigned int cpu = apicd->prev_cpu; local 1076 unsigned int cpu = smp_processor_id(); local 1171 unsigned int rsvd, avl, tomove, cpu = smp_processor_id(); local 1316 int cpu; local [all...] |