Lines Matching refs:cpu
24 int cpu;
35 * Size of per-cpu log buffers. Firmware requires that the buffer does
43 * When CONFIG_VIRT_CPU_ACCOUNTING_NATIVE = y, the cpu accounting code controls
62 * The cpu accounting code controls the DTL ring buffer, and we get
93 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu);
104 lppaca_of(dtl->cpu).dtl_enable_mask |= dtl_event_mask;
113 struct dtl_ring *dtlr = &per_cpu(dtl_rings, dtl->cpu);
121 lppaca_of(dtl->cpu).dtl_enable_mask = DTL_LOG_PREEMPT;
129 return per_cpu(dtl_rings, dtl->cpu).write_index;
143 hwcpu = get_hard_smp_processor_id(dtl->cpu);
147 printk(KERN_WARNING "%s: DTL registration for cpu %d (hw %d) "
148 "failed with %d\n", __func__, dtl->cpu, hwcpu, ret);
153 lppaca_of(dtl->cpu).dtl_idx = 0;
160 lppaca_of(dtl->cpu).dtl_enable_mask = dtl_event_mask;
167 int hwcpu = get_hard_smp_processor_id(dtl->cpu);
169 lppaca_of(dtl->cpu).dtl_enable_mask = 0x0;
176 return be64_to_cpu(lppaca_of(dtl->cpu).dtl_idx);
198 buf = kmem_cache_alloc_node(dtl_cache, GFP_KERNEL, cpu_to_node(dtl->cpu));
200 printk(KERN_WARNING "%s: buffer alloc failed for cpu %d\n",
201 __func__, dtl->cpu);
337 sprintf(name, "cpu-%d", dtl->cpu);
356 /* set up the per-cpu log structures */
360 dtl->cpu = i;