Lines Matching refs:vpe
50 #define ltq_icu_w32(vpe, m, x, y) \
51 ltq_w32((x), ltq_icu_membase[vpe] + m*LTQ_ICU_IM_SIZE + (y))
53 #define ltq_icu_r32(vpe, m, x) \
54 ltq_r32(ltq_icu_membase[vpe] + m*LTQ_ICU_IM_SIZE + (x))
83 int vpe;
88 for_each_present_cpu(vpe) {
89 ltq_icu_w32(vpe, im,
90 ltq_icu_r32(vpe, im, LTQ_ICU_IER) & ~BIT(offset),
101 int vpe;
106 for_each_present_cpu(vpe) {
107 ltq_icu_w32(vpe, im,
108 ltq_icu_r32(vpe, im, LTQ_ICU_IER) & ~BIT(offset),
110 ltq_icu_w32(vpe, im, BIT(offset), LTQ_ICU_ISR);
120 int vpe;
125 for_each_present_cpu(vpe) {
126 ltq_icu_w32(vpe, im, BIT(offset), LTQ_ICU_ISR);
136 int vpe;
140 vpe = cpumask_first(irq_data_get_effective_affinity_mask(d));
143 if (unlikely(vpe >= nr_cpu_ids))
144 vpe = smp_processor_id();
148 ltq_icu_w32(vpe, im, ltq_icu_r32(vpe, im, LTQ_ICU_IER) | BIT(offset),
291 int vpe = smp_processor_id();
293 irq = ltq_icu_r32(vpe, module, LTQ_ICU_IOSR);
342 int i, ret, vpe;
345 for_each_possible_cpu(vpe) {
346 if (of_address_to_resource(node, vpe, &res))
347 panic("Failed to get icu%i memory range", vpe);
351 pr_err("Failed to request icu%i memory\n", vpe);
353 ltq_icu_membase[vpe] = ioremap(res.start,
356 if (!ltq_icu_membase[vpe])
357 panic("Failed to remap icu%i memory", vpe);
361 for_each_possible_cpu(vpe) {
364 ltq_icu_w32(vpe, i, 0, LTQ_ICU_IER);
367 ltq_icu_w32(vpe, i, ~0, LTQ_ICU_ISR);
368 ltq_icu_w32(vpe, i, ~0, LTQ_ICU_IMR);
371 ltq_icu_w32(vpe, i, 0, LTQ_ICU_IRSR);