/linux-master/drivers/hwtracing/coresight/ |
H A D | coresight-cpu-debug.c | 10 #include <linux/cpu.h> 92 int cpu; member in struct:debug_drvdata 175 __func__, drvdata->cpu); 383 int cpu; local 395 for_each_possible_cpu(cpu) { 396 drvdata = per_cpu(debug_drvdata, cpu); 400 dev_emerg(drvdata->dev, "CPU[%d]:\n", drvdata->cpu); 418 int cpu, ret = 0; local 427 for_each_possible_cpu(cpu) { 428 drvdata = per_cpu(debug_drvdata, cpu); 457 int cpu, ret, err = 0; local [all...] |
/linux-master/drivers/irqchip/ |
H A D | irq-sifive-plic.c | 6 #include <linux/cpu.h> 116 int cpu; local 118 for_each_cpu(cpu, mask) { 119 struct plic_handler *handler = per_cpu_ptr(&plic_handlers, cpu); 166 unsigned int cpu; local 170 cpu = cpumask_first_and(&priv->lmask, mask_val); 172 cpu = cpumask_first_and_and(&priv->lmask, mask_val, cpu_online_mask); 174 if (cpu >= nr_cpu_ids) 179 irq_data_update_effective_affinity(d, cpumask_of(cpu)); 243 unsigned int i, cpu; local 275 unsigned int i, index, cpu; local 395 plic_dying_cpu(unsigned int cpu) argument 403 plic_starting_cpu(unsigned int cpu) argument 485 int error = 0, nr_contexts, nr_handlers = 0, cpu, i; local [all...] |
H A D | irq-bcm7038-l1.c | 122 struct bcm7038_l1_cpu *cpu; local 127 cpu = intc->cpus[cpu_logical_map(smp_processor_id())]; 129 cpu = intc->cpus[0]; 140 pending = l1_readl(cpu->map_base + reg_status(intc, idx)) & 141 ~cpu->mask_cache[idx]; 228 struct bcm7038_l1_cpu *cpu; local 252 cpu = intc->cpus[idx] = kzalloc(struct_size(cpu, mask_cache, n_words), 254 if (!cpu) 257 cpu 444 struct bcm7038_l1_cpu *cpu = intc->cpus[idx]; local [all...] |
H A D | irq-mips-gic.c | 111 static void gic_send_ipi(struct irq_data *d, unsigned int cpu) argument 157 /* Get per-cpu bitmaps */ 190 unsigned int cpu; local 195 cpu = cpumask_first(irq_data_get_effective_affinity_mask(d)); 196 set_bit(intr, per_cpu_ptr(pcpu_masks, cpu)); 264 unsigned int cpu; local 266 cpu = cpumask_first_and(cpumask, cpu_online_mask); 267 if (cpu >= NR_CPUS) 274 write_gic_map_vp(irq, BIT(mips_cm_vp_id(cpu))); 279 set_bit(irq, per_cpu_ptr(pcpu_masks, cpu)); 354 int intr, cpu; local 372 int intr, cpu; local 431 gic_shared_irq_domain_map(struct irq_domain *d, unsigned int virq, irq_hw_number_t hw, unsigned int cpu) argument 474 int err, cpu; local 594 int cpu, ret, i; local 719 gic_cpu_startup(unsigned int cpu) argument [all...] |
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | amd.c | 19 #include <linux/cpu.h> 134 enum smca_bank_types smca_get_bank_type(unsigned int cpu, unsigned int bank) argument 141 b = &per_cpu(smca_banks, cpu)[bank]; 244 static void smca_set_misc_banks_map(unsigned int bank, unsigned int cpu) argument 262 per_cpu(smca_misc_banks_map, cpu) |= BIT_ULL(bank); 266 static void smca_configure(unsigned int bank, unsigned int cpu) argument 306 smca_set_misc_banks_map(bank, cpu); 389 pr_err(FW_BUG "cpu %d, failed to setup threshold interrupt " 390 "for bank %d, block %d (MSR%08X=0x%x%08x)\n", b->cpu, 404 pr_err(FW_BUG "cpu 519 smca_get_block_address(unsigned int bank, unsigned int block, unsigned int cpu) argument 531 get_block_address(u32 current_addr, u32 low, u32 high, unsigned int bank, unsigned int block, unsigned int cpu) argument 563 unsigned int cpu = smp_processor_id(); local 674 unsigned int bank, block, cpu = smp_processor_id(); local 922 unsigned int bank, cpu = smp_processor_id(); local 1090 get_name(unsigned int cpu, unsigned int bank, struct threshold_block *b) argument 1120 allocate_threshold_blocks(unsigned int cpu, struct threshold_bank *tb, unsigned int bank, unsigned int block, u32 address) argument 1221 threshold_create_bank(struct threshold_bank **bp, unsigned int cpu, unsigned int bank) argument 1364 mce_threshold_remove_device(unsigned int cpu) argument 1392 mce_threshold_create_device(unsigned int cpu) argument [all...] |
/linux-master/tools/testing/selftests/rseq/ |
H A D | rseq-riscv-bits.h | 9 int RSEQ_TEMPLATE_IDENTIFIER(rseq_cmpeqv_storev)(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument 32 : [cpu_id] "r" (cpu), 63 off_t voffp, intptr_t *load, int cpu) 89 : [cpu_id] "r" (cpu), 119 int RSEQ_TEMPLATE_IDENTIFIER(rseq_addv)(intptr_t *v, intptr_t count, int cpu) argument 139 : [cpu_id] "r" (cpu), 165 intptr_t newv, int cpu) 192 : [cpu_id] "r" (cpu), 232 int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, off_t off, intptr_t inc, int cpu) argument 250 : [cpu_id] "r" (cpu), 62 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, off_t voffp, intptr_t *load, int cpu) argument 163 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument 281 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument 344 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument [all...] |
H A D | rseq-x86-bits.h | 16 int RSEQ_TEMPLATE_IDENTIFIER(rseq_cmpeqv_storev)(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument 45 : [cpu_id] "r" (cpu), 82 long voffp, intptr_t *load, int cpu) 116 : [cpu_id] "r" (cpu), 150 int RSEQ_TEMPLATE_IDENTIFIER(rseq_addv)(intptr_t *v, intptr_t count, int cpu) argument 172 : [cpu_id] "r" (cpu), 204 int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, long off, intptr_t inc, int cpu) argument 231 : [cpu_id] "r" (cpu), 257 intptr_t newv, int cpu) 292 : [cpu_id] "r" (cpu), 81 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, long voffp, intptr_t *load, int cpu) argument 255 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument 337 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument 407 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument 528 rseq_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t newv, int cpu) argument 593 rseq_cmpnev_storeoffp_load(intptr_t *v, intptr_t expectnot, long voffp, intptr_t *load, int cpu) argument 662 rseq_addv(intptr_t *v, intptr_t count, int cpu) argument 710 rseq_cmpeqv_cmpeqv_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t expect2, intptr_t newv, int cpu) argument 793 rseq_cmpeqv_trystorev_storev(intptr_t *v, intptr_t expect, intptr_t *v2, intptr_t newv2, intptr_t newv, int cpu) argument 870 rseq_cmpeqv_trymemcpy_storev(intptr_t *v, intptr_t expect, void *dst, void *src, size_t len, intptr_t newv, int cpu) argument [all...] |
/linux-master/lib/ |
H A D | percpu_counter.c | 9 #include <linux/cpu.h> 62 int cpu; local 66 for_each_possible_cpu(cpu) { 67 s32 *pcount = per_cpu_ptr(fbc->counters, cpu); 126 * Add up all the per-cpu counts, return the result. This is a more accurate 129 * We use the cpu mask of (cpu_online_mask | cpu_dying_mask) to capture sums 140 int cpu; local 145 for_each_cpu_or(cpu, cpu_online_mask, cpu_dying_mask) { 146 s32 *pcount = per_cpu_ptr(fbc->counters, cpu); 224 static int compute_batch_value(unsigned int cpu) argument 232 percpu_counter_cpu_dead(unsigned int cpu) argument 335 int cpu; local [all...] |
/linux-master/arch/loongarch/kvm/ |
H A D | main.c | 195 static void kvm_update_vpid(struct kvm_vcpu *vcpu, int cpu) argument 200 context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu); 219 int cpu; local 224 cpu = smp_processor_id(); 229 context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu); 230 migrated = (vcpu->cpu != cpu); 242 kvm_update_vpid(vcpu, cpu); 244 vcpu->cpu = cpu; 315 int cpu, order; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | gic_v3.c | 53 static inline volatile void *gicr_base_cpu(uint32_t cpu) argument 56 return GICR_BASE_GVA + cpu * SZ_64K * 2; 59 static void gicv3_gicr_wait_for_rwp(uint32_t cpu) argument 63 while (readl(gicr_base_cpu(cpu) + GICR_CTLR) & GICR_CTLR_RWP) { 172 uint32_t cpu = guest_get_vcpuid(); local 193 cpu_or_dist = (intid_range == SPI_RANGE) ? DIST_BIT : cpu; 236 uint32_t cpu = guest_get_vcpuid(); local 239 gicv3_wait_for_rwp(is_spi ? DIST_BIT : cpu); 245 uint32_t cpu = guest_get_vcpuid(); local 248 gicv3_wait_for_rwp(is_spi ? DIST_BIT : cpu); 296 gicv3_cpu_init(unsigned int cpu) argument [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grufile.c | 311 static unsigned long gru_chiplet_cpu_to_mmr(int chiplet, int cpu, int *corep) argument 322 core = uv_cpu_core_number(cpu) + UV_MAX_INT_CORES * uv_cpu_socket_number(cpu); 323 if (core >= GRU_NUM_TFM || uv_cpu_ht_number(cpu)) 341 irq_handler_t irq_handler, int cpu, int blade) 347 mmr = gru_chiplet_cpu_to_mmr(chiplet, cpu, &core); 351 irq = uv_setup_irq(irq_name, cpu, blade, mmr, UV_AFFINITY_CPU); 369 static void gru_chiplet_teardown_tlb_irq(int chiplet, int cpu, int blade) argument 374 mmr = gru_chiplet_cpu_to_mmr(chiplet, cpu, &core); 387 int cpu; local 340 gru_chiplet_setup_tlb_irq(int chiplet, char *irq_name, irq_handler_t irq_handler, int cpu, int blade) argument 405 int cpu; local [all...] |
/linux-master/kernel/printk/ |
H A D | nbcon.c | 23 * - The 'cpu' field denotes on which CPU the console is locked. It is used 229 unsigned int cpu = smp_processor_id(); local 253 new.cpu = cpu; 309 unsigned int cpu = smp_processor_id(); local 338 new.cpu = cpu; 392 unsigned int cpu = smp_processor_id(); local 406 if (cur->cpu == cpu) 498 unsigned int cpu = smp_processor_id(); local 544 unsigned int cpu = smp_processor_id(); local 606 unsigned int cpu = smp_processor_id(); local 659 unsigned int cpu = smp_processor_id(); local [all...] |
/linux-master/arch/parisc/kernel/ |
H A D | irq.c | 72 int cpu = smp_processor_id(); local 75 per_cpu(local_ack_eiem, cpu) &= ~mask; 78 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu)); 87 int cpu = smp_processor_id(); local 90 per_cpu(local_ack_eiem, cpu) |= mask; 93 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu)); 311 unsigned long txn_affinity_addr(unsigned int irq, int cpu) argument 315 irq_data_update_affinity(d, cpumask_of(cpu)); 318 return per_cpu(cpu_data, cpu).txn_addr; 385 int cpu local 493 int irq, cpu = smp_processor_id(); local 518 int cpu = cpumask_first(&dest); local [all...] |
/linux-master/kernel/rcu/ |
H A D | tasks.h | 36 * @cpu: CPU number corresponding to this entry. 51 int cpu; member in struct:rcu_tasks_percpu 245 int cpu; local 265 for_each_possible_cpu(cpu) { 266 struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu); 269 if (cpu) 274 rtpcp->cpu = cpu; 401 int cpu; local 415 for_each_possible_cpu(cpu) { 436 int cpu; local 514 int cpu; local 604 int cpu; local 696 int cpu; local 907 int cpu; local 947 int cpu; local 1012 int cpu; local 1612 int cpu = task_cpu(t); local 1660 int cpu; local 1742 int cpu; local 1828 int cpu; local 1859 int cpu; local 1907 int cpu; local [all...] |
/linux-master/include/linux/ |
H A D | workqueue.h | 119 int cpu; member in struct:delayed_work 371 WQ_UNBOUND = 1 << 1, /* not bound to any cpu */ 375 WQ_CPU_INTENSIVE = 1 << 5, /* cpu intensive workqueue */ 379 * Per-cpu workqueues are generally preferred because they tend to 380 * show better performance thanks to cache locality. Per-cpu 387 * however, for example, a per-cpu work item scheduled from an 393 * Workqueues marked with WQ_POWER_EFFICIENT are per-cpu by default 395 * specified. Per-cpu workqueues which are identified to 469 void workqueue_softirq_dead(unsigned int cpu); 478 * For a per-cpu workqueu 661 schedule_work_on(int cpu, struct work_struct *work) argument 759 schedule_delayed_work_on(int cpu, struct delayed_work *dwork, unsigned long delay) argument 780 work_on_cpu(int cpu, long (*fn)(void *), void *arg) argument 784 work_on_cpu_safe(int cpu, long (*fn)(void *), void *arg) argument 833 wq_watchdog_touch(int cpu) argument [all...] |
/linux-master/drivers/macintosh/ |
H A D | rack-meter.c | 67 struct rackmeter_cpu cpu[2]; member in struct:rackmeter 81 static inline u64 get_cpu_idle_time(unsigned int cpu) argument 83 struct kernel_cpustat *kcpustat = &kcpustat_cpu(cpu); 90 retval += kcpustat_field(kcpustat, CPUTIME_NICE, cpu); 217 unsigned int cpu = smp_processor_id(); local 226 total_idle_nsecs = get_cpu_idle_time(cpu); 236 offset = cpu << 3; 246 pause = (rm->cpu[0].zero && rm->cpu[1].zero); 249 pause = (rm->cpu[ 259 unsigned int cpu; local [all...] |
/linux-master/kernel/power/ |
H A D | energy_model.c | 12 #include <linux/cpu.h> 221 int i, cpu; local 227 cpu = cpumask_first(em_span_cpus(pd)); 235 max_cap = (u64) arch_scale_cpu_capacity(cpu); 405 int cpu, ret, num_cpus; local 440 for_each_cpu(cpu, cpus) { 441 cpu_dev = get_cpu_device(cpu); 462 int i, cpu; local 468 cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask); 469 if (cpu > 523 em_cpu_get(int cpu) argument 564 int cpu, ret; local 751 int cpu; local [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | setup-common.c | 26 #include <linux/cpu.h> 114 /* This keeps a track of which one is the crashing cpu. */ 267 * rate of a given cpu. The rest can use ppc_proc_freq to 324 /* If this is the last cpu, print the summary */ 333 if (*pos == 0) /* just in case, cpu 0 is not the first */ 415 static int assign_threads(unsigned int cpu, unsigned int nthreads, bool present, argument 418 for (int i = 0; i < nthreads && cpu < nr_cpu_ids; i++) { 423 DBG(" thread %d -> cpu %d (hard id %d)\n", i, cpu, hwid); 425 set_cpu_present(cpu, presen 455 int cpu = 0; local 889 int cpu; local [all...] |
/linux-master/drivers/virt/nitro_enclaves/ |
H A D | ne_misc_dev.c | 13 #include <linux/cpu.h> 36 * NE_CPUS_SIZE - Size for max 128 CPUs, for now, in a cpu-list string, comma 101 /* https://www.kernel.org/doc/html/latest/admin-guide/kernel-parameters.html#cpu-lists */ 102 MODULE_PARM_DESC(ne_cpus, "<cpu-list> - CPU pool used for Nitro Enclaves"); 182 unsigned int cpu = 0; local 201 cpu = cpumask_any(cpu_pool); 202 if (cpu >= nr_cpu_ids) { 214 for_each_cpu(cpu, cpu_pool) 215 if (cpu_is_offline(cpu)) { 217 ne_misc_dev.name, cpu); 388 unsigned int cpu = 0; local 480 ne_donated_cpu(struct ne_enclave *ne_enclave, unsigned int cpu) argument 530 unsigned int cpu = 0; local 576 unsigned int cpu = 0; local 1076 unsigned int cpu = 0; local 1419 unsigned int cpu = 0; local [all...] |
/linux-master/kernel/ |
H A D | torture.c | 32 #include <linux/cpu.h> 201 bool torture_offline(int cpu, long *n_offl_attempts, long *n_offl_successes, argument 209 if (!cpu_online(cpu) || !cpu_is_hotpluggable(cpu)) 217 torture_type, cpu); 220 ret = remove_cpu(cpu); 231 torture_type, cpu, s, ret); 236 torture_type, cpu); 263 bool torture_online(int cpu, long *n_onl_attempts, long *n_onl_successes, argument 271 if (cpu_online(cpu) || !cpu_is_hotpluggabl 320 int cpu; local 342 int cpu; local [all...] |
/linux-master/arch/x86/kernel/cpu/microcode/ |
H A D | amd.c | 35 #include <asm/cpu.h> 432 * patch container file in initrd, traverse equivalent cpu table, look for a 436 * load_microcode_amd() to save equivalent cpu table and microcode patches in 593 static struct ucode_patch *find_patch(unsigned int cpu) argument 595 struct ucode_cpu_info *uci = ucode_cpu_info + cpu; 605 void reload_ucode_amd(unsigned int cpu) argument 611 p = find_patch(cpu); 625 static int collect_cpu_info_amd(int cpu, struct cpu_signature *csig) argument 627 struct cpuinfo_x86 *c = &cpu_data(cpu); 628 struct ucode_cpu_info *uci = ucode_cpu_info + cpu; 645 apply_microcode_amd(int cpu) argument 695 unsigned int cpu = smp_processor_id(); local 824 unsigned int nid, cpu; local 870 request_microcode_amd(int cpu, struct device *device) argument 901 microcode_fini_cpu_amd(int cpu) argument [all...] |
/linux-master/tools/tracing/rtla/src/ |
H A D | timerlat_aa.c | 25 * Per-cpu data statistics and data. 137 * timerlat_aa_get_data - Get the per-cpu data from the timerlat context 140 *timerlat_aa_get_data(struct timerlat_aa_context *taa_ctx, int cpu) argument 142 return &taa_ctx->taa_data[cpu]; 245 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 268 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 311 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 407 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 440 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 485 struct timerlat_aa_data *taa_data = timerlat_aa_get_data(taa_ctx, record->cpu); 560 timerlat_thread_analysis(struct timerlat_aa_data *taa_data, int cpu, int irq_thresh, int thread_thresh) argument 735 int cpu; local [all...] |
/linux-master/net/core/ |
H A D | gen_stats.c | 127 struct gnet_stats_basic_sync __percpu *cpu) 133 struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i); 150 struct gnet_stats_basic_sync __percpu *cpu, 157 WARN_ON_ONCE((cpu || running) && in_hardirq()); 159 if (cpu) { 160 gnet_stats_add_basic_cpu(bstats, cpu); 175 struct gnet_stats_basic_sync __percpu *cpu, 180 if (cpu) { 185 struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i); 212 struct gnet_stats_basic_sync __percpu *cpu, 126 gnet_stats_add_basic_cpu(struct gnet_stats_basic_sync *bstats, struct gnet_stats_basic_sync __percpu *cpu) argument 149 gnet_stats_add_basic(struct gnet_stats_basic_sync *bstats, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument 174 gnet_stats_read_basic(u64 *ret_bytes, u64 *ret_packets, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument 211 ___gnet_stats_copy_basic(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, int type, bool running) argument 260 gnet_stats_copy_basic(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument 287 gnet_stats_copy_basic_hw(struct gnet_dump *d, struct gnet_stats_basic_sync __percpu *cpu, struct gnet_stats_basic_sync *b, bool running) argument 356 gnet_stats_add_queue(struct gnet_stats_queue *qstats, const struct gnet_stats_queue __percpu *cpu, const struct gnet_stats_queue *q) argument [all...] |
/linux-master/arch/x86/kernel/ |
H A D | irq.c | 5 #include <linux/cpu.h> 200 u64 arch_irq_stat_cpu(unsigned int cpu) argument 202 u64 sum = irq_stats(cpu)->__nmi_count; 205 sum += irq_stats(cpu)->apic_timer_irqs; 206 sum += irq_stats(cpu)->irq_spurious_count; 207 sum += irq_stats(cpu)->apic_perf_irqs; 208 sum += irq_stats(cpu)->apic_irq_work_irqs; 209 sum += irq_stats(cpu)->icr_read_retry_count; 211 sum += irq_stats(cpu)->x86_platform_ipis; 214 sum += irq_stats(cpu) [all...] |
/linux-master/drivers/thermal/ |
H A D | cpufreq_cooling.c | 13 #include <linux/cpu.h> 44 * @time: previous reading of the absolute time that this cpu was idle 153 * get_load() - get load for a cpu 154 * @cpufreq_cdev: struct cpufreq_cooling_device for the cpu 155 * @cpu: cpu number 156 * @cpu_idx: index of the cpu in time_in_idle array 158 * Return: The average load of cpu @cpu in percentage since this 162 static u32 get_load(struct cpufreq_cooling_device *cpufreq_cdev, int cpu, argument 170 get_load(struct cpufreq_cooling_device *cpufreq_cdev, int cpu, int cpu_idx) argument 236 int i = 0, cpu; local [all...] |