/linux-master/arch/arm/mm/ |
H A D | cache-l2x0-pmu.c | 120 u64 prev_count, new_count, mask; local 123 prev_count = local64_read(&hw->prev_count); 125 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); 128 local64_add((new_count - prev_count) & mask, &event->count); 147 local64_set(&hw->prev_count, 0);
|
/linux-master/arch/alpha/kernel/ |
H A D | perf_event.c | 283 local64_set(&hwc->prev_count, (unsigned long)(-left)); 314 prev_raw_count = local64_read(&hwc->prev_count); 317 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
|
/linux-master/arch/arc/kernel/ |
H A D | perf_event.c | 283 u64 prev_raw_count = local64_read(&hwc->prev_count); 288 * We aren't afraid of hwc->prev_count changing beneath our feet 291 local64_set(&hwc->prev_count, new_raw_count); 432 local64_set(&hwc->prev_count, value); 550 local64_set(&hwc->prev_count, 0);
|
/linux-master/arch/arm/mach-imx/ |
H A D | mmdc.c | 314 prev_raw_count = local64_read(&hwc->prev_count); 317 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, 342 local64_set(&hwc->prev_count, 0); 377 local64_set(&hwc->prev_count, mmdc_pmu_read_counter(pmu_mmdc, cfg));
|
/linux-master/arch/csky/kernel/ |
H A D | perf_event.c | 906 local64_set(&hwc->prev_count, (u64)(-left)); 922 uint64_t prev_raw_count = local64_read(&hwc->prev_count); 932 * We aren't afraid of hwc->prev_count changing beneath our feet 935 local64_set(&hwc->prev_count, new_raw_count);
|
/linux-master/arch/loongarch/kernel/ |
H A D | perf_event.c | 326 local64_set(&hwc->prev_count, loongarch_pmu.overflow - left); 343 prev_raw_count = local64_read(&hwc->prev_count); 346 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
|
/linux-master/arch/mips/kernel/ |
H A D | perf_event_mipsxx.c | 436 local64_set(&hwc->prev_count, mipspmu.overflow - left); 457 prev_raw_count = local64_read(&hwc->prev_count); 460 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
|
/linux-master/arch/powerpc/perf/ |
H A D | 8xx-pmu.c | 115 local64_set(&event->hw.prev_count, val); 128 prev = local64_read(&event->hw.prev_count); 149 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);
|
H A D | core-book3s.c | 1169 local64_set(&event->hw.prev_count, val); 1179 prev = local64_read(&event->hw.prev_count); 1185 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); 1229 prev = local64_read(&event->hw.prev_count); 1248 prev = local64_read(&event->hw.prev_count); 1250 local64_set(&event->hw.prev_count, val); 1548 val = local64_read(&event->hw.prev_count); 1556 local64_set(&event->hw.prev_count, val); 2172 * of the sample_period logic. We use hw.prev_count for this. 2175 local64_set(&event->hw.prev_count, [all...] |
H A D | core-fsl-emb.c | 175 prev = local64_read(&event->hw.prev_count); 178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); 314 local64_set(&event->hw.prev_count, val); 602 prev = local64_read(&event->hw.prev_count); 625 local64_set(&event->hw.prev_count, val);
|
H A D | hv-24x7.c | 1446 (void)local64_xchg(&event->hw.prev_count, ct); 1466 prev = local64_xchg(&event->hw.prev_count, now); 1520 local64_set(&event->hw.prev_count, h_24x7_get_value(event));
|
H A D | hv-gpci.c | 750 prev = local64_xchg(&event->hw.prev_count, now); 756 local64_set(&event->hw.prev_count, h_gpci_get_value(event));
|
H A D | imc-pmu.c | 1078 local64_set(&event->hw.prev_count, data); 1087 counter_prev = local64_read(&event->hw.prev_count);
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | papr_scm.c | 453 local64_set(&event->hw.prev_count, count); 472 prev = local64_xchg(&event->hw.prev_count, now);
|
/linux-master/arch/s390/kernel/ |
H A D | perf_cpum_cf.c | 873 prev = local64_read(&event->hw.prev_count); 885 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); 896 prev = local64_read(&event->hw.prev_count); 900 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); 931 * Because all counters in a set are active, the event->hw.prev_count
|
H A D | perf_pai_crypto.c | 316 prev = local64_read(&event->hw.prev_count); 318 local64_set(&event->hw.prev_count, new); 332 local64_set(&event->hw.prev_count, sum);
|
H A D | perf_pai_ext.c | 320 prev = local64_read(&event->hw.prev_count); 322 local64_set(&event->hw.prev_count, new); 335 local64_set(&event->hw.prev_count, sum);
|
/linux-master/arch/sh/kernel/ |
H A D | perf_event.c | 178 prev_raw_count = local64_read(&hwc->prev_count); 181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
|
/linux-master/arch/sparc/kernel/ |
H A D | perf_event.c | 871 prev_raw_count = local64_read(&hwc->prev_count); 874 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, 914 local64_set(&hwc->prev_count, (u64)-left);
|
/linux-master/arch/x86/events/amd/ |
H A D | ibs.c | 158 prev_raw_count = local64_read(&hwc->prev_count); 159 if (!local64_try_cmpxchg(&hwc->prev_count, 341 local64_set(&hwc->prev_count, 0);
|
H A D | uncore.c | 99 * we do not have to worry about prev_count changing on us 102 prev = local64_read(&hwc->prev_count); 113 local64_set(&hwc->prev_count, new); 124 wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count)); 876 wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count));
|
/linux-master/arch/x86/events/ |
H A D | core.c | 132 prev_raw_count = local64_read(&hwc->prev_count); 135 } while (!local64_try_cmpxchg(&hwc->prev_count, 1408 local64_set(&hwc->prev_count, (u64)-left);
|
/linux-master/arch/x86/events/intel/ |
H A D | core.c | 2852 local64_set(&event->hw.prev_count, 0);
|
H A D | cstate.c | 388 prev_raw_count = local64_read(&hwc->prev_count); 391 } while (!local64_try_cmpxchg(&hwc->prev_count, 399 local64_set(&event->hw.prev_count, cstate_pmu_read_counter(event));
|
H A D | ds.c | 1993 prev_raw_count = local64_read(&hwc->prev_count); 1995 local64_set(&hwc->prev_count, new_raw_count);
|