Searched refs:pmus (Results 1 - 11 of 11) sorted by relevance

/linux-master/tools/perf/arch/arm/util/
H A Dauxtrace.c17 #include "../../../util/pmus.h"
102 static struct perf_pmu *find_pmu_for_event(struct perf_pmu **pmus, argument
107 if (!pmus)
111 if (evsel->core.attr.type == pmus[i]->type)
112 return pmus[i];
/linux-master/arch/x86/events/amd/
H A Duncore.c75 struct amd_uncore_pmu *pmus; member in struct:amd_uncore
421 pmu = &uncore->pmus[i];
451 pmu = &uncore->pmus[i];
515 pmu = &uncore->pmus[i];
663 uncore->pmus = kzalloc(sizeof(*uncore->pmus), GFP_KERNEL);
664 if (!uncore->pmus) {
674 pmu = &uncore->pmus[0];
794 uncore->pmus = kzalloc(sizeof(*uncore->pmus), GFP_KERNE
[all...]
/linux-master/arch/x86/events/intel/
H A Duncore.c958 struct intel_uncore_pmu *pmu = type->pmus;
969 kfree(type->pmus);
970 type->pmus = NULL;
988 struct intel_uncore_pmu *pmus; local
992 pmus = kcalloc(type->num_boxes, sizeof(*pmus), GFP_KERNEL);
993 if (!pmus)
999 pmus[i].func_id = setid ? i : -1;
1000 pmus[i].pmu_idx = i;
1001 pmus[
[all...]
H A Duncore.h82 struct intel_uncore_pmu *pmus; member in struct:intel_uncore_type
H A Dcore.c6057 static __always_inline int intel_pmu_init_hybrid(enum hybrid_pmu_type pmus) argument
6059 unsigned long pmus_mask = pmus;
/linux-master/arch/x86/events/
H A Drapl.c118 struct rapl_pmu *pmus[] __counted_by(maxdie);
150 return dieid < rapl_pmus->maxdie ? rapl_pmus->pmus[dieid] : NULL;
581 rapl_pmus->pmus[topology_logical_die_id(cpu)] = pmu;
662 kfree(rapl_pmus->pmus[i]);
/linux-master/tools/perf/util/
H A Dpmu.h147 /** @list: Element on pmus list in pmu.c. */
272 struct perf_pmu *perf_pmu__lookup(struct list_head *pmus, int dirfd, const char *lookup_name);
H A Dpmu.c21 #include "pmus.h"
996 struct perf_pmu *perf_pmu__lookup(struct list_head *pmus, int dirfd, const char *name) argument
1039 list_add_tail(&pmu->list, pmus);
/linux-master/tools/perf/pmu-events/
H A Djevents.py505 pmus = set()
515 pmus.add((event.pmu, pmu_name))
525 for (pmu, tbl_pmu) in sorted(pmus):
560 pmus = set()
570 pmus.add((metric.pmu, pmu_name))
580 for (pmu, tbl_pmu) in sorted(pmus):
671 const struct pmu_table_entry *pmus;
677 const struct pmu_table_entry *pmus;
708 \t\t.pmus = pmu_events__test_soc_cpu,
712 \t\t.pmus
[all...]
/linux-master/drivers/perf/
H A Dxgene_pmu.c1931 xgene_pmu_dev_cleanup(struct xgene_pmu *xgene_pmu, struct list_head *pmus) argument
1935 list_for_each_entry(ctx, pmus, next) {
/linux-master/kernel/events/
H A Dcore.c406 static LIST_HEAD(pmus);
5102 * The following implement mutual exclusion of events on "exclusive" pmus
5103 * (PERF_PMU_CAP_EXCLUSIVE). Such pmus can only have one event scheduled
11589 list_add_rcu(&pmu->entry, &pmus);
11702 * pmus overwrites event->attr.type to forward event to another pmu.
11750 list_for_each_entry_rcu(pmu, &pmus, entry, lockdep_is_held(&pmus_srcu)) {
12660 /* Don't allow group of multiple hw events from different pmus */
13802 list_for_each_entry(pmu, &pmus, entry) {

Completed in 263 milliseconds