Searched refs:kvm_pmc (Results 1 - 9 of 9) sorted by relevance
/linux-master/arch/x86/kvm/ |
H A D | pmu.h | 31 struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu, 33 struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr); 79 static inline struct kvm_pmc *kvm_pmc_idx_to_pmc(struct kvm_pmu *pmu, int idx) 97 static inline u64 pmc_bitmask(struct kvm_pmc *pmc) 104 static inline u64 pmc_read_counter(struct kvm_pmc *pmc) 117 void pmc_write_counter(struct kvm_pmc *pmc, u64 val); 119 static inline bool pmc_is_gp(struct kvm_pmc *pmc) 124 static inline bool pmc_is_fixed(struct kvm_pmc *pmc) 139 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, 153 static inline struct kvm_pmc *get_fixed_pm [all...] |
H A D | pmu.c | 51 * - Each perf counter is defined as "struct kvm_pmc"; 69 * code. Each pmc, stored in kvm_pmc.idx field, is unique across 99 static inline void __kvm_perf_overflow(struct kvm_pmc *pmc, bool in_pmi) 130 struct kvm_pmc *pmc = perf_event->overflow_handler_context; 145 static u64 pmc_get_pebs_precise_level(struct kvm_pmc *pmc) 167 static u64 get_sample_period(struct kvm_pmc *pmc, u64 counter_value) 176 static int pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type, u64 config, 230 static bool pmc_pause_counter(struct kvm_pmc *pmc) 256 static bool pmc_resume_counter(struct kvm_pmc *pmc) 278 static void pmc_release_perf_event(struct kvm_pmc *pm [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | pmu.c | 28 static struct kvm_pmc *amd_pmu_get_pmc(struct kvm_pmu *pmu, int pmc_idx) 38 static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr, 87 static struct kvm_pmc *amd_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu, 93 static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr) 96 struct kvm_pmc *pmc; 130 struct kvm_pmc *pmc; 152 struct kvm_pmc *pmc;
|
/linux-master/arch/arm64/kvm/ |
H A D | pmu-emul.c | 26 static void kvm_pmu_create_perf_event(struct kvm_pmc *pmc); 27 static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc); 29 static struct kvm_vcpu *kvm_pmc_to_vcpu(const struct kvm_pmc *pmc) 34 static struct kvm_pmc *kvm_vcpu_idx_to_pmc(struct kvm_vcpu *vcpu, int cnt_idx) 83 static bool kvm_pmc_is_64bit(struct kvm_pmc *pmc) 91 static bool kvm_pmc_has_64bit_overflow(struct kvm_pmc *pmc) 99 static bool kvm_pmu_counter_can_chain(struct kvm_pmc *pmc) 115 static u64 kvm_pmu_get_pmc_value(struct kvm_pmc *pmc) 150 static void kvm_pmu_set_pmc_value(struct kvm_pmc *pmc, u64 val, bool force) 195 static void kvm_pmu_release_perf_event(struct kvm_pmc *pm [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | kvm_vcpu_pmu.h | 30 struct kvm_pmc { struct 44 struct kvm_pmc pmc[RISCV_KVM_MAX_COUNTERS];
|
/linux-master/arch/riscv/kvm/ |
H A D | vcpu_pmu.c | 37 static u64 kvm_pmu_get_sample_period(struct kvm_pmc *pmc) 78 static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc) 204 struct kvm_pmc *pmc; 234 struct kvm_pmc *pmc; 272 struct kvm_pmc *pmc = perf_event->overflow_handler_context; 309 static long kvm_pmu_create_perf_event(struct kvm_pmc *pmc, struct perf_event_attr *attr, 496 struct kvm_pmc *pmc; 573 struct kvm_pmc *pmc; 674 struct kvm_pmc *pmc = NULL; 771 struct kvm_pmc *pm [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 39 struct kvm_pmc *pmc; 58 static struct kvm_pmc *intel_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu, 63 struct kvm_pmc *counters; 124 static inline struct kvm_pmc *get_fw_gp_pmc(struct kvm_pmu *pmu, u32 msr) 181 static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr) 184 struct kvm_pmc *pmc; 301 struct kvm_pmc *pmc; 344 struct kvm_pmc *pmc; 704 struct kvm_pmc *pmc = NULL;
|
/linux-master/include/kvm/ |
H A D | arm_pmu.h | 16 struct kvm_pmc { struct 29 struct kvm_pmc pmc[ARMV8_PMU_MAX_COUNTERS];
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 502 struct kvm_pmc { struct 557 struct kvm_pmc gp_counters[KVM_INTEL_PMC_MAX_GENERIC]; 558 struct kvm_pmc fixed_counters[KVM_PMC_MAX_FIXED];
|
Completed in 143 milliseconds