/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | vmx_msrs_test.c | 15 static void vmx_fixed1_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index, argument 18 uint64_t val = vcpu_get_msr(vcpu, msr_index); 24 vcpu_set_msr(vcpu, msr_index, val & ~BIT_ULL(bit)); 25 vcpu_set_msr(vcpu, msr_index, val); 29 static void vmx_fixed0_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index, argument 32 uint64_t val = vcpu_get_msr(vcpu, msr_index); 38 vcpu_set_msr(vcpu, msr_index, val | BIT_ULL(bit)); 39 vcpu_set_msr(vcpu, msr_index, val); 43 static void vmx_fixed0and1_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index) argument 45 vmx_fixed0_msr_test(vcpu, msr_index, GENMASK_UL [all...] |
H A D | userspace_msr_exit_test.c | 392 static void process_rdmsr(struct kvm_vcpu *vcpu, uint32_t msr_index) argument 399 TEST_ASSERT(run->msr.index == msr_index, 401 run->msr.index, msr_index); 424 static void process_wrmsr(struct kvm_vcpu *vcpu, uint32_t msr_index) argument 431 TEST_ASSERT(run->msr.index == msr_index, 433 run->msr.index, msr_index); 490 uint32_t msr_index) 493 process_rdmsr(vcpu, msr_index); 497 uint32_t msr_index) 500 process_wrmsr(vcpu, msr_index); 489 run_guest_then_process_rdmsr(struct kvm_vcpu *vcpu, uint32_t msr_index) argument 496 run_guest_then_process_wrmsr(struct kvm_vcpu *vcpu, uint32_t msr_index) argument [all...] |
/linux-master/arch/x86/xen/ |
H A D | pmu.c | 147 static bool is_intel_pmu_msr(u32 msr_index, int *type, int *index) argument 156 switch (msr_index) { 171 if ((msr_index >= MSR_CORE_PERF_FIXED_CTR0) && 172 (msr_index < MSR_CORE_PERF_FIXED_CTR0 + 174 *index = msr_index - MSR_CORE_PERF_FIXED_CTR0; 179 if ((msr_index >= MSR_P6_EVNTSEL0) && 180 (msr_index < MSR_P6_EVNTSEL0 + intel_num_arch_counters)) { 181 *index = msr_index - MSR_P6_EVNTSEL0; 186 msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK;
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | hyperv.h | 59 void nested_evmcs_filter_control_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata);
|
H A D | hyperv.c | 111 void nested_evmcs_filter_control_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata) argument 121 switch (msr_index) {
|
H A D | nested.h | 32 int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); 33 int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata);
|
H A D | nested.c | 975 u32 msr_index, 985 if (msr_index == MSR_IA32_TSC) { 997 if (kvm_get_msr(vcpu, msr_index, data)) { 999 msr_index); 1055 static bool nested_msr_store_list_has_msr(struct kvm_vcpu *vcpu, u32 msr_index) argument 1067 if (e.index == msr_index) 1074 u32 msr_index) 1083 msr_autostore_slot = vmx_find_loadstore_msr_slot(autostore, msr_index); 1085 in_vmcs12_store_list = nested_msr_store_list_has_msr(vcpu, msr_index); 1098 msr_index); 974 nested_vmx_get_vmexit_msr_value(struct kvm_vcpu *vcpu, u32 msr_index, u64 *data) argument 1073 prepare_vmx_msr_autostore_list(struct kvm_vcpu *vcpu, u32 msr_index) argument 1259 vmx_get_control_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u32 **low, u32 **high) argument 1289 vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) argument 1361 vmx_get_fixed0_msr(struct nested_vmx_msrs *msrs, u32 msr_index) argument 1373 vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) argument 1393 vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data) argument 1458 vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata) argument 6024 u32 msr_index = kvm_rcx_read(vcpu); local [all...] |
H A D | vmx.c | 2161 u32 msr_index = msr_info->index; local 2165 switch (msr_index) { 2224 kvm_pr_unimpl_wrmsr(vcpu, msr_index, data); 2352 [msr_index - MSR_IA32_SGXLEPUBKEYHASH0] = data; 2359 return vmx_set_vmx_msr(vcpu, msr_index, data); 2445 msr = vmx_find_uret_msr(vmx, msr_index); 2453 if (msr_index == MSR_IA32_ARCH_CAPABILITIES)
|
/linux-master/drivers/thermal/intel/ |
H A D | intel_powerclamp.c | 316 int msr_index; member in struct:pkg_cstate_info 321 .msr_index = MSR_PKG_C##id##_RESIDENCY, \ 342 while (info->msr_index) { 343 if (!rdmsrl_safe(info->msr_index, &val)) 357 while (info->msr_index) { 359 if (!rdmsrl_safe(info->msr_index, &val))
|
/linux-master/arch/x86/kvm/ |
H A D | kvm_emulate.h | 208 int (*set_msr_with_filter)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data); 209 int (*get_msr_with_filter)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata); 210 int (*get_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata);
|
H A D | x86.c | 3773 static bool kvm_is_msr_to_save(u32 msr_index) argument 3778 if (msrs_to_save[i] == msr_index) 7338 static void kvm_probe_feature_msr(u32 msr_index) argument 7341 .index = msr_index, 7347 msr_based_features[num_msr_based_features++] = msr_index; 7350 static void kvm_probe_msr_to_save(u32 msr_index) argument 7354 if (rdmsr_safe(msr_index, &dummy[0], &dummy[1])) 7361 switch (msr_index) { 7394 (msr_index - MSR_IA32_RTIT_ADDR0_A >= 7399 if (msr_index 8410 emulator_get_msr_with_filter(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata) argument 8433 emulator_set_msr_with_filter(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data) argument 8456 emulator_get_msr(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata) argument [all...] |
H A D | emulate.c | 3345 u64 msr_index = reg_read(ctxt, VCPU_REGS_RCX); local 3351 r = ctxt->ops->set_msr_with_filter(ctxt, msr_index, msr_data); 3361 u64 msr_index = reg_read(ctxt, VCPU_REGS_RCX); local 3365 r = ctxt->ops->get_msr_with_filter(ctxt, msr_index, &msr_data);
|
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | processor.c | 710 uint64_t kvm_get_feature_msr(uint64_t msr_index) argument 719 buffer.entry.index = msr_index; 830 uint64_t vcpu_get_msr(struct kvm_vcpu *vcpu, uint64_t msr_index) argument 838 buffer.entry.index = msr_index; 845 int _vcpu_set_msr(struct kvm_vcpu *vcpu, uint64_t msr_index, uint64_t msr_value) argument 854 buffer.entry.index = msr_index; 963 bool kvm_msr_is_in_save_restore_list(uint32_t msr_index) argument 969 if (list->indices[i] == msr_index)
|
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | processor.h | 839 bool kvm_msr_is_in_save_restore_list(uint32_t msr_index); 840 uint64_t kvm_get_feature_msr(uint64_t msr_index); 1070 uint64_t vcpu_get_msr(struct kvm_vcpu *vcpu, uint64_t msr_index); 1071 int _vcpu_set_msr(struct kvm_vcpu *vcpu, uint64_t msr_index, uint64_t msr_value);
|