/linux-master/Documentation/trace/postprocess/ |
H A D | decode_msr.py | 7 msrs = dict() variable 13 msrs[int(m.group(2), 16)] = m.group(1) 26 if num in msrs: 27 r = msrs[num]
|
/linux-master/arch/x86/lib/ |
H A D | msr.c | 11 struct msr __percpu *msrs = NULL; local 13 msrs = alloc_percpu(struct msr); 14 if (!msrs) { 15 pr_warn("%s: error allocating msrs\n", __func__); 19 return msrs; 23 void msrs_free(struct msr __percpu *msrs) argument 25 free_percpu(msrs);
|
H A D | msr-smp.c | 13 if (rv->msrs) 14 reg = this_cpu_ptr(rv->msrs); 26 if (rv->msrs) 27 reg = this_cpu_ptr(rv->msrs); 98 struct msr __percpu *msrs, 106 rv.msrs = msrs; 122 * @msrs: array of MSR values 125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) argument 127 __rwmsr_on_cpus(mask, msr_no, msrs, __rdmsr_on_cp 97 __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs, void (*msr_func) (void *info)) argument 139 wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) argument [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | nested.h | 33 int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata); 101 return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low); 111 return to_vmx(vcpu)->nested.msrs.misc_low & 117 return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS; 122 return to_vmx(vcpu)->nested.msrs.procbased_ctls_high & 128 return to_vmx(vcpu)->nested.msrs.secondary_ctls_high & 259 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; 260 u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1; 263 if (to_vmx(vcpu)->nested.msrs.secondary_ctls_high & 273 u64 fixed0 = to_vmx(vcpu)->nested.msrs [all...] |
H A D | nested.c | 443 bool execonly = vmx->nested.msrs.ept_caps & VMX_EPT_EXECUTE_ONLY_BIT; 444 int ept_lpage_level = ept_caps_to_lpage_level(vmx->nested.msrs.ept_caps); 923 u64 vmx_misc = vmx_control_msr(vmx->nested.msrs.misc_low, 924 vmx->nested.msrs.misc_high); 1255 vmx->nested.msrs.basic = data; 1259 static void vmx_get_control_msr(struct nested_vmx_msrs *msrs, u32 msr_index, argument 1264 *low = &msrs->pinbased_ctls_low; 1265 *high = &msrs->pinbased_ctls_high; 1268 *low = &msrs->procbased_ctls_low; 1269 *high = &msrs 1361 vmx_get_fixed0_msr(struct nested_vmx_msrs *msrs, u32 msr_index) argument 1458 vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata) argument 6806 nested_vmx_setup_pinbased_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6823 nested_vmx_setup_exit_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6846 nested_vmx_setup_entry_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6866 nested_vmx_setup_cpubased_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6902 nested_vmx_setup_secondary_ctls(u32 ept_caps, struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6985 nested_vmx_setup_misc_data(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument 6997 nested_vmx_setup_basic(struct nested_vmx_msrs *msrs) argument 7015 nested_vmx_setup_cr_fixed(struct nested_vmx_msrs *msrs) argument 7047 struct nested_vmx_msrs *msrs = &vmcs_conf->nested; local [all...] |
H A D | vmx.c | 2051 if (vmx_get_vmx_msr(&vmx->nested.msrs, msr_info->index, 2254 ((vmx->nested.msrs.entry_ctls_high & VM_ENTRY_LOAD_BNDCFGS) || 2255 (vmx->nested.msrs.exit_ctls_high & VM_EXIT_CLEAR_BNDCFGS))) 4550 vmx->nested.msrs.secondary_ctls_high |= control; 4552 vmx->nested.msrs.secondary_ctls_high &= ~control; 4827 memcpy(&vmx->nested.msrs, &vmcs_config.nested, sizeof(vmx->nested.msrs)); 7154 struct perf_guest_switch_msr *msrs; local 7162 msrs = perf_guest_get_msrs(&nr_msrs, (void *)pmu); 7163 if (!msrs) [all...] |
H A D | vmx.h | 237 struct nested_vmx_msrs msrs; member in struct:nested_vmx
|
/linux-master/arch/x86/include/asm/ |
H A D | msr.h | 20 struct msr __percpu *msrs; member in struct:msr_info 329 void msrs_free(struct msr __percpu *msrs); 338 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs); 339 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs); 368 struct msr __percpu *msrs) 370 rdmsr_on_cpu(0, msr_no, raw_cpu_ptr(&msrs->l), raw_cpu_ptr(&msrs->h)); 373 struct msr __percpu *msrs) 375 wrmsr_on_cpu(0, msr_no, raw_cpu_read(msrs->l), raw_cpu_read(msrs 367 rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, struct msr __percpu *msrs) argument 372 wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, struct msr __percpu *msrs) argument [all...] |
/linux-master/tools/testing/selftests/kvm/include/x86_64/ |
H A D | processor.h | 417 struct kvm_msrs msrs; member in struct:kvm_x86_state 843 struct kvm_msrs *msrs) 845 int r = __vcpu_ioctl(vcpu, KVM_GET_MSRS, msrs); 847 TEST_ASSERT(r == msrs->nmsrs, 849 r, r < 0 || r >= msrs->nmsrs ? -1 : msrs->entries[r].index); 851 static inline void vcpu_msrs_set(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs) argument 853 int r = __vcpu_ioctl(vcpu, KVM_SET_MSRS, msrs); 855 TEST_ASSERT(r == msrs->nmsrs, 857 r, r < 0 || r >= msrs 842 vcpu_msrs_get(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs) argument [all...] |
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | amd.c | 637 u32 msrs[NR_BLOCKS]; local 640 msrs[0] = 0x00000413; /* MC4_MISC0 */ 641 msrs[1] = 0xc0000408; /* MC4_MISC1 */ 649 msrs[0] = MSR_AMD64_SMCA_MCx_MISC(bank); 664 msr_clear_bit(msrs[i], 62);
|
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | processor.c | 1013 state = malloc(sizeof(*state) + msr_list->nmsrs * sizeof(state->msrs.entries[0])); 1037 state->msrs.nmsrs = msr_list->nmsrs; 1039 state->msrs.entries[i].index = msr_list->indices[i]; 1040 vcpu_msrs_get(vcpu, &state->msrs); 1050 vcpu_msrs_set(vcpu, &state->msrs);
|
/linux-master/drivers/edac/ |
H A D | amd64_edac.c | 15 static struct msr __percpu *msrs; variable in typeref:struct:__percpu 3199 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs); 3202 struct msr *reg = per_cpu_ptr(msrs, cpu); 3231 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); 3235 struct msr *reg = per_cpu_ptr(msrs, cpu); 3250 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); 4154 msrs = msrs_alloc(); 4155 if (!msrs) 4191 msrs_free(msrs); 4192 msrs [all...] |
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 365 struct kvm_user_return_msrs *msrs local 375 if (msrs->registered) { 376 msrs->registered = false; 381 values = &msrs->values[slot]; 431 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local 437 msrs->values[i].host = value; 438 msrs->values[i].curr = value; 445 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local 448 value = (value & mask) | (msrs->values[slot].host & ~mask); 449 if (value == msrs 468 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local 4546 __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs, struct kvm_msr_entry *entries, int (*do_msr)(struct kvm_vcpu *vcpu, unsigned index, u64 *data)) argument 4570 struct kvm_msrs msrs; local [all...] |