Searched refs:msrs (Results 1 - 13 of 13) sorted by relevance

/linux-master/Documentation/trace/postprocess/
H A Ddecode_msr.py7 msrs = dict() variable
13 msrs[int(m.group(2), 16)] = m.group(1)
26 if num in msrs:
27 r = msrs[num]
/linux-master/arch/x86/lib/
H A Dmsr.c11 struct msr __percpu *msrs = NULL; local
13 msrs = alloc_percpu(struct msr);
14 if (!msrs) {
15 pr_warn("%s: error allocating msrs\n", __func__);
19 return msrs;
23 void msrs_free(struct msr __percpu *msrs) argument
25 free_percpu(msrs);
H A Dmsr-smp.c13 if (rv->msrs)
14 reg = this_cpu_ptr(rv->msrs);
26 if (rv->msrs)
27 reg = this_cpu_ptr(rv->msrs);
98 struct msr __percpu *msrs,
106 rv.msrs = msrs;
122 * @msrs: array of MSR values
125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) argument
127 __rwmsr_on_cpus(mask, msr_no, msrs, __rdmsr_on_cp
97 __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs, void (*msr_func) (void *info)) argument
139 wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) argument
[all...]
/linux-master/arch/x86/kvm/vmx/
H A Dnested.h33 int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata);
101 return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low);
111 return to_vmx(vcpu)->nested.msrs.misc_low &
117 return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS;
122 return to_vmx(vcpu)->nested.msrs.procbased_ctls_high &
128 return to_vmx(vcpu)->nested.msrs.secondary_ctls_high &
259 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0;
260 u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1;
263 if (to_vmx(vcpu)->nested.msrs.secondary_ctls_high &
273 u64 fixed0 = to_vmx(vcpu)->nested.msrs
[all...]
H A Dnested.c443 bool execonly = vmx->nested.msrs.ept_caps & VMX_EPT_EXECUTE_ONLY_BIT;
444 int ept_lpage_level = ept_caps_to_lpage_level(vmx->nested.msrs.ept_caps);
923 u64 vmx_misc = vmx_control_msr(vmx->nested.msrs.misc_low,
924 vmx->nested.msrs.misc_high);
1255 vmx->nested.msrs.basic = data;
1259 static void vmx_get_control_msr(struct nested_vmx_msrs *msrs, u32 msr_index, argument
1264 *low = &msrs->pinbased_ctls_low;
1265 *high = &msrs->pinbased_ctls_high;
1268 *low = &msrs->procbased_ctls_low;
1269 *high = &msrs
1361 vmx_get_fixed0_msr(struct nested_vmx_msrs *msrs, u32 msr_index) argument
1458 vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata) argument
6806 nested_vmx_setup_pinbased_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6823 nested_vmx_setup_exit_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6846 nested_vmx_setup_entry_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6866 nested_vmx_setup_cpubased_ctls(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6902 nested_vmx_setup_secondary_ctls(u32 ept_caps, struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6985 nested_vmx_setup_misc_data(struct vmcs_config *vmcs_conf, struct nested_vmx_msrs *msrs) argument
6997 nested_vmx_setup_basic(struct nested_vmx_msrs *msrs) argument
7015 nested_vmx_setup_cr_fixed(struct nested_vmx_msrs *msrs) argument
7047 struct nested_vmx_msrs *msrs = &vmcs_conf->nested; local
[all...]
H A Dvmx.c2051 if (vmx_get_vmx_msr(&vmx->nested.msrs, msr_info->index,
2254 ((vmx->nested.msrs.entry_ctls_high & VM_ENTRY_LOAD_BNDCFGS) ||
2255 (vmx->nested.msrs.exit_ctls_high & VM_EXIT_CLEAR_BNDCFGS)))
4550 vmx->nested.msrs.secondary_ctls_high |= control;
4552 vmx->nested.msrs.secondary_ctls_high &= ~control;
4827 memcpy(&vmx->nested.msrs, &vmcs_config.nested, sizeof(vmx->nested.msrs));
7154 struct perf_guest_switch_msr *msrs; local
7162 msrs = perf_guest_get_msrs(&nr_msrs, (void *)pmu);
7163 if (!msrs)
[all...]
H A Dvmx.h237 struct nested_vmx_msrs msrs; member in struct:nested_vmx
/linux-master/arch/x86/include/asm/
H A Dmsr.h20 struct msr __percpu *msrs; member in struct:msr_info
329 void msrs_free(struct msr __percpu *msrs);
338 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs);
339 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs);
368 struct msr __percpu *msrs)
370 rdmsr_on_cpu(0, msr_no, raw_cpu_ptr(&msrs->l), raw_cpu_ptr(&msrs->h));
373 struct msr __percpu *msrs)
375 wrmsr_on_cpu(0, msr_no, raw_cpu_read(msrs->l), raw_cpu_read(msrs
367 rdmsr_on_cpus(const struct cpumask *m, u32 msr_no, struct msr __percpu *msrs) argument
372 wrmsr_on_cpus(const struct cpumask *m, u32 msr_no, struct msr __percpu *msrs) argument
[all...]
/linux-master/tools/testing/selftests/kvm/include/x86_64/
H A Dprocessor.h417 struct kvm_msrs msrs; member in struct:kvm_x86_state
843 struct kvm_msrs *msrs)
845 int r = __vcpu_ioctl(vcpu, KVM_GET_MSRS, msrs);
847 TEST_ASSERT(r == msrs->nmsrs,
849 r, r < 0 || r >= msrs->nmsrs ? -1 : msrs->entries[r].index);
851 static inline void vcpu_msrs_set(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs) argument
853 int r = __vcpu_ioctl(vcpu, KVM_SET_MSRS, msrs);
855 TEST_ASSERT(r == msrs->nmsrs,
857 r, r < 0 || r >= msrs
842 vcpu_msrs_get(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs) argument
[all...]
/linux-master/arch/x86/kernel/cpu/mce/
H A Damd.c637 u32 msrs[NR_BLOCKS]; local
640 msrs[0] = 0x00000413; /* MC4_MISC0 */
641 msrs[1] = 0xc0000408; /* MC4_MISC1 */
649 msrs[0] = MSR_AMD64_SMCA_MCx_MISC(bank);
664 msr_clear_bit(msrs[i], 62);
/linux-master/tools/testing/selftests/kvm/lib/x86_64/
H A Dprocessor.c1013 state = malloc(sizeof(*state) + msr_list->nmsrs * sizeof(state->msrs.entries[0]));
1037 state->msrs.nmsrs = msr_list->nmsrs;
1039 state->msrs.entries[i].index = msr_list->indices[i];
1040 vcpu_msrs_get(vcpu, &state->msrs);
1050 vcpu_msrs_set(vcpu, &state->msrs);
/linux-master/drivers/edac/
H A Damd64_edac.c15 static struct msr __percpu *msrs; variable in typeref:struct:__percpu
3199 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs);
3202 struct msr *reg = per_cpu_ptr(msrs, cpu);
3231 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
3235 struct msr *reg = per_cpu_ptr(msrs, cpu);
3250 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
4154 msrs = msrs_alloc();
4155 if (!msrs)
4191 msrs_free(msrs);
4192 msrs
[all...]
/linux-master/arch/x86/kvm/
H A Dx86.c365 struct kvm_user_return_msrs *msrs local
375 if (msrs->registered) {
376 msrs->registered = false;
381 values = &msrs->values[slot];
431 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local
437 msrs->values[i].host = value;
438 msrs->values[i].curr = value;
445 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local
448 value = (value & mask) | (msrs->values[slot].host & ~mask);
449 if (value == msrs
468 struct kvm_user_return_msrs *msrs = per_cpu_ptr(user_return_msrs, cpu); local
4546 __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs, struct kvm_msr_entry *entries, int (*do_msr)(struct kvm_vcpu *vcpu, unsigned index, u64 *data)) argument
4570 struct kvm_msrs msrs; local
[all...]

Completed in 234 milliseconds