/linux-master/arch/x86/pci/ |
H A D | mmconfig-shared.c | 192 u64 base, msr; local 203 msr = high; 204 msr <<= 32; 205 msr |= low; 208 if (!(msr & FAM10H_MMIO_CONF_ENABLE)) 211 base = msr & (FAM10H_MMIO_CONF_BASE_MASK<<FAM10H_MMIO_CONF_BASE_SHIFT); 213 busnbits = (msr >> FAM10H_MMIO_CONF_BUSRANGE_SHIFT) &
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.c | 373 u64 msr; local 378 msr = __rdmsr(MSR_IA32_MCU_OPT_CTRL); 379 msr |= FB_CLEAR_DIS; 380 native_wrmsrl(MSR_IA32_MCU_OPT_CTRL, msr); 382 vmx->msr_ia32_mcu_opt_ctrl = msr; 675 static int vmx_get_passthrough_msr_slot(u32 msr) argument 679 switch (msr) { 701 if (vmx_possible_passthrough_msrs[i] == msr) 705 WARN(1, "Invalid MSR %x, please adapt vmx_possible_passthrough_msrs[]", msr); 709 struct vmx_uret_msr *vmx_find_uret_msr(struct vcpu_vmx *vmx, u32 msr) argument 719 vmx_set_guest_uret_msr(struct vcpu_vmx *vmx, struct vmx_uret_msr *msr, u64 data) argument 932 msr_write_intercepted(struct vcpu_vmx *vmx, u32 msr) argument 965 vmx_find_loadstore_msr_slot(struct vmx_msrs *m, u32 msr) argument 976 clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr) argument 1028 add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr, u64 guest_val, u64 host_val, bool entry_only) argument 1836 vmx_setup_uret_msr(struct vcpu_vmx *vmx, unsigned int msr, bool load_into_hardware) argument 1936 is_vmx_feature_control_msr_valid(struct vcpu_vmx *vmx, struct msr_data *msr) argument 1960 vmx_get_msr_feature(struct kvm_msr_entry *msr) argument 1980 struct vmx_uret_msr *msr; local 2161 struct vmx_uret_msr *msr; local 2539 adjust_vmx_controls(u32 ctl_min, u32 ctl_opt, u32 msr, u32 *result) argument 2557 adjust_vmx_controls64(u64 ctl_opt, u32 msr) argument 2786 u64 msr; local 3956 vmx_disable_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type) argument 3998 vmx_enable_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type) argument 4141 u32 msr = vmx_possible_passthrough_msrs[i]; local 7828 struct vmx_uret_msr *msr; local [all...] |
/linux-master/arch/x86/events/intel/ |
H A D | uncore_nhmex.c | 212 unsigned msr = uncore_msr_box_ctl(box); local 215 if (msr) { 216 rdmsrl(msr, config); 221 wrmsrl(msr, config); 227 unsigned msr = uncore_msr_box_ctl(box); local 230 if (msr) { 231 rdmsrl(msr, config); 236 wrmsrl(msr, config); 308 /* msr offset for each instance of cbox */ 771 unsigned msr; local [all...] |
H A D | cstate.c | 147 u64 msr; member in struct:perf_cstate_msr 338 event->hw.event_base = core_msr[cfg].msr; 350 event->hw.event_base = pkg_msr[cfg].msr; 359 event->hw.event_base = module_msr[cfg].msr; 779 pkg_msr[PERF_CSTATE_PKG_C6_RES].msr = MSR_PKG_C7_RESIDENCY; 783 pkg_msr[PERF_CSTATE_CORE_C6_RES].msr = MSR_KNL_CORE_C6_RESIDENCY;
|
/linux-master/drivers/tty/ |
H A D | mxser.c | 550 u8 msr) 552 bool cts = msr & UART_MSR_CTS; 689 u8 msr = inb(port->ioaddr + UART_MSR); local 691 if (!(msr & UART_MSR_ANY_DELTA)) 692 return msr; 695 if (msr & UART_MSR_TERI) 697 if (msr & UART_MSR_DDSR) 699 if (msr & UART_MSR_DDCD) 701 if (msr & UART_MSR_DCTS) 705 if (tty_port_check_carrier(&port->port) && (msr 549 mxser_handle_cts(struct tty_struct *tty, struct mxser_port *info, u8 msr) argument 1102 u8 msr; local [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 264 u32 svm_msrpm_offset(u32 msr) argument 270 if (msr < msrpm_ranges[i] || 271 msr >= msrpm_ranges[i] + MSRS_IN_RANGE) 274 offset = (msr - msrpm_ranges[i]) / 4; /* 4 msrs per u8 */ 630 * to avoid having stale value in the msr 757 static int direct_access_msr_slot(u32 msr) argument 762 if (direct_access_msrs[i].index == msr) 768 static void set_shadow_msr_intercept(struct kvm_vcpu *vcpu, u32 msr, int read, argument 772 int slot = direct_access_msr_slot(msr); 794 static bool msr_write_intercepted(struct kvm_vcpu *vcpu, u32 msr) argument 822 set_msr_interception_bitmap(struct kvm_vcpu *vcpu, u32 *msrpm, u32 msr, int read, int write) argument 859 set_msr_interception(struct kvm_vcpu *vcpu, u32 *msrpm, u32 msr, int read, int write) argument 931 u32 msr = direct_access_msrs[i].index; local 2808 svm_get_msr_feature(struct kvm_msr_entry *msr) argument 2971 svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr) argument 5058 u64 msr, mask; local [all...] |
/linux-master/arch/x86/kvm/ |
H A D | trace.h | 1185 TP_PROTO(int vcpu_id, u32 msr, u64 data, bool host), 1186 TP_ARGS(vcpu_id, msr, data, host), 1190 __field(u32, msr) 1197 __entry->msr = msr; 1202 TP_printk("vcpu_id %d msr 0x%x data 0x%llx host %d", 1203 __entry->vcpu_id, __entry->msr, __entry->data, __entry->host) 1693 TP_PROTO(int vcpu_id, u32 vp_index, u32 msr, u64 data), 1694 TP_ARGS(vcpu_id, vp_index, msr, data), 1699 __field(u32, msr) [all...] |
H A D | pmu.c | 59 * 1. MSR (named msr): For example Intel has MSR_IA32_PERFCTRn and AMD 615 bool kvm_pmu_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr) argument 617 switch (msr) { 625 return static_call(kvm_x86_pmu_msr_idx_to_pmc)(vcpu, msr) || 626 static_call(kvm_x86_pmu_is_valid_msr)(vcpu, msr); 629 static void kvm_pmu_mark_pmc_in_use(struct kvm_vcpu *vcpu, u32 msr) argument 632 struct kvm_pmc *pmc = static_call(kvm_x86_pmu_msr_idx_to_pmc)(vcpu, msr); 641 u32 msr = msr_info->index; local 643 switch (msr) { 666 u32 msr local [all...] |
/linux-master/arch/arm/kernel/ |
H A D | entry-armv.S | 316 ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) 318 THUMB( msr cpsr_c, r0 ) 321 ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) 323 THUMB( msr cpsr_c, r0 ) 330 ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) 332 THUMB( msr cpsr_c, r0 ) 334 msr spsr_cxsf, r2 @ Restore spsr_abt 335 ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) 337 THUMB( msr cpsr_c, r0 ) 874 msr spsr_cxs [all...] |
H A D | entry-header.S | 159 msr psp, r2 180 msr cpsr_c, \rtemp @ switch to the SYS mode 186 msr cpsr_c, \rtemp @ switch back to the SVC mode 192 msr cpsr_c, \rtemp @ switch to the SYS mode 198 msr cpsr_c, \rtemp @ switch back to the SVC mode 224 msr spsr_cxsf, \rpsr 270 msr cpsr_c, #FIQ_MODE | PSR_I_BIT | PSR_F_BIT 273 msr spsr_cxsf, r9 284 msr cpsr_c, r1 316 msr spsr_cxs [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_nested.c | 303 if (MSR_TM_TRANSACTIONAL(vcpu->arch.shregs.msr)) 333 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) { 334 if (!MSR_TM_ACTIVE(l2_regs.msr)) 337 if (l2_regs.msr & MSR_TS_MASK) 339 if (WARN_ON_ONCE(vcpu->arch.shregs.msr & MSR_TS_MASK)) 354 vcpu->arch.regs.msr = vcpu->arch.shregs.msr; 370 vcpu->arch.shregs.msr = (vcpu->arch.regs.msr | MSR_ME) & ~MSR_HV; 383 l2_regs.msr [all...] |
H A D | book3s_32_mmu.c | 354 u64 msr = kvmppc_get_msr(vcpu); local 356 if (msr & (MSR_DR|MSR_IR)) { 365 switch (msr & (MSR_DR|MSR_IR)) { 385 if (msr & MSR_PR)
|
/linux-master/drivers/usb/serial/ |
H A D | ch341.c | 98 u8 msr; member in struct:ch341_private 303 priv->msr = (~(*buffer)) & CH341_BITS_MODEM_STAT; 419 if (priv->msr & CH341_BIT_DCD) 718 delta = status ^ priv->msr; 719 priv->msr = status; 792 status = priv->msr;
|
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | pmu_event_filter_test.c | 81 static void check_msr(uint32_t msr, uint64_t bits_to_flip) argument 83 uint64_t v = rdmsr(msr) ^ bits_to_flip; 85 wrmsr(msr, v); 86 if (rdmsr(msr) != v) 90 wrmsr(msr, v); 91 if (rdmsr(msr) != v)
|
/linux-master/drivers/platform/x86/intel/speed_select_if/ |
H A D | isst_if_common.c | 496 static bool match_punit_msr_white_list(int msr) argument 501 if (punit_msr_white_list[i] == msr) 515 if (!match_punit_msr_white_list(msr_cmd->msr)) 526 msr_cmd->msr, 530 ret = isst_store_cmd(0, msr_cmd->msr, 537 msr_cmd->msr, &data);
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_pgtable.c | 434 unsigned long msr, tmp, flags; local 444 "mfmsr %[msr] ;" 446 "andc %[tmp], %[msr], %[tmp] ;" 463 "mtmsrd %[msr] ;" 466 [msr] "=&r" (msr), [tmp] "=&b" (tmp), "+m" (*p)
|
/linux-master/arch/powerpc/kernel/ptrace/ |
H A D | ptrace.c | 348 BUILD_BUG_ON(offsetof(struct pt_regs, msr) != 349 offsetof(struct user_pt_regs, msr)); 420 CHECK_REG(PT_MSR, msr);
|
/linux-master/arch/x86/kernel/apic/ |
H A D | apic.c | 1691 u64 msr; local 1695 rdmsrl(MSR_IA32_XAPIC_DISABLE_STATUS, msr); 1696 return (msr & LEGACY_XAPIC_DISABLED); 1703 u64 msr; local 1708 rdmsrl(MSR_IA32_APICBASE, msr); 1709 if (!(msr & X2APIC_ENABLE)) 1712 wrmsrl(MSR_IA32_APICBASE, msr & ~(X2APIC_ENABLE | XAPIC_ENABLE)); 1713 wrmsrl(MSR_IA32_APICBASE, msr & ~X2APIC_ENABLE); 1719 u64 msr; local 1721 rdmsrl(MSR_IA32_APICBASE, msr); [all...] |
/linux-master/drivers/bluetooth/ |
H A D | dtl1_cs.c | 293 unsigned char msr; local 340 msr = inb(iobase + UART_MSR); 342 if (info->ri_latch ^ (msr & UART_MSR_RI)) { 343 info->ri_latch = msr & UART_MSR_RI;
|
/linux-master/arch/m68k/include/asm/ |
H A D | bvme6000hw.h | 51 pad_a[3], msr, member in struct:__anon467
|
/linux-master/drivers/cpufreq/ |
H A D | elanfreq.c | 24 #include <asm/msr.h>
|
/linux-master/arch/x86/kernel/ |
H A D | tsc_msr.c | 15 #include <asm/msr.h>
|
/linux-master/arch/powerpc/include/uapi/asm/ |
H A D | ptrace.h | 40 unsigned long msr; member in struct:user_pt_regs
|
/linux-master/arch/x86/lib/ |
H A D | msr-reg.S | 5 #include <asm/msr.h>
|
/linux-master/arch/arm64/include/asm/ |
H A D | asm_pointer_auth.h | 71 msr sctlr_el1, \tmp2
|