/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 561 __u32 msr; member in struct:kvm_xen_hvm_config
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 34 #include <asm/msr-index.h> 1634 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr); 1635 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr); 2080 int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr); 2081 int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr); 2186 static inline unsigned long read_msr(unsigned long msr) argument 2190 rdmsrl(msr, value); 2240 int kvm_add_user_return_msr(u32 msr); 2241 int kvm_find_user_return_msr(u32 msr); 2244 static inline bool kvm_is_supported_user_return_msr(u32 msr) argument [all...] |
/linux-master/arch/powerpc/xmon/ |
H A D | xmon.c | 554 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) 703 if (regs->msr & MSR_DE) { 711 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { 764 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) 797 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) 812 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) 837 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { 1192 regs_set_return_msr(regs, regs->msr | MSR_DE); 1208 if ((regs->msr & (MSR_64BIT|MSR_PR|MSR_IR)) == (MSR_64BIT|MSR_IR)) { 1225 regs_set_return_msr(regs, regs->msr | MSR_S 2048 unsigned long msr; local [all...] |
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | rtas-fadump.c | 317 regs->msr = (unsigned long)reg_val;
|
/linux-master/arch/powerpc/mm/ |
H A D | cacheflush.c | 89 unsigned long msr, msr0; local 93 msr = msr0 & ~MSR_DR; 114 : "r" (nb), "r" (msr), "i" (bytes), "r" (msr0)
|
/linux-master/arch/powerpc/kvm/ |
H A D | e500_mmu_host.c | 311 u32 pr = vcpu->arch.shared->msr & MSR_PR; 642 addr_space = (vcpu->arch.shared->msr & MSR_IS) >> MSR_IR_LG; 672 pr = vcpu->arch.shared->msr & MSR_PR;
|
H A D | book3s_pr.c | 54 ulong msr); 69 ulong msr = kvmppc_get_msr(vcpu); local 70 return (msr & (MSR_IR|MSR_DR)) == MSR_DR; 75 ulong msr = kvmppc_get_msr(vcpu); local 79 if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) 109 unsigned long msr, pc, new_msr, new_pc; local 113 msr = kvmppc_get_msr(vcpu); 120 if (MSR_TM_TRANSACTIONAL(msr)) 123 new_msr |= msr & MSR_TS_MASK; 127 kvmppc_set_srr1(vcpu, (msr 466 kvmppc_set_msr_pr(struct kvm_vcpu *vcpu, u64 msr) argument 799 kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr) argument 861 kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr, ulong msr) argument 1060 u64 msr = kvmppc_get_msr(vcpu); local 1069 u64 msr = kvmppc_get_msr(vcpu); local [all...] |
H A D | book3s_hv_nestedv2.c | 308 vcpu->arch.shregs.msr); 541 vcpu->arch.shregs.msr = kvmppc_gse_get_u64(gse);
|
H A D | book3s_hv.c | 488 pr_err("pc = %.16lx msr = %.16llx trap = %x\n", 489 vcpu->arch.regs.nip, vcpu->arch.shregs.msr, vcpu->arch.trap); 810 dt->srr1 = cpu_to_be64(vcpu->arch.shregs.msr); 1645 printk(KERN_EMERG "trap=0x%x | pc=0x%lx | msr=0x%llx\n", 1647 vcpu->arch.shregs.msr); 1947 printk(KERN_EMERG "trap=0x%x | pc=0x%lx | msr=0x%llx\n", 1975 pr_emerg("trap=0x%x | pc=0x%lx | msr=0x%llx\n", 4115 unsigned long msr, i; local 4121 msr = mfmsr(); 4122 kvmppc_msr_hard_disable_set_facilities(vcpu, msr); 4166 unsigned long msr; local 4924 unsigned long msr; local [all...] |
H A D | book3s.h | 34 extern void kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 msr);
|
H A D | book3s.c | 561 regs->msr = kvmppc_get_msr(vcpu); 589 kvmppc_set_msr(vcpu, regs->msr); 800 void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 msr) argument 802 vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr);
|
/linux-master/arch/powerpc/kernel/ptrace/ |
H A D | ptrace-view.c | 56 REG_OFFSET_NAME(msr), 111 return task->thread.regs->msr | task->thread.fpexc_mode; 114 static __always_inline int set_user_msr(struct task_struct *task, unsigned long msr) argument 116 unsigned long newmsr = (task->thread.regs->msr & ~MSR_DEBUGCHANGE) | 117 (msr & MSR_DEBUGCHANGE); 221 struct membuf to_msr = membuf_at(&to, offsetof(struct pt_regs, msr)); 261 offsetof(struct pt_regs, msr) + sizeof(long));
|
H A D | ptrace-tm.c | 34 return task->thread.ckpt_regs.msr | task->thread.fpexc_mode; 37 static int set_user_ckpt_msr(struct task_struct *task, unsigned long msr) argument 39 task->thread.ckpt_regs.msr &= ~MSR_DEBUGCHANGE; 40 task->thread.ckpt_regs.msr |= msr & MSR_DEBUGCHANGE; 63 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) 89 struct membuf to_msr = membuf_at(&to, offsetof(struct pt_regs, msr)); 97 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) 144 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) 164 offsetof(struct pt_regs, msr) [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | process.c | 94 MSR_TM_ACTIVE(tsk->thread.regs->msr) && 97 tsk->thread.regs->msr); 155 unsigned long msr; local 158 msr = tsk->thread.regs->msr; 159 msr &= ~(MSR_FP|MSR_FE0|MSR_FE1); 161 msr &= ~MSR_VSX; 162 regs_set_return_msr(tsk->thread.regs, msr); 184 * another process could get scheduled after the regs->msr 191 if (tsk->thread.regs->msr 238 unsigned long msr; local 304 unsigned long msr = tsk->thread.regs->msr; local 519 unsigned long msr; local [all...] |
H A D | kprobes.c | 228 kcb->kprobe_saved_msr = regs->msr; 293 (!(regs->msr & MSR_IR) || !(regs->msr & MSR_DR))) 337 (regs->msr & ~MSR_SINGLESTEP) | 469 regs_set_return_msr(regs, regs->msr | kcb->kprobe_saved_msr); 481 * if somebody else is singlestepping across a probe point, msr 485 if (regs->msr & MSR_SINGLESTEP) 511 (regs->msr & ~MSR_SINGLESTEP) |
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_ppc.h | 272 void (*set_msr)(struct kvm_vcpu *vcpu, u64 msr); 310 void (*giveup_ext)(struct kvm_vcpu *vcpu, ulong msr); 1029 KVMPPC_VCPU_SHARED_REGS_ACCESSOR_GET(msr, 64, KVMPPC_GSID_MSR) 1033 vcpu->arch.shared->msr = cpu_to_be64(val); 1035 vcpu->arch.shared->msr = cpu_to_le64(val);
|
H A D | interrupt.h | 167 INT_SOFT_MASK_BUG_ON(regs, !(regs->msr & MSR_EE)); 199 !(regs->msr & MSR_EE)); 296 if (!(regs->msr & MSR_EE) || is_implicit_soft_masked(regs)) {
|
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_asm.h | 263 "2: msr spsr_el2, %1\n" \ 264 " msr elr_el2, %2\n" \ 366 msr sp_el0, \tmp
|
/linux-master/arch/arm/mm/ |
H A D | proc-xscale.S | 149 msr cpsr_c, r1 @ reset CPSR
|
H A D | proc-xsc3.S | 111 msr cpsr_c, r1 @ reset CPSR 458 msr cpsr_c, r0
|
H A D | proc-v7m.S | 147 msr control, r1 @ Thread mode has unpriviledged access
|
H A D | proc-feroceon.S | 265 msr cpsr_c, r3 @ disable interrupts 268 msr cpsr_c, r2 @ restore interrupts 312 msr cpsr_c, r3 @ disable interrupts 315 msr cpsr_c, r2 @ restore interrupts 344 msr cpsr_c, r3 @ disable interrupts 347 msr cpsr_c, r2 @ restore interrupts 376 msr cpsr_c, r3 @ disable interrupts 379 msr cpsr_c, r2 @ restore interrupts
|
H A D | proc-arm926.S | 98 msr cpsr_c, ip 102 msr cpsr_c, r3 @ Restore FIQ state
|
H A D | cache-v6.S | 47 msr cpsr_cx, r1 @ restore interrupts
|
/linux-master/arch/arm/include/asm/ |
H A D | assembler.h | 107 msr cpsr_c, #PSR_I_BIT | SVC_MODE 111 msr cpsr_c, #SVC_MODE 190 msr primask, \oldcpsr 192 msr cpsr_c, \oldcpsr 425 msr cpsr_c, \reg 429 msr cpsr_c, #\mode 451 msr spsr_cxsf, \reg 454 1: msr cpsr_c, \reg
|