Lines Matching defs:efer
984 if ((vcpu->arch.efer & EFER_LME) && !is_paging(vcpu) &&
995 if (!(vcpu->arch.efer & EFER_LME) && (cr0 & X86_CR0_PG) &&
1720 static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer)
1722 if (efer & EFER_AUTOIBRS && !guest_cpuid_has(vcpu, X86_FEATURE_AUTOIBRS))
1725 if (efer & EFER_FFXSR && !guest_cpuid_has(vcpu, X86_FEATURE_FXSR_OPT))
1728 if (efer & EFER_SVME && !guest_cpuid_has(vcpu, X86_FEATURE_SVM))
1731 if (efer & (EFER_LME | EFER_LMA) &&
1735 if (efer & EFER_NX && !guest_cpuid_has(vcpu, X86_FEATURE_NX))
1741 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer)
1743 if (efer & efer_reserved_bits)
1746 return __kvm_valid_efer(vcpu, efer);
1752 u64 old_efer = vcpu->arch.efer;
1753 u64 efer = msr_info->data;
1756 if (efer & efer_reserved_bits)
1760 if (!__kvm_valid_efer(vcpu, efer))
1764 (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME))
1768 efer &= ~EFER_LMA;
1769 efer |= vcpu->arch.efer & EFER_LMA;
1771 r = static_call(kvm_x86_set_efer)(vcpu, efer);
1777 if ((efer ^ old_efer) & KVM_MMU_EFER_ROLE_BITS)
1781 (efer & EFER_SVME))
4366 msr_info->data = vcpu->arch.efer;
11567 sregs->efer = vcpu->arch.efer;
11709 if ((sregs->efer & EFER_LME) && (sregs->cr0 & X86_CR0_PG)) {
11715 if (!(sregs->cr4 & X86_CR4_PAE) || !(sregs->efer & EFER_LMA))
11724 if (sregs->efer & EFER_LMA || sregs->cs.l)
11765 *mmu_reset_needed |= vcpu->arch.efer != sregs->efer;
11766 static_call(kvm_x86_set_efer)(vcpu, sregs->efer);
11835 !(sregs2->efer & EFER_LMA);