Lines Matching defs:efer
296 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer)
299 u64 old_efer = vcpu->arch.efer;
300 vcpu->arch.efer = efer;
304 efer |= EFER_NX;
306 if (!(efer & EFER_LMA))
307 efer &= ~EFER_LME;
310 if ((old_efer & EFER_SVME) != (efer & EFER_SVME)) {
311 if (!(efer & EFER_SVME)) {
330 vcpu->arch.efer = old_efer;
343 svm->vmcb->save.efer = efer | EFER_SVME;
574 uint64_t efer;
577 rdmsrl(MSR_EFER, efer);
578 if (efer & EFER_SVME) {
584 wrmsrl(MSR_EFER, efer & ~EFER_SVME);
610 uint64_t efer;
613 rdmsrl(MSR_EFER, efer);
614 if (efer & EFER_SVME)
623 wrmsrl(MSR_EFER, efer | EFER_SVME);
1860 if (vcpu->arch.efer & EFER_LME) {
1862 vcpu->arch.efer |= EFER_LMA;
1864 svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
1868 vcpu->arch.efer &= ~EFER_LMA;
1870 svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
2964 /* check for svm_disable while efer.svme is set */
2965 if (svm_dis && (vcpu->arch.efer & EFER_SVME))
3402 pr_err("vmpl: %d cpl: %d efer: %016llx\n",
3403 save->vmpl, save->cpl, save->efer);
4668 if (!(smram64->efer & EFER_SVME))