Lines Matching defs:state

80 #define AMD_CPUID_SVM_VMCB_CLEAN	BIT(5)  /* VMCB state caching */
125 * SVM host state saved area of size 4KB for each core.
416 struct vmcb_state *state;
421 state = svm_get_vmcb_state(sc, vcpu);
510 state->dbgctl = BIT(0);
513 state->efer = EFER_SVM;
515 /* Set up the PAT to power-on state */
516 state->g_pat = PAT_VALUE(0, PAT_WRITE_BACK) |
525 /* Set up DR6/7 to power-on state */
526 state->dr6 = DBREG_DR6_RESERVED1;
527 state->dr7 = DBREG_DR7_RESERVED1;
620 svm_cpl(struct vmcb_state *state)
628 return (state->cpl);
635 struct vmcb_state *state;
638 state = &vmcb->state;
640 if (state->efer & EFER_LMA) {
653 } else if (state->cr0 & CR0_PE) {
736 struct vmcb_state *state;
738 state = &vmcb->state;
739 paging->cr3 = state->cr3;
740 paging->cpl = svm_cpl(state);
742 paging->paging_mode = svm_paging_mode(state->cr0, state->cr4,
743 state->efer);
755 struct vmcb_state *state;
761 state = svm_get_vmcb_state(svm_sc, vcpu);
784 vmexit->u.inout.eax = (uint32_t)(state->rax);
790 vis->rflags = state->rflags;
791 vis->cr0 = state->cr0;
1126 struct vmcb_state *state;
1130 state = svm_get_vmcb_state(sc, vcpu);
1132 oldval = state->efer;
1143 if (state->cr0 & CR0_PG)
1148 if ((newval & EFER_LME) != 0 && (state->cr0 & CR0_PG) != 0)
1209 struct vmcb_state *state;
1220 state = svm_get_vmcb_state(sc, vcpu);
1222 state->rax = result & 0xffffffff;
1305 struct vmcb_state *state;
1315 state = &vmcb->state;
1324 vmexit->rip = state->rip;
1325 vmexit->inst_length = nrip_valid(code) ? ctrl->nrip - state->rip : 0;
1331 * in an inconsistent state and can trigger assertions that would
1440 eax = state->rax;
1481 (uint32_t *)&state->rax,
1489 vmexit->u.hlt.rflags = state->rflags;
1508 info2, info1, state->rip);
1514 info2, info1, state->rip);
1549 state->rip = vmexit->rip;
1593 struct vmcb_state *state;
1599 state = svm_get_vmcb_state(sc, vcpu);
1605 if (vcpustate->nextrip != state->rip) {
1609 vcpustate->nextrip, state->rip);
1690 if ((state->rflags & PSL_I) == 0) {
1692 "rflags %#lx", vector, state->rflags);
1759 (state->rflags & PSL_I) == 0 || ctrl->intr_shadow,
1762 ctrl->eventinj, ctrl->intr_shadow, state->rflags));
1967 struct vmcb_state *state;
1980 state = svm_get_vmcb_state(svm_sc, vcpu);
1995 * Invalidate the VMCB state cache by marking all fields dirty.
2016 state->rip = rip;
2021 * loading of guest state. This includes not only the state
2022 * loaded by the "vmrun" instruction but also software state
2024 * state, NPT generation number, vlapic interrupts etc.
2030 vm_exit_suspended(vm, vcpu, state->rip);
2036 vm_exit_rendezvous(vm, vcpu, state->rip);
2042 vm_exit_reqidle(vm, vcpu, state->rip);
2049 vm_exit_astpending(vm, vcpu, state->rip);
2078 VCPU_CTR1(vm, vcpu, "Resume execution at %#lx", state->rip);
2099 vcpustate->nextrip = state->rip;