/freebsd-11-stable/sys/amd64/vmm/amd/ |
H A D | svm_softc.h | 47 struct vmcb vmcb; /* hardware saved vcpu context */ 48 struct svm_regctx swctx; /* software saved vcpu context */ 51 int lastcpu; /* host cpu that the vcpu last ran on */ 53 long eptgen; /* pmap->pm_eptgen when the vcpu last ran */ 62 struct svm_vcpu vcpu[VM_MAXCPU]; member in struct:svm_softc 72 svm_get_vcpu(struct svm_softc *sc, int vcpu) argument 75 return (&(sc->vcpu[vcpu])); 79 svm_get_vmcb(struct svm_softc *sc, int vcpu) argument 82 return (&(sc->vcpu[vcp 86 svm_get_vmcb_state(struct svm_softc *sc, int vcpu) argument 93 svm_get_vmcb_ctrl(struct svm_softc *sc, int vcpu) argument 100 svm_get_guest_regctx(struct svm_softc *sc, int vcpu) argument 107 svm_set_dirty(struct svm_softc *sc, int vcpu, uint32_t dirtybits) argument [all...] |
H A D | svm_msr.h | 37 void svm_msr_guest_init(struct svm_softc *sc, int vcpu); 38 void svm_msr_guest_enter(struct svm_softc *sc, int vcpu); 39 void svm_msr_guest_exit(struct svm_softc *sc, int vcpu); 41 int svm_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, 43 int svm_rdmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t *result,
|
H A D | svm.c | 133 static int svm_setreg(void *arg, int vcpu, int ident, uint64_t val); 327 * Allow vcpu to read or write the 'msr' without trapping into the hypervisor. 363 svm_get_intercept(struct svm_softc *sc, int vcpu, int idx, uint32_t bitmask) argument 369 ctrl = svm_get_vmcb_ctrl(sc, vcpu); 374 svm_set_intercept(struct svm_softc *sc, int vcpu, int idx, uint32_t bitmask, argument 382 ctrl = svm_get_vmcb_ctrl(sc, vcpu); 391 svm_set_dirty(sc, vcpu, VMCB_CACHE_I); 392 VCPU_CTR3(sc->vm, vcpu, "intercept[%d] modified " 398 svm_disable_intercept(struct svm_softc *sc, int vcpu, int off, uint32_t bitmask) argument 401 svm_set_intercept(sc, vcpu, of 405 svm_enable_intercept(struct svm_softc *sc, int vcpu, int off, uint32_t bitmask) argument 412 vmcb_init(struct svm_softc *sc, int vcpu, uint64_t iopm_base_pa, uint64_t msrpm_base_pa, uint64_t np_pml4) argument 537 struct svm_vcpu *vcpu; local 698 svm_inout_str_seginfo(struct svm_softc *svm_sc, int vcpu, int64_t info1, int in, struct vm_inout_str *vis) argument 752 svm_handle_io(struct svm_softc *svm_sc, int vcpu, struct vm_exit *vmexit) argument 910 svm_eventinject(struct svm_softc *sc, int vcpu, int intr_type, int vector, uint32_t error, bool ec_valid) argument 949 svm_update_virqinfo(struct svm_softc *sc, int vcpu) argument 968 svm_save_intinfo(struct svm_softc *svm_sc, int vcpu) argument 992 vintr_intercept_enabled(struct svm_softc *sc, int vcpu) argument 1001 enable_intr_window_exiting(struct svm_softc *sc, int vcpu) argument 1023 disable_intr_window_exiting(struct svm_softc *sc, int vcpu) argument 1043 svm_modify_intr_shadow(struct svm_softc *sc, int vcpu, uint64_t val) argument 1059 svm_get_intr_shadow(struct svm_softc *sc, int vcpu, uint64_t *val) argument 1074 nmi_blocked(struct svm_softc *sc, int vcpu) argument 1084 enable_nmi_blocking(struct svm_softc *sc, int vcpu) argument 1093 clear_nmi_blocking(struct svm_softc *sc, int vcpu) argument 1123 svm_write_efer(struct svm_softc *sc, int vcpu, uint64_t newval, bool *retu) argument 1191 emulate_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, bool *retu) argument 1207 emulate_rdmsr(struct svm_softc *sc, int vcpu, u_int num, bool *retu) argument 1302 svm_vmexit(struct svm_softc *svm_sc, int vcpu, struct vm_exit *vmexit) argument 1568 svm_inj_intinfo(struct svm_softc *svm_sc, int vcpu) argument 1590 svm_inj_interrupts(struct svm_softc *sc, int vcpu, struct vlapic *vlapic) argument 1961 svm_vmrun(void *arg, int vcpu, register_t rip, pmap_t pmap, struct vm_eventinfo *evinfo) argument 2167 svm_getreg(void *arg, int vcpu, int ident, uint64_t *val) argument 2194 svm_setreg(void *arg, int vcpu, int ident, uint64_t val) argument 2230 svm_setcap(void *arg, int vcpu, int type, int val) argument 2259 svm_getcap(void *arg, int vcpu, int type, int *retval) argument [all...] |
H A D | vmcb.c | 116 vmcb_access(struct svm_softc *softc, int vcpu, int write, int ident, argument 123 vmcb = svm_get_vmcb(softc, vcpu); 145 VCPU_CTR1(softc->vm, vcpu, 152 svm_set_dirty(softc, vcpu, 0xffffffff); 161 vmcb_read(struct svm_softc *sc, int vcpu, int ident, uint64_t *retval) argument 168 vmcb = svm_get_vmcb(sc, vcpu); 173 return (vmcb_access(sc, vcpu, 0, ident, retval)); 251 vmcb_write(struct svm_softc *sc, int vcpu, int ident, uint64_t val) argument 258 vmcb = svm_get_vmcb(sc, vcpu); 264 return (vmcb_access(sc, vcpu, 364 vmcb_setdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 416 vmcb_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument [all...] |
H A D | svm_msr.c | 73 svm_msr_guest_init(struct svm_softc *sc, int vcpu) argument 87 svm_msr_guest_enter(struct svm_softc *sc, int vcpu) argument 95 svm_msr_guest_exit(struct svm_softc *sc, int vcpu) argument 109 svm_rdmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t *result, argument 138 svm_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, bool *retu) argument 147 vm_inject_gp(sc->vm, vcpu);
|
/freebsd-11-stable/usr.sbin/bhyve/ |
H A D | xmsr.h | 35 int emulate_wrmsr(struct vmctx *ctx, int vcpu, uint32_t code, uint64_t val); 36 int emulate_rdmsr(struct vmctx *ctx, int vcpu, uint32_t code, uint64_t *val);
|
H A D | spinup_ap.h | 34 int spinup_ap(struct vmctx *ctx, int vcpu, int newcpu, uint64_t rip);
|
H A D | task_switch.c | 103 GETREG(struct vmctx *ctx, int vcpu, int reg) argument 108 error = vm_get_register(ctx, vcpu, reg, &val); 114 SETREG(struct vmctx *ctx, int vcpu, int reg, uint64_t val) argument 118 error = vm_set_register(ctx, vcpu, reg, val); 154 sel_exception(struct vmctx *ctx, int vcpu, int vector, uint16_t sel, int ext) argument 168 vm_inject_fault(ctx, vcpu, vector, 1, sel); 176 desc_table_limit_check(struct vmctx *ctx, int vcpu, uint16_t sel) argument 183 error = vm_get_desc(ctx, vcpu, reg, &base, &limit, &access); 206 desc_table_rw(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, argument 216 error = vm_get_desc(ctx, vcpu, re 234 desc_table_read(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 241 desc_table_write(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 255 read_tss_descriptor(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 311 validate_seg_desc(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, int segment, struct seg_desc *seg_desc, int *faultptr) argument 437 tss32_save(struct vmctx *ctx, int vcpu, struct vm_task_switch *task_switch, uint32_t eip, struct tss32 *tss, struct iovec *iov) argument 470 update_seg_desc(struct vmctx *ctx, int vcpu, int reg, struct seg_desc *sd) argument 482 tss32_restore(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, uint16_t ot_sel, struct tss32 *tss, struct iovec *iov, int *faultptr) argument 626 push_errcode(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, int task_type, uint32_t errcode, int *faultptr) argument 716 int error, ext, fault, minlimit, nt_type, ot_type, vcpu; local [all...] |
H A D | mem.h | 38 typedef int (*mem_func_t)(struct vmctx *ctx, int vcpu, int dir, uint64_t addr, 56 int emulate_mem(struct vmctx *, int vcpu, uint64_t paddr, struct vie *vie,
|
H A D | post.c | 42 post_data_handler(struct vmctx *ctx, int vcpu, int in, int port, int bytes, argument
|
H A D | inout.c | 68 default_inout(struct vmctx *ctx, int vcpu, int in, int port, int bytes, argument 106 emulate_inout(struct vmctx *ctx, int vcpu, struct vm_exit *vmexit, int strict) argument 163 vm_inject_gp(ctx, vcpu); 167 error = vm_copy_setup(ctx, vcpu, &vis->paging, gla, 179 vm_inject_ac(ctx, vcpu, 0); 185 vm_copyin(ctx, vcpu, iov, &val, bytes); 187 retval = handler(ctx, vcpu, in, port, bytes, &val, arg); 192 vm_copyout(ctx, vcpu, &val, iov, bytes); 205 error = vie_update_register(ctx, vcpu, idxreg, index, addrsize); 213 error = vie_update_register(ctx, vcpu, VM_REG_GUEST_RC [all...] |
H A D | inout.h | 42 typedef int (*inout_func_t)(struct vmctx *ctx, int vcpu, int in, int port, 75 int emulate_inout(struct vmctx *, int vcpu, struct vm_exit *vmexit,
|
H A D | mem.c | 143 mem_read(void *ctx, int vcpu, uint64_t gpa, uint64_t *rval, int size, void *arg) argument 148 error = (*mr->handler)(ctx, vcpu, MEM_F_READ, gpa, size, 154 mem_write(void *ctx, int vcpu, uint64_t gpa, uint64_t wval, int size, void *arg) argument 159 error = (*mr->handler)(ctx, vcpu, MEM_F_WRITE, gpa, size, 165 emulate_mem(struct vmctx *ctx, int vcpu, uint64_t paddr, struct vie *vie, argument 176 if (mmio_hint[vcpu] && 177 paddr >= mmio_hint[vcpu]->mr_base && 178 paddr <= mmio_hint[vcpu]->mr_end) { 179 entry = mmio_hint[vcpu]; 186 mmio_hint[vcpu] [all...] |
H A D | bhyverun.c | 161 typedef int (*vmexit_handler_t)(struct vmctx *, struct vm_exit *, int *vcpu); 162 extern int vmexit_task_switch(struct vmctx *, struct vm_exit *, int *vcpu); 186 static void vm_loop(struct vmctx *ctx, int vcpu, uint64_t rip); 217 " %*s [-m mem] [-p vcpu:hostcpu] [-s <pci>] [-U uuid] <vm>\n" 229 " -p: pin 'vcpu' to 'hostcpu'\n" 373 int vcpu, pcpu; local 375 if (sscanf(opt, "%d:%d", &vcpu, &pcpu) != 2) { 380 if (vcpu < 0 || vcpu >= VM_MAXCPU) { 381 fprintf(stderr, "vcpu ' 404 vm_inject_fault(void *arg, int vcpu, int vector, int errcode_valid, int errcode) argument 451 int vcpu; local 501 fbsdrun_deletecpu(struct vmctx *ctx, int vcpu) argument 530 int vcpu; local 824 vm_loop(struct vmctx *ctx, int vcpu, uint64_t startrip) argument [all...] |
/freebsd-11-stable/lib/libvmmapi/ |
H A D | vmmapi_freebsd.c | 68 * Setup the 'vcpu' register set such that it will begin execution at 72 vm_setup_freebsd_registers_i386(struct vmctx *vmctx, int vcpu, uint32_t eip, argument 82 error = vm_get_capability(vmctx, vcpu, VM_CAP_UNRESTRICTED_GUEST, &tmp); 85 error = vm_set_capability(vmctx, vcpu, VM_CAP_UNRESTRICTED_GUEST, 1); 90 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_CR0, cr0)) != 0) 93 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_CR4, 0)) != 0) 100 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_EFER, 0))) 109 error = vm_set_desc(vmctx, vcpu, VM_REG_GUEST_GDTR, 119 error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_RFLAGS, rflags); 126 error = vm_set_desc(vmctx, vcpu, VM_REG_GUEST_C 222 vm_setup_freebsd_registers(struct vmctx *vmctx, int vcpu, uint64_t rip, uint64_t cr3, uint64_t gdtbase, uint64_t rsp) argument [all...] |
H A D | vmmapi.h | 120 int vm_set_desc(struct vmctx *ctx, int vcpu, int reg, 122 int vm_get_desc(struct vmctx *ctx, int vcpu, int reg, 124 int vm_get_seg_desc(struct vmctx *ctx, int vcpu, int reg, 126 int vm_set_register(struct vmctx *ctx, int vcpu, int reg, uint64_t val); 127 int vm_get_register(struct vmctx *ctx, int vcpu, int reg, uint64_t *retval); 128 int vm_run(struct vmctx *ctx, int vcpu, struct vm_exit *ret_vmexit); 132 int vm_inject_exception(struct vmctx *ctx, int vcpu, int vector, 134 int vm_lapic_irq(struct vmctx *ctx, int vcpu, int vector); 135 int vm_lapic_local_irq(struct vmctx *ctx, int vcpu, int vector); 146 int vm_inject_nmi(struct vmctx *ctx, int vcpu); [all...] |
H A D | vmmapi.c | 510 vm_set_desc(struct vmctx *ctx, int vcpu, int reg, argument 517 vmsegdesc.cpuid = vcpu; 528 vm_get_desc(struct vmctx *ctx, int vcpu, int reg, argument 535 vmsegdesc.cpuid = vcpu; 548 vm_get_seg_desc(struct vmctx *ctx, int vcpu, int reg, struct seg_desc *seg_desc) argument 552 error = vm_get_desc(ctx, vcpu, reg, &seg_desc->base, &seg_desc->limit, 558 vm_set_register(struct vmctx *ctx, int vcpu, int reg, uint64_t val) argument 564 vmreg.cpuid = vcpu; 573 vm_get_register(struct vmctx *ctx, int vcpu, int reg, uint64_t *ret_val) argument 579 vmreg.cpuid = vcpu; 588 vm_run(struct vmctx *ctx, int vcpu, struct vm_exit *vmexit) argument 619 vm_inject_exception(struct vmctx *ctx, int vcpu, int vector, int errcode_valid, uint32_t errcode, int restart_instruction) argument 644 vm_lapic_irq(struct vmctx *ctx, int vcpu, int vector) argument 656 vm_lapic_local_irq(struct vmctx *ctx, int vcpu, int vector) argument 769 vm_inject_nmi(struct vmctx *ctx, int vcpu) argument 818 vm_get_capability(struct vmctx *ctx, int vcpu, enum vm_cap_type cap, int *retval) argument 834 vm_set_capability(struct vmctx *ctx, int vcpu, enum vm_cap_type cap, int val) argument 890 vm_setup_pptdev_msi(struct vmctx *ctx, int vcpu, int bus, int slot, int func, uint64_t addr, uint64_t msg, int numvec) argument 908 vm_setup_pptdev_msix(struct vmctx *ctx, int vcpu, int bus, int slot, int func, int idx, uint64_t addr, uint64_t msg, uint32_t vector_control) argument 927 vm_get_stats(struct vmctx *ctx, int vcpu, struct timeval *ret_tv, int *ret_entries) argument 960 vm_get_x2apic_state(struct vmctx *ctx, int vcpu, enum x2apic_state *state) argument 974 vm_set_x2apic_state(struct vmctx *ctx, int vcpu, enum x2apic_state state) argument 993 vcpu_reset(struct vmctx *vmctx, int vcpu) argument 1179 vm_gla2gpa(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint64_t gla, int prot, uint64_t *gpa, int *fault) argument 1204 vm_copy_setup(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint64_t gla, size_t len, int prot, struct iovec *iov, int iovcnt, int *fault) argument 1242 vm_copy_teardown(struct vmctx *ctx, int vcpu, struct iovec *iov, int iovcnt) argument 1249 vm_copyin(struct vmctx *ctx, int vcpu, struct iovec *iov, void *vp, size_t len) argument 1269 vm_copyout(struct vmctx *ctx, int vcpu, const void *vp, struct iovec *iov, size_t len) argument 1319 vm_activate_cpu(struct vmctx *ctx, int vcpu) argument 1331 vm_get_intinfo(struct vmctx *ctx, int vcpu, uint64_t *info1, uint64_t *info2) argument 1347 vm_set_intinfo(struct vmctx *ctx, int vcpu, uint64_t info1) argument 1412 vm_restart_instruction(void *arg, int vcpu) argument [all...] |
/freebsd-11-stable/usr.sbin/bhyvectl/ |
H A D | bhyvectl.c | 328 dump_vm_run_exitcode(struct vm_exit *vmexit, int vcpu) argument 330 printf("vm exit[%d]\n", vcpu); 412 print_msr_pm(uint64_t msr, int vcpu, int readable, int writeable) argument 416 printf("%-20s[%d]\t\t%c%c\n", msr_name(msr), vcpu, 425 dump_amd_msr_pm(const char *bitmap, int vcpu) argument 437 print_msr_pm(msr, vcpu, readable, writeable); 443 print_msr_pm(msr + MSR_AMD6TH_START, vcpu, readable, 450 print_msr_pm(msr + MSR_AMD7TH_START, vcpu, readable, 459 dump_intel_msr_pm(const char *bitmap, int vcpu) argument 471 print_msr_pm(msr, vcpu, readabl 483 dump_msr_bitmap(int vcpu, uint64_t addr, bool cpu_intel) argument 524 vm_get_vmcs_field(struct vmctx *ctx, int vcpu, int field, uint64_t *ret_val) argument 531 vm_get_vmcb_field(struct vmctx *ctx, int vcpu, int off, int bytes, uint64_t *ret_val) argument 652 get_all_registers(struct vmctx *ctx, int vcpu) argument 839 get_all_segments(struct vmctx *ctx, int vcpu) argument 986 get_misc_vmcs(struct vmctx *ctx, int vcpu) argument 1220 get_misc_vmcb(struct vmctx *ctx, int vcpu) argument 1662 int error, ch, vcpu, ptenum; local [all...] |
/freebsd-11-stable/sys/amd64/vmm/ |
H A D | vmm.c | 88 * (a) allocated when vcpu is created 89 * (i) initialized when vcpu is created and when it is reinitialized 90 * (o) initialized the first time the vcpu is created 93 struct vcpu { struct 95 enum vcpu_state state; /* (o) vcpu state */ 96 int hostcpu; /* (o) vcpu's host cpu */ 97 int reqidle; /* (i) request vcpu to idle */ 115 #define vcpu_lock_init(v) mtx_init(&((v)->mtx), "vcpu lock", 0, MTX_SPIN) 165 struct vcpu vcpu[VM_MAXCP member in struct:vm 262 struct vcpu *vcpu = &vm->vcpu[i]; local 274 struct vcpu *vcpu; local 313 struct vcpu *vcpu; local 1006 vm_get_register(struct vm *vm, int vcpu, int reg, uint64_t *retval) argument 1021 struct vcpu *vcpu; local 1074 vm_get_seg_desc(struct vm *vm, int vcpu, int reg, struct seg_desc *desc) argument 1088 vm_set_seg_desc(struct vm *vm, int vcpu, int reg, struct seg_desc *desc) argument 1101 restore_guest_fpustate(struct vcpu *vcpu) argument 1123 save_guest_fpustate(struct vcpu *vcpu) argument 1147 struct vcpu *vcpu; local 1298 struct vcpu *vcpu; local 1382 struct vcpu *vcpu; local 1423 struct vcpu *vcpu; local 1501 struct vcpu *vcpu; local 1552 struct vcpu *vcpu = &vm->vcpu[vcpuid]; local 1647 struct vcpu *vcpu; local 1751 struct vcpu *vcpu; local 1793 struct vcpu *vcpu; local 1915 vcpu_exception_intinfo(struct vcpu *vcpu) argument 1933 struct vcpu *vcpu; local 1976 struct vcpu *vcpu; local 1991 struct vcpu *vcpu; local 2084 struct vcpu *vcpu; local 2099 struct vcpu *vcpu; local 2112 struct vcpu *vcpu; local 2131 struct vcpu *vcpu; local 2146 struct vcpu *vcpu; local 2159 struct vcpu *vcpu; local 2174 vm_get_capability(struct vm *vm, int vcpu, int type, int *retval) argument 2186 vm_set_capability(struct vm *vm, int vcpu, int type, int val) argument 2271 struct vcpu *vcpu; local 2288 struct vcpu *vcpu; local 2376 vcpu_notify_event_locked(struct vcpu *vcpu, bool lapic_intr) argument 2409 struct vcpu *vcpu = &vm->vcpu[vcpuid]; local 2624 vm_get_rescnt(struct vm *vm, int vcpu, struct vmm_stat_type *stat) argument 2634 vm_get_wiredcnt(struct vm *vm, int vcpu, struct vmm_stat_type *stat) argument [all...] |
H A D | vmm_stat.h | 46 typedef void (*vmm_stat_func_t)(struct vm *vm, int vcpu, 91 int vmm_stat_copy(struct vm *vm, int vcpu, int *num_stats, uint64_t *buf); 95 vmm_stat_array_incr(struct vm *vm, int vcpu, struct vmm_stat_type *vst, argument 101 stats = vcpu_stats(vm, vcpu); 109 vmm_stat_array_set(struct vm *vm, int vcpu, struct vmm_stat_type *vst, argument 115 stats = vcpu_stats(vm, vcpu); 123 vmm_stat_incr(struct vm *vm, int vcpu, struct vmm_stat_type *vst, uint64_t x) argument 127 vmm_stat_array_incr(vm, vcpu, vst, 0, x); 132 vmm_stat_set(struct vm *vm, int vcpu, struct vmm_stat_type *vst, uint64_t val) argument 136 vmm_stat_array_set(vm, vcpu, vs [all...] |
H A D | vmm_stat.c | 83 vmm_stat_copy(struct vm *vm, int vcpu, int *num_stats, uint64_t *buf) argument 89 if (vcpu < 0 || vcpu >= vm_get_maxcpus(vm)) 96 (*vst->func)(vm, vcpu, vst); 100 stats = vcpu_stats(vm, vcpu); 150 VMM_STAT(VCPU_MIGRATIONS, "vcpu migration across host cpus");
|
/freebsd-11-stable/sys/amd64/vmm/intel/ |
H A D | vmx.c | 289 static int vmx_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc); 290 static int vmx_getreg(void *arg, int vcpu, int reg, uint64_t *retval); 510 * Allocate a unique VPID for each vcpu from the unit number allocator. 1023 panic("vmx_vminit: vmclear error %d on vcpu %d\n", 1112 vmx_handle_cpuid(struct vm *vm, int vcpu, struct vmxctx *vmxctx) argument 1118 handled = x86_emulate_cpuid(vm, vcpu, 1127 vmx_run_trace(struct vmx *vmx, int vcpu) argument 1130 VCPU_CTR1(vmx->vm, vcpu, "Resume execution at %#lx", vmcs_guest_rip()); 1135 vmx_exit_trace(struct vmx *vmx, int vcpu, uint64_t rip, uint32_t exit_reason, argument 1139 VCPU_CTR3(vmx->vm, vcpu, " 1146 vmx_astpending_trace(struct vmx *vmx, int vcpu, uint64_t rip) argument 1160 vmx_invvpid(struct vmx *vmx, int vcpu, pmap_t pmap, int running) argument 1217 vmx_set_pcpu_defaults(struct vmx *vmx, int vcpu, pmap_t pmap) argument 1241 vmx_set_int_window_exiting(struct vmx *vmx, int vcpu) argument 1252 vmx_clear_int_window_exiting(struct vmx *vmx, int vcpu) argument 1263 vmx_set_nmi_window_exiting(struct vmx *vmx, int vcpu) argument 1274 vmx_clear_nmi_window_exiting(struct vmx *vmx, int vcpu) argument 1285 vmx_set_tsc_offset(struct vmx *vmx, int vcpu, uint64_t offset) argument 1306 vmx_inject_nmi(struct vmx *vmx, int vcpu) argument 1332 vmx_inject_interrupts(struct vmx *vmx, int vcpu, struct vlapic *vlapic, uint64_t guestrip) argument 1561 vmx_emulate_xsetbv(struct vmx *vmx, int vcpu, struct vm_exit *vmexit) argument 1637 vmx_get_guest_reg(struct vmx *vmx, int vcpu, int ident) argument 1682 vmx_set_guest_reg(struct vmx *vmx, int vcpu, int ident, uint64_t regval) argument 1743 vmx_emulate_cr0_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 1781 vmx_emulate_cr4_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 1801 vmx_emulate_cr8_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 2240 vmx_exit_process(struct vmx *vmx, int vcpu, struct vm_exit *vmexit) argument 2847 vmx_run(void *arg, int vcpu, register_t rip, pmap_t pmap, struct vm_eventinfo *evinfo) argument 3121 vmx_get_intr_shadow(struct vmx *vmx, int vcpu, int running, uint64_t *retval) argument 3133 vmx_modify_intr_shadow(struct vmx *vmx, int vcpu, int running, uint64_t val) argument 3182 vmx_getreg(void *arg, int vcpu, int reg, uint64_t *retval) argument 3201 vmx_setreg(void *arg, int vcpu, int reg, uint64_t val) argument 3268 vmx_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 3281 vmx_setdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 3294 vmx_getcap(void *arg, int vcpu, int type, int *retval) argument 3336 vmx_setcap(void *arg, int vcpu, int type, int val) argument [all...] |
/freebsd-11-stable/sys/amd64/vmm/io/ |
H A D | ppt.h | 35 int ppt_setup_msi(struct vm *vm, int vcpu, int bus, int slot, int func, 37 int ppt_setup_msix(struct vm *vm, int vcpu, int bus, int slot, int func,
|
/freebsd-11-stable/sys/xen/interface/ |
H A D | event_channel.h | 129 * vcpu. 131 * 1. Virtual IRQs are classified as per-vcpu or global. See the VIRQ list 135 * 3. Per-vcpu VIRQs may be bound to at most one event channel per vcpu. 136 * The allocated event channel is bound to the specified vcpu and the 142 uint32_t vcpu; member in struct:evtchn_bind_virq 167 * 1. The allocated event channel is bound to the specified vcpu. The binding 171 uint32_t vcpu; member in struct:evtchn_bind_ipi 218 uint32_t vcpu; /* VCPU to which this channel is bound. */ member in struct:evtchn_status 234 * EVTCHNOP_bind_vcpu: Specify which vcpu 248 uint32_t vcpu; member in struct:evtchn_bind_vcpu 289 uint32_t vcpu; member in struct:evtchn_init_control [all...] |
/freebsd-11-stable/sys/amd64/include/ |
H A D | vmm.h | 142 typedef int (*vmi_run_func_t)(void *vmi, int vcpu, register_t rip, 145 typedef int (*vmi_get_register_t)(void *vmi, int vcpu, int num, 147 typedef int (*vmi_set_register_t)(void *vmi, int vcpu, int num, 149 typedef int (*vmi_get_desc_t)(void *vmi, int vcpu, int num, 151 typedef int (*vmi_set_desc_t)(void *vmi, int vcpu, int num, 153 typedef int (*vmi_get_cap_t)(void *vmi, int vcpu, int num, int *retval); 154 typedef int (*vmi_set_cap_t)(void *vmi, int vcpu, int num, int val); 157 typedef struct vlapic * (*vmi_vlapic_init)(void *vmi, int vcpu); 206 * APIs that inspect the guest memory map require only a *single* vcpu to 220 int vm_get_register(struct vm *vm, int vcpu, in 314 vcpu_is_running(struct vm *vm, int vcpu, int *hostcpu) argument 321 vcpu_should_yield(struct vm *vm, int vcpu) argument 644 int vcpu; member in struct:vm_exit::__anon4079::__anon4085 [all...] |