/freebsd-10-stable/sys/amd64/vmm/amd/ |
H A D | svm_softc.h | 45 struct vmcb vmcb; /* hardware saved vcpu context */ 46 struct svm_regctx swctx; /* software saved vcpu context */ 49 int lastcpu; /* host cpu that the vcpu last ran on */ 51 long eptgen; /* pmap->pm_eptgen when the vcpu last ran */ 60 struct svm_vcpu vcpu[VM_MAXCPU]; member in struct:svm_softc 70 svm_get_vcpu(struct svm_softc *sc, int vcpu) argument 73 return (&(sc->vcpu[vcpu])); 77 svm_get_vmcb(struct svm_softc *sc, int vcpu) argument 80 return (&(sc->vcpu[vcp 84 svm_get_vmcb_state(struct svm_softc *sc, int vcpu) argument 91 svm_get_vmcb_ctrl(struct svm_softc *sc, int vcpu) argument 98 svm_get_guest_regctx(struct svm_softc *sc, int vcpu) argument 105 svm_set_dirty(struct svm_softc *sc, int vcpu, uint32_t dirtybits) argument [all...] |
H A D | svm_msr.h | 35 void svm_msr_guest_init(struct svm_softc *sc, int vcpu); 36 void svm_msr_guest_enter(struct svm_softc *sc, int vcpu); 37 void svm_msr_guest_exit(struct svm_softc *sc, int vcpu); 39 int svm_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, 41 int svm_rdmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t *result,
|
H A D | svm.c | 130 static int svm_setreg(void *arg, int vcpu, int ident, uint64_t val); 324 * Allow vcpu to read or write the 'msr' without trapping into the hypervisor. 360 svm_get_intercept(struct svm_softc *sc, int vcpu, int idx, uint32_t bitmask) argument 366 ctrl = svm_get_vmcb_ctrl(sc, vcpu); 371 svm_set_intercept(struct svm_softc *sc, int vcpu, int idx, uint32_t bitmask, argument 379 ctrl = svm_get_vmcb_ctrl(sc, vcpu); 388 svm_set_dirty(sc, vcpu, VMCB_CACHE_I); 389 VCPU_CTR3(sc->vm, vcpu, "intercept[%d] modified " 395 svm_disable_intercept(struct svm_softc *sc, int vcpu, int off, uint32_t bitmask) argument 398 svm_set_intercept(sc, vcpu, of 402 svm_enable_intercept(struct svm_softc *sc, int vcpu, int off, uint32_t bitmask) argument 409 vmcb_init(struct svm_softc *sc, int vcpu, uint64_t iopm_base_pa, uint64_t msrpm_base_pa, uint64_t np_pml4) argument 517 struct svm_vcpu *vcpu; local 676 svm_inout_str_seginfo(struct svm_softc *svm_sc, int vcpu, int64_t info1, int in, struct vm_inout_str *vis) argument 730 svm_handle_io(struct svm_softc *svm_sc, int vcpu, struct vm_exit *vmexit) argument 888 svm_eventinject(struct svm_softc *sc, int vcpu, int intr_type, int vector, uint32_t error, bool ec_valid) argument 927 svm_update_virqinfo(struct svm_softc *sc, int vcpu) argument 946 svm_save_intinfo(struct svm_softc *svm_sc, int vcpu) argument 970 vintr_intercept_enabled(struct svm_softc *sc, int vcpu) argument 979 enable_intr_window_exiting(struct svm_softc *sc, int vcpu) argument 1001 disable_intr_window_exiting(struct svm_softc *sc, int vcpu) argument 1021 svm_modify_intr_shadow(struct svm_softc *sc, int vcpu, uint64_t val) argument 1037 svm_get_intr_shadow(struct svm_softc *sc, int vcpu, uint64_t *val) argument 1052 nmi_blocked(struct svm_softc *sc, int vcpu) argument 1062 enable_nmi_blocking(struct svm_softc *sc, int vcpu) argument 1071 clear_nmi_blocking(struct svm_softc *sc, int vcpu) argument 1101 svm_write_efer(struct svm_softc *sc, int vcpu, uint64_t newval, bool *retu) argument 1169 emulate_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, bool *retu) argument 1185 emulate_rdmsr(struct svm_softc *sc, int vcpu, u_int num, bool *retu) argument 1278 svm_vmexit(struct svm_softc *svm_sc, int vcpu, struct vm_exit *vmexit) argument 1530 svm_inj_intinfo(struct svm_softc *svm_sc, int vcpu) argument 1552 svm_inj_interrupts(struct svm_softc *sc, int vcpu, struct vlapic *vlapic) argument 1869 svm_vmrun(void *arg, int vcpu, register_t rip, pmap_t pmap, struct vm_eventinfo *evinfo) argument 2052 svm_getreg(void *arg, int vcpu, int ident, uint64_t *val) argument 2079 svm_setreg(void *arg, int vcpu, int ident, uint64_t val) argument 2112 svm_setcap(void *arg, int vcpu, int type, int val) argument 2141 svm_getcap(void *arg, int vcpu, int type, int *retval) argument [all...] |
H A D | vmcb.c | 114 vmcb_access(struct svm_softc *softc, int vcpu, int write, int ident, argument 121 vmcb = svm_get_vmcb(softc, vcpu); 143 VCPU_CTR1(softc->vm, vcpu, 150 svm_set_dirty(softc, vcpu, 0xffffffff); 159 vmcb_read(struct svm_softc *sc, int vcpu, int ident, uint64_t *retval) argument 166 vmcb = svm_get_vmcb(sc, vcpu); 171 return (vmcb_access(sc, vcpu, 0, ident, retval)); 245 vmcb_write(struct svm_softc *sc, int vcpu, int ident, uint64_t val) argument 252 vmcb = svm_get_vmcb(sc, vcpu); 258 return (vmcb_access(sc, vcpu, 352 vmcb_setdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 404 vmcb_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument [all...] |
H A D | svm_msr.c | 71 svm_msr_guest_init(struct svm_softc *sc, int vcpu) argument 85 svm_msr_guest_enter(struct svm_softc *sc, int vcpu) argument 93 svm_msr_guest_exit(struct svm_softc *sc, int vcpu) argument 107 svm_rdmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t *result, argument 137 svm_wrmsr(struct svm_softc *sc, int vcpu, u_int num, uint64_t val, bool *retu) argument 146 vm_inject_gp(sc->vm, vcpu);
|
/freebsd-10-stable/usr.sbin/bhyve/ |
H A D | xmsr.h | 33 int emulate_wrmsr(struct vmctx *ctx, int vcpu, uint32_t code, uint64_t val); 34 int emulate_rdmsr(struct vmctx *ctx, int vcpu, uint32_t code, uint64_t *val);
|
H A D | spinup_ap.h | 32 int spinup_ap(struct vmctx *ctx, int vcpu, int newcpu, uint64_t rip);
|
H A D | task_switch.c | 101 GETREG(struct vmctx *ctx, int vcpu, int reg) argument 106 error = vm_get_register(ctx, vcpu, reg, &val); 112 SETREG(struct vmctx *ctx, int vcpu, int reg, uint64_t val) argument 116 error = vm_set_register(ctx, vcpu, reg, val); 152 sel_exception(struct vmctx *ctx, int vcpu, int vector, uint16_t sel, int ext) argument 166 vm_inject_fault(ctx, vcpu, vector, 1, sel); 174 desc_table_limit_check(struct vmctx *ctx, int vcpu, uint16_t sel) argument 181 error = vm_get_desc(ctx, vcpu, reg, &base, &limit, &access); 204 desc_table_rw(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, argument 214 error = vm_get_desc(ctx, vcpu, re 232 desc_table_read(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 239 desc_table_write(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 253 read_tss_descriptor(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, uint16_t sel, struct user_segment_descriptor *desc, int *faultptr) argument 309 validate_seg_desc(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, int segment, struct seg_desc *seg_desc, int *faultptr) argument 435 tss32_save(struct vmctx *ctx, int vcpu, struct vm_task_switch *task_switch, uint32_t eip, struct tss32 *tss, struct iovec *iov) argument 468 update_seg_desc(struct vmctx *ctx, int vcpu, int reg, struct seg_desc *sd) argument 480 tss32_restore(struct vmctx *ctx, int vcpu, struct vm_task_switch *ts, uint16_t ot_sel, struct tss32 *tss, struct iovec *iov, int *faultptr) argument 624 push_errcode(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, int task_type, uint32_t errcode, int *faultptr) argument 714 int error, ext, fault, minlimit, nt_type, ot_type, vcpu; local [all...] |
H A D | mem.h | 36 typedef int (*mem_func_t)(struct vmctx *ctx, int vcpu, int dir, uint64_t addr, 54 int emulate_mem(struct vmctx *, int vcpu, uint64_t paddr, struct vie *vie,
|
H A D | post.c | 40 post_data_handler(struct vmctx *ctx, int vcpu, int in, int port, int bytes, argument
|
H A D | bhyverun.c | 77 typedef int (*vmexit_handler_t)(struct vmctx *, struct vm_exit *, int *vcpu); 78 extern int vmexit_task_switch(struct vmctx *, struct vm_exit *, int *vcpu); 99 static void vm_loop(struct vmctx *ctx, int vcpu, uint64_t rip); 128 " %*s [-m mem] [-p vcpu:hostcpu] [-s <pci>] [-U uuid] <vm>\n" 139 " -p: pin 'vcpu' to 'hostcpu'\n" 157 int vcpu, pcpu; local 159 if (sscanf(opt, "%d:%d", &vcpu, &pcpu) != 2) { 164 if (vcpu < 0 || vcpu >= VM_MAXCPU) { 165 fprintf(stderr, "vcpu ' 188 vm_inject_fault(void *arg, int vcpu, int vector, int errcode_valid, int errcode) argument 235 int vcpu; local 285 fbsdrun_deletecpu(struct vmctx *ctx, int vcpu) argument 314 int vcpu; local 597 vm_loop(struct vmctx *ctx, int vcpu, uint64_t startrip) argument [all...] |
H A D | inout.c | 66 default_inout(struct vmctx *ctx, int vcpu, int in, int port, int bytes, argument 104 emulate_inout(struct vmctx *ctx, int vcpu, struct vm_exit *vmexit, int strict) argument 161 vm_inject_gp(ctx, vcpu); 165 error = vm_copy_setup(ctx, vcpu, &vis->paging, gla, 177 vm_inject_ac(ctx, vcpu, 0); 183 vm_copyin(ctx, vcpu, iov, &val, bytes); 185 retval = handler(ctx, vcpu, in, port, bytes, &val, arg); 190 vm_copyout(ctx, vcpu, &val, iov, bytes); 203 error = vie_update_register(ctx, vcpu, idxreg, index, addrsize); 211 error = vie_update_register(ctx, vcpu, VM_REG_GUEST_RC [all...] |
H A D | atkbdc.c | 51 atkbdc_data_handler(struct vmctx *ctx, int vcpu, int in, int port, int bytes, argument 63 atkbdc_sts_ctl_handler(struct vmctx *ctx, int vcpu, int in, int port, argument
|
H A D | inout.h | 40 typedef int (*inout_func_t)(struct vmctx *ctx, int vcpu, int in, int port, 73 int emulate_inout(struct vmctx *, int vcpu, struct vm_exit *vmexit,
|
/freebsd-10-stable/usr.sbin/bhyvectl/ |
H A D | bhyvectl.c | 307 dump_vm_run_exitcode(struct vm_exit *vmexit, int vcpu) argument 309 printf("vm exit[%d]\n", vcpu); 391 print_msr_pm(uint64_t msr, int vcpu, int readable, int writeable) argument 395 printf("%-20s[%d]\t\t%c%c\n", msr_name(msr), vcpu, 404 dump_amd_msr_pm(const char *bitmap, int vcpu) argument 416 print_msr_pm(msr, vcpu, readable, writeable); 422 print_msr_pm(msr + MSR_AMD6TH_START, vcpu, readable, 429 print_msr_pm(msr + MSR_AMD7TH_START, vcpu, readable, 438 dump_intel_msr_pm(const char *bitmap, int vcpu) argument 450 print_msr_pm(msr, vcpu, readabl 462 dump_msr_bitmap(int vcpu, uint64_t addr, bool cpu_intel) argument 503 vm_get_vmcs_field(struct vmctx *ctx, int vcpu, int field, uint64_t *ret_val) argument 510 vm_set_vmcs_field(struct vmctx *ctx, int vcpu, int field, uint64_t val) argument 517 vm_get_vmcb_field(struct vmctx *ctx, int vcpu, int off, int bytes, uint64_t *ret_val) argument 525 vm_set_vmcb_field(struct vmctx *ctx, int vcpu, int off, int bytes, uint64_t val) argument 642 get_all_registers(struct vmctx *ctx, int vcpu) argument 792 get_all_segments(struct vmctx *ctx, int vcpu) argument 939 get_misc_vmcs(struct vmctx *ctx, int vcpu) argument 1173 get_misc_vmcb(struct vmctx *ctx, int vcpu) argument 1606 int error, ch, vcpu, ptenum; local [all...] |
/freebsd-10-stable/lib/libvmmapi/ |
H A D | vmmapi_freebsd.c | 68 * Setup the 'vcpu' register set such that it will begin execution at 72 vm_setup_freebsd_registers_i386(struct vmctx *vmctx, int vcpu, uint32_t eip, argument 82 error = vm_get_capability(vmctx, vcpu, VM_CAP_UNRESTRICTED_GUEST, &tmp); 85 error = vm_set_capability(vmctx, vcpu, VM_CAP_UNRESTRICTED_GUEST, 1); 90 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_CR0, cr0)) != 0) 93 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_CR4, 0)) != 0) 100 if ((error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_EFER, 0))) 109 error = vm_set_desc(vmctx, vcpu, VM_REG_GUEST_GDTR, 119 error = vm_set_register(vmctx, vcpu, VM_REG_GUEST_RFLAGS, rflags); 126 error = vm_set_desc(vmctx, vcpu, VM_REG_GUEST_C 222 vm_setup_freebsd_registers(struct vmctx *vmctx, int vcpu, uint64_t rip, uint64_t cr3, uint64_t gdtbase, uint64_t rsp) argument [all...] |
H A D | vmmapi.h | 119 int vm_set_desc(struct vmctx *ctx, int vcpu, int reg, 121 int vm_get_desc(struct vmctx *ctx, int vcpu, int reg, 123 int vm_get_seg_desc(struct vmctx *ctx, int vcpu, int reg, 125 int vm_set_register(struct vmctx *ctx, int vcpu, int reg, uint64_t val); 126 int vm_get_register(struct vmctx *ctx, int vcpu, int reg, uint64_t *retval); 127 int vm_run(struct vmctx *ctx, int vcpu, struct vm_exit *ret_vmexit); 131 int vm_inject_exception(struct vmctx *ctx, int vcpu, int vector, 133 int vm_lapic_irq(struct vmctx *ctx, int vcpu, int vector); 134 int vm_lapic_local_irq(struct vmctx *ctx, int vcpu, int vector); 145 int vm_inject_nmi(struct vmctx *ctx, int vcpu); [all...] |
H A D | vmmapi.c | 512 vm_set_desc(struct vmctx *ctx, int vcpu, int reg, argument 519 vmsegdesc.cpuid = vcpu; 530 vm_get_desc(struct vmctx *ctx, int vcpu, int reg, argument 537 vmsegdesc.cpuid = vcpu; 550 vm_get_seg_desc(struct vmctx *ctx, int vcpu, int reg, struct seg_desc *seg_desc) argument 554 error = vm_get_desc(ctx, vcpu, reg, &seg_desc->base, &seg_desc->limit, 560 vm_set_register(struct vmctx *ctx, int vcpu, int reg, uint64_t val) argument 566 vmreg.cpuid = vcpu; 575 vm_get_register(struct vmctx *ctx, int vcpu, int reg, uint64_t *ret_val) argument 581 vmreg.cpuid = vcpu; 590 vm_run(struct vmctx *ctx, int vcpu, struct vm_exit *vmexit) argument 621 vm_inject_exception(struct vmctx *ctx, int vcpu, int vector, int errcode_valid, uint32_t errcode, int restart_instruction) argument 646 vm_lapic_irq(struct vmctx *ctx, int vcpu, int vector) argument 658 vm_lapic_local_irq(struct vmctx *ctx, int vcpu, int vector) argument 771 vm_inject_nmi(struct vmctx *ctx, int vcpu) argument 820 vm_get_capability(struct vmctx *ctx, int vcpu, enum vm_cap_type cap, int *retval) argument 836 vm_set_capability(struct vmctx *ctx, int vcpu, enum vm_cap_type cap, int val) argument 892 vm_setup_pptdev_msi(struct vmctx *ctx, int vcpu, int bus, int slot, int func, uint64_t addr, uint64_t msg, int numvec) argument 910 vm_setup_pptdev_msix(struct vmctx *ctx, int vcpu, int bus, int slot, int func, int idx, uint64_t addr, uint64_t msg, uint32_t vector_control) argument 929 vm_get_stats(struct vmctx *ctx, int vcpu, struct timeval *ret_tv, int *ret_entries) argument 962 vm_get_x2apic_state(struct vmctx *ctx, int vcpu, enum x2apic_state *state) argument 976 vm_set_x2apic_state(struct vmctx *ctx, int vcpu, enum x2apic_state state) argument 995 vcpu_reset(struct vmctx *vmctx, int vcpu) argument 1181 vm_gla2gpa(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint64_t gla, int prot, uint64_t *gpa, int *fault) argument 1206 vm_copy_setup(struct vmctx *ctx, int vcpu, struct vm_guest_paging *paging, uint64_t gla, size_t len, int prot, struct iovec *iov, int iovcnt, int *fault) argument 1244 vm_copy_teardown(struct vmctx *ctx, int vcpu, struct iovec *iov, int iovcnt) argument 1251 vm_copyin(struct vmctx *ctx, int vcpu, struct iovec *iov, void *vp, size_t len) argument 1271 vm_copyout(struct vmctx *ctx, int vcpu, const void *vp, struct iovec *iov, size_t len) argument 1321 vm_activate_cpu(struct vmctx *ctx, int vcpu) argument 1333 vm_get_intinfo(struct vmctx *ctx, int vcpu, uint64_t *info1, uint64_t *info2) argument 1349 vm_set_intinfo(struct vmctx *ctx, int vcpu, uint64_t info1) argument 1414 vm_restart_instruction(void *arg, int vcpu) argument [all...] |
/freebsd-10-stable/sys/amd64/vmm/ |
H A D | vmm.c | 90 * (a) allocated when vcpu is created 91 * (i) initialized when vcpu is created and when it is reinitialized 92 * (o) initialized the first time the vcpu is created 95 struct vcpu { struct 97 enum vcpu_state state; /* (o) vcpu state */ 98 int hostcpu; /* (o) vcpu's host cpu */ 99 int reqidle; /* (i) request vcpu to idle */ 117 #define vcpu_lock_init(v) mtx_init(&((v)->mtx), "vcpu lock", 0, MTX_SPIN) 167 struct vcpu vcpu[VM_MAXCP member in struct:vm 258 struct vcpu *vcpu = &vm->vcpu[i]; local 270 struct vcpu *vcpu; local 309 struct vcpu *vcpu; local 958 vm_get_register(struct vm *vm, int vcpu, int reg, uint64_t *retval) argument 973 struct vcpu *vcpu; local 1026 vm_get_seg_desc(struct vm *vm, int vcpu, int reg, struct seg_desc *desc) argument 1040 vm_set_seg_desc(struct vm *vm, int vcpu, int reg, struct seg_desc *desc) argument 1053 restore_guest_fpustate(struct vcpu *vcpu) argument 1075 save_guest_fpustate(struct vcpu *vcpu) argument 1099 struct vcpu *vcpu; local 1250 struct vcpu *vcpu; local 1334 struct vcpu *vcpu; local 1375 struct vcpu *vcpu; local 1453 struct vcpu *vcpu; local 1504 struct vcpu *vcpu = &vm->vcpu[vcpuid]; local 1599 struct vcpu *vcpu; local 1702 struct vcpu *vcpu; local 1744 struct vcpu *vcpu; local 1866 vcpu_exception_intinfo(struct vcpu *vcpu) argument 1884 struct vcpu *vcpu; local 1926 struct vcpu *vcpu; local 1941 struct vcpu *vcpu; local 2034 struct vcpu *vcpu; local 2049 struct vcpu *vcpu; local 2062 struct vcpu *vcpu; local 2081 struct vcpu *vcpu; local 2096 struct vcpu *vcpu; local 2109 struct vcpu *vcpu; local 2124 vm_get_capability(struct vm *vm, int vcpu, int type, int *retval) argument 2136 vm_set_capability(struct vm *vm, int vcpu, int type, int val) argument 2221 struct vcpu *vcpu; local 2238 struct vcpu *vcpu; local 2326 vcpu_notify_event_locked(struct vcpu *vcpu, bool lapic_intr) argument 2359 struct vcpu *vcpu = &vm->vcpu[vcpuid]; local 2574 vm_get_rescnt(struct vm *vm, int vcpu, struct vmm_stat_type *stat) argument 2584 vm_get_wiredcnt(struct vm *vm, int vcpu, struct vmm_stat_type *stat) argument [all...] |
H A D | vmm_stat.h | 46 typedef void (*vmm_stat_func_t)(struct vm *vm, int vcpu, 91 int vmm_stat_copy(struct vm *vm, int vcpu, int *num_stats, uint64_t *buf); 95 vmm_stat_array_incr(struct vm *vm, int vcpu, struct vmm_stat_type *vst, argument 101 stats = vcpu_stats(vm, vcpu); 109 vmm_stat_array_set(struct vm *vm, int vcpu, struct vmm_stat_type *vst, argument 115 stats = vcpu_stats(vm, vcpu); 123 vmm_stat_incr(struct vm *vm, int vcpu, struct vmm_stat_type *vst, uint64_t x) argument 127 vmm_stat_array_incr(vm, vcpu, vst, 0, x); 132 vmm_stat_set(struct vm *vm, int vcpu, struct vmm_stat_type *vst, uint64_t val) argument 136 vmm_stat_array_set(vm, vcpu, vs [all...] |
H A D | vmm_stat.c | 83 vmm_stat_copy(struct vm *vm, int vcpu, int *num_stats, uint64_t *buf) argument 89 if (vcpu < 0 || vcpu >= VM_MAXCPU) 96 (*vst->func)(vm, vcpu, vst); 100 stats = vcpu_stats(vm, vcpu); 150 VMM_STAT(VCPU_MIGRATIONS, "vcpu migration across host cpus");
|
/freebsd-10-stable/sys/amd64/vmm/intel/ |
H A D | vmx.c | 194 static int vmx_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc); 195 static int vmx_getreg(void *arg, int vcpu, int reg, uint64_t *retval); 415 * Allocate a unique VPID for each vcpu from the unit number allocator. 889 panic("vmx_vminit: vmclear error %d on vcpu %d\n", 963 vmx_handle_cpuid(struct vm *vm, int vcpu, struct vmxctx *vmxctx) argument 969 handled = x86_emulate_cpuid(vm, vcpu, 978 vmx_run_trace(struct vmx *vmx, int vcpu) argument 981 VCPU_CTR1(vmx->vm, vcpu, "Resume execution at %#lx", vmcs_guest_rip()); 986 vmx_exit_trace(struct vmx *vmx, int vcpu, uint64_t rip, uint32_t exit_reason, argument 990 VCPU_CTR3(vmx->vm, vcpu, " 997 vmx_astpending_trace(struct vmx *vmx, int vcpu, uint64_t rip) argument 1011 vmx_invvpid(struct vmx *vmx, int vcpu, pmap_t pmap, int running) argument 1068 vmx_set_pcpu_defaults(struct vmx *vmx, int vcpu, pmap_t pmap) argument 1092 vmx_set_int_window_exiting(struct vmx *vmx, int vcpu) argument 1103 vmx_clear_int_window_exiting(struct vmx *vmx, int vcpu) argument 1114 vmx_set_nmi_window_exiting(struct vmx *vmx, int vcpu) argument 1125 vmx_clear_nmi_window_exiting(struct vmx *vmx, int vcpu) argument 1136 vmx_set_tsc_offset(struct vmx *vmx, int vcpu, uint64_t offset) argument 1157 vmx_inject_nmi(struct vmx *vmx, int vcpu) argument 1183 vmx_inject_interrupts(struct vmx *vmx, int vcpu, struct vlapic *vlapic, uint64_t guestrip) argument 1412 vmx_emulate_xsetbv(struct vmx *vmx, int vcpu, struct vm_exit *vmexit) argument 1488 vmx_get_guest_reg(struct vmx *vmx, int vcpu, int ident) argument 1533 vmx_set_guest_reg(struct vmx *vmx, int vcpu, int ident, uint64_t regval) argument 1594 vmx_emulate_cr0_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 1632 vmx_emulate_cr4_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 1652 vmx_emulate_cr8_access(struct vmx *vmx, int vcpu, uint64_t exitqual) argument 2091 vmx_exit_process(struct vmx *vmx, int vcpu, struct vm_exit *vmexit) argument 2574 vmx_run(void *arg, int vcpu, register_t rip, pmap_t pmap, struct vm_eventinfo *evinfo) argument 2802 vmx_get_intr_shadow(struct vmx *vmx, int vcpu, int running, uint64_t *retval) argument 2814 vmx_modify_intr_shadow(struct vmx *vmx, int vcpu, int running, uint64_t val) argument 2863 vmx_getreg(void *arg, int vcpu, int reg, uint64_t *retval) argument 2882 vmx_setreg(void *arg, int vcpu, int reg, uint64_t val) argument 2945 vmx_getdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 2958 vmx_setdesc(void *arg, int vcpu, int reg, struct seg_desc *desc) argument 2971 vmx_getcap(void *arg, int vcpu, int type, int *retval) argument 3013 vmx_setcap(void *arg, int vcpu, int type, int val) argument [all...] |
/freebsd-10-stable/sys/amd64/vmm/io/ |
H A D | ppt.h | 35 int ppt_setup_msi(struct vm *vm, int vcpu, int bus, int slot, int func, 37 int ppt_setup_msix(struct vm *vm, int vcpu, int bus, int slot, int func,
|
/freebsd-10-stable/sys/xen/interface/ |
H A D | event_channel.h | 118 * vcpu. 120 * 1. Virtual IRQs are classified as per-vcpu or global. See the VIRQ list 124 * 3. Per-vcpu VIRQs may be bound to at most one event channel per vcpu. 125 * The allocated event channel is bound to the specified vcpu and the 131 uint32_t vcpu; member in struct:evtchn_bind_virq 156 * 1. The allocated event channel is bound to the specified vcpu. The binding 160 uint32_t vcpu; member in struct:evtchn_bind_ipi 207 uint32_t vcpu; /* VCPU to which this channel is bound. */ member in struct:evtchn_status 223 * EVTCHNOP_bind_vcpu: Specify which vcpu 237 uint32_t vcpu; member in struct:evtchn_bind_vcpu [all...] |
/freebsd-10-stable/lib/libkse/thread/ |
H A D | thr_concurrency.c | 165 int vcpu; local 169 len = sizeof(vcpu); 170 ret = sysctlbyname("kern.threads.virtual_cpu", &vcpu, &len, NULL, 0); 171 if (ret == 0 && vcpu > 0) 172 ret = _thr_setconcurrency(vcpu);
|