Lines Matching refs:vie

236 /* struct vie.mod */
242 /* struct vie.rm */
285 vie_calc_bytereg(struct vie *vie, enum vm_reg_name *reg, int *lhbr)
288 *reg = gpr_map[vie->reg];
302 if (!vie->rex_present) {
303 if (vie->reg & 0x4) {
305 *reg = gpr_map[vie->reg & 0x3];
311 vie_read_bytereg(void *vm, int vcpuid, struct vie *vie, uint8_t *rval)
317 vie_calc_bytereg(vie, &reg, &lhbr);
332 vie_write_bytereg(void *vm, int vcpuid, struct vie *vie, uint8_t byte)
338 vie_calc_bytereg(vie, &reg, &lhbr);
494 emulate_mov(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
502 size = vie->opsize;
505 switch (vie->op.op_byte) {
513 error = vie_read_bytereg(vm, vcpuid, vie, &byte);
524 reg = gpr_map[vie->reg];
540 error = vie_write_bytereg(vm, vcpuid, vie, val);
551 reg = gpr_map[vie->reg];
588 error = memwrite(vm, vcpuid, gpa, vie->immediate, size, arg);
597 val = vie->immediate & size2mask[size];
608 emulate_movx(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
616 size = vie->opsize;
619 switch (vie->op.op_byte) {
636 reg = gpr_map[vie->reg];
656 reg = gpr_map[vie->reg];
679 reg = gpr_map[vie->reg];
697 get_gla(void *vm, int vcpuid, struct vie *vie, struct vm_guest_paging *paging,
750 emulate_movs(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
763 opsize = (vie->op.op_byte == 0xA4) ? 1 : vie->opsize;
774 repeat = vie->repz_present | vie->repnz_present;
784 if ((rcx & vie_size2mask(vie->addrsize)) == 0) {
806 seg = vie->segment_override ? vie->segment_register : VM_REG_GUEST_DS;
807 error = get_gla(vm, vcpuid, vie, paging, opsize, vie->addrsize,
832 error = get_gla(vm, vcpuid, vie, paging, opsize, vie->addrsize,
906 vie->addrsize);
910 vie->addrsize);
916 rcx, vie->addrsize);
922 if ((rcx & vie_size2mask(vie->addrsize)) != 0)
932 emulate_stos(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
940 opsize = (vie->op.op_byte == 0xAA) ? 1 : vie->opsize;
941 repeat = vie->repz_present | vie->repnz_present;
951 if ((rcx & vie_size2mask(vie->addrsize)) == 0)
974 vie->addrsize);
980 rcx, vie->addrsize);
986 if ((rcx & vie_size2mask(vie->addrsize)) != 0)
994 emulate_and(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1001 size = vie->opsize;
1004 switch (vie->op.op_byte) {
1016 reg = gpr_map[vie->reg];
1054 result = val1 & vie->immediate;
1082 emulate_or(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1089 size = vie->opsize;
1092 switch (vie->op.op_byte) {
1104 reg = gpr_map[vie->reg];
1142 result = val1 | vie->immediate;
1170 emulate_cmp(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1177 size = vie->opsize;
1178 switch (vie->op.op_byte) {
1197 reg = gpr_map[vie->reg];
1207 if (vie->op.op_byte == 0x3B) {
1238 if (vie->op.op_byte == 0x80)
1246 rflags2 = getcc(size, op1, vie->immediate);
1262 emulate_test(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1268 size = vie->opsize;
1271 switch (vie->op.op_byte) {
1283 if ((vie->reg & 7) != 0)
1290 rflags2 = getandflags(size, op1, vie->immediate);
1311 emulate_add(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1318 size = vie->opsize;
1321 switch (vie->op.op_byte) {
1332 reg = gpr_map[vie->reg];
1367 emulate_sub(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1374 size = vie->opsize;
1377 switch (vie->op.op_byte) {
1388 reg = gpr_map[vie->reg];
1423 emulate_stack_op(void *vm, int vcpuid, uint64_t mmio_gpa, struct vie *vie,
1437 size = vie->opsize;
1438 pushop = (vie->op.op_type == VIE_OP_TYPE_PUSH) ? 1 : 0;
1454 size = vie->opsize_override ? 2 : 8;
1525 emulate_push(void *vm, int vcpuid, uint64_t mmio_gpa, struct vie *vie,
1537 if ((vie->reg & 7) != 6)
1540 error = emulate_stack_op(vm, vcpuid, mmio_gpa, vie, paging, memread,
1546 emulate_pop(void *vm, int vcpuid, uint64_t mmio_gpa, struct vie *vie,
1558 if ((vie->reg & 7) != 0)
1561 error = emulate_stack_op(vm, vcpuid, mmio_gpa, vie, paging, memread,
1567 emulate_group1(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1573 switch (vie->reg & 7) {
1575 error = emulate_or(vm, vcpuid, gpa, vie,
1579 error = emulate_and(vm, vcpuid, gpa, vie,
1583 error = emulate_cmp(vm, vcpuid, gpa, vie,
1595 emulate_bittest(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1607 if ((vie->reg & 7) != 4)
1613 error = memread(vm, vcpuid, gpa, &val, vie->opsize, memarg);
1621 bitmask = vie->opsize * 8 - 1;
1622 bitoff = vie->immediate & bitmask;
1637 emulate_twob_group15(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1643 switch (vie->reg & 7) {
1645 if (vie->mod == 0x3) {
1668 vmm_emulate_instruction(void *vm, int vcpuid, uint64_t gpa, struct vie *vie,
1674 if (!vie->decoded)
1677 switch (vie->op.op_type) {
1679 error = emulate_group1(vm, vcpuid, gpa, vie, paging, memread,
1683 error = emulate_pop(vm, vcpuid, gpa, vie, paging, memread,
1687 error = emulate_push(vm, vcpuid, gpa, vie, paging, memread,
1691 error = emulate_cmp(vm, vcpuid, gpa, vie,
1695 error = emulate_mov(vm, vcpuid, gpa, vie,
1700 error = emulate_movx(vm, vcpuid, gpa, vie,
1704 error = emulate_movs(vm, vcpuid, gpa, vie, paging, memread,
1708 error = emulate_stos(vm, vcpuid, gpa, vie, paging, memread,
1712 error = emulate_and(vm, vcpuid, gpa, vie,
1716 error = emulate_or(vm, vcpuid, gpa, vie,
1720 error = emulate_sub(vm, vcpuid, gpa, vie,
1724 error = emulate_bittest(vm, vcpuid, gpa, vie,
1728 error = emulate_twob_group15(vm, vcpuid, gpa, vie,
1732 error = emulate_add(vm, vcpuid, gpa, vie, memread,
1736 error = emulate_test(vm, vcpuid, gpa, vie,
1899 vie_init(struct vie *vie, const char *inst_bytes, int inst_length)
1904 bzero(vie, sizeof(struct vie));
1906 vie->base_register = VM_REG_LAST;
1907 vie->index_register = VM_REG_LAST;
1908 vie->segment_register = VM_REG_LAST;
1911 bcopy(inst_bytes, vie->inst, inst_length);
1912 vie->num_valid = inst_length;
2143 uint64_t rip, int inst_length, struct vie *vie, int *faultptr)
2157 vm_copyin(vm, vcpuid, copyinfo, vie->inst, inst_length);
2159 vie->num_valid = inst_length;
2164 vie_peek(struct vie *vie, uint8_t *x)
2167 if (vie->num_processed < vie->num_valid) {
2168 *x = vie->inst[vie->num_processed];
2175 vie_advance(struct vie *vie)
2178 vie->num_processed++;
2211 decode_prefixes(struct vie *vie, enum vm_cpu_mode cpu_mode, int cs_d)
2216 if (vie_peek(vie, &x))
2220 vie->opsize_override = 1;
2222 vie->addrsize_override = 1;
2224 vie->repz_present = 1;
2226 vie->repnz_present = 1;
2227 else if (segment_override(x, &vie->segment_register))
2228 vie->segment_override = 1;
2232 vie_advance(vie);
2244 vie->rex_present = 1;
2245 vie->rex_w = x & 0x8 ? 1 : 0;
2246 vie->rex_r = x & 0x4 ? 1 : 0;
2247 vie->rex_x = x & 0x2 ? 1 : 0;
2248 vie->rex_b = x & 0x1 ? 1 : 0;
2249 vie_advance(vie);
2260 vie->addrsize = vie->addrsize_override ? 4 : 8;
2261 if (vie->rex_w)
2262 vie->opsize = 8;
2263 else if (vie->opsize_override)
2264 vie->opsize = 2;
2266 vie->opsize = 4;
2269 vie->addrsize = vie->addrsize_override ? 2 : 4;
2270 vie->opsize = vie->opsize_override ? 2 : 4;
2273 vie->addrsize = vie->addrsize_override ? 4 : 2;
2274 vie->opsize = vie->opsize_override ? 4 : 2;
2280 decode_two_byte_opcode(struct vie *vie)
2284 if (vie_peek(vie, &x))
2287 vie->op = two_byte_opcodes[x];
2289 if (vie->op.op_type == VIE_OP_TYPE_NONE)
2292 vie_advance(vie);
2297 decode_opcode(struct vie *vie)
2301 if (vie_peek(vie, &x))
2304 vie->op = one_byte_opcodes[x];
2306 if (vie->op.op_type == VIE_OP_TYPE_NONE)
2309 vie_advance(vie);
2311 if (vie->op.op_type == VIE_OP_TYPE_TWO_BYTE)
2312 return (decode_two_byte_opcode(vie));
2318 decode_modrm(struct vie *vie, enum vm_cpu_mode cpu_mode)
2322 if (vie->op.op_flags & VIE_OP_F_NO_MODRM)
2328 if (vie_peek(vie, &x))
2331 vie->mod = (x >> 6) & 0x3;
2332 vie->rm = (x >> 0) & 0x7;
2333 vie->reg = (x >> 3) & 0x7;
2340 if (vie->mod == VIE_MOD_DIRECT)
2343 if ((vie->mod == VIE_MOD_INDIRECT && vie->rm == VIE_RM_DISP32) ||
2344 (vie->mod != VIE_MOD_DIRECT && vie->rm == VIE_RM_SIB)) {
2358 vie->rm |= (vie->rex_b << 3);
2361 vie->reg |= (vie->rex_r << 3);
2364 if (vie->mod != VIE_MOD_DIRECT && vie->rm == VIE_RM_SIB)
2367 vie->base_register = gpr_map[vie->rm];
2369 switch (vie->mod) {
2371 vie->disp_bytes = 1;
2374 vie->disp_bytes = 4;
2377 if (vie->rm == VIE_RM_DISP32) {
2378 vie->disp_bytes = 4;
2387 vie->base_register = VM_REG_GUEST_RIP;
2389 vie->base_register = VM_REG_LAST;
2395 vie_advance(vie);
2401 decode_sib(struct vie *vie)
2406 if (vie->mod == VIE_MOD_DIRECT || vie->rm != VIE_RM_SIB)
2409 if (vie_peek(vie, &x))
2413 vie->ss = (x >> 6) & 0x3;
2414 vie->index = (x >> 3) & 0x7;
2415 vie->base = (x >> 0) & 0x7;
2418 vie->index |= vie->rex_x << 3;
2419 vie->base |= vie->rex_b << 3;
2421 switch (vie->mod) {
2423 vie->disp_bytes = 1;
2426 vie->disp_bytes = 4;
2430 if (vie->mod == VIE_MOD_INDIRECT &&
2431 (vie->base == 5 || vie->base == 13)) {
2440 vie->disp_bytes = 4;
2442 vie->base_register = gpr_map[vie->base];
2452 if (vie->index != 4)
2453 vie->index_register = gpr_map[vie->index];
2456 if (vie->index_register < VM_REG_LAST)
2457 vie->scale = 1 << vie->ss;
2459 vie_advance(vie);
2465 decode_displacement(struct vie *vie)
2476 if ((n = vie->disp_bytes) == 0)
2483 if (vie_peek(vie, &x))
2487 vie_advance(vie);
2491 vie->displacement = u.signed8; /* sign-extended */
2493 vie->displacement = u.signed32; /* sign-extended */
2499 decode_immediate(struct vie *vie)
2511 if (vie->op.op_flags & VIE_OP_F_IMM) {
2519 if (vie->opsize == 4 || vie->opsize == 8)
2520 vie->imm_bytes = 4;
2522 vie->imm_bytes = 2;
2523 } else if (vie->op.op_flags & VIE_OP_F_IMM8) {
2524 vie->imm_bytes = 1;
2527 if ((n = vie->imm_bytes) == 0)
2534 if (vie_peek(vie, &x))
2538 vie_advance(vie);
2543 vie->immediate = u.signed8;
2545 vie->immediate = u.signed16;
2547 vie->immediate = u.signed32;
2553 decode_moffset(struct vie *vie)
2562 if ((vie->op.op_flags & VIE_OP_F_MOFFSET) == 0)
2569 n = vie->addrsize;
2574 if (vie_peek(vie, &x))
2578 vie_advance(vie);
2580 vie->displacement = u.u64;
2589 verify_gla(struct vm *vm, int cpuid, uint64_t gla, struct vie *vie,
2602 if (vie->base_register != VM_REG_LAST) {
2603 error = vm_get_register(vm, cpuid, vie->base_register, &base);
2606 error, vie->base_register);
2614 if (vie->base_register == VM_REG_GUEST_RIP)
2615 base += vie->num_processed;
2619 if (vie->index_register != VM_REG_LAST) {
2620 error = vm_get_register(vm, cpuid, vie->index_register, &idx);
2623 error, vie->index_register);
2641 if (vie->segment_override)
2642 seg = vie->segment_register;
2643 else if (vie->base_register == VM_REG_GUEST_RSP ||
2644 vie->base_register == VM_REG_GUEST_RBP)
2656 vie->segment_register);
2662 gla2 = segbase + base + vie->scale * idx + vie->displacement;
2663 gla2 &= size2mask[vie->addrsize];
2668 segbase, base, vie->scale, idx, vie->displacement,
2678 enum vm_cpu_mode cpu_mode, int cs_d, struct vie *vie)
2681 if (decode_prefixes(vie, cpu_mode, cs_d))
2684 if (decode_opcode(vie))
2687 if (decode_modrm(vie, cpu_mode))
2690 if (decode_sib(vie))
2693 if (decode_displacement(vie))
2696 if (decode_immediate(vie))
2699 if (decode_moffset(vie))
2702 if ((vie->op.op_flags & VIE_OP_F_NO_GLA_VERIFICATION) == 0) {
2703 if (verify_gla(vm, cpuid, gla, vie, cpu_mode))
2707 vie->decoded = 1; /* success */