/linux-master/arch/x86/kvm/ |
H A D | mmu.h | 125 int bytes);
|
H A D | smm.c | 283 memset(smram.bytes, 0, sizeof(smram.bytes)); 574 ret = kvm_vcpu_read_guest(vcpu, smbase + 0xfe00, smram.bytes, sizeof(smram));
|
H A D | smm.h | 140 u8 bytes[512]; member in union:kvm_smram
|
H A D | x86.c | 7568 static int kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, argument 7576 while (bytes) { 7579 unsigned toread = min(bytes, (unsigned)PAGE_SIZE - offset); 7591 bytes -= toread; 7601 gva_t addr, void *val, unsigned int bytes, 7617 if (WARN_ON(offset + bytes > PAGE_SIZE)) 7618 bytes = (unsigned)PAGE_SIZE - offset; 7620 offset, bytes); 7628 gva_t addr, void *val, unsigned int bytes, 7640 return kvm_read_guest_virt_helper(addr, val, bytes, vcp 7600 kvm_fetch_guest_virt(struct x86_emulate_ctxt *ctxt, gva_t addr, void *val, unsigned int bytes, struct x86_exception *exception) argument 7627 kvm_read_guest_virt(struct kvm_vcpu *vcpu, gva_t addr, void *val, unsigned int bytes, struct x86_exception *exception) argument 7645 emulator_read_std(struct x86_emulate_ctxt *ctxt, gva_t addr, void *val, unsigned int bytes, struct x86_exception *exception, bool system) argument 7660 kvm_write_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, struct kvm_vcpu *vcpu, u64 access, struct x86_exception *exception) argument 7690 emulator_write_std(struct x86_emulate_ctxt *ctxt, gva_t addr, void *val, unsigned int bytes, struct x86_exception *exception, bool system) argument 7706 kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, gva_t addr, void *val, unsigned int bytes, struct x86_exception *exception) argument 7796 emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, const void *val, int bytes) argument 7820 read_prepare(struct kvm_vcpu *vcpu, void *val, int bytes) argument 7832 read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, void *val, int bytes) argument 7838 write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, void *val, int bytes) argument 7844 write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) argument 7850 read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, void *val, int bytes) argument 7857 write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, void *val, int bytes) argument 7880 emulator_read_write_onepage(unsigned long addr, void *val, unsigned int bytes, struct x86_exception *exception, struct kvm_vcpu *vcpu, const struct read_write_emulator_ops *ops) argument 7931 emulator_read_write(struct x86_emulate_ctxt *ctxt, unsigned long addr, void *val, unsigned int bytes, struct x86_exception *exception, const struct read_write_emulator_ops *ops) argument 7985 emulator_read_emulated(struct x86_emulate_ctxt *ctxt, unsigned long addr, void *val, unsigned int bytes, struct x86_exception *exception) argument 7995 emulator_write_emulated(struct x86_emulate_ctxt *ctxt, unsigned long addr, const void *val, unsigned int bytes, struct x86_exception *exception) argument 8008 emulator_cmpxchg_emulated(struct x86_emulate_ctxt *ctxt, unsigned long addr, const void *old, const void *new, unsigned int bytes, struct x86_exception *exception) argument 13725 kvm_sev_es_mmio_write(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned int bytes, void *data) argument 13764 kvm_sev_es_mmio_read(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned int bytes, void *data) argument [all...] |
H A D | x86.h | 300 gva_t addr, void *val, unsigned int bytes, 304 gva_t addr, void *val, unsigned int bytes, 533 int kvm_sev_es_mmio_write(struct kvm_vcpu *vcpu, gpa_t src, unsigned int bytes, 535 int kvm_sev_es_mmio_read(struct kvm_vcpu *vcpu, gpa_t src, unsigned int bytes,
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 5696 int *bytes) 5706 if (is_pae(vcpu) && *bytes == 4) { 5709 *bytes = 8; 5712 if (*bytes == 4 || *bytes == 8) { 5713 r = kvm_vcpu_read_guest_atomic(vcpu, *gpa, &gentry, *bytes); 5743 int bytes) 5751 * Sometimes, the OS only writes the last one bytes to update status 5754 if (!(offset & (pte_size - 1)) && bytes == 1) 5757 misaligned = (offset ^ (offset + bytes 5695 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, int *bytes) argument 5742 detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, int bytes) argument 5794 kvm_mmu_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, int bytes) argument [all...] |
H A D | page_track.c | 270 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes) argument 285 n->track_write(gpa, new, bytes, n);
|
H A D | page_track.h | 30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes); 42 const u8 *new, int bytes) { } 51 const u8 *new, int bytes) 53 __kvm_page_track_write(vcpu->kvm, gpa, new, bytes); 55 kvm_mmu_track_write(vcpu, gpa, new, bytes); 41 __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes) argument 50 kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, int bytes) argument
|
/linux-master/arch/x86/kvm/svm/ |
H A D | sev.c | 2968 int bytes; local 2975 if (unlikely(check_mul_overflow(count, size, &bytes))) 2978 r = setup_vmgexit_scratch(svm, in, bytes);
|
H A D | svm.c | 4504 u32 bytes; local 4510 bytes = info->dst_bytes; 4513 bytes = info->src_bytes; 4523 bytes = min(bytes, 4u); 4525 exit_info |= bytes << SVM_IOIO_SIZE_SHIFT; 4756 * buffer containing the bytes of the intercepted instruction is 4787 * will be empty if the DecodeAssist microcode cannot fetch bytes for 4804 * gives up and does not fill the instruction bytes buffer. 4822 * retrieve the instruction bytes i [all...] |
/linux-master/arch/x86/lib/ |
H A D | checksum_32.S | 58 jz 10f # Jump if alignment is boundary of 2 bytes. 70 subl $2, %ecx # Alignment uses up two bytes. 71 jae 1f # Jump if we had at least two bytes. 187 addw (%esi), %ax # csumming 2 bytes, 2-aligned 233 # Handle the last 1-3 bytes without jumping 284 subl $2, %ecx # Alignment uses up two bytes. 285 jae 1f # Jump if we had at least two bytes.
|
H A D | insn-eval.c | 44 switch (insn->opcode.bytes[0]) { 649 * Segment descriptors have a size of 8 bytes. Thus, the index is 742 * In protected mode, the limit of the segment descriptor in bytes. 787 * descriptor. In virtual-8086 mode, the default is always two bytes for both 982 /* Ignore bytes that are outside the address size. */ 1223 * address referred by the instruction's ModRM and displacement bytes. Also, the 1284 * @insn: Instruction with ModRM, SIB bytes and displacement 1289 * displacement bytes and segment base address, as applicable. If in protected 1373 * @insn: Instruction struct with ModRM and SIB bytes and displacement 1378 * displacement bytes an 1600 insn_decode_mmio(struct insn *insn, int *bytes) argument [all...] |
H A D | insn.c | 36 /* Verify next sizeof(t) bytes can be on the same instruction */ 64 * Instructions longer than MAX_INSN_SIZE (15 bytes) are invalid 113 * insn_get_prefixes - scan x86 instruction prefix bytes 143 if (prefixes->bytes[i] == b) 148 prefixes->bytes[nb++] = b; 167 if (lb && lb != insn->prefixes.bytes[3]) { 168 if (unlikely(insn->prefixes.bytes[3])) { 170 b = insn->prefixes.bytes[3]; 172 if (prefixes->bytes[i] == lb) 254 * If necessary, first collects any preceding (prefix) bytes [all...] |
/linux-master/arch/x86/mm/ |
H A D | extable.c | 76 if (insn.opcode.bytes[0] != 0x8b) 249 * 1) RSP = %rsp - (IA32_FRED_CONFIG & 0x1c0) // Reserve N*64 bytes 253 * Here is an example with N*64 (N=1) bytes reserved: 282 * FRED stack frame ((N*64) bytes may be reserved between), and it is
|
/linux-master/arch/x86/net/ |
H A D | bpf_jit_comp.c | 24 static u8 *emit_code(u8 *ptr, u32 bytes, unsigned int len) argument 27 *ptr = bytes; 29 *(u16 *)ptr = bytes; 31 *(u32 *)ptr = bytes; 37 #define EMIT(bytes, len) \ 38 do { prog = emit_code(prog, bytes, len); } while (0) 269 /* Maximum number of bytes emitted while JITing one eBPF insn */ 273 /* Number of bytes emit_patch() needs to generate instructions */ 275 /* Number of bytes that will be skipped on tailcall */ 408 * bpf_tail_call helper will skip the first X86_TAIL_CALL_OFFSET bytes [all...] |
H A D | bpf_jit_comp32.c | 51 static u8 *emit_code(u8 *ptr, u32 bytes, unsigned int len) argument 54 *ptr = bytes; 56 *(u16 *)ptr = bytes; 58 *(u32 *)ptr = bytes; 64 #define EMIT(bytes, len) \ 65 do { prog = emit_code(prog, bytes, len); cnt += len; } while (0) 233 /* dst = imm (4 bytes)*/ 265 /* zero out high 4 bytes */ 382 /* Emit 'ror %ax, 8' to swap lower 2 bytes */ 394 /* Emit 'bswap eax' to swap lower 4 bytes */ [all...] |
/linux-master/arch/x86/tools/ |
H A D | insn_decoder_test.c | 63 fprintf(fp, "%s\t.value = %d, bytes[] = {%x, %x, %x, %x},\n", 64 indent, field->value, field->bytes[0], field->bytes[1], 65 field->bytes[2], field->bytes[3]); 161 pr_warn("objdump says %d bytes, but insn_get_length() "
|
H A D | insn_sanity.c | 56 fprintf(fp, "%s\t.value = %d, bytes[] = {%x, %x, %x, %x},\n", 57 indent, field->value, field->bytes[0], field->bytes[1], 58 field->bytes[2], field->bytes[3]); 229 /* Prepare stop bytes with NOPs */
|
/linux-master/arch/xtensa/lib/ |
H A D | memcopy.S | 33 * copying 1 and 2 bytes. 35 * do 16 bytes with a loop, and then finish up with 101 # copy 2 bytes 129 # copy 16 bytes per iteration for word-aligned dst and word-aligned src 153 # copy 8 bytes 166 # copy 4 bytes 175 # copy 2 bytes 195 # copy 16 bytes per iteration for word-aligned dst and unaligned src 233 # copy 8 bytes 245 # copy 4 bytes [all...] |
H A D | memset.S | 40 slli a7, a3, 8 # duplicate character in all bytes of word 56 # set 16 bytes per iteration for word-aligned dst 78 # set 8 bytes 84 # set 4 bytes 89 # set 2 bytes 116 # set 2 bytes
|
H A D | strncpy_user.S | 125 # copy 2 bytes
|
H A D | strnlen_user.S | 83 EX(10f) l32i a9, a4, 4 # load 4 bytes for remaining checks 86 # check two more bytes (bytes 0, 1 of word) 127 EX(10f) l32i a9, a4, 0 # get word with first two bytes of string
|
H A D | usercopy.S | 26 * The returned value is the number of bytes not copied. Implies zero 101 # copy 2 bytes 134 movi a2, 0 # return success for len bytes copied 140 # copy 16 bytes per iteration for word-aligned dst and word-aligned src 167 # copy 8 bytes 176 # copy 4 bytes 183 # copy 2 bytes 194 movi a2, 0 # return success for len bytes copied 205 # copy 16 bytes per iteration for word-aligned dst and unaligned src 247 # copy 8 bytes [all...] |
/linux-master/block/ |
H A D | bfq-cgroup.c | 350 blkg_rwstat_add(&bfqg->stats.bytes, rq->cmd_flags, blk_rq_bytes(rq)); 438 blkg_rwstat_exit(&stats->bytes); 457 if (blkg_rwstat_init(&stats->bytes, gfp) || 1227 u64 sum = blkg_rwstat_total(&bfqg->stats.bytes); 1245 offsetof(struct bfq_group, stats.bytes), &tmp); 1326 .private = offsetof(struct bfq_group, stats.bytes), 1369 .private = offsetof(struct bfq_group, stats.bytes),
|
H A D | bfq-iosched.h | 922 struct blkg_rwstat bytes; member in struct:bfqg_stats
|