/linux-master/arch/riscv/include/asm/ |
H A D | kvm_vcpu_insn.h | 10 struct kvm_run; 36 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 37 int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 40 int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run, 43 int kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run, 46 int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
H A D | kvm_vcpu_sbi.h | 47 int (*handler)(struct kvm_vcpu *vcpu, struct kvm_run *run, 54 void kvm_riscv_vcpu_sbi_forward(struct kvm_vcpu *vcpu, struct kvm_run *run); 56 struct kvm_run *run, 58 int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 70 int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
/linux-master/tools/testing/selftests/kvm/lib/s390x/ |
H A D | ucall.c | 11 struct kvm_run *run = vcpu->run;
|
H A D | diag318_test_handler.c | 30 struct kvm_run *run;
|
/linux-master/tools/testing/selftests/kvm/lib/riscv/ |
H A D | ucall.c | 15 struct kvm_run *run = vcpu->run;
|
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | ucall.c | 47 struct kvm_run *run = vcpu->run;
|
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | ucall.c | 24 struct kvm_run *run = vcpu->run;
|
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | sync_regs_test.c | 9 * in kvm_run.s.regs when kvm_valid_regs and kvm_dirty_regs are toggled. 94 struct kvm_run *run = (struct kvm_run *)arg; 117 struct kvm_run *run = (struct kvm_run *)arg; 139 struct kvm_run *run = (struct kvm_run *)arg; 163 struct kvm_run *run; 212 struct kvm_run *run = vcpu->run; 233 struct kvm_run *ru [all...] |
H A D | vmx_close_while_nested_test.c | 63 volatile struct kvm_run *run = vcpu->run;
|
H A D | flds_emulation.h | 22 struct kvm_run *run = vcpu->run;
|
H A D | vmx_invalid_nested_guest_state.c | 58 struct kvm_run *run;
|
H A D | userspace_io_test.c | 57 struct kvm_run *run;
|
H A D | hyperv_extended_hypercalls.c | 41 struct kvm_run *run;
|
/linux-master/arch/riscv/kvm/ |
H A D | vcpu_sbi_replace.c | 17 static int kvm_sbi_ext_time_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, 45 static int kvm_sbi_ext_ipi_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, 83 static int kvm_sbi_ext_rfence_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, 138 struct kvm_run *run, 180 struct kvm_run *run,
|
H A D | vcpu_sbi_base.c | 16 static int kvm_sbi_ext_base_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, 76 struct kvm_run *run,
|
H A D | vcpu_insn.c | 150 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn); 153 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 169 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 199 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) 243 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) 263 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) 306 /* Update CSR details in kvm_run struct */ 383 static int system_opcode_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 423 int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 466 int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *ru [all...] |
H A D | vcpu_exit.c | 13 static int gstage_page_fault(struct kvm_vcpu *vcpu, struct kvm_run *run, 172 int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
H A D | vcpu_sbi_hsm.c | 69 static int kvm_sbi_ext_hsm_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
H A D | vcpu_sbi_pmu.c | 16 static int kvm_sbi_ext_pmu_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
/linux-master/tools/testing/selftests/kvm/s390x/ |
H A D | sync_regs_test.c | 78 struct kvm_run *run = vcpu->run; 99 struct kvm_run *run = vcpu->run; 120 struct kvm_run *run = vcpu->run; 146 struct kvm_run *run = vcpu->run; 185 struct kvm_run *run = vcpu->run;
|
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_host.h | 61 int (*enter_guest)(struct kvm_run *run, struct kvm_vcpu *vcpu); 91 EMULATE_DO_MMIO, /* kvm_run filled with MMIO request */ 112 int (*handle_exit)(struct kvm_run *run, struct kvm_vcpu *vcpu); 242 int kvm_enter_guest(struct kvm_run *run, struct kvm_vcpu *vcpu);
|
H A D | kvm_vcpu.h | 44 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run); 45 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
/linux-master/include/linux/ |
H A D | kvm_types.h | 14 struct kvm_run;
|
/linux-master/arch/arm64/kvm/ |
H A D | mmio.c | 94 struct kvm_run *run = vcpu->run; 125 struct kvm_run *run = vcpu->run; 178 /* Now prepare kvm_run for the potential return to userland. */
|
/linux-master/arch/loongarch/kvm/ |
H A D | exit.c | 107 int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu) 170 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run) 216 struct kvm_run *run = vcpu->run; 320 struct kvm_run *run = vcpu->run; 430 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run) 474 struct kvm_run *run = vcpu->run; 583 struct kvm_run *run = vcpu->run; 635 struct kvm_run *run = vcpu->run;
|