/linux-master/arch/arm64/kvm/ |
H A D | arm.c | 918 struct kvm_run *run = vcpu->run; 973 * return with return value 0 and with the kvm_run structure filled in with the 978 struct kvm_run *run = vcpu->run;
|
H A D | handle_exit.c | 168 struct kvm_run *run = vcpu->run; 324 struct kvm_run *run = vcpu->run;
|
H A D | hypercalls.c | 247 struct kvm_run *run = vcpu->run;
|
H A D | mmio.c | 94 struct kvm_run *run = vcpu->run; 125 struct kvm_run *run = vcpu->run; 178 /* Now prepare kvm_run for the potential return to userland. */
|
H A D | psci.c | 210 struct kvm_run *run = vcpu->run;
|
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_csr.h | 184 int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu);
|
H A D | kvm_host.h | 61 int (*enter_guest)(struct kvm_run *run, struct kvm_vcpu *vcpu); 91 EMULATE_DO_MMIO, /* kvm_run filled with MMIO request */ 112 int (*handle_exit)(struct kvm_run *run, struct kvm_vcpu *vcpu); 242 int kvm_enter_guest(struct kvm_run *run, struct kvm_vcpu *vcpu);
|
H A D | kvm_vcpu.h | 44 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run); 45 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
/linux-master/arch/loongarch/kvm/ |
H A D | exit.c | 107 int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu) 170 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run) 216 struct kvm_run *run = vcpu->run; 320 struct kvm_run *run = vcpu->run; 430 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run) 474 struct kvm_run *run = vcpu->run; 583 struct kvm_run *run = vcpu->run; 635 struct kvm_run *run = vcpu->run;
|
H A D | vcpu.c | 121 static int kvm_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu) 1153 struct kvm_run *run = vcpu->run;
|
/linux-master/arch/mips/kvm/ |
H A D | emulate.c | 975 struct kvm_run *run = vcpu->run; 1271 struct kvm_run *run = vcpu->run; 1511 struct kvm_run *run = vcpu->run;
|
H A D | mips.c | 1173 struct kvm_run *run = vcpu->run; 1196 kvm_debug("kvm_mips_handle_exit: cause: %#x, PC: %p, kvm_run: %p, kvm_vcpu: %p\n",
|
H A D | vz.c | 1625 struct kvm_run *run = vcpu->run; 1672 struct kvm_run *run = vcpu->run;
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv.c | 1630 struct kvm_run *run = vcpu->run; 4584 struct kvm_run *run = vcpu->run; 4716 struct kvm_run *run = vcpu->run; 4920 struct kvm_run *run = vcpu->run;
|
H A D | book3s_pr.c | 1145 struct kvm_run *run = vcpu->run;
|
H A D | booke.c | 866 struct kvm_run *run = vcpu->run; 1013 struct kvm_run *run = vcpu->run;
|
H A D | powerpc.c | 1141 struct kvm_run *run = vcpu->run; 1253 struct kvm_run *run = vcpu->run; 1339 struct kvm_run *run = vcpu->run; 1484 struct kvm_run *run = vcpu->run; 1663 struct kvm_run *run = vcpu->run; 1795 struct kvm_run *run = vcpu->run;
|
/linux-master/arch/riscv/include/asm/ |
H A D | kvm_host.h | 358 int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
H A D | kvm_vcpu_insn.h | 10 struct kvm_run; 36 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 37 int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 40 int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run, 43 int kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run, 46 int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
H A D | kvm_vcpu_sbi.h | 47 int (*handler)(struct kvm_vcpu *vcpu, struct kvm_run *run, 54 void kvm_riscv_vcpu_sbi_forward(struct kvm_vcpu *vcpu, struct kvm_run *run); 56 struct kvm_run *run, 58 int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 70 int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
/linux-master/arch/riscv/kvm/ |
H A D | aia_imsic.c | 737 struct kvm_run *run = vcpu->run;
|
H A D | vcpu.c | 682 struct kvm_run *run = vcpu->run;
|
H A D | vcpu_exit.c | 13 static int gstage_page_fault(struct kvm_vcpu *vcpu, struct kvm_run *run, 172 int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
|
H A D | vcpu_insn.c | 150 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn); 153 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 169 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 199 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) 243 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) 263 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) 306 /* Update CSR details in kvm_run struct */ 383 static int system_opcode_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 423 int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 466 int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *ru [all...] |
H A D | vcpu_sbi.c | 115 void kvm_riscv_vcpu_sbi_forward(struct kvm_vcpu *vcpu, struct kvm_run *run) 135 struct kvm_run *run, 152 int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run) 416 int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run)
|