/linux-master/arch/riscv/kernel/ |
H A D | asm-offsets.c | 129 OFFSET(KVM_ARCH_GUEST_ZERO, kvm_vcpu_arch, guest_context.zero); 130 OFFSET(KVM_ARCH_GUEST_RA, kvm_vcpu_arch, guest_context.ra); 131 OFFSET(KVM_ARCH_GUEST_SP, kvm_vcpu_arch, guest_context.sp); 132 OFFSET(KVM_ARCH_GUEST_GP, kvm_vcpu_arch, guest_context.gp); 133 OFFSET(KVM_ARCH_GUEST_TP, kvm_vcpu_arch, guest_context.tp); 134 OFFSET(KVM_ARCH_GUEST_T0, kvm_vcpu_arch, guest_context.t0); 135 OFFSET(KVM_ARCH_GUEST_T1, kvm_vcpu_arch, guest_context.t1); 136 OFFSET(KVM_ARCH_GUEST_T2, kvm_vcpu_arch, guest_context.t2); 137 OFFSET(KVM_ARCH_GUEST_S0, kvm_vcpu_arch, guest_context.s0); 138 OFFSET(KVM_ARCH_GUEST_S1, kvm_vcpu_arch, guest_contex [all...] |
/linux-master/arch/mips/kernel/ |
H A D | asm-offsets.c | 362 OFFSET(VCPU_FPR0, kvm_vcpu_arch, fpu.fpr[0]); 363 OFFSET(VCPU_FPR1, kvm_vcpu_arch, fpu.fpr[1]); 364 OFFSET(VCPU_FPR2, kvm_vcpu_arch, fpu.fpr[2]); 365 OFFSET(VCPU_FPR3, kvm_vcpu_arch, fpu.fpr[3]); 366 OFFSET(VCPU_FPR4, kvm_vcpu_arch, fpu.fpr[4]); 367 OFFSET(VCPU_FPR5, kvm_vcpu_arch, fpu.fpr[5]); 368 OFFSET(VCPU_FPR6, kvm_vcpu_arch, fpu.fpr[6]); 369 OFFSET(VCPU_FPR7, kvm_vcpu_arch, fpu.fpr[7]); 370 OFFSET(VCPU_FPR8, kvm_vcpu_arch, fpu.fpr[8]); 371 OFFSET(VCPU_FPR9, kvm_vcpu_arch, fp [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_host.h | 103 struct kvm_vcpu_arch { struct 183 static inline bool kvm_guest_has_fpu(struct kvm_vcpu_arch *arch) 188 static inline bool kvm_guest_has_lsx(struct kvm_vcpu_arch *arch) 193 static inline bool kvm_guest_has_lasx(struct kvm_vcpu_arch *arch) 211 static inline void update_pc(struct kvm_vcpu_arch *arch) 223 static inline bool kvm_is_ifetch_fault(struct kvm_vcpu_arch *arch)
|
/linux-master/arch/loongarch/kernel/ |
H A D | asm-offsets.c | 298 OFFSET(VCPU_FCC, kvm_vcpu_arch, fpu.fcc); 299 OFFSET(VCPU_FCSR0, kvm_vcpu_arch, fpu.fcsr); 307 OFFSET(KVM_ARCH_HSP, kvm_vcpu_arch, host_sp); 308 OFFSET(KVM_ARCH_HTP, kvm_vcpu_arch, host_tp); 309 OFFSET(KVM_ARCH_HPGD, kvm_vcpu_arch, host_pgd); 310 OFFSET(KVM_ARCH_HANDLE_EXIT, kvm_vcpu_arch, handle_exit); 311 OFFSET(KVM_ARCH_HEENTRY, kvm_vcpu_arch, host_eentry); 312 OFFSET(KVM_ARCH_GEENTRY, kvm_vcpu_arch, guest_eentry); 313 OFFSET(KVM_ARCH_GPC, kvm_vcpu_arch, pc); 314 OFFSET(KVM_ARCH_GGPR, kvm_vcpu_arch, gpr [all...] |
/linux-master/arch/mips/kvm/ |
H A D | entry.c | 190 UASM_i_SW(&p, GPR_SP, offsetof(struct kvm_vcpu_arch, host_stack), GPR_K1); 193 UASM_i_SW(&p, GPR_GP, offsetof(struct kvm_vcpu_arch, host_gp), GPR_K1); 204 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, guest_ebase), GPR_K1); 246 UASM_i_LW(&p, GPR_T0, offsetof(struct kvm_vcpu_arch, pc), GPR_K1); 254 UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, host_pgd), GPR_K1); 304 UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, host_entryhi), 347 UASM_i_LW(&p, i, offsetof(struct kvm_vcpu_arch, gprs[i]), GPR_K1); 352 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, hi), GPR_K1); 355 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, lo), GPR_K1); 360 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, gpr [all...] |
H A D | vz.c | 1081 struct kvm_vcpu_arch *arch = &vcpu->arch; 1216 struct kvm_vcpu_arch *arch = &vcpu->arch; 1298 struct kvm_vcpu_arch *arch = &vcpu->arch;
|
H A D | emulate.c | 45 struct kvm_vcpu_arch *arch = &vcpu->arch;
|
/linux-master/arch/mips/include/asm/ |
H A D | kvm_host.h | 290 struct kvm_vcpu_arch { struct 692 static inline bool kvm_mips_guest_can_have_fpu(struct kvm_vcpu_arch *vcpu) 698 static inline bool kvm_mips_guest_has_fpu(struct kvm_vcpu_arch *vcpu) 704 static inline bool kvm_mips_guest_can_have_msa(struct kvm_vcpu_arch *vcpu) 710 static inline bool kvm_mips_guest_has_msa(struct kvm_vcpu_arch *vcpu) 776 void __kvm_save_fpu(struct kvm_vcpu_arch *vcpu); 777 void __kvm_restore_fpu(struct kvm_vcpu_arch *vcpu); 778 void __kvm_restore_fcsr(struct kvm_vcpu_arch *vcpu); 779 void __kvm_save_msa(struct kvm_vcpu_arch *vcpu); 780 void __kvm_restore_msa(struct kvm_vcpu_arch *vcp [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | kvm_host.h | 178 struct kvm_vcpu_arch { struct 361 void __kvm_riscv_switch_to(struct kvm_vcpu_arch *vcpu_arch);
|
/linux-master/arch/riscv/kvm/ |
H A D | tlb.c | 207 struct kvm_vcpu_arch *varch = &vcpu->arch; 232 struct kvm_vcpu_arch *varch = &vcpu->arch;
|
H A D | vcpu.c | 348 struct kvm_vcpu_arch *v = &vcpu->arch;
|
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_host.h | 567 struct kvm_vcpu_arch { struct 1155 static inline void kvm_arm_pvtime_vcpu_init(struct kvm_vcpu_arch *vcpu_arch) 1160 static inline bool kvm_arm_is_pvtime_enabled(struct kvm_vcpu_arch *vcpu_arch)
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_host.h | 528 struct kvm_vcpu_arch { struct in struct:kvmppc_slb
|
/linux-master/arch/s390/include/asm/ |
H A D | kvm_host.h | 740 struct kvm_vcpu_arch { struct
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_pr.c | 522 struct kvm_vcpu_arch *a = &vcpu->arch;
|
/linux-master/include/linux/ |
H A D | kvm_host.h | 387 struct kvm_vcpu_arch arch;
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 742 struct kvm_vcpu_arch { struct 862 * when "struct kvm_vcpu_arch" is no longer defined in an
|
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 3166 struct kvm_vcpu_arch *vcpu = &v->arch; 3219 struct kvm_vcpu_arch *vcpu = &v->arch;
|