Lines Matching defs:vcpu

80 /* x86-specific vcpu->requests bit members */
450 unsigned long (*get_guest_pgd)(struct kvm_vcpu *vcpu);
451 u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
452 int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
453 void (*inject_page_fault)(struct kvm_vcpu *vcpu,
455 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
458 int (*sync_spte)(struct kvm_vcpu *vcpu,
526 struct kvm_vcpu *vcpu;
589 * pmc_in_use only once in a vcpu time slice.
669 /* Hyper-V per vcpu emulation context */
671 struct kvm_vcpu *vcpu;
711 /* Xen HVM per vcpu emulation context */
785 * Paging state of the vcpu
787 * If the vcpu runs in guest mode with two level paging this still saves
890 int (*complete_userspace_io)(struct kvm_vcpu *vcpu);
1617 void (*vcpu_after_set_cpuid)(struct kvm_vcpu *vcpu);
1625 int (*vcpu_create)(struct kvm_vcpu *vcpu);
1626 void (*vcpu_free)(struct kvm_vcpu *vcpu);
1627 void (*vcpu_reset)(struct kvm_vcpu *vcpu, bool init_event);
1629 void (*prepare_switch_to_guest)(struct kvm_vcpu *vcpu);
1630 void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu);
1631 void (*vcpu_put)(struct kvm_vcpu *vcpu);
1633 void (*update_exception_bitmap)(struct kvm_vcpu *vcpu);
1634 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
1635 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
1636 u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg);
1637 void (*get_segment)(struct kvm_vcpu *vcpu,
1639 int (*get_cpl)(struct kvm_vcpu *vcpu);
1640 void (*set_segment)(struct kvm_vcpu *vcpu,
1642 void (*get_cs_db_l_bits)(struct kvm_vcpu *vcpu, int *db, int *l);
1643 bool (*is_valid_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0);
1644 void (*set_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0);
1645 void (*post_set_cr3)(struct kvm_vcpu *vcpu, unsigned long cr3);
1646 bool (*is_valid_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
1647 void (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
1648 int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer);
1649 void (*get_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
1650 void (*set_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
1651 void (*get_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
1652 void (*set_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
1653 void (*sync_dirty_debug_regs)(struct kvm_vcpu *vcpu);
1654 void (*set_dr7)(struct kvm_vcpu *vcpu, unsigned long value);
1655 void (*cache_reg)(struct kvm_vcpu *vcpu, enum kvm_reg reg);
1656 unsigned long (*get_rflags)(struct kvm_vcpu *vcpu);
1657 void (*set_rflags)(struct kvm_vcpu *vcpu, unsigned long rflags);
1658 bool (*get_if_flag)(struct kvm_vcpu *vcpu);
1660 void (*flush_tlb_all)(struct kvm_vcpu *vcpu);
1661 void (*flush_tlb_current)(struct kvm_vcpu *vcpu);
1674 void (*flush_tlb_gva)(struct kvm_vcpu *vcpu, gva_t addr);
1680 void (*flush_tlb_guest)(struct kvm_vcpu *vcpu);
1682 int (*vcpu_pre_run)(struct kvm_vcpu *vcpu);
1683 enum exit_fastpath_completion (*vcpu_run)(struct kvm_vcpu *vcpu,
1685 int (*handle_exit)(struct kvm_vcpu *vcpu,
1687 int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu);
1688 void (*update_emulated_instruction)(struct kvm_vcpu *vcpu);
1689 void (*set_interrupt_shadow)(struct kvm_vcpu *vcpu, int mask);
1690 u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu);
1691 void (*patch_hypercall)(struct kvm_vcpu *vcpu,
1693 void (*inject_irq)(struct kvm_vcpu *vcpu, bool reinjected);
1694 void (*inject_nmi)(struct kvm_vcpu *vcpu);
1695 void (*inject_exception)(struct kvm_vcpu *vcpu);
1696 void (*cancel_injection)(struct kvm_vcpu *vcpu);
1697 int (*interrupt_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
1698 int (*nmi_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
1699 bool (*get_nmi_mask)(struct kvm_vcpu *vcpu);
1700 void (*set_nmi_mask)(struct kvm_vcpu *vcpu, bool masked);
1702 bool (*is_vnmi_pending)(struct kvm_vcpu *vcpu);
1707 bool (*set_vnmi_pending)(struct kvm_vcpu *vcpu);
1708 void (*enable_nmi_window)(struct kvm_vcpu *vcpu);
1709 void (*enable_irq_window)(struct kvm_vcpu *vcpu);
1710 void (*update_cr8_intercept)(struct kvm_vcpu *vcpu, int tpr, int irr);
1714 void (*refresh_apicv_exec_ctrl)(struct kvm_vcpu *vcpu);
1715 void (*hwapic_irr_update)(struct kvm_vcpu *vcpu, int max_irr);
1717 bool (*guest_apic_has_interrupt)(struct kvm_vcpu *vcpu);
1718 void (*load_eoi_exitmap)(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap);
1719 void (*set_virtual_apic_mode)(struct kvm_vcpu *vcpu);
1720 void (*set_apic_access_page_addr)(struct kvm_vcpu *vcpu);
1723 int (*sync_pir_to_irr)(struct kvm_vcpu *vcpu);
1726 u8 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
1728 void (*load_mmu_pgd)(struct kvm_vcpu *vcpu, hpa_t root_hpa,
1733 u64 (*get_l2_tsc_offset)(struct kvm_vcpu *vcpu);
1734 u64 (*get_l2_tsc_multiplier)(struct kvm_vcpu *vcpu);
1735 void (*write_tsc_offset)(struct kvm_vcpu *vcpu);
1736 void (*write_tsc_multiplier)(struct kvm_vcpu *vcpu);
1742 void (*get_exit_info)(struct kvm_vcpu *vcpu, u32 *reason,
1746 int (*check_intercept)(struct kvm_vcpu *vcpu,
1750 void (*handle_exit_irqoff)(struct kvm_vcpu *vcpu);
1752 void (*sched_in)(struct kvm_vcpu *vcpu, int cpu);
1759 void (*update_cpu_dirty_logging)(struct kvm_vcpu *vcpu);
1763 void (*vcpu_blocking)(struct kvm_vcpu *vcpu);
1764 void (*vcpu_unblocking)(struct kvm_vcpu *vcpu);
1769 void (*apicv_pre_state_restore)(struct kvm_vcpu *vcpu);
1770 void (*apicv_post_state_restore)(struct kvm_vcpu *vcpu);
1771 bool (*dy_apicv_has_pending_interrupt)(struct kvm_vcpu *vcpu);
1773 int (*set_hv_timer)(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc,
1775 void (*cancel_hv_timer)(struct kvm_vcpu *vcpu);
1777 void (*setup_mce)(struct kvm_vcpu *vcpu);
1780 int (*smi_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
1781 int (*enter_smm)(struct kvm_vcpu *vcpu, union kvm_smram *smram);
1782 int (*leave_smm)(struct kvm_vcpu *vcpu, const union kvm_smram *smram);
1783 void (*enable_smi_window)(struct kvm_vcpu *vcpu);
1796 int (*check_emulate_instruction)(struct kvm_vcpu *vcpu, int emul_type,
1799 bool (*apic_init_signal_blocked)(struct kvm_vcpu *vcpu);
1800 int (*enable_l2_tlb_flush)(struct kvm_vcpu *vcpu);
1802 void (*migrate_timers)(struct kvm_vcpu *vcpu);
1803 void (*msr_filter_changed)(struct kvm_vcpu *vcpu);
1804 int (*complete_emulated_msr)(struct kvm_vcpu *vcpu, int err);
1806 void (*vcpu_deliver_sipi_vector)(struct kvm_vcpu *vcpu, u8 vector);
1811 unsigned long (*vcpu_get_apicv_inhibit_reasons)(struct kvm_vcpu *vcpu);
1813 gva_t (*get_untagged_addr)(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags);
1814 void *(*alloc_apic_backing_page)(struct kvm_vcpu *vcpu);
1818 void (*leave_nested)(struct kvm_vcpu *vcpu);
1819 bool (*is_exception_vmexit)(struct kvm_vcpu *vcpu, u8 vector,
1821 int (*check_events)(struct kvm_vcpu *vcpu);
1822 bool (*has_events)(struct kvm_vcpu *vcpu);
1823 void (*triple_fault)(struct kvm_vcpu *vcpu);
1824 int (*get_state)(struct kvm_vcpu *vcpu,
1827 int (*set_state)(struct kvm_vcpu *vcpu,
1830 bool (*get_nested_state_pages)(struct kvm_vcpu *vcpu);
1831 int (*write_log_dirty)(struct kvm_vcpu *vcpu, gpa_t l2_gpa);
1833 int (*enable_evmcs)(struct kvm_vcpu *vcpu,
1835 uint16_t (*get_evmcs_version)(struct kvm_vcpu *vcpu);
1836 void (*hv_inject_synthetic_vmexit_post_tlb_flush)(struct kvm_vcpu *vcpu);
1908 #define kvm_arch_pmi_in_guest(vcpu) \
1909 ((vcpu) && (vcpu)->arch.handling_intr_from_guest && \
1910 (!!in_nmi() == ((vcpu)->arch.handling_intr_from_guest == KVM_HANDLING_NMI)))
1916 void kvm_mmu_destroy(struct kvm_vcpu *vcpu);
1917 int kvm_mmu_create(struct kvm_vcpu *vcpu);
1924 void kvm_mmu_after_set_cpuid(struct kvm_vcpu *vcpu);
1925 void kvm_mmu_reset_context(struct kvm_vcpu *vcpu);
1943 int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3);
1945 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
1963 u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu);
2033 int kvm_emulate_instruction(struct kvm_vcpu *vcpu, int emulation_type);
2034 int kvm_emulate_instruction_from_buffer(struct kvm_vcpu *vcpu,
2036 void __kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu,
2038 void kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu);
2041 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
2042 int __kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data, bool host_initiated);
2043 int kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data);
2044 int kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data);
2045 int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu);
2046 int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu);
2047 int kvm_emulate_as_nop(struct kvm_vcpu *vcpu);
2048 int kvm_emulate_invd(struct kvm_vcpu *vcpu);
2049 int kvm_emulate_mwait(struct kvm_vcpu *vcpu);
2050 int kvm_handle_invalid_op(struct kvm_vcpu *vcpu);
2051 int kvm_emulate_monitor(struct kvm_vcpu *vcpu);
2053 int kvm_fast_pio(struct kvm_vcpu *vcpu, int size, unsigned short port, int in);
2054 int kvm_emulate_cpuid(struct kvm_vcpu *vcpu);
2055 int kvm_emulate_halt(struct kvm_vcpu *vcpu);
2056 int kvm_emulate_halt_noskip(struct kvm_vcpu *vcpu);
2057 int kvm_emulate_ap_reset_hold(struct kvm_vcpu *vcpu);
2058 int kvm_emulate_wbinvd(struct kvm_vcpu *vcpu);
2060 void kvm_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
2061 void kvm_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
2062 int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg);
2063 void kvm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector);
2065 int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int idt_index,
2068 void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0);
2069 void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4);
2070 int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
2071 int kvm_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3);
2072 int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
2073 int kvm_set_cr8(struct kvm_vcpu *vcpu, unsigned long cr8);
2074 int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val);
2075 unsigned long kvm_get_dr(struct kvm_vcpu *vcpu, int dr);
2076 unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu);
2077 void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw);
2078 int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu);
2080 int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
2081 int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
2083 unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu);
2084 void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
2085 int kvm_emulate_rdpmc(struct kvm_vcpu *vcpu);
2087 void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr);
2088 void kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code);
2089 void kvm_queue_exception_p(struct kvm_vcpu *vcpu, unsigned nr, unsigned long payload);
2090 void kvm_requeue_exception(struct kvm_vcpu *vcpu, unsigned nr);
2091 void kvm_requeue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code);
2092 void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault);
2093 void kvm_inject_emulated_page_fault(struct kvm_vcpu *vcpu,
2095 bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl);
2096 bool kvm_require_dr(struct kvm_vcpu *vcpu, int dr);
2113 void kvm_inject_nmi(struct kvm_vcpu *vcpu);
2114 int kvm_get_nr_pending_nmis(struct kvm_vcpu *vcpu);
2116 void kvm_update_dr7(struct kvm_vcpu *vcpu);
2122 gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
2124 gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
2126 gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
2130 bool kvm_vcpu_apicv_activated(struct kvm_vcpu *vcpu);
2131 void __kvm_vcpu_update_apicv(struct kvm_vcpu *vcpu);
2149 unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
2153 int kvm_emulate_hypercall(struct kvm_vcpu *vcpu);
2155 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
2157 void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
2158 void kvm_mmu_invalidate_addr(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
2160 void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid);
2161 void kvm_mmu_new_pgd(struct kvm_vcpu *vcpu, gpa_t new_pgd);
2195 static inline void kvm_inject_gp(struct kvm_vcpu *vcpu, u32 error_code)
2197 kvm_queue_exception_e(vcpu, GP_VECTOR, error_code);
2223 # define kvm_arch_vcpu_memslots_id(vcpu) ((vcpu)->arch.hflags & HF_SMM_MASK ? 1 : 0)
2230 int kvm_cpu_has_interrupt(struct kvm_vcpu *vcpu);
2232 int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu);
2234 void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
2250 u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc);
2254 unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu);
2255 bool kvm_is_linear_rip(struct kvm_vcpu *vcpu, unsigned long linear_rip);
2261 bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
2263 void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
2265 void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
2267 void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu);
2268 bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu);
2269 extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
2271 int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu);
2272 int kvm_complete_insn_gp(struct kvm_vcpu *vcpu, int err);
2276 bool kvm_vcpu_is_reset_bsp(struct kvm_vcpu *vcpu);
2277 bool kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu);
2292 static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu)
2294 static_call_cond(kvm_x86_vcpu_blocking)(vcpu);
2297 static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu)
2299 static_call_cond(kvm_x86_vcpu_unblocking)(vcpu);