H A D | mmu.c | 158 u64 *sptep; member in struct:kvm_shadow_walk_iterator 177 ({ spte = mmu_spte_get_lockless(_walker.sptep); 1; }); \ 184 static void mmu_spte_set(u64 *sptep, u64 spte); 284 static void kvm_flush_remote_tlbs_sptep(struct kvm *kvm, u64 *sptep) argument 286 struct kvm_mmu_page *sp = sptep_to_sp(sptep); 287 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); 292 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, argument 297 trace_mark_mmio_spte(sptep, gfn, spte); 298 mmu_spte_set(sptep, spte); 337 static void __set_spte(u64 *sptep, u6 argument 342 __update_clear_spte_fast(u64 *sptep, u64 spte) argument 347 __update_clear_spte_slow(u64 *sptep, u64 spte) argument 352 __get_spte_lockless(u64 *sptep) argument 365 count_spte_clear(u64 *sptep, u64 spte) argument 377 __set_spte(u64 *sptep, u64 spte) argument 396 __update_clear_spte_fast(u64 *sptep, u64 spte) argument 415 __update_clear_spte_slow(u64 *sptep, u64 spte) argument 449 __get_spte_lockless(u64 *sptep) argument 479 mmu_spte_set(u64 *sptep, u64 new_spte) argument 489 mmu_spte_update_no_track(u64 *sptep, u64 new_spte) argument 520 mmu_spte_update(u64 *sptep, u64 new_spte) argument 561 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) argument 604 mmu_spte_clear_no_track(u64 *sptep) argument 609 mmu_spte_get_lockless(u64 *sptep) argument 615 mmu_spte_age(u64 *sptep) argument 1037 kvm_zap_one_rmap_spte(struct kvm *kvm, struct kvm_rmap_head *rmap_head, u64 *sptep) argument 1139 u64 *sptep; local 1165 u64 *sptep; local 1195 drop_spte(struct kvm *kvm, u64 *sptep) argument 1203 drop_large_spte(struct kvm *kvm, u64 *sptep, bool flush) argument 1229 spte_write_protect(u64 *sptep, bool pt_protect) argument 1247 u64 *sptep; local 1257 spte_clear_dirty(u64 *sptep) argument 1266 spte_wrprot_for_clear_dirty(u64 *sptep) argument 1285 u64 *sptep; local 1461 u64 *sptep; local 1616 u64 *sptep; local 1630 u64 *sptep; local 1779 u64 *sptep; local 2307 kvm_mmu_child_role(u64 *sptep, bool direct, unsigned int access) argument 2353 kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, bool direct, unsigned int access) argument 2429 __link_shadow_page(struct kvm *kvm, struct kvm_mmu_memory_cache *cache, u64 *sptep, struct kvm_mmu_page *sp, bool flush) argument 2464 link_shadow_page(struct kvm_vcpu *vcpu, u64 *sptep, struct kvm_mmu_page *sp) argument 2470 validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned direct_access) argument 2538 u64 *sptep; local 2906 mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, u64 *sptep, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, struct kvm_page_fault *fault) argument 3005 __direct_pte_prefetch(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, u64 *sptep) argument 3030 direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep) argument 3381 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault, u64 *sptep, u64 old_spte, u64 new_spte) argument 3431 u64 *sptep = NULL; local 3449 u64 *sptep; local 4138 get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep) argument 4812 sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, unsigned int access) argument 6526 u64 *sptep, spte; local 6721 u64 *sptep; local [all...] |