Searched refs:sptep (Results 1 - 10 of 10) sorted by relevance

/linux-master/arch/x86/kvm/mmu/
H A Dtdp_iter.h17 static inline u64 kvm_tdp_mmu_read_spte(tdp_ptep_t sptep) argument
19 return READ_ONCE(*rcu_dereference(sptep));
22 static inline u64 kvm_tdp_mmu_write_spte_atomic(tdp_ptep_t sptep, u64 new_spte) argument
24 return xchg(rcu_dereference(sptep), new_spte);
27 static inline void __kvm_tdp_mmu_write_spte(tdp_ptep_t sptep, u64 new_spte) argument
29 WRITE_ONCE(*rcu_dereference(sptep), new_spte);
51 static inline u64 kvm_tdp_mmu_write_spte(tdp_ptep_t sptep, u64 old_spte, argument
55 return kvm_tdp_mmu_write_spte_atomic(sptep, new_spte);
57 __kvm_tdp_mmu_write_spte(sptep, new_spte);
61 static inline u64 tdp_mmu_clear_spte_bits(tdp_ptep_t sptep, u6 argument
93 tdp_ptep_t sptep; member in struct:tdp_iter
[all...]
H A Dmmutrace.h212 TP_PROTO(u64 *sptep, gfn_t gfn, u64 spte),
213 TP_ARGS(sptep, gfn, spte),
216 __field(void *, sptep)
223 __entry->sptep = sptep;
229 TP_printk("sptep:%p gfn %llx access %x gen %x", __entry->sptep,
257 u64 *sptep, u64 old_spte, int ret),
258 TP_ARGS(vcpu, fault, sptep, old_spte, ret),
264 __field(u64 *, sptep)
[all...]
H A Dtdp_iter.c14 iter->sptep = iter->pt_path[iter->level - 1] +
16 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep);
89 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep);
122 iter->sptep++;
123 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep);
H A Dmmu.c158 u64 *sptep; member in struct:kvm_shadow_walk_iterator
177 ({ spte = mmu_spte_get_lockless(_walker.sptep); 1; }); \
184 static void mmu_spte_set(u64 *sptep, u64 spte);
284 static void kvm_flush_remote_tlbs_sptep(struct kvm *kvm, u64 *sptep) argument
286 struct kvm_mmu_page *sp = sptep_to_sp(sptep);
287 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep));
292 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, argument
297 trace_mark_mmio_spte(sptep, gfn, spte);
298 mmu_spte_set(sptep, spte);
337 static void __set_spte(u64 *sptep, u6 argument
342 __update_clear_spte_fast(u64 *sptep, u64 spte) argument
347 __update_clear_spte_slow(u64 *sptep, u64 spte) argument
352 __get_spte_lockless(u64 *sptep) argument
365 count_spte_clear(u64 *sptep, u64 spte) argument
377 __set_spte(u64 *sptep, u64 spte) argument
396 __update_clear_spte_fast(u64 *sptep, u64 spte) argument
415 __update_clear_spte_slow(u64 *sptep, u64 spte) argument
449 __get_spte_lockless(u64 *sptep) argument
479 mmu_spte_set(u64 *sptep, u64 new_spte) argument
489 mmu_spte_update_no_track(u64 *sptep, u64 new_spte) argument
520 mmu_spte_update(u64 *sptep, u64 new_spte) argument
561 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) argument
604 mmu_spte_clear_no_track(u64 *sptep) argument
609 mmu_spte_get_lockless(u64 *sptep) argument
615 mmu_spte_age(u64 *sptep) argument
1037 kvm_zap_one_rmap_spte(struct kvm *kvm, struct kvm_rmap_head *rmap_head, u64 *sptep) argument
1139 u64 *sptep; local
1165 u64 *sptep; local
1195 drop_spte(struct kvm *kvm, u64 *sptep) argument
1203 drop_large_spte(struct kvm *kvm, u64 *sptep, bool flush) argument
1229 spte_write_protect(u64 *sptep, bool pt_protect) argument
1247 u64 *sptep; local
1257 spte_clear_dirty(u64 *sptep) argument
1266 spte_wrprot_for_clear_dirty(u64 *sptep) argument
1285 u64 *sptep; local
1461 u64 *sptep; local
1616 u64 *sptep; local
1630 u64 *sptep; local
1779 u64 *sptep; local
2307 kvm_mmu_child_role(u64 *sptep, bool direct, unsigned int access) argument
2353 kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, bool direct, unsigned int access) argument
2429 __link_shadow_page(struct kvm *kvm, struct kvm_mmu_memory_cache *cache, u64 *sptep, struct kvm_mmu_page *sp, bool flush) argument
2464 link_shadow_page(struct kvm_vcpu *vcpu, u64 *sptep, struct kvm_mmu_page *sp) argument
2470 validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned direct_access) argument
2538 u64 *sptep; local
2906 mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, u64 *sptep, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, struct kvm_page_fault *fault) argument
3005 __direct_pte_prefetch(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, u64 *sptep) argument
3030 direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep) argument
3381 fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault, u64 *sptep, u64 old_spte, u64 new_spte) argument
3431 u64 *sptep = NULL; local
3449 u64 *sptep; local
4138 get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep) argument
4812 sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, unsigned int access) argument
6526 u64 *sptep, spte; local
6721 u64 *sptep; local
[all...]
H A Dtdp_mmu.c197 static void tdp_mmu_init_sp(struct kvm_mmu_page *sp, tdp_ptep_t sptep, argument
206 sp->ptep = sptep;
218 parent_sp = sptep_to_sp(rcu_dereference(iter->sptep));
223 tdp_mmu_init_sp(child_sp, iter->sptep, iter->gfn, role);
354 tdp_ptep_t sptep = pt + i; local
368 old_spte = kvm_tdp_mmu_write_spte_atomic(sptep, REMOVED_SPTE);
383 old_spte = kvm_tdp_mmu_read_spte(sptep);
415 old_spte = kvm_tdp_mmu_write_spte(sptep, old_spte,
554 u64 *sptep = rcu_dereference(iter->sptep); local
625 tdp_mmu_set_spte(struct kvm *kvm, int as_id, tdp_ptep_t sptep, u64 old_spte, u64 new_spte, gfn_t gfn, int level) argument
1833 tdp_ptep_t sptep = NULL; local
[all...]
H A Dpaging_tmpl.h585 u64 *sptep)
592 sp = sptep_to_sp(sptep);
605 return __direct_pte_prefetch(vcpu, sp, sptep);
607 i = spte_index(sptep) & ~(PTE_PREFETCH_NUM - 1);
611 if (spte == sptep)
668 clear_sp_write_flooding_count(it.sptep);
674 sp = kvm_mmu_get_child_sp(vcpu, it.sptep, table_gfn,
707 link_shadow_page(vcpu, it.sptep, sp);
729 disallowed_hugepage_adjust(fault, *it.sptep, it.level);
735 validate_direct_spte(vcpu, it.sptep, direct_acces
584 pte_prefetch(struct kvm_vcpu *vcpu, struct guest_walker *gw, u64 *sptep) argument
907 u64 *sptep, spte; local
[all...]
H A Dspte.h208 static inline int spte_index(u64 *sptep) argument
210 return ((unsigned long)sptep / sizeof(*sptep)) & (SPTE_ENT_PER_PAGE - 1);
235 static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep) argument
237 return to_shadow_page(__pa(sptep));
/linux-master/arch/s390/mm/
H A Dpgtable.c679 pte_t *sptep, pte_t *tptep, pte_t pte)
687 spgste = pgste_get_lock(sptep);
688 spte = *sptep;
701 pgste_set_unlock(sptep, spgste);
678 ptep_shadow_pte(struct mm_struct *mm, unsigned long saddr, pte_t *sptep, pte_t *tptep, pte_t pte) argument
H A Dgmap.c2143 pte_t *sptep, *tptep; local
2167 sptep = gmap_pte_op_walk(parent, paddr, &ptl);
2168 if (sptep) {
2174 gmap_pte_op_end(sptep, ptl);
2178 rc = ptep_shadow_pte(sg->mm, saddr, sptep, tptep, pte);
2185 gmap_pte_op_end(sptep, ptl);
/linux-master/arch/s390/include/asm/
H A Dpgtable.h1289 pte_t *sptep, pte_t *tptep, pte_t pte);

Completed in 166 milliseconds