Lines Matching defs:gpa
303 u64 gpa = spte & shadow_nonpresent_or_rsvd_lower_gfn_mask;
305 gpa |= (spte >> SHADOW_NONPRESENT_OR_RSVD_MASK_LEN)
308 return gpa >> PAGE_SHIFT;
2777 gpa_t gpa;
2783 gpa = kvm_mmu_gva_to_gpa_read(vcpu, gva, NULL);
2785 r = kvm_mmu_unprotect_page(vcpu->kvm, gpa >> PAGE_SHIFT);
3420 * gpa, and sets *spte to the spte value. This spte may be non-preset. If no
3427 static u64 *fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gpa_t gpa, u64 *spte)
3433 for_each_shadow_entry_lockless(vcpu, gpa, iterator, old_spte) {
5695 static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa,
5708 *gpa &= ~(gpa_t)7;
5713 r = kvm_vcpu_read_guest_atomic(vcpu, *gpa, &gentry, *bytes);
5742 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa,
5747 offset = offset_in_page(gpa);
5763 static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte)
5769 page_offset = offset_in_page(gpa);
5794 void kvm_mmu_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
5797 gfn_t gfn = gpa >> PAGE_SHIFT;
5813 gentry = mmu_pte_write_fetch_gpte(vcpu, &gpa, &bytes);
5818 if (detect_write_misaligned(sp, gpa, bytes) ||
5825 spte = get_written_sptes(sp, gpa, &npte);