Searched refs:gfn (Results 1 - 25 of 84) sorted by last modified time

1234

/linux-master/arch/x86/kernel/
H A Dsev-shared.c1094 vaddr = (unsigned long)pfn_to_kaddr(e->gfn);
H A Dsev.c828 e->gfn = pfn;
/linux-master/virt/kvm/
H A Dpfncache.c291 gfn_t gfn = gpa_to_gfn(gpa); local
295 gpc->memslot = __gfn_to_memslot(slots, gfn);
296 gpc->uhva = gfn_to_hva_memslot(gpc->memslot, gfn);
H A Dkvm_main.c367 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument
369 if (!kvm_arch_flush_remote_tlbs_range(kvm, gfn, nr_pages))
639 * {gfn(page) | page intersects with [hva_start, hva_end)} =
1579 * If the memslot gfn is unchanged, rb_replace_node() can be used to
1580 * switch the node in the gfn tree instead of removing the old and
1835 * memslot will be created. Validation of sp->gfn happens in:
2541 /* Set @attributes for the gfn range [@start, @end). */
2627 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) argument
2629 return __gfn_to_memslot(kvm_memslots(kvm), gfn);
2633 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) argument
2666 kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) argument
2674 kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument
2682 kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) argument
2711 __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages, bool write) argument
2726 gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages) argument
2732 gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) argument
2739 gfn_to_hva(struct kvm *kvm, gfn_t gfn) argument
2745 kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn) argument
2759 gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, bool *writable) argument
2770 gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) argument
2777 kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable) argument
3028 __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable, hva_t *hva) argument
3060 gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, bool *writable) argument
3068 gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) argument
3075 gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn) argument
3082 kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn) argument
3088 gfn_to_pfn(struct kvm *kvm, gfn_t gfn) argument
3094 kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument
3100 gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, struct page **pages, int nr_pages) argument
3123 gfn_to_page(struct kvm *kvm, gfn_t gfn) argument
3149 kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument
3310 __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, int len) argument
3325 kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, int len) argument
3334 kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset, int len) argument
3345 gfn_t gfn = gpa >> PAGE_SHIFT; local
3365 gfn_t gfn = gpa >> PAGE_SHIFT; local
3383 __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, unsigned long len) argument
3403 gfn_t gfn = gpa >> PAGE_SHIFT; local
3411 __kvm_write_guest_page(struct kvm *kvm, struct kvm_memory_slot *memslot, gfn_t gfn, const void *data, int offset, int len) argument
3428 kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data, int offset, int len) argument
3437 kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data, int offset, int len) argument
3449 gfn_t gfn = gpa >> PAGE_SHIFT; local
3470 gfn_t gfn = gpa >> PAGE_SHIFT; local
3615 gfn_t gfn = gpa >> PAGE_SHIFT; local
3632 mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn) argument
3657 mark_page_dirty(struct kvm *kvm, gfn_t gfn) argument
3666 kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn) argument
[all...]
/linux-master/arch/x86/kvm/
H A Dx86.c5119 * Take the srcu lock as memslots will be accessed to check the gfn
8859 * table, unprotect the gfn (zap any relevant SPTEs) and re-enter the
8869 * the issue by unprotecting the gfn, as zapping the shadow page will
8872 * a SPTE and write-protect the gfn to resolve the !PRESENT fault, and
8873 * then zap the SPTE to unprotect the gfn, and then do it all over
10650 unsigned long gfn = gpa_to_gfn(APIC_DEFAULT_PHYS_BASE); local
10653 kvm_zap_gfn_range(kvm, gfn, gfn+1);
12814 * If the gfn and userspace address are not aligned wrt each
13199 static inline u32 kvm_async_pf_hash_fn(gfn_t gfn) argument
13211 kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument
13221 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) argument
13234 kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument
13239 kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument
[all...]
H A Dxen.c732 gfn_t gfn = data->u.shared_info.gfn; local
734 if (gfn == KVM_XEN_INVALID_GFN) {
739 gfn_to_gpa(gfn), PAGE_SIZE);
817 data->u.shared_info.gfn = gpa_to_gfn(kvm->arch.xen.shinfo_cache.gpa);
819 data->u.shared_info.gfn = KVM_XEN_INVALID_GFN;
H A Dx86.h219 gva_t gva, gfn_t gfn, unsigned access)
232 vcpu->arch.mmio_gfn = gfn;
313 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn);
316 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
218 vcpu_cache_mmio_info(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, unsigned access) argument
/linux-master/arch/x86/kvm/vmx/
H A Dvmx.c4113 WARN_ON_ONCE(!vmx->nested.virtual_apic_map.gfn))
6764 const gfn_t gfn = APIC_DEFAULT_PHYS_BASE >> PAGE_SHIFT; local
6805 pfn = gfn_to_pfn_memslot(slot, gfn);
6810 if (mmu_invalidate_retry_gfn(kvm, mmu_seq, gfn)) {
7617 static u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) argument
7651 return kvm_mtrr_get_guest_memory_type(vcpu, gfn) << VMX_EPT_MT_EPTE_SHIFT;
H A Dnested.c3841 gfn_t gfn; local
3849 gfn = vmcs12->virtual_apic_page_addr >> PAGE_SHIFT;
3850 kvm_vcpu_mark_page_dirty(vcpu, gfn);
3854 gfn = vmcs12->posted_intr_desc_addr >> PAGE_SHIFT;
3855 kvm_vcpu_mark_page_dirty(vcpu, gfn);
/linux-master/arch/x86/kvm/mmu/
H A Dtdp_mmu.c198 gfn_t gfn, union kvm_mmu_page_role role)
205 sp->gfn = gfn;
223 tdp_mmu_init_sp(child_sp, iter->sptep, iter->gfn, role);
290 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
346 gfn_t base_gfn = sp->gfn;
355 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); local
418 handle_changed_spte(kvm, kvm_mmu_page_as_id(sp), gfn,
429 * @gfn: the base GFN that was mapped by the SPTE
441 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, argument
197 tdp_mmu_init_sp(struct kvm_mmu_page *sp, tdp_ptep_t sptep, gfn_t gfn, union kvm_mmu_page_role role) argument
625 tdp_mmu_set_spte(struct kvm *kvm, int as_id, tdp_ptep_t sptep, u64 old_spte, u64 new_spte, gfn_t gfn, int level) argument
1615 clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, unsigned long mask, bool wrprot) argument
1661 kvm_tdp_mmu_clear_dirty_pt_masked(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn, unsigned long mask, bool wrprot) argument
1742 write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root, gfn_t gfn, int min_level) argument
1778 kvm_tdp_mmu_write_protect_gfn(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn, int min_level) argument
1803 gfn_t gfn = addr >> PAGE_SHIFT; local
1832 gfn_t gfn = addr >> PAGE_SHIFT; local
[all...]
H A Dmmu.c287 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); local
289 kvm_flush_remote_tlbs_gfn(kvm, gfn, sp->role.level);
292 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, argument
295 u64 spte = make_mmio_spte(vcpu, gfn, access);
297 trace_mark_mmio_spte(sptep, gfn, spte);
720 return sp->gfn;
725 return sp->gfn + (index << ((sp->role.level - 1) * SPTE_LEVEL_BITS));
755 gfn_t gfn, unsigned int access)
758 sp->shadowed_translation[index] = (gfn << PAGE_SHIFT) | access;
765 sp->gfn, kvm_mmu_page_get_acces
754 kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index, gfn_t gfn, unsigned int access) argument
776 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); local
785 lpage_info_slot(gfn_t gfn, const struct kvm_memory_slot *slot, int level) argument
802 update_gfn_disallow_lpage_count(const struct kvm_memory_slot *slot, gfn_t gfn, int count) argument
817 kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) argument
822 kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) argument
831 gfn_t gfn; local
879 gfn_t gfn; local
907 gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, bool no_dirty_log) argument
1086 gfn_to_rmap(gfn_t gfn, int level, const struct kvm_memory_slot *slot) argument
1100 gfn_t gfn; local
1414 kvm_mmu_slot_gfn_write_protect(struct kvm *kvm, struct kvm_memory_slot *slot, u64 gfn, int min_level) argument
1436 kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) argument
1450 kvm_zap_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, struct kvm_memory_slot *slot, gfn_t gfn, int level, pte_t unused) argument
1457 kvm_set_pte_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, struct kvm_memory_slot *slot, gfn_t gfn, int level, pte_t pte) argument
1503 gfn_t gfn; member in struct:slot_rmap_walk_iterator
1612 kvm_age_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, struct kvm_memory_slot *slot, gfn_t gfn, int level, pte_t unused) argument
1626 kvm_test_age_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head, struct kvm_memory_slot *slot, gfn_t gfn, int level, pte_t unused) argument
1641 __rmap_add(struct kvm *kvm, struct kvm_mmu_memory_cache *cache, const struct kvm_memory_slot *slot, u64 *spte, gfn_t gfn, unsigned int access) argument
1665 rmap_add(struct kvm_vcpu *vcpu, const struct kvm_memory_slot *slot, u64 *spte, gfn_t gfn, unsigned int access) argument
1749 kvm_page_table_hashfn(gfn_t gfn) argument
2151 kvm_mmu_find_shadow_page(struct kvm *kvm, struct kvm_vcpu *vcpu, gfn_t gfn, struct hlist_head *sp_list, union kvm_mmu_page_role role) argument
2236 kvm_mmu_alloc_shadow_page(struct kvm *kvm, struct shadow_page_caches *caches, gfn_t gfn, struct hlist_head *sp_list, union kvm_mmu_page_role role) argument
2272 __kvm_mmu_get_shadow_page(struct kvm *kvm, struct kvm_vcpu *vcpu, struct shadow_page_caches *caches, gfn_t gfn, union kvm_mmu_page_role role) argument
2294 kvm_mmu_get_shadow_page(struct kvm_vcpu *vcpu, gfn_t gfn, union kvm_mmu_page_role role) argument
2353 kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, bool direct, unsigned int access) argument
2757 kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) argument
2805 mmu_try_to_unsync_pages(struct kvm *kvm, const struct kvm_memory_slot *slot, gfn_t gfn, bool can_unsync, bool prefetch) argument
2906 mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot, u64 *sptep, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, struct kvm_page_fault *fault) argument
2985 gfn_t gfn; local
3082 host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, const struct kvm_memory_slot *slot) argument
3146 __kvm_mmu_max_mapping_level(struct kvm *kvm, const struct kvm_memory_slot *slot, gfn_t gfn, int max_level, bool is_private) argument
3170 kvm_mmu_max_mapping_level(struct kvm *kvm, const struct kvm_memory_slot *slot, gfn_t gfn, int max_level) argument
3281 kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) argument
3683 mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, u8 level) argument
4200 gfn_t gfn = get_mmio_spte_gfn(spte); local
4262 kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, gfn_t gfn) argument
4812 sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, unsigned int access) argument
5797 gfn_t gfn = gpa >> PAGE_SHIFT; local
6496 gfn_t gfn; local
6527 gfn_t gfn; local
6569 gfn_t gfn; local
7333 hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, int level) argument
7339 hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn, int level) argument
7345 hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn, int level) argument
7351 hugepage_has_attrs(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn, int level, unsigned long attrs) argument
7393 gfn_t gfn = gfn_round_for_level(range->start, level); local
7452 gfn_t gfn; local
[all...]
H A Dtdp_mmu.h42 gfn_t gfn, unsigned long mask,
48 struct kvm_memory_slot *slot, gfn_t gfn,
H A Dpage_track.c75 static void update_gfn_write_track(struct kvm_memory_slot *slot, gfn_t gfn, argument
80 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K);
91 gfn_t gfn)
101 update_gfn_write_track(slot, gfn, 1);
107 kvm_mmu_gfn_disallow_lpage(slot, gfn);
109 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K))
114 struct kvm_memory_slot *slot, gfn_t gfn)
124 update_gfn_write_track(slot, gfn, -1);
130 kvm_mmu_gfn_allow_lpage(slot, gfn);
137 const struct kvm_memory_slot *slot, gfn_t gfn)
90 __kvm_write_track_add_gfn(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn) argument
113 __kvm_write_track_remove_gfn(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn) argument
136 kvm_gfn_is_write_tracked(struct kvm *kvm, const struct kvm_memory_slot *slot, gfn_t gfn) argument
319 kvm_write_track_add_gfn(struct kvm *kvm, gfn_t gfn) argument
349 kvm_write_track_remove_gfn(struct kvm *kvm, gfn_t gfn) argument
[all...]
/linux-master/arch/x86/include/asm/
H A Dkvm_host.h301 * allocating 2 bytes per gfn instead of 4 bytes per gfn.
305 * not create more than 2^16-1 upper-level shadow pages at a single gfn,
308 * A unique shadow page (SP) for a gfn is created if and only if an existing SP
332 * single gfn is a bit less than 2^13.
1660 int (*flush_remote_tlbs_range)(struct kvm *kvm, gfn_t gfn,
1722 u8 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
1844 gfn_t gfn; member in struct:kvm_arch_async_pf
1885 static inline int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, argument
1891 return static_call(kvm_x86_flush_remote_tlbs_range)(kvm, gfn, nr_page
[all...]
/linux-master/tools/arch/x86/include/uapi/asm/
H A Dkvm.h572 __u64 gfn; member in union:kvm_xen_hvm_attr::__anon140::__anon141
/linux-master/drivers/net/
H A Dxen-netfront.c473 static void xennet_tx_setup_grant(unsigned long gfn, unsigned int offset, argument
491 gfn, GNTMAP_readonly);
526 static void xennet_make_one_txreq(unsigned long gfn, unsigned int offset, argument
533 xennet_tx_setup_grant(gfn, offset, len, data);
/linux-master/arch/x86/include/uapi/asm/
H A Dkvm.h572 __u64 gfn; member in union:kvm_xen_hvm_attr::__anon11::__anon12
/linux-master/arch/arm64/kvm/
H A Dmmu.c176 gfn_t gfn, u64 nr_pages)
179 gfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT);
1178 * @gfn_offset: The gfn offset in memory slot
1390 gfn_t gfn; local
1472 gfn = fault_ipa >> PAGE_SHIFT;
1491 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL,
1588 mark_page_dirty_in_slot(kvm, memslot, gfn);
1632 gfn_t gfn; local
1687 gfn = fault_ipa >> PAGE_SHIFT;
1688 memslot = gfn_to_memslot(vcpu->kvm, gfn);
175 kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument
[all...]
H A Dguest.c1023 gpa_t gfn; local
1043 gfn = gpa_to_gfn(guest_ipa);
1048 kvm_pfn_t pfn = gfn_to_pfn_prot(kvm, gfn, write, NULL);
1098 gfn++;
/linux-master/tools/testing/selftests/kvm/x86_64/
H A Dxen_shinfo_test.c415 .u.shared_info.gfn = SHINFO_REGION_GPA / PAGE_SIZE
420 .u.shared_info.gfn = KVM_XEN_INVALID_GFN
523 ha.u.shared_info.gfn = SHINFO_ADDR / PAGE_SIZE;
/linux-master/include/linux/
H A Dkvm_host.h102 * error pfns indicate that the gfn is in slot but faild to
120 * error_noslot pfns indicate that the gfn can not be
129 /* noslot pfn indicates that the gfn is not in slot. */
300 kvm_pfn_t gfn; member in struct:kvm_host_map
1072 /* Iterator used for walking memslots that overlap a gfn range. */
1100 * its key strictly greater than the searched one (the start gfn in our case).
1114 * Find the slot with the lowest gfn that can possibly intersect with
1120 * already has a higher start gfn.
1204 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
1207 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
1567 kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument
1702 try_get_memslot(struct kvm_memory_slot *slot, gfn_t gfn) argument
1721 search_memslots(struct kvm_memslots *slots, gfn_t gfn, bool approx) argument
1742 ____gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn, bool approx) argument
1766 __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) argument
1772 __gfn_to_hva_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) argument
1785 memslot_id(struct kvm *kvm, gfn_t gfn) argument
1798 gfn_to_gpa(gfn_t gfn) argument
2037 mmu_invalidate_retry_gfn(struct kvm *kvm, unsigned long mmu_seq, gfn_t gfn) argument
2073 mmu_invalidate_retry_gfn_unsafe(struct kvm *kvm, unsigned long mmu_seq, gfn_t gfn) argument
2411 kvm_get_memory_attributes(struct kvm *kvm, gfn_t gfn) argument
2423 kvm_mem_is_private(struct kvm *kvm, gfn_t gfn) argument
2429 kvm_mem_is_private(struct kvm *kvm, gfn_t gfn) argument
2439 kvm_gmem_get_pfn(struct kvm *kvm, struct kvm_memory_slot *slot, gfn_t gfn, kvm_pfn_t *pfn, int *max_order) argument
[all...]
/linux-master/arch/s390/kvm/
H A Dpriv.c1160 unsigned long gfn, hva, res, pgstev, ptev; local
1169 gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT;
1170 hva = gfn_to_hva(vcpu->kvm, gfn);
1205 cbrlo[entries] = gfn << PAGE_SHIFT;
1209 struct kvm_memory_slot *ms = gfn_to_memslot(vcpu->kvm, gfn);
1212 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms)))
H A Dkvm-s390.c2215 gfn_t gfn)
2217 return ____gfn_to_memslot(slots, gfn, true);
2214 gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn) argument
H A Dgaccess.c874 * caller needs to ensure that gfn is accessible, so we can
985 const gfn_t gfn = gpa_to_gfn(gpa); local
989 rc = kvm_write_guest_page(kvm, gfn, data, offset, len);
991 rc = kvm_read_guest_page(kvm, gfn, data, offset, len);
1001 gfn_t gfn; local
1005 gfn = gpa >> PAGE_SHIFT;
1006 slot = gfn_to_memslot(kvm, gfn);
1007 hva = gfn_to_hva_memslot_prot(slot, gfn, &writable);
1025 mark_page_dirty_in_slot(kvm, slot, gfn);
1176 gfn_t gfn local
[all...]
/linux-master/arch/arm64/kvm/vgic/
H A Dvgic-its.c930 gfn_t gfn = gpa >> PAGE_SHIFT; local
935 ret = kvm_is_visible_gfn(its->dev->kvm, gfn);

Completed in 467 milliseconds

1234