Searched refs:gfn (Results 1 - 25 of 84) sorted by path

1234

/linux-master/arch/arm64/kvm/
H A Dguest.c1023 gpa_t gfn; local
1043 gfn = gpa_to_gfn(guest_ipa);
1048 kvm_pfn_t pfn = gfn_to_pfn_prot(kvm, gfn, write, NULL);
1098 gfn++;
H A Dmmu.c176 gfn_t gfn, u64 nr_pages)
179 gfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT);
1178 * @gfn_offset: The gfn offset in memory slot
1390 gfn_t gfn; local
1472 gfn = fault_ipa >> PAGE_SHIFT;
1491 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL,
1588 mark_page_dirty_in_slot(kvm, memslot, gfn);
1632 gfn_t gfn; local
1687 gfn = fault_ipa >> PAGE_SHIFT;
1688 memslot = gfn_to_memslot(vcpu->kvm, gfn);
175 kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument
[all...]
/linux-master/arch/arm64/kvm/vgic/
H A Dvgic-its.c930 gfn_t gfn = gpa >> PAGE_SHIFT; local
935 ret = kvm_is_visible_gfn(its->dev->kvm, gfn);
/linux-master/arch/loongarch/include/asm/
H A Dkvm_mmu.h34 unsigned long gfn; member in struct:kvm_ptw_ctx
/linux-master/arch/loongarch/kvm/
H A Dmmu.c68 offset = (addr >> PAGE_SHIFT) - ctx->gfn;
354 * @gfn_offset: The gfn offset in memory slot
372 ctx.gfn = base_gfn;
572 gfn_t gfn = gpa >> PAGE_SHIFT; local
602 slot = gfn_to_memslot(kvm, gfn);
628 mark_page_dirty(kvm, gfn);
671 * Lookup the mapping level for @gfn in the current mm.
695 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, argument
714 hva = __gfn_to_hva_memslot(slot, gfn);
757 static kvm_pte_t *kvm_split_huge(struct kvm_vcpu *vcpu, kvm_pte_t *ptep, gfn_t gfn) argument
807 gfn_t gfn = gpa >> PAGE_SHIFT; local
[all...]
/linux-master/arch/mips/kvm/
H A Dmmu.c407 * @gfn_offset: The gfn offset in memory slot
515 gfn_t gfn = gpa >> PAGE_SHIFT; local
546 mark_page_dirty(kvm, gfn);
592 gfn_t gfn = gpa >> PAGE_SHIFT; local
632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable);
660 mark_page_dirty(kvm, gfn);
/linux-master/arch/powerpc/include/asm/
H A Dkvm_book3s.h215 unsigned long gfn);
217 unsigned long gfn);
219 unsigned long gfn);
242 unsigned long gfn, unsigned long psize);
H A Dkvm_book3s_uvmem.h22 int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn);
81 static inline int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn) argument
/linux-master/arch/powerpc/kvm/
H A Dbook3s.c432 gfn_t gfn = gpa >> PAGE_SHIFT; local
450 return gfn_to_pfn_prot(vcpu->kvm, gfn, writing, writable);
H A Dbook3s_64_mmu_host.c89 unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; local
128 mark_page_dirty(vcpu->kvm, gfn);
H A Dbook3s_64_mmu_hv.c515 unsigned long gpa, gfn, hva, pfn, hpa; local
578 gfn = gpa >> PAGE_SHIFT;
579 memslot = gfn_to_memslot(kvm, gfn);
604 hva = gfn_to_hva_memslot(memslot, gfn);
616 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL,
781 unsigned long *rmapp, unsigned long gfn)
805 hpte_rpn(ptel, psize) == gfn) {
813 kvmppc_update_dirty_map(memslot, gfn, psize);
822 unsigned long gfn)
828 rmapp = &memslot->arch.rmap[gfn
779 kvmppc_unmap_hpte(struct kvm *kvm, unsigned long i, struct kvm_memory_slot *memslot, unsigned long *rmapp, unsigned long gfn) argument
821 kvm_unmap_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
859 gfn_t gfn; local
875 unsigned long gfn; local
899 kvm_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
956 gfn_t gfn; local
970 kvm_test_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
1119 unsigned long gfn; local
1160 unsigned long gfn = gpa >> PAGE_SHIFT; local
1192 unsigned long gfn; local
1276 unsigned long gfn = hpte_rpn(guest_rpte, apsize); local
[all...]
H A Dbook3s_64_mmu_radix.c430 unsigned long gfn = gpa >> PAGE_SHIFT; local
442 memslot = gfn_to_memslot(kvm, gfn);
459 kvmppc_update_dirty_map(memslot, gfn, page_size);
830 unsigned long hva, gfn = gpa >> PAGE_SHIFT; local
848 hva = gfn_to_hva_memslot(memslot, gfn);
855 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL,
949 unsigned long gpa, gfn; local
972 gfn = gpa >> PAGE_SHIFT;
977 return kvmppc_send_page_to_uv(kvm, gfn);
980 memslot = gfn_to_memslot(kvm, gfn);
1032 kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
1051 kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
1079 kvm_test_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument
1101 unsigned long gfn = memslot->base_gfn + pagenum; local
[all...]
H A Dbook3s_64_vio.c358 unsigned long gfn = tce >> PAGE_SHIFT; local
361 memslot = __gfn_to_memslot(kvm_memslots(kvm), gfn);
365 *ua = __gfn_to_hva_memslot(memslot, gfn) |
H A Dbook3s_hv_nested.c1033 unsigned long gfn, end_gfn; local
1038 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn;
1039 end_gfn = gfn + (nbytes >> PAGE_SHIFT);
1044 for (; gfn < end_gfn; gfn++) {
1045 unsigned long *rmap = &memslot->arch.rmap[gfn];
1527 unsigned long n_gpa, gpa, gfn, perm = 0UL; local
1585 gfn = gpa >> PAGE_SHIFT;
1589 memslot = gfn_to_memslot(kvm, gfn);
1639 /* Align gfn t
[all...]
H A Dbook3s_hv_rm_mmu.c97 unsigned long gfn, unsigned long psize)
104 gfn -= memslot->base_gfn;
105 set_dirty_bits_atomic(memslot->dirty_bitmap, gfn, npages);
113 unsigned long gfn; local
117 gfn = hpte_rpn(hpte_gr, psize);
118 memslot = __gfn_to_memslot(kvm_memslots_raw(kvm), gfn);
120 kvmppc_update_dirty_map(memslot, gfn, psize);
131 unsigned long gfn; local
133 gfn = hpte_rpn(hpte_gr, kvmppc_actual_pgsz(hpte_v, hpte_gr));
134 memslot = __gfn_to_memslot(kvm_memslots_raw(kvm), gfn);
96 kvmppc_update_dirty_map(const struct kvm_memory_slot *memslot, unsigned long gfn, unsigned long psize) argument
156 unsigned long gfn; local
188 unsigned long i, pa, gpa, gfn, psize; local
889 unsigned long gfn, hva, pa, psize = PAGE_SHIFT; local
[all...]
H A Dbook3s_hv_uvmem.c289 static void kvmppc_mark_gfn(unsigned long gfn, struct kvm *kvm, argument
295 if (gfn >= p->base_pfn && gfn < p->base_pfn + p->nr_pfns) {
296 unsigned long index = gfn - p->base_pfn;
308 static void kvmppc_gfn_secure_uvmem_pfn(unsigned long gfn, argument
311 kvmppc_mark_gfn(gfn, kvm, KVMPPC_GFN_UVMEM_PFN, uvmem_pfn);
315 static void kvmppc_gfn_secure_mem_pfn(unsigned long gfn, struct kvm *kvm) argument
317 kvmppc_mark_gfn(gfn, kvm, KVMPPC_GFN_MEM_PFN, 0);
321 static void kvmppc_gfn_shared(unsigned long gfn, struct kvm *kvm) argument
323 kvmppc_mark_gfn(gfn, kv
327 kvmppc_gfn_remove(unsigned long gfn, struct kvm *kvm) argument
333 kvmppc_gfn_is_uvmem_pfn(unsigned long gfn, struct kvm *kvm, unsigned long *uvmem_pfn) argument
361 kvmppc_next_nontransitioned_gfn(const struct kvm_memory_slot *memslot, struct kvm *kvm, unsigned long *gfn) argument
394 unsigned long gfn = memslot->base_gfn; local
617 unsigned long uvmem_pfn, gfn; local
797 unsigned long gfn = memslot->base_gfn; local
885 unsigned long gfn = gpa >> page_shift; local
943 unsigned long gfn = gpa >> page_shift; local
1049 unsigned long gfn = gpa >> page_shift; local
1084 kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn) argument
[all...]
H A Dbook3s_xive_native.c573 gfn_t gfn; local
647 gfn = gpa_to_gfn(kvm_eq.qaddr);
649 page_size = kvm_host_page_size(vcpu, gfn);
656 page = gfn_to_page(kvm, gfn);
H A Dbooke.c1287 gfn_t gfn; local
1316 gfn = gpaddr >> PAGE_SHIFT;
1318 if (kvm_is_visible_gfn(vcpu->kvm, gfn)) {
1344 gfn_t gfn; local
1364 gfn = gpaddr >> PAGE_SHIFT;
1366 if (kvm_is_visible_gfn(vcpu->kvm, gfn)) {
H A De500_mmu_host.c323 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe,
353 slot = gfn_to_memslot(vcpu_e500->vcpu.kvm, gfn);
354 hva = gfn_to_hva_memslot(slot, gfn);
381 slot_start = pfn - (gfn - slot->base_gfn);
400 * requested) that will cover gfn, stay within the
401 * range, and for which gfn and pfn are mutually
409 gfn_start = gfn & ~(tsize_pages - 1);
412 if (gfn_start + pfn - gfn < start)
414 if (gfn_end + pfn - gfn > end)
416 if ((gfn
322 kvmppc_e500_shadow_map(struct kvmppc_vcpu_e500 *vcpu_e500, u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, int tlbsel, struct kvm_book3e_206_tlb_entry *stlbe, struct tlbe_ref *ref) argument
559 kvmppc_e500_tlb1_map(struct kvmppc_vcpu_e500 *vcpu_e500, u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, struct kvm_book3e_206_tlb_entry *stlbe, int esel) argument
612 gfn_t gfn = gpaddr >> PAGE_SHIFT; local
[all...]
/linux-master/arch/riscv/kvm/
H A Dmmu.c617 gfn_t gfn = gpa >> PAGE_SHIFT; local
650 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT;
670 hfn = gfn_to_pfn_prot(kvm, gfn, is_write, &writable);
693 mark_page_dirty(kvm, gfn);
H A Dvcpu_exit.c19 gfn_t gfn; local
23 gfn = fault_addr >> PAGE_SHIFT;
24 memslot = gfn_to_memslot(vcpu->kvm, gfn);
25 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable);
H A Dvcpu_sbi_sta.c36 gfn_t gfn; local
46 gfn = shmem >> PAGE_SHIFT;
47 hva = kvm_vcpu_gfn_to_hva(vcpu, gfn);
78 kvm_vcpu_mark_page_dirty(vcpu, gfn);
/linux-master/arch/s390/kvm/
H A Dgaccess.c874 * caller needs to ensure that gfn is accessible, so we can
985 const gfn_t gfn = gpa_to_gfn(gpa); local
989 rc = kvm_write_guest_page(kvm, gfn, data, offset, len);
991 rc = kvm_read_guest_page(kvm, gfn, data, offset, len);
1001 gfn_t gfn; local
1005 gfn = gpa >> PAGE_SHIFT;
1006 slot = gfn_to_memslot(kvm, gfn);
1007 hva = gfn_to_hva_memslot_prot(slot, gfn, &writable);
1025 mark_page_dirty_in_slot(kvm, slot, gfn);
1176 gfn_t gfn local
[all...]
H A Dkvm-s390.c2215 gfn_t gfn)
2217 return ____gfn_to_memslot(slots, gfn, true);
2214 gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn) argument
H A Dpriv.c1160 unsigned long gfn, hva, res, pgstev, ptev; local
1169 gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT;
1170 hva = gfn_to_hva(vcpu->kvm, gfn);
1205 cbrlo[entries] = gfn << PAGE_SHIFT;
1209 struct kvm_memory_slot *ms = gfn_to_memslot(vcpu->kvm, gfn);
1212 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms)))

Completed in 353 milliseconds

1234