/linux-master/arch/arm64/kvm/ |
H A D | guest.c | 1023 gpa_t gfn; local 1043 gfn = gpa_to_gfn(guest_ipa); 1048 kvm_pfn_t pfn = gfn_to_pfn_prot(kvm, gfn, write, NULL); 1098 gfn++;
|
H A D | mmu.c | 176 gfn_t gfn, u64 nr_pages) 179 gfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT); 1178 * @gfn_offset: The gfn offset in memory slot 1390 gfn_t gfn; local 1472 gfn = fault_ipa >> PAGE_SHIFT; 1491 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL, 1588 mark_page_dirty_in_slot(kvm, memslot, gfn); 1632 gfn_t gfn; local 1687 gfn = fault_ipa >> PAGE_SHIFT; 1688 memslot = gfn_to_memslot(vcpu->kvm, gfn); 175 kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument [all...] |
/linux-master/arch/arm64/kvm/vgic/ |
H A D | vgic-its.c | 930 gfn_t gfn = gpa >> PAGE_SHIFT; local 935 ret = kvm_is_visible_gfn(its->dev->kvm, gfn);
|
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_mmu.h | 34 unsigned long gfn; member in struct:kvm_ptw_ctx
|
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 68 offset = (addr >> PAGE_SHIFT) - ctx->gfn; 354 * @gfn_offset: The gfn offset in memory slot 372 ctx.gfn = base_gfn; 572 gfn_t gfn = gpa >> PAGE_SHIFT; local 602 slot = gfn_to_memslot(kvm, gfn); 628 mark_page_dirty(kvm, gfn); 671 * Lookup the mapping level for @gfn in the current mm. 695 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, argument 714 hva = __gfn_to_hva_memslot(slot, gfn); 757 static kvm_pte_t *kvm_split_huge(struct kvm_vcpu *vcpu, kvm_pte_t *ptep, gfn_t gfn) argument 807 gfn_t gfn = gpa >> PAGE_SHIFT; local [all...] |
/linux-master/arch/mips/kvm/ |
H A D | mmu.c | 407 * @gfn_offset: The gfn offset in memory slot 515 gfn_t gfn = gpa >> PAGE_SHIFT; local 546 mark_page_dirty(kvm, gfn); 592 gfn_t gfn = gpa >> PAGE_SHIFT; local 632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); 660 mark_page_dirty(kvm, gfn);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_book3s.h | 215 unsigned long gfn); 217 unsigned long gfn); 219 unsigned long gfn); 242 unsigned long gfn, unsigned long psize);
|
H A D | kvm_book3s_uvmem.h | 22 int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn); 81 static inline int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn) argument
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s.c | 432 gfn_t gfn = gpa >> PAGE_SHIFT; local 450 return gfn_to_pfn_prot(vcpu->kvm, gfn, writing, writable);
|
H A D | book3s_64_mmu_host.c | 89 unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; local 128 mark_page_dirty(vcpu->kvm, gfn);
|
H A D | book3s_64_mmu_hv.c | 515 unsigned long gpa, gfn, hva, pfn, hpa; local 578 gfn = gpa >> PAGE_SHIFT; 579 memslot = gfn_to_memslot(kvm, gfn); 604 hva = gfn_to_hva_memslot(memslot, gfn); 616 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL, 781 unsigned long *rmapp, unsigned long gfn) 805 hpte_rpn(ptel, psize) == gfn) { 813 kvmppc_update_dirty_map(memslot, gfn, psize); 822 unsigned long gfn) 828 rmapp = &memslot->arch.rmap[gfn 779 kvmppc_unmap_hpte(struct kvm *kvm, unsigned long i, struct kvm_memory_slot *memslot, unsigned long *rmapp, unsigned long gfn) argument 821 kvm_unmap_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 859 gfn_t gfn; local 875 unsigned long gfn; local 899 kvm_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 956 gfn_t gfn; local 970 kvm_test_age_rmapp(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 1119 unsigned long gfn; local 1160 unsigned long gfn = gpa >> PAGE_SHIFT; local 1192 unsigned long gfn; local 1276 unsigned long gfn = hpte_rpn(guest_rpte, apsize); local [all...] |
H A D | book3s_64_mmu_radix.c | 430 unsigned long gfn = gpa >> PAGE_SHIFT; local 442 memslot = gfn_to_memslot(kvm, gfn); 459 kvmppc_update_dirty_map(memslot, gfn, page_size); 830 unsigned long hva, gfn = gpa >> PAGE_SHIFT; local 848 hva = gfn_to_hva_memslot(memslot, gfn); 855 pfn = __gfn_to_pfn_memslot(memslot, gfn, false, false, NULL, 949 unsigned long gpa, gfn; local 972 gfn = gpa >> PAGE_SHIFT; 977 return kvmppc_send_page_to_uv(kvm, gfn); 980 memslot = gfn_to_memslot(kvm, gfn); 1032 kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 1051 kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 1079 kvm_test_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot, unsigned long gfn) argument 1101 unsigned long gfn = memslot->base_gfn + pagenum; local [all...] |
H A D | book3s_64_vio.c | 358 unsigned long gfn = tce >> PAGE_SHIFT; local 361 memslot = __gfn_to_memslot(kvm_memslots(kvm), gfn); 365 *ua = __gfn_to_hva_memslot(memslot, gfn) |
|
H A D | book3s_hv_nested.c | 1033 unsigned long gfn, end_gfn; local 1038 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; 1039 end_gfn = gfn + (nbytes >> PAGE_SHIFT); 1044 for (; gfn < end_gfn; gfn++) { 1045 unsigned long *rmap = &memslot->arch.rmap[gfn]; 1527 unsigned long n_gpa, gpa, gfn, perm = 0UL; local 1585 gfn = gpa >> PAGE_SHIFT; 1589 memslot = gfn_to_memslot(kvm, gfn); 1639 /* Align gfn t [all...] |
H A D | book3s_hv_rm_mmu.c | 97 unsigned long gfn, unsigned long psize) 104 gfn -= memslot->base_gfn; 105 set_dirty_bits_atomic(memslot->dirty_bitmap, gfn, npages); 113 unsigned long gfn; local 117 gfn = hpte_rpn(hpte_gr, psize); 118 memslot = __gfn_to_memslot(kvm_memslots_raw(kvm), gfn); 120 kvmppc_update_dirty_map(memslot, gfn, psize); 131 unsigned long gfn; local 133 gfn = hpte_rpn(hpte_gr, kvmppc_actual_pgsz(hpte_v, hpte_gr)); 134 memslot = __gfn_to_memslot(kvm_memslots_raw(kvm), gfn); 96 kvmppc_update_dirty_map(const struct kvm_memory_slot *memslot, unsigned long gfn, unsigned long psize) argument 156 unsigned long gfn; local 188 unsigned long i, pa, gpa, gfn, psize; local 889 unsigned long gfn, hva, pa, psize = PAGE_SHIFT; local [all...] |
H A D | book3s_hv_uvmem.c | 289 static void kvmppc_mark_gfn(unsigned long gfn, struct kvm *kvm, argument 295 if (gfn >= p->base_pfn && gfn < p->base_pfn + p->nr_pfns) { 296 unsigned long index = gfn - p->base_pfn; 308 static void kvmppc_gfn_secure_uvmem_pfn(unsigned long gfn, argument 311 kvmppc_mark_gfn(gfn, kvm, KVMPPC_GFN_UVMEM_PFN, uvmem_pfn); 315 static void kvmppc_gfn_secure_mem_pfn(unsigned long gfn, struct kvm *kvm) argument 317 kvmppc_mark_gfn(gfn, kvm, KVMPPC_GFN_MEM_PFN, 0); 321 static void kvmppc_gfn_shared(unsigned long gfn, struct kvm *kvm) argument 323 kvmppc_mark_gfn(gfn, kv 327 kvmppc_gfn_remove(unsigned long gfn, struct kvm *kvm) argument 333 kvmppc_gfn_is_uvmem_pfn(unsigned long gfn, struct kvm *kvm, unsigned long *uvmem_pfn) argument 361 kvmppc_next_nontransitioned_gfn(const struct kvm_memory_slot *memslot, struct kvm *kvm, unsigned long *gfn) argument 394 unsigned long gfn = memslot->base_gfn; local 617 unsigned long uvmem_pfn, gfn; local 797 unsigned long gfn = memslot->base_gfn; local 885 unsigned long gfn = gpa >> page_shift; local 943 unsigned long gfn = gpa >> page_shift; local 1049 unsigned long gfn = gpa >> page_shift; local 1084 kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn) argument [all...] |
H A D | book3s_xive_native.c | 573 gfn_t gfn; local 647 gfn = gpa_to_gfn(kvm_eq.qaddr); 649 page_size = kvm_host_page_size(vcpu, gfn); 656 page = gfn_to_page(kvm, gfn);
|
H A D | booke.c | 1287 gfn_t gfn; local 1316 gfn = gpaddr >> PAGE_SHIFT; 1318 if (kvm_is_visible_gfn(vcpu->kvm, gfn)) { 1344 gfn_t gfn; local 1364 gfn = gpaddr >> PAGE_SHIFT; 1366 if (kvm_is_visible_gfn(vcpu->kvm, gfn)) {
|
H A D | e500_mmu_host.c | 323 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, 353 slot = gfn_to_memslot(vcpu_e500->vcpu.kvm, gfn); 354 hva = gfn_to_hva_memslot(slot, gfn); 381 slot_start = pfn - (gfn - slot->base_gfn); 400 * requested) that will cover gfn, stay within the 401 * range, and for which gfn and pfn are mutually 409 gfn_start = gfn & ~(tsize_pages - 1); 412 if (gfn_start + pfn - gfn < start) 414 if (gfn_end + pfn - gfn > end) 416 if ((gfn 322 kvmppc_e500_shadow_map(struct kvmppc_vcpu_e500 *vcpu_e500, u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, int tlbsel, struct kvm_book3e_206_tlb_entry *stlbe, struct tlbe_ref *ref) argument 559 kvmppc_e500_tlb1_map(struct kvmppc_vcpu_e500 *vcpu_e500, u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, struct kvm_book3e_206_tlb_entry *stlbe, int esel) argument 612 gfn_t gfn = gpaddr >> PAGE_SHIFT; local [all...] |
/linux-master/arch/riscv/kvm/ |
H A D | mmu.c | 617 gfn_t gfn = gpa >> PAGE_SHIFT; local 650 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; 670 hfn = gfn_to_pfn_prot(kvm, gfn, is_write, &writable); 693 mark_page_dirty(kvm, gfn);
|
H A D | vcpu_exit.c | 19 gfn_t gfn; local 23 gfn = fault_addr >> PAGE_SHIFT; 24 memslot = gfn_to_memslot(vcpu->kvm, gfn); 25 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable);
|
H A D | vcpu_sbi_sta.c | 36 gfn_t gfn; local 46 gfn = shmem >> PAGE_SHIFT; 47 hva = kvm_vcpu_gfn_to_hva(vcpu, gfn); 78 kvm_vcpu_mark_page_dirty(vcpu, gfn);
|
/linux-master/arch/s390/kvm/ |
H A D | gaccess.c | 874 * caller needs to ensure that gfn is accessible, so we can 985 const gfn_t gfn = gpa_to_gfn(gpa); local 989 rc = kvm_write_guest_page(kvm, gfn, data, offset, len); 991 rc = kvm_read_guest_page(kvm, gfn, data, offset, len); 1001 gfn_t gfn; local 1005 gfn = gpa >> PAGE_SHIFT; 1006 slot = gfn_to_memslot(kvm, gfn); 1007 hva = gfn_to_hva_memslot_prot(slot, gfn, &writable); 1025 mark_page_dirty_in_slot(kvm, slot, gfn); 1176 gfn_t gfn local [all...] |
H A D | kvm-s390.c | 2215 gfn_t gfn) 2217 return ____gfn_to_memslot(slots, gfn, true); 2214 gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn) argument
|
H A D | priv.c | 1160 unsigned long gfn, hva, res, pgstev, ptev; local 1169 gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT; 1170 hva = gfn_to_hva(vcpu->kvm, gfn); 1205 cbrlo[entries] = gfn << PAGE_SHIFT; 1209 struct kvm_memory_slot *ms = gfn_to_memslot(vcpu->kvm, gfn); 1212 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms)))
|