Lines Matching defs:memslot
410 * memslot->userspace_addr:
415 * memslot->base_gfn << PAGE_SIZE:
637 static bool fault_supports_huge_mapping(struct kvm_memory_slot *memslot,
643 if (kvm_slot_dirty_track_enabled(memslot) && write)
646 if (kvm_hugepage_capable(memslot))
649 if (kvm_hugepage_incapable(memslot))
652 start = memslot->userspace_addr;
653 end = start + memslot->npages * PAGE_SIZE;
657 * by the memslot. This means we have to prohibit block size mappings
707 * Note, using the already-retrieved memslot and __gfn_to_hva_memslot()
712 * read-only memslot.
809 struct kvm_memory_slot *memslot;
818 memslot = gfn_to_memslot(kvm, gfn);
819 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writeable);
893 if (!fault_supports_huge_mapping(memslot, hva, write)) {
896 level = host_pfn_mapping_level(kvm, gfn, memslot);
923 mark_page_dirty_in_slot(kvm, memslot, gfn);
948 void kvm_arch_sync_dirty_log(struct kvm *kvm, struct kvm_memory_slot *memslot)
953 const struct kvm_memory_slot *memslot)