H A D | kvm_main.c | 358 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) argument 360 if (!kvm_arch_flush_remote_tlbs_range(kvm, gfn, nr_pages)) 640 * {gfn(page) | page intersects with [hva_start, hva_end)} = 1540 * If the memslot gfn is unchanged, rb_replace_node() can be used to 1541 * switch the node in the gfn tree instead of removing the old and 1796 * memslot will be created. Validation of sp->gfn happens in: 2502 /* Set @attributes for the gfn range [@start, @end). */ 2588 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) argument 2590 return __gfn_to_memslot(kvm_memslots(kvm), gfn); 2594 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) argument 2627 kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) argument 2635 kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument 2643 kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) argument 2672 __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages, bool write) argument 2687 gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t *nr_pages) argument 2693 gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) argument 2700 gfn_to_hva(struct kvm *kvm, gfn_t gfn) argument 2706 kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn) argument 2720 gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, bool *writable) argument 2731 gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) argument 2738 kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable) argument 2989 __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable, hva_t *hva) argument 3017 gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, bool *writable) argument 3025 gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn) argument 3032 gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn) argument 3039 kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn) argument 3045 gfn_to_pfn(struct kvm *kvm, gfn_t gfn) argument 3051 kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn) argument 3057 gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, struct page **pages, int nr_pages) argument 3080 gfn_to_page(struct kvm *kvm, gfn_t gfn) argument 3106 kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument 3268 __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, int len) argument 3283 kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, int len) argument 3292 kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset, int len) argument 3303 gfn_t gfn = gpa >> PAGE_SHIFT; local 3323 gfn_t gfn = gpa >> PAGE_SHIFT; local 3341 __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn, void *data, int offset, unsigned long len) argument 3361 gfn_t gfn = gpa >> PAGE_SHIFT; local 3370 __kvm_write_guest_page(struct kvm *kvm, struct kvm_memory_slot *memslot, gfn_t gfn, const void *data, int offset, int len) argument 3387 kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data, int offset, int len) argument 3396 kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data, int offset, int len) argument 3408 gfn_t gfn = gpa >> PAGE_SHIFT; local 3429 gfn_t gfn = gpa >> PAGE_SHIFT; local 3574 gfn_t gfn = gpa >> PAGE_SHIFT; local 3591 mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn) argument 3616 mark_page_dirty(struct kvm *kvm, gfn_t gfn) argument 3625 kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn) argument [all...] |