/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
H A D | page_track.h | 47 struct intel_vgpu *vgpu, unsigned long gfn); 50 unsigned long gfn, gvt_page_track_handler_t handler, 53 unsigned long gfn); 55 int intel_vgpu_enable_page_track(struct intel_vgpu *vgpu, unsigned long gfn); 56 int intel_vgpu_disable_page_track(struct intel_vgpu *vgpu, unsigned long gfn);
|
H A D | page_track.c | 34 * @gfn: the gfn of guest page 40 struct intel_vgpu *vgpu, unsigned long gfn) 42 return radix_tree_lookup(&vgpu->page_track_tree, gfn); 48 * @gfn: the gfn of guest page 55 int intel_vgpu_register_page_track(struct intel_vgpu *vgpu, unsigned long gfn, argument 61 track = intel_vgpu_find_page_track(vgpu, gfn); 72 ret = radix_tree_insert(&vgpu->page_track_tree, gfn, track); 84 * @gfn 39 intel_vgpu_find_page_track( struct intel_vgpu *vgpu, unsigned long gfn) argument 87 intel_vgpu_unregister_page_track(struct intel_vgpu *vgpu, unsigned long gfn) argument 108 intel_vgpu_enable_page_track(struct intel_vgpu *vgpu, unsigned long gfn) argument 135 intel_vgpu_disable_page_track(struct intel_vgpu *vgpu, unsigned long gfn) argument [all...] |
H A D | hypercall.h | 59 int (*enable_page_track)(unsigned long handle, u64 gfn); 60 int (*disable_page_track)(unsigned long handle, u64 gfn); 65 unsigned long (*gfn_to_mfn)(unsigned long handle, unsigned long gfn); 67 int (*dma_map_guest_page)(unsigned long handle, unsigned long gfn, 73 int (*map_gfn_to_mfn)(unsigned long handle, unsigned long gfn, 81 bool (*is_valid_gfn)(unsigned long handle, unsigned long gfn);
|
H A D | mpt.h | 159 * @gfn: the gfn of guest 165 struct intel_vgpu *vgpu, unsigned long gfn) 167 return intel_gvt_host.mpt->enable_page_track(vgpu->handle, gfn); 173 * @gfn: the gfn of guest 179 struct intel_vgpu *vgpu, unsigned long gfn) 181 return intel_gvt_host.mpt->disable_page_track(vgpu->handle, gfn); 225 struct intel_vgpu *vgpu, unsigned long gfn) 227 return intel_gvt_host.mpt->gfn_to_mfn(vgpu->handle, gfn); 164 intel_gvt_hypervisor_enable_page_track( struct intel_vgpu *vgpu, unsigned long gfn) argument 178 intel_gvt_hypervisor_disable_page_track( struct intel_vgpu *vgpu, unsigned long gfn) argument 224 intel_gvt_hypervisor_gfn_to_mfn( struct intel_vgpu *vgpu, unsigned long gfn) argument 240 intel_gvt_hypervisor_dma_map_guest_page( struct intel_vgpu *vgpu, unsigned long gfn, unsigned long size, dma_addr_t *dma_addr) argument 285 intel_gvt_hypervisor_map_gfn_to_mfn( struct intel_vgpu *vgpu, unsigned long gfn, unsigned long mfn, unsigned int nr, bool map) argument 388 intel_gvt_hypervisor_is_valid_gfn( struct intel_vgpu *vgpu, unsigned long gfn) argument [all...] |
H A D | kvmgt.c | 92 gfn_t gfn; member in struct:kvmgt_pgfn 110 gfn_t gfn; member in struct:gvt_dma 125 static void gvt_unpin_guest_page(struct intel_vgpu *vgpu, unsigned long gfn, argument 135 unsigned long cur_gfn = gfn + npage; 143 static int gvt_pin_guest_page(struct intel_vgpu *vgpu, unsigned long gfn, argument 157 unsigned long cur_gfn = gfn + npage; 163 gvt_vgpu_err("vfio_pin_pages failed for gfn 0x%lx, ret %d\n", 188 gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZE); 192 static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn, argument 199 ret = gvt_pin_guest_page(vgpu, gfn, siz 215 gvt_dma_unmap_page(struct intel_vgpu *vgpu, unsigned long gfn, dma_addr_t dma_addr, unsigned long size) argument 243 __gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) argument 261 __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, dma_addr_t dma_addr, unsigned long size) argument 364 __kvmgt_protect_table_find(struct kvmgt_guest_info *info, gfn_t gfn) argument 378 kvmgt_gfn_is_write_protected(struct kvmgt_guest_info *info, gfn_t gfn) argument 387 kvmgt_protect_table_add(struct kvmgt_guest_info *info, gfn_t gfn) argument 402 kvmgt_protect_table_del(struct kvmgt_guest_info *info, gfn_t gfn) argument 1622 kvmgt_page_track_add(unsigned long handle, u64 gfn) argument 1656 kvmgt_page_track_remove(unsigned long handle, u64 gfn) argument 1707 gfn_t gfn; local 1855 kvmgt_gfn_to_pfn(unsigned long handle, unsigned long gfn) argument 1872 kvmgt_dma_map_guest_page(unsigned long handle, unsigned long gfn, unsigned long size, dma_addr_t *dma_addr) argument 2022 kvmgt_is_valid_gfn(unsigned long handle, unsigned long gfn) argument [all...] |
H A D | gtt.c | 666 spt->guest_page.gfn << I915_GTT_PAGE_SHIFT, 695 spt->guest_page.gfn << I915_GTT_PAGE_SHIFT, 751 if (spt->guest_page.gfn) { 755 intel_vgpu_unregister_page_track(spt->vgpu, spt->guest_page.gfn); 801 /* Find a spt by guest gfn. */ 803 struct intel_vgpu *vgpu, unsigned long gfn) 807 track = intel_vgpu_find_page_track(vgpu, gfn); 873 /* Allocate shadow page table associated with specific gfn. */ 876 unsigned long gfn, bool guest_pde_ips) 888 ret = intel_vgpu_register_page_track(vgpu, gfn, 802 intel_vgpu_find_spt_by_gfn( struct intel_vgpu *vgpu, unsigned long gfn) argument 874 ppgtt_alloc_spt_gfn( struct intel_vgpu *vgpu, enum intel_gvt_gtt_type type, unsigned long gfn, bool guest_pde_ips) argument 1255 unsigned long gfn, page_size = PAGE_SIZE; local 1310 unsigned long gfn, i; local 2219 unsigned long gma, gfn; local [all...] |
H A D | opregion.c | 277 vgpu_opregion(vgpu)->gfn[i], 310 vgpu_opregion(vgpu)->gfn[i] = (gpa >> PAGE_SHIFT) + i; 321 vgpu_opregion(vgpu)->gfn[i] = (gpa >> PAGE_SHIFT) + i; 487 scic_pa = (vgpu_opregion(vgpu)->gfn[0] << PAGE_SHIFT) + 489 parm_pa = (vgpu_opregion(vgpu)->gfn[0] << PAGE_SHIFT) +
|
H A D | trace.h | 66 TP_printk("VM%d [alloc] spt %p type %d mfn 0x%lx gfn 0x%lx\n", 155 TP_PROTO(int id, char *action, void *spt, unsigned long gfn, 158 TP_ARGS(id, action, spt, gfn, type), 166 "VM%d [%s] spt %p gfn 0x%lx type %d\n", 167 id, action, spt, gfn, type);
|
H A D | gtt.h | 249 unsigned long gfn; member in struct:intel_vgpu_ppgtt_spt::__anon779
|
H A D | gvt.h | 124 u32 gfn[INTEL_GVT_OPREGION_PAGES]; member in struct:intel_vgpu_opregion
|
/netbsd-current/sys/external/mit/xen-include-public/dist/xen/include/public/ |
H A D | vm_event.h | 236 uint64_t gfn; member in struct:vm_event_mem_access 251 uint64_t gfn; member in struct:vm_event_singlestep 255 uint64_t gfn; member in struct:vm_event_debug 308 uint64_t gfn; member in struct:vm_event_paging 314 uint64_t gfn; member in struct:vm_event_sharing
|
H A D | memory.h | 402 uint64_aligned_t gfn; /* IN: gfn of page being operated on */ member in struct:xen_mem_paging_op 509 uint64_aligned_t gfn; /* IN: gfn to nominate */ member in union:xen_mem_sharing_op::__anon301::mem_sharing_op_nominate::__anon302 515 uint64_aligned_t source_gfn; /* IN: the gfn of the source page */ 517 uint64_aligned_t client_gfn; /* IN: the client gfn */ 522 uint64_aligned_t first_gfn; /* IN: the first gfn */ 523 uint64_aligned_t last_gfn; /* IN: the last gfn */ 530 uint64_aligned_t gfn; /* IN: gfn t member in union:xen_mem_sharing_op::__anon301::mem_sharing_op_debug::__anon303 [all...] |
H A D | domctl.h | 1083 uint64_aligned_t gfn; /* IN - guest gfn to be used as a member in struct:xen_domctl_vuart_op
|
/netbsd-current/sys/external/mit/xen-include-public/dist/xen/include/public/hvm/ |
H A D | hvm_op.h | 221 /* #VE info area gfn */ 222 uint64_t gfn; member in struct:xen_hvm_altp2m_vcpu_enable_notify 243 /* gfn */ 244 uint64_t gfn; member in struct:xen_hvm_altp2m_set_mem_access 271 /* old gfn */ 273 /* new gfn, INVALID_GFN (~0UL) means revert */ 295 /* Change a p2m entry to have a different gfn->mfn mapping */
|