Lines Matching refs:gpa
227 * Map segment 'segid' starting at 'off' into guest address range [gpa,gpa+len).
230 vm_mmap_memseg(struct vmctx *ctx, vm_paddr_t gpa, int segid, vm_ooffset_t off,
236 memmap.gpa = gpa;
250 error = vm_mmap_getnext(ctx, &gpa, &segid, &off, &len, &prot, &flags);
251 if (error == 0 && gpa == memmap.gpa) {
277 vm_munmap_memseg(struct vmctx *ctx, vm_paddr_t gpa, size_t len)
282 munmap.gpa = gpa;
290 vm_mmap_getnext(struct vmctx *ctx, vm_paddr_t *gpa, int *segid,
297 memmap.gpa = *gpa;
300 *gpa = memmap.gpa;
391 setup_memory_segment(struct vmctx *ctx, vm_paddr_t gpa, size_t len, char *base)
396 /* Map 'len' bytes starting at 'gpa' in the guest address space */
397 error = vm_mmap_memseg(ctx, gpa, VM_SYSMEM, gpa, len, PROT_ALL);
406 ptr = mmap(base + gpa, len, PROT_RW, flags, ctx->fd, gpa);
417 vm_paddr_t gpa;
452 gpa = VM_HIGHMEM_BASE;
454 error = setup_memory_segment(ctx, gpa, len, baseaddr);
460 gpa = 0;
462 error = setup_memory_segment(ctx, gpa, len, baseaddr);
803 vm_get_gpa_pmap(struct vmctx *ctx, uint64_t gpa, uint64_t *pte, int *num)
809 gpapte.gpa = gpa;
824 uint64_t gla, int prot, uint64_t *gpa, int *fault)
837 *gpa = gg.gpa;
845 uint64_t gla, int prot, uint64_t *gpa, int *fault)
858 *gpa = gg.gpa;
874 uint64_t gpa, off;
884 error = vm_gla2gpa(vcpu, paging, gla, prot, &gpa, fault);
888 off = gpa & PAGE_MASK;
891 va = vm_map_gpa(vcpu->ctx, gpa, n);