Searched refs:vm (Results 151 - 175 of 485) sorted by relevance

1234567891011>>

/linux-master/tools/testing/selftests/kvm/aarch64/
H A Dhypercalls.c160 vm_userspace_mem_region_add(vcpu->vm, VM_MEM_SRC_ANONYMOUS, ST_GPA_BASE, 1, gpages, 0);
236 struct kvm_vm *vm; local
238 vm = vm_create_with_one_vcpu(vcpu, guest_code);
242 return vm;
245 static void test_guest_stage(struct kvm_vm **vm, struct kvm_vcpu **vcpu) argument
253 sync_global_to_guest(*vm, stage);
261 kvm_vm_free(*vm);
262 *vm = test_vm_create(vcpu);
275 struct kvm_vm *vm; local
279 vm
[all...]
H A Dvgic_irq.c83 static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc,
516 static void kvm_irq_line_check(struct kvm_vm *vm, uint32_t intid, int level, argument
522 kvm_arm_irq_line(vm, intid, level);
528 ret = _kvm_arm_irq_line(vm, intid, level);
559 static void kvm_set_gsi_routing_irqchip_check(struct kvm_vm *vm, argument
574 kvm_gsi_routing_write(vm, routing);
576 ret = _kvm_gsi_routing_write(vm, routing);
604 static void kvm_routing_and_irqfd_check(struct kvm_vm *vm, argument
621 kvm_set_gsi_routing_irqchip_check(vm, intid, num,
641 vm_ioctl(vm, KVM_IRQF
670 struct kvm_vm *vm = vcpu->vm; local
720 kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, struct kvm_inject_args *args) argument
743 struct kvm_vm *vm; local
[all...]
/linux-master/tools/testing/selftests/kvm/x86_64/
H A Duserspace_msr_exit_test.c519 struct kvm_vm *vm = vcpu->vm; local
523 sync_global_to_guest(vm, fep_available);
527 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MSR, KVM_MSR_EXIT_REASON_FILTER);
532 vm_ioctl(vm, KVM_X86_SET_MSR_FILTER, &filter_allow);
534 vm_init_descriptor_tables(vm);
537 vm_install_exception_handler(vm, GP_VECTOR, guest_gp_handler);
556 vm_install_exception_handler(vm, GP_VECTOR, guest_fep_gp_handler);
587 vm_ioctl(vcpu->vm, KVM_X86_SET_MSR_FILTER, &no_filter_deny);
637 struct kvm_vm *vm local
678 struct kvm_vm *vm = vcpu->vm; local
709 run_user_space_msr_flag_test(struct kvm_vm *vm) argument
726 run_msr_filter_flag_test(struct kvm_vm *vm) argument
766 struct kvm_vm *vm = vcpu->vm; local
[all...]
H A Dprivate_mem_conversions_test.c298 struct kvm_vm *vm = vcpu->vm; local
305 vm_guest_mem_fallocate(vm, gpa, size, map_shared);
308 vm_set_memory_attributes(vm, gpa, size,
319 struct kvm_vm *vm = vcpu->vm; local
349 for (i = 0; i < size; i += vm->page_size) {
350 size_t nr_bytes = min_t(size_t, vm->page_size, size - i);
351 uint8_t *hva = addr_gpa2hva(vm, gpa + i);
383 struct kvm_vm *vm; local
[all...]
H A Dxapic_state_test.c166 struct kvm_vm *vm; local
169 vm = vm_create_with_vcpus(NR_VCPUS, NULL, vcpus);
170 vm_enable_cap(vm, KVM_CAP_X2APIC_API, KVM_X2APIC_API_USE_32BIT_IDS);
185 kvm_vm_free(vm);
194 struct kvm_vm *vm; local
196 vm = vm_create_with_one_vcpu(&x.vcpu, x2apic_guest_code);
198 kvm_vm_free(vm);
205 vm = vm_create_with_one_vcpu(&x.vcpu, xapic_guest_code);
210 virt_pg_map(vm, APIC_DEFAULT_GPA, APIC_DEFAULT_GPA);
212 kvm_vm_free(vm);
[all...]
H A Dxen_shinfo_test.c411 struct kvm_vm *vm = (struct kvm_vm *)arg; local
436 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_activate_gfn);
438 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_deactivate_gfn);
441 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_activate_hva);
443 __vm_ioctl(vm, KVM_XEN_HVM_SET_ATTR, &cache_deactivate_hva);
454 struct kvm_vm *vm; local
473 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
476 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS,
478 virt_map(vm, SHINFO_REGION_GVA, SHINFO_REGION_GPA, 3);
480 shinfo = addr_gpa2hva(vm, SHINFO_VADD
[all...]
/linux-master/tools/testing/selftests/kvm/
H A Dsteal_time.c74 sync_global_to_guest(vcpu->vm, st_gva[i]);
83 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
85 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]);
171 struct kvm_vm *vm = vcpu->vm; local
185 sync_global_to_guest(vm, st_gva[i]);
198 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
200 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]);
284 st_gpa[i] = addr_gva2gpa(vcpu->vm, (vm_vaddr_t)st_gva[i]);
285 sync_global_to_guest(vcpu->vm, st_gv
289 steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
344 struct kvm_vm *vm; local
[all...]
H A Dguest_memfd_test.c125 static void test_create_guest_memfd_invalid(struct kvm_vm *vm) argument
133 fd = __vm_create_guest_memfd(vm, size, 0);
140 fd = __vm_create_guest_memfd(vm, page_size, flag);
147 static void test_create_guest_memfd_multiple(struct kvm_vm *vm) argument
152 fd1 = __vm_create_guest_memfd(vm, 4096, 0);
159 fd2 = __vm_create_guest_memfd(vm, 8192, 0);
180 struct kvm_vm *vm; local
187 vm = vm_create_barebones();
189 test_create_guest_memfd_invalid(vm);
190 test_create_guest_memfd_multiple(vm);
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_csa.c65 int amdgpu_map_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
74 r = amdgpu_vm_lock_pd(vm, &exec, 0);
84 *bo_va = amdgpu_vm_bo_add(adev, vm, bo);
105 int amdgpu_unmap_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
114 r = amdgpu_vm_lock_pd(vm, &exec, 0);
H A Damdgpu_seq64.c55 * @vm: vm pointer
63 int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
77 r = amdgpu_vm_lock_pd(vm, &exec, 0);
85 *bo_va = amdgpu_vm_bo_add(adev, vm, bo);
122 struct amdgpu_vm *vm; local
134 vm = &fpriv->vm;
138 r = amdgpu_vm_lock_pd(vm, &exec, 0);
H A Damdgpu_ids.c262 * @vm: vm to allocate id for
270 static int amdgpu_vmid_grab_reserved(struct amdgpu_vm *vm, argument
280 bool needs_flush = vm->use_cpu_for_update;
281 uint64_t updates = amdgpu_vm_tlb_seq(vm);
285 if ((*id)->owner != vm->immediate.fence_context ||
323 * @vm: vm to allocate id for
331 static int amdgpu_vmid_grab_used(struct amdgpu_vm *vm, argument
341 uint64_t updates = amdgpu_vm_tlb_seq(vm);
394 amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring, struct amdgpu_job *job, struct dma_fence **fence) argument
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_exec_queue.c37 struct xe_vm *vm,
89 if (vm)
90 q->vm = xe_vm_get(vm);
102 if (q->vm)
103 xe_vm_put(q->vm);
113 err = xe_lrc_init(q->lrc + i, q->hwe, q, q->vm, SZ_16K);
123 * Normally the user vm holds an rpm ref to keep the device
124 * awake, and the context holds a ref for the vm, however for
125 * some engines we use the kernels migrate vm underneat
36 __xe_exec_queue_alloc(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument
141 xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument
173 xe_exec_queue_create_class(struct xe_device *xe, struct xe_gt *gt, struct xe_vm *vm, enum xe_engine_class class, u32 flags) argument
555 struct xe_vm *vm, *migrate_vm; local
820 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue *q, struct xe_vm *vm) argument
834 xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) argument
867 xe_exec_queue_last_fence_get(struct xe_exec_queue *q, struct xe_vm *vm) argument
892 xe_exec_queue_last_fence_set(struct xe_exec_queue *q, struct xe_vm *vm, struct dma_fence *fence) argument
[all...]
H A Dxe_trace.h86 __field(struct xe_vm *, vm)
92 __entry->vm = bo->vm;
95 TP_printk("size=%zu, flags=0x%02x, vm=%p",
96 __entry->size, __entry->flags, __entry->vm)
487 TP_PROTO(struct xe_vm *vm),
488 TP_ARGS(vm),
491 __field(struct xe_vm *, vm)
496 __entry->vm = vm;
[all...]
H A Dxe_bo.h17 * xe_vm_assert_held(vm) - Assert that the vm's reservation object is held.
18 * @vm: The vm
20 #define xe_vm_assert_held(vm) dma_resv_assert_held(xe_vm_resv(vm))
105 struct xe_tile *tile, struct xe_vm *vm,
109 struct xe_vm *vm, size_t size,
112 struct xe_vm *vm, size_t size,
115 struct xe_vm *vm, size_
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_trace.h64 __field(struct i915_address_space *, vm)
72 __entry->vm = vma->vm;
78 TP_printk("obj=%p, offset=0x%016llx size=0x%llx%s vm=%p",
81 __entry->vm)
90 __field(struct i915_address_space *, vm)
97 __entry->vm = vma->vm;
102 TP_printk("obj=%p, offset=0x%016llx size=0x%llx vm=%p",
103 __entry->obj, __entry->offset, __entry->size, __entry->vm)
[all...]
H A Di915_vgpu.c151 struct drm_i915_private *dev_priv = ggtt->vm.i915;
161 ggtt->vm.reserved -= node->size;
174 struct drm_i915_private *dev_priv = ggtt->vm.i915;
177 if (!intel_vgpu_active(ggtt->vm.i915))
190 struct drm_i915_private *dev_priv = ggtt->vm.i915;
200 ret = i915_gem_gtt_reserve(&ggtt->vm, NULL, node,
204 ggtt->vm.reserved += size;
255 struct drm_i915_private *dev_priv = ggtt->vm.i915;
257 unsigned long ggtt_end = ggtt->vm.total;
263 if (!intel_vgpu_active(ggtt->vm
[all...]
/linux-master/drivers/gpu/drm/exynos/
H A Dexynos_drm_dpi.c33 struct videomode *vm; member in struct:exynos_dpi
69 if (ctx->vm) {
78 drm_display_mode_from_videomode(ctx->vm, mode);
167 struct videomode *vm; local
172 vm = devm_kzalloc(dev, sizeof(*ctx->vm), GFP_KERNEL);
173 if (!vm)
176 ret = of_get_videomode(dn, vm, 0);
178 devm_kfree(dev, vm);
182 ctx->vm
[all...]
/linux-master/arch/s390/kernel/
H A Dearly.c110 if (!memcmp(vmms->vm[0].cpi, "\xd2\xe5\xd4", 3))
112 else if (!memcmp(vmms->vm[0].cpi, "\xa9\x61\xe5\xd4", 4))
137 struct sysinfo_3_2_2 *vm = (struct sysinfo_3_2_2 *)&sysinfo_page; local
150 if (stsi(vm, 3, 2, 2) == 0 && vm->count) {
151 EBCASC(vm->vm[0].cpi, sizeof(vm->vm[0].cpi));
152 sprintf(hvstr, "%-16.16s", vm
[all...]
H A Dlgr.c44 } vm[VM_LEVEL_MAX]; member in struct:lgr_info
105 cpascii(lgr_info->vm[i].name, si->vm[i].name,
106 sizeof(si->vm[i].name));
107 cpascii(lgr_info->vm[i].cpi, si->vm[i].cpi,
108 sizeof(si->vm[i].cpi));
/linux-master/drivers/gpu/drm/i915/gt/
H A Dselftest_tlb.c36 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length),
41 i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE);
50 batch = i915_gem_object_create_internal(ce->vm->i915, 4096);
54 vma = i915_vma_instance(batch, ce->vm, NULL);
67 addr = igt_random_offset(prng, addr, min(ce->vm->total, BIT_ULL(48)),
168 ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags);
171 tlbinv(ce->vm, addr & -length, length);
230 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length))
283 va = i915_vma_instance(A, &ppgtt->vm, NUL
32 pte_tlbinv(struct intel_context *ce, struct i915_vma *va, struct i915_vma *vb, u64 align, void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), u64 length, struct rnd_state *prng) argument
228 mem_tlbinv(struct intel_gt *gt, struct drm_i915_gem_object *(*create_fn)(struct intel_gt *), void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) argument
363 tlbinv_full(struct i915_address_space *vm, u64 addr, u64 length) argument
[all...]
/linux-master/drivers/gpu/drm/xe/tests/
H A Dxe_bo.c190 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); local
198 xe_vm_lock(vm, false);
199 bo = xe_bo_create_user(xe, NULL, vm, 0x10000,
203 xe_vm_unlock(vm);
275 down_read(&vm->lock);
276 xe_vm_lock(vm, false);
277 err = xe_bo_validate(bo, bo->vm, false);
278 xe_vm_unlock(vm);
279 up_read(&vm->lock);
321 xe_vm_put(vm);
[all...]
/linux-master/drivers/gpu/drm/arm/
H A Dhdlcd_crtc.c131 struct videomode vm; local
134 vm.vfront_porch = m->crtc_vsync_start - m->crtc_vdisplay;
135 vm.vback_porch = m->crtc_vtotal - m->crtc_vsync_end;
136 vm.vsync_len = m->crtc_vsync_end - m->crtc_vsync_start;
137 vm.hfront_porch = m->crtc_hsync_start - m->crtc_hdisplay;
138 vm.hback_porch = m->crtc_htotal - m->crtc_hsync_end;
139 vm.hsync_len = m->crtc_hsync_end - m->crtc_hsync_start;
153 hdlcd_write(hdlcd, HDLCD_REG_V_BACK_PORCH, vm.vback_porch - 1);
154 hdlcd_write(hdlcd, HDLCD_REG_V_FRONT_PORCH, vm.vfront_porch - 1);
155 hdlcd_write(hdlcd, HDLCD_REG_V_SYNC, vm
[all...]
/linux-master/drivers/gpu/drm/lima/
H A Dlima_gem.c21 int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm) argument
89 if (vm) {
90 ret = lima_vm_map_bo(vm, bo, old_size >> PAGE_SHIFT);
167 struct lima_vm *vm = priv->vm; local
169 return lima_vm_bo_add(vm, bo, true);
176 struct lima_vm *vm = priv->vm; local
178 lima_vm_bo_del(vm, bo);
246 struct lima_vm *vm local
302 struct lima_vm *vm = priv->vm; local
[all...]
/linux-master/tools/testing/selftests/mm/
H A Drun_vmtests.sh161 nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages)
166 echo 3 > /proc/sys/vm/drop_caches
167 if ! echo $((lackpgs + nr_hugepgs)) > /proc/sys/vm/nr_hugepages; then
216 echo 3 > /proc/sys/vm/drop_caches
218 echo 1 > /proc/sys/vm/compact_memory
265 nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages)
267 echo 1 > /proc/sys/vm/nr_hugepages
271 echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages
304 echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages
338 prev_policy=$(cat /proc/sys/vm/overcommit_memor
[all...]
/linux-master/drivers/gpu/drm/omapdrm/dss/
H A Dhdmi5_core.c239 video_cfg->v_fc_config.vm = cfg->vm;
243 video_cfg->hblank = cfg->vm.hfront_porch +
244 cfg->vm.hback_porch + cfg->vm.hsync_len;
246 video_cfg->vblank = cfg->vm.vsync_len + cfg->vm.vfront_porch +
247 cfg->vm.vback_porch;
250 if (cfg->vm.flags & DISPLAY_FLAGS_INTERLACED) {
255 video_cfg->v_fc_config.vm
276 const struct videomode *vm = &cfg->v_fc_config.vm; local
568 struct videomode vm; local
[all...]

Completed in 180 milliseconds

1234567891011>>