Searched refs:vm (Results 151 - 175 of 497) sorted by relevance

1234567891011>>

/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_context.h160 return rcu_dereference_protected(ctx->vm, lockdep_is_held(&ctx->mutex));
165 GEM_BUG_ON(!!ctx->vm != HAS_FULL_PPGTT(ctx->i915));
167 return !!ctx->vm;
173 struct i915_address_space *vm; local
175 vm = ctx->vm;
176 if (!vm)
177 vm = &to_gt(ctx->i915)->ggtt->vm;
178 vm
[all...]
/linux-master/tools/testing/selftests/kvm/x86_64/
H A Dvmx_invalid_nested_guest_state.c14 static struct kvm_vm *vm; variable in typeref:struct:kvm_vm
63 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code);
66 vcpu_alloc_vmx(vm, &vmx_pages_gva);
H A Drecalc_apic_map_test.c39 struct kvm_vm *vm; local
51 vm = vm_create_with_vcpus(KVM_MAX_VCPUS, NULL, vcpus);
71 kvm_vm_free(vm);
H A Duserspace_io_test.c58 struct kvm_vm *vm; local
61 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
101 kvm_vm_free(vm);
H A Dkvm_pv_test.c139 struct kvm_vm *vm; local
148 vm = vm_create_with_one_vcpu(&vcpu, guest_main);
160 vm_enable_cap(vm, KVM_CAP_X86_DISABLE_EXITS, KVM_X86_DISABLE_EXITS_HLT);
170 kvm_vm_free(vm);
176 struct kvm_vm *vm; local
180 vm = vm_create_with_one_vcpu(&vcpu, guest_main);
187 kvm_vm_free(vm);
H A Duserspace_msr_exit_test.c515 struct kvm_vm *vm = vcpu->vm; local
521 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MSR, KVM_MSR_EXIT_REASON_FILTER);
526 vm_ioctl(vm, KVM_X86_SET_MSR_FILTER, &filter_allow);
528 vm_install_exception_handler(vm, GP_VECTOR, guest_gp_handler);
547 vm_install_exception_handler(vm, GP_VECTOR, guest_fep_gp_handler);
578 vm_ioctl(vcpu->vm, KVM_X86_SET_MSR_FILTER, &no_filter_deny);
628 struct kvm_vm *vm = vcpu->vm; local
634 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MS
669 struct kvm_vm *vm = vcpu->vm; local
700 run_user_space_msr_flag_test(struct kvm_vm *vm) argument
717 run_msr_filter_flag_test(struct kvm_vm *vm) argument
757 struct kvm_vm *vm = vcpu->vm; local
[all...]
H A Dprivate_mem_conversions_test.c297 struct kvm_vm *vm = vcpu->vm; local
304 vm_guest_mem_fallocate(vm, gpa, size, map_shared);
307 vm_set_memory_attributes(vm, gpa, size,
318 struct kvm_vm *vm = vcpu->vm; local
348 for (i = 0; i < size; i += vm->page_size) {
349 size_t nr_bytes = min_t(size_t, vm->page_size, size - i);
350 uint8_t *hva = addr_gpa2hva(vm, gpa + i);
382 struct kvm_vm *vm; local
[all...]
/linux-master/tools/testing/selftests/kvm/lib/s390x/
H A Ddiag318_test_handler.c29 struct kvm_vm *vm; local
34 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
49 kvm_vm_free(vm);
/linux-master/tools/testing/selftests/kvm/aarch64/
H A Dhypercalls.c160 vm_userspace_mem_region_add(vcpu->vm, VM_MEM_SRC_ANONYMOUS, ST_GPA_BASE, 1, gpages, 0);
236 struct kvm_vm *vm; local
238 vm = vm_create_with_one_vcpu(vcpu, guest_code);
242 return vm;
245 static void test_guest_stage(struct kvm_vm **vm, struct kvm_vcpu **vcpu) argument
253 sync_global_to_guest(*vm, stage);
261 kvm_vm_free(*vm);
262 *vm = test_vm_create(vcpu);
275 struct kvm_vm *vm; local
279 vm
[all...]
H A Dvgic_irq.c77 static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc,
510 static void kvm_irq_line_check(struct kvm_vm *vm, uint32_t intid, int level, argument
516 kvm_arm_irq_line(vm, intid, level);
522 ret = _kvm_arm_irq_line(vm, intid, level);
553 static void kvm_set_gsi_routing_irqchip_check(struct kvm_vm *vm, argument
568 kvm_gsi_routing_write(vm, routing);
570 ret = _kvm_gsi_routing_write(vm, routing);
598 static void kvm_routing_and_irqfd_check(struct kvm_vm *vm, argument
615 kvm_set_gsi_routing_irqchip_check(vm, intid, num,
635 vm_ioctl(vm, KVM_IRQF
664 struct kvm_vm *vm = vcpu->vm; local
714 kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, struct kvm_inject_args *args) argument
737 struct kvm_vm *vm; local
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_csa.c65 int amdgpu_map_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
74 r = amdgpu_vm_lock_pd(vm, &exec, 0);
84 *bo_va = amdgpu_vm_bo_add(adev, vm, bo);
105 int amdgpu_unmap_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
114 r = amdgpu_vm_lock_pd(vm, &exec, 0);
H A Damdgpu_seq64.c55 * @vm: vm pointer
63 int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument
77 r = amdgpu_vm_lock_pd(vm, &exec, 0);
85 *bo_va = amdgpu_vm_bo_add(adev, vm, bo);
122 struct amdgpu_vm *vm; local
134 vm = &fpriv->vm;
138 r = amdgpu_vm_lock_pd(vm, &exec, 0);
/linux-master/tools/testing/selftests/kvm/
H A Dsteal_time.c76 sync_global_to_guest(vcpu->vm, st_gva[i]);
85 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
87 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]);
171 struct kvm_vm *vm = vcpu->vm; local
185 sync_global_to_guest(vm, st_gva[i]);
198 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
200 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]);
284 st_gpa[i] = addr_gva2gpa(vcpu->vm, (vm_vaddr_t)st_gva[i]);
285 sync_global_to_guest(vcpu->vm, st_gv
289 steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument
344 struct kvm_vm *vm; local
[all...]
H A Dguest_memfd_test.c123 static void test_create_guest_memfd_invalid(struct kvm_vm *vm) argument
131 fd = __vm_create_guest_memfd(vm, size, 0);
138 fd = __vm_create_guest_memfd(vm, page_size, flag);
145 static void test_create_guest_memfd_multiple(struct kvm_vm *vm) argument
150 fd1 = __vm_create_guest_memfd(vm, 4096, 0);
157 fd2 = __vm_create_guest_memfd(vm, 8192, 0);
178 struct kvm_vm *vm; local
185 vm = vm_create_barebones();
187 test_create_guest_memfd_invalid(vm);
188 test_create_guest_memfd_multiple(vm);
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_exec_queue.c38 if (q->vm)
39 xe_vm_put(q->vm);
44 struct xe_vm *vm,
84 if (vm)
85 q->vm = xe_vm_get(vm);
113 err = xe_lrc_init(q->lrc + i, q->hwe, q, q->vm, SZ_16K);
123 * Normally the user vm holds an rpm ref to keep the device
124 * awake, and the context holds a ref for the vm, however for
125 * some engines we use the kernels migrate vm underneat
43 __xe_exec_queue_alloc(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument
141 xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument
173 xe_exec_queue_create_class(struct xe_device *xe, struct xe_gt *gt, struct xe_vm *vm, enum xe_engine_class class, u32 flags) argument
547 struct xe_vm *vm, *migrate_vm; local
812 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue *q, struct xe_vm *vm) argument
826 xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) argument
859 xe_exec_queue_last_fence_get(struct xe_exec_queue *q, struct xe_vm *vm) argument
884 xe_exec_queue_last_fence_set(struct xe_exec_queue *q, struct xe_vm *vm, struct dma_fence *fence) argument
[all...]
H A Dxe_trace.h86 __field(struct xe_vm *, vm)
92 __entry->vm = bo->vm;
95 TP_printk("size=%zu, flags=0x%02x, vm=%p",
96 __entry->size, __entry->flags, __entry->vm)
487 TP_PROTO(struct xe_vm *vm),
488 TP_ARGS(vm),
491 __field(struct xe_vm *, vm)
496 __entry->vm = vm;
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_trace.h64 __field(struct i915_address_space *, vm)
72 __entry->vm = vma->vm;
78 TP_printk("obj=%p, offset=0x%016llx size=0x%llx%s vm=%p",
81 __entry->vm)
90 __field(struct i915_address_space *, vm)
97 __entry->vm = vma->vm;
102 TP_printk("obj=%p, offset=0x%016llx size=0x%llx vm=%p",
103 __entry->obj, __entry->offset, __entry->size, __entry->vm)
[all...]
H A Di915_vgpu.c151 struct drm_i915_private *dev_priv = ggtt->vm.i915;
161 ggtt->vm.reserved -= node->size;
174 struct drm_i915_private *dev_priv = ggtt->vm.i915;
177 if (!intel_vgpu_active(ggtt->vm.i915))
190 struct drm_i915_private *dev_priv = ggtt->vm.i915;
200 ret = i915_gem_gtt_reserve(&ggtt->vm, NULL, node,
204 ggtt->vm.reserved += size;
255 struct drm_i915_private *dev_priv = ggtt->vm.i915;
257 unsigned long ggtt_end = ggtt->vm.total;
263 if (!intel_vgpu_active(ggtt->vm
[all...]
/linux-master/drivers/gpu/drm/exynos/
H A Dexynos_drm_dpi.c33 struct videomode *vm; member in struct:exynos_dpi
69 if (ctx->vm) {
78 drm_display_mode_from_videomode(ctx->vm, mode);
167 struct videomode *vm; local
172 vm = devm_kzalloc(dev, sizeof(*ctx->vm), GFP_KERNEL);
173 if (!vm)
176 ret = of_get_videomode(dn, vm, 0);
178 devm_kfree(dev, vm);
182 ctx->vm
[all...]
/linux-master/arch/s390/kernel/
H A Dearly.c110 if (!memcmp(vmms->vm[0].cpi, "\xd2\xe5\xd4", 3))
112 else if (!memcmp(vmms->vm[0].cpi, "\xa9\x61\xe5\xd4", 4))
137 struct sysinfo_3_2_2 *vm = (struct sysinfo_3_2_2 *)&sysinfo_page; local
150 if (stsi(vm, 3, 2, 2) == 0 && vm->count) {
151 EBCASC(vm->vm[0].cpi, sizeof(vm->vm[0].cpi));
152 sprintf(hvstr, "%-16.16s", vm
[all...]
H A Dlgr.c44 } vm[VM_LEVEL_MAX]; member in struct:lgr_info
105 cpascii(lgr_info->vm[i].name, si->vm[i].name,
106 sizeof(si->vm[i].name));
107 cpascii(lgr_info->vm[i].cpi, si->vm[i].cpi,
108 sizeof(si->vm[i].cpi));
/linux-master/drivers/gpu/drm/i915/gt/
H A Dselftest_tlb.c36 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length),
41 i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE);
50 batch = i915_gem_object_create_internal(ce->vm->i915, 4096);
54 vma = i915_vma_instance(batch, ce->vm, NULL);
67 addr = igt_random_offset(prng, addr, min(ce->vm->total, BIT_ULL(48)),
168 ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags);
171 tlbinv(ce->vm, addr & -length, length);
230 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length))
283 va = i915_vma_instance(A, &ppgtt->vm, NUL
32 pte_tlbinv(struct intel_context *ce, struct i915_vma *va, struct i915_vma *vb, u64 align, void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), u64 length, struct rnd_state *prng) argument
228 mem_tlbinv(struct intel_gt *gt, struct drm_i915_gem_object *(*create_fn)(struct intel_gt *), void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) argument
363 tlbinv_full(struct i915_address_space *vm, u64 addr, u64 length) argument
[all...]
/linux-master/tools/testing/selftests/kvm/riscv/
H A Dsbi_pmu_test.c525 void test_vm_destroy(struct kvm_vm *vm) argument
529 kvm_vm_free(vm);
534 struct kvm_vm *vm; local
537 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
540 vm_init_vector_tables(vm);
542 vm_install_exception_handler(vm, EXC_INST_ILLEGAL, guest_illegal_exception_handler);
547 test_vm_destroy(vm);
552 struct kvm_vm *vm = NULL; local
555 vm = vm_create_with_one_vcpu(&vcpu, guest_code);
560 test_vm_destroy(vm);
563 test_vm_setup_snapshot_mem(struct kvm_vm *vm, struct kvm_vcpu *vcpu) argument
578 struct kvm_vm *vm = NULL; local
594 struct kvm_vm *vm = NULL; local
[all...]
/linux-master/drivers/gpu/drm/xe/tests/
H A Dxe_bo.c190 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); local
198 xe_vm_lock(vm, false);
199 bo = xe_bo_create_user(xe, NULL, vm, 0x10000,
203 xe_vm_unlock(vm);
275 down_read(&vm->lock);
276 xe_vm_lock(vm, false);
277 err = xe_bo_validate(bo, bo->vm, false);
278 xe_vm_unlock(vm);
279 up_read(&vm->lock);
321 xe_vm_put(vm);
[all...]
/linux-master/drivers/gpu/drm/radeon/
H A Dradeon_ib.c52 * @vm: requested vm
61 struct radeon_ib *ib, struct radeon_vm *vm,
77 ib->vm = vm;
78 if (vm) {
146 /* grab a vm id if necessary */
147 if (ib->vm) {
149 vm_id_fence = radeon_vm_grab_id(rdev, ib->vm, ib->ring);
161 if (ib->vm)
60 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) argument
[all...]

Completed in 322 milliseconds

1234567891011>>