/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.h | 160 return rcu_dereference_protected(ctx->vm, lockdep_is_held(&ctx->mutex)); 165 GEM_BUG_ON(!!ctx->vm != HAS_FULL_PPGTT(ctx->i915)); 167 return !!ctx->vm; 173 struct i915_address_space *vm; local 175 vm = ctx->vm; 176 if (!vm) 177 vm = &to_gt(ctx->i915)->ggtt->vm; 178 vm [all...] |
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | vmx_invalid_nested_guest_state.c | 14 static struct kvm_vm *vm; variable in typeref:struct:kvm_vm 63 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 66 vcpu_alloc_vmx(vm, &vmx_pages_gva);
|
H A D | recalc_apic_map_test.c | 39 struct kvm_vm *vm; local 51 vm = vm_create_with_vcpus(KVM_MAX_VCPUS, NULL, vcpus); 71 kvm_vm_free(vm);
|
H A D | userspace_io_test.c | 58 struct kvm_vm *vm; local 61 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 101 kvm_vm_free(vm);
|
H A D | kvm_pv_test.c | 139 struct kvm_vm *vm; local 148 vm = vm_create_with_one_vcpu(&vcpu, guest_main); 160 vm_enable_cap(vm, KVM_CAP_X86_DISABLE_EXITS, KVM_X86_DISABLE_EXITS_HLT); 170 kvm_vm_free(vm); 176 struct kvm_vm *vm; local 180 vm = vm_create_with_one_vcpu(&vcpu, guest_main); 187 kvm_vm_free(vm);
|
H A D | userspace_msr_exit_test.c | 515 struct kvm_vm *vm = vcpu->vm; local 521 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MSR, KVM_MSR_EXIT_REASON_FILTER); 526 vm_ioctl(vm, KVM_X86_SET_MSR_FILTER, &filter_allow); 528 vm_install_exception_handler(vm, GP_VECTOR, guest_gp_handler); 547 vm_install_exception_handler(vm, GP_VECTOR, guest_fep_gp_handler); 578 vm_ioctl(vcpu->vm, KVM_X86_SET_MSR_FILTER, &no_filter_deny); 628 struct kvm_vm *vm = vcpu->vm; local 634 vm_enable_cap(vm, KVM_CAP_X86_USER_SPACE_MS 669 struct kvm_vm *vm = vcpu->vm; local 700 run_user_space_msr_flag_test(struct kvm_vm *vm) argument 717 run_msr_filter_flag_test(struct kvm_vm *vm) argument 757 struct kvm_vm *vm = vcpu->vm; local [all...] |
H A D | private_mem_conversions_test.c | 297 struct kvm_vm *vm = vcpu->vm; local 304 vm_guest_mem_fallocate(vm, gpa, size, map_shared); 307 vm_set_memory_attributes(vm, gpa, size, 318 struct kvm_vm *vm = vcpu->vm; local 348 for (i = 0; i < size; i += vm->page_size) { 349 size_t nr_bytes = min_t(size_t, vm->page_size, size - i); 350 uint8_t *hva = addr_gpa2hva(vm, gpa + i); 382 struct kvm_vm *vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/s390x/ |
H A D | diag318_test_handler.c | 29 struct kvm_vm *vm; local 34 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 49 kvm_vm_free(vm);
|
/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | hypercalls.c | 160 vm_userspace_mem_region_add(vcpu->vm, VM_MEM_SRC_ANONYMOUS, ST_GPA_BASE, 1, gpages, 0); 236 struct kvm_vm *vm; local 238 vm = vm_create_with_one_vcpu(vcpu, guest_code); 242 return vm; 245 static void test_guest_stage(struct kvm_vm **vm, struct kvm_vcpu **vcpu) argument 253 sync_global_to_guest(*vm, stage); 261 kvm_vm_free(*vm); 262 *vm = test_vm_create(vcpu); 275 struct kvm_vm *vm; local 279 vm [all...] |
H A D | vgic_irq.c | 77 static void kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, 510 static void kvm_irq_line_check(struct kvm_vm *vm, uint32_t intid, int level, argument 516 kvm_arm_irq_line(vm, intid, level); 522 ret = _kvm_arm_irq_line(vm, intid, level); 553 static void kvm_set_gsi_routing_irqchip_check(struct kvm_vm *vm, argument 568 kvm_gsi_routing_write(vm, routing); 570 ret = _kvm_gsi_routing_write(vm, routing); 598 static void kvm_routing_and_irqfd_check(struct kvm_vm *vm, argument 615 kvm_set_gsi_routing_irqchip_check(vm, intid, num, 635 vm_ioctl(vm, KVM_IRQF 664 struct kvm_vm *vm = vcpu->vm; local 714 kvm_inject_get_call(struct kvm_vm *vm, struct ucall *uc, struct kvm_inject_args *args) argument 737 struct kvm_vm *vm; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_csa.c | 65 int amdgpu_map_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 74 r = amdgpu_vm_lock_pd(vm, &exec, 0); 84 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); 105 int amdgpu_unmap_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 114 r = amdgpu_vm_lock_pd(vm, &exec, 0);
|
H A D | amdgpu_seq64.c | 55 * @vm: vm pointer 63 int amdgpu_seq64_map(struct amdgpu_device *adev, struct amdgpu_vm *vm, argument 77 r = amdgpu_vm_lock_pd(vm, &exec, 0); 85 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); 122 struct amdgpu_vm *vm; local 134 vm = &fpriv->vm; 138 r = amdgpu_vm_lock_pd(vm, &exec, 0);
|
/linux-master/tools/testing/selftests/kvm/ |
H A D | steal_time.c | 76 sync_global_to_guest(vcpu->vm, st_gva[i]); 85 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 87 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); 171 struct kvm_vm *vm = vcpu->vm; local 185 sync_global_to_guest(vm, st_gva[i]); 198 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 200 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); 284 st_gpa[i] = addr_gva2gpa(vcpu->vm, (vm_vaddr_t)st_gva[i]); 285 sync_global_to_guest(vcpu->vm, st_gv 289 steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) argument 344 struct kvm_vm *vm; local [all...] |
H A D | guest_memfd_test.c | 123 static void test_create_guest_memfd_invalid(struct kvm_vm *vm) argument 131 fd = __vm_create_guest_memfd(vm, size, 0); 138 fd = __vm_create_guest_memfd(vm, page_size, flag); 145 static void test_create_guest_memfd_multiple(struct kvm_vm *vm) argument 150 fd1 = __vm_create_guest_memfd(vm, 4096, 0); 157 fd2 = __vm_create_guest_memfd(vm, 8192, 0); 178 struct kvm_vm *vm; local 185 vm = vm_create_barebones(); 187 test_create_guest_memfd_invalid(vm); 188 test_create_guest_memfd_multiple(vm); [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_exec_queue.c | 38 if (q->vm) 39 xe_vm_put(q->vm); 44 struct xe_vm *vm, 84 if (vm) 85 q->vm = xe_vm_get(vm); 113 err = xe_lrc_init(q->lrc + i, q->hwe, q, q->vm, SZ_16K); 123 * Normally the user vm holds an rpm ref to keep the device 124 * awake, and the context holds a ref for the vm, however for 125 * some engines we use the kernels migrate vm underneat 43 __xe_exec_queue_alloc(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument 141 xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, u32 logical_mask, u16 width, struct xe_hw_engine *hwe, u32 flags, u64 extensions) argument 173 xe_exec_queue_create_class(struct xe_device *xe, struct xe_gt *gt, struct xe_vm *vm, enum xe_engine_class class, u32 flags) argument 547 struct xe_vm *vm, *migrate_vm; local 812 xe_exec_queue_last_fence_lockdep_assert(struct xe_exec_queue *q, struct xe_vm *vm) argument 826 xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) argument 859 xe_exec_queue_last_fence_get(struct xe_exec_queue *q, struct xe_vm *vm) argument 884 xe_exec_queue_last_fence_set(struct xe_exec_queue *q, struct xe_vm *vm, struct dma_fence *fence) argument [all...] |
H A D | xe_trace.h | 86 __field(struct xe_vm *, vm) 92 __entry->vm = bo->vm; 95 TP_printk("size=%zu, flags=0x%02x, vm=%p", 96 __entry->size, __entry->flags, __entry->vm) 487 TP_PROTO(struct xe_vm *vm), 488 TP_ARGS(vm), 491 __field(struct xe_vm *, vm) 496 __entry->vm = vm; [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_trace.h | 64 __field(struct i915_address_space *, vm) 72 __entry->vm = vma->vm; 78 TP_printk("obj=%p, offset=0x%016llx size=0x%llx%s vm=%p", 81 __entry->vm) 90 __field(struct i915_address_space *, vm) 97 __entry->vm = vma->vm; 102 TP_printk("obj=%p, offset=0x%016llx size=0x%llx vm=%p", 103 __entry->obj, __entry->offset, __entry->size, __entry->vm) [all...] |
H A D | i915_vgpu.c | 151 struct drm_i915_private *dev_priv = ggtt->vm.i915; 161 ggtt->vm.reserved -= node->size; 174 struct drm_i915_private *dev_priv = ggtt->vm.i915; 177 if (!intel_vgpu_active(ggtt->vm.i915)) 190 struct drm_i915_private *dev_priv = ggtt->vm.i915; 200 ret = i915_gem_gtt_reserve(&ggtt->vm, NULL, node, 204 ggtt->vm.reserved += size; 255 struct drm_i915_private *dev_priv = ggtt->vm.i915; 257 unsigned long ggtt_end = ggtt->vm.total; 263 if (!intel_vgpu_active(ggtt->vm [all...] |
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dpi.c | 33 struct videomode *vm; member in struct:exynos_dpi 69 if (ctx->vm) { 78 drm_display_mode_from_videomode(ctx->vm, mode); 167 struct videomode *vm; local 172 vm = devm_kzalloc(dev, sizeof(*ctx->vm), GFP_KERNEL); 173 if (!vm) 176 ret = of_get_videomode(dn, vm, 0); 178 devm_kfree(dev, vm); 182 ctx->vm [all...] |
/linux-master/arch/s390/kernel/ |
H A D | early.c | 110 if (!memcmp(vmms->vm[0].cpi, "\xd2\xe5\xd4", 3)) 112 else if (!memcmp(vmms->vm[0].cpi, "\xa9\x61\xe5\xd4", 4)) 137 struct sysinfo_3_2_2 *vm = (struct sysinfo_3_2_2 *)&sysinfo_page; local 150 if (stsi(vm, 3, 2, 2) == 0 && vm->count) { 151 EBCASC(vm->vm[0].cpi, sizeof(vm->vm[0].cpi)); 152 sprintf(hvstr, "%-16.16s", vm [all...] |
H A D | lgr.c | 44 } vm[VM_LEVEL_MAX]; member in struct:lgr_info 105 cpascii(lgr_info->vm[i].name, si->vm[i].name, 106 sizeof(si->vm[i].name)); 107 cpascii(lgr_info->vm[i].cpi, si->vm[i].cpi, 108 sizeof(si->vm[i].cpi));
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_tlb.c | 36 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), 41 i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE); 50 batch = i915_gem_object_create_internal(ce->vm->i915, 4096); 54 vma = i915_vma_instance(batch, ce->vm, NULL); 67 addr = igt_random_offset(prng, addr, min(ce->vm->total, BIT_ULL(48)), 168 ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags); 171 tlbinv(ce->vm, addr & -length, length); 230 void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) 283 va = i915_vma_instance(A, &ppgtt->vm, NUL 32 pte_tlbinv(struct intel_context *ce, struct i915_vma *va, struct i915_vma *vb, u64 align, void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length), u64 length, struct rnd_state *prng) argument 228 mem_tlbinv(struct intel_gt *gt, struct drm_i915_gem_object *(*create_fn)(struct intel_gt *), void (*tlbinv)(struct i915_address_space *vm, u64 addr, u64 length)) argument 363 tlbinv_full(struct i915_address_space *vm, u64 addr, u64 length) argument [all...] |
/linux-master/tools/testing/selftests/kvm/riscv/ |
H A D | sbi_pmu_test.c | 525 void test_vm_destroy(struct kvm_vm *vm) argument 529 kvm_vm_free(vm); 534 struct kvm_vm *vm; local 537 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 540 vm_init_vector_tables(vm); 542 vm_install_exception_handler(vm, EXC_INST_ILLEGAL, guest_illegal_exception_handler); 547 test_vm_destroy(vm); 552 struct kvm_vm *vm = NULL; local 555 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 560 test_vm_destroy(vm); 563 test_vm_setup_snapshot_mem(struct kvm_vm *vm, struct kvm_vcpu *vcpu) argument 578 struct kvm_vm *vm = NULL; local 594 struct kvm_vm *vm = NULL; local [all...] |
/linux-master/drivers/gpu/drm/xe/tests/ |
H A D | xe_bo.c | 190 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); local 198 xe_vm_lock(vm, false); 199 bo = xe_bo_create_user(xe, NULL, vm, 0x10000, 203 xe_vm_unlock(vm); 275 down_read(&vm->lock); 276 xe_vm_lock(vm, false); 277 err = xe_bo_validate(bo, bo->vm, false); 278 xe_vm_unlock(vm); 279 up_read(&vm->lock); 321 xe_vm_put(vm); [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_ib.c | 52 * @vm: requested vm 61 struct radeon_ib *ib, struct radeon_vm *vm, 77 ib->vm = vm; 78 if (vm) { 146 /* grab a vm id if necessary */ 147 if (ib->vm) { 149 vm_id_fence = radeon_vm_grab_id(rdev, ib->vm, ib->ring); 161 if (ib->vm) 60 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) argument [all...] |