/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | igt_gem_utils.c | 48 const int ver = GRAPHICS_VER(vma->vm->i915); 55 obj = i915_gem_object_create_internal(vma->vm->i915, size); 92 intel_gt_chipset_flush(vma->vm->gt); 94 vma = i915_vma_instance(obj, vma->vm, NULL); 142 if (GRAPHICS_VER(ce->vm->i915) <= 5)
|
H A D | i915_gem_context.c | 448 GEM_BUG_ON(obj->base.size > ce->vm->total); 451 vma = i915_vma_instance(obj, ce->vm, NULL); 578 create_test_object(struct i915_address_space *vm, argument 587 intel_gt_retire_requests(vm->gt); 589 size = min(vm->total / 2, 1024ull * DW_PER_PAGE * PAGE_SIZE); 592 obj = huge_gem_object(vm->i915, DW_PER_PAGE * PAGE_SIZE, size); 714 obj = create_test_object(ce->vm, file, &objects); 812 if (!parent->vm) { /* not full-ppgtt; nothing to share */ 837 ctx = kernel_context(i915, parent->vm); 847 obj = create_test_object(parent->vm, 1373 struct i915_address_space *vm; local 1490 check_scratch(struct i915_address_space *vm, u64 offset) argument 1514 struct i915_address_space *vm; local 1602 struct i915_address_space *vm; local 1746 struct i915_address_space *vm; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/ |
H A D | ucall_common.c | 25 void ucall_init(struct kvm_vm *vm, vm_paddr_t mmio_gpa) argument 32 vaddr = vm_vaddr_alloc_shared(vm, sizeof(*hdr), KVM_UTIL_MIN_VADDR, 34 hdr = (struct ucall_header *)addr_gva2hva(vm, vaddr); 42 write_guest_global(vm, ucall_pool, (struct ucall_header *)vaddr); 44 ucall_arch_init(vm, mmio_gpa);
|
/linux-master/tools/testing/selftests/kvm/include/aarch64/ |
H A D | processor.h | 64 struct kvm_vcpu *aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, 125 void vm_init_descriptor_tables(struct kvm_vm *vm); 129 void vm_install_exception_handler(struct kvm_vm *vm, 131 void vm_install_sync_handler(struct kvm_vm *vm, 134 uint64_t *virt_get_pte_hva(struct kvm_vm *vm, vm_vaddr_t gva);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_ring_submission.c | 145 static struct i915_address_space *vm_alias(struct i915_address_space *vm) argument 147 if (i915_is_ggtt(vm)) 148 vm = &i915_vm_to_ggtt(vm)->alias->vm; 150 return vm; 153 static u32 pp_dir(struct i915_address_space *vm) argument 155 return to_gen6_ppgtt(i915_vm_to_ppgtt(vm))->pp_dir; 160 struct i915_address_space *vm = vm_alias(engine->gt->vm); local 491 struct i915_address_space *vm; local 510 struct i915_address_space *vm; local 654 load_pd_dir(struct i915_request *rq, struct i915_address_space *vm, u32 valid) argument 856 switch_mm(struct i915_request *rq, struct i915_address_space *vm) argument [all...] |
/linux-master/drivers/gpu/drm/bridge/ |
H A D | tc358768.c | 677 struct videomode vm; local 711 drm_display_mode_to_videomode(mode, &vm); 721 hact = vm.hactive * 3; 726 hact = vm.hactive * 3; 732 hact = vm.hactive * 18 / 8; 738 hact = vm.hactive * 2; 763 dpi_htot = vm.hactive + vm.hfront_porch + vm.hsync_len + vm [all...] |
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | xapic_ipi_test.c | 402 struct kvm_vm *vm; local 411 vm = vm_create_with_one_vcpu(¶ms[0].vcpu, halter_guest_code); 413 vm_init_descriptor_tables(vm); 415 vm_install_exception_handler(vm, IPI_VECTOR, guest_ipi_handler); 417 virt_pg_map(vm, APIC_DEFAULT_GPA, APIC_DEFAULT_GPA); 419 params[1].vcpu = vm_vcpu_add(vm, 1, sender_guest_code); 421 test_data_page_vaddr = vm_vaddr_alloc_page(vm); 422 data = addr_gva2hva(vm, test_data_page_vaddr); 430 pipis_rcvd = (uint64_t *)addr_gva2hva(vm, (uint64_t)&ipis_rcvd); 488 kvm_vm_free(vm); [all...] |
H A D | hyperv_features.c | 134 struct kvm_vm *vm; local 142 vm = vm_create_with_one_vcpu(&vcpu, guest_msr); 144 msr_gva = vm_vaddr_alloc_page(vm); 145 memset(addr_gva2hva(vm, msr_gva), 0x0, getpagesize()); 146 msr = addr_gva2hva(vm, msr_gva); 159 vm_init_descriptor_tables(vm); 491 kvm_vm_free(vm); 518 kvm_vm_free(vm); 526 struct kvm_vm *vm; local 533 vm [all...] |
H A D | hyperv_tlb_flush.c | 583 struct kvm_vm *vm; local 595 vm = vm_create_with_one_vcpu(&vcpu[0], sender_guest_code); 598 test_data_page = vm_vaddr_alloc_page(vm); 599 data = (struct test_data *)addr_gva2hva(vm, test_data_page); 602 data->hcall_gva = vm_vaddr_alloc_pages(vm, 2); 603 data->hcall_gpa = addr_gva2gpa(vm, data->hcall_gva); 604 memset(addr_gva2hva(vm, data->hcall_gva), 0x0, 2 * PAGE_SIZE); 611 data->test_pages = vm_vaddr_alloc_pages(vm, NTEST_PAGES + 1); 613 memset(addr_gva2hva(vm, data->test_pages + PAGE_SIZE * i), 615 set_expected_val(addr_gva2hva(vm, dat [all...] |
H A D | cpuid_test.c | 126 struct kvm_cpuid2 *vcpu_alloc_cpuid(struct kvm_vm *vm, vm_vaddr_t *p_gva, struct kvm_cpuid2 *cpuid) argument 129 vm_vaddr_t gva = vm_vaddr_alloc(vm, size, KVM_UTIL_MIN_VADDR); 130 struct kvm_cpuid2 *guest_cpuids = addr_gva2hva(vm, gva); 189 struct kvm_vm *vm; local 192 vm = vm_create_with_one_vcpu(&vcpu, guest_main); 196 vcpu_alloc_cpuid(vm, &cpuid_gva, vcpu->cpuid); 207 kvm_vm_free(vm);
|
H A D | vmx_nested_tsc_scaling_test.c | 122 struct kvm_vm *vm; local 154 vm = vm_create_with_one_vcpu(&vcpu, l1_guest_code); 155 vcpu_alloc_vmx(vm, &vmx_pages_gva); 204 kvm_vm_free(vm);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_modes.c | 1062 * drm_display_mode_from_videomode - fill in @dmode using @vm, 1063 * @vm: videomode structure to use as source 1066 * Fills out @dmode using the display mode specified in @vm. 1068 void drm_display_mode_from_videomode(const struct videomode *vm, argument 1071 dmode->hdisplay = vm->hactive; 1072 dmode->hsync_start = dmode->hdisplay + vm->hfront_porch; 1073 dmode->hsync_end = dmode->hsync_start + vm->hsync_len; 1074 dmode->htotal = dmode->hsync_end + vm->hback_porch; 1076 dmode->vdisplay = vm->vactive; 1077 dmode->vsync_start = dmode->vdisplay + vm 1109 drm_display_mode_to_videomode(const struct drm_display_mode *dmode, struct videomode *vm) argument 1153 drm_bus_flags_from_videomode(const struct videomode *vm, u32 *bus_flags) argument 1192 struct videomode vm; local 1228 struct videomode vm; local [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_bo.c | 470 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); local 473 if (!xe_vm_in_fault_mode(vm)) { 1063 if (bo->vm && xe_bo_is_user(bo)) 1064 xe_vm_put(bo->vm); 1098 if (bo->vm && !xe_vm_in_fault_mode(bo->vm)) { 1368 struct xe_tile *tile, struct xe_vm *vm, 1375 if (vm) 1376 xe_vm_assert_held(vm); 1367 __xe_bo_create_locked(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, u64 start, u64 end, u16 cpu_caching, enum ttm_bo_type type, u32 flags) argument 1436 xe_bo_create_locked_range(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, u64 start, u64 end, enum ttm_bo_type type, u32 flags) argument 1444 xe_bo_create_locked(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, enum ttm_bo_type type, u32 flags) argument 1451 xe_bo_create_user(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, u16 cpu_caching, enum ttm_bo_type type, u32 flags) argument 1466 xe_bo_create(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, enum ttm_bo_type type, u32 flags) argument 1478 xe_bo_create_pin_map_at(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, u64 offset, enum ttm_bo_type type, u32 flags) argument 1517 xe_bo_create_pin_map(struct xe_device *xe, struct xe_tile *tile, struct xe_vm *vm, size_t size, enum ttm_bo_type type, u32 flags) argument 1792 xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict) argument 1910 struct xe_vm *vm = NULL; local [all...] |
H A D | xe_bb.c | 89 xe_gt_assert(q->gt, q->vm->flags & XE_VM_FLAG_MIGRATION); 99 xe_gt_assert(q->gt, !(q->vm && q->vm->flags & XE_VM_FLAG_MIGRATION));
|
H A D | xe_sync.h | 34 struct xe_exec_queue *q, struct xe_vm *vm);
|
H A D | xe_bo_types.h | 33 /** @vm: VM this BO is attached to, for extobj this will be NULL */ 34 struct xe_vm *vm; member in struct:xe_bo
|
/linux-master/tools/testing/selftests/kvm/ |
H A D | guest_print_test.c | 59 sync_global_to_guest(vcpu->vm, vals); \ 156 struct kvm_vm *vm; local 159 vm = vm_create_with_one_vcpu(&vcpu, guest_code_limits); 171 kvm_vm_free(vm); 177 struct kvm_vm *vm; local 179 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 214 kvm_vm_free(vm);
|
H A D | rseq_test.c | 192 struct kvm_vm *vm; local 211 vm = vm_create_with_one_vcpu(&vcpu, guest_code); 263 kvm_vm_free(vm);
|
H A D | system_counter_offset_test.c | 120 struct kvm_vm *vm; local 122 vm = vm_create_with_one_vcpu(&vcpu, guest_main); 126 kvm_vm_free(vm);
|
/linux-master/drivers/gpu/drm/mediatek/ |
H A D | mtk_dpi.c | 514 struct videomode vm = { 0 }; local 520 drm_display_mode_to_videomode(mode, &vm); 521 pll_rate = vm.pixelclock * factor; 524 pll_rate, vm.pixelclock); 534 vm.pixelclock = pll_rate / factor; 535 vm.pixelclock /= dpi->conf->pixels_per_iter; 539 clk_set_rate(dpi->pixel_clk, vm.pixelclock * 2); 541 clk_set_rate(dpi->pixel_clk, vm.pixelclock); 544 vm.pixelclock = clk_get_rate(dpi->pixel_clk); 547 pll_rate, vm [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 67 err = mutex_lock_interruptible(&ggtt->vm.mutex); 72 err = drm_mm_insert_node_in_range(&ggtt->vm.mm, node, 77 mutex_unlock(&ggtt->vm.mutex); 85 mutex_lock(&ggtt->vm.mutex); 87 mutex_unlock(&ggtt->vm.mutex); 100 if (mutex_lock_interruptible(&ggtt->vm.mutex)) 103 pinned = ggtt->vm.reserved; 104 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link) 108 mutex_unlock(&ggtt->vm.mutex); 110 args->aper_size = ggtt->vm [all...] |
/linux-master/tools/testing/selftests/kvm/aarch64/ |
H A D | vgic_init.c | 28 struct kvm_vm *vm; member in struct:vm_gic 81 v.vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus); 82 v.gic_fd = kvm_create_device(v.vm, gic_dev_type); 92 v.vm = vm_create_barebones(); 93 v.gic_fd = kvm_create_device(v.vm, gic_dev_type); 102 kvm_vm_free(v->vm); 347 vcpus[i] = vm_vcpu_add(v.vm, i, guest_code); 466 v.vm = vm_create(NR_VCPUS); 467 (void)vm_vcpu_add(v.vm, 0, guest_code); 469 v.gic_fd = kvm_create_device(v.vm, KVM_DEV_TYPE_ARM_VGIC_V [all...] |
/linux-master/drivers/accel/habanalabs/common/ |
H A D | memory.c | 91 struct hl_vm *vm = &hdev->vm; local 116 paddr = (uintptr_t) gen_pool_dma_alloc_align(vm->dram_pg_pool, 119 paddr = gen_pool_alloc(vm->dram_pg_pool, total_size); 155 (uintptr_t)gen_pool_dma_alloc_align(vm->dram_pg_pool, 159 phys_pg_pack->pages[i] = gen_pool_alloc(vm->dram_pg_pool, 173 spin_lock(&vm->idr_lock); 174 handle = idr_alloc(&vm->phys_pg_pack_handles, phys_pg_pack, 1, 0, 176 spin_unlock(&vm->idr_lock); 185 kref_get(&vm 290 struct hl_vm *vm = container_of(ref, struct hl_vm, local 315 struct hl_vm *vm = &hdev->vm; local 356 struct hl_vm *vm = &hdev->vm; local 1057 struct hl_vm *vm = &hdev->vm; local 1998 struct hl_vm *vm = &hdev->vm; local 2757 struct hl_vm *vm = &hdev->vm; local 2840 struct hl_vm *vm = &hdev->vm; local 2893 struct hl_vm *vm = &hdev->vm; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_job.h | 49 struct amdgpu_vm *vm; member in struct:amdgpu_job 88 int amdgpu_job_alloc(struct amdgpu_device *adev, struct amdgpu_vm *vm,
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | mem_protect.h | 77 int kvm_guest_prepare_stage2(struct pkvm_hyp_vm *vm, void *pgd); 82 void reclaim_guest_pages(struct pkvm_hyp_vm *vm, struct kvm_hyp_memcache *mc);
|