Searched refs:vm (Results 126 - 150 of 496) sorted by last modified time

1234567891011>>

/linux-master/drivers/gpu/drm/panthor/
H A Dpanthor_heap.h23 panthor_heap_pool_create(struct panthor_device *ptdev, struct panthor_vm *vm);
H A Dpanthor_heap.c89 /** @vm: VM this pool is bound to. */
90 struct panthor_vm *vm; member in struct:panthor_heap_pool
121 static void panthor_free_heap_chunk(struct panthor_vm *vm, argument
130 panthor_kernel_bo_destroy(vm, chunk->bo);
135 struct panthor_vm *vm,
147 chunk->bo = panthor_kernel_bo_create(ptdev, vm, heap->chunk_size,
186 panthor_kernel_bo_destroy(vm, chunk->bo);
194 static void panthor_free_heap_chunks(struct panthor_vm *vm, argument
200 panthor_free_heap_chunk(vm, heap, chunk);
204 struct panthor_vm *vm,
134 panthor_alloc_heap_chunk(struct panthor_device *ptdev, struct panthor_vm *vm, struct panthor_heap *heap, bool initial_chunk) argument
203 panthor_alloc_heap_chunks(struct panthor_device *ptdev, struct panthor_vm *vm, struct panthor_heap *heap, u32 chunk_count) argument
277 struct panthor_vm *vm; local
521 panthor_heap_pool_create(struct panthor_device *ptdev, struct panthor_vm *vm) argument
[all...]
H A Dpanthor_gem.c29 * @vm: The VM this BO was mapped to.
33 void panthor_kernel_bo_destroy(struct panthor_vm *vm, argument
44 to_panthor_bo(bo->obj)->exclusive_vm_root_gem != panthor_vm_root_gem(vm)))
47 ret = panthor_vm_unmap_range(vm, bo->va_node.start,
52 panthor_vm_free_va(vm, &bo->va_node);
62 * @vm: VM to map the GEM to. If NULL, the kernel object is not GPU mapped.
74 panthor_kernel_bo_create(struct panthor_device *ptdev, struct panthor_vm *vm, argument
83 if (drm_WARN_ON(&ptdev->base, !vm))
101 ret = panthor_vm_alloc_va(vm, gpu_va, size, &kbo->va_node);
105 ret = panthor_vm_map_bo_range(vm, b
[all...]
H A Dpanthor_gem.h135 panthor_kernel_bo_create(struct panthor_device *ptdev, struct panthor_vm *vm,
139 void panthor_kernel_bo_destroy(struct panthor_vm *vm,
H A Dpanthor_fw.c220 /** @vm: MCU VM. */
221 struct panthor_vm *vm; member in struct:panthor_fw
269 return ptdev->fw->vm;
446 mem = panthor_kernel_bo_create(ptdev, ptdev->fw->vm, SZ_8K,
1074 ret = panthor_vm_active(ptdev->fw->vm);
1144 panthor_vm_put(ptdev->fw->vm);
1329 fw->vm = panthor_vm_create(ptdev, true,
1333 if (IS_ERR(fw->vm)) {
1334 ret = PTR_ERR(fw->vm);
1335 fw->vm
[all...]
H A Dpanthor_drv.c827 struct panthor_vm *vm = NULL; local
840 vm = panthor_vm_pool_get_vm(pfile->vms, args->exclusive_vm_id);
841 if (!vm) {
847 ret = panthor_gem_create_with_handle(file, ddev, vm, &args->size,
850 panthor_vm_put(vm);
949 struct panthor_vm *vm = panthor_job_vm(ctx.jobs[0].job); local
952 ret = panthor_vm_prepare_mapped_bos_resvs(&ctx.exec, vm,
1039 struct panthor_vm *vm; local
1042 vm = panthor_vm_pool_get_vm(pfile->vms, args->vm_id);
1043 if (!vm)
1082 struct panthor_vm *vm; local
1113 struct panthor_vm *vm; local
1181 struct panthor_vm *vm; local
1233 struct panthor_vm *vm; local
[all...]
/linux-master/drivers/gpu/drm/panel/
H A Dpanel-simple.c174 struct videomode vm; local
176 videomode_from_timing(dt, &vm);
184 drm_display_mode_from_videomode(&vm, mode);
468 struct videomode vm; local
495 vm.flags = timing->flags;
496 drm_bus_flags_from_videomode(&vm, &bus_flags);
515 struct videomode vm; local
543 videomode_from_timing(ot, &vm);
544 drm_display_mode_from_videomode(&vm, &panel->override_mode);
H A Dpanel-edp.c265 struct videomode vm; local
267 videomode_from_timing(dt, &vm);
275 drm_display_mode_from_videomode(&vm, mode);
725 struct videomode vm; local
753 videomode_from_timing(ot, &vm);
754 drm_display_mode_from_videomode(&vm, &panel->override_mode);
/linux-master/drivers/gpu/drm/mediatek/
H A Dmtk_dsi.c212 struct videomode vm; member in struct:mtk_dsi
400 ps_wc = FIELD_PREP(DSI_PS_WC, dsi->vm.hactive * dsi_buf_bpp);
422 vact_nl = FIELD_PREP(VACT_NL, dsi->vm.vactive);
440 struct videomode *vm = &dsi->vm; local
447 writel(vm->vsync_len, dsi->regs + DSI_VSA_NL);
448 writel(vm->vback_porch, dsi->regs + DSI_VBP_NL);
449 writel(vm->vfront_porch, dsi->regs + DSI_VFP_NL);
450 writel(vm->vactive, dsi->regs + DSI_VACT_NL);
453 writel(FIELD_PREP(DSI_HEIGHT, vm
[all...]
H A Dmtk_dpi.c514 struct videomode vm = { 0 }; local
520 drm_display_mode_to_videomode(mode, &vm);
521 pll_rate = vm.pixelclock * factor;
524 pll_rate, vm.pixelclock);
534 vm.pixelclock = pll_rate / factor;
535 vm.pixelclock /= dpi->conf->pixels_per_iter;
539 clk_set_rate(dpi->pixel_clk, vm.pixelclock * 2);
541 clk_set_rate(dpi->pixel_clk, vm.pixelclock);
544 vm.pixelclock = clk_get_rate(dpi->pixel_clk);
547 pll_rate, vm
[all...]
H A Dmtk_dp.c88 struct videomode vm; member in struct:mtk_dp_info
393 struct videomode *vm = &mtk_dp->info.vm; local
395 drm_display_mode_from_videomode(vm, &mode);
401 vm->hsync_len + vm->hback_porch,
404 vm->hsync_len, HSW_SW_DP_ENC0_P0_MASK);
408 vm->hactive, HWIDTH_SW_DP_ENC0_P0_MASK);
414 vm->vsync_len + vm
1335 struct videomode *vm = &mtk_dp->info.vm; local
1370 struct videomode *vm = &mtk_dp->info.vm; local
[all...]
/linux-master/drivers/gpu/drm/lima/
H A Dlima_sched.h23 struct lima_vm *vm; member in struct:lima_sched_task
91 struct lima_vm *vm);
H A Dlima_sched.c116 struct lima_vm *vm)
127 err = drm_sched_job_init(&task->base, &context->base, 1, vm);
136 task->vm = lima_vm_get(vm);
153 lima_vm_put(task->vm);
252 pipe->current_vm = lima_vm_get(task->vm);
356 buffer_chunk->va = lima_vm_get_va(task->vm, bo);
475 struct lima_vm *vm = task->vm; local
482 lima_vm_bo_del(vm, bo
113 lima_sched_task_init(struct lima_sched_task *task, struct lima_sched_context *context, struct lima_bo **bos, int num_bos, struct lima_vm *vm) argument
[all...]
H A Dlima_mmu.c134 void lima_mmu_switch_vm(struct lima_ip *ip, struct lima_vm *vm) argument
143 mmu_write(LIMA_MMU_DTE_ADDR, vm->pd.dma);
H A Dlima_drv.c222 priv->vm = lima_vm_create(ldev);
223 if (!priv->vm) {
243 lima_vm_put(priv->vm);
H A Dlima_gp.c142 lima_vm_get_va(task->vm, bo) ==
256 ret = lima_heap_alloc(task->heap, task->vm);
H A Dlima_drv.h24 struct lima_vm *vm; member in struct:lima_drm_priv
/linux-master/drivers/gpu/drm/i915/
H A Di915_vma.c51 * We may be forced to unbind when the vm is dead, to clean it up.
55 if (kref_read(&vma->vm->ref))
124 intel_gt_pm_get_untracked(vma->vm->gt);
139 intel_gt_pm_put_async_untracked(vma->vm->gt);
147 struct i915_address_space *vm,
156 GEM_BUG_ON(vm == &vm->gt->ggtt->alias->vm);
162 vma->ops = &vm->vma_ops;
199 if (unlikely(vma->size > vm
146 vma_create(struct drm_i915_gem_object *obj, struct i915_address_space *vm, const struct i915_gtt_view *view) argument
285 i915_vma_lookup(struct drm_i915_gem_object *obj, struct i915_address_space *vm, const struct i915_gtt_view *view) argument
323 i915_vma_instance(struct drm_i915_gem_object *obj, struct i915_address_space *vm, const struct i915_gtt_view *view) argument
346 struct i915_address_space *vm; member in struct:i915_vma_work
1373 vma_invalidate_tlb(struct i915_address_space *vm, u32 *tlb) argument
1624 struct i915_address_space *vm = vma->vm; local
1844 struct i915_address_space *vm = vma->vm; local
1863 struct i915_address_space *vm = vma->vm; local
2159 struct i915_address_space *vm = vma->vm; local
2198 struct i915_address_space *vm = vma->vm; local
[all...]
H A Di915_perf.c1372 scratch = __vm_create_scratch_for_read_pinned(&ce->engine->gt->ggtt->vm, 4);
1871 vma = i915_vma_instance(bo, &gt->ggtt->vm, NULL);
1982 vma = i915_vma_instance(bo, &gt->ggtt->vm, NULL);
2248 &stream->engine->gt->ggtt->vm,
H A Di915_gpu_error.c1129 if (ggtt->vm.raw_insert_page)
1130 ggtt->vm.raw_insert_page(&ggtt->vm, dma, slot,
1135 ggtt->vm.insert_page(&ggtt->vm, dma, slot,
1148 ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE);
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_uc_fw.c1055 if (ggtt->vm.raw_insert_entries)
1056 ggtt->vm.raw_insert_entries(&ggtt->vm, vma_res,
1057 i915_gem_get_pat_index(ggtt->vm.i915,
1061 ggtt->vm.insert_entries(&ggtt->vm, vma_res,
1062 i915_gem_get_pat_index(ggtt->vm.i915,
1075 ggtt->vm.clear_range(&ggtt->vm, vma_res->start, vma_res->node_size);
H A Dintel_guc.c789 vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
/linux-master/drivers/gpu/drm/i915/gt/
H A Dselftest_reset.c87 ggtt->vm.insert_page(&ggtt->vm, dma,
111 ggtt->vm.clear_range(&ggtt->vm, ggtt->error_capture.start, PAGE_SIZE);
130 ggtt->vm.insert_page(&ggtt->vm, dma,
160 ggtt->vm.clear_range(&ggtt->vm, ggtt->error_capture.start, PAGE_SIZE);
H A Dselftest_hangcheck.c107 struct i915_address_space *vm = i915_gem_context_get_eb_vm(h->ctx); local
118 i915_vm_put(vm);
125 i915_vm_put(vm);
135 vma = i915_vma_instance(h->obj, vm, NULL);
137 i915_vm_put(vm);
141 hws = i915_vma_instance(h->hws, vm, NULL);
143 i915_vm_put(vm);
149 i915_vm_put(vm);
249 i915_vm_put(vm);
1381 struct i915_address_space *vm local
1426 __igt_reset_evict_vma(struct intel_gt *gt, struct i915_address_space *vm, int (*fn)(void *), unsigned int flags) argument
[all...]
H A Dintel_workarounds.c3004 vma = __vm_create_scratch_for_read(&ce->engine->gt->ggtt->vm,

Completed in 292 milliseconds

1234567891011>>