/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma_resource.h | 141 bool i915_vma_resource_hold(struct i915_vma_resource *vma_res, 144 void i915_vma_resource_unhold(struct i915_vma_resource *vma_res, 149 void i915_vma_resource_free(struct i915_vma_resource *vma_res); 151 struct dma_fence *i915_vma_resource_unbind(struct i915_vma_resource *vma_res, 154 void __i915_vma_resource_init(struct i915_vma_resource *vma_res); 158 * @vma_res: The vma resource on which to take a reference. 160 * Return: The @vma_res pointer 163 *i915_vma_resource_get(struct i915_vma_resource *vma_res) argument 165 dma_fence_get(&vma_res->unbind_fence); 166 return vma_res; 173 i915_vma_resource_put(struct i915_vma_resource *vma_res) argument 201 i915_vma_resource_init(struct i915_vma_resource *vma_res, struct i915_address_space *vm, struct sg_table *pages, const struct i915_page_sizes *page_sizes, struct i915_refct_sgt *pages_rsgt, bool readonly, bool lmem, struct intel_memory_region *mr, const struct i915_vma_ops *ops, void *private, u64 start, u64 node_size, u64 size, u32 guard) argument 235 i915_vma_resource_fini(struct i915_vma_resource *vma_res) argument [all...] |
H A D | i915_vma_resource.c | 53 struct i915_vma_resource *vma_res = local 56 return vma_res ? vma_res : ERR_PTR(-ENOMEM); 61 * @vma_res: The vma resource to free. 63 void i915_vma_resource_free(struct i915_vma_resource *vma_res) argument 65 if (vma_res) 66 kmem_cache_free(slab_vma_resources, vma_res); 81 struct i915_vma_resource *vma_res = local 82 container_of(head, typeof(*vma_res), unbind_fence.rcu); 84 i915_vma_resource_free(vma_res); 89 struct i915_vma_resource *vma_res = local 103 __i915_vma_resource_unhold(struct i915_vma_resource *vma_res) argument 135 i915_vma_resource_unhold(struct i915_vma_resource *vma_res, bool lockdep_cookie) argument 162 i915_vma_resource_hold(struct i915_vma_resource *vma_res, bool *lockdep_cookie) argument 175 struct i915_vma_resource *vma_res = local 193 struct i915_vma_resource *vma_res = local 230 i915_vma_resource_unbind(struct i915_vma_resource *vma_res, u32 *tlb) argument 262 __i915_vma_resource_init(struct i915_vma_resource *vma_res) argument [all...] |
H A D | i915_vma.c | 348 struct i915_vma_resource *vma_res; member in struct:i915_vma_work 358 struct i915_vma_resource *vma_res = vw->vma_res; local 369 vma_res->ops->bind_vma(vma_res->vm, &vw->stash, 370 vma_res, vw->pat_index, vw->flags); 381 if (vw->vma_res) 382 i915_vma_resource_put(vw->vma_res); 447 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, argument 452 i915_vma_resource_init(vma_res, vm 472 i915_vma_bind(struct i915_vma *vma, unsigned int pat_index, u32 flags, struct i915_vma_work *work, struct i915_vma_resource *vma_res) argument 1437 struct i915_vma_resource *vma_res = NULL; local 2015 struct i915_vma_resource *vma_res = vma->resource; local [all...] |
H A D | i915_gpu_error.c | 1087 const struct i915_vma_resource *vma_res, 1100 if (!vma_res || !vma_res->bi.pages || !compress) 1116 dst->gtt_offset = vma_res->start; 1117 dst->gtt_size = vma_res->node_size; 1118 dst->gtt_page_sizes = vma_res->page_sizes_gtt; 1126 for_each_sgt_daddr(dma, iter, vma_res->bi.pages) { 1152 } else if (vma_res->bi.lmem) { 1153 struct intel_memory_region *mem = vma_res->mr; 1156 for_each_sgt_daddr(dma, iter, vma_res 1086 i915_vma_coredump_create(const struct intel_gt *gt, const struct i915_vma_resource *vma_res, struct i915_vma_compress *compress, const char *name) argument 1436 struct i915_vma_resource *vma_res; member in struct:intel_engine_capture_vma 1442 capture_vma_snapshot(struct intel_engine_capture_vma *next, struct i915_vma_resource *vma_res, gfp_t gfp, const char *name) argument 1517 struct i915_vma_resource *vma_res; local 1618 struct i915_vma_resource *vma_res = this->vma_res; local [all...] |
H A D | i915_vma.h | 256 struct i915_vma_resource *vma_res); 488 void i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res,
|
H A D | i915_request.h | 55 struct i915_vma_resource *vma_res; member in struct:i915_capture_list
|
H A D | i915_request.c | 336 i915_vma_resource_put(capture->vma_res);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | mock_gtt.c | 36 struct i915_vma_resource *vma_res, 43 struct i915_vma_resource *vma_res, 48 vma_res->bound_flags |= flags; 52 struct i915_vma_resource *vma_res) 96 struct i915_vma_resource *vma_res, 103 struct i915_vma_resource *vma_res) 35 mock_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 41 mock_bind_ppgtt(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 51 mock_unbind_ppgtt(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 94 mock_bind_ggtt(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 102 mock_unbind_ggtt(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument
|
H A D | i915_gem_gtt.c | 1507 struct i915_vma_resource *vma_res; local 1511 vma_res = i915_vma_resource_alloc(); 1512 if (IS_ERR(vma_res)) 1513 return PTR_ERR(vma_res); 1521 i915_vma_resource_init_from_vma(vma_res, vma); 1522 vma->resource = vma_res; 1524 kfree(vma_res); 1688 struct i915_vma_resource *vma_res; local 1692 vma_res = i915_vma_resource_alloc(); 1693 if (IS_ERR(vma_res)) [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_ppgtt.c | 184 struct i915_vma_resource *vma_res, 190 if (!vma_res->allocated) { 191 vm->allocate_va_range(vm, stash, vma_res->start, 192 vma_res->vma_size); 193 vma_res->allocated = true; 198 if (vma_res->bi.readonly) 200 if (vma_res->bi.lmem) 203 vm->insert_entries(vm, vma_res, pat_index, pte_flags); 208 struct i915_vma_resource *vma_res) 210 if (!vma_res 182 ppgtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 207 ppgtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument [all...] |
H A D | intel_ggtt.c | 470 struct i915_vma_resource *vma_res, 487 gte += (vma_res->start - vma_res->guard) / I915_GTT_PAGE_SIZE; 488 end = gte + vma_res->guard / I915_GTT_PAGE_SIZE; 491 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; 493 for_each_sgt_daddr(addr, iter, vma_res->bi.pages) 509 struct i915_vma_resource *vma_res, 518 start = (vma_res->start - vma_res 469 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 508 __gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 539 gen8_ggtt_insert_entries_bind(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 615 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 696 struct i915_vma_resource *vma_res; member in struct:insert_entries 712 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 743 intel_ggtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 767 intel_ggtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument 920 aliasing_gtt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 943 aliasing_gtt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument [all...] |
H A D | intel_ggtt_gmch.c | 31 struct i915_vma_resource *vma_res, 38 intel_gmch_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start >> PAGE_SHIFT, 30 gmch_ggtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 unused) argument
|
H A D | gen6_ppgtt.c | 111 struct i915_vma_resource *vma_res, 117 unsigned int first_entry = vma_res->start / I915_GTT_PAGE_SIZE; 121 struct sgt_dma iter = sgt_dma(vma_res); 147 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; 282 struct i915_vma_resource *vma_res, 287 struct gen6_ppgtt *ppgtt = vma_res->private; 288 u32 ggtt_offset = vma_res->start / I915_GTT_PAGE_SIZE; 297 struct i915_vma_resource *vma_res) 299 struct gen6_ppgtt *ppgtt = vma_res->private; 110 gen6_ppgtt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 280 pd_vma_bind(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 unused) argument 296 pd_vma_unbind(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument
|
H A D | intel_gtt.h | 235 struct i915_vma_resource *vma_res, 243 struct i915_vma_resource *vma_res); 331 struct i915_vma_resource *vma_res, 340 struct i915_vma_resource *vma_res, 587 struct i915_vma_resource *vma_res, 591 struct i915_vma_resource *vma_res); 665 struct i915_vma_resource *vma_res, 669 struct i915_vma_resource *vma_res); 692 } sgt_dma(struct i915_vma_resource *vma_res) { argument 693 struct scatterlist *sg = vma_res [all...] |
H A D | gen8_ppgtt.c | 504 struct i915_vma_resource *vma_res, 511 u64 start = vma_res->start; 512 u64 end = start + vma_res->vma_size; 531 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && 544 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_64K) { 607 vma_res->page_sizes_gtt |= page_size; 612 struct i915_vma_resource *vma_res, 619 u64 start = vma_res->start; 634 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && 651 vma_res 503 xehpsdv_ppgtt_insert_huge(struct i915_address_space *vm, struct i915_vma_resource *vma_res, struct sgt_dma *iter, unsigned int pat_index, u32 flags) argument 611 gen8_ppgtt_insert_huge(struct i915_address_space *vm, struct i915_vma_resource *vma_res, struct sgt_dma *iter, unsigned int pat_index, u32 flags) argument 735 gen8_ppgtt_insert(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument [all...] |
H A D | intel_engine_cs.c | 2247 struct i915_vma_resource *vma_res = rq->batch_res; local 2254 vma_res ? upper_32_bits(vma_res->start) : ~0u, 2255 vma_res ? lower_32_bits(vma_res->start) : ~0u);
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.c | 55 struct i915_vma_resource *vma_res, 71 i = vma_res->start / I915_GTT_PAGE_SIZE; 72 for_each_sgt_daddr(addr, sgt_iter, vma_res->bi.pages) 83 struct i915_vma_resource *vma_res, 89 if (vma_res->bound_flags) 94 if (vm->has_read_only && vma_res->bi.readonly) 96 if (vma_res->bi.lmem) 99 vm->insert_entries(vm, vma_res, pat_index, pte_flags); 101 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; 108 vma_res 54 dpt_insert_entries(struct i915_address_space *vm, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 81 dpt_bind_vma(struct i915_address_space *vm, struct i915_vm_pt_stash *stash, struct i915_vma_resource *vma_res, unsigned int pat_index, u32 flags) argument 111 dpt_unbind_vma(struct i915_address_space *vm, struct i915_vma_resource *vma_res) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_uc_fw.c | 1036 struct i915_vma_resource *vma_res = &uc_fw->vma_res; local 1042 vma_res->start = uc_fw_ggtt_offset(uc_fw); 1043 vma_res->node_size = obj->base.size; 1044 vma_res->bi.pages = obj->mm.pages; 1050 drm_clflush_sg(vma_res->bi.pages); 1056 ggtt->vm.raw_insert_entries(&ggtt->vm, vma_res, 1061 ggtt->vm.insert_entries(&ggtt->vm, vma_res, 1070 struct i915_vma_resource *vma_res = &uc_fw->vma_res; local [all...] |
H A D | intel_uc_fw.h | 110 * @vma_res: A vma resource used in binding the uc fw to ggtt. The fw is 115 struct i915_vma_resource vma_res; member in struct:intel_uc_fw
|
H A D | intel_huc_fw.c | 61 msg_in->huc_in.huc_base_address = huc->fw.vma_res.start;
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 2030 capture->vma_res = i915_vma_resource_get(vma->resource);
|