/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
H A D | mock_gtt.c | 70 struct i915_ppgtt *ppgtt; local 72 ppgtt = kzalloc(sizeof(*ppgtt), GFP_KERNEL); 73 if (!ppgtt) 76 ppgtt->vm.gt = &i915->gt; 77 ppgtt->vm.i915 = i915; 78 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); 79 ppgtt->vm.file = ERR_PTR(-ENODEV); 81 i915_address_space_init(&ppgtt->vm, VM_CLASS_PPGTT); 83 ppgtt [all...] |
H A D | i915_gem_gtt.c | 151 struct i915_ppgtt *ppgtt; local 160 ppgtt = i915_ppgtt_create(&dev_priv->gt); 161 if (IS_ERR(ppgtt)) 162 return PTR_ERR(ppgtt); 164 if (!ppgtt->vm.allocate_va_range) 175 limit = min(ppgtt->vm.total, limit); 179 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, size); 191 ppgtt->vm.clear_range(&ppgtt 1006 struct i915_ppgtt *ppgtt; local [all...] |
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | gen6_ppgtt.c | 21 static inline void gen6_write_pde(const struct gen6_ppgtt *ppgtt, argument 28 bus_space_write_4(ppgtt->pd_bst, ppgtt->pd_bsh, pde*sizeof(gen6_pte_t), 32 ppgtt->pd_addr + pde); 92 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local 101 i915_pt_entry(ppgtt->base.pd, pde++); 111 ppgtt->scan_for_unused_pt = true; 133 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local 134 struct i915_page_directory * const pd = ppgtt->pd; 190 static void gen6_flush_pd(struct gen6_ppgtt *ppgtt, u6 argument 219 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local 273 gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) argument 298 gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) argument 313 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local 344 struct gen6_ppgtt *ppgtt = vma->private; local 383 struct gen6_ppgtt *ppgtt = vma->private; local 412 pd_vma_create(struct gen6_ppgtt *ppgtt, int size) argument 445 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local 479 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local 488 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local 500 struct gen6_ppgtt *ppgtt; local [all...] |
H A D | intel_ppgtt.c | 157 struct i915_ppgtt *ppgtt; local 159 ppgtt = __ppgtt_create(gt); 160 if (IS_ERR(ppgtt)) 161 return ppgtt; 163 trace_i915_ppgtt_create(&ppgtt->vm); 165 return ppgtt; 213 void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt) argument 217 ppgtt->vm.gt = gt; 218 ppgtt->vm.i915 = i915; 220 ppgtt [all...] |
H A D | intel_lrc_reg.h | 37 #define ASSIGN_CTX_PDP(ppgtt, reg_state, n) do { \ 39 const u64 addr__ = i915_page_dir_dma_addr((ppgtt), (n)); \ 44 #define ASSIGN_CTX_PML4(ppgtt, reg_state) do { \ 46 const u64 addr__ = px_dma(ppgtt->pd); \
|
H A D | gen8_ppgtt.c | 33 static void gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create) argument 35 struct drm_i915_private *i915 = ppgtt->vm.i915; 36 struct intel_uncore *uncore = ppgtt->vm.gt->uncore; 41 atomic_inc(px_used(ppgtt->pd)); /* never remove */ 43 atomic_dec(px_used(ppgtt->pd)); 47 if (i915_vm_is_4lvl(&ppgtt->vm)) { 48 const u64 daddr = px_dma(ppgtt->pd); 60 const u64 daddr = i915_page_dir_dma_addr(ppgtt, i); 135 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); local 138 return ppgtt 170 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local 384 gen8_ppgtt_insert_pte(struct i915_ppgtt *ppgtt, struct i915_page_directory *pdp, struct sgt_dma *iter, u64 idx, enum i915_cache_level cache_level, u32 flags) argument 638 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); local 703 gen8_preallocate_top_level_pdp(struct i915_ppgtt *ppgtt) argument 760 struct i915_ppgtt *ppgtt; local [all...] |
H A D | intel_ggtt.c | 745 struct i915_ppgtt *ppgtt; local 748 ppgtt = i915_ppgtt_create(ggtt->vm.gt); 749 if (IS_ERR(ppgtt)) 750 return PTR_ERR(ppgtt); 752 if (GEM_WARN_ON(ppgtt->vm.total < ggtt->vm.total)) { 763 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, ggtt->vm.total); 767 ggtt->alias = ppgtt; 768 ggtt->vm.bind_async_flags |= ppgtt->vm.bind_async_flags; 779 i915_vm_put(&ppgtt 785 struct i915_ppgtt *ppgtt; local [all...] |
H A D | intel_gtt.h | 278 * Each active user context has its own address space (in full-ppgtt). 544 i915_page_dir_dma_addr(const struct i915_ppgtt *ppgtt, const unsigned int n) argument 546 struct i915_page_dma *pt = ppgtt->pd->entry[n]; 548 return px_dma(pt ?: px_base(&ppgtt->vm.scratch[ppgtt->vm.top])); 551 void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt);
|
H A D | selftest_hangcheck.c | 1313 struct i915_ppgtt *ppgtt; local 1320 ppgtt = i915_ppgtt_create(gt); 1321 if (IS_ERR(ppgtt)) 1322 return PTR_ERR(ppgtt); 1324 err = __igt_reset_evict_vma(gt, &ppgtt->vm, 1326 i915_vm_put(&ppgtt->vm);
|
H A D | intel_lrc.c | 47 * contained there mean you don't need to do a ppgtt->switch_mm yourself, 4536 static void init_ppgtt_reg_state(u32 *regs, const struct i915_ppgtt *ppgtt) argument 4538 if (i915_vm_is_4lvl(&ppgtt->vm)) { 4543 ASSIGN_CTX_PML4(ppgtt, regs); 4545 ASSIGN_CTX_PDP(ppgtt, regs, 3); 4546 ASSIGN_CTX_PDP(ppgtt, regs, 2); 4547 ASSIGN_CTX_PDP(ppgtt, regs, 1); 4548 ASSIGN_CTX_PDP(ppgtt, regs, 0);
|
H A D | intel_ring_submission.c | 839 * aliasing ppgtt), but skip over the hung batch. 1373 const struct i915_ppgtt *ppgtt, 1389 *cs++ = px_base(ppgtt->pd)->ggtt_offset << 10; 1372 load_pd_dir(struct i915_request *rq, const struct i915_ppgtt *ppgtt, u32 valid) argument
|
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
H A D | mock_context.c | 43 struct i915_ppgtt *ppgtt; local 47 ppgtt = mock_ppgtt(i915, name); 48 if (!ppgtt) 52 __set_ppgtt(ctx, &ppgtt->vm); 55 i915_vm_put(&ppgtt->vm);
|
H A D | huge_pages.c | 384 struct i915_ppgtt *ppgtt = arg; local 385 struct drm_i915_private *i915 = ppgtt->vm.i915; 421 vma = i915_vma_instance(obj, &ppgtt->vm, NULL); 464 struct i915_ppgtt *ppgtt = arg; local 465 struct drm_i915_private *i915 = ppgtt->vm.i915; 492 vma = i915_vma_instance(obj, &ppgtt->vm, NULL); 545 struct i915_ppgtt *ppgtt = arg; local 546 struct drm_i915_private *i915 = ppgtt->vm.i915; 588 vma = i915_vma_instance(obj, &ppgtt->vm, NULL); 673 struct i915_ppgtt *ppgtt) 672 close_object_list(struct list_head *objects, struct i915_ppgtt *ppgtt) argument 693 struct i915_ppgtt *ppgtt = arg; local 814 struct i915_ppgtt *ppgtt = arg; local 1862 struct i915_ppgtt *ppgtt; local [all...] |
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
H A D | scheduler.c | 374 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(ce->vm); local 378 px_dma(ppgtt->pd) = mm->ppgtt_mm.shadow_pdps[0]; 382 i915_pd_entry(ppgtt->pd, i); 470 if (bb->ppgtt) { 627 gvt_vgpu_err("workload shadow ppgtt isn't ready\n"); 1136 struct i915_ppgtt *ppgtt) 1140 if (i915_vm_is_4lvl(&ppgtt->vm)) { 1141 px_dma(ppgtt->pd) = s->i915_context_pml4; 1145 i915_pd_entry(ppgtt->pd, i); 1197 struct i915_ppgtt *ppgtt) 1135 i915_context_ppgtt_root_restore(struct intel_vgpu_submission *s, struct i915_ppgtt *ppgtt) argument 1196 i915_context_ppgtt_root_save(struct intel_vgpu_submission *s, struct i915_ppgtt *ppgtt) argument 1228 struct i915_ppgtt *ppgtt; local [all...] |
H A D | scheduler.h | 131 bool ppgtt; member in struct:intel_vgpu_shadow_bb
|
H A D | cmd_parser.c | 1663 /* Check post-sync and ppgtt bit */ 1766 "ppgtt" : "ggtt", s->ring_id, s->workload); 1778 "ppgtt" : "ggtt", s->ring_id, s->workload); 1812 "ppgtt" : "ggtt", s->ring_id, s->workload); 1849 bb->ppgtt = (s->buf_addr_type == GTT_BUFFER) ? false : true; 1862 if (bb->ppgtt) 2680 "ppgtt" : "ggtt", s->ring_id, s->workload);
|
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gem/ |
H A D | i915_gem_context.c | 193 * We do not allow downgrading from full-ppgtt [to a shared 201 * This ppgtt may have be reallocated between 204 * of this ppgtt with that third context (and not 206 * ppgtt after passing through the strong memory 210 * Once we have acquired the current ppgtt of ctx, 539 * the ppgtt). 734 struct i915_ppgtt *ppgtt; local 736 ppgtt = i915_ppgtt_create(&i915->gt); 737 if (IS_ERR(ppgtt)) { 739 PTR_ERR(ppgtt)); 895 struct i915_ppgtt *ppgtt; local 1118 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local 1135 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local [all...] |