Searched refs:ppgtt (Results 1 - 17 of 17) sorted by relevance

/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/selftests/
H A Dmock_gtt.c70 struct i915_ppgtt *ppgtt; local
72 ppgtt = kzalloc(sizeof(*ppgtt), GFP_KERNEL);
73 if (!ppgtt)
76 ppgtt->vm.gt = &i915->gt;
77 ppgtt->vm.i915 = i915;
78 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE);
79 ppgtt->vm.file = ERR_PTR(-ENODEV);
81 i915_address_space_init(&ppgtt->vm, VM_CLASS_PPGTT);
83 ppgtt
[all...]
H A Di915_gem_gtt.c151 struct i915_ppgtt *ppgtt; local
160 ppgtt = i915_ppgtt_create(&dev_priv->gt);
161 if (IS_ERR(ppgtt))
162 return PTR_ERR(ppgtt);
164 if (!ppgtt->vm.allocate_va_range)
175 limit = min(ppgtt->vm.total, limit);
179 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, size);
191 ppgtt->vm.clear_range(&ppgtt
1006 struct i915_ppgtt *ppgtt; local
[all...]
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gt/
H A Dgen6_ppgtt.c21 static inline void gen6_write_pde(const struct gen6_ppgtt *ppgtt, argument
28 bus_space_write_4(ppgtt->pd_bst, ppgtt->pd_bsh, pde*sizeof(gen6_pte_t),
32 ppgtt->pd_addr + pde);
92 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local
101 i915_pt_entry(ppgtt->base.pd, pde++);
111 ppgtt->scan_for_unused_pt = true;
133 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local
134 struct i915_page_directory * const pd = ppgtt->pd;
190 static void gen6_flush_pd(struct gen6_ppgtt *ppgtt, u6 argument
219 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local
273 gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) argument
298 gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) argument
313 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); local
344 struct gen6_ppgtt *ppgtt = vma->private; local
383 struct gen6_ppgtt *ppgtt = vma->private; local
412 pd_vma_create(struct gen6_ppgtt *ppgtt, int size) argument
445 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local
479 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local
488 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(base); local
500 struct gen6_ppgtt *ppgtt; local
[all...]
H A Dintel_ppgtt.c157 struct i915_ppgtt *ppgtt; local
159 ppgtt = __ppgtt_create(gt);
160 if (IS_ERR(ppgtt))
161 return ppgtt;
163 trace_i915_ppgtt_create(&ppgtt->vm);
165 return ppgtt;
213 void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt) argument
217 ppgtt->vm.gt = gt;
218 ppgtt->vm.i915 = i915;
220 ppgtt
[all...]
H A Dintel_lrc_reg.h37 #define ASSIGN_CTX_PDP(ppgtt, reg_state, n) do { \
39 const u64 addr__ = i915_page_dir_dma_addr((ppgtt), (n)); \
44 #define ASSIGN_CTX_PML4(ppgtt, reg_state) do { \
46 const u64 addr__ = px_dma(ppgtt->pd); \
H A Dgen8_ppgtt.c33 static void gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create) argument
35 struct drm_i915_private *i915 = ppgtt->vm.i915;
36 struct intel_uncore *uncore = ppgtt->vm.gt->uncore;
41 atomic_inc(px_used(ppgtt->pd)); /* never remove */
43 atomic_dec(px_used(ppgtt->pd));
47 if (i915_vm_is_4lvl(&ppgtt->vm)) {
48 const u64 daddr = px_dma(ppgtt->pd);
60 const u64 daddr = i915_page_dir_dma_addr(ppgtt, i);
135 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); local
138 return ppgtt
170 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local
384 gen8_ppgtt_insert_pte(struct i915_ppgtt *ppgtt, struct i915_page_directory *pdp, struct sgt_dma *iter, u64 idx, enum i915_cache_level cache_level, u32 flags) argument
638 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); local
703 gen8_preallocate_top_level_pdp(struct i915_ppgtt *ppgtt) argument
760 struct i915_ppgtt *ppgtt; local
[all...]
H A Dintel_ggtt.c745 struct i915_ppgtt *ppgtt; local
748 ppgtt = i915_ppgtt_create(ggtt->vm.gt);
749 if (IS_ERR(ppgtt))
750 return PTR_ERR(ppgtt);
752 if (GEM_WARN_ON(ppgtt->vm.total < ggtt->vm.total)) {
763 err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, ggtt->vm.total);
767 ggtt->alias = ppgtt;
768 ggtt->vm.bind_async_flags |= ppgtt->vm.bind_async_flags;
779 i915_vm_put(&ppgtt
785 struct i915_ppgtt *ppgtt; local
[all...]
H A Dintel_gtt.h278 * Each active user context has its own address space (in full-ppgtt).
544 i915_page_dir_dma_addr(const struct i915_ppgtt *ppgtt, const unsigned int n) argument
546 struct i915_page_dma *pt = ppgtt->pd->entry[n];
548 return px_dma(pt ?: px_base(&ppgtt->vm.scratch[ppgtt->vm.top]));
551 void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt);
H A Dselftest_hangcheck.c1313 struct i915_ppgtt *ppgtt; local
1320 ppgtt = i915_ppgtt_create(gt);
1321 if (IS_ERR(ppgtt))
1322 return PTR_ERR(ppgtt);
1324 err = __igt_reset_evict_vma(gt, &ppgtt->vm,
1326 i915_vm_put(&ppgtt->vm);
H A Dintel_lrc.c47 * contained there mean you don't need to do a ppgtt->switch_mm yourself,
4536 static void init_ppgtt_reg_state(u32 *regs, const struct i915_ppgtt *ppgtt) argument
4538 if (i915_vm_is_4lvl(&ppgtt->vm)) {
4543 ASSIGN_CTX_PML4(ppgtt, regs);
4545 ASSIGN_CTX_PDP(ppgtt, regs, 3);
4546 ASSIGN_CTX_PDP(ppgtt, regs, 2);
4547 ASSIGN_CTX_PDP(ppgtt, regs, 1);
4548 ASSIGN_CTX_PDP(ppgtt, regs, 0);
H A Dintel_ring_submission.c839 * aliasing ppgtt), but skip over the hung batch.
1373 const struct i915_ppgtt *ppgtt,
1389 *cs++ = px_base(ppgtt->pd)->ggtt_offset << 10;
1372 load_pd_dir(struct i915_request *rq, const struct i915_ppgtt *ppgtt, u32 valid) argument
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
H A Dmock_context.c43 struct i915_ppgtt *ppgtt; local
47 ppgtt = mock_ppgtt(i915, name);
48 if (!ppgtt)
52 __set_ppgtt(ctx, &ppgtt->vm);
55 i915_vm_put(&ppgtt->vm);
H A Dhuge_pages.c384 struct i915_ppgtt *ppgtt = arg; local
385 struct drm_i915_private *i915 = ppgtt->vm.i915;
421 vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
464 struct i915_ppgtt *ppgtt = arg; local
465 struct drm_i915_private *i915 = ppgtt->vm.i915;
492 vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
545 struct i915_ppgtt *ppgtt = arg; local
546 struct drm_i915_private *i915 = ppgtt->vm.i915;
588 vma = i915_vma_instance(obj, &ppgtt->vm, NULL);
673 struct i915_ppgtt *ppgtt)
672 close_object_list(struct list_head *objects, struct i915_ppgtt *ppgtt) argument
693 struct i915_ppgtt *ppgtt = arg; local
814 struct i915_ppgtt *ppgtt = arg; local
1862 struct i915_ppgtt *ppgtt; local
[all...]
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gvt/
H A Dscheduler.c374 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(ce->vm); local
378 px_dma(ppgtt->pd) = mm->ppgtt_mm.shadow_pdps[0];
382 i915_pd_entry(ppgtt->pd, i);
470 if (bb->ppgtt) {
627 gvt_vgpu_err("workload shadow ppgtt isn't ready\n");
1136 struct i915_ppgtt *ppgtt)
1140 if (i915_vm_is_4lvl(&ppgtt->vm)) {
1141 px_dma(ppgtt->pd) = s->i915_context_pml4;
1145 i915_pd_entry(ppgtt->pd, i);
1197 struct i915_ppgtt *ppgtt)
1135 i915_context_ppgtt_root_restore(struct intel_vgpu_submission *s, struct i915_ppgtt *ppgtt) argument
1196 i915_context_ppgtt_root_save(struct intel_vgpu_submission *s, struct i915_ppgtt *ppgtt) argument
1228 struct i915_ppgtt *ppgtt; local
[all...]
H A Dscheduler.h131 bool ppgtt; member in struct:intel_vgpu_shadow_bb
H A Dcmd_parser.c1663 /* Check post-sync and ppgtt bit */
1766 "ppgtt" : "ggtt", s->ring_id, s->workload);
1778 "ppgtt" : "ggtt", s->ring_id, s->workload);
1812 "ppgtt" : "ggtt", s->ring_id, s->workload);
1849 bb->ppgtt = (s->buf_addr_type == GTT_BUFFER) ? false : true;
1862 if (bb->ppgtt)
2680 "ppgtt" : "ggtt", s->ring_id, s->workload);
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gem/
H A Di915_gem_context.c193 * We do not allow downgrading from full-ppgtt [to a shared
201 * This ppgtt may have be reallocated between
204 * of this ppgtt with that third context (and not
206 * ppgtt after passing through the strong memory
210 * Once we have acquired the current ppgtt of ctx,
539 * the ppgtt).
734 struct i915_ppgtt *ppgtt; local
736 ppgtt = i915_ppgtt_create(&i915->gt);
737 if (IS_ERR(ppgtt)) {
739 PTR_ERR(ppgtt));
895 struct i915_ppgtt *ppgtt; local
1118 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local
1135 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); local
[all...]

Completed in 294 milliseconds