Lines Matching defs:ggtt

63 insert_mappable_node(struct i915_ggtt *ggtt, struct drm_mm_node *node, u32 size)
67 err = mutex_lock_interruptible(&ggtt->vm.mutex);
72 err = drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
74 0, ggtt->mappable_end,
77 mutex_unlock(&ggtt->vm.mutex);
83 remove_mappable_node(struct i915_ggtt *ggtt, struct drm_mm_node *node)
85 mutex_lock(&ggtt->vm.mutex);
87 mutex_unlock(&ggtt->vm.mutex);
95 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
100 if (mutex_lock_interruptible(&ggtt->vm.mutex))
103 pinned = ggtt->vm.reserved;
104 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link)
108 mutex_unlock(&ggtt->vm.mutex);
110 args->aper_size = ggtt->vm.total;
308 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
336 ret = insert_mappable_node(ggtt, node, PAGE_SIZE);
346 ggtt->vm.clear_range(&ggtt->vm, node->start, node->size);
347 remove_mappable_node(ggtt, node);
369 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
373 ggtt->vm.clear_range(&ggtt->vm, node->start, node->size);
374 remove_mappable_node(ggtt, node);
385 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
421 ggtt->vm.insert_page(&ggtt->vm,
431 if (gtt_user_read(&ggtt->iomap, page_base, page_offset,
548 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
601 intel_gt_flush_ggtt_writes(ggtt->vm.gt);
602 ggtt->vm.insert_page(&ggtt->vm,
618 if (ggtt_write(&ggtt->iomap, page_base, page_offset,
629 intel_gt_flush_ggtt_writes(ggtt->vm.gt);
857 &to_gt(i915)->ggtt->userfault_list, userfault_link)
869 for (i = 0; i < to_gt(i915)->ggtt->num_fences; i++) {
870 struct i915_fence_reg *reg = &to_gt(i915)->ggtt->fence_regs[i];
911 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
927 if (obj->base.size > ggtt->mappable_end)
947 obj->base.size > ggtt->mappable_end / 2)
952 vma = i915_vma_instance(obj, &ggtt->vm, view);
972 (vma->fence_size > ggtt->mappable_end / 2 ||
993 mutex_lock(&ggtt->vm.mutex);
995 mutex_unlock(&ggtt->vm.mutex);
1239 i915_ggtt_resume(to_gt(dev_priv)->ggtt);