/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_evict.c | 58 I915_GTT_PAGE_SIZE); 123 I915_GTT_PAGE_SIZE, 0, 0, 138 I915_GTT_PAGE_SIZE, 0, 0, 170 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); 247 .start = I915_GTT_PAGE_SIZE * 2, 248 .size = I915_GTT_PAGE_SIZE, 265 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); 274 I915_GTT_PAGE_SIZE | flags); 281 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); 291 (I915_GTT_PAGE_SIZE * [all...] |
H A D | i915_gem_gtt.c | 122 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); 733 obj = i915_gem_object_create_internal(vm->i915, 2 * I915_GTT_PAGE_SIZE); 1547 total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; 1548 total += 2 * I915_GTT_PAGE_SIZE) { 1581 vma->node.size != 2*I915_GTT_PAGE_SIZE) { 1584 total, 2*I915_GTT_PAGE_SIZE); 1591 for (total = I915_GTT_PAGE_SIZE; 1592 total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; 1593 total += 2 * I915_GTT_PAGE_SIZE) { 1627 vma->node.size != 2*I915_GTT_PAGE_SIZE) { [all...] |
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | aperture_gm.c | 54 start = ALIGN(gvt_hidden_gmadr_base(gvt), I915_GTT_PAGE_SIZE); 55 end = ALIGN(gvt_hidden_gmadr_end(gvt), I915_GTT_PAGE_SIZE); 60 start = ALIGN(gvt_aperture_gmadr_base(gvt), I915_GTT_PAGE_SIZE); 61 end = ALIGN(gvt_aperture_gmadr_end(gvt), I915_GTT_PAGE_SIZE); 68 size, I915_GTT_PAGE_SIZE, 263 vgpu_aperture_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE); 274 vgpu_hidden_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE);
|
H A D | reg.h | 115 I915_GTT_PAGE_SIZE)
|
H A D | scheduler.c | 186 I915_GTT_PAGE_SIZE - RING_CTX_SIZE); 240 gpa_size += I915_GTT_PAGE_SIZE; 250 gpa_size = I915_GTT_PAGE_SIZE; 1002 gpa_size += I915_GTT_PAGE_SIZE; 1012 gpa_size = I915_GTT_PAGE_SIZE; 1044 I915_GTT_PAGE_SIZE - sizeof(*shadow_ring_context));
|
H A D | gtt.c | 906 (I915_GTT_PAGE_SIZE >> pt_entry_size_shift(spt)) 1417 for (index = 0; index < (I915_GTT_PAGE_SIZE >> 1470 oos_page->mem, I915_GTT_PAGE_SIZE); 2358 int page_entry_num = I915_GTT_PAGE_SIZE >>
|
H A D | cmd_parser.c | 1570 if (guest_gma >= I915_GTT_PAGE_SIZE) { 1775 offset = gma & (I915_GTT_PAGE_SIZE - 1); 1777 copy_len = (end_gma - gma) >= (I915_GTT_PAGE_SIZE - offset) ? 1778 I915_GTT_PAGE_SIZE - offset : end_gma - gma; 2841 if (WARN_ON(!IS_ALIGNED(workload->rb_start, I915_GTT_PAGE_SIZE))) 2885 I915_GTT_PAGE_SIZE)))
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_gtt.c | 77 * must be #I915_GTT_PAGE_SIZE aligned 106 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); 161 * must be #I915_GTT_PAGE_SIZE aligned 167 * must be #I915_GTT_PAGE_SIZE aligned 169 * must be #I915_GTT_PAGE_SIZE aligned if not U64_MAX 203 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); 207 GEM_BUG_ON(start > 0 && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); 208 GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); 230 BUILD_BUG_ON(I915_GTT_MIN_ALIGNMENT > I915_GTT_PAGE_SIZE);
|
H A D | i915_gem_evict.c | 341 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); 342 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); 365 start -= I915_GTT_PAGE_SIZE; 368 end += I915_GTT_PAGE_SIZE;
|
H A D | i915_vma_resource.c | 278 *start -= I915_GTT_PAGE_SIZE; 279 *end += I915_GTT_PAGE_SIZE;
|
H A D | i915_vma.c | 202 GEM_BUG_ON(!IS_ALIGNED(vma->size, I915_GTT_PAGE_SIZE)); 807 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); 824 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); 830 end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE); 831 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); 883 vma->page_sizes.sg > I915_GTT_PAGE_SIZE && 987 sg_set_page(sg, NULL, I915_GTT_PAGE_SIZE, 0); 990 sg_dma_len(sg) = I915_GTT_PAGE_SIZE; 995 left = (dst_stride - height) * I915_GTT_PAGE_SIZE; 1070 sg_set_page(sg, NULL, count * I915_GTT_PAGE_SIZE, [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_ggtt.c | 43 *start += I915_GTT_PAGE_SIZE; 53 *end -= I915_GTT_PAGE_SIZE; 447 (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE; 487 gte += (vma_res->start - vma_res->guard) / I915_GTT_PAGE_SIZE; 488 end = gte + vma_res->guard / I915_GTT_PAGE_SIZE; 491 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; 518 start = (vma_res->start - vma_res->guard) / I915_GTT_PAGE_SIZE; 519 end = start + vma_res->guard / I915_GTT_PAGE_SIZE; 524 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; 526 vma_res->node_size / I915_GTT_PAGE_SIZE, pte_encod [all...] |
H A D | gen6_ppgtt.c | 78 const unsigned int first_entry = start / I915_GTT_PAGE_SIZE; 82 unsigned int num_entries = length / I915_GTT_PAGE_SIZE; 117 unsigned int first_entry = vma_res->start / I915_GTT_PAGE_SIZE; 128 GEM_BUG_ON(sg_dma_len(iter.sg) < I915_GTT_PAGE_SIZE); 131 iter.dma += I915_GTT_PAGE_SIZE; 147 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; 288 u32 ggtt_offset = vma_res->start / I915_GTT_PAGE_SIZE;
|
H A D | selftest_context.c | 112 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; 113 memset(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE); 136 if (memchr_inv(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE)) { 178 engine->context_size += I915_GTT_PAGE_SIZE; 182 engine->context_size -= I915_GTT_PAGE_SIZE;
|
H A D | intel_gtt.h | 49 #define I915_GTT_PAGE_SIZE I915_GTT_PAGE_SIZE_4K macro 52 #define I915_GTT_PAGE_MASK -I915_GTT_PAGE_SIZE 54 #define I915_GTT_MIN_ALIGNMENT I915_GTT_PAGE_SIZE 172 __for_each_sgt_daddr(__dp, __iter, __sgt, I915_GTT_PAGE_SIZE) 175 __for_each_daddr_next(__dp, __iter, I915_GTT_PAGE_SIZE)
|
H A D | gen8_ppgtt.c | 469 GEM_BUG_ON(sg_dma_len(iter->sg) < I915_GTT_PAGE_SIZE); 472 iter->dma += I915_GTT_PAGE_SIZE; 542 page_size = I915_GTT_PAGE_SIZE; 583 I915_GTT_PAGE_SIZE); 648 page_size = I915_GTT_PAGE_SIZE; 654 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE)) 682 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE))) 743 if (vma_res->bi.page_sizes.sg > I915_GTT_PAGE_SIZE) { 759 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE;
|
H A D | intel_lrc.c | 1013 memset(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE); 1024 if (memchr_inv(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE)) 1120 context_size = round_up(engine->context_size, I915_GTT_PAGE_SIZE); 1123 context_size += I915_GTT_PAGE_SIZE; /* for redzone */ 1498 GEM_BUG_ON(cs - start > I915_GTT_PAGE_SIZE / sizeof(*cs)); 1512 GEM_BUG_ON(cs - start > I915_GTT_PAGE_SIZE / sizeof(*cs));
|
H A D | intel_migrate.c | 380 page_size = I915_GTT_PAGE_SIZE;
|
H A D | intel_engine_cs.c | 278 BUILD_BUG_ON(I915_GTT_PAGE_SIZE != PAGE_SIZE);
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.c | 50 gen8_set_pte(base + offset / I915_GTT_PAGE_SIZE, 71 i = vma_res->start / I915_GTT_PAGE_SIZE; 101 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; 255 size = DIV_ROUND_UP_ULL(obj->size, I915_GTT_PAGE_SIZE); 257 size = round_up(size * sizeof(gen8_pte_t), I915_GTT_PAGE_SIZE); 290 vm->total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE;
|
H A D | intel_plane_initial.c | 62 gte += base / I915_GTT_PAGE_SIZE;
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 111 GEM_BUG_ON(!IS_ALIGNED(dma_size, I915_GTT_PAGE_SIZE));
|
H A D | i915_gem_context.c | 1520 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); 1610 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); 1859 I915_GTT_PAGE_SIZE, vm_total,
|
H A D | huge_pages.c | 334 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE));
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_domain.c | 21 #define VTD_GUARD (168u * I915_GTT_PAGE_SIZE) /* 168 or tile-row PTE padding */
|