1// SPDX-License-Identifier: MIT 2/* 3 * Copyright �� 2022 Intel Corporation 4 */ 5 6#include "intel_ggtt_gmch.h" 7 8#include <drm/intel/intel-gtt.h> 9 10#include <linux/agp_backend.h> 11 12#include "i915_drv.h" 13#include "i915_utils.h" 14#include "intel_gtt.h" 15#include "intel_gt_regs.h" 16#include "intel_gt.h" 17 18static void gmch_ggtt_insert_page(struct i915_address_space *vm, 19 dma_addr_t addr, 20 u64 offset, 21 unsigned int pat_index, 22 u32 unused) 23{ 24 unsigned int flags = (pat_index == I915_CACHE_NONE) ? 25 AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY; 26 27 intel_gmch_gtt_insert_page(addr, offset >> PAGE_SHIFT, flags); 28} 29 30static void gmch_ggtt_insert_entries(struct i915_address_space *vm, 31 struct i915_vma_resource *vma_res, 32 unsigned int pat_index, 33 u32 unused) 34{ 35 unsigned int flags = (pat_index == I915_CACHE_NONE) ? 36 AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY; 37 38 intel_gmch_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start >> PAGE_SHIFT, 39 flags); 40} 41 42static void gmch_ggtt_invalidate(struct i915_ggtt *ggtt) 43{ 44 intel_gmch_gtt_flush(); 45} 46 47static void gmch_ggtt_clear_range(struct i915_address_space *vm, 48 u64 start, u64 length) 49{ 50 intel_gmch_gtt_clear_range(start >> PAGE_SHIFT, length >> PAGE_SHIFT); 51} 52 53static void gmch_ggtt_remove(struct i915_address_space *vm) 54{ 55 intel_gmch_remove(); 56} 57 58/* 59 * Certain Gen5 chipsets require idling the GPU before unmapping anything from 60 * the GTT when VT-d is enabled. 61 */ 62static bool needs_idle_maps(struct drm_i915_private *i915) 63{ 64 /* 65 * Query intel_iommu to see if we need the workaround. Presumably that 66 * was loaded first. 67 */ 68 if (!i915_vtd_active(i915)) 69 return false; 70 71 if (GRAPHICS_VER(i915) == 5 && IS_MOBILE(i915)) 72 return true; 73 74 return false; 75} 76 77int intel_ggtt_gmch_probe(struct i915_ggtt *ggtt) 78{ 79 struct drm_i915_private *i915 = ggtt->vm.i915; 80 phys_addr_t gmadr_base; 81 int ret; 82 83 ret = intel_gmch_probe(i915->gmch.pdev, to_pci_dev(i915->drm.dev), NULL); 84 if (!ret) { 85 drm_err(&i915->drm, "failed to set up gmch\n"); 86 return -EIO; 87 } 88 89 intel_gmch_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end); 90 91 ggtt->gmadr = DEFINE_RES_MEM(gmadr_base, ggtt->mappable_end); 92 93 ggtt->vm.alloc_pt_dma = alloc_pt_dma; 94 ggtt->vm.alloc_scratch_dma = alloc_pt_dma; 95 96 if (needs_idle_maps(i915)) { 97 drm_notice(&i915->drm, 98 "Flushing DMA requests before IOMMU unmaps; performance may be degraded\n"); 99 ggtt->do_idle_maps = true; 100 } 101 102 ggtt->vm.insert_page = gmch_ggtt_insert_page; 103 ggtt->vm.insert_entries = gmch_ggtt_insert_entries; 104 ggtt->vm.clear_range = gmch_ggtt_clear_range; 105 ggtt->vm.scratch_range = gmch_ggtt_clear_range; 106 ggtt->vm.cleanup = gmch_ggtt_remove; 107 108 ggtt->invalidate = gmch_ggtt_invalidate; 109 110 ggtt->vm.vma_ops.bind_vma = intel_ggtt_bind_vma; 111 ggtt->vm.vma_ops.unbind_vma = intel_ggtt_unbind_vma; 112 113 if (unlikely(ggtt->do_idle_maps)) 114 drm_notice(&i915->drm, 115 "Applying Ironlake quirks for intel_iommu\n"); 116 117 return 0; 118} 119 120int intel_ggtt_gmch_enable_hw(struct drm_i915_private *i915) 121{ 122 if (!intel_gmch_enable_gtt()) 123 return -EIO; 124 125 return 0; 126} 127 128void intel_ggtt_gmch_flush(void) 129{ 130 intel_gmch_gtt_flush(); 131} 132