Lines Matching defs:mmu

113 			       struct panfrost_mmu *mmu,
119 ret = mmu_hw_do_operation_locked(pfdev, mmu->as, iova, size, op);
124 static void panfrost_mmu_enable(struct panfrost_device *pfdev, struct panfrost_mmu *mmu)
126 int as_nr = mmu->as;
127 struct io_pgtable_cfg *cfg = &mmu->pgtbl_cfg;
158 u32 panfrost_mmu_as_get(struct panfrost_device *pfdev, struct panfrost_mmu *mmu)
164 as = mmu->as;
166 int en = atomic_inc_return(&mmu->as_count);
175 list_move(&mmu->list, &pfdev->as_lru_list);
185 panfrost_mmu_enable(pfdev, mmu);
210 mmu->as = as;
212 atomic_set(&mmu->as_count, 1);
213 list_add(&mmu->list, &pfdev->as_lru_list);
215 dev_dbg(pfdev->dev, "Assigned AS%d to mmu %p, alloc_mask=%lx", as, mmu, pfdev->as_alloc_mask);
217 panfrost_mmu_enable(pfdev, mmu);
224 void panfrost_mmu_as_put(struct panfrost_device *pfdev, struct panfrost_mmu *mmu)
226 atomic_dec(&mmu->as_count);
227 WARN_ON(atomic_read(&mmu->as_count) < 0);
232 struct panfrost_mmu *mmu, *mmu_tmp;
241 list_for_each_entry_safe(mmu, mmu_tmp, &pfdev->as_lru_list, list) {
242 mmu->as = -1;
243 atomic_set(&mmu->as_count, 0);
244 list_del_init(&mmu->list);
275 struct panfrost_mmu *mmu,
278 if (mmu->as < 0)
285 mmu_hw_do_operation(pfdev, mmu, iova, size, AS_COMMAND_FLUSH_PT);
290 static int mmu_map_sg(struct panfrost_device *pfdev, struct panfrost_mmu *mmu,
295 struct io_pgtable_ops *ops = mmu->pgtbl_ops;
302 dev_dbg(pfdev->dev, "map: as=%d, iova=%llx, paddr=%lx, len=%zx", mmu->as, iova, paddr, len);
318 panfrost_mmu_flush_range(pfdev, mmu, start_iova, iova - start_iova);
342 mmu_map_sg(pfdev, mapping->mmu, mapping->mmnode.start << PAGE_SHIFT,
354 struct io_pgtable_ops *ops = mapping->mmu->pgtbl_ops;
363 mapping->mmu->as, iova, len);
379 panfrost_mmu_flush_range(pfdev, mapping->mmu,
389 //struct panfrost_mmu *mmu = cookie;
410 struct panfrost_mmu *mmu;
413 list_for_each_entry(mmu, &pfdev->as_lru_list, list) {
414 if (as == mmu->as)
421 spin_lock(&mmu->mm_lock);
423 drm_mm_for_each_node(node, &mmu->mm) {
433 spin_unlock(&mmu->mm_lock);
464 WARN_ON(bomapping->mmu->as != as);
530 mmu_map_sg(pfdev, bomapping->mmu, addr,
556 struct panfrost_mmu *mmu = container_of(kref, struct panfrost_mmu,
558 struct panfrost_device *pfdev = mmu->pfdev;
561 if (mmu->as >= 0) {
564 panfrost_mmu_disable(pfdev, mmu->as);
567 clear_bit(mmu->as, &pfdev->as_alloc_mask);
568 clear_bit(mmu->as, &pfdev->as_in_use_mask);
569 list_del(&mmu->list);
573 free_io_pgtable_ops(mmu->pgtbl_ops);
574 drm_mm_takedown(&mmu->mm);
575 kfree(mmu);
578 void panfrost_mmu_ctx_put(struct panfrost_mmu *mmu)
580 kref_put(&mmu->refcount, panfrost_mmu_release_ctx);
583 struct panfrost_mmu *panfrost_mmu_ctx_get(struct panfrost_mmu *mmu)
585 kref_get(&mmu->refcount);
587 return mmu;
618 struct panfrost_mmu *mmu;
620 mmu = kzalloc(sizeof(*mmu), GFP_KERNEL);
621 if (!mmu)
624 mmu->pfdev = pfdev;
625 spin_lock_init(&mmu->mm_lock);
628 drm_mm_init(&mmu->mm, SZ_32M >> PAGE_SHIFT, (SZ_4G - SZ_32M) >> PAGE_SHIFT);
629 mmu->mm.color_adjust = panfrost_drm_mm_color_adjust;
631 INIT_LIST_HEAD(&mmu->list);
632 mmu->as = -1;
634 mmu->pgtbl_cfg = (struct io_pgtable_cfg) {
643 mmu->pgtbl_ops = alloc_io_pgtable_ops(ARM_MALI_LPAE, &mmu->pgtbl_cfg,
644 mmu);
645 if (!mmu->pgtbl_ops) {
646 kfree(mmu);
650 kref_init(&mmu->refcount);
652 return mmu;
771 pfdev->mmu_irq = platform_get_irq_byname(to_platform_device(pfdev->dev), "mmu");
778 IRQF_SHARED, KBUILD_MODNAME "-mmu",
782 dev_err(pfdev->dev, "failed to request mmu irq");