/linux-master/drivers/accel/habanalabs/common/mmu/ |
H A D | Makefile | 2 HL_COMMON_MMU_FILES := common/mmu/mmu.o common/mmu/mmu_v1.o \ 3 common/mmu/mmu_v2.o common/mmu/mmu_v2_hr.o
|
/linux-master/arch/arm64/include/asm/ |
H A D | stage2_pgtable.h | 24 #define kvm_stage2_levels(mmu) VTCR_EL2_LVLS((mmu)->vtcr) 31 #define kvm_mmu_cache_min_pages(mmu) (kvm_stage2_levels(mmu) - 1)
|
/linux-master/drivers/gpu/drm/nouveau/nvif/ |
H A D | mmu.c | 22 #include <nvif/mmu.h> 28 nvif_mmu_dtor(struct nvif_mmu *mmu) argument 30 if (!nvif_object_constructed(&mmu->object)) 33 kfree(mmu->kind); 34 kfree(mmu->type); 35 kfree(mmu->heap); 36 nvif_object_dtor(&mmu->object); 41 struct nvif_mmu *mmu) 53 mmu->heap = NULL; 54 mmu 40 nvif_mmu_ctor(struct nvif_object *parent, const char *name, s32 oclass, struct nvif_mmu *mmu) argument [all...] |
/linux-master/fs/ramfs/ |
H A D | Makefile | 8 file-mmu-y := file-nommu.o 9 file-mmu-$(CONFIG_MMU) := file-mmu.o 10 ramfs-objs += inode.o $(file-mmu-y)
|
/linux-master/arch/arm64/include/asm/xen/ |
H A D | page.h | 2 #include <asm/mmu.h>
|
/linux-master/arch/m68k/include/asm/ |
H A D | mmu.h | 9 #include <asm-generic/mmu.h>
|
/linux-master/arch/powerpc/include/asm/nohash/ |
H A D | mmu.h | 7 #include <asm/nohash/32/mmu-40x.h> 10 #include <asm/nohash/32/mmu-44x.h> 13 #include <asm/nohash/mmu-e500.h> 16 #include <asm/nohash/32/mmu-8xx.h>
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_mmu.h | 13 void (*detach)(struct msm_mmu *mmu); 14 int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt, 16 int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len); 17 void (*destroy)(struct msm_mmu *mmu); 18 void (*resume_translation)(struct msm_mmu *mmu); 35 static inline void msm_mmu_init(struct msm_mmu *mmu, struct device *dev, argument 38 mmu->dev = dev; 39 mmu->funcs = funcs; 40 mmu->type = type; 47 static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, voi argument [all...] |
H A D | msm_gpummu.c | 24 static void msm_gpummu_detach(struct msm_mmu *mmu) argument 28 static int msm_gpummu_map(struct msm_mmu *mmu, uint64_t iova, argument 31 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); 56 static int msm_gpummu_unmap(struct msm_mmu *mmu, uint64_t iova, size_t len) argument 58 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); 71 static void msm_gpummu_resume_translation(struct msm_mmu *mmu) argument 75 static void msm_gpummu_destroy(struct msm_mmu *mmu) argument 77 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); 79 dma_free_attrs(mmu->dev, TABLE_SIZE, gpummu->table, gpummu->pt_base, 114 void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_ argument [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | base.c | 42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) argument 51 list_add(&ptp->head, &mmu->ptp.list); 56 nvkm_mmu_ptc_put(mmu, force, &ptp->pt); 65 nvkm_mmu_ptp_get(struct nvkm_mmu *mmu, u32 size, bool zero) argument 74 ptp = list_first_entry_or_null(&mmu->ptp.list, typeof(*ptp), head); 82 ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false); 93 list_add(&ptp->head, &mmu->ptp.list); 120 nvkm_mmu_ptc_find(struct nvkm_mmu *mmu, u32 size) argument 124 list_for_each_entry(ptc, &mmu->ptc.list, head) { 134 list_add(&ptc->head, &mmu 141 nvkm_mmu_ptc_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt **ppt) argument 167 nvkm_mmu_ptc_get(struct nvkm_mmu *mmu, u32 size, u32 align, bool zero) argument 220 nvkm_mmu_ptc_dump(struct nvkm_mmu *mmu) argument 234 nvkm_mmu_ptc_fini(struct nvkm_mmu *mmu) argument 246 nvkm_mmu_ptc_init(struct nvkm_mmu *mmu) argument 255 nvkm_mmu_type(struct nvkm_mmu *mmu, int heap, u8 type) argument 265 nvkm_mmu_heap(struct nvkm_mmu *mmu, u8 type, u64 size) argument 278 nvkm_mmu_host(struct nvkm_mmu *mmu) argument 315 nvkm_mmu_vram(struct nvkm_mmu *mmu) argument 370 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local 391 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local 400 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local 421 nvkm_mmu_ctor(const struct nvkm_mmu_func *func, struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_mmu *mmu) argument [all...] |
H A D | ummu.c | 35 struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu; local 37 if (mmu->func->mem.user.oclass) { 39 oclass->base = mmu->func->mem.user; 45 if (mmu->func->vmm.user.oclass) { 47 oclass->base = mmu->func->vmm.user; 59 struct nvkm_mmu *mmu = ummu->mmu; local 67 if ((index = args->v0.index) >= mmu->heap_nr) 69 args->v0.size = mmu 79 struct nvkm_mmu *mmu = ummu->mmu; local 108 struct nvkm_mmu *mmu = ummu->mmu; local 159 struct nvkm_mmu *mmu = device->mmu; local [all...] |
H A D | nv41.c | 32 nv41_mmu_init(struct nvkm_mmu *mmu) argument 34 struct nvkm_device *device = mmu->subdev.device; 35 nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr); 44 .mmu = {{ -1, -1, NVIF_CLASS_MMU_NV04}},
|
/linux-master/drivers/staging/media/ipu3/ |
H A D | ipu3-mmu.c | 21 #include "ipu3-mmu.h" 73 * @mmu: MMU to perform the invalidate operation on 78 static void imgu_mmu_tlb_invalidate(struct imgu_mmu *mmu) argument 80 writel(TLB_INVALIDATE, mmu->base + REG_TLB_INVALIDATE); 83 static void call_if_imgu_is_powered(struct imgu_mmu *mmu, argument 84 void (*func)(struct imgu_mmu *mmu)) 86 if (!pm_runtime_get_if_in_use(mmu->dev)) 89 func(mmu); 90 pm_runtime_put(mmu->dev); 95 * @mmu 101 imgu_mmu_set_halt(struct imgu_mmu *mmu, bool halt) argument 168 imgu_mmu_get_l2pt(struct imgu_mmu *mmu, u32 l1pt_idx) argument 210 __imgu_mmu_map(struct imgu_mmu *mmu, unsigned long iova, phys_addr_t paddr) argument 254 struct imgu_mmu *mmu = to_imgu_mmu(info); local 302 struct imgu_mmu *mmu = to_imgu_mmu(info); local 338 __imgu_mmu_unmap(struct imgu_mmu *mmu, unsigned long iova, size_t size) argument 382 struct imgu_mmu *mmu = to_imgu_mmu(info); local 429 struct imgu_mmu *mmu; local 505 struct imgu_mmu *mmu = to_imgu_mmu(info); local 520 struct imgu_mmu *mmu = to_imgu_mmu(info); local 527 struct imgu_mmu *mmu = to_imgu_mmu(info); local [all...] |
/linux-master/drivers/staging/media/atomisp/pci/mmu/ |
H A D | sh_mmu_mrfld.c | 21 #include "mmu/isp_mmu.h" 22 #include "mmu/sh_mmu_mrfld.h" 31 static unsigned int sh_phys_to_pte(struct isp_mmu *mmu, argument 37 static phys_addr_t sh_pte_to_phys(struct isp_mmu *mmu, argument 40 unsigned int mask = mmu->driver->pte_valid_mask; 45 static unsigned int sh_get_pd_base(struct isp_mmu *mmu, argument 48 unsigned int pte = sh_phys_to_pte(mmu, phys); 64 static void sh_tlb_flush(struct isp_mmu *mmu) argument
|
H A D | isp_mmu.c | 41 #include "mmu/isp_mmu.h" 57 static void free_mmu_map(struct isp_mmu *mmu, unsigned int start_isp_virt, 79 static phys_addr_t isp_pte_to_pgaddr(struct isp_mmu *mmu, argument 82 return mmu->driver->pte_to_phys(mmu, pte); 85 static unsigned int isp_pgaddr_to_pte_valid(struct isp_mmu *mmu, argument 88 unsigned int pte = mmu->driver->phys_to_pte(mmu, phys); 90 return (unsigned int)(pte | ISP_PTE_VALID_MASK(mmu)); 97 static phys_addr_t alloc_page_table(struct isp_mmu *mmu) argument 125 free_page_table(struct isp_mmu *mmu, phys_addr_t page) argument 142 mmu_remap_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int l2_idx, unsigned int isp_virt, phys_addr_t old_phys, phys_addr_t new_phys) argument 161 mmu_unmap_l2_pte_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int l2_idx, unsigned int isp_virt, unsigned int pte) argument 177 mmu_unmap_l1_pte_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, unsigned int isp_virt, unsigned int pte) argument 189 mmu_unmap_l1_pt_error(struct isp_mmu *mmu, unsigned int pte) argument 199 mmu_l2_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument 244 mmu_l1_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument 317 mmu_map(struct isp_mmu *mmu, unsigned int isp_virt, phys_addr_t phys, unsigned int pgnr) argument 363 mmu_l2_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end) argument 401 mmu_l1_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end) argument 448 mmu_unmap(struct isp_mmu *mmu, unsigned int isp_virt, unsigned int pgnr) argument 474 free_mmu_map(struct isp_mmu *mmu, unsigned int start_isp_virt, unsigned int end_isp_virt) argument 486 isp_mmu_map(struct isp_mmu *mmu, unsigned int isp_virt, phys_addr_t phys, unsigned int pgnr) argument 492 isp_mmu_unmap(struct isp_mmu *mmu, unsigned int isp_virt, unsigned int pgnr) argument 498 isp_mmu_flush_tlb_range_default(struct isp_mmu *mmu, unsigned int start, unsigned int size) argument 506 isp_mmu_init(struct isp_mmu *mmu, struct isp_mmu_client *driver) argument 539 isp_mmu_exit(struct isp_mmu *mmu) argument [all...] |
/linux-master/drivers/staging/media/atomisp/include/mmu/ |
H A D | isp_mmu.h | 88 unsigned int (*get_pd_base)(struct isp_mmu *mmu, phys_addr_t pd_base); 100 void (*tlb_flush_range)(struct isp_mmu *mmu, 102 void (*tlb_flush_all)(struct isp_mmu *mmu); 103 unsigned int (*phys_to_pte)(struct isp_mmu *mmu, 105 phys_addr_t (*pte_to_phys)(struct isp_mmu *mmu, 120 #define ISP_PTE_VALID_MASK(mmu) \ 121 ((mmu)->driver->pte_valid_mask) 123 #define ISP_PTE_VALID(mmu, pte) \ 124 ((pte) & ISP_PTE_VALID_MASK(mmu)) 130 * init mmu wit 154 isp_mmu_flush_tlb_all(struct isp_mmu *mmu) argument 162 isp_mmu_flush_tlb_range(struct isp_mmu *mmu, unsigned int start, unsigned int size) argument [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | mmu.h | 13 #include <asm-generic/mmu.h>
|
/linux-master/arch/x86/kernel/ |
H A D | paravirt.c | 191 .mmu.flush_tlb_user = native_flush_tlb_local, 192 .mmu.flush_tlb_kernel = native_flush_tlb_global, 193 .mmu.flush_tlb_one_user = native_flush_tlb_one_user, 194 .mmu.flush_tlb_multi = native_flush_tlb_multi, 195 .mmu.tlb_remove_table = native_tlb_remove_table, 197 .mmu.exit_mmap = paravirt_nop, 198 .mmu.notify_page_enc_status_changed = paravirt_nop, 201 .mmu.read_cr2 = __PV_IS_CALLEE_SAVE(pv_native_read_cr2), 202 .mmu.write_cr2 = pv_native_write_cr2, 203 .mmu [all...] |
/linux-master/drivers/iommu/ |
H A D | ipmmu-vmsa.c | 71 struct ipmmu_vmsa_device *mmu; member in struct:ipmmu_vmsa_domain 149 static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) argument 151 return mmu->root == mmu; 156 struct ipmmu_vmsa_device *mmu = dev_get_drvdata(dev); local 159 if (ipmmu_is_root(mmu)) 160 *rootp = mmu; 177 static u32 ipmmu_read(struct ipmmu_vmsa_device *mmu, unsigned int offset) argument 179 return ioread32(mmu->base + offset); 182 static void ipmmu_write(struct ipmmu_vmsa_device *mmu, unsigne argument 188 ipmmu_ctx_reg(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg) argument 199 ipmmu_ctx_read(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg) argument 205 ipmmu_ctx_write(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg, u32 data) argument 232 ipmmu_utlb_reg(struct ipmmu_vmsa_device *mmu, unsigned int reg) argument 237 ipmmu_imuasid_write(struct ipmmu_vmsa_device *mmu, unsigned int utlb, u32 data) argument 243 ipmmu_imuctr_write(struct ipmmu_vmsa_device *mmu, unsigned int utlb, u32 data) argument 282 struct ipmmu_vmsa_device *mmu = domain->mmu; local 303 struct ipmmu_vmsa_device *mmu = domain->mmu; local 331 ipmmu_domain_allocate_context(struct ipmmu_vmsa_device *mmu, struct ipmmu_vmsa_domain *domain) argument 351 ipmmu_domain_free_context(struct ipmmu_vmsa_device *mmu, unsigned int context_id) argument 490 struct ipmmu_vmsa_device *mmu = domain->mmu; local 539 struct ipmmu_vmsa_device *mmu = dev; local 595 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local 792 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local 836 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local 861 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local 903 ipmmu_device_reset(struct ipmmu_vmsa_device *mmu) argument 1007 struct ipmmu_vmsa_device *mmu; local 1114 struct ipmmu_vmsa_device *mmu = platform_get_drvdata(pdev); local 1126 struct ipmmu_vmsa_device *mmu = dev_get_drvdata(dev); local [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | mmu.h | 21 #include <asm/mmu-arcv2.h>
|
/linux-master/drivers/gpu/drm/panfrost/ |
H A D | panfrost_mmu.h | 19 u32 panfrost_mmu_as_get(struct panfrost_device *pfdev, struct panfrost_mmu *mmu); 20 void panfrost_mmu_as_put(struct panfrost_device *pfdev, struct panfrost_mmu *mmu); 22 struct panfrost_mmu *panfrost_mmu_ctx_get(struct panfrost_mmu *mmu); 23 void panfrost_mmu_ctx_put(struct panfrost_mmu *mmu);
|
/linux-master/drivers/xen/ |
H A D | grant-dma-iommu.c | 36 struct grant_dma_iommu_device *mmu; local 39 mmu = devm_kzalloc(&pdev->dev, sizeof(*mmu), GFP_KERNEL); 40 if (!mmu) 43 mmu->dev = &pdev->dev; 45 ret = iommu_device_register(&mmu->iommu, &grant_dma_iommu_ops, &pdev->dev); 49 platform_set_drvdata(pdev, mmu); 56 struct grant_dma_iommu_device *mmu = platform_get_drvdata(pdev); local 59 iommu_device_unregister(&mmu->iommu);
|
/linux-master/arch/arm64/kvm/hyp/vhe/ |
H A D | tlb.c | 14 struct kvm_s2_mmu *mmu; member in struct:tlb_inv_context 20 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, argument 28 if (vcpu && mmu != vcpu->arch.hw_mmu) 29 cxt->mmu = vcpu->arch.hw_mmu; 31 cxt->mmu = NULL; 63 __load_stage2(mmu, mmu->arch); 80 if (cxt->mmu) 81 __load_stage2(cxt->mmu, cxt->mmu 92 __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa, int level) argument 124 __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, phys_addr_t ipa, int level) argument 156 __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, phys_addr_t start, unsigned long pages) argument 185 __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) argument 201 __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu) argument [all...] |
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | tlb.c | 17 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, argument 65 __load_stage2(mmu, kern_hyp_va(mmu->arch)); 81 void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, argument 87 __tlb_switch_to_guest(mmu, &cxt, false); 111 void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, argument 117 __tlb_switch_to_guest(mmu, &cxt, true); 141 void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, argument 155 __tlb_switch_to_guest(mmu, &cxt, false); 168 void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) argument 182 __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu) argument [all...] |
/linux-master/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | mmu.h | 39 nvif_mmu_kind_valid(struct nvif_mmu *mmu, u8 kind) argument 42 if (kind >= mmu->kind_nr || mmu->kind[kind] == mmu->kind_inv) 49 nvif_mmu_type(struct nvif_mmu *mmu, u8 mask) argument 52 for (i = 0; i < mmu->type_nr; i++) { 53 if ((mmu->type[i].type & mask) == mask)
|