Searched refs:mmu (Results 1 - 25 of 406) sorted by relevance

1234567891011>>

/linux-master/drivers/accel/habanalabs/common/mmu/
H A DMakefile2 HL_COMMON_MMU_FILES := common/mmu/mmu.o common/mmu/mmu_v1.o \
3 common/mmu/mmu_v2.o common/mmu/mmu_v2_hr.o
/linux-master/arch/arm64/include/asm/
H A Dstage2_pgtable.h24 #define kvm_stage2_levels(mmu) VTCR_EL2_LVLS((mmu)->vtcr)
31 #define kvm_mmu_cache_min_pages(mmu) (kvm_stage2_levels(mmu) - 1)
/linux-master/drivers/gpu/drm/nouveau/nvif/
H A Dmmu.c22 #include <nvif/mmu.h>
28 nvif_mmu_dtor(struct nvif_mmu *mmu) argument
30 if (!nvif_object_constructed(&mmu->object))
33 kfree(mmu->kind);
34 kfree(mmu->type);
35 kfree(mmu->heap);
36 nvif_object_dtor(&mmu->object);
41 struct nvif_mmu *mmu)
53 mmu->heap = NULL;
54 mmu
40 nvif_mmu_ctor(struct nvif_object *parent, const char *name, s32 oclass, struct nvif_mmu *mmu) argument
[all...]
/linux-master/fs/ramfs/
H A DMakefile8 file-mmu-y := file-nommu.o
9 file-mmu-$(CONFIG_MMU) := file-mmu.o
10 ramfs-objs += inode.o $(file-mmu-y)
/linux-master/arch/arm64/include/asm/xen/
H A Dpage.h2 #include <asm/mmu.h>
/linux-master/arch/m68k/include/asm/
H A Dmmu.h9 #include <asm-generic/mmu.h>
/linux-master/arch/powerpc/include/asm/nohash/
H A Dmmu.h7 #include <asm/nohash/32/mmu-40x.h>
10 #include <asm/nohash/32/mmu-44x.h>
13 #include <asm/nohash/mmu-e500.h>
16 #include <asm/nohash/32/mmu-8xx.h>
/linux-master/drivers/gpu/drm/msm/
H A Dmsm_mmu.h13 void (*detach)(struct msm_mmu *mmu);
14 int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt,
16 int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len);
17 void (*destroy)(struct msm_mmu *mmu);
18 void (*resume_translation)(struct msm_mmu *mmu);
35 static inline void msm_mmu_init(struct msm_mmu *mmu, struct device *dev, argument
38 mmu->dev = dev;
39 mmu->funcs = funcs;
40 mmu->type = type;
47 static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, voi argument
[all...]
H A Dmsm_gpummu.c24 static void msm_gpummu_detach(struct msm_mmu *mmu) argument
28 static int msm_gpummu_map(struct msm_mmu *mmu, uint64_t iova, argument
31 struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
56 static int msm_gpummu_unmap(struct msm_mmu *mmu, uint64_t iova, size_t len) argument
58 struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
71 static void msm_gpummu_resume_translation(struct msm_mmu *mmu) argument
75 static void msm_gpummu_destroy(struct msm_mmu *mmu) argument
77 struct msm_gpummu *gpummu = to_msm_gpummu(mmu);
79 dma_free_attrs(mmu->dev, TABLE_SIZE, gpummu->table, gpummu->pt_base,
114 void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_ argument
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dbase.c42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) argument
51 list_add(&ptp->head, &mmu->ptp.list);
56 nvkm_mmu_ptc_put(mmu, force, &ptp->pt);
65 nvkm_mmu_ptp_get(struct nvkm_mmu *mmu, u32 size, bool zero) argument
74 ptp = list_first_entry_or_null(&mmu->ptp.list, typeof(*ptp), head);
82 ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false);
93 list_add(&ptp->head, &mmu->ptp.list);
120 nvkm_mmu_ptc_find(struct nvkm_mmu *mmu, u32 size) argument
124 list_for_each_entry(ptc, &mmu->ptc.list, head) {
134 list_add(&ptc->head, &mmu
141 nvkm_mmu_ptc_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt **ppt) argument
167 nvkm_mmu_ptc_get(struct nvkm_mmu *mmu, u32 size, u32 align, bool zero) argument
220 nvkm_mmu_ptc_dump(struct nvkm_mmu *mmu) argument
234 nvkm_mmu_ptc_fini(struct nvkm_mmu *mmu) argument
246 nvkm_mmu_ptc_init(struct nvkm_mmu *mmu) argument
255 nvkm_mmu_type(struct nvkm_mmu *mmu, int heap, u8 type) argument
265 nvkm_mmu_heap(struct nvkm_mmu *mmu, u8 type, u64 size) argument
278 nvkm_mmu_host(struct nvkm_mmu *mmu) argument
315 nvkm_mmu_vram(struct nvkm_mmu *mmu) argument
370 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local
391 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local
400 struct nvkm_mmu *mmu = nvkm_mmu(subdev); local
421 nvkm_mmu_ctor(const struct nvkm_mmu_func *func, struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_mmu *mmu) argument
[all...]
H A Dummu.c35 struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu; local
37 if (mmu->func->mem.user.oclass) {
39 oclass->base = mmu->func->mem.user;
45 if (mmu->func->vmm.user.oclass) {
47 oclass->base = mmu->func->vmm.user;
59 struct nvkm_mmu *mmu = ummu->mmu; local
67 if ((index = args->v0.index) >= mmu->heap_nr)
69 args->v0.size = mmu
79 struct nvkm_mmu *mmu = ummu->mmu; local
108 struct nvkm_mmu *mmu = ummu->mmu; local
159 struct nvkm_mmu *mmu = device->mmu; local
[all...]
H A Dnv41.c32 nv41_mmu_init(struct nvkm_mmu *mmu) argument
34 struct nvkm_device *device = mmu->subdev.device;
35 nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr);
44 .mmu = {{ -1, -1, NVIF_CLASS_MMU_NV04}},
/linux-master/drivers/staging/media/ipu3/
H A Dipu3-mmu.c21 #include "ipu3-mmu.h"
73 * @mmu: MMU to perform the invalidate operation on
78 static void imgu_mmu_tlb_invalidate(struct imgu_mmu *mmu) argument
80 writel(TLB_INVALIDATE, mmu->base + REG_TLB_INVALIDATE);
83 static void call_if_imgu_is_powered(struct imgu_mmu *mmu, argument
84 void (*func)(struct imgu_mmu *mmu))
86 if (!pm_runtime_get_if_in_use(mmu->dev))
89 func(mmu);
90 pm_runtime_put(mmu->dev);
95 * @mmu
101 imgu_mmu_set_halt(struct imgu_mmu *mmu, bool halt) argument
168 imgu_mmu_get_l2pt(struct imgu_mmu *mmu, u32 l1pt_idx) argument
210 __imgu_mmu_map(struct imgu_mmu *mmu, unsigned long iova, phys_addr_t paddr) argument
254 struct imgu_mmu *mmu = to_imgu_mmu(info); local
302 struct imgu_mmu *mmu = to_imgu_mmu(info); local
338 __imgu_mmu_unmap(struct imgu_mmu *mmu, unsigned long iova, size_t size) argument
382 struct imgu_mmu *mmu = to_imgu_mmu(info); local
429 struct imgu_mmu *mmu; local
505 struct imgu_mmu *mmu = to_imgu_mmu(info); local
520 struct imgu_mmu *mmu = to_imgu_mmu(info); local
527 struct imgu_mmu *mmu = to_imgu_mmu(info); local
[all...]
/linux-master/drivers/staging/media/atomisp/pci/mmu/
H A Dsh_mmu_mrfld.c21 #include "mmu/isp_mmu.h"
22 #include "mmu/sh_mmu_mrfld.h"
31 static unsigned int sh_phys_to_pte(struct isp_mmu *mmu, argument
37 static phys_addr_t sh_pte_to_phys(struct isp_mmu *mmu, argument
40 unsigned int mask = mmu->driver->pte_valid_mask;
45 static unsigned int sh_get_pd_base(struct isp_mmu *mmu, argument
48 unsigned int pte = sh_phys_to_pte(mmu, phys);
64 static void sh_tlb_flush(struct isp_mmu *mmu) argument
H A Disp_mmu.c41 #include "mmu/isp_mmu.h"
57 static void free_mmu_map(struct isp_mmu *mmu, unsigned int start_isp_virt,
79 static phys_addr_t isp_pte_to_pgaddr(struct isp_mmu *mmu, argument
82 return mmu->driver->pte_to_phys(mmu, pte);
85 static unsigned int isp_pgaddr_to_pte_valid(struct isp_mmu *mmu, argument
88 unsigned int pte = mmu->driver->phys_to_pte(mmu, phys);
90 return (unsigned int)(pte | ISP_PTE_VALID_MASK(mmu));
97 static phys_addr_t alloc_page_table(struct isp_mmu *mmu) argument
125 free_page_table(struct isp_mmu *mmu, phys_addr_t page) argument
142 mmu_remap_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int l2_idx, unsigned int isp_virt, phys_addr_t old_phys, phys_addr_t new_phys) argument
161 mmu_unmap_l2_pte_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int l2_idx, unsigned int isp_virt, unsigned int pte) argument
177 mmu_unmap_l1_pte_error(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, unsigned int isp_virt, unsigned int pte) argument
189 mmu_unmap_l1_pt_error(struct isp_mmu *mmu, unsigned int pte) argument
199 mmu_l2_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument
244 mmu_l1_map(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end, phys_addr_t phys) argument
317 mmu_map(struct isp_mmu *mmu, unsigned int isp_virt, phys_addr_t phys, unsigned int pgnr) argument
363 mmu_l2_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int l1_idx, phys_addr_t l2_pt, unsigned int start, unsigned int end) argument
401 mmu_l1_unmap(struct isp_mmu *mmu, phys_addr_t l1_pt, unsigned int start, unsigned int end) argument
448 mmu_unmap(struct isp_mmu *mmu, unsigned int isp_virt, unsigned int pgnr) argument
474 free_mmu_map(struct isp_mmu *mmu, unsigned int start_isp_virt, unsigned int end_isp_virt) argument
486 isp_mmu_map(struct isp_mmu *mmu, unsigned int isp_virt, phys_addr_t phys, unsigned int pgnr) argument
492 isp_mmu_unmap(struct isp_mmu *mmu, unsigned int isp_virt, unsigned int pgnr) argument
498 isp_mmu_flush_tlb_range_default(struct isp_mmu *mmu, unsigned int start, unsigned int size) argument
506 isp_mmu_init(struct isp_mmu *mmu, struct isp_mmu_client *driver) argument
539 isp_mmu_exit(struct isp_mmu *mmu) argument
[all...]
/linux-master/drivers/staging/media/atomisp/include/mmu/
H A Disp_mmu.h88 unsigned int (*get_pd_base)(struct isp_mmu *mmu, phys_addr_t pd_base);
100 void (*tlb_flush_range)(struct isp_mmu *mmu,
102 void (*tlb_flush_all)(struct isp_mmu *mmu);
103 unsigned int (*phys_to_pte)(struct isp_mmu *mmu,
105 phys_addr_t (*pte_to_phys)(struct isp_mmu *mmu,
120 #define ISP_PTE_VALID_MASK(mmu) \
121 ((mmu)->driver->pte_valid_mask)
123 #define ISP_PTE_VALID(mmu, pte) \
124 ((pte) & ISP_PTE_VALID_MASK(mmu))
130 * init mmu wit
154 isp_mmu_flush_tlb_all(struct isp_mmu *mmu) argument
162 isp_mmu_flush_tlb_range(struct isp_mmu *mmu, unsigned int start, unsigned int size) argument
[all...]
/linux-master/arch/xtensa/include/asm/
H A Dmmu.h13 #include <asm-generic/mmu.h>
/linux-master/arch/x86/kernel/
H A Dparavirt.c191 .mmu.flush_tlb_user = native_flush_tlb_local,
192 .mmu.flush_tlb_kernel = native_flush_tlb_global,
193 .mmu.flush_tlb_one_user = native_flush_tlb_one_user,
194 .mmu.flush_tlb_multi = native_flush_tlb_multi,
195 .mmu.tlb_remove_table = native_tlb_remove_table,
197 .mmu.exit_mmap = paravirt_nop,
198 .mmu.notify_page_enc_status_changed = paravirt_nop,
201 .mmu.read_cr2 = __PV_IS_CALLEE_SAVE(pv_native_read_cr2),
202 .mmu.write_cr2 = pv_native_write_cr2,
203 .mmu
[all...]
/linux-master/drivers/iommu/
H A Dipmmu-vmsa.c71 struct ipmmu_vmsa_device *mmu; member in struct:ipmmu_vmsa_domain
149 static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) argument
151 return mmu->root == mmu;
156 struct ipmmu_vmsa_device *mmu = dev_get_drvdata(dev); local
159 if (ipmmu_is_root(mmu))
160 *rootp = mmu;
177 static u32 ipmmu_read(struct ipmmu_vmsa_device *mmu, unsigned int offset) argument
179 return ioread32(mmu->base + offset);
182 static void ipmmu_write(struct ipmmu_vmsa_device *mmu, unsigne argument
188 ipmmu_ctx_reg(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg) argument
199 ipmmu_ctx_read(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg) argument
205 ipmmu_ctx_write(struct ipmmu_vmsa_device *mmu, unsigned int context_id, unsigned int reg, u32 data) argument
232 ipmmu_utlb_reg(struct ipmmu_vmsa_device *mmu, unsigned int reg) argument
237 ipmmu_imuasid_write(struct ipmmu_vmsa_device *mmu, unsigned int utlb, u32 data) argument
243 ipmmu_imuctr_write(struct ipmmu_vmsa_device *mmu, unsigned int utlb, u32 data) argument
282 struct ipmmu_vmsa_device *mmu = domain->mmu; local
303 struct ipmmu_vmsa_device *mmu = domain->mmu; local
331 ipmmu_domain_allocate_context(struct ipmmu_vmsa_device *mmu, struct ipmmu_vmsa_domain *domain) argument
351 ipmmu_domain_free_context(struct ipmmu_vmsa_device *mmu, unsigned int context_id) argument
490 struct ipmmu_vmsa_device *mmu = domain->mmu; local
539 struct ipmmu_vmsa_device *mmu = dev; local
595 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local
792 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local
836 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local
861 struct ipmmu_vmsa_device *mmu = to_ipmmu(dev); local
903 ipmmu_device_reset(struct ipmmu_vmsa_device *mmu) argument
1007 struct ipmmu_vmsa_device *mmu; local
1114 struct ipmmu_vmsa_device *mmu = platform_get_drvdata(pdev); local
1126 struct ipmmu_vmsa_device *mmu = dev_get_drvdata(dev); local
[all...]
/linux-master/arch/arc/include/asm/
H A Dmmu.h21 #include <asm/mmu-arcv2.h>
/linux-master/drivers/gpu/drm/panfrost/
H A Dpanfrost_mmu.h19 u32 panfrost_mmu_as_get(struct panfrost_device *pfdev, struct panfrost_mmu *mmu);
20 void panfrost_mmu_as_put(struct panfrost_device *pfdev, struct panfrost_mmu *mmu);
22 struct panfrost_mmu *panfrost_mmu_ctx_get(struct panfrost_mmu *mmu);
23 void panfrost_mmu_ctx_put(struct panfrost_mmu *mmu);
/linux-master/drivers/xen/
H A Dgrant-dma-iommu.c36 struct grant_dma_iommu_device *mmu; local
39 mmu = devm_kzalloc(&pdev->dev, sizeof(*mmu), GFP_KERNEL);
40 if (!mmu)
43 mmu->dev = &pdev->dev;
45 ret = iommu_device_register(&mmu->iommu, &grant_dma_iommu_ops, &pdev->dev);
49 platform_set_drvdata(pdev, mmu);
56 struct grant_dma_iommu_device *mmu = platform_get_drvdata(pdev); local
59 iommu_device_unregister(&mmu->iommu);
/linux-master/arch/arm64/kvm/hyp/vhe/
H A Dtlb.c14 struct kvm_s2_mmu *mmu; member in struct:tlb_inv_context
20 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, argument
28 if (vcpu && mmu != vcpu->arch.hw_mmu)
29 cxt->mmu = vcpu->arch.hw_mmu;
31 cxt->mmu = NULL;
63 __load_stage2(mmu, mmu->arch);
80 if (cxt->mmu)
81 __load_stage2(cxt->mmu, cxt->mmu
92 __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa, int level) argument
124 __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, phys_addr_t ipa, int level) argument
156 __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, phys_addr_t start, unsigned long pages) argument
185 __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) argument
201 __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu) argument
[all...]
/linux-master/arch/arm64/kvm/hyp/nvhe/
H A Dtlb.c17 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, argument
65 __load_stage2(mmu, kern_hyp_va(mmu->arch));
81 void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, argument
87 __tlb_switch_to_guest(mmu, &cxt, false);
111 void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, argument
117 __tlb_switch_to_guest(mmu, &cxt, true);
141 void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, argument
155 __tlb_switch_to_guest(mmu, &cxt, false);
168 void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) argument
182 __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu) argument
[all...]
/linux-master/drivers/gpu/drm/nouveau/include/nvif/
H A Dmmu.h39 nvif_mmu_kind_valid(struct nvif_mmu *mmu, u8 kind) argument
42 if (kind >= mmu->kind_nr || mmu->kind[kind] == mmu->kind_inv)
49 nvif_mmu_type(struct nvif_mmu *mmu, u8 mask) argument
52 for (i = 0; i < mmu->type_nr; i++) {
53 if ((mmu->type[i].type & mask) == mask)

Completed in 490 milliseconds

1234567891011>>