/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmgm200.c | 144 struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, 168 return nvkm_vmm_new_(func, mmu, 0, managed, addr, size, key, name, pvmm); 172 gm200_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, argument 176 return gm200_vmm_new_(&gm200_vmm_16, &gm200_vmm_17, mmu, managed, addr, 181 gm200_vmm_new_fixed(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, argument 185 return gf100_vmm_new_(&gm200_vmm_16, &gm200_vmm_17, mmu, managed, addr, 142 gm200_vmm_new_(const struct nvkm_vmm_func *func_16, const struct nvkm_vmm_func *func_17, struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, void *argv, u32 argc, struct lock_class_key *key, const char *name, struct nvkm_vmm **pvmm) argument
|
H A D | memnv50.c | 34 nv50_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv, argument 42 struct nvkm_device *device = mmu->subdev.device; 66 nv50_mem_new(struct nvkm_mmu *mmu, int type, u8 page, u64 size, argument 86 return nvkm_ram_get(mmu->subdev.device, NVKM_RAM_MM_NORMAL, type,
|
H A D | gp100.c | 32 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
|
H A D | gp10b.c | 32 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
|
H A D | vmmgk104.c | 98 gk104_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, argument 102 return gf100_vmm_new_(&gk104_vmm_16, &gk104_vmm_17, mmu, managed, addr,
|
H A D | r535.c | 35 ret = nvkm_gsp_client_device_ctor(vmm->mmu->subdev.device->gsp, 93 r535_mmu_dtor(struct nvkm_mmu *mmu) argument 95 kfree(mmu->func); 111 rm->mmu = hw->mmu;
|
H A D | vmmgf100.c | 183 struct nvkm_device *device = vmm->mmu->subdev.device; 190 struct nvkm_device *device = vmm->mmu->subdev.device; 194 mutex_lock(&vmm->mmu->mutex); 224 mutex_unlock(&vmm->mmu->mutex); 247 struct nvkm_device *device = vmm->mmu->subdev.device; 276 kindm = vmm->mmu->func->kind(vmm->mmu, &kindn, &kind_inv); 404 struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, 408 switch (mmu->subdev.device->fb->page) { 409 case 16: return nv04_vmm_new_(func_16, mmu, 402 gf100_vmm_new_(const struct nvkm_vmm_func *func_16, const struct nvkm_vmm_func *func_17, struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, void *argv, u32 argc, struct lock_class_key *key, const char *name, struct nvkm_vmm **pvmm) argument 420 gf100_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, void *argv, u32 argc, struct lock_class_key *key, const char *name, struct nvkm_vmm **pvmm) argument [all...] |
H A D | vmmnv04.c | 102 nv04_vmm_new_(const struct nvkm_vmm_func *func, struct nvkm_mmu *mmu, argument 112 ret = nvkm_vmm_new_(func, mmu, pd_header, managed, addr, size, 121 nv04_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, argument 129 ret = nv04_vmm_new_(&nv04_vmm, mmu, 8, managed, addr, size,
|
/linux-master/arch/x86/kvm/ |
H A D | Makefile | 10 debugfs.o mmu/mmu.o mmu/page_track.o \ 11 mmu/spte.o 13 kvm-$(CONFIG_X86_64) += mmu/tdp_iter.o mmu/tdp_mmu.o
|
/linux-master/arch/sh/mm/ |
H A D | Makefile | 18 mmu-y := nommu.o extable_32.o 19 mmu-$(CONFIG_MMU) := extable_32.o fault.o ioremap.o kmap.o \ 22 obj-y += $(mmu-y)
|
/linux-master/drivers/gpu/drm/panfrost/ |
H A D | panfrost_gem.c | 63 if (iter->mmu == priv->mmu) { 80 spin_lock(&mapping->mmu->mm_lock); 83 spin_unlock(&mapping->mmu->mm_lock); 94 panfrost_mmu_ctx_put(mapping->mmu); 144 mapping->mmu = panfrost_mmu_ctx_get(priv->mmu); 145 spin_lock(&mapping->mmu->mm_lock); 146 ret = drm_mm_insert_node_generic(&mapping->mmu->mm, &mapping->mmnode, 148 spin_unlock(&mapping->mmu [all...] |
/linux-master/include/linux/fsl/bestcomm/ |
H A D | sram.h | 16 #include <asm/mmu.h>
|
/linux-master/arch/powerpc/sysdev/ |
H A D | fsl_soc.h | 6 #include <asm/mmu.h>
|
/linux-master/arch/nios2/include/asm/ |
H A D | cpuinfo.h | 15 bool mmu; member in struct:cpuinfo
|
/linux-master/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 574 lockdep_assert_held(&ptdev->mmu->as.slots_lock); 595 mutex_lock(&ptdev->mmu->as.slots_lock); 597 mutex_unlock(&ptdev->mmu->as.slots_lock); 680 lockdep_assert_held(&ptdev->mmu->as.slots_lock); 685 ptdev->mmu->as.slots[vm->as.id].vm = NULL; 686 clear_bit(vm->as.id, &ptdev->mmu->as.alloc_mask); 714 mutex_lock(&ptdev->mmu->as.slots_lock); 724 if (ptdev->mmu->as.faulty_mask & panthor_mmu_as_fault_mask(ptdev, as)) 732 drm_WARN_ON(&ptdev->base, ptdev->mmu->as.alloc_mask & BIT(0)); 735 as = ffz(ptdev->mmu 2659 struct panthor_mmu *mmu; local [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_32_mmu.c | 399 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; local 401 mmu->mtsrin = kvmppc_mmu_book3s_32_mtsrin; 402 mmu->mfsrin = kvmppc_mmu_book3s_32_mfsrin; 403 mmu->xlate = kvmppc_mmu_book3s_32_xlate; 404 mmu->tlbie = kvmppc_mmu_book3s_32_tlbie; 405 mmu->esid_to_vsid = kvmppc_mmu_book3s_32_esid_to_vsid; 406 mmu->ea_to_vp = kvmppc_mmu_book3s_32_ea_to_vp; 407 mmu->is_dcbz32 = kvmppc_mmu_book3s_32_is_dcbz32; 409 mmu [all...] |
/linux-master/drivers/accel/habanalabs/common/mmu/ |
H A D | mmu_v2_hr.c | 9 #include "../../include/hw_ip/mmu/mmu_general.h" 86 * hl_mmu_v2_hr_ctx_fini - disable a ctx from using the mmu module 378 * hl_mmu_v2_prepare - prepare mmu_if for working with mmu v2 381 * @mmu_if: pointer to the mmu interface structure 383 void hl_mmu_v2_hr_set_funcs(struct hl_device *hdev, struct hl_mmu_funcs *mmu) argument 385 mmu->init = hl_mmu_v2_hr_init; 386 mmu->fini = hl_mmu_v2_hr_fini; 387 mmu->ctx_init = hl_mmu_v2_hr_ctx_init; 388 mmu->ctx_fini = hl_mmu_v2_hr_ctx_fini; 389 mmu [all...] |
H A D | mmu_v2.c | 9 #include "../../include/hw_ip/mmu/mmu_general.h" 10 #include "../../include/hw_ip/mmu/mmu_v2_0.h" 30 * hl_mmu_v2_ctx_fini - disable a ctx from using the mmu module 321 * hl_mmu_v2_prepare - prepare mmu_if for working with mmu v2 324 * @mmu_if: pointer to the mmu interface structure 326 void hl_mmu_v2_set_funcs(struct hl_device *hdev, struct hl_mmu_funcs *mmu) argument 328 mmu->init = hl_mmu_dr_init; 329 mmu->fini = hl_mmu_dr_fini; 330 mmu->ctx_init = hl_mmu_v2_ctx_init; 331 mmu [all...] |
/linux-master/arch/powerpc/boot/dts/fsl/ |
H A D | e500v2_power_isa.dtsi | 50 mmu-type = "power-embedded";
|
/linux-master/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | lut.c | 67 nv50_lut_init(struct nv50_disp *disp, struct nvif_mmu *mmu, argument 73 int ret = nvif_mem_ctor_map(mmu, "kmsLut", NVIF_MEM_VRAM,
|
/linux-master/arch/mips/kvm/ |
H A D | Makefile | 15 kvm-y += mmu.o
|
/linux-master/arch/loongarch/kvm/ |
H A D | Makefile | 15 kvm-y += mmu.o
|
/linux-master/drivers/staging/media/ipu3/ |
H A D | ipu3-dmamap.c | 14 #include "ipu3-mmu.h" 108 imgu->mmu->aperture_end >> shift, 0); 119 rval = imgu_mmu_map(imgu->mmu, iovaddr, 142 imgu_mmu_unmap(imgu->mmu, iova_dma_addr(&imgu->iova_domain, iova), 160 imgu_mmu_unmap(imgu->mmu, iova_dma_addr(&imgu->iova_domain, iova), 208 imgu->mmu->aperture_end >> shift, 0); 215 if (imgu_mmu_map_sg(imgu->mmu, iova_dma_addr(&imgu->iova_domain, iova), 240 base_pfn = max_t(unsigned long, 1, imgu->mmu->aperture_start >> order);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
H A D | usernv04.c | 29 #include <subdev/mmu/vmm.h> 53 device->mmu->vmm->pd->pt[0]->memory; 98 if (device->mmu->func == &nv04_mmu)
|
/linux-master/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 585 pgt->mmu = NULL; 627 struct kvm_s2_mmu *mmu; member in struct:stage2_map_data 695 void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, argument 701 kvm_call_hyp(__kvm_tlb_flush_vmid, mmu); 708 kvm_call_hyp(__kvm_tlb_flush_vmid_range, mmu, addr, inval_pages); 816 * @mmu: stage-2 mmu 825 struct kvm_s2_mmu *mmu) 850 kvm_tlb_flush_vmid_range(mmu, addr, size); 852 kvm_call_hyp(__kvm_tlb_flush_vmid_ipa, mmu, 824 stage2_try_break_pte(const struct kvm_pgtable_visit_ctx *ctx, struct kvm_s2_mmu *mmu) argument 888 stage2_unmap_put_pte(const struct kvm_pgtable_visit_ctx *ctx, struct kvm_s2_mmu *mmu, struct kvm_pgtable_mm_ops *mm_ops) argument 1154 struct kvm_s2_mmu *mmu = pgt->mmu; local 1490 struct kvm_s2_mmu *mmu; local 1566 __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu, struct kvm_pgtable_mm_ops *mm_ops, enum kvm_pgtable_stage2_flags flags, kvm_pgtable_force_pte_cb_t force_pte_cb) argument [all...] |