/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | pkvm.c | 321 hyp_vm->kvm.arch.mmu.vtcr = host_mmu.arch.mmu.vtcr; 345 hyp_vcpu->vcpu.arch.hw_mmu = &hyp_vm->kvm.arch.mmu; 374 struct kvm_s2_mmu *mmu = &hyp_vm->kvm.arch.mmu; local 394 atomic64_set(&mmu->vmid.id, idx + 1); 396 mmu->arch = &hyp_vm->kvm.arch; 397 mmu->pgt = &hyp_vm->pgt; 502 pgd_size = kvm_pgtable_stage2_pgd_size(host_mmu.arch.mmu.vtcr); 628 __kvm_tlb_flush_vmid(&hyp_vm->kvm.arch.mmu); [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_pgtable.h | 367 * @mmu: Stage-2 KVM MMU struct. Unused for stage-1 page-tables. 379 struct kvm_s2_mmu *mmu; member in struct:kvm_pgtable 474 * @mmu: S2 MMU context for this S2 translation 482 int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu, 487 #define kvm_pgtable_stage2_init(pgt, mmu, mm_ops) \ 488 __kvm_pgtable_stage2_init(pgt, mmu, mm_ops, 0, NULL) 783 * @mmu: Stage-2 KVM MMU struct 787 void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
|
H A D | kvm_asm.h | 225 extern void __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu); 226 extern void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa, 228 extern void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, 231 extern void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, 233 extern void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | gf100.c | 33 gf100_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) argument 79 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
|
H A D | uvmm.c | 535 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; local 560 if (!mmu->vmm) { 561 ret = mmu->func->vmm.ctor(mmu, managed || raw, addr, size, 571 uvmm->vmm = nvkm_vmm_ref(mmu->vmm); 575 if (mmu->func->promote_vmm) { 576 ret = mmu->func->promote_vmm(uvmm->vmm);
|
H A D | gm200.c | 30 gm200_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) argument 75 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}}, 85 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
|
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_drv.h | 33 struct etnaviv_iommu_context *mmu; member in struct:etnaviv_file_private 78 struct etnaviv_iommu_context *mmu,
|
/linux-master/arch/nios2/kernel/ |
H A D | cpuinfo.c | 59 cpuinfo.mmu = of_property_read_bool(cpu, "altr,has-mmu"); 133 cpuinfo.mmu ? "present" : "none",
|
/linux-master/arch/microblaze/include/asm/ |
H A D | mmu_context_mm.h | 16 #include <asm/mmu.h> 75 * Get a new mmu context for the address space described by `mm'.
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/acr/ |
H A D | gp108.c | 24 #include <subdev/mmu.h>
|
/linux-master/arch/x86/xen/ |
H A D | Makefile | 15 obj-y += mmu.o
|
/linux-master/arch/powerpc/platforms/44x/ |
H A D | iss4xx.c | 29 #include <asm/mmu.h>
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | mem_protect.h | 89 __load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
|
/linux-master/arch/sh/mm/ |
H A D | ioremap.c | 27 #include <asm/mmu.h>
|
/linux-master/arch/powerpc/mm/ |
H A D | init_32.c | 33 #include <asm/mmu.h>
|
H A D | pageattr.c | 13 #include <asm/mmu.h>
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/device/ |
H A D | priv.h | 22 #include <subdev/mmu.h>
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | suspend.c | 13 #include <asm/mmu.h>
|
/linux-master/arch/arc/include/asm/ |
H A D | entry.h | 14 #include <asm/mmu.h>
|
/linux-master/drivers/gpu/drm/lima/ |
H A D | lima_sched.h | 58 struct lima_ip *mmu[LIMA_SCHED_PIPE_MAX_MMU]; member in struct:lima_sched_pipe
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu_internal.h | 121 * Used out of the mmu-lock to avoid reading spte values while an 203 /* Derived from mmu and global state. */ 255 * Return values of handle_mmio_page_fault(), mmu.page_fault(), fast_page_fault(), 302 .is_tdp = likely(vcpu->arch.mmu->page_fault == kvm_tdp_page_fault), 316 if (vcpu->arch.mmu->root_role.direct) { 332 r = vcpu->arch.mmu->page_fault(vcpu, &fault);
|
H A D | tdp_mmu.c | 4 #include "mmu.h" 228 struct kvm_mmu *mmu = vcpu->arch.mmu; local 229 union kvm_mmu_page_role role = mmu->root_role; 285 mmu->root.hpa = __pa(root->spt); 286 mmu->root.pgd = 0; 1107 struct kvm_mmu *mmu = vcpu->arch.mmu; local 1119 tdp_mmu_for_each_pte(iter, mmu, fault->gfn, fault->gfn + 1) { 1779 struct kvm_mmu *mmu local 1808 struct kvm_mmu *mmu = vcpu->arch.mmu; local [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | mmu.h | 385 #include <asm/book3s/64/mmu.h> 404 #include <asm/book3s/32/mmu-hash.h> 406 #include <asm/nohash/mmu.h>
|
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | adreno_gpu.c | 207 struct msm_mmu *mmu; local 211 mmu = msm_iommu_gpu_new(&pdev->dev, gpu, quirks); 212 if (IS_ERR_OR_NULL(mmu)) 213 return ERR_CAST(mmu); 215 geometry = msm_iommu_get_geometry(mmu); 227 aspace = msm_gem_address_space_create(mmu, "gpu", 230 if (IS_ERR(aspace) && !IS_ERR(mmu)) 231 mmu->funcs->destroy(mmu); 265 gpu->aspace->mmu [all...] |
/linux-master/drivers/staging/media/atomisp/pci/hmm/ |
H A D | hmm_bo.c | 342 ret = isp_mmu_init(&bdev->mmu, mmu_driver); 365 isp_mmu_exit(&bdev->mmu); 372 isp_mmu_exit(&bdev->mmu); 380 isp_mmu_exit(&bdev->mmu); 536 isp_mmu_exit(&bdev->mmu); 793 isp_mmu_map(&bdev->mmu, virt, 812 isp_mmu_flush_tlb_range(&bdev->mmu, bo->start, 825 isp_mmu_unmap(&bdev->mmu, virt, 1); 868 isp_mmu_unmap(&bdev->mmu, virt, 1); 876 isp_mmu_flush_tlb_range(&bdev->mmu, b [all...] |