Searched refs:mmu (Results 151 - 175 of 416) sorted by relevance

1234567891011>>

/linux-master/arch/arm64/kvm/hyp/nvhe/
H A Dpkvm.c321 hyp_vm->kvm.arch.mmu.vtcr = host_mmu.arch.mmu.vtcr;
345 hyp_vcpu->vcpu.arch.hw_mmu = &hyp_vm->kvm.arch.mmu;
374 struct kvm_s2_mmu *mmu = &hyp_vm->kvm.arch.mmu; local
394 atomic64_set(&mmu->vmid.id, idx + 1);
396 mmu->arch = &hyp_vm->kvm.arch;
397 mmu->pgt = &hyp_vm->pgt;
502 pgd_size = kvm_pgtable_stage2_pgd_size(host_mmu.arch.mmu.vtcr);
628 __kvm_tlb_flush_vmid(&hyp_vm->kvm.arch.mmu);
[all...]
/linux-master/arch/arm64/include/asm/
H A Dkvm_pgtable.h367 * @mmu: Stage-2 KVM MMU struct. Unused for stage-1 page-tables.
379 struct kvm_s2_mmu *mmu; member in struct:kvm_pgtable
474 * @mmu: S2 MMU context for this S2 translation
482 int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
487 #define kvm_pgtable_stage2_init(pgt, mmu, mm_ops) \
488 __kvm_pgtable_stage2_init(pgt, mmu, mm_ops, 0, NULL)
783 * @mmu: Stage-2 KVM MMU struct
787 void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
H A Dkvm_asm.h225 extern void __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu);
226 extern void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa,
228 extern void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu,
231 extern void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
233 extern void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu);
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dgf100.c33 gf100_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) argument
79 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
H A Duvmm.c535 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; local
560 if (!mmu->vmm) {
561 ret = mmu->func->vmm.ctor(mmu, managed || raw, addr, size,
571 uvmm->vmm = nvkm_vmm_ref(mmu->vmm);
575 if (mmu->func->promote_vmm) {
576 ret = mmu->func->promote_vmm(uvmm->vmm);
H A Dgm200.c30 gm200_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) argument
75 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
85 .mmu = {{ -1, -1, NVIF_CLASS_MMU_GF100}},
/linux-master/drivers/gpu/drm/etnaviv/
H A Detnaviv_drv.h33 struct etnaviv_iommu_context *mmu; member in struct:etnaviv_file_private
78 struct etnaviv_iommu_context *mmu,
/linux-master/arch/nios2/kernel/
H A Dcpuinfo.c59 cpuinfo.mmu = of_property_read_bool(cpu, "altr,has-mmu");
133 cpuinfo.mmu ? "present" : "none",
/linux-master/arch/microblaze/include/asm/
H A Dmmu_context_mm.h16 #include <asm/mmu.h>
75 * Get a new mmu context for the address space described by `mm'.
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/acr/
H A Dgp108.c24 #include <subdev/mmu.h>
/linux-master/arch/x86/xen/
H A DMakefile15 obj-y += mmu.o
/linux-master/arch/powerpc/platforms/44x/
H A Diss4xx.c29 #include <asm/mmu.h>
/linux-master/arch/arm64/kvm/hyp/include/nvhe/
H A Dmem_protect.h89 __load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
/linux-master/arch/sh/mm/
H A Dioremap.c27 #include <asm/mmu.h>
/linux-master/arch/powerpc/mm/
H A Dinit_32.c33 #include <asm/mmu.h>
H A Dpageattr.c13 #include <asm/mmu.h>
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/device/
H A Dpriv.h22 #include <subdev/mmu.h>
/linux-master/arch/powerpc/platforms/pseries/
H A Dsuspend.c13 #include <asm/mmu.h>
/linux-master/arch/arc/include/asm/
H A Dentry.h14 #include <asm/mmu.h>
/linux-master/drivers/gpu/drm/lima/
H A Dlima_sched.h58 struct lima_ip *mmu[LIMA_SCHED_PIPE_MAX_MMU]; member in struct:lima_sched_pipe
/linux-master/arch/x86/kvm/mmu/
H A Dmmu_internal.h121 * Used out of the mmu-lock to avoid reading spte values while an
203 /* Derived from mmu and global state. */
255 * Return values of handle_mmio_page_fault(), mmu.page_fault(), fast_page_fault(),
302 .is_tdp = likely(vcpu->arch.mmu->page_fault == kvm_tdp_page_fault),
316 if (vcpu->arch.mmu->root_role.direct) {
332 r = vcpu->arch.mmu->page_fault(vcpu, &fault);
H A Dtdp_mmu.c4 #include "mmu.h"
228 struct kvm_mmu *mmu = vcpu->arch.mmu; local
229 union kvm_mmu_page_role role = mmu->root_role;
285 mmu->root.hpa = __pa(root->spt);
286 mmu->root.pgd = 0;
1107 struct kvm_mmu *mmu = vcpu->arch.mmu; local
1119 tdp_mmu_for_each_pte(iter, mmu, fault->gfn, fault->gfn + 1) {
1779 struct kvm_mmu *mmu local
1808 struct kvm_mmu *mmu = vcpu->arch.mmu; local
[all...]
/linux-master/arch/powerpc/include/asm/
H A Dmmu.h385 #include <asm/book3s/64/mmu.h>
404 #include <asm/book3s/32/mmu-hash.h>
406 #include <asm/nohash/mmu.h>
/linux-master/drivers/gpu/drm/msm/adreno/
H A Dadreno_gpu.c207 struct msm_mmu *mmu; local
211 mmu = msm_iommu_gpu_new(&pdev->dev, gpu, quirks);
212 if (IS_ERR_OR_NULL(mmu))
213 return ERR_CAST(mmu);
215 geometry = msm_iommu_get_geometry(mmu);
227 aspace = msm_gem_address_space_create(mmu, "gpu",
230 if (IS_ERR(aspace) && !IS_ERR(mmu))
231 mmu->funcs->destroy(mmu);
265 gpu->aspace->mmu
[all...]
/linux-master/drivers/staging/media/atomisp/pci/hmm/
H A Dhmm_bo.c342 ret = isp_mmu_init(&bdev->mmu, mmu_driver);
365 isp_mmu_exit(&bdev->mmu);
372 isp_mmu_exit(&bdev->mmu);
380 isp_mmu_exit(&bdev->mmu);
536 isp_mmu_exit(&bdev->mmu);
793 isp_mmu_map(&bdev->mmu, virt,
812 isp_mmu_flush_tlb_range(&bdev->mmu, bo->start,
825 isp_mmu_unmap(&bdev->mmu, virt, 1);
868 isp_mmu_unmap(&bdev->mmu, virt, 1);
876 isp_mmu_flush_tlb_range(&bdev->mmu, b
[all...]

Completed in 229 milliseconds

1234567891011>>