Searched refs:vmm (Results 26 - 50 of 96) sorted by relevance

1234

/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/bar/
H A Dg84.c50 .bar1.vmm = nv50_bar_bar1_vmm,
54 .bar2.vmm = nv50_bar_bar2_vmm,
H A Dgm107.c53 .bar1.vmm = gf100_bar_bar1_vmm,
57 .bar2.vmm = gf100_bar_bar2_vmm,
H A Dgf100.h9 struct nvkm_vmm *vmm; member in struct:gf100_barN
H A Dgk20a.c30 .bar1.vmm = gf100_bar_bar1_vmm,
H A Dgm20b.c30 .bar1.vmm = gf100_bar_bar1_vmm,
H A Dpriv.h22 struct nvkm_vmm *(*vmm)(struct nvkm_bar *); member in struct:nvkm_bar_func::__anon249
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dnv41.c25 #include "vmm.h"
35 nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr);
46 .vmm = {{ -1, -1, NVIF_CLASS_VMM_NV04}, nv41_vmm_new, true },
H A Dgv100.c23 #include "vmm.h"
34 .vmm = {{ -1, 0, NVIF_CLASS_VMM_GP100}, gv100_vmm_new },
H A Dmcp77.c23 #include "vmm.h"
32 .vmm = {{ -1, -1, NVIF_CLASS_VMM_NV50}, mcp77_vmm_new, false, 0x0200 },
H A Dg84.c23 #include "vmm.h"
32 .vmm = {{ -1, -1, NVIF_CLASS_VMM_NV50}, nv50_vmm_new, false, 0x0200 },
H A Dnv44.c25 #include "vmm.h"
35 struct nvkm_memory *pt = mmu->vmm->pd->pt[0]->memory;
46 nvkm_wr32(device, 0x100818, mmu->vmm->null);
61 .vmm = {{ -1, -1, NVIF_CLASS_VMM_NV04}, nv44_vmm_new, true },
H A Dgp100.c23 #include "vmm.h"
34 .vmm = {{ -1, 0, NVIF_CLASS_VMM_GP100}, gp100_vmm_new },
H A Dgp10b.c23 #include "vmm.h"
34 .vmm = {{ -1, 0, NVIF_CLASS_VMM_GP100}, gp10b_vmm_new },
H A Dvmmgk104.c22 #include "vmm.h"
25 gk104_vmm_lpt_invalid(struct nvkm_vmm *vmm, argument
29 VMM_FO064(pt, vmm, ptei * 8, BIT_ULL(1) /* PRIV. */, ptes);
H A Dtu102.c24 #include "vmm.h"
49 .vmm = {{ -1, 0, NVIF_CLASS_VMM_GP100}, tu102_vmm_new },
H A Dvmmgv100.c22 #include "vmm.h"
31 gv100_vmm_join(struct nvkm_vmm *vmm, struct nvkm_memory *inst) argument
34 int ret = gp100_vmm_join(vmm, inst), i;
/linux-master/arch/xtensa/kernel/
H A Dsyscall.c60 struct vm_area_struct *vmm; local
84 for_each_vma(vmi, vmm) {
85 /* At this point: (addr < vmm->vm_end). */
86 if (addr + len <= vm_start_gap(vmm))
89 addr = vmm->vm_end;
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dr535.c26 #include <subdev/mmu/vmm.h>
65 struct nvkm_vmm *vmm; member in struct:r535_gr_chan
115 nvkm_vmm_put(grc->vmm, &grc->vma[i]);
119 nvkm_vmm_unref(&grc->vmm);
129 r535_gr_promote_ctx(struct r535_gr *gr, bool golden, struct nvkm_vmm *vmm, argument
137 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.subdevice,
143 ctrl->hChanClient = vmm->rm.client.object.handle;
180 mutex_lock(&vmm->mutex.vmm);
181 ret = nvkm_vmm_get_locked(vmm, fals
254 struct nvkm_vmm *vmm; member in struct:__anon247
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dnv50.c121 nv50_instobj_kmap(struct nv50_instobj *iobj, struct nvkm_vmm *vmm) argument
138 while ((ret = nvkm_vmm_get(vmm, 12, size, &bar))) {
159 nvkm_vmm_put(vmm, &ebar);
163 ret = nvkm_memory_map(memory, 0, vmm, bar, NULL, 0);
168 nvkm_vmm_put(vmm, &bar);
179 nvkm_vmm_put(vmm, &iobj->bar);
184 nv50_instobj_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm, argument
188 return nvkm_memory_map(memory, offset, vmm, vma, argv, argc);
221 struct nvkm_vmm *vmm; local
241 if ((vmm
266 nv50_instobj_boot(struct nvkm_memory *memory, struct nvkm_vmm *vmm) argument
332 struct nvkm_vmm *vmm = nvkm_bar_bar2_vmm(imem->subdev.device); local
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_mem.h7 #include <nvif/vmm.h>
40 struct nvif_vmm *vmm,
H A Dnouveau_svm.c29 #include <nvif/vmm.h>
105 NV_DEBUG((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a)
107 NV_WARN((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a)
210 mutex_lock(&svmm->vmm->cli->drm->svm->mutex);
211 ivmm = nouveau_ivmm_find(svmm->vmm->cli->drm->svm, inst);
216 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex);
231 mutex_lock(&svmm->vmm->cli->drm->svm->mutex);
232 list_add(&ivmm->head, &svmm->vmm->cli->drm->svm->inst);
233 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex);
243 nvif_object_mthd(&svmm->vmm
[all...]
H A Dnouveau_mem.c38 struct nvif_vmm *vmm, struct nvif_vma *vma)
46 switch (vmm->object.oclass) {
75 return nvif_vmm_map(vmm, vma->addr, mem->mem.size, &args, argc, &mem->mem, 0);
81 nvif_vmm_put(&mem->cli->drm->client.vmm.vmm, &mem->vma[1]);
82 nvif_vmm_put(&mem->cli->drm->client.vmm.vmm, &mem->vma[0]);
37 nouveau_mem_map(struct nouveau_mem *mem, struct nvif_vmm *vmm, struct nvif_vma *vma) argument
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Dchan.c103 cctx->vctx->ectx->engn == engn && cctx->vctx->vmm == chan->vmm);
288 if (chan->vmm) {
289 nvkm_vmm_part(chan->vmm, chan->inst->memory);
290 nvkm_vmm_unref(&chan->vmm);
350 struct nvkm_cgrp *cgrp, const char *name, bool priv, u32 devm, struct nvkm_vmm *vmm,
361 (!func->inst->vmm != !vmm) ||
366 RUNL_DEBUG(runl, "args runq:%d:%d vmm:%d:%p userd:%d:%p "
368 runl->func->runqs, runq, func->inst->vmm, vm
349 nvkm_chan_new_(const struct nvkm_chan_func *func, struct nvkm_runl *runl, int runq, struct nvkm_cgrp *cgrp, const char *name, bool priv, u32 devm, struct nvkm_vmm *vmm, struct nvkm_dmaobj *dmaobj, u64 offset, u64 length, struct nvkm_memory *userd, u64 ouserd, struct nvkm_chan **pchan) argument
[all...]
/linux-master/drivers/gpu/drm/
H A Ddrm_gem_vram_helper.c191 struct drm_vram_mm *vmm = dev->vram_mm; local
195 if (WARN_ONCE(!vmm, "VRAM MM not initialized"))
219 bdev = &vmm->bdev;
911 struct drm_vram_mm *vmm = drm_vram_mm_of_bdev(bdev); local
917 mem->bus.offset = (mem->start << PAGE_SHIFT) + vmm->vram_base;
945 struct drm_vram_mm *vmm = entry->dev->vram_mm; local
946 struct ttm_resource_manager *man = ttm_manager_type(&vmm->bdev, TTM_PL_VRAM);
970 static int drm_vram_mm_init(struct drm_vram_mm *vmm, struct drm_device *dev, argument
975 vmm->vram_base = vram_base;
976 vmm
993 drm_vram_mm_cleanup(struct drm_vram_mm *vmm) argument
1081 struct drm_vram_mm *vmm = dev->vram_mm; local
[all...]
/linux-master/drivers/gpu/drm/nouveau/include/nvif/
H A Dif0020.h14 __u64 vmm; member in struct:nvif_chan_args::nvif_chan_v0

Completed in 608 milliseconds

1234