Lines Matching refs:chan

44 nouveau_channel_kill(struct nouveau_channel *chan)
46 atomic_set(&chan->killed, 1);
47 if (chan->fence)
48 nouveau_fence_context_kill(chan->fence, -ENODEV);
54 struct nouveau_channel *chan = container_of(event, typeof(*chan), kill);
55 struct nouveau_cli *cli = (void *)chan->user.client;
57 NV_PRINTK(warn, cli, "channel %d killed!\n", chan->chid);
59 if (unlikely(!atomic_read(&chan->killed)))
60 nouveau_channel_kill(chan);
66 nouveau_channel_idle(struct nouveau_channel *chan)
68 if (likely(chan && chan->fence && !atomic_read(&chan->killed))) {
69 struct nouveau_cli *cli = (void *)chan->user.client;
73 ret = nouveau_fence_new(&fence, chan);
81 chan->chid, nvxx_client(&cli->base)->name);
91 struct nouveau_channel *chan = *pchan;
92 if (chan) {
93 struct nouveau_cli *cli = (void *)chan->user.client;
95 if (chan->fence)
96 nouveau_fence(chan->drm)->context_del(chan);
99 nouveau_svmm_part(chan->vmm->svmm, chan->inst);
101 nvif_object_dtor(&chan->blit);
102 nvif_object_dtor(&chan->nvsw);
103 nvif_object_dtor(&chan->gart);
104 nvif_object_dtor(&chan->vram);
105 nvif_event_dtor(&chan->kill);
106 nvif_object_dtor(&chan->user);
107 nvif_mem_dtor(&chan->mem_userd);
108 nvif_object_dtor(&chan->push.ctxdma);
109 nouveau_vma_del(&chan->push.vma);
110 nouveau_bo_unmap(chan->push.buffer);
111 if (chan->push.buffer && chan->push.buffer->bo.pin_count)
112 nouveau_bo_unpin(chan->push.buffer);
113 nouveau_bo_ref(NULL, &chan->push.buffer);
114 kfree(chan);
122 struct nouveau_channel *chan = container_of(push, typeof(*chan), chan._push);
123 chan->dma.cur = chan->dma.cur + (chan->chan._push.cur - chan->chan._push.bgn);
124 FIRE_RING(chan);
125 chan->chan._push.bgn = chan->chan._push.cur;
131 struct nouveau_channel *chan = container_of(push, typeof(*chan), chan._push);
133 chan->dma.cur = chan->dma.cur + (chan->chan._push.cur - chan->chan._push.bgn);
134 ret = RING_SPACE(chan, size);
136 chan->chan._push.bgn = chan->chan._push.mem.object.map.ptr;
137 chan->chan._push.bgn = chan->chan._push.bgn + chan->dma.cur;
138 chan->chan._push.cur = chan->chan._push.bgn;
139 chan->chan._push.end = chan->chan._push.bgn + size;
150 struct nouveau_channel *chan;
154 chan = *pchan = kzalloc(sizeof(*chan), GFP_KERNEL);
155 if (!chan)
158 chan->device = device;
159 chan->drm = drm;
160 chan->vmm = nouveau_cli_vmm(cli);
161 atomic_set(&chan->killed, 0);
169 &chan->push.buffer);
171 ret = nouveau_bo_pin(chan->push.buffer, target, false);
173 ret = nouveau_bo_map(chan->push.buffer);
181 chan->chan._push.mem.object.parent = cli->base.object.parent;
182 chan->chan._push.mem.object.client = &cli->base;
183 chan->chan._push.mem.object.name = "chanPush";
184 chan->chan._push.mem.object.map.ptr = chan->push.buffer->kmap.virtual;
185 chan->chan._push.wait = nouveau_channel_wait;
186 chan->chan._push.kick = nouveau_channel_kick;
187 chan->chan.push = &chan->chan._push;
193 chan->push.addr = chan->push.buffer->offset;
196 ret = nouveau_vma_new(chan->push.buffer, chan->vmm,
197 &chan->push.vma);
203 chan->push.addr = chan->push.vma->addr;
211 args.limit = chan->vmm->vmm.limit - 1;
213 if (chan->push.buffer->bo.resource->mem_type == TTM_PL_VRAM) {
231 if (chan->drm->agp.bridge) {
234 args.start = chan->drm->agp.base;
235 args.limit = chan->drm->agp.base +
236 chan->drm->agp.size - 1;
241 args.limit = chan->vmm->vmm.limit - 1;
247 &chan->push.ctxdma);
279 struct nvif_chan_v0 chan;
283 struct nouveau_channel *chan;
301 ret = nouveau_channel_prep(drm, device, size, &chan);
302 *pchan = chan;
307 args.chan.version = 0;
308 args.chan.namelen = sizeof(args.name);
309 args.chan.runlist = __ffs64(runm);
310 args.chan.runq = 0;
311 args.chan.priv = priv;
312 args.chan.devm = BIT(0);
314 args.chan.vmm = 0;
315 args.chan.ctxdma = nvif_handle(&chan->push.ctxdma);
316 args.chan.offset = chan->push.addr;
317 args.chan.length = 0;
319 args.chan.vmm = nvif_handle(&chan->vmm->vmm.object);
321 args.chan.ctxdma = nvif_handle(&chan->push.ctxdma);
323 args.chan.ctxdma = 0;
324 args.chan.offset = ioffset + chan->push.addr;
325 args.chan.length = ilength;
327 args.chan.huserd = 0;
328 args.chan.ouserd = 0;
334 0, PAGE_SIZE, NULL, 0, &chan->mem_userd);
338 args.chan.huserd = nvif_handle(&chan->mem_userd.object);
339 args.chan.ouserd = 0;
341 chan->userd = &chan->mem_userd.object;
343 chan->userd = &chan->user;
350 &args, sizeof(args), &chan->user);
356 chan->runlist = args.chan.runlist;
357 chan->chid = args.chan.chid;
358 chan->inst = args.chan.inst;
359 chan->token = args.chan.token;
364 nouveau_channel_init(struct nouveau_channel *chan, u32 vram, u32 gart)
366 struct nvif_device *device = chan->device;
367 struct nouveau_drm *drm = chan->drm;
371 ret = nvif_object_map(chan->userd, NULL, 0);
375 if (chan->user.oclass >= FERMI_CHANNEL_GPFIFO) {
384 ret = nvif_event_ctor(&chan->user, "abi16ChanKilled", chan->chid,
386 &args.base, sizeof(args), &chan->kill);
388 ret = nvif_event_allow(&chan->kill);
402 args.limit = chan->vmm->vmm.limit - 1;
410 ret = nvif_object_ctor(&chan->user, "abi16ChanVramCtxDma", vram,
412 &chan->vram);
420 args.limit = chan->vmm->vmm.limit - 1;
422 if (chan->drm->agp.bridge) {
425 args.start = chan->drm->agp.base;
426 args.limit = chan->drm->agp.base +
427 chan->drm->agp.size - 1;
432 args.limit = chan->vmm->vmm.limit - 1;
435 ret = nvif_object_ctor(&chan->user, "abi16ChanGartCtxDma", gart,
437 &chan->gart);
443 switch (chan->user.oclass) {
448 chan->user_put = 0x40;
449 chan->user_get = 0x44;
450 chan->dma.max = (0x10000 / 4) - 2;
453 chan->user_put = 0x40;
454 chan->user_get = 0x44;
455 chan->user_get_hi = 0x60;
456 chan->dma.ib_base = 0x10000 / 4;
457 chan->dma.ib_max = NV50_DMA_IB_MAX;
458 chan->dma.ib_put = 0;
459 chan->dma.ib_free = chan->dma.ib_max - chan->dma.ib_put;
460 chan->dma.max = chan->dma.ib_base;
464 chan->dma.put = 0;
465 chan->dma.cur = chan->dma.put;
466 chan->dma.free = chan->dma.max - chan->dma.cur;
468 ret = PUSH_WAIT(chan->chan.push, NOUVEAU_DMA_SKIPS);
473 PUSH_DATA(chan->chan.push, 0x00000000);
477 ret = nvif_object_ctor(&chan->user, "abi16NvswFence", 0x006e,
479 NULL, 0, &chan->nvsw);
483 ret = PUSH_WAIT(chan->chan.push, 2);
487 PUSH_NVSQ(chan->chan.push, NV_SW, 0x0000, chan->nvsw.handle);
488 PUSH_KICK(chan->chan.push);
492 return nouveau_fence(chan->drm)->context_new(chan);