Lines Matching refs:vc

197  * @vc: wrapped virtual channel
204 struct virt_dma_chan vc;
302 return container_of(chan, struct owl_dma_vchan, vc.chan);
388 struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
537 struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
538 struct virt_dma_desc *vd = vchan_next_desc(&vchan->vc);
576 dev_dbg(chan2dev(&vchan->vc.chan), "starting pchan %d\n", pchan->id);
652 spin_lock(&vchan->vc.lock);
664 if (vchan_next_desc(&vchan->vc))
670 spin_unlock(&vchan->vc.lock);
704 spin_lock_irqsave(&vchan->vc.lock, flags);
714 vchan_get_all_descriptors(&vchan->vc, &head);
716 spin_unlock_irqrestore(&vchan->vc.lock, flags);
718 vchan_dma_desc_free_list(&vchan->vc, &head);
743 spin_lock_irqsave(&vchan->vc.lock, flags);
747 spin_unlock_irqrestore(&vchan->vc.lock, flags);
760 dev_dbg(chan2dev(chan), "vchan %p: resume\n", &vchan->vc);
762 spin_lock_irqsave(&vchan->vc.lock, flags);
766 spin_unlock_irqrestore(&vchan->vc.lock, flags);
820 spin_lock_irqsave(&vchan->vc.lock, flags);
822 vd = vchan_find_desc(&vchan->vc, cookie);
831 spin_unlock_irqrestore(&vchan->vc.lock, flags);
840 struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
858 spin_lock_irqsave(&vchan->vc.lock, flags);
859 if (vchan_issue_pending(&vchan->vc)) {
863 spin_unlock_irqrestore(&vchan->vc.lock, flags);
908 return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
972 return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
1034 return vchan_tx_prep(&vchan->vc, &txd->vd, flags);
1047 vchan_free_chan_resources(&vchan->vc);
1056 next, &od->dma.channels, vc.chan.device_node) {
1057 list_del(&vchan->vc.chan.device_node);
1058 tasklet_kill(&vchan->vc.task);
1193 vchan->vc.desc_free = owl_dma_desc_free;
1194 vchan_init(&vchan->vc, &od->dma);