Searched refs:vchan (Results 1 - 25 of 34) sorted by relevance

12

/linux-master/drivers/dma/
H A Dsun4i-dma.c138 /* vchan currently being serviced */
139 struct sun4i_dma_vchan *vchan; member in struct:sun4i_dma_pchan
223 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
225 vchan_free_chan_resources(&vchan->vc);
229 struct sun4i_dma_vchan *vchan)
239 if (vchan->is_dedicated) {
250 pchan->vchan = vchan;
267 pchan->vchan = NULL;
322 * Execute pending operations on a vchan
228 find_and_use_pchan(struct sun4i_dma_dev *priv, struct sun4i_dma_vchan *vchan) argument
330 __execute_vchan_pending(struct sun4i_dma_dev *priv, struct sun4i_dma_vchan *vchan) argument
625 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
676 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
796 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
890 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
926 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
937 struct sun4i_dma_vchan *vchan; local
967 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
1015 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local
1034 struct sun4i_dma_vchan *vchan; local
1223 struct sun4i_dma_vchan *vchan = &priv->vchans[i]; local
[all...]
H A Dowl-dma.c187 * @vchan: the virtual channel currently being served by this physical channel
192 struct owl_dma_vchan *vchan; member in struct:owl_dma_pchan
381 static inline int owl_dma_cfg_lli(struct owl_dma_vchan *vchan, argument
388 struct owl_dma *od = to_owl_dma(vchan->vc.chan.device);
401 mode |= OWL_DMA_MODE_TS(vchan->drq)
414 mode |= OWL_DMA_MODE_TS(vchan->drq)
469 struct owl_dma_vchan *vchan)
479 if (!pchan->vchan) {
480 pchan->vchan = vchan;
468 owl_dma_get_pchan(struct owl_dma *od, struct owl_dma_vchan *vchan) argument
535 owl_dma_start_next_txd(struct owl_dma_vchan *vchan) argument
584 owl_dma_phy_free(struct owl_dma *od, struct owl_dma_vchan *vchan) argument
595 struct owl_dma_vchan *vchan; local
700 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
726 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
740 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
754 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
771 owl_dma_getbytes_chan(struct owl_dma_vchan *vchan) argument
808 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
838 owl_dma_phy_alloc_and_start(struct owl_dma_vchan *vchan) argument
855 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
872 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
923 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
988 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
1044 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local
1052 struct owl_dma_vchan *vchan = NULL; local
1066 struct owl_dma_vchan *vchan; local
1191 struct owl_dma_vchan *vchan = &od->vchans[i]; local
[all...]
H A Dsun6i-dma.c180 struct sun6i_vchan *vchan; member in struct:sun6i_pchan
395 static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan, argument
399 dev_dbg(chan2dev(&vchan->vc.chan),
434 static int sun6i_dma_start_desc(struct sun6i_vchan *vchan) argument
436 struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device);
437 struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc);
438 struct sun6i_pchan *pchan = vchan->phy;
455 sun6i_dma_dump_lli(vchan, pchan->desc->v_lli, pchan->desc->p_lli);
460 vchan->irq_type = vchan
480 struct sun6i_vchan *vchan; local
545 struct sun6i_vchan *vchan; local
653 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
706 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
792 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
857 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
867 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
887 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
912 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
956 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
993 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
1020 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local
1034 struct sun6i_vchan *vchan; local
1072 struct sun6i_vchan *vchan = &sdev->vchans[i]; local
1406 struct sun6i_vchan *vchan = &sdc->vchans[i]; local
[all...]
H A Dst_fdma.c25 return container_of(c, struct st_fdma_chan, vchan.chan);
80 vdesc = vchan_next_desc(&fchan->vchan);
86 cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
95 dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
102 int ch_id = fchan->vchan.chan.chan_id;
140 spin_lock(&fchan->vchan.lock);
158 spin_unlock(&fchan->vchan.lock);
283 fchan->vchan.chan.chan_id, fchan->cfg.type);
295 __func__, fchan->vchan.chan.chan_id);
300 spin_lock_irqsave(&fchan->vchan
[all...]
H A Dfsl-edma-common.c47 spin_lock(&fsl_chan->vchan.lock);
51 spin_unlock(&fsl_chan->vchan.lock);
68 spin_unlock(&fsl_chan->vchan.lock);
112 u32 ch = fsl_chan->vchan.chan.chan_id;
146 u32 ch = fsl_chan->vchan.chan.chan_id;
192 u32 ch = fsl_chan->vchan.chan.chan_id;
202 ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux;
245 spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
249 vchan_get_all_descriptors(&fsl_chan->vchan, &head);
250 spin_unlock_irqrestore(&fsl_chan->vchan
[all...]
H A Dls2x-apb-dma.c117 * @vchan: virtual dma channel entry.
124 struct virt_dma_chan vchan; member in struct:ls2x_dma_chan
156 return container_of(chan, struct ls2x_dma_chan, vchan.chan);
186 struct ls2x_dma_priv *priv = to_ldma_priv(lchan->vchan.chan.device);
196 struct ls2x_dma_priv *priv = to_ldma_priv(lchan->vchan.chan.device);
202 vdesc = vchan_next_desc(&lchan->vchan);
367 return vchan_tx_prep(&lchan->vchan, &desc->vdesc, flags);
435 return vchan_tx_prep(&lchan->vchan, &desc->vdesc, flags);
466 spin_lock_irqsave(&lchan->vchan.lock, flags);
467 if (vchan_issue_pending(&lchan->vchan)
[all...]
H A Ddma-jz4780.c130 struct virt_dma_chan vchan; member in struct:jz4780_dma_chan
167 return container_of(chan, struct jz4780_dma_chan, vchan.chan);
179 return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev,
403 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
455 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
483 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags);
494 vdesc = vchan_next_desc(&jzchan->vchan);
566 spin_lock_irqsave(&jzchan->vchan.lock, flags);
568 if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc)
571 spin_unlock_irqrestore(&jzchan->vchan
[all...]
H A Didma64.c107 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device);
114 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device);
118 vdesc = vchan_next_desc(&idma64c->vchan);
143 stat = this_cpu_ptr(idma64c->vchan.chan.local);
145 spin_lock(&idma64c->vchan.lock);
163 spin_unlock(&idma64c->vchan.lock);
334 return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags);
342 spin_lock_irqsave(&idma64c->vchan.lock, flags);
343 if (vchan_issue_pending(&idma64c->vchan) && !idma64c->desc)
345 spin_unlock_irqrestore(&idma64c->vchan
[all...]
H A Dst_fdma.h125 struct virt_dma_chan vchan; member in struct:st_fdma_chan
186 + (fchan)->vchan.chan.chan_id * 0x4 \
191 + (fchan)->vchan.chan.chan_id * 0x4 \
208 + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \
213 + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \
H A Dstm32-dma.c210 struct virt_dma_chan vchan; member in struct:stm32_dma_chan
237 return container_of(chan->vchan.chan.device, struct stm32_dma_device,
243 return container_of(c, struct stm32_dma_chan, vchan.chan);
253 return &chan->vchan.chan.dev->device;
505 spin_lock_irqsave(&chan->vchan.lock, flags);
515 vchan_get_all_descriptors(&chan->vchan, &head);
516 spin_unlock_irqrestore(&chan->vchan.lock, flags);
517 vchan_dma_desc_free_list(&chan->vchan, &head);
526 vchan_synchronize(&chan->vchan);
570 vdesc = vchan_next_desc(&chan->vchan);
[all...]
H A Dstm32-mdma.c237 struct virt_dma_chan vchan; member in struct:stm32_mdma_chan
265 return container_of(chan->vchan.chan.device, struct stm32_mdma_device,
271 return container_of(c, struct stm32_mdma_chan, vchan.chan);
281 return &chan->vchan.chan.dev->device;
830 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
926 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
1112 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
1149 vdesc = vchan_next_desc(&chan->vchan);
1190 dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan);
[all...]
H A Dfsl-qdma.c181 struct virt_dma_chan vchan; member in struct:fsl_qdma_chan
301 return container_of(chan, struct fsl_qdma_chan, vchan.chan);
318 spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
319 vchan_get_all_descriptors(&fsl_chan->vchan, &head);
320 spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
322 vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
729 spin_lock(&fsl_comp->qchan->vchan.lock);
732 spin_unlock(&fsl_comp->qchan->vchan.lock);
985 return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
999 vdesc = vchan_next_desc(&fsl_chan->vchan);
[all...]
H A Ddma-axi-dmac.c143 struct virt_dma_chan vchan; member in struct:axi_dmac_chan
176 return container_of(chan->vchan.chan.device, struct axi_dmac,
182 return container_of(c, struct axi_dmac_chan, vchan.chan);
245 vdesc = vchan_next_desc(&chan->vchan);
470 spin_lock(&dmac->chan.vchan.lock);
481 spin_unlock(&dmac->chan.vchan.lock);
493 spin_lock_irqsave(&chan->vchan.lock, flags);
496 vchan_get_all_descriptors(&chan->vchan, &head);
498 spin_unlock_irqrestore(&chan->vchan.lock, flags);
500 vchan_dma_desc_free_list(&chan->vchan,
[all...]
H A Dpxa_dma.c100 struct pxad_chan *vchan; member in struct:pxad_phy
151 dev_vdbg(&phy->vchan->vc.chan.dev->device, \
159 dev_vdbg(&phy->vchan->vc.chan.dev->device, \
166 dev_vdbg(&phy->vchan->vc.chan.dev->device, \
390 if (!phy->vchan) {
391 phy->vchan = pchan;
425 chan->phy->vchan = NULL;
455 if (!phy->vchan)
458 dev_dbg(&phy->vchan->vc.chan.dev->device,
462 pdev = to_pxad_dev(phy->vchan
[all...]
H A Dmmp_pdma.c118 struct mmp_pdma_chan *vchan; member in struct:mmp_pdma_phy
154 if (!phy->vchan)
157 reg = DRCMR(phy->vchan->drcmr);
161 if (phy->vchan->byte_align)
194 if ((dcsr & DCSR_BUSERR) && (phy->vchan))
195 dev_warn(phy->vchan->dev, "DCSR_BUSERR\n");
207 tasklet_schedule(&phy->vchan->tasklet);
259 if (!phy->vchan) {
260 phy->vchan = pchan;
286 pchan->phy->vchan
[all...]
/linux-master/drivers/dma/hsu/
H A Dhsu.c121 vdesc = vchan_next_desc(&hsuc->vchan);
168 spin_lock_irqsave(&hsuc->vchan.lock, flags);
170 spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
222 stat = this_cpu_ptr(hsuc->vchan.chan.local);
224 spin_lock_irqsave(&hsuc->vchan.lock, flags);
238 spin_unlock_irqrestore(&hsuc->vchan.lock, flags);
295 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags);
303 spin_lock_irqsave(&hsuc->vchan.lock, flags);
304 if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc)
306 spin_unlock_irqrestore(&hsuc->vchan
[all...]
H A Dhsu.h87 struct virt_dma_chan vchan; member in struct:hsu_dma_chan
100 return container_of(chan, struct hsu_dma_chan, vchan.chan);
/linux-master/drivers/dma/sf-pdma/
H A Dsf-pdma.c47 return container_of(dchan, struct sf_pdma_chan, vchan.chan);
103 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
105 spin_lock_irqsave(&chan->vchan.lock, iflags);
107 spin_unlock_irqrestore(&chan->vchan.lock, iflags);
146 spin_lock_irqsave(&chan->vchan.lock, flags);
150 vchan_get_all_descriptors(&chan->vchan, &head);
152 spin_unlock_irqrestore(&chan->vchan.lock, flags);
153 vchan_dma_desc_free_list(&chan->vchan, &head);
166 spin_lock_irqsave(&chan->vchan.lock, flags);
168 list_for_each_entry(vd, &chan->vchan
243 struct virt_dma_chan *vchan = &chan->vchan; local
[all...]
H A Dsf-pdma.h92 struct virt_dma_chan vchan; member in struct:sf_pdma_chan
/linux-master/drivers/dma/lgm/
H A Dlgm-dma.c195 struct virt_dma_chan vchan; member in struct:ldma_chan
292 return container_of(chan, struct ldma_chan, vchan.chan);
521 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
550 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
578 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
596 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
615 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
638 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
662 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device);
696 struct ldma_dev *d = to_ldma_dev(c->vchan
[all...]
/linux-master/drivers/dma/xilinx/
H A Dxilinx_dpdma.c210 * @vchan: virtual DMA channel
218 * @vchan.lock, if both are to be held.
227 struct virt_dma_chan vchan; member in struct:xilinx_dpdma_chan
249 container_of(_chan, struct xilinx_dpdma_chan, vchan.chan)
847 vdesc = vchan_next_desc(&chan->vchan);
1101 spin_lock(&chan->vchan.lock);
1108 spin_unlock(&chan->vchan.lock);
1179 list_empty(&chan->vchan.desc_issued)) {
1182 &chan->vchan.desc_issued);
1216 vchan_tx_prep(&chan->vchan,
[all...]
H A Dxdma.c56 * @vchan: Virtual channel
66 struct virt_dma_chan vchan; member in struct:xdma_chan
210 return container_of(chan, struct xdma_chan, vchan.chan);
327 struct virt_dma_desc *vd = vchan_next_desc(&xchan->vchan);
336 * vchan lock should be held where this function is called.
480 xchan->vchan.desc_free = xdma_free_desc;
481 vchan_init(&xchan->vchan, &xdev->dma_dev);
501 spin_lock_irqsave(&xdma_chan->vchan.lock, flags);
502 if (vchan_issue_pending(&xdma_chan->vchan))
504 spin_unlock_irqrestore(&xdma_chan->vchan
[all...]
/linux-master/drivers/dma/fsl-dpaa2-qdma/
H A Ddpaa2-qdma.c21 return container_of(chan, struct dpaa2_qdma_chan, vchan.chan);
73 spin_lock_irqsave(&dpaa2_chan->vchan.lock, flags);
74 vchan_get_all_descriptors(&dpaa2_chan->vchan, &head);
75 spin_unlock_irqrestore(&dpaa2_chan->vchan.lock, flags);
77 vchan_dma_desc_free_list(&dpaa2_chan->vchan, &head);
269 return vchan_tx_prep(&dpaa2_chan->vchan, &dpaa2_comp->vdesc, flags);
282 spin_lock(&dpaa2_chan->vchan.lock);
283 if (vchan_issue_pending(&dpaa2_chan->vchan)) {
284 vdesc = vchan_next_desc(&dpaa2_chan->vchan);
301 spin_unlock(&dpaa2_chan->vchan
[all...]
H A Ddpaa2-qdma.h71 struct virt_dma_chan vchan; member in struct:dpaa2_qdma_chan
/linux-master/drivers/dma/ti/
H A Dedma.c216 struct virt_dma_chan vchan; member in struct:edma_chan
704 return container_of(c, struct edma_chan, vchan.chan);
723 struct device *dev = echan->vchan.chan.device->dev;
728 vdesc = vchan_next_desc(&echan->vchan);
817 spin_lock_irqsave(&echan->vchan.lock, flags);
834 vchan_get_all_descriptors(&echan->vchan, &head);
835 spin_unlock_irqrestore(&echan->vchan.lock, flags);
836 vchan_dma_desc_free_list(&echan->vchan, &head);
845 vchan_synchronize(&echan->vchan);
1102 return vchan_tx_prep(&echan->vchan,
[all...]

Completed in 299 milliseconds

12