/linux-master/drivers/dma/ |
H A D | sun4i-dma.c | 138 /* vchan currently being serviced */ 139 struct sun4i_dma_vchan *vchan; member in struct:sun4i_dma_pchan 223 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 225 vchan_free_chan_resources(&vchan->vc); 229 struct sun4i_dma_vchan *vchan) 239 if (vchan->is_dedicated) { 250 pchan->vchan = vchan; 267 pchan->vchan = NULL; 322 * Execute pending operations on a vchan 228 find_and_use_pchan(struct sun4i_dma_dev *priv, struct sun4i_dma_vchan *vchan) argument 330 __execute_vchan_pending(struct sun4i_dma_dev *priv, struct sun4i_dma_vchan *vchan) argument 625 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 676 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 796 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 890 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 926 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 937 struct sun4i_dma_vchan *vchan; local 967 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 1015 struct sun4i_dma_vchan *vchan = to_sun4i_dma_vchan(chan); local 1034 struct sun4i_dma_vchan *vchan; local 1223 struct sun4i_dma_vchan *vchan = &priv->vchans[i]; local [all...] |
H A D | owl-dma.c | 187 * @vchan: the virtual channel currently being served by this physical channel 192 struct owl_dma_vchan *vchan; member in struct:owl_dma_pchan 381 static inline int owl_dma_cfg_lli(struct owl_dma_vchan *vchan, argument 388 struct owl_dma *od = to_owl_dma(vchan->vc.chan.device); 401 mode |= OWL_DMA_MODE_TS(vchan->drq) 414 mode |= OWL_DMA_MODE_TS(vchan->drq) 469 struct owl_dma_vchan *vchan) 479 if (!pchan->vchan) { 480 pchan->vchan = vchan; 468 owl_dma_get_pchan(struct owl_dma *od, struct owl_dma_vchan *vchan) argument 535 owl_dma_start_next_txd(struct owl_dma_vchan *vchan) argument 584 owl_dma_phy_free(struct owl_dma *od, struct owl_dma_vchan *vchan) argument 595 struct owl_dma_vchan *vchan; local 700 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 726 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 740 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 754 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 771 owl_dma_getbytes_chan(struct owl_dma_vchan *vchan) argument 808 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 838 owl_dma_phy_alloc_and_start(struct owl_dma_vchan *vchan) argument 855 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 872 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 923 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 988 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 1044 struct owl_dma_vchan *vchan = to_owl_vchan(chan); local 1052 struct owl_dma_vchan *vchan = NULL; local 1066 struct owl_dma_vchan *vchan; local 1191 struct owl_dma_vchan *vchan = &od->vchans[i]; local [all...] |
H A D | sun6i-dma.c | 180 struct sun6i_vchan *vchan; member in struct:sun6i_pchan 395 static inline void sun6i_dma_dump_lli(struct sun6i_vchan *vchan, argument 399 dev_dbg(chan2dev(&vchan->vc.chan), 434 static int sun6i_dma_start_desc(struct sun6i_vchan *vchan) argument 436 struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vchan->vc.chan.device); 437 struct virt_dma_desc *desc = vchan_next_desc(&vchan->vc); 438 struct sun6i_pchan *pchan = vchan->phy; 455 sun6i_dma_dump_lli(vchan, pchan->desc->v_lli, pchan->desc->p_lli); 460 vchan->irq_type = vchan 480 struct sun6i_vchan *vchan; local 545 struct sun6i_vchan *vchan; local 653 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 706 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 792 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 857 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 867 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 887 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 912 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 956 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 993 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 1020 struct sun6i_vchan *vchan = to_sun6i_vchan(chan); local 1034 struct sun6i_vchan *vchan; local 1072 struct sun6i_vchan *vchan = &sdev->vchans[i]; local 1406 struct sun6i_vchan *vchan = &sdc->vchans[i]; local [all...] |
H A D | st_fdma.c | 25 return container_of(c, struct st_fdma_chan, vchan.chan); 80 vdesc = vchan_next_desc(&fchan->vchan); 86 cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id); 95 dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id); 102 int ch_id = fchan->vchan.chan.chan_id; 140 spin_lock(&fchan->vchan.lock); 158 spin_unlock(&fchan->vchan.lock); 283 fchan->vchan.chan.chan_id, fchan->cfg.type); 295 __func__, fchan->vchan.chan.chan_id); 300 spin_lock_irqsave(&fchan->vchan [all...] |
H A D | fsl-edma-common.c | 47 spin_lock(&fsl_chan->vchan.lock); 51 spin_unlock(&fsl_chan->vchan.lock); 68 spin_unlock(&fsl_chan->vchan.lock); 112 u32 ch = fsl_chan->vchan.chan.chan_id; 146 u32 ch = fsl_chan->vchan.chan.chan_id; 192 u32 ch = fsl_chan->vchan.chan.chan_id; 202 ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux; 245 spin_lock_irqsave(&fsl_chan->vchan.lock, flags); 249 vchan_get_all_descriptors(&fsl_chan->vchan, &head); 250 spin_unlock_irqrestore(&fsl_chan->vchan [all...] |
H A D | ls2x-apb-dma.c | 117 * @vchan: virtual dma channel entry. 124 struct virt_dma_chan vchan; member in struct:ls2x_dma_chan 156 return container_of(chan, struct ls2x_dma_chan, vchan.chan); 186 struct ls2x_dma_priv *priv = to_ldma_priv(lchan->vchan.chan.device); 196 struct ls2x_dma_priv *priv = to_ldma_priv(lchan->vchan.chan.device); 202 vdesc = vchan_next_desc(&lchan->vchan); 367 return vchan_tx_prep(&lchan->vchan, &desc->vdesc, flags); 435 return vchan_tx_prep(&lchan->vchan, &desc->vdesc, flags); 466 spin_lock_irqsave(&lchan->vchan.lock, flags); 467 if (vchan_issue_pending(&lchan->vchan) [all...] |
H A D | dma-jz4780.c | 130 struct virt_dma_chan vchan; member in struct:jz4780_dma_chan 167 return container_of(chan, struct jz4780_dma_chan, vchan.chan); 179 return container_of(jzchan->vchan.chan.device, struct jz4780_dma_dev, 403 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); 455 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); 483 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); 494 vdesc = vchan_next_desc(&jzchan->vchan); 566 spin_lock_irqsave(&jzchan->vchan.lock, flags); 568 if (vchan_issue_pending(&jzchan->vchan) && !jzchan->desc) 571 spin_unlock_irqrestore(&jzchan->vchan [all...] |
H A D | idma64.c | 107 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); 114 struct idma64 *idma64 = to_idma64(idma64c->vchan.chan.device); 118 vdesc = vchan_next_desc(&idma64c->vchan); 143 stat = this_cpu_ptr(idma64c->vchan.chan.local); 145 spin_lock(&idma64c->vchan.lock); 163 spin_unlock(&idma64c->vchan.lock); 334 return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); 342 spin_lock_irqsave(&idma64c->vchan.lock, flags); 343 if (vchan_issue_pending(&idma64c->vchan) && !idma64c->desc) 345 spin_unlock_irqrestore(&idma64c->vchan [all...] |
H A D | st_fdma.h | 125 struct virt_dma_chan vchan; member in struct:st_fdma_chan 186 + (fchan)->vchan.chan.chan_id * 0x4 \ 191 + (fchan)->vchan.chan.chan_id * 0x4 \ 208 + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \ 213 + (fchan)->vchan.chan.chan_id * FDMA_NODE_SZ \
|
H A D | stm32-dma.c | 210 struct virt_dma_chan vchan; member in struct:stm32_dma_chan 237 return container_of(chan->vchan.chan.device, struct stm32_dma_device, 243 return container_of(c, struct stm32_dma_chan, vchan.chan); 253 return &chan->vchan.chan.dev->device; 505 spin_lock_irqsave(&chan->vchan.lock, flags); 515 vchan_get_all_descriptors(&chan->vchan, &head); 516 spin_unlock_irqrestore(&chan->vchan.lock, flags); 517 vchan_dma_desc_free_list(&chan->vchan, &head); 526 vchan_synchronize(&chan->vchan); 570 vdesc = vchan_next_desc(&chan->vchan); [all...] |
H A D | stm32-mdma.c | 237 struct virt_dma_chan vchan; member in struct:stm32_mdma_chan 265 return container_of(chan->vchan.chan.device, struct stm32_mdma_device, 271 return container_of(c, struct stm32_mdma_chan, vchan.chan); 281 return &chan->vchan.chan.dev->device; 830 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); 926 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); 1112 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); 1149 vdesc = vchan_next_desc(&chan->vchan); 1190 dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan); [all...] |
H A D | fsl-qdma.c | 181 struct virt_dma_chan vchan; member in struct:fsl_qdma_chan 301 return container_of(chan, struct fsl_qdma_chan, vchan.chan); 318 spin_lock_irqsave(&fsl_chan->vchan.lock, flags); 319 vchan_get_all_descriptors(&fsl_chan->vchan, &head); 320 spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags); 322 vchan_dma_desc_free_list(&fsl_chan->vchan, &head); 729 spin_lock(&fsl_comp->qchan->vchan.lock); 732 spin_unlock(&fsl_comp->qchan->vchan.lock); 985 return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags); 999 vdesc = vchan_next_desc(&fsl_chan->vchan); [all...] |
H A D | dma-axi-dmac.c | 143 struct virt_dma_chan vchan; member in struct:axi_dmac_chan 176 return container_of(chan->vchan.chan.device, struct axi_dmac, 182 return container_of(c, struct axi_dmac_chan, vchan.chan); 245 vdesc = vchan_next_desc(&chan->vchan); 470 spin_lock(&dmac->chan.vchan.lock); 481 spin_unlock(&dmac->chan.vchan.lock); 493 spin_lock_irqsave(&chan->vchan.lock, flags); 496 vchan_get_all_descriptors(&chan->vchan, &head); 498 spin_unlock_irqrestore(&chan->vchan.lock, flags); 500 vchan_dma_desc_free_list(&chan->vchan, [all...] |
H A D | pxa_dma.c | 100 struct pxad_chan *vchan; member in struct:pxad_phy 151 dev_vdbg(&phy->vchan->vc.chan.dev->device, \ 159 dev_vdbg(&phy->vchan->vc.chan.dev->device, \ 166 dev_vdbg(&phy->vchan->vc.chan.dev->device, \ 390 if (!phy->vchan) { 391 phy->vchan = pchan; 425 chan->phy->vchan = NULL; 455 if (!phy->vchan) 458 dev_dbg(&phy->vchan->vc.chan.dev->device, 462 pdev = to_pxad_dev(phy->vchan [all...] |
H A D | mmp_pdma.c | 118 struct mmp_pdma_chan *vchan; member in struct:mmp_pdma_phy 154 if (!phy->vchan) 157 reg = DRCMR(phy->vchan->drcmr); 161 if (phy->vchan->byte_align) 194 if ((dcsr & DCSR_BUSERR) && (phy->vchan)) 195 dev_warn(phy->vchan->dev, "DCSR_BUSERR\n"); 207 tasklet_schedule(&phy->vchan->tasklet); 259 if (!phy->vchan) { 260 phy->vchan = pchan; 286 pchan->phy->vchan [all...] |
/linux-master/drivers/dma/hsu/ |
H A D | hsu.c | 121 vdesc = vchan_next_desc(&hsuc->vchan); 168 spin_lock_irqsave(&hsuc->vchan.lock, flags); 170 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); 222 stat = this_cpu_ptr(hsuc->vchan.chan.local); 224 spin_lock_irqsave(&hsuc->vchan.lock, flags); 238 spin_unlock_irqrestore(&hsuc->vchan.lock, flags); 295 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); 303 spin_lock_irqsave(&hsuc->vchan.lock, flags); 304 if (vchan_issue_pending(&hsuc->vchan) && !hsuc->desc) 306 spin_unlock_irqrestore(&hsuc->vchan [all...] |
H A D | hsu.h | 87 struct virt_dma_chan vchan; member in struct:hsu_dma_chan 100 return container_of(chan, struct hsu_dma_chan, vchan.chan);
|
/linux-master/drivers/dma/sf-pdma/ |
H A D | sf-pdma.c | 47 return container_of(dchan, struct sf_pdma_chan, vchan.chan); 103 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); 105 spin_lock_irqsave(&chan->vchan.lock, iflags); 107 spin_unlock_irqrestore(&chan->vchan.lock, iflags); 146 spin_lock_irqsave(&chan->vchan.lock, flags); 150 vchan_get_all_descriptors(&chan->vchan, &head); 152 spin_unlock_irqrestore(&chan->vchan.lock, flags); 153 vchan_dma_desc_free_list(&chan->vchan, &head); 166 spin_lock_irqsave(&chan->vchan.lock, flags); 168 list_for_each_entry(vd, &chan->vchan 243 struct virt_dma_chan *vchan = &chan->vchan; local [all...] |
H A D | sf-pdma.h | 92 struct virt_dma_chan vchan; member in struct:sf_pdma_chan
|
/linux-master/drivers/dma/lgm/ |
H A D | lgm-dma.c | 195 struct virt_dma_chan vchan; member in struct:ldma_chan 292 return container_of(chan, struct ldma_chan, vchan.chan); 521 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 550 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 578 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 596 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 615 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 638 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 662 struct ldma_dev *d = to_ldma_dev(c->vchan.chan.device); 696 struct ldma_dev *d = to_ldma_dev(c->vchan [all...] |
/linux-master/drivers/dma/xilinx/ |
H A D | xilinx_dpdma.c | 210 * @vchan: virtual DMA channel 218 * @vchan.lock, if both are to be held. 227 struct virt_dma_chan vchan; member in struct:xilinx_dpdma_chan 249 container_of(_chan, struct xilinx_dpdma_chan, vchan.chan) 847 vdesc = vchan_next_desc(&chan->vchan); 1101 spin_lock(&chan->vchan.lock); 1108 spin_unlock(&chan->vchan.lock); 1179 list_empty(&chan->vchan.desc_issued)) { 1182 &chan->vchan.desc_issued); 1216 vchan_tx_prep(&chan->vchan, [all...] |
H A D | xdma.c | 56 * @vchan: Virtual channel 66 struct virt_dma_chan vchan; member in struct:xdma_chan 210 return container_of(chan, struct xdma_chan, vchan.chan); 327 struct virt_dma_desc *vd = vchan_next_desc(&xchan->vchan); 336 * vchan lock should be held where this function is called. 480 xchan->vchan.desc_free = xdma_free_desc; 481 vchan_init(&xchan->vchan, &xdev->dma_dev); 501 spin_lock_irqsave(&xdma_chan->vchan.lock, flags); 502 if (vchan_issue_pending(&xdma_chan->vchan)) 504 spin_unlock_irqrestore(&xdma_chan->vchan [all...] |
/linux-master/drivers/dma/fsl-dpaa2-qdma/ |
H A D | dpaa2-qdma.c | 21 return container_of(chan, struct dpaa2_qdma_chan, vchan.chan); 73 spin_lock_irqsave(&dpaa2_chan->vchan.lock, flags); 74 vchan_get_all_descriptors(&dpaa2_chan->vchan, &head); 75 spin_unlock_irqrestore(&dpaa2_chan->vchan.lock, flags); 77 vchan_dma_desc_free_list(&dpaa2_chan->vchan, &head); 269 return vchan_tx_prep(&dpaa2_chan->vchan, &dpaa2_comp->vdesc, flags); 282 spin_lock(&dpaa2_chan->vchan.lock); 283 if (vchan_issue_pending(&dpaa2_chan->vchan)) { 284 vdesc = vchan_next_desc(&dpaa2_chan->vchan); 301 spin_unlock(&dpaa2_chan->vchan [all...] |
H A D | dpaa2-qdma.h | 71 struct virt_dma_chan vchan; member in struct:dpaa2_qdma_chan
|
/linux-master/drivers/dma/ti/ |
H A D | edma.c | 216 struct virt_dma_chan vchan; member in struct:edma_chan 704 return container_of(c, struct edma_chan, vchan.chan); 723 struct device *dev = echan->vchan.chan.device->dev; 728 vdesc = vchan_next_desc(&echan->vchan); 817 spin_lock_irqsave(&echan->vchan.lock, flags); 834 vchan_get_all_descriptors(&echan->vchan, &head); 835 spin_unlock_irqrestore(&echan->vchan.lock, flags); 836 vchan_dma_desc_free_list(&echan->vchan, &head); 845 vchan_synchronize(&echan->vchan); 1102 return vchan_tx_prep(&echan->vchan, [all...] |