/linux-master/drivers/mtd/hyperbus/ |
H A D | hbmc-am654.c | 29 struct dma_chan *rx_chan; member in struct:am654_hbmc_device_priv 78 struct dma_chan *rx_chan = priv->rx_chan; local 84 if (!priv->rx_chan || !virt_addr_valid(to) || object_is_on_stack(to)) 87 dma_dst = dma_map_single(rx_chan->device->dev, to, len, DMA_FROM_DEVICE); 88 if (dma_mapping_error(rx_chan->device->dev, dma_dst)) { 94 tx = dmaengine_prep_dma_memcpy(rx_chan, dma_dst, dma_src, len, flags); 112 dma_async_issue_pending(rx_chan); 114 dmaengine_terminate_sync(rx_chan); 120 dma_unmap_single(rx_chan 140 struct dma_chan *rx_chan; local [all...] |
/linux-master/drivers/soc/xilinx/ |
H A D | zynqmp_power.c | 34 static struct mbox_chan *rx_chan; variable in typeref:struct:mbox_chan 116 ret = mbox_send_message(rx_chan, NULL); 241 rx_chan = mbox_request_channel_byname(client, "rx"); 242 if (IS_ERR(rx_chan)) { 244 return PTR_ERR(rx_chan); 286 if (!rx_chan) 287 mbox_free_channel(rx_chan);
|
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | bcm63xx_enet.c | 268 enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); 270 enet_dmac_writel(priv, 1, ENETDMAC_BUFALLOC, priv->rx_chan); 416 ENETDMAC_CHANCFG, priv->rx_chan); 497 ENETDMAC_IR, priv->rx_chan); 519 ENETDMAC_IRMASK, priv->rx_chan); 564 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->rx_chan); 782 val |= ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); 784 val &= ~ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); 935 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->rx_chan); 1011 ENETDMA_BUFALLOC_REG(priv->rx_chan)); [all...] |
H A D | bcm63xx_enet.h | 216 int rx_chan; member in struct:bcm_enet_priv
|
/linux-master/arch/mips/include/asm/mach-bcm63xx/ |
H A D | bcm63xx_dev_enet.h | 60 int rx_chan; member in struct:bcm63xx_enet_platform_data
|
/linux-master/drivers/net/ethernet/allwinner/ |
H A D | sun4i-emac.c | 90 struct dma_chan *rx_chan; member in struct:emac_board_info 290 desc = dmaengine_prep_slave_single(db->rx_chan, rxbuf, count, 316 dma_async_issue_pending(db->rx_chan); 734 if (rxlen >= dev->mtu && db->rx_chan) { 930 db->rx_chan = dma_request_chan(&pdev->dev, "rx"); 931 if (IS_ERR(db->rx_chan)) { 934 err = PTR_ERR(db->rx_chan); 946 err = dmaengine_slave_config(db->rx_chan, &conf); 956 dma_release_channel(db->rx_chan); 959 db->rx_chan [all...] |
/linux-master/drivers/usb/renesas_usbhs/ |
H A D | fifo.h | 26 struct dma_chan *rx_chan; member in struct:usbhs_fifo
|
H A D | fifo.c | 774 return fifo->rx_chan; 1252 if (fifo->rx_chan) 1253 dma_release_channel(fifo->rx_chan); 1256 fifo->rx_chan = NULL; 1270 fifo->rx_chan = dma_request_channel(mask, usbhsf_dma_filter, 1290 fifo->rx_chan = dma_request_chan(dev, name); 1291 if (IS_ERR(fifo->rx_chan)) 1292 fifo->rx_chan = NULL; 1306 if (fifo->tx_chan || fifo->rx_chan) 1310 fifo->rx_chan [all...] |
/linux-master/drivers/spi/ |
H A D | spi-ti-qspi.c | 57 struct dma_chan *rx_chan; member in struct:ti_qspi 444 struct dma_chan *chan = qspi->rx_chan; 632 if (qspi->rx_chan) { 745 if (qspi->rx_chan) 746 dma_release_channel(qspi->rx_chan); 861 qspi->rx_chan = dma_request_chan_by_mask(&mask); 862 if (IS_ERR(qspi->rx_chan)) { 865 qspi->rx_chan = NULL; 876 dma_release_channel(qspi->rx_chan); 879 host->dma_rx = qspi->rx_chan; [all...] |
H A D | spi-cadence-quadspi.c | 82 struct dma_chan *rx_chan; member in struct:cqspi_st 1311 if (!cqspi->rx_chan || !virt_addr_valid(buf)) { 1316 ddev = cqspi->rx_chan->device->dev; 1322 tx = dmaengine_prep_dma_memcpy(cqspi->rx_chan, dma_dst, dma_src, 1342 dma_async_issue_pending(cqspi->rx_chan); 1345 dmaengine_terminate_sync(cqspi->rx_chan); 1591 cqspi->rx_chan = dma_request_chan_by_mask(&mask); 1592 if (IS_ERR(cqspi->rx_chan)) { 1593 int ret = PTR_ERR(cqspi->rx_chan); 1595 cqspi->rx_chan [all...] |
/linux-master/drivers/tty/serial/ |
H A D | samsung_tty.c | 108 struct dma_chan *rx_chan; member in struct:s3c24xx_uart_dma 519 dma_sync_single_for_cpu(dma->rx_chan->device->dev, dma->rx_addr, 561 if (dma && dma->rx_chan) { 563 dma_status = dmaengine_tx_status(dma->rx_chan, 568 dmaengine_terminate_all(dma->rx_chan); 617 dmaengine_tx_status(dma->rx_chan, dma->rx_cookie, &state); 640 dma_sync_single_for_device(dma->rx_chan->device->dev, dma->rx_addr, 643 dma->rx_desc = dmaengine_prep_slave_single(dma->rx_chan, 656 dma_async_issue_pending(dma->rx_chan); 735 dmaengine_pause(dma->rx_chan); [all...] |
/linux-master/arch/mips/bcm63xx/ |
H A D | dev-enet.c | 277 dpd->rx_chan = 0; 280 dpd->rx_chan = 2;
|
/linux-master/drivers/remoteproc/ |
H A D | xlnx_r5_remoteproc.c | 64 * @rx_chan: mailbox rx channel 73 struct mbox_chan *rx_chan; member in struct:mbox_info 205 if (mbox_send_message(ipi->rx_chan, NULL) < 0) 206 dev_err(cl->dev, "ack failed to mbox rx_chan\n"); 245 ipi->rx_chan = mbox_request_channel_byname(mbox_cl, "rx"); 246 if (IS_ERR(ipi->rx_chan)) { 248 ipi->rx_chan = NULL; 270 if (ipi->rx_chan) { 271 mbox_free_channel(ipi->rx_chan); 272 ipi->rx_chan [all...] |
/linux-master/drivers/mmc/host/ |
H A D | au1xmmc.c | 112 u32 rx_chan; member in struct:au1xmmc_host 151 (((h)->flags & HOST_F_XMIT) ? (h)->tx_chan : (h)->rx_chan) 885 host->rx_chan = au1xxx_dbdma_chan_alloc(rxid, memid, 887 if (!host->rx_chan) { 894 au1xxx_dbdma_set_devwidth(host->rx_chan, 8); 897 au1xxx_dbdma_ring_alloc(host->rx_chan, AU1XMMC_DESCRIPTOR_COUNT); 910 au1xxx_dbdma_chan_free(host->rx_chan);
|
H A D | omap_hsmmc.c | 196 struct dma_chan *rx_chan; member in struct:omap_hsmmc_host 813 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; 1894 host->rx_chan = dma_request_chan(&pdev->dev, "rx"); 1895 if (IS_ERR(host->rx_chan)) { 1897 ret = PTR_ERR(host->rx_chan); 1916 dma_get_max_seg_size(host->rx_chan->device->dev), 1970 if (!IS_ERR_OR_NULL(host->rx_chan)) 1971 dma_release_channel(host->rx_chan); 1990 dma_release_channel(host->rx_chan);
|
H A D | sdhci.c | 1224 host->rx_chan = dma_request_chan(mmc_dev(mmc), "rx"); 1225 if (IS_ERR(host->rx_chan)) { 1231 ret = PTR_ERR(host->rx_chan); 1234 host->rx_chan = NULL; 1243 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; 1308 if (host->rx_chan) { 1309 dma_release_channel(host->rx_chan); 1310 host->rx_chan = NULL;
|
H A D | sdhci.h | 578 struct dma_chan *rx_chan; member in struct:sdhci_host
|
/linux-master/drivers/net/ethernet/xilinx/ |
H A D | xilinx_axienet_main.c | 1071 dma_async_issue_pending(lp->rx_chan); 1341 dma_rx_desc = dmaengine_prep_slave_sg(lp->rx_chan, skbuf_dma->sgl, 1383 lp->rx_chan = dma_request_chan(lp->dev, "rx_chan0"); 1384 if (IS_ERR(lp->rx_chan)) { 1385 ret = PTR_ERR(lp->rx_chan); 1426 dma_async_issue_pending(lp->rx_chan); 1439 dma_release_channel(lp->rx_chan); 1601 dmaengine_terminate_sync(lp->rx_chan); 1602 dmaengine_synchronize(lp->rx_chan); 1611 dma_release_channel(lp->rx_chan); [all...] |
H A D | xilinx_axienet.h | 457 * @rx_chan: RX DMA channel. 530 struct dma_chan *rx_chan; member in struct:axienet_local
|
/linux-master/drivers/dma/qcom/ |
H A D | gpi.c | 2070 struct gchan *tx_chan, *rx_chan; local 2079 rx_chan = &gpi_dev->gpiis[gpii].gchan[GPI_RX_CHAN]; 2081 if (rx_chan->vc.chan.client_count && rx_chan->seid == seid) 2093 rx_chan = &gpi_dev->gpiis[gpii].gchan[GPI_RX_CHAN]; 2097 rx_chan->vc.chan.client_count)
|
/linux-master/drivers/mtd/nand/raw/ |
H A D | qcom_nandc.c | 412 struct dma_chan *rx_chan; member in struct:qcom_nand_controller::__anon27::__anon28 1316 ret = prepare_bam_async_desc(nandc, nandc->rx_chan, 0); 1349 dma_async_issue_pending(nandc->rx_chan); 3030 if (nandc->rx_chan) 3031 dma_release_channel(nandc->rx_chan); 3093 nandc->rx_chan = dma_request_chan(nandc->dev, "rx"); 3094 if (IS_ERR(nandc->rx_chan)) { 3095 ret = PTR_ERR(nandc->rx_chan); 3096 nandc->rx_chan = NULL;
|
/linux-master/drivers/scsi/cxgbi/ |
H A D | libcxgbi.h | 123 unsigned int rx_chan; member in struct:cxgbi_sock
|
/linux-master/drivers/ntb/ |
H A D | ntb_transport.c | 2408 struct dma_chan *rx_chan, *tx_chan; local 2413 rx_chan = qp->rx_dma_chan; 2416 copy_align = max(rx_chan ? rx_chan->device->copy_align : 0,
|
/linux-master/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls.h | 304 u32 rx_chan; member in struct:chtls_sock
|
/linux-master/drivers/net/ethernet/ti/ |
H A D | am65-cpsw-nuss.c | 2796 struct am65_cpsw_rx_chn *rx_chan = &common->rx_chns; local 2821 k3_udma_glue_reset_rx_chn(rx_chan->rx_chn, i, rx_chan, 2824 k3_udma_glue_disable_rx_chn(rx_chan->rx_chn);
|