Lines Matching refs:dma_dev

133 	struct dma_device dma_dev;
182 return mdma->dma_dev.dev;
816 list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) {
912 dma_cap_zero(mdma->dma_dev.cap_mask);
913 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask);
914 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask);
915 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask);
916 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask);
946 mdma->dma_dev.dev = &pdev->dev;
947 mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg;
948 mdma->dma_dev.device_prep_dma_cyclic = mdc_prep_dma_cyclic;
949 mdma->dma_dev.device_prep_dma_memcpy = mdc_prep_dma_memcpy;
950 mdma->dma_dev.device_alloc_chan_resources = mdc_alloc_chan_resources;
951 mdma->dma_dev.device_free_chan_resources = mdc_free_chan_resources;
952 mdma->dma_dev.device_tx_status = mdc_tx_status;
953 mdma->dma_dev.device_issue_pending = mdc_issue_pending;
954 mdma->dma_dev.device_terminate_all = mdc_terminate_all;
955 mdma->dma_dev.device_synchronize = mdc_synchronize;
956 mdma->dma_dev.device_config = mdc_slave_config;
958 mdma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
959 mdma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
961 mdma->dma_dev.src_addr_widths |= BIT(i);
962 mdma->dma_dev.dst_addr_widths |= BIT(i);
965 INIT_LIST_HEAD(&mdma->dma_dev.channels);
982 vchan_init(&mchan->vc, &mdma->dma_dev);
998 ret = dma_async_device_register(&mdma->dma_dev);
1012 dma_async_device_unregister(&mdma->dma_dev);
1026 dma_async_device_unregister(&mdma->dma_dev);
1028 list_for_each_entry_safe(mchan, next, &mdma->dma_dev.channels,