Searched refs:dma_dev (Results 51 - 75 of 177) sorted by relevance

12345678

/linux-master/drivers/net/wireless/broadcom/b43/
H A Dbus.c104 dev->dma_dev = core->dma_dev;
203 dev->dma_dev = sdev->dma_dev;
H A Dbus.h40 struct device *dma_dev; member in struct:b43_bus_dev
/linux-master/drivers/dma/dw/
H A Dacpi.c13 .dma_dev = dma_spec->dev,
/linux-master/include/linux/
H A Dvirtio_ring.h94 struct device *dma_dev);
/linux-master/drivers/gpu/drm/mediatek/
H A Dmtk_drm_gem.c76 mtk_gem->cookie = dma_alloc_attrs(priv->dma_dev, obj->size,
108 dma_free_attrs(priv->dma_dev, obj->size, mtk_gem->cookie,
177 ret = dma_mmap_attrs(priv->dma_dev, vma, mtk_gem->cookie,
200 ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie,
H A Dmtk_drm_drv.h53 struct device *dma_dev; member in struct:mtk_drm_private
/linux-master/drivers/dma/sh/
H A Dshdma.h59 struct sh_dmae_device, shdma_dev.dma_dev)
/linux-master/arch/s390/include/asm/
H A Dcio.h366 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
368 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
371 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev);
372 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
/linux-master/sound/soc/
H A Dsoc-generic-dmaengine-pcm.c103 struct device *dma_dev = dmaengine_dma_dev(pcm, substream); local
128 hw.period_bytes_max = dma_get_max_seg_size(dma_dev);
357 !(config->dma_dev && config->dma_dev->of_node)))
360 if (config->dma_dev) {
368 dev_name(config->dma_dev));
369 dev = config->dma_dev;
/linux-master/drivers/thunderbolt/
H A Ddma_test.c235 struct device *dma_dev = tb_ring_dma_device(dt->rx_ring); local
237 dma_unmap_single(dma_dev, tf->frame.buffer_phy, DMA_TEST_FRAME_SIZE,
263 struct device *dma_dev = tb_ring_dma_device(dt->rx_ring); local
280 dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE,
282 if (dma_mapping_error(dma_dev, dma_addr)) {
304 struct device *dma_dev = tb_ring_dma_device(dt->tx_ring); local
306 dma_unmap_single(dma_dev, tf->frame.buffer_phy, DMA_TEST_FRAME_SIZE,
314 struct device *dma_dev = tb_ring_dma_device(dt->tx_ring); local
334 dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE,
336 if (dma_mapping_error(dma_dev, dma_add
[all...]
/linux-master/drivers/spi/
H A Dspi-dw-dma.c30 if (s->dma_dev != chan->device->dev)
115 struct pci_dev *dma_dev; local
123 dma_dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x0827, NULL);
124 if (!dma_dev)
131 rx->dma_dev = &dma_dev->dev;
137 tx->dma_dev = &dma_dev->dev;
153 pci_dev_put(dma_dev);
164 pci_dev_put(dma_dev);
[all...]
/linux-master/drivers/dma/
H A Dfsl-qdma.c214 struct dma_device dma_dev; member in struct:fsl_qdma_engine
723 dev_err(fsl_qdma->dma_dev.dev,
755 dev_err(fsl_qdma->dma_dev.dev,
773 dev_err(fsl_qdma->dma_dev.dev,
791 dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n");
880 dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
1205 INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
1214 vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
1217 dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
1219 fsl_qdma->dma_dev
[all...]
H A Dtegra20-apb-dma.c211 struct dma_device dma_dev; member in struct:tegra_dma
1357 chan = dma_get_any_slave_channel(&tdma->dma_dev);
1481 INIT_LIST_HEAD(&tdma->dma_dev.channels);
1506 tdc->dma_chan.device = &tdma->dma_dev;
1509 &tdma->dma_dev.channels);
1524 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask);
1525 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask);
1526 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask);
1529 tdma->dma_dev.dev = &pdev->dev;
1530 tdma->dma_dev
[all...]
/linux-master/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_rx.c58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE,
112 if (unlikely(dma_mapping_error(q->dma_dev, rb->dma_addr))) {
130 dma_unmap_page(q->dma_dev, rb->dma_addr, PAGE_SIZE,
269 dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
319 dma_sync_single_for_cpu(q->dma_dev,
636 q->dma_dev = &fp->pdev->dev;
638 q->rqes = fun_alloc_ring_mem(q->dma_dev, nrq
[all...]
/linux-master/drivers/net/ethernet/broadcom/
H A Db44.c151 dma_sync_single_for_device(sdev->dma_dev, dma_base + offset,
160 dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset,
606 dma_unmap_single(bp->sdev->dma_dev,
651 mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
657 if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
660 if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
661 dma_unmap_single(bp->sdev->dma_dev, mapping,
667 mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
670 if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
672 if (!dma_mapping_error(bp->sdev->dma_dev, mappin
[all...]
/linux-master/drivers/virtio/
H A Dvirtio_ring.c214 struct device *dma_dev; member in struct:vring_virtqueue
234 struct device *dma_dev);
313 struct device *dma_dev)
316 return dma_alloc_coherent(dma_dev, size,
347 struct device *dma_dev)
350 dma_free_coherent(dma_dev, size, queue, dma_handle);
362 return vq->dma_dev;
1079 struct virtio_device *vdev, struct device *dma_dev)
1084 dma_dev);
1095 struct device *dma_dev)
311 vring_alloc_queue(struct virtio_device *vdev, size_t size, dma_addr_t *dma_handle, gfp_t flag, struct device *dma_dev) argument
345 vring_free_queue(struct virtio_device *vdev, size_t size, void *queue, dma_addr_t dma_handle, struct device *dma_dev) argument
1078 vring_free_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, struct device *dma_dev) argument
1090 vring_alloc_queue_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, u32 num, unsigned int vring_align, bool may_reduce_num, struct device *dma_dev) argument
1141 vring_create_virtqueue_split( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
1909 vring_free_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, struct device *dma_dev) argument
1935 vring_alloc_queue_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, u32 num, struct device *dma_dev) argument
2053 vring_create_virtqueue_packed( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
2602 __vring_new_virtqueue(unsigned int index, struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, bool weak_barriers, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
2689 vring_create_virtqueue_dma( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
[all...]
/linux-master/drivers/soc/ti/
H A Dk3-ringacc.c155 * @dma_dev: device to be used for DMA API (allocation, mapping)
177 struct device *dma_dev; member in struct:k3_ring
615 dma_free_coherent(ring->dma_dev,
620 ring->dma_dev = NULL;
726 ring->dma_dev = cfg->dma_dev;
727 if (!ring->dma_dev) {
728 dev_warn(ringacc->dev, "dma_dev is not provided for ring%d\n",
730 ring->dma_dev = ringacc->dev;
737 ring->ring_mem_virt = dma_alloc_coherent(ring->dma_dev,
[all...]
/linux-master/drivers/dma/fsl-dpaa2-qdma/
H A Ddpaa2-qdma.c641 INIT_LIST_HEAD(&dpaa2_qdma->dma_dev.channels);
647 vchan_init(&dpaa2_chan->vchan, &dpaa2_qdma->dma_dev);
729 dma_cap_set(DMA_PRIVATE, dpaa2_qdma->dma_dev.cap_mask);
730 dma_cap_set(DMA_SLAVE, dpaa2_qdma->dma_dev.cap_mask);
731 dma_cap_set(DMA_MEMCPY, dpaa2_qdma->dma_dev.cap_mask);
733 dpaa2_qdma->dma_dev.dev = dev;
734 dpaa2_qdma->dma_dev.device_alloc_chan_resources =
736 dpaa2_qdma->dma_dev.device_free_chan_resources =
738 dpaa2_qdma->dma_dev.device_tx_status = dma_cookie_status;
739 dpaa2_qdma->dma_dev
[all...]
/linux-master/drivers/dma/ioat/
H A Dsysfs.c77 struct dma_device *dma = &ioat_dma->dma_dev;
98 struct dma_device *dma = &ioat_dma->dma_dev;
/linux-master/drivers/net/ethernet/sfc/falcon/
H A Dtx.c57 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local
60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
63 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
210 struct device *dma_dev = &efx->pci_dev->dev; local
221 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
226 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
255 dma_addr = skb_frag_dma_map(dma_dev, fragment,
261 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
/linux-master/drivers/gpu/drm/exynos/
H A Dexynos_drm_drv.h201 struct device *dma_dev; member in struct:exynos_drm_private
214 return priv->dma_dev;
/linux-master/drivers/net/ethernet/ti/
H A Dam65-cpsw-nuss.h72 struct device *dma_dev; member in struct:am65_cpsw_tx_chn
89 struct device *dma_dev; member in struct:am65_cpsw_rx_chn
/linux-master/include/linux/soc/ti/
H A Dk3-ringacc.h70 * @dma_dev: Master device which is using and accessing to the ring
82 struct device *dma_dev; member in struct:k3_ring_cfg
/linux-master/include/linux/iio/
H A Dbuffer-dma.h141 struct device *dma_dev, const struct iio_dma_buffer_ops *ops);
/linux-master/drivers/dma/sf-pdma/
H A Dsf-pdma.h113 struct dma_device dma_dev; member in struct:sf_pdma

Completed in 324 milliseconds

12345678