/linux-master/drivers/net/wireless/broadcom/b43/ |
H A D | bus.c | 104 dev->dma_dev = core->dma_dev; 203 dev->dma_dev = sdev->dma_dev;
|
H A D | bus.h | 40 struct device *dma_dev; member in struct:b43_bus_dev
|
/linux-master/drivers/dma/dw/ |
H A D | acpi.c | 13 .dma_dev = dma_spec->dev,
|
/linux-master/include/linux/ |
H A D | virtio_ring.h | 94 struct device *dma_dev);
|
/linux-master/drivers/gpu/drm/mediatek/ |
H A D | mtk_drm_gem.c | 76 mtk_gem->cookie = dma_alloc_attrs(priv->dma_dev, obj->size, 108 dma_free_attrs(priv->dma_dev, obj->size, mtk_gem->cookie, 177 ret = dma_mmap_attrs(priv->dma_dev, vma, mtk_gem->cookie, 200 ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie,
|
H A D | mtk_drm_drv.h | 53 struct device *dma_dev; member in struct:mtk_drm_private
|
/linux-master/drivers/dma/sh/ |
H A D | shdma.h | 59 struct sh_dmae_device, shdma_dev.dma_dev)
|
/linux-master/arch/s390/include/asm/ |
H A D | cio.h | 366 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, 368 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, 371 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev); 372 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
|
/linux-master/sound/soc/ |
H A D | soc-generic-dmaengine-pcm.c | 103 struct device *dma_dev = dmaengine_dma_dev(pcm, substream); local 128 hw.period_bytes_max = dma_get_max_seg_size(dma_dev); 357 !(config->dma_dev && config->dma_dev->of_node))) 360 if (config->dma_dev) { 368 dev_name(config->dma_dev)); 369 dev = config->dma_dev;
|
/linux-master/drivers/thunderbolt/ |
H A D | dma_test.c | 235 struct device *dma_dev = tb_ring_dma_device(dt->rx_ring); local 237 dma_unmap_single(dma_dev, tf->frame.buffer_phy, DMA_TEST_FRAME_SIZE, 263 struct device *dma_dev = tb_ring_dma_device(dt->rx_ring); local 280 dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE, 282 if (dma_mapping_error(dma_dev, dma_addr)) { 304 struct device *dma_dev = tb_ring_dma_device(dt->tx_ring); local 306 dma_unmap_single(dma_dev, tf->frame.buffer_phy, DMA_TEST_FRAME_SIZE, 314 struct device *dma_dev = tb_ring_dma_device(dt->tx_ring); local 334 dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE, 336 if (dma_mapping_error(dma_dev, dma_add [all...] |
/linux-master/drivers/spi/ |
H A D | spi-dw-dma.c | 30 if (s->dma_dev != chan->device->dev) 115 struct pci_dev *dma_dev; local 123 dma_dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x0827, NULL); 124 if (!dma_dev) 131 rx->dma_dev = &dma_dev->dev; 137 tx->dma_dev = &dma_dev->dev; 153 pci_dev_put(dma_dev); 164 pci_dev_put(dma_dev); [all...] |
/linux-master/drivers/dma/ |
H A D | fsl-qdma.c | 214 struct dma_device dma_dev; member in struct:fsl_qdma_engine 723 dev_err(fsl_qdma->dma_dev.dev, 755 dev_err(fsl_qdma->dma_dev.dev, 773 dev_err(fsl_qdma->dma_dev.dev, 791 dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n"); 880 dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!"); 1205 INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels); 1214 vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev); 1217 dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask); 1219 fsl_qdma->dma_dev [all...] |
H A D | tegra20-apb-dma.c | 211 struct dma_device dma_dev; member in struct:tegra_dma 1357 chan = dma_get_any_slave_channel(&tdma->dma_dev); 1481 INIT_LIST_HEAD(&tdma->dma_dev.channels); 1506 tdc->dma_chan.device = &tdma->dma_dev; 1509 &tdma->dma_dev.channels); 1524 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); 1525 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); 1526 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); 1529 tdma->dma_dev.dev = &pdev->dev; 1530 tdma->dma_dev [all...] |
/linux-master/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_rx.c | 58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, 77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, 88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, 110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE, 112 if (unlikely(dma_mapping_error(q->dma_dev, rb->dma_addr))) { 130 dma_unmap_page(q->dma_dev, rb->dma_addr, PAGE_SIZE, 269 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, 319 dma_sync_single_for_cpu(q->dma_dev, 636 q->dma_dev = &fp->pdev->dev; 638 q->rqes = fun_alloc_ring_mem(q->dma_dev, nrq [all...] |
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | b44.c | 151 dma_sync_single_for_device(sdev->dma_dev, dma_base + offset, 160 dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset, 606 dma_unmap_single(bp->sdev->dma_dev, 651 mapping = dma_map_single(bp->sdev->dma_dev, skb->data, 657 if (dma_mapping_error(bp->sdev->dma_dev, mapping) || 660 if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) 661 dma_unmap_single(bp->sdev->dma_dev, mapping, 667 mapping = dma_map_single(bp->sdev->dma_dev, skb->data, 670 if (dma_mapping_error(bp->sdev->dma_dev, mapping) || 672 if (!dma_mapping_error(bp->sdev->dma_dev, mappin [all...] |
/linux-master/drivers/virtio/ |
H A D | virtio_ring.c | 214 struct device *dma_dev; member in struct:vring_virtqueue 234 struct device *dma_dev); 313 struct device *dma_dev) 316 return dma_alloc_coherent(dma_dev, size, 347 struct device *dma_dev) 350 dma_free_coherent(dma_dev, size, queue, dma_handle); 362 return vq->dma_dev; 1079 struct virtio_device *vdev, struct device *dma_dev) 1084 dma_dev); 1095 struct device *dma_dev) 311 vring_alloc_queue(struct virtio_device *vdev, size_t size, dma_addr_t *dma_handle, gfp_t flag, struct device *dma_dev) argument 345 vring_free_queue(struct virtio_device *vdev, size_t size, void *queue, dma_addr_t dma_handle, struct device *dma_dev) argument 1078 vring_free_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, struct device *dma_dev) argument 1090 vring_alloc_queue_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, u32 num, unsigned int vring_align, bool may_reduce_num, struct device *dma_dev) argument 1141 vring_create_virtqueue_split( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 1909 vring_free_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, struct device *dma_dev) argument 1935 vring_alloc_queue_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, u32 num, struct device *dma_dev) argument 2053 vring_create_virtqueue_packed( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 2602 __vring_new_virtqueue(unsigned int index, struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, bool weak_barriers, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 2689 vring_create_virtqueue_dma( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument [all...] |
/linux-master/drivers/soc/ti/ |
H A D | k3-ringacc.c | 155 * @dma_dev: device to be used for DMA API (allocation, mapping) 177 struct device *dma_dev; member in struct:k3_ring 615 dma_free_coherent(ring->dma_dev, 620 ring->dma_dev = NULL; 726 ring->dma_dev = cfg->dma_dev; 727 if (!ring->dma_dev) { 728 dev_warn(ringacc->dev, "dma_dev is not provided for ring%d\n", 730 ring->dma_dev = ringacc->dev; 737 ring->ring_mem_virt = dma_alloc_coherent(ring->dma_dev, [all...] |
/linux-master/drivers/dma/fsl-dpaa2-qdma/ |
H A D | dpaa2-qdma.c | 641 INIT_LIST_HEAD(&dpaa2_qdma->dma_dev.channels); 647 vchan_init(&dpaa2_chan->vchan, &dpaa2_qdma->dma_dev); 729 dma_cap_set(DMA_PRIVATE, dpaa2_qdma->dma_dev.cap_mask); 730 dma_cap_set(DMA_SLAVE, dpaa2_qdma->dma_dev.cap_mask); 731 dma_cap_set(DMA_MEMCPY, dpaa2_qdma->dma_dev.cap_mask); 733 dpaa2_qdma->dma_dev.dev = dev; 734 dpaa2_qdma->dma_dev.device_alloc_chan_resources = 736 dpaa2_qdma->dma_dev.device_free_chan_resources = 738 dpaa2_qdma->dma_dev.device_tx_status = dma_cookie_status; 739 dpaa2_qdma->dma_dev [all...] |
/linux-master/drivers/dma/ioat/ |
H A D | sysfs.c | 77 struct dma_device *dma = &ioat_dma->dma_dev; 98 struct dma_device *dma = &ioat_dma->dma_dev;
|
/linux-master/drivers/net/ethernet/sfc/falcon/ |
H A D | tx.c | 57 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local 60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, 63 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, 210 struct device *dma_dev = &efx->pci_dev->dev; local 221 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); 226 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) 255 dma_addr = skb_frag_dma_map(dma_dev, fragment, 261 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
|
/linux-master/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_drv.h | 201 struct device *dma_dev; member in struct:exynos_drm_private 214 return priv->dma_dev;
|
/linux-master/drivers/net/ethernet/ti/ |
H A D | am65-cpsw-nuss.h | 72 struct device *dma_dev; member in struct:am65_cpsw_tx_chn 89 struct device *dma_dev; member in struct:am65_cpsw_rx_chn
|
/linux-master/include/linux/soc/ti/ |
H A D | k3-ringacc.h | 70 * @dma_dev: Master device which is using and accessing to the ring 82 struct device *dma_dev; member in struct:k3_ring_cfg
|
/linux-master/include/linux/iio/ |
H A D | buffer-dma.h | 141 struct device *dma_dev, const struct iio_dma_buffer_ops *ops);
|
/linux-master/drivers/dma/sf-pdma/ |
H A D | sf-pdma.h | 113 struct dma_device dma_dev; member in struct:sf_pdma
|