Searched refs:dma_dev (Results 1 - 25 of 177) sorted by last modified time

12345678

/linux-master/drivers/scsi/
H A Dscsi_lib.c1969 struct device *dev = shost->dma_dev;
2074 tag_set->numa_node = dev_to_node(shost->dma_dev);
H A Dhosts.c210 * @dma_dev: dma device for the host
220 struct device *dma_dev)
244 if (!dma_dev)
245 dma_dev = shost->shost_gendev.parent;
247 shost->dma_dev = dma_dev;
249 if (dma_dev->dma_mask) {
251 dma_max_mapping_size(dma_dev) >> SECTOR_SHIFT);
219 scsi_add_host_with_dma(struct Scsi_Host *shost, struct device *dev, struct device *dma_dev) argument
H A Dscsi_debug.c387 dev_to_sdebug_host(shost->dma_dev)
/linux-master/drivers/net/ethernet/ti/
H A Dam65-cpsw-nuss.c325 buf_dma = dma_map_single(rx_chn->dma_dev, skb->data, pkt_len,
327 if (unlikely(dma_mapping_error(rx_chn->dma_dev, buf_dma))) {
387 dma_unmap_single(rx_chn->dma_dev, buf_dma, buf_dma_len, DMA_FROM_DEVICE);
406 dma_unmap_single(tx_chn->dma_dev, buf_dma, buf_dma_len, DMA_TO_DEVICE);
416 dma_unmap_page(tx_chn->dma_dev, buf_dma, buf_dma_len,
850 dma_unmap_single(rx_chn->dma_dev, buf_dma, buf_dma_len, DMA_FROM_DEVICE);
1170 buf_dma = dma_map_single(tx_chn->dma_dev, skb->data, pkt_len,
1172 if (unlikely(dma_mapping_error(tx_chn->dma_dev, buf_dma))) {
1181 dma_unmap_single(tx_chn->dma_dev, buf_dma, pkt_len,
1228 buf_dma = skb_frag_dma_map(tx_chn->dma_dev, fra
[all...]
/linux-master/drivers/net/ethernet/mediatek/
H A Dmtk_wed.c1151 hw->eth->dma_dev != hw->eth->dev)
2420 if (hw->eth->dma_dev == hw->eth->dev &&
H A Dmtk_eth_soc.c1141 eth->scratch_ring = dma_alloc_coherent(eth->dma_dev,
1152 dma_addr = dma_map_single(eth->dma_dev,
1155 if (unlikely(dma_mapping_error(eth->dma_dev, dma_addr)))
1219 dma_unmap_single(eth->dma_dev,
1224 dma_unmap_page(eth->dma_dev,
1231 dma_unmap_page(eth->dma_dev,
1238 dma_unmap_page(eth->dma_dev,
1422 txd_info.addr = dma_map_single(eth->dma_dev, skb->data, txd_info.size,
1424 if (unlikely(dma_mapping_error(eth->dma_dev, txd_info.addr)))
1464 txd_info.addr = skb_frag_dma_map(eth->dma_dev, fra
4667 mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev) argument
[all...]
/linux-master/drivers/acpi/
H A Dscan.c1524 struct device *dma_dev = dev; local
1533 adev = ACPI_COMPANION(dma_dev);
1537 dma_dev = dma_dev->parent;
1538 } while (dma_dev);
1540 if (!dma_dev)
1563 dev_dbg(dma_dev, "Invalid DMA regions configuration\n");
/linux-master/include/scsi/
H A Dscsi_host.h711 struct device *dma_dev; member in struct:Scsi_Host
/linux-master/drivers/tty/serial/
H A Dpch_uart.c658 if ((chan->chan_id == param->chan_id) && (param->dma_dev ==
671 struct pci_dev *dma_dev; local
679 dma_dev = pci_get_slot(priv->pdev->bus,
684 param->dma_dev = &dma_dev->dev;
692 pci_dev_put(dma_dev);
699 param->dma_dev = &dma_dev->dev;
709 pci_dev_put(dma_dev);
718 pci_dev_put(dma_dev);
[all...]
H A Damba-pl011.c589 struct dma_device *dma_dev = chan->device; local
632 dmatx->dma = dma_map_single(dma_dev->dev, dmatx->buf, count,
643 dma_unmap_single(dma_dev->dev, dmatx->dma, dmatx->len, DMA_TO_DEVICE);
661 dma_dev->device_issue_pending(chan);
/linux-master/drivers/mcb/
H A Dmcb-core.c233 dev->dma_dev = bus->carrier;
/linux-master/drivers/s390/cio/
H A Dcss.c1068 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages) argument
1079 cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr,
1099 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev) argument
1104 gen_pool_for_each_chunk(gp_dma, __gp_dma_free_dma, dma_dev);
1117 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, argument
1129 addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP);
1140 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, argument
1143 return __cio_gp_dma_zalloc(gp_dma, dma_dev, size, NULL);
/linux-master/arch/s390/include/asm/
H A Dcio.h366 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
368 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev,
371 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev);
372 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
/linux-master/include/linux/
H A Dvdpa.h73 * @dma_dev: the actual device that is performing DMA
90 struct device *dma_dev; member in struct:vdpa_device
525 return vdev->dma_dev;
/linux-master/drivers/virtio/
H A Dvirtio_ring.c214 struct device *dma_dev; member in struct:vring_virtqueue
234 struct device *dma_dev);
313 struct device *dma_dev)
316 return dma_alloc_coherent(dma_dev, size,
347 struct device *dma_dev)
350 dma_free_coherent(dma_dev, size, queue, dma_handle);
362 return vq->dma_dev;
1079 struct virtio_device *vdev, struct device *dma_dev)
1084 dma_dev);
1095 struct device *dma_dev)
311 vring_alloc_queue(struct virtio_device *vdev, size_t size, dma_addr_t *dma_handle, gfp_t flag, struct device *dma_dev) argument
345 vring_free_queue(struct virtio_device *vdev, size_t size, void *queue, dma_addr_t dma_handle, struct device *dma_dev) argument
1078 vring_free_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, struct device *dma_dev) argument
1090 vring_alloc_queue_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, u32 num, unsigned int vring_align, bool may_reduce_num, struct device *dma_dev) argument
1141 vring_create_virtqueue_split( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
1909 vring_free_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, struct device *dma_dev) argument
1935 vring_alloc_queue_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, u32 num, struct device *dma_dev) argument
2053 vring_create_virtqueue_packed( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
2602 __vring_new_virtqueue(unsigned int index, struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, bool weak_barriers, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
2689 vring_create_virtqueue_dma( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument
[all...]
H A Dvirtio_vdpa.c150 struct device *dma_dev; local
205 dma_dev = ops->get_vq_dma_dev(vdpa, index);
207 dma_dev = vdpa_get_dma_dev(vdpa);
210 notify, callback, name, dma_dev);
/linux-master/drivers/vhost/
H A Dvdpa.c1314 struct device *dma_dev = vdpa_get_dma_dev(vdpa); local
1322 bus = dma_dev->bus;
1326 if (!device_iommu_capable(dma_dev, IOMMU_CAP_CACHE_COHERENCY)) {
1336 ret = iommu_attach_device(v->domain, dma_dev);
1351 struct device *dma_dev = vdpa_get_dma_dev(vdpa); local
1354 iommu_detach_device(v->domain, dma_dev);
/linux-master/drivers/vdpa/vdpa_user/
H A Dvduse_dev.c2009 vdev->vdpa.dma_dev = &vdev->vdpa.dev;
/linux-master/drivers/vdpa/virtio_pci/
H A Dvp_vdpa.c501 vp_vdpa->vdpa.dma_dev = &pdev->dev;
/linux-master/drivers/vdpa/vdpa_sim/
H A Dvdpa_sim.c275 vdpasim->vdpa.dma_dev = dev;
/linux-master/drivers/vdpa/pds/
H A Dvdpa_dev.c618 struct device *dma_dev; local
645 dma_dev = &pdev->dev;
646 pdsv->vdpa_dev.dma_dev = dma_dev;
/linux-master/drivers/vdpa/mlx5/net/
H A Dmlx5_vnet.c3170 return mvdev->vdev.dma_dev;
3725 mvdev->vdev.dma_dev = &mdev->pdev->dev;
/linux-master/drivers/vdpa/ifcvf/
H A Difcvf_main.c716 adapter->vdpa.dma_dev = &pdev->dev;
/linux-master/drivers/vdpa/alibaba/
H A Deni_vdpa.c499 eni_vdpa->vdpa.dma_dev = &pdev->dev;
/linux-master/drivers/scsi/ibmvscsi_tgt/
H A Dibmvscsi_tgt.c261 vio_disable_interrupts(vscsi->dma_dev);
818 rc = vio_enable_interrupts(vscsi->dma_dev);
890 vio_enable_interrupts(vscsi->dma_dev);
1231 vio_enable_interrupts(vscsi->dma_dev);
1517 info = dma_alloc_coherent(&vscsi->dma_dev->dev, sizeof(*info), &token,
1596 dma_free_coherent(&vscsi->dma_dev->dev, sizeof(*info), info, token);
1636 cap = dma_alloc_coherent(&vscsi->dma_dev->dev, olen, &token,
1648 strscpy_pad(cap->name, dev_name(&vscsi->dma_dev->dev),
1712 dma_free_coherent(&vscsi->dma_dev->dev, olen, cap, token);
1939 rc = h_send_crq(vscsi->dma_dev
[all...]

Completed in 915 milliseconds

12345678