/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 1969 struct device *dev = shost->dma_dev; 2074 tag_set->numa_node = dev_to_node(shost->dma_dev);
|
H A D | hosts.c | 210 * @dma_dev: dma device for the host 220 struct device *dma_dev) 244 if (!dma_dev) 245 dma_dev = shost->shost_gendev.parent; 247 shost->dma_dev = dma_dev; 249 if (dma_dev->dma_mask) { 251 dma_max_mapping_size(dma_dev) >> SECTOR_SHIFT); 219 scsi_add_host_with_dma(struct Scsi_Host *shost, struct device *dev, struct device *dma_dev) argument
|
H A D | scsi_debug.c | 387 dev_to_sdebug_host(shost->dma_dev)
|
/linux-master/drivers/net/ethernet/ti/ |
H A D | am65-cpsw-nuss.c | 325 buf_dma = dma_map_single(rx_chn->dma_dev, skb->data, pkt_len, 327 if (unlikely(dma_mapping_error(rx_chn->dma_dev, buf_dma))) { 387 dma_unmap_single(rx_chn->dma_dev, buf_dma, buf_dma_len, DMA_FROM_DEVICE); 406 dma_unmap_single(tx_chn->dma_dev, buf_dma, buf_dma_len, DMA_TO_DEVICE); 416 dma_unmap_page(tx_chn->dma_dev, buf_dma, buf_dma_len, 850 dma_unmap_single(rx_chn->dma_dev, buf_dma, buf_dma_len, DMA_FROM_DEVICE); 1170 buf_dma = dma_map_single(tx_chn->dma_dev, skb->data, pkt_len, 1172 if (unlikely(dma_mapping_error(tx_chn->dma_dev, buf_dma))) { 1181 dma_unmap_single(tx_chn->dma_dev, buf_dma, pkt_len, 1228 buf_dma = skb_frag_dma_map(tx_chn->dma_dev, fra [all...] |
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_wed.c | 1151 hw->eth->dma_dev != hw->eth->dev) 2420 if (hw->eth->dma_dev == hw->eth->dev &&
|
H A D | mtk_eth_soc.c | 1141 eth->scratch_ring = dma_alloc_coherent(eth->dma_dev, 1152 dma_addr = dma_map_single(eth->dma_dev, 1155 if (unlikely(dma_mapping_error(eth->dma_dev, dma_addr))) 1219 dma_unmap_single(eth->dma_dev, 1224 dma_unmap_page(eth->dma_dev, 1231 dma_unmap_page(eth->dma_dev, 1238 dma_unmap_page(eth->dma_dev, 1422 txd_info.addr = dma_map_single(eth->dma_dev, skb->data, txd_info.size, 1424 if (unlikely(dma_mapping_error(eth->dma_dev, txd_info.addr))) 1464 txd_info.addr = skb_frag_dma_map(eth->dma_dev, fra 4667 mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev) argument [all...] |
/linux-master/drivers/acpi/ |
H A D | scan.c | 1524 struct device *dma_dev = dev; local 1533 adev = ACPI_COMPANION(dma_dev); 1537 dma_dev = dma_dev->parent; 1538 } while (dma_dev); 1540 if (!dma_dev) 1563 dev_dbg(dma_dev, "Invalid DMA regions configuration\n");
|
/linux-master/include/scsi/ |
H A D | scsi_host.h | 711 struct device *dma_dev; member in struct:Scsi_Host
|
/linux-master/drivers/tty/serial/ |
H A D | pch_uart.c | 658 if ((chan->chan_id == param->chan_id) && (param->dma_dev == 671 struct pci_dev *dma_dev; local 679 dma_dev = pci_get_slot(priv->pdev->bus, 684 param->dma_dev = &dma_dev->dev; 692 pci_dev_put(dma_dev); 699 param->dma_dev = &dma_dev->dev; 709 pci_dev_put(dma_dev); 718 pci_dev_put(dma_dev); [all...] |
H A D | amba-pl011.c | 589 struct dma_device *dma_dev = chan->device; local 632 dmatx->dma = dma_map_single(dma_dev->dev, dmatx->buf, count, 643 dma_unmap_single(dma_dev->dev, dmatx->dma, dmatx->len, DMA_TO_DEVICE); 661 dma_dev->device_issue_pending(chan);
|
/linux-master/drivers/mcb/ |
H A D | mcb-core.c | 233 dev->dma_dev = bus->carrier;
|
/linux-master/drivers/s390/cio/ |
H A D | css.c | 1068 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages) argument 1079 cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr, 1099 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev) argument 1104 gen_pool_for_each_chunk(gp_dma, __gp_dma_free_dma, dma_dev); 1117 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, argument 1129 addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP); 1140 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, argument 1143 return __cio_gp_dma_zalloc(gp_dma, dma_dev, size, NULL);
|
/linux-master/arch/s390/include/asm/ |
H A D | cio.h | 366 void *cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, 368 void *__cio_gp_dma_zalloc(struct gen_pool *gp_dma, struct device *dma_dev, 371 void cio_gp_dma_destroy(struct gen_pool *gp_dma, struct device *dma_dev); 372 struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
|
/linux-master/include/linux/ |
H A D | vdpa.h | 73 * @dma_dev: the actual device that is performing DMA 90 struct device *dma_dev; member in struct:vdpa_device 525 return vdev->dma_dev;
|
/linux-master/drivers/virtio/ |
H A D | virtio_ring.c | 214 struct device *dma_dev; member in struct:vring_virtqueue 234 struct device *dma_dev); 313 struct device *dma_dev) 316 return dma_alloc_coherent(dma_dev, size, 347 struct device *dma_dev) 350 dma_free_coherent(dma_dev, size, queue, dma_handle); 362 return vq->dma_dev; 1079 struct virtio_device *vdev, struct device *dma_dev) 1084 dma_dev); 1095 struct device *dma_dev) 311 vring_alloc_queue(struct virtio_device *vdev, size_t size, dma_addr_t *dma_handle, gfp_t flag, struct device *dma_dev) argument 345 vring_free_queue(struct virtio_device *vdev, size_t size, void *queue, dma_addr_t dma_handle, struct device *dma_dev) argument 1078 vring_free_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, struct device *dma_dev) argument 1090 vring_alloc_queue_split(struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, u32 num, unsigned int vring_align, bool may_reduce_num, struct device *dma_dev) argument 1141 vring_create_virtqueue_split( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 1909 vring_free_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, struct device *dma_dev) argument 1935 vring_alloc_queue_packed(struct vring_virtqueue_packed *vring_packed, struct virtio_device *vdev, u32 num, struct device *dma_dev) argument 2053 vring_create_virtqueue_packed( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 2602 __vring_new_virtqueue(unsigned int index, struct vring_virtqueue_split *vring_split, struct virtio_device *vdev, bool weak_barriers, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument 2689 vring_create_virtqueue_dma( unsigned int index, unsigned int num, unsigned int vring_align, struct virtio_device *vdev, bool weak_barriers, bool may_reduce_num, bool context, bool (*notify)(struct virtqueue *), void (*callback)(struct virtqueue *), const char *name, struct device *dma_dev) argument [all...] |
H A D | virtio_vdpa.c | 150 struct device *dma_dev; local 205 dma_dev = ops->get_vq_dma_dev(vdpa, index); 207 dma_dev = vdpa_get_dma_dev(vdpa); 210 notify, callback, name, dma_dev);
|
/linux-master/drivers/vhost/ |
H A D | vdpa.c | 1314 struct device *dma_dev = vdpa_get_dma_dev(vdpa); local 1322 bus = dma_dev->bus; 1326 if (!device_iommu_capable(dma_dev, IOMMU_CAP_CACHE_COHERENCY)) { 1336 ret = iommu_attach_device(v->domain, dma_dev); 1351 struct device *dma_dev = vdpa_get_dma_dev(vdpa); local 1354 iommu_detach_device(v->domain, dma_dev);
|
/linux-master/drivers/vdpa/vdpa_user/ |
H A D | vduse_dev.c | 2009 vdev->vdpa.dma_dev = &vdev->vdpa.dev;
|
/linux-master/drivers/vdpa/virtio_pci/ |
H A D | vp_vdpa.c | 501 vp_vdpa->vdpa.dma_dev = &pdev->dev;
|
/linux-master/drivers/vdpa/vdpa_sim/ |
H A D | vdpa_sim.c | 275 vdpasim->vdpa.dma_dev = dev;
|
/linux-master/drivers/vdpa/pds/ |
H A D | vdpa_dev.c | 618 struct device *dma_dev; local 645 dma_dev = &pdev->dev; 646 pdsv->vdpa_dev.dma_dev = dma_dev;
|
/linux-master/drivers/vdpa/mlx5/net/ |
H A D | mlx5_vnet.c | 3170 return mvdev->vdev.dma_dev; 3725 mvdev->vdev.dma_dev = &mdev->pdev->dev;
|
/linux-master/drivers/vdpa/ifcvf/ |
H A D | ifcvf_main.c | 716 adapter->vdpa.dma_dev = &pdev->dev;
|
/linux-master/drivers/vdpa/alibaba/ |
H A D | eni_vdpa.c | 499 eni_vdpa->vdpa.dma_dev = &pdev->dev;
|
/linux-master/drivers/scsi/ibmvscsi_tgt/ |
H A D | ibmvscsi_tgt.c | 261 vio_disable_interrupts(vscsi->dma_dev); 818 rc = vio_enable_interrupts(vscsi->dma_dev); 890 vio_enable_interrupts(vscsi->dma_dev); 1231 vio_enable_interrupts(vscsi->dma_dev); 1517 info = dma_alloc_coherent(&vscsi->dma_dev->dev, sizeof(*info), &token, 1596 dma_free_coherent(&vscsi->dma_dev->dev, sizeof(*info), info, token); 1636 cap = dma_alloc_coherent(&vscsi->dma_dev->dev, olen, &token, 1648 strscpy_pad(cap->name, dev_name(&vscsi->dma_dev->dev), 1712 dma_free_coherent(&vscsi->dma_dev->dev, olen, cap, token); 1939 rc = h_send_crq(vscsi->dma_dev [all...] |