Searched refs:dma_dev (Results 76 - 100 of 177) sorted by path

12345678

/linux-master/drivers/iommu/
H A Drockchip-iommu.c123 static struct device *dma_dev; variable in typeref:struct:device
132 dma_sync_single_for_device(dma_dev, dma, size, DMA_TO_DEVICE);
734 pt_dma = dma_map_single(dma_dev, page_table, SPAGE_SIZE, DMA_TO_DEVICE);
735 if (dma_mapping_error(dma_dev, pt_dma)) {
736 dev_err(dma_dev, "DMA mapping error while allocating page table\n");
1052 if (!dma_dev)
1068 rk_domain->dt_dma = dma_map_single(dma_dev, rk_domain->dt,
1070 if (dma_mapping_error(dma_dev, rk_domain->dt_dma)) {
1071 dev_err(dma_dev, "DMA map error for DT\n");
1105 dma_unmap_single(dma_dev, pt_phy
[all...]
/linux-master/drivers/mcb/
H A Dmcb-core.c233 dev->dma_dev = bus->carrier;
/linux-master/drivers/media/usb/stk1160/
H A Dstk1160-video.c299 struct device *dma_dev = stk1160_get_dmadev(dev); local
316 dma_sync_sgtable_for_cpu(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE);
326 dma_sync_sgtable_for_device(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE);
364 struct device *dma_dev = stk1160_get_dmadev(dev); local
366 dma_vunmap_noncontiguous(dma_dev, stk_urb->transfer_buffer);
367 dma_free_noncontiguous(dma_dev, stk_urb->urb->transfer_buffer_length,
409 struct device *dma_dev = stk1160_get_dmadev(dev); local
414 stk_urb->sgt = dma_alloc_noncontiguous(dma_dev, sb_size,
424 stk_urb->transfer_buffer = dma_vmap_noncontiguous(dma_dev, sb_size,
433 dma_free_noncontiguous(dma_dev, sb_siz
[all...]
/linux-master/drivers/media/usb/uvc/
H A Duvc_video.c1654 struct device *dma_dev = uvc_stream_to_dmadev(stream); local
1661 dma_vunmap_noncontiguous(dma_dev, uvc_urb->buffer);
1662 dma_free_noncontiguous(dma_dev, stream->urb_size, uvc_urb->sgt,
1675 struct device *dma_dev = uvc_stream_to_dmadev(stream); local
1677 uvc_urb->sgt = dma_alloc_noncontiguous(dma_dev, stream->urb_size,
1684 uvc_urb->buffer = dma_vmap_noncontiguous(dma_dev, stream->urb_size,
1687 dma_free_noncontiguous(dma_dev, stream->urb_size,
/linux-master/drivers/mtd/nand/raw/
H A Dcadence-nand-controller.c1828 struct dma_device *dma_dev; local
1834 dma_dev = chan->device;
1836 buf_dma = dma_map_single(dma_dev->dev, buf, len, dir);
1837 if (dma_mapping_error(dma_dev->dev, buf_dma)) {
H A Dfsmc_nand.c489 struct dma_device *dma_dev; local
504 dma_dev = chan->device;
505 dma_addr = dma_map_single(dma_dev->dev, buffer, len, direction);
515 tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src,
548 dma_unmap_single(dma_dev->dev, dma_addr, len, direction);
/linux-master/drivers/net/ethernet/broadcom/
H A Db44.c151 dma_sync_single_for_device(sdev->dma_dev, dma_base + offset,
160 dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset,
606 dma_unmap_single(bp->sdev->dma_dev,
651 mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
657 if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
660 if (!dma_mapping_error(bp->sdev->dma_dev, mapping))
661 dma_unmap_single(bp->sdev->dma_dev, mapping,
667 mapping = dma_map_single(bp->sdev->dma_dev, skb->data,
670 if (dma_mapping_error(bp->sdev->dma_dev, mapping) ||
672 if (!dma_mapping_error(bp->sdev->dma_dev, mappin
[all...]
H A Dbgmac-bcma.c138 bgmac->dma_dev = core->dma_dev;
H A Dbgmac-platform.c193 bgmac->dma_dev = &pdev->dev;
H A Dbgmac.c135 struct device *dma_dev = bgmac->dma_dev; local
162 slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb),
164 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
180 slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0,
182 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
211 dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb),
220 dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE);
237 struct device *dma_dev = bgmac->dma_dev; local
336 struct device *dma_dev = bgmac->dma_dev; local
399 bgmac_dma_rx_poison_buf(struct device *dma_dev, struct bgmac_slot_info *slot) argument
425 struct device *dma_dev = bgmac->dma_dev; local
528 struct device *dma_dev = bgmac->dma_dev; local
555 struct device *dma_dev = bgmac->dma_dev; local
576 struct device *dma_dev = bgmac->dma_dev; local
614 struct device *dma_dev = bgmac->dma_dev; local
[all...]
H A Dbgmac.h455 struct device *dma_dev; member in struct:bgmac
/linux-master/drivers/net/ethernet/freescale/enetc/
H A Denetc.h178 struct device *dma_dev; member in struct:enetc_cbdr
466 data = dma_alloc_coherent(ring->dma_dev,
470 dev_err(ring->dma_dev, "CBD alloc data memory failed!\n");
489 dma_free_coherent(ring->dma_dev, size + ENETC_CBD_DATA_MEM_ALIGN,
H A Denetc_cbdr.c25 cbdr->dma_dev = dev;
56 dma_free_coherent(cbdr->dma_dev, size, cbdr->bd_base,
59 cbdr->dma_dev = NULL;
74 dev_warn(ring->dma_dev, "CMD err %04x for cmd %04x\n",
199 dev_err(ring->dma_dev, "FS entry add failed (%d)!", err);
235 dev_err(ring->dma_dev, "RSS cmd failed (%d)!", err);
/linux-master/drivers/net/ethernet/fungible/funcore/
H A Dfun_queue.c19 void *fun_alloc_ring_mem(struct device *dma_dev, size_t depth, argument
24 int dev_node = dev_to_node(dma_dev);
36 set_dev_node(dma_dev, numa_node);
37 va = dma_alloc_coherent(dma_dev, dma_sz, dma_addr, GFP_KERNEL);
38 set_dev_node(dma_dev, dev_node);
46 dma_free_coherent(dma_dev, dma_sz, va, *dma_addr);
57 void fun_free_ring_mem(struct device *dma_dev, size_t depth, size_t hw_desc_sz, argument
65 dma_free_coherent(dma_dev, sz, hw_va, dma_addr);
237 struct device *dma_dev = funq->fdev->dev; local
251 dma_sync_single_for_cpu(dma_dev, rqinf
[all...]
H A Dfun_queue.h142 void *fun_alloc_ring_mem(struct device *dma_dev, size_t depth,
146 void fun_free_ring_mem(struct device *dma_dev, size_t depth, size_t hw_desc_sz,
/linux-master/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_rx.c58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE,
112 if (unlikely(dma_mapping_error(q->dma_dev, rb->dma_addr))) {
130 dma_unmap_page(q->dma_dev, rb->dma_addr, PAGE_SIZE,
269 dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
319 dma_sync_single_for_cpu(q->dma_dev,
636 q->dma_dev = &fp->pdev->dev;
638 q->rqes = fun_alloc_ring_mem(q->dma_dev, nrq
[all...]
H A Dfuneth_tx.c165 if (unlikely(fun_map_pkt(q->dma_dev, shinfo, skb->data,
401 dma_unmap_single(q->dma_dev, be64_to_cpu(gle->sgl_data),
405 dma_unmap_page(q->dma_dev, be64_to_cpu(gle->sgl_data),
410 dma_unmap_page(q->dma_dev, be64_to_cpu(gle->sgl_data),
539 if (unlikely(fun_map_pkt(q->dma_dev, si, xdpf->data, xdpf->len, dma,
641 q->dma_dev = &fp->pdev->dev;
642 q->desc = fun_alloc_ring_mem(q->dma_dev, ndesc, FUNETH_SQE_SIZE,
669 fun_free_ring_mem(q->dma_dev, q->mask + 1, FUNETH_SQE_SIZE, true,
H A Dfuneth_txrx.h111 struct device *dma_dev; /* device for DMA mappings */ member in struct:funeth_txq
166 struct device *dma_dev; /* device for DMA mappings */ member in struct:funeth_rxq
/linux-master/drivers/net/ethernet/marvell/prestera/
H A Dprestera_rxtx.c470 struct device *dma_dev = sdma->sw->dev->dev; local
473 dma = dma_map_single(dma_dev, skb->data, skb->len, DMA_TO_DEVICE);
474 if (dma_mapping_error(dma_dev, dma))
486 struct device *dma_dev = sdma->sw->dev->dev; local
488 dma_unmap_single(dma_dev, buf->buf_dma, buf->skb->len, DMA_TO_DEVICE);
716 struct device *dma_dev = sdma->sw->dev->dev; local
741 dma_sync_single_for_device(dma_dev, buf->buf_dma, skb->len,
/linux-master/drivers/net/ethernet/mediatek/
H A Dmtk_eth_soc.c1141 eth->scratch_ring = dma_alloc_coherent(eth->dma_dev,
1152 dma_addr = dma_map_single(eth->dma_dev,
1155 if (unlikely(dma_mapping_error(eth->dma_dev, dma_addr)))
1219 dma_unmap_single(eth->dma_dev,
1224 dma_unmap_page(eth->dma_dev,
1231 dma_unmap_page(eth->dma_dev,
1238 dma_unmap_page(eth->dma_dev,
1422 txd_info.addr = dma_map_single(eth->dma_dev, skb->data, txd_info.size,
1424 if (unlikely(dma_mapping_error(eth->dma_dev, txd_info.addr)))
1464 txd_info.addr = skb_frag_dma_map(eth->dma_dev, fra
4667 mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev) argument
[all...]
H A Dmtk_eth_soc.h1239 struct device *dma_dev; member in struct:mtk_eth
1441 void mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev);
H A Dmtk_wed.c1151 hw->eth->dma_dev != hw->eth->dev)
2420 if (hw->eth->dma_dev == hw->eth->dev &&
/linux-master/drivers/net/ethernet/sfc/falcon/
H A Dtx.c57 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local
60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
63 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
210 struct device *dma_dev = &efx->pci_dev->dev; local
221 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
226 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
255 dma_addr = skb_frag_dma_map(dma_dev, fragment,
261 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
/linux-master/drivers/net/ethernet/sfc/siena/
H A Dtx_common.c128 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local
132 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
135 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
333 struct device *dma_dev = &efx->pci_dev->dev; local
344 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
349 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
392 dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len,
398 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
/linux-master/drivers/net/ethernet/sfc/
H A Dtx_common.c156 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local
160 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
163 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
352 struct device *dma_dev = &efx->pci_dev->dev; local
363 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
368 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
411 dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len,
417 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))

Completed in 580 milliseconds

12345678