/linux-master/drivers/iommu/ |
H A D | rockchip-iommu.c | 123 static struct device *dma_dev; variable in typeref:struct:device 132 dma_sync_single_for_device(dma_dev, dma, size, DMA_TO_DEVICE); 734 pt_dma = dma_map_single(dma_dev, page_table, SPAGE_SIZE, DMA_TO_DEVICE); 735 if (dma_mapping_error(dma_dev, pt_dma)) { 736 dev_err(dma_dev, "DMA mapping error while allocating page table\n"); 1052 if (!dma_dev) 1068 rk_domain->dt_dma = dma_map_single(dma_dev, rk_domain->dt, 1070 if (dma_mapping_error(dma_dev, rk_domain->dt_dma)) { 1071 dev_err(dma_dev, "DMA map error for DT\n"); 1105 dma_unmap_single(dma_dev, pt_phy [all...] |
/linux-master/drivers/mcb/ |
H A D | mcb-core.c | 233 dev->dma_dev = bus->carrier;
|
/linux-master/drivers/media/usb/stk1160/ |
H A D | stk1160-video.c | 299 struct device *dma_dev = stk1160_get_dmadev(dev); local 316 dma_sync_sgtable_for_cpu(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE); 326 dma_sync_sgtable_for_device(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE); 364 struct device *dma_dev = stk1160_get_dmadev(dev); local 366 dma_vunmap_noncontiguous(dma_dev, stk_urb->transfer_buffer); 367 dma_free_noncontiguous(dma_dev, stk_urb->urb->transfer_buffer_length, 409 struct device *dma_dev = stk1160_get_dmadev(dev); local 414 stk_urb->sgt = dma_alloc_noncontiguous(dma_dev, sb_size, 424 stk_urb->transfer_buffer = dma_vmap_noncontiguous(dma_dev, sb_size, 433 dma_free_noncontiguous(dma_dev, sb_siz [all...] |
/linux-master/drivers/media/usb/uvc/ |
H A D | uvc_video.c | 1654 struct device *dma_dev = uvc_stream_to_dmadev(stream); local 1661 dma_vunmap_noncontiguous(dma_dev, uvc_urb->buffer); 1662 dma_free_noncontiguous(dma_dev, stream->urb_size, uvc_urb->sgt, 1675 struct device *dma_dev = uvc_stream_to_dmadev(stream); local 1677 uvc_urb->sgt = dma_alloc_noncontiguous(dma_dev, stream->urb_size, 1684 uvc_urb->buffer = dma_vmap_noncontiguous(dma_dev, stream->urb_size, 1687 dma_free_noncontiguous(dma_dev, stream->urb_size,
|
/linux-master/drivers/mtd/nand/raw/ |
H A D | cadence-nand-controller.c | 1828 struct dma_device *dma_dev; local 1834 dma_dev = chan->device; 1836 buf_dma = dma_map_single(dma_dev->dev, buf, len, dir); 1837 if (dma_mapping_error(dma_dev->dev, buf_dma)) {
|
H A D | fsmc_nand.c | 489 struct dma_device *dma_dev; local 504 dma_dev = chan->device; 505 dma_addr = dma_map_single(dma_dev->dev, buffer, len, direction); 515 tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src, 548 dma_unmap_single(dma_dev->dev, dma_addr, len, direction);
|
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | b44.c | 151 dma_sync_single_for_device(sdev->dma_dev, dma_base + offset, 160 dma_sync_single_for_cpu(sdev->dma_dev, dma_base + offset, 606 dma_unmap_single(bp->sdev->dma_dev, 651 mapping = dma_map_single(bp->sdev->dma_dev, skb->data, 657 if (dma_mapping_error(bp->sdev->dma_dev, mapping) || 660 if (!dma_mapping_error(bp->sdev->dma_dev, mapping)) 661 dma_unmap_single(bp->sdev->dma_dev, mapping, 667 mapping = dma_map_single(bp->sdev->dma_dev, skb->data, 670 if (dma_mapping_error(bp->sdev->dma_dev, mapping) || 672 if (!dma_mapping_error(bp->sdev->dma_dev, mappin [all...] |
H A D | bgmac-bcma.c | 138 bgmac->dma_dev = core->dma_dev;
|
H A D | bgmac-platform.c | 193 bgmac->dma_dev = &pdev->dev;
|
H A D | bgmac.c | 135 struct device *dma_dev = bgmac->dma_dev; local 162 slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb), 164 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) 180 slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0, 182 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) 211 dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb), 220 dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE); 237 struct device *dma_dev = bgmac->dma_dev; local 336 struct device *dma_dev = bgmac->dma_dev; local 399 bgmac_dma_rx_poison_buf(struct device *dma_dev, struct bgmac_slot_info *slot) argument 425 struct device *dma_dev = bgmac->dma_dev; local 528 struct device *dma_dev = bgmac->dma_dev; local 555 struct device *dma_dev = bgmac->dma_dev; local 576 struct device *dma_dev = bgmac->dma_dev; local 614 struct device *dma_dev = bgmac->dma_dev; local [all...] |
H A D | bgmac.h | 455 struct device *dma_dev; member in struct:bgmac
|
/linux-master/drivers/net/ethernet/freescale/enetc/ |
H A D | enetc.h | 178 struct device *dma_dev; member in struct:enetc_cbdr 466 data = dma_alloc_coherent(ring->dma_dev, 470 dev_err(ring->dma_dev, "CBD alloc data memory failed!\n"); 489 dma_free_coherent(ring->dma_dev, size + ENETC_CBD_DATA_MEM_ALIGN,
|
H A D | enetc_cbdr.c | 25 cbdr->dma_dev = dev; 56 dma_free_coherent(cbdr->dma_dev, size, cbdr->bd_base, 59 cbdr->dma_dev = NULL; 74 dev_warn(ring->dma_dev, "CMD err %04x for cmd %04x\n", 199 dev_err(ring->dma_dev, "FS entry add failed (%d)!", err); 235 dev_err(ring->dma_dev, "RSS cmd failed (%d)!", err);
|
/linux-master/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_queue.c | 19 void *fun_alloc_ring_mem(struct device *dma_dev, size_t depth, argument 24 int dev_node = dev_to_node(dma_dev); 36 set_dev_node(dma_dev, numa_node); 37 va = dma_alloc_coherent(dma_dev, dma_sz, dma_addr, GFP_KERNEL); 38 set_dev_node(dma_dev, dev_node); 46 dma_free_coherent(dma_dev, dma_sz, va, *dma_addr); 57 void fun_free_ring_mem(struct device *dma_dev, size_t depth, size_t hw_desc_sz, argument 65 dma_free_coherent(dma_dev, sz, hw_va, dma_addr); 237 struct device *dma_dev = funq->fdev->dev; local 251 dma_sync_single_for_cpu(dma_dev, rqinf [all...] |
H A D | fun_queue.h | 142 void *fun_alloc_ring_mem(struct device *dma_dev, size_t depth, 146 void fun_free_ring_mem(struct device *dma_dev, size_t depth, size_t hw_desc_sz,
|
/linux-master/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_rx.c | 58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, 77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, 88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, 110 rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE, 112 if (unlikely(dma_mapping_error(q->dma_dev, rb->dma_addr))) { 130 dma_unmap_page(q->dma_dev, rb->dma_addr, PAGE_SIZE, 269 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, 319 dma_sync_single_for_cpu(q->dma_dev, 636 q->dma_dev = &fp->pdev->dev; 638 q->rqes = fun_alloc_ring_mem(q->dma_dev, nrq [all...] |
H A D | funeth_tx.c | 165 if (unlikely(fun_map_pkt(q->dma_dev, shinfo, skb->data, 401 dma_unmap_single(q->dma_dev, be64_to_cpu(gle->sgl_data), 405 dma_unmap_page(q->dma_dev, be64_to_cpu(gle->sgl_data), 410 dma_unmap_page(q->dma_dev, be64_to_cpu(gle->sgl_data), 539 if (unlikely(fun_map_pkt(q->dma_dev, si, xdpf->data, xdpf->len, dma, 641 q->dma_dev = &fp->pdev->dev; 642 q->desc = fun_alloc_ring_mem(q->dma_dev, ndesc, FUNETH_SQE_SIZE, 669 fun_free_ring_mem(q->dma_dev, q->mask + 1, FUNETH_SQE_SIZE, true,
|
H A D | funeth_txrx.h | 111 struct device *dma_dev; /* device for DMA mappings */ member in struct:funeth_txq 166 struct device *dma_dev; /* device for DMA mappings */ member in struct:funeth_rxq
|
/linux-master/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_rxtx.c | 470 struct device *dma_dev = sdma->sw->dev->dev; local 473 dma = dma_map_single(dma_dev, skb->data, skb->len, DMA_TO_DEVICE); 474 if (dma_mapping_error(dma_dev, dma)) 486 struct device *dma_dev = sdma->sw->dev->dev; local 488 dma_unmap_single(dma_dev, buf->buf_dma, buf->skb->len, DMA_TO_DEVICE); 716 struct device *dma_dev = sdma->sw->dev->dev; local 741 dma_sync_single_for_device(dma_dev, buf->buf_dma, skb->len,
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_eth_soc.c | 1141 eth->scratch_ring = dma_alloc_coherent(eth->dma_dev, 1152 dma_addr = dma_map_single(eth->dma_dev, 1155 if (unlikely(dma_mapping_error(eth->dma_dev, dma_addr))) 1219 dma_unmap_single(eth->dma_dev, 1224 dma_unmap_page(eth->dma_dev, 1231 dma_unmap_page(eth->dma_dev, 1238 dma_unmap_page(eth->dma_dev, 1422 txd_info.addr = dma_map_single(eth->dma_dev, skb->data, txd_info.size, 1424 if (unlikely(dma_mapping_error(eth->dma_dev, txd_info.addr))) 1464 txd_info.addr = skb_frag_dma_map(eth->dma_dev, fra 4667 mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev) argument [all...] |
H A D | mtk_eth_soc.h | 1239 struct device *dma_dev; member in struct:mtk_eth 1441 void mtk_eth_set_dma_device(struct mtk_eth *eth, struct device *dma_dev);
|
H A D | mtk_wed.c | 1151 hw->eth->dma_dev != hw->eth->dev) 2420 if (hw->eth->dma_dev == hw->eth->dev &&
|
/linux-master/drivers/net/ethernet/sfc/falcon/ |
H A D | tx.c | 57 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local 60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, 63 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, 210 struct device *dma_dev = &efx->pci_dev->dev; local 221 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); 226 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) 255 dma_addr = skb_frag_dma_map(dma_dev, fragment, 261 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
|
/linux-master/drivers/net/ethernet/sfc/siena/ |
H A D | tx_common.c | 128 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local 132 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, 135 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, 333 struct device *dma_dev = &efx->pci_dev->dev; local 344 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); 349 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) 392 dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len, 398 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
|
/linux-master/drivers/net/ethernet/sfc/ |
H A D | tx_common.c | 156 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; local 160 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, 163 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, 352 struct device *dma_dev = &efx->pci_dev->dev; local 363 dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE); 368 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) 411 dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len, 417 if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
|