Searched refs:dma_addr (Results 151 - 175 of 819) sorted by relevance

1234567891011>>

/linux-master/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_dma.c48 dma_addr_t dma_addr; local
80 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
81 writel(lower_32_bits(dma_addr),
84 dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
85 writel(lower_32_bits(dma_addr),
/linux-master/drivers/net/ethernet/cavium/thunder/
H A Dnicvf_queues.h218 u64 dma_addr; member in struct:pgcache
320 static inline u64 nicvf_iova_to_phys(struct nicvf *nic, dma_addr_t dma_addr) argument
324 return iommu_iova_to_phys(nic->iommu_domain, dma_addr);
325 return dma_addr;
346 u64 bufaddr, u64 dma_addr, u16 len);
/linux-master/drivers/infiniband/hw/efa/
H A Defa_com.c134 dma_alloc_coherent(aq->dmadev, size, &sq->dma_addr, GFP_KERNEL);
146 addr_high = upper_32_bits(sq->dma_addr);
147 addr_low = lower_32_bits(sq->dma_addr);
171 dma_alloc_coherent(aq->dmadev, size, &cq->dma_addr, GFP_KERNEL);
180 addr_high = upper_32_bits(cq->dma_addr);
181 addr_low = lower_32_bits(cq->dma_addr);
211 aenq->entries = dma_alloc_coherent(edev->dmadev, size, &aenq->dma_addr,
221 addr_low = lower_32_bits(aenq->dma_addr);
222 addr_high = upper_32_bits(aenq->dma_addr);
673 dma_free_coherent(edev->dmadev, size, sq->entries, sq->dma_addr);
[all...]
/linux-master/drivers/net/ethernet/brocade/bna/
H A Dbfa_ioc.h50 #define bfa_dma_be_addr_set(dma_addr, pa) \
51 __bfa_dma_be_addr_set(&dma_addr, (u64)pa)
53 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) argument
55 dma_addr->a32.addr_lo = (u32) htonl(pa);
56 dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa));
/linux-master/drivers/net/ethernet/sfc/falcon/
H A Drx.c88 dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len,
131 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
159 dma_addr_t dma_addr; local
171 dma_addr =
176 dma_addr))) {
181 state->dma_addr = dma_addr;
184 dma_addr = state->dma_addr;
187 dma_addr
[all...]
/linux-master/drivers/net/ethernet/mscc/
H A Docelot_fdma.c111 dma_addr_t dma_addr,
114 u32 offset = dma_addr & 0x3;
117 dcb->datap = ALIGN_DOWN(dma_addr, 4);
141 rxb->dma_addr = mapping;
152 dma_addr_t dma_addr; local
172 dma_addr = rxb->dma_addr + rxb->page_offset;
173 ocelot_fdma_dcb_set_data(dcb, dma_addr, OCELOT_FDMA_RXB_SIZE);
198 dma_unmap_addr_set(tx_buf, dma_addr, mapping);
309 dma_sync_single_range_for_device(ocelot->dev, old_rxb->dma_addr,
110 ocelot_fdma_dcb_set_data(struct ocelot_fdma_dcb *dcb, dma_addr_t dma_addr, size_t size) argument
[all...]
/linux-master/drivers/crypto/ccree/
H A Dcc_buffer_mgr.c227 dev_dbg(dev, "MLLI params: virt_addr=%pK dma_addr=%pad mlli_len=0x%X\n",
954 dma_addr_t dma_addr; local
977 dma_addr = dma_map_single(dev, areq_ctx->mac_buf, MAX_MAC_SIZE,
979 if (dma_mapping_error(dev, dma_addr)) {
985 areq_ctx->mac_buf_dma_addr = dma_addr;
990 dma_addr = dma_map_single(dev, addr, AES_BLOCK_SIZE,
993 if (dma_mapping_error(dev, dma_addr)) {
1000 areq_ctx->ccm_iv0_dma_addr = dma_addr;
1009 dma_addr = dma_map_single(dev, areq_ctx->hkey, AES_BLOCK_SIZE,
1011 if (dma_mapping_error(dev, dma_addr)) {
[all...]
/linux-master/drivers/gpu/drm/omapdrm/
H A Domap_gem.c45 * dma_addr contains the buffer DMA address. It is valid for
65 dma_addr_t dma_addr; member in struct:omap_gem_object
114 dma_addr_t dma_addr; member in struct:omap_drm_usergart_entry
367 pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff;
463 pfn = entry->dma_addr >> PAGE_SHIFT;
776 omap_obj->dma_addr = tiler_ssptr(block);
779 DBG("got dma address: %pad", &omap_obj->dma_addr);
788 * @dma_addr: the DMA address
790 * Pin the given GEM object in memory and fill the dma_addr pointer with the
799 int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr) argument
898 omap_gem_rotated_dma_addr(struct drm_gem_object *obj, u32 orient, int x, int y, dma_addr_t *dma_addr) argument
[all...]
H A Domap_fb.c46 dma_addr_t dma_addr; member in struct:plane
56 /* lock for pinning (pin_count and planes.dma_addr) */
95 return plane->dma_addr + offset;
272 ret = omap_gem_pin(fb->obj[i], &plane->dma_addr);
288 plane->dma_addr = 0;
314 plane->dma_addr = 0;
439 plane->dma_addr = 0;
/linux-master/arch/arm/mm/
H A Ddma-mapping.c967 dma_addr_t dma_addr, iova; local
970 dma_addr = __alloc_iova(mapping, size);
971 if (dma_addr == DMA_MAPPING_ERROR)
972 return dma_addr;
974 iova = dma_addr;
995 return dma_addr;
997 iommu_unmap(mapping->domain, dma_addr, iova-dma_addr);
998 __free_iova(mapping, dma_addr, size);
1116 void *cpu_addr, dma_addr_t dma_addr, size_
1115 arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, dma_addr_t dma_addr, size_t size, unsigned long attrs) argument
1168 arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, size_t size, unsigned long attrs) argument
1373 dma_addr_t dma_addr; local
1438 dma_addr_t dma_addr; local
[all...]
/linux-master/drivers/net/ethernet/sfc/
H A Drx_common.c60 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
166 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
331 state->dma_addr,
379 dma_addr_t dma_addr; local
391 dma_addr =
396 dma_addr))) {
401 state->dma_addr = dma_addr;
404 dma_addr = state->dma_addr;
[all...]
H A Dtx.c47 buffer->dma_addr = page_buf->dma_addr + offset;
427 dma_addr_t dma_addr; local
479 dma_addr = dma_map_single(&efx->pci_dev->dev,
482 if (dma_mapping_error(&efx->pci_dev->dev, dma_addr))
486 tx_buffer = efx_tx_map_chunk(tx_queue, dma_addr, len);
/linux-master/drivers/net/ethernet/sfc/siena/
H A Drx_common.c63 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
169 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
331 state->dma_addr,
379 dma_addr_t dma_addr; local
391 dma_addr =
396 dma_addr))) {
401 state->dma_addr = dma_addr;
404 dma_addr = state->dma_addr;
[all...]
/linux-master/drivers/net/wan/
H A Dfsl_qmc_hdlc.c31 dma_addr_t dma_addr; member in struct:qmc_hdlc_desc
257 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE);
296 desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data,
298 ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr);
302 ret = qmc_chan_read_submit(qmc_hdlc->qmc_chan, desc->dma_addr, desc->dma_size,
310 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE);
328 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_TO_DEVICE);
345 desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data,
347 ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr);
353 ret = qmc_chan_write_submit(qmc_hdlc->qmc_chan, desc->dma_addr, des
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c142 struct page *dpage, dma_addr_t *dma_addr)
148 *dma_addr = dma_map_page(dev, dpage, 0, PAGE_SIZE, DMA_BIDIRECTIONAL);
149 if (dma_mapping_error(dev, *dma_addr))
152 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr,
154 dma_unmap_page(dev, *dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL);
169 dma_addr_t dma_addr = 0; local
205 ret = nouveau_dmem_copy_one(drm, spage, dpage, &dma_addr);
215 dma_unmap_page(drm->dev->dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL);
617 dma_addr_t *dma_addr, u64 *pfn)
633 *dma_addr
141 nouveau_dmem_copy_one(struct nouveau_drm *drm, struct page *spage, struct page *dpage, dma_addr_t *dma_addr) argument
615 nouveau_dmem_migrate_copy_one(struct nouveau_drm *drm, struct nouveau_svmm *svmm, unsigned long src, dma_addr_t *dma_addr, u64 *pfn) argument
[all...]
/linux-master/drivers/net/ethernet/altera/
H A Daltera_tse_main.c202 rxbuffer->dma_addr = dma_map_single(priv->device, rxbuffer->skb->data,
206 if (dma_mapping_error(priv->device, rxbuffer->dma_addr)) {
211 rxbuffer->dma_addr &= (dma_addr_t)~3;
219 dma_addr_t dma_addr = rxbuffer->dma_addr; local
223 if (dma_addr)
224 dma_unmap_single(priv->device, dma_addr,
229 rxbuffer->dma_addr = 0;
238 if (buffer->dma_addr) {
240 dma_unmap_page(priv->device, buffer->dma_addr,
550 dma_addr_t dma_addr; local
[all...]
/linux-master/drivers/net/ethernet/cisco/enic/
H A Denic_main.c329 dma_unmap_single(&enic->pdev->dev, buf->dma_addr, buf->len,
332 dma_unmap_page(&enic->pdev->dev, buf->dma_addr, buf->len,
549 dma_addr_t dma_addr; local
554 dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag, 0,
557 if (unlikely(enic_dma_map_check(enic, dma_addr)))
559 enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag),
574 dma_addr_t dma_addr; local
577 dma_addr = dma_map_single(&enic->pdev->dev, skb->data, head_len,
579 if (unlikely(enic_dma_map_check(enic, dma_addr)))
587 enic_queue_wq_desc(wq, skb, dma_addr, head_le
605 dma_addr_t dma_addr; local
678 dma_addr_t dma_addr; local
752 dma_addr_t dma_addr; local
1242 dma_addr_t dma_addr; local
[all...]
/linux-master/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_dp.c11 * @dma_addr: Pointer to storage for DMA address (output param)
17 void *nfp_net_rx_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr) argument
34 *dma_addr = nfp_net_dma_map_rx(dp, frag);
35 if (dma_mapping_error(dp->dev, *dma_addr)) {
148 nfp_net_dma_unmap_rx(dp, rx_ring->rxbufs[i].dma_addr);
150 rx_ring->rxbufs[i].dma_addr = 0;
173 rxbufs[i].frag = nfp_net_rx_alloc_one(dp, &rxbufs[i].dma_addr);
/linux-master/drivers/media/platform/mediatek/vcodec/decoder/vdec/
H A Dvdec_h264_if.c150 inst->vsi->pred_buf_dma = inst->pred_buf.dma_addr;
181 inst->vsi->mv_buf_dma[i] = mem->dma_addr;
233 fb->base_y.va, (u64)fb->base_y.dma_addr);
342 uint64_t y_fb_dma = fb ? (u64)fb->base_y.dma_addr : 0;
343 uint64_t c_fb_dma = fb ? (u64)fb->base_c.dma_addr : 0;
375 inst->vsi->dec.bs_dma = (uint64_t)bs->dma_addr;
H A Dvdec_av1_req_lat_if.c112 * @buf: dma_addr padding
113 * @dma_addr: buffer address
121 dma_addr_t dma_addr; member in union:vdec_av1_slice_mem::__anon330
1000 vsi->cdf_tmp.buf = instance->cdf_temp.dma_addr;
1592 vsi->bs.dma_addr = bs->dma_addr;
1595 vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
1597 vsi->trans.dma_addr = lat_buf->ctx->msg_queue.wdma_wptr_addr;
1600 vsi->err_map.dma_addr
[all...]
/linux-master/drivers/crypto/intel/keembay/
H A Dkeembay-ocs-aes-core.c246 rctx->src_dll.dma_addr = DMA_MAPPING_ERROR;
247 rctx->dst_dll.dma_addr = DMA_MAPPING_ERROR;
248 rctx->aad_src_dll.dma_addr = DMA_MAPPING_ERROR;
249 rctx->aad_dst_dll.dma_addr = DMA_MAPPING_ERROR;
368 dma_free_coherent(dev, dll->size, dll->vaddr, dll->dma_addr);
371 dll->dma_addr = DMA_MAPPING_ERROR;
438 * If descriptor creation was successful, set the src_dll.dma_addr to
439 * the value of dst_dll.dma_addr, as we do in-place AES operation on
442 rctx->src_dll.dma_addr = rctx->dst_dll.dma_addr;
[all...]
/linux-master/drivers/infiniband/hw/qib/
H A Dqib_user_sdma.c259 dma_addr_t dma_addr, u16 dma_length)
269 pkt->addr[i].addr = dma_addr;
274 size_t len, dma_addr_t *dma_addr)
280 dma_addr);
289 *dma_addr = 0;
308 dma_addr_t dma_addr = local
313 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) {
368 dma_addr, len); /* dma addr, dma length */
809 dma_addr_t dma_addr; local
836 pbc = qib_user_sdma_alloc_header(pq, len, &dma_addr);
254 qib_user_sdma_init_frag(struct qib_user_sdma_pkt *pkt, int i, u16 offset, u16 len, u16 first_desc, u16 last_desc, u16 put_page, u16 dma_mapped, struct page *page, void *kvaddr, dma_addr_t dma_addr, u16 dma_length) argument
273 qib_user_sdma_alloc_header(struct qib_user_sdma_queue *pq, size_t len, dma_addr_t *dma_addr) argument
[all...]
/linux-master/drivers/net/wireless/mediatek/mt76/
H A Ddma.c62 t->dma_addr = addr;
166 dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
272 entry->dma_addr[0] = buf->addr;
312 entry->dma_addr[0] = buf[0].addr;
320 entry->dma_addr[1] = buf[1].addr;
360 dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0],
364 dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1],
453 dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr,
458 t->dma_addr = 0;
465 dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_tx.c181 dma_addr_t dma_addr = 0; local
186 dma_addr = dma_map_single(sq->pdev, skb_data, headlen,
188 if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
191 dseg->addr = cpu_to_be64(dma_addr);
195 mlx5e_dma_push(sq, dma_addr, headlen, MLX5E_DMA_MAP_SINGLE);
204 dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
206 if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
209 dseg->addr = cpu_to_be64(dma_addr);
213 mlx5e_dma_push(sq, dma_addr, fsz, MLX5E_DMA_MAP_PAGE);
558 dseg->addr = cpu_to_be64(txd->dma_addr);
[all...]
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c109 otx2_dma_unmap_page(pfvf, sg->dma_addr[seg],
126 pa = otx2_iova_to_phys(pfvf->iommu_domain, sg->dma_addr[0]);
127 otx2_dma_unmap_page(pfvf, sg->dma_addr[0],
618 u64 dma_addr, *iova = NULL; local
640 dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
641 if (dma_mapping_error(pfvf->dev, dma_addr))
646 *iova++ = dma_addr;
649 sq->sg[sq->head].dma_addr[seg] = dma_addr;
804 u64 dma_addr; local
1347 otx2_xdp_sqe_add_sg(struct otx2_snd_queue *sq, u64 dma_addr, int len, int *offset) argument
[all...]

Completed in 239 milliseconds

1234567891011>>