/linux-master/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_dma.c | 48 dma_addr_t dma_addr; local 80 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); 81 writel(lower_32_bits(dma_addr), 84 dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES); 85 writel(lower_32_bits(dma_addr),
|
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.h | 218 u64 dma_addr; member in struct:pgcache 320 static inline u64 nicvf_iova_to_phys(struct nicvf *nic, dma_addr_t dma_addr) argument 324 return iommu_iova_to_phys(nic->iommu_domain, dma_addr); 325 return dma_addr; 346 u64 bufaddr, u64 dma_addr, u16 len);
|
/linux-master/drivers/infiniband/hw/efa/ |
H A D | efa_com.c | 134 dma_alloc_coherent(aq->dmadev, size, &sq->dma_addr, GFP_KERNEL); 146 addr_high = upper_32_bits(sq->dma_addr); 147 addr_low = lower_32_bits(sq->dma_addr); 171 dma_alloc_coherent(aq->dmadev, size, &cq->dma_addr, GFP_KERNEL); 180 addr_high = upper_32_bits(cq->dma_addr); 181 addr_low = lower_32_bits(cq->dma_addr); 211 aenq->entries = dma_alloc_coherent(edev->dmadev, size, &aenq->dma_addr, 221 addr_low = lower_32_bits(aenq->dma_addr); 222 addr_high = upper_32_bits(aenq->dma_addr); 673 dma_free_coherent(edev->dmadev, size, sq->entries, sq->dma_addr); [all...] |
/linux-master/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc.h | 50 #define bfa_dma_be_addr_set(dma_addr, pa) \ 51 __bfa_dma_be_addr_set(&dma_addr, (u64)pa) 53 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) argument 55 dma_addr->a32.addr_lo = (u32) htonl(pa); 56 dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa));
|
/linux-master/drivers/net/ethernet/sfc/falcon/ |
H A D | rx.c | 88 dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len, 131 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr, 159 dma_addr_t dma_addr; local 171 dma_addr = 176 dma_addr))) { 181 state->dma_addr = dma_addr; 184 dma_addr = state->dma_addr; 187 dma_addr [all...] |
/linux-master/drivers/net/ethernet/mscc/ |
H A D | ocelot_fdma.c | 111 dma_addr_t dma_addr, 114 u32 offset = dma_addr & 0x3; 117 dcb->datap = ALIGN_DOWN(dma_addr, 4); 141 rxb->dma_addr = mapping; 152 dma_addr_t dma_addr; local 172 dma_addr = rxb->dma_addr + rxb->page_offset; 173 ocelot_fdma_dcb_set_data(dcb, dma_addr, OCELOT_FDMA_RXB_SIZE); 198 dma_unmap_addr_set(tx_buf, dma_addr, mapping); 309 dma_sync_single_range_for_device(ocelot->dev, old_rxb->dma_addr, 110 ocelot_fdma_dcb_set_data(struct ocelot_fdma_dcb *dcb, dma_addr_t dma_addr, size_t size) argument [all...] |
/linux-master/drivers/crypto/ccree/ |
H A D | cc_buffer_mgr.c | 227 dev_dbg(dev, "MLLI params: virt_addr=%pK dma_addr=%pad mlli_len=0x%X\n", 954 dma_addr_t dma_addr; local 977 dma_addr = dma_map_single(dev, areq_ctx->mac_buf, MAX_MAC_SIZE, 979 if (dma_mapping_error(dev, dma_addr)) { 985 areq_ctx->mac_buf_dma_addr = dma_addr; 990 dma_addr = dma_map_single(dev, addr, AES_BLOCK_SIZE, 993 if (dma_mapping_error(dev, dma_addr)) { 1000 areq_ctx->ccm_iv0_dma_addr = dma_addr; 1009 dma_addr = dma_map_single(dev, areq_ctx->hkey, AES_BLOCK_SIZE, 1011 if (dma_mapping_error(dev, dma_addr)) { [all...] |
/linux-master/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem.c | 45 * dma_addr contains the buffer DMA address. It is valid for 65 dma_addr_t dma_addr; member in struct:omap_gem_object 114 dma_addr_t dma_addr; member in struct:omap_drm_usergart_entry 367 pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff; 463 pfn = entry->dma_addr >> PAGE_SHIFT; 776 omap_obj->dma_addr = tiler_ssptr(block); 779 DBG("got dma address: %pad", &omap_obj->dma_addr); 788 * @dma_addr: the DMA address 790 * Pin the given GEM object in memory and fill the dma_addr pointer with the 799 int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr) argument 898 omap_gem_rotated_dma_addr(struct drm_gem_object *obj, u32 orient, int x, int y, dma_addr_t *dma_addr) argument [all...] |
H A D | omap_fb.c | 46 dma_addr_t dma_addr; member in struct:plane 56 /* lock for pinning (pin_count and planes.dma_addr) */ 95 return plane->dma_addr + offset; 272 ret = omap_gem_pin(fb->obj[i], &plane->dma_addr); 288 plane->dma_addr = 0; 314 plane->dma_addr = 0; 439 plane->dma_addr = 0;
|
/linux-master/arch/arm/mm/ |
H A D | dma-mapping.c | 967 dma_addr_t dma_addr, iova; local 970 dma_addr = __alloc_iova(mapping, size); 971 if (dma_addr == DMA_MAPPING_ERROR) 972 return dma_addr; 974 iova = dma_addr; 995 return dma_addr; 997 iommu_unmap(mapping->domain, dma_addr, iova-dma_addr); 998 __free_iova(mapping, dma_addr, size); 1116 void *cpu_addr, dma_addr_t dma_addr, size_ 1115 arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, dma_addr_t dma_addr, size_t size, unsigned long attrs) argument 1168 arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, size_t size, unsigned long attrs) argument 1373 dma_addr_t dma_addr; local 1438 dma_addr_t dma_addr; local [all...] |
/linux-master/drivers/net/ethernet/sfc/ |
H A D | rx_common.c | 60 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr, 166 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr, 331 state->dma_addr, 379 dma_addr_t dma_addr; local 391 dma_addr = 396 dma_addr))) { 401 state->dma_addr = dma_addr; 404 dma_addr = state->dma_addr; [all...] |
H A D | tx.c | 47 buffer->dma_addr = page_buf->dma_addr + offset; 427 dma_addr_t dma_addr; local 479 dma_addr = dma_map_single(&efx->pci_dev->dev, 482 if (dma_mapping_error(&efx->pci_dev->dev, dma_addr)) 486 tx_buffer = efx_tx_map_chunk(tx_queue, dma_addr, len);
|
/linux-master/drivers/net/ethernet/sfc/siena/ |
H A D | rx_common.c | 63 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr, 169 dma_unmap_page(&efx->pci_dev->dev, state->dma_addr, 331 state->dma_addr, 379 dma_addr_t dma_addr; local 391 dma_addr = 396 dma_addr))) { 401 state->dma_addr = dma_addr; 404 dma_addr = state->dma_addr; [all...] |
/linux-master/drivers/net/wan/ |
H A D | fsl_qmc_hdlc.c | 31 dma_addr_t dma_addr; member in struct:qmc_hdlc_desc 257 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE); 296 desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data, 298 ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr); 302 ret = qmc_chan_read_submit(qmc_hdlc->qmc_chan, desc->dma_addr, desc->dma_size, 310 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE); 328 dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_TO_DEVICE); 345 desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data, 347 ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr); 353 ret = qmc_chan_write_submit(qmc_hdlc->qmc_chan, desc->dma_addr, des [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 142 struct page *dpage, dma_addr_t *dma_addr) 148 *dma_addr = dma_map_page(dev, dpage, 0, PAGE_SIZE, DMA_BIDIRECTIONAL); 149 if (dma_mapping_error(dev, *dma_addr)) 152 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, 154 dma_unmap_page(dev, *dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL); 169 dma_addr_t dma_addr = 0; local 205 ret = nouveau_dmem_copy_one(drm, spage, dpage, &dma_addr); 215 dma_unmap_page(drm->dev->dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL); 617 dma_addr_t *dma_addr, u64 *pfn) 633 *dma_addr 141 nouveau_dmem_copy_one(struct nouveau_drm *drm, struct page *spage, struct page *dpage, dma_addr_t *dma_addr) argument 615 nouveau_dmem_migrate_copy_one(struct nouveau_drm *drm, struct nouveau_svmm *svmm, unsigned long src, dma_addr_t *dma_addr, u64 *pfn) argument [all...] |
/linux-master/drivers/net/ethernet/altera/ |
H A D | altera_tse_main.c | 202 rxbuffer->dma_addr = dma_map_single(priv->device, rxbuffer->skb->data, 206 if (dma_mapping_error(priv->device, rxbuffer->dma_addr)) { 211 rxbuffer->dma_addr &= (dma_addr_t)~3; 219 dma_addr_t dma_addr = rxbuffer->dma_addr; local 223 if (dma_addr) 224 dma_unmap_single(priv->device, dma_addr, 229 rxbuffer->dma_addr = 0; 238 if (buffer->dma_addr) { 240 dma_unmap_page(priv->device, buffer->dma_addr, 550 dma_addr_t dma_addr; local [all...] |
/linux-master/drivers/net/ethernet/cisco/enic/ |
H A D | enic_main.c | 329 dma_unmap_single(&enic->pdev->dev, buf->dma_addr, buf->len, 332 dma_unmap_page(&enic->pdev->dev, buf->dma_addr, buf->len, 549 dma_addr_t dma_addr; local 554 dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag, 0, 557 if (unlikely(enic_dma_map_check(enic, dma_addr))) 559 enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag), 574 dma_addr_t dma_addr; local 577 dma_addr = dma_map_single(&enic->pdev->dev, skb->data, head_len, 579 if (unlikely(enic_dma_map_check(enic, dma_addr))) 587 enic_queue_wq_desc(wq, skb, dma_addr, head_le 605 dma_addr_t dma_addr; local 678 dma_addr_t dma_addr; local 752 dma_addr_t dma_addr; local 1242 dma_addr_t dma_addr; local [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_dp.c | 11 * @dma_addr: Pointer to storage for DMA address (output param) 17 void *nfp_net_rx_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr) argument 34 *dma_addr = nfp_net_dma_map_rx(dp, frag); 35 if (dma_mapping_error(dp->dev, *dma_addr)) { 148 nfp_net_dma_unmap_rx(dp, rx_ring->rxbufs[i].dma_addr); 150 rx_ring->rxbufs[i].dma_addr = 0; 173 rxbufs[i].frag = nfp_net_rx_alloc_one(dp, &rxbufs[i].dma_addr);
|
/linux-master/drivers/media/platform/mediatek/vcodec/decoder/vdec/ |
H A D | vdec_h264_if.c | 150 inst->vsi->pred_buf_dma = inst->pred_buf.dma_addr; 181 inst->vsi->mv_buf_dma[i] = mem->dma_addr; 233 fb->base_y.va, (u64)fb->base_y.dma_addr); 342 uint64_t y_fb_dma = fb ? (u64)fb->base_y.dma_addr : 0; 343 uint64_t c_fb_dma = fb ? (u64)fb->base_c.dma_addr : 0; 375 inst->vsi->dec.bs_dma = (uint64_t)bs->dma_addr;
|
H A D | vdec_av1_req_lat_if.c | 112 * @buf: dma_addr padding 113 * @dma_addr: buffer address 121 dma_addr_t dma_addr; member in union:vdec_av1_slice_mem::__anon330 1000 vsi->cdf_tmp.buf = instance->cdf_temp.dma_addr; 1592 vsi->bs.dma_addr = bs->dma_addr; 1595 vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr; 1597 vsi->trans.dma_addr = lat_buf->ctx->msg_queue.wdma_wptr_addr; 1600 vsi->err_map.dma_addr [all...] |
/linux-master/drivers/crypto/intel/keembay/ |
H A D | keembay-ocs-aes-core.c | 246 rctx->src_dll.dma_addr = DMA_MAPPING_ERROR; 247 rctx->dst_dll.dma_addr = DMA_MAPPING_ERROR; 248 rctx->aad_src_dll.dma_addr = DMA_MAPPING_ERROR; 249 rctx->aad_dst_dll.dma_addr = DMA_MAPPING_ERROR; 368 dma_free_coherent(dev, dll->size, dll->vaddr, dll->dma_addr); 371 dll->dma_addr = DMA_MAPPING_ERROR; 438 * If descriptor creation was successful, set the src_dll.dma_addr to 439 * the value of dst_dll.dma_addr, as we do in-place AES operation on 442 rctx->src_dll.dma_addr = rctx->dst_dll.dma_addr; [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_user_sdma.c | 259 dma_addr_t dma_addr, u16 dma_length) 269 pkt->addr[i].addr = dma_addr; 274 size_t len, dma_addr_t *dma_addr) 280 dma_addr); 289 *dma_addr = 0; 308 dma_addr_t dma_addr = local 313 if (dma_mapping_error(&dd->pcidev->dev, dma_addr)) { 368 dma_addr, len); /* dma addr, dma length */ 809 dma_addr_t dma_addr; local 836 pbc = qib_user_sdma_alloc_header(pq, len, &dma_addr); 254 qib_user_sdma_init_frag(struct qib_user_sdma_pkt *pkt, int i, u16 offset, u16 len, u16 first_desc, u16 last_desc, u16 put_page, u16 dma_mapped, struct page *page, void *kvaddr, dma_addr_t dma_addr, u16 dma_length) argument 273 qib_user_sdma_alloc_header(struct qib_user_sdma_queue *pq, size_t len, dma_addr_t *dma_addr) argument [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | dma.c | 62 t->dma_addr = addr; 166 dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size, 272 entry->dma_addr[0] = buf->addr; 312 entry->dma_addr[0] = buf[0].addr; 320 entry->dma_addr[1] = buf[1].addr; 360 dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0], 364 dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1], 453 dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, 458 t->dma_addr = 0; 465 dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[ [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_tx.c | 181 dma_addr_t dma_addr = 0; local 186 dma_addr = dma_map_single(sq->pdev, skb_data, headlen, 188 if (unlikely(dma_mapping_error(sq->pdev, dma_addr))) 191 dseg->addr = cpu_to_be64(dma_addr); 195 mlx5e_dma_push(sq, dma_addr, headlen, MLX5E_DMA_MAP_SINGLE); 204 dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz, 206 if (unlikely(dma_mapping_error(sq->pdev, dma_addr))) 209 dseg->addr = cpu_to_be64(dma_addr); 213 mlx5e_dma_push(sq, dma_addr, fsz, MLX5E_DMA_MAP_PAGE); 558 dseg->addr = cpu_to_be64(txd->dma_addr); [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 109 otx2_dma_unmap_page(pfvf, sg->dma_addr[seg], 126 pa = otx2_iova_to_phys(pfvf->iommu_domain, sg->dma_addr[0]); 127 otx2_dma_unmap_page(pfvf, sg->dma_addr[0], 618 u64 dma_addr, *iova = NULL; local 640 dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len); 641 if (dma_mapping_error(pfvf->dev, dma_addr)) 646 *iova++ = dma_addr; 649 sq->sg[sq->head].dma_addr[seg] = dma_addr; 804 u64 dma_addr; local 1347 otx2_xdp_sqe_add_sg(struct otx2_snd_queue *sq, u64 dma_addr, int len, int *offset) argument [all...] |