/linux-master/drivers/crypto/qce/ |
H A D | dma.c | 53 struct scatterlist *sg = sgt->sgl, *sg_last = NULL;
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmnv41.c | 71 .sgl = nv41_vmm_pgt_sgl,
|
/linux-master/kernel/dma/ |
H A D | ops_helpers.c | 27 sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
|
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_lio.c | 423 struct efct_scsi_sgl *sgl = io->sgl; local 448 sgl[cnt].addr = sg_dma_address(sg); 449 sgl[cnt].dif_addr = 0; 450 sgl[cnt].len = sg_dma_len(sg); 451 length += sgl[cnt].len; 458 return efct_scsi_recv_wr_data(io, flags, sgl, curcnt, length, 468 struct efct_scsi_sgl *sgl = io->sgl; local 500 sgl[cn [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | r535.c | 95 struct scatterlist *sgl; local 98 for_each_sgtable_dma_sg(sgt, sgl, idx) { 99 for (int i = 0; i < sg_dma_len(sgl) / GSP_PAGE_SIZE; i++) 100 rpc->pteDesc.pte_pde[pte++].pte = (sg_dma_address(sgl) >> 12) + i;
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | mock_dmabuf.c | 26 sg = st->sgl;
|
/linux-master/arch/mips/jazz/ |
H A D | jazzdma.c | 588 struct scatterlist *sgl, int nents, enum dma_data_direction dir) 593 for_each_sg(sgl, sg, nents, i) 598 struct scatterlist *sgl, int nents, enum dma_data_direction dir) 603 for_each_sg(sgl, sg, nents, i) 587 jazz_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction dir) argument 597 jazz_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction dir) argument
|
/linux-master/drivers/scsi/ |
H A D | myrs.c | 163 union myrs_sgl *sgl; local 186 sgl = &mbox->ctlr_info.dma_addr; 187 sgl->sge[0].sge_addr = ctlr_info_addr; 188 sgl->sge[0].sge_count = mbox->ctlr_info.dma_size; 226 union myrs_sgl *sgl; local 245 sgl = &mbox->ldev_info.dma_addr; 246 sgl->sge[0].sge_addr = ldev_info_addr; 247 sgl->sge[0].sge_count = mbox->ldev_info.dma_size; 313 union myrs_sgl *sgl; local 333 sgl 383 union myrs_sgl *sgl; local 426 union myrs_sgl *sgl; local 458 union myrs_sgl *sgl; local 1594 struct scatterlist *sgl; local [all...] |
H A D | 3w-sas.h | 245 TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; member in struct:TW_Command::__anon2502::__anon2503 249 TW_SG_Entry_ISO sgl[TW_LIBERATOR_MAX_SGL_LENGTH_OLD]; member in struct:TW_Command::__anon2502::__anon2504
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gpu_error.c | 92 struct scatterlist *sgl; local 94 sgl = (typeof(sgl))__get_free_page(ALLOW_FAIL); 95 if (!sgl) { 104 (unsigned long)sgl | SG_CHAIN; 106 e->sgl = sgl; 109 e->cur = sgl; 110 e->end = sgl + SG_MAX_SINGLE_ALLOC - 1; 707 static void err_free_sgl(struct scatterlist *sgl) argument [all...] |
H A D | i915_gem_gtt.c | 33 pages->sgl, pages->nents, 67 dma_unmap_sg(i915->drm.dev, pages->sgl, pages->nents,
|
/linux-master/drivers/hwtracing/intel_th/ |
H A D | msu.c | 280 return win->sgt->sgl; 330 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 425 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 662 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 932 sg_set_buf(msc->single_sgt.sgl, page_address(page), size); 934 ret = dma_map_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl, 1, 941 msc->base_addr = sg_dma_address(msc->single_sgt.sgl); 963 dma_unmap_sg(msc_dev(msc)->parent->parent, msc->single_sgt.sgl, 1004 for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { 1017 for_each_sg(win->sgt->sgl, sg_pt [all...] |
/linux-master/drivers/iommu/ |
H A D | dma-iommu.c | 994 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) 998 ret = iommu_map_sg(domain, iova, sgt->sgl, sgt->orig_nents, ioprot, 1003 sgt->sgl->dma_address = iova; 1004 sgt->sgl->dma_length = size; 1028 *dma_handle = sgt.sgl->dma_address; 1066 __iommu_dma_unmap(dev, sgt->sgl->dma_address, size); 1105 struct scatterlist *sgl, int nelems, 1111 if (sg_dma_is_swiotlb(sgl)) 1112 for_each_sg(sgl, sg, nelems, i) 1116 for_each_sg(sgl, s 1104 iommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nelems, enum dma_data_direction dir) argument 1120 iommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nelems, enum dma_data_direction dir) argument [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 810 * @sgl: starting location for writing the SGL 818 * main body except for the first @start bytes. @sgl must be 16-byte 821 * wrap around, i.e., @end > @sgl. 824 struct ulptx_sgl *sgl, u64 *end, unsigned int start, 835 sgl->len0 = htonl(len); 836 sgl->addr0 = cpu_to_be64(addr[0] + start); 839 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); 840 sgl->addr0 = cpu_to_be64(addr[1]); 843 sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | 852 to = (u8 *)end > (u8 *)q->stat ? buf : sgl 823 cxgb4_write_sgl(const struct sk_buff *skb, struct sge_txq *q, struct ulptx_sgl *sgl, u64 *end, unsigned int start, const dma_addr_t *addr) argument 892 cxgb4_write_partial_sgl(const struct sk_buff *skb, struct sge_txq *q, struct ulptx_sgl *sgl, u64 *end, const dma_addr_t *addr, u32 start, u32 len) argument 1501 u64 cntrl, *end, *sgl; local 2050 struct ulptx_sgl *sgl = (struct ulptx_sgl *)(cpl + 1); local 2255 u64 cntrl, *start, *end, *sgl; local 2661 u8 *sgl; local [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_db.c | 44 db->dma = sg_dma_address(page->umem->sgt_append.sgt.sgl) + offset; 45 db->virt_addr = sg_virt(page->umem->sgt_append.sgt.sgl) + offset;
|
/linux-master/drivers/media/platform/nvidia/tegra-vde/ |
H A D | dmabuf-cache.c | 93 *addrp = sg_dma_address(entry->sgt->sgl); 131 *addrp = sg_dma_address(sgt->sgl);
|
/linux-master/drivers/crypto/hisilicon/sec/ |
H A D | sec_algs.c | 176 struct scatterlist *sgl, 190 for_each_sg(sgl, sg, count, i) { 568 static int sec_map_and_split_sg(struct scatterlist *sgl, size_t *split_sizes, argument 576 count = dma_map_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); 592 ret = sg_split(sgl, count, 0, steps, split_sizes, 606 dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); 615 static void sec_unmap_sg_on_err(struct scatterlist *sgl, int steps, argument 626 dma_unmap_sg(dev, sgl, sgl_len_in, DMA_BIDIRECTIONAL); 174 sec_alloc_and_fill_hw_sgl(struct sec_hw_sgl **sec_sgl, dma_addr_t *psec_sgl, struct scatterlist *sgl, int count, struct sec_dev_info *info, gfp_t gfp) argument
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_mem.c | 113 args.sgl = tt->sg->sgl;
|
/linux-master/drivers/media/pci/saa7134/ |
H A D | saa7134-vbi.c | 114 if (dma->sgl->offset) { 124 return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
|
/linux-master/include/linux/ |
H A D | dmaengine.h | 140 * @src_sgl: If the 'icg' of sgl[] applies to Source (scattered read). 143 * @dst_sgl: If the 'icg' of sgl[] applies to Destination (scattered write). 147 * @frame_size: Number of chunks in a frame i.e, size of sgl[]. 148 * @sgl: Array of {chunk,icg} pairs that make up a frame. 160 struct data_chunk sgl[]; member in struct:dma_interleaved_template 914 struct dma_chan *chan, struct scatterlist *sgl, 977 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, 983 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, 990 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, 997 return chan->device->device_prep_slave_sg(chan, sgl, sg_le 976 dmaengine_prep_slave_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags) argument 989 dmaengine_prep_rio_sg( struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, struct rio_dma_ext *rio_ext) argument 1300 dmaengine_get_icg(bool inc, bool sgl, size_t icg, size_t dir_icg) argument [all...] |
/linux-master/net/smc/ |
H A D | smc_ib.c | 708 buf_slot->sgt[link_idx].sgl, 747 for_each_sg(buf_slot->sgt[lnk->link_idx].sgl, sg, 774 for_each_sg(buf_slot->sgt[lnk->link_idx].sgl, sg, 797 for_each_sg(buf_slot->sgt[lnk->link_idx].sgl, sg, 816 buf_slot->sgt[lnk->link_idx].sgl, 829 if (!buf_slot->sgt[lnk->link_idx].sgl->dma_address) 833 buf_slot->sgt[lnk->link_idx].sgl, 836 buf_slot->sgt[lnk->link_idx].sgl->dma_address = 0;
|
/linux-master/include/scsi/ |
H A D | scsi_cmnd.h | 188 return cmd->sdb.table.sgl; 323 return cmd->prot_sdb ? cmd->prot_sdb->table.sgl : NULL;
|
/linux-master/drivers/spi/ |
H A D | spi-pl022.c | 686 dma_unmap_sg(pl022->dma_tx_channel->device->dev, pl022->sgt_tx.sgl, 688 dma_unmap_sg(pl022->dma_rx_channel->device->dev, pl022->sgt_rx.sgl, 698 BUG_ON(!pl022->sgt_rx.sgl); 712 pl022->sgt_rx.sgl, 716 for_each_sg(pl022->sgt_rx.sgl, sg, pl022->sgt_rx.nents, i) { 726 for_each_sg(pl022->sgt_tx.sgl, sg, pl022->sgt_tx.nents, i) { 756 for_each_sg(sgtab->sgl, sg, sgtab->nents, i) { 777 for_each_sg(sgtab->sgl, sg, sgtab->nents, i) { 931 rx_sglen = dma_map_sg(rxchan->device->dev, pl022->sgt_rx.sgl, 936 tx_sglen = dma_map_sg(txchan->device->dev, pl022->sgt_tx.sgl, [all...] |
/linux-master/drivers/crypto/marvell/cesa/ |
H A D | cesa.h | 875 struct scatterlist *sgl, unsigned int nents, 880 struct scatterlist *sgl, 885 return mv_cesa_sg_copy(engine, sgl, nents, sram_off, buflen, skip, 890 struct scatterlist *sgl, 895 return mv_cesa_sg_copy(engine, sgl, nents, sram_off, buflen, skip, 879 mv_cesa_sg_copy_to_sram(struct mv_cesa_engine *engine, struct scatterlist *sgl, unsigned int nents, unsigned int sram_off, size_t buflen, off_t skip) argument 889 mv_cesa_sg_copy_from_sram(struct mv_cesa_engine *engine, struct scatterlist *sgl, unsigned int nents, unsigned int sram_off, size_t buflen, off_t skip) argument
|
/linux-master/drivers/gpu/drm/armada/ |
H A D | armada_gem.c | 426 sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); 434 sg_dma_address(sgt->sgl) = dobj->dev_addr; 435 sg_dma_len(sgt->sgl) = dobj->obj.size; 554 if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) { 558 dobj->dev_addr = sg_dma_address(dobj->sgt->sgl);
|