/linux-master/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | bcmsdh.c | 755 uint nents; local 771 nents = max_t(uint, BRCMF_DEFAULT_RXGLOM_SIZE, 773 nents += (nents >> 4) + 1; 775 WARN_ON(nents > sdiodev->max_segment_count); 777 brcmf_dbg(TRACE, "nents=%d\n", nents); 778 err = sg_alloc_table(&sdiodev->sgtable, nents, GFP_KERNEL);
|
/linux-master/crypto/ |
H A D | af_alg.c | 540 sg_unmark_end(sgl_prev->sgt.sgl + sgl_prev->sgt.nents - 1); 541 sg_chain(sgl_prev->sgt.sgl, sgl_prev->sgt.nents + 1, sgl_new->sgt.sgl); 550 for (i = 0; i < sgl->sgt.nents; i++) 1044 .nents = sgl->cur, 1055 for (; sgl->cur < sgtable.nents; sgl->cur++) 1264 rsgl->sgl.sgt.nents = 0; 1276 sg_mark_end(rsgl->sgl.sgt.sgl + rsgl->sgl.sgt.nents - 1);
|
H A D | testmgr.c | 548 unsigned int nents; member in struct:test_sglist 599 tsgl->nents = 0; 607 partitions[tsgl->nents].div = &divs[i]; 608 partitions[tsgl->nents].length = len_this_sg; 609 tsgl->nents++; 613 if (tsgl->nents == 0) { 614 partitions[tsgl->nents].div = &divs[0]; 615 partitions[tsgl->nents].length = 0; 616 tsgl->nents++; 618 partitions[tsgl->nents [all...] |
H A D | algif_hash.c | 107 ctx->sgl.sgt.nents = 0; 124 sg_mark_end(ctx->sgl.sgt.sgl + ctx->sgl.sgt.nents - 1);
|
/linux-master/drivers/crypto/chelsio/ |
H A D | chcr_algo.c | 138 int nents = 0; local 155 nents += DIV_ROUND_UP(less, entlen); 160 return nents; 379 walk->nents = 0; 398 CPL_RX_PHYS_DSGL_NOOFSGENTR_V(walk->nents)); 413 j = walk->nents; 419 walk->nents = j; 429 unsigned int j = walk->nents; 466 walk->nents = j; 473 walk->nents 812 int nents; local 1571 unsigned int nents = 0, transhdr_len; local [all...] |
/linux-master/drivers/fpga/ |
H A D | zynq-fpga.c | 400 for_each_sg(sgt->sgl, sg, sgt->nents, i) { 409 dma_map_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE); 481 dma_unmap_sg(mgr->dev.parent, sgt->sgl, sgt->nents, DMA_TO_DEVICE);
|
/linux-master/drivers/hsi/ |
H A D | hsi_core.c | 527 * @nents: Number of memory entries 530 * nents can be 0. This mainly makes sense for read transfer. 536 struct hsi_msg *hsi_alloc_msg(unsigned int nents, gfp_t flags) argument 545 if (!nents) 548 err = sg_alloc_table(&msg->sgt, nents, flags); 564 * fields set beforehand. If nents > 0 then the client has to initialize
|
/linux-master/drivers/crypto/marvell/cesa/ |
H A D | tdma.c | 355 struct scatterlist *sgl, unsigned int nents, 368 sg_miter_start(&miter, sgl, nents, sg_flags); 354 mv_cesa_sg_copy(struct mv_cesa_engine *engine, struct scatterlist *sgl, unsigned int nents, unsigned int sram_off, size_t buflen, off_t skip, bool to_sram) argument
|
/linux-master/drivers/gpu/drm/virtio/ |
H A D | virtgpu_vq.c | 629 uint32_t nents, 640 cmd_p->nr_entries = cpu_to_le32(nents); 643 vbuf->data_size = sizeof(*ents) * nents; 1104 unsigned int nents) 1107 ents, nents, NULL); 1246 uint32_t nents) 1261 cmd_p->nr_entries = cpu_to_le32(nents); 1264 vbuf->data_size = sizeof(*ents) * nents; 626 virtio_gpu_cmd_resource_attach_backing(struct virtio_gpu_device *vgdev, uint32_t resource_id, struct virtio_gpu_mem_entry *ents, uint32_t nents, struct virtio_gpu_fence *fence) argument 1101 virtio_gpu_object_attach(struct virtio_gpu_device *vgdev, struct virtio_gpu_object *obj, struct virtio_gpu_mem_entry *ents, unsigned int nents) argument 1242 virtio_gpu_cmd_resource_create_blob(struct virtio_gpu_device *vgdev, struct virtio_gpu_object *bo, struct virtio_gpu_object_params *params, struct virtio_gpu_mem_entry *ents, uint32_t nents) argument
|
H A D | virtgpu_vram.c | 119 if (sgt->nents) {
|
/linux-master/drivers/dma-buf/ |
H A D | udmabuf.c | 162 dma_sync_sg_for_cpu(dev, ubuf->sg->sgl, ubuf->sg->nents, 178 dma_sync_sg_for_device(dev, ubuf->sg->sgl, ubuf->sg->nents, direction);
|
/linux-master/include/rdma/ |
H A D | ib_verbs.h | 2865 unsigned int nents, 2884 * @nents: maximum number of sg entries to iterate over 2890 #define rdma_for_each_block(sglist, biter, nents, pgsz) \ 2891 for (__rdma_block_iter_start(biter, sglist, nents, \ 4128 int ib_dma_virt_map_sg(struct ib_device *dev, struct scatterlist *sg, int nents); 4130 struct scatterlist *sg, int nents, 4135 return ib_dma_virt_map_sg(dev, sg, nents); 4136 return dma_map_sg_attrs(dev->dma_device, sg, nents, direction, 4141 struct scatterlist *sg, int nents, 4146 dma_unmap_sg_attrs(dev->dma_device, sg, nents, directio 4129 ib_dma_map_sg_attrs(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction, unsigned long dma_attrs) argument 4140 ib_dma_unmap_sg_attrs(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction, unsigned long dma_attrs) argument 4162 int nents; local 4190 ib_dma_map_sg(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument 4204 ib_dma_unmap_sg(struct ib_device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma.c | 981 st->nents++; 1000 st->nents++; 1036 st->nents = 0; 1063 st->nents++; 1111 st->nents++; 1159 st->nents++; 1246 st->nents = 0; 1284 st->nents = 0;
|
/linux-master/fs/smb/client/ |
H A D | cifsglob.h | 2206 unsigned int nents = 0; local 2247 nents += DIV_ROUND_UP(offset_in_page(addr) + len, 2250 nents++; 2255 nents += iov_iter_npages(&rqst[i].rq_iter, INT_MAX); 2257 nents += DIV_ROUND_UP(offset_in_page(sig) + SMB2_SIGNATURE_SIZE, PAGE_SIZE); 2258 return nents; 2276 sg_set_page(&sgtable->sgl[sgtable->nents++], 2284 sg_set_page(&sgtable->sgl[sgtable->nents++],
|
/linux-master/arch/arm/mm/ |
H A D | dma-mapping.c | 1232 * @nents: number of buffers to map 1241 int nents, enum dma_data_direction dir, unsigned long attrs) 1249 for (i = 1; i < nents; i++) { 1291 * @nents: number of buffers to unmap (same as was passed to dma_map_sg) 1298 struct scatterlist *sg, int nents, 1305 for_each_sg(sg, s, nents, i) { 1319 * @nents: number of buffers to map (returned from dma_map_sg) 1324 int nents, enum dma_data_direction dir) 1332 for_each_sg(sg, s, nents, i) 1341 * @nents 1240 arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir, unsigned long attrs) argument 1297 arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir, unsigned long attrs) argument 1322 arm_iommu_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) argument 1344 arm_iommu_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) argument [all...] |
/linux-master/drivers/scsi/be2iscsi/ |
H A D | be_mgmt.h | 168 unsigned int nents);
|
/linux-master/drivers/net/ethernet/chelsio/libcxgb/ |
H A D | libcxgb_ppm.h | 93 int nents; member in struct:cxgbi_task_tag_info
|
/linux-master/drivers/char/agp/ |
H A D | intel-gtt.c | 114 if (!dma_map_sg(&intel_private.pcidev->dev, st->sgl, st->nents, 134 st.orig_nents = st.nents = num_sg; 867 for_each_sg(st->sgl, sg, st->nents, i) { 936 mem->num_sg = st.nents;
|
/linux-master/drivers/media/platform/nvidia/tegra-vde/ |
H A D | dmabuf-cache.c | 112 if (!vde->domain && sgt->nents > 1) {
|
/linux-master/drivers/infiniband/core/ |
H A D | umem_dmabuf.c | 63 umem_dmabuf->umem.sgt_append.sgt.nents = nmap;
|
H A D | umem.c | 119 if (i != (umem->sgt_append.sgt.nents - 1))
|
/linux-master/drivers/media/pci/saa7134/ |
H A D | saa7134-vbi.c | 124 return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
|
/linux-master/drivers/crypto/ |
H A D | img-hash.c | 97 size_t nents; member in struct:img_hash_request_ctx 405 ctx->bufcnt = sg_pcopy_to_buffer(ctx->sgfirst, ctx->nents, 413 tbc = sg_pcopy_to_buffer(ctx->sgfirst, ctx->nents, 671 ctx->nents = sg_nents(ctx->sg);
|
/linux-master/drivers/spi/ |
H A D | spi-davinci.c | 624 t->rx_sg.sgl, t->rx_sg.nents, DMA_DEV_TO_MEM, 636 t->tx_sg.nents = t->rx_sg.nents; 640 t->tx_sg.sgl, t->tx_sg.nents, DMA_MEM_TO_DEV,
|
H A D | spi-pic32.c | 314 xfer->rx_sg.nents, 324 xfer->tx_sg.nents, 524 if (transfer->rx_sg.nents && transfer->tx_sg.nents) {
|