Searched refs:sgl (Results 1 - 25 of 504) sorted by relevance

1234567891011>>

/linux-master/drivers/misc/genwqe/
H A Dcard_utils.c284 * genwqe_alloc_sync_sgl() - Allocate memory for sgl and overlapping pages
286 * Allocates memory for sgl and overlapping pages. Pages which might
291 int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, argument
297 sgl->fpage_offs = offset_in_page((unsigned long)user_addr);
298 sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size);
299 sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE);
300 sgl->lpage_size = (user_size - sgl
374 genwqe_setup_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, dma_addr_t *dma_list) argument
473 genwqe_free_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl) argument
[all...]
/linux-master/drivers/crypto/hisilicon/
H A Dsgl.c23 /* use default sgl head size 64B */
37 struct hisi_acc_hw_sgl *sgl; member in struct:hisi_acc_sgl_pool::mem_block
49 * hisi_acc_create_sgl_pool() - Create a hw sgl pool.
50 * @dev: The device which hw sgl pool belongs to.
54 * This function creates a hw sgl pool, after this user can get hw sgl memory
92 block[i].sgl = dma_alloc_coherent(dev, block_size,
95 if (!block[i].sgl) {
104 block[i].sgl = dma_alloc_coherent(dev, remain_sgl * sgl_size,
107 if (!block[i].sgl) {
175 sg_map_to_hw_sg(struct scatterlist *sgl, struct acc_hw_sge *hw_sge) argument
221 hisi_acc_sg_buf_map_to_hw_sgl(struct device *dev, struct scatterlist *sgl, struct hisi_acc_sgl_pool *pool, u32 index, dma_addr_t *hw_sgl_dma) argument
279 hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl, struct hisi_acc_hw_sgl *hw_sgl) argument
[all...]
H A DMakefile6 hisi_qm-objs = qm.o sgl.o debugfs.o
/linux-master/mm/
H A Dpage_reporting.c106 struct scatterlist *sgl, unsigned int nents, bool reported)
108 struct scatterlist *sg = sgl;
137 sg_init_table(sgl, nents);
148 struct scatterlist *sgl, unsigned int *offset)
208 sg_set_page(&sgl[*offset], page, page_len, 0);
225 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY);
237 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err);
261 struct scatterlist *sgl, struct zone *zone)
286 sgl, &offset);
295 sgl
105 page_reporting_drain(struct page_reporting_dev_info *prdev, struct scatterlist *sgl, unsigned int nents, bool reported) argument
146 page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, unsigned int order, unsigned int mt, struct scatterlist *sgl, unsigned int *offset) argument
260 page_reporting_process_zone(struct page_reporting_dev_info *prdev, struct scatterlist *sgl, struct zone *zone) argument
313 struct scatterlist *sgl; local
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_res_cursor.h47 struct scatterlist *sgl; member in struct:xe_res_cursor
73 cur->sgl = NULL;
131 struct scatterlist *sgl = cur->sgl; local
134 while (start >= sg_dma_len(sgl)) {
135 start -= sg_dma_len(sgl);
136 sgl = sg_next(sgl);
137 XE_WARN_ON(!sgl);
141 cur->size = sg_dma_len(sgl)
[all...]
/linux-master/include/linux/
H A Dscatterlist.h40 struct scatterlist *sgl; /* the list */ member in struct:sg_table
202 for_each_sg((sgt)->sgl, sg, (sgt)->orig_nents, i)
210 for_each_sg((sgt)->sgl, sg, (sgt)->nents, i)
213 struct scatterlist *sgl)
225 chain_sg->page_link = ((unsigned long) sgl | SG_CHAIN) & ~SG_END;
232 * @sgl: Second scatterlist
235 * Links @prv@ and @sgl@ together, to form a longer scatterlist.
239 struct scatterlist *sgl)
241 __sg_chain(&prv[prv_nents - 1], sgl);
409 * @sgl
212 __sg_chain(struct scatterlist *chain_sg, struct scatterlist *sgl) argument
238 sg_chain(struct scatterlist *prv, unsigned int prv_nents, struct scatterlist *sgl) argument
413 sg_init_marker(struct scatterlist *sgl, unsigned int nents) argument
[all...]
/linux-master/drivers/gpu/drm/nouveau/include/nvif/
H A Dif000a.h20 struct scatterlist *sgl; member in struct:nvif_mem_ram_v0
/linux-master/crypto/
H A Daf_alg.c539 sg_unmark_end(sgl_prev->sgt.sgl + sgl_prev->sgt.nents - 1);
540 sg_chain(sgl_prev->sgt.sgl, sgl_prev->sgt.nents + 1, sgl_new->sgt.sgl);
543 void af_alg_free_sg(struct af_alg_sgl *sgl) argument
547 if (sgl->sgt.sgl) {
548 if (sgl->need_unpin)
549 for (i = 0; i < sgl->sgt.nents; i++)
550 unpin_user_page(sg_page(&sgl->sgt.sgl[
608 struct af_alg_tsgl *sgl; local
649 const struct af_alg_tsgl *sgl; local
706 struct af_alg_tsgl *sgl; local
943 struct af_alg_tsgl *sgl; local
[all...]
H A Dalgif_hash.c20 struct af_alg_sgl sgl; member in struct:hash_ctx
106 ctx->sgl.sgt.sgl = ctx->sgl.sgl;
107 ctx->sgl.sgt.nents = 0;
108 ctx->sgl.sgt.orig_nents = 0;
115 sg_init_table(ctx->sgl.sgl, npages);
117 ctx->sgl
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/
H A Dfirmware.c121 case NVKM_FIRMWARE_IMG_DMA: return &fw->mem.sgl;
122 case NVKM_FIRMWARE_IMG_SGT: return fw->mem.sgt.sgl;
139 .sgl = nvkm_firmware_mem_sgl(memory),
142 if (!map.sgl)
151 struct scatterlist *sgl = nvkm_firmware_mem_sgl(memory); local
153 return sgl ? sg_dma_len(sgl) : 0;
208 dma_free_coherent(fw->device->dev, sg_dma_len(&fw->mem.sgl), fw->img, fw->phys);
248 sg_init_one(&fw->mem.sgl, fw->img, len);
249 sg_dma_address(&fw->mem.sgl)
265 struct scatterlist *sgl; local
[all...]
/linux-master/tools/virtio/linux/
H A Dscatterlist.h85 * @sgl: Second scatterlist
88 * Links @prv@ and @sgl@ together, to form a longer scatterlist.
92 struct scatterlist *sgl)
104 prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02;
150 static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents) argument
152 memset(sgl, 0, sizeof(*sgl) * nents);
153 sg_mark_end(&sgl[nents - 1]);
91 sg_chain(struct scatterlist *prv, unsigned int prv_nents, struct scatterlist *sgl) argument
/linux-master/lib/
H A Dscatterlist.c91 * @sgl: First entry in the scatterlist
98 * Note that the @sgl@ pointer passed in need not be the first one,
100 * exist from @sgl@.
103 struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents) argument
108 for_each_sg(sgl, sg, nents, i)
118 * @sgl: The SG table
126 void sg_init_table(struct scatterlist *sgl, unsigned int nents) argument
128 memset(sgl, 0, sizeof(*sgl) * nents);
129 sg_init_marker(sgl, nent
199 struct scatterlist *sgl, *next; local
615 struct scatterlist *sgl, *sg; local
685 sgl_free_n_order(struct scatterlist *sgl, int nents, int order) argument
707 sgl_free_order(struct scatterlist *sgl, int order) argument
717 sgl_free(struct scatterlist *sgl) argument
800 sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl, unsigned int nents, unsigned int flags) argument
962 sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, size_t buflen, off_t skip, bool to_buffer) argument
1008 sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents, const void *buf, size_t buflen) argument
1025 sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, size_t buflen) argument
1043 sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents, const void *buf, size_t buflen, off_t skip) argument
1061 sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents, void *buf, size_t buflen, off_t skip) argument
1077 sg_zero_buffer(struct scatterlist *sgl, unsigned int nents, size_t buflen, off_t skip) argument
[all...]
H A Dsg_pool.c54 static void sg_pool_free(struct scatterlist *sgl, unsigned int nents) argument
59 mempool_free(sgl, sgp->pool);
122 sg_init_table(table->sgl, nents);
/linux-master/drivers/gpu/drm/i915/
H A Di915_mm.h33 struct scatterlist *sgl, resource_size_t iobase);
H A Di915_scatterlist.h31 } __sgt_iter(struct scatterlist *sgl, bool dma) { argument
32 struct sgt_iter s = { .sgp = sgl };
90 for ((__iter) = __sgt_iter((__sgt)->sgl, true); \
112 for ((__iter) = __sgt_iter((__sgt)->sgl, false); \
179 * @table->sgl = NULL can be used to determine whether a scatter-gather table
225 rsgt->table.sgl = NULL;
/linux-master/drivers/target/iscsi/cxgbit/
H A Dcxgbit_ddp.c133 struct scatterlist *sg = ttinfo->sgl;
176 struct scatterlist *sgl = ttinfo->sgl; local
178 unsigned int sg_offset = sgl->offset;
188 if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0)
191 ttinfo->nr_pages = (xferlen + sgl->offset +
203 sgl->offset = 0;
204 ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE);
205 sgl->offset = sg_offset;
207 pr_debug("%s: 0x%x, xfer %u, sgl
[all...]
/linux-master/drivers/scsi/qedf/
H A Ddrv_scsi_fw_funcs.c18 /* no need to check for sgl_task_params->sgl validity */
34 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.lo);
36 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_addr.hi);
38 val = cpu_to_le32(sgl_task_params->sgl[sge_index].sge_len);
/linux-master/kernel/dma/
H A Ddirect.h21 int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents,
28 void dma_direct_sync_sg_for_device(struct device *dev, struct scatterlist *sgl,
32 struct scatterlist *sgl, int nents, enum dma_data_direction dir)
40 void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl,
43 struct scatterlist *sgl, int nents, enum dma_data_direction dir);
46 struct scatterlist *sgl, int nents, enum dma_data_direction dir,
51 struct scatterlist *sgl, int nents, enum dma_data_direction dir)
31 dma_direct_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction dir) argument
45 dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction dir, unsigned long attrs) argument
50 dma_direct_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction dir) argument
H A Ddummy.c21 static int dma_dummy_map_sg(struct device *dev, struct scatterlist *sgl, argument
/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dqat_bl.c51 struct scatterlist *sgl,
62 int n = sg_nents(sgl);
89 bufl_dma_dir = sgl != sglout ? DMA_TO_DEVICE : DMA_BIDIRECTIONAL;
96 for_each_sg(sgl, sg, n, i) {
126 if (sgl != sglout) {
215 n = sg_nents(sgl);
230 struct scatterlist *sgl,
248 return __qat_bl_sgl_to_bufl(accel_dev, sgl, sglout, buf,
267 struct scatterlist *sgl,
277 n = sg_nents(sgl);
50 __qat_bl_sgl_to_bufl(struct adf_accel_dev *accel_dev, struct scatterlist *sgl, struct scatterlist *sglout, struct qat_request_buffs *buf, dma_addr_t extra_dst_buff, size_t sz_extra_dst_buff, unsigned int sskip, unsigned int dskip, gfp_t flags) argument
229 qat_bl_sgl_to_bufl(struct adf_accel_dev *accel_dev, struct scatterlist *sgl, struct scatterlist *sglout, struct qat_request_buffs *buf, struct qat_sgl_to_bufl_params *params, gfp_t flags) argument
266 qat_bl_sgl_map(struct adf_accel_dev *accel_dev, struct scatterlist *sgl, struct qat_alg_buf_list **bl) argument
320 qat_bl_sgl_free_unmap(struct adf_accel_dev *accel_dev, struct scatterlist *sgl, struct qat_alg_buf_list *bl, bool free_bl) argument
335 qat_bl_sgl_alloc_map(struct adf_accel_dev *accel_dev, struct scatterlist **sgl, struct qat_alg_buf_list **bl, unsigned int dlen, gfp_t gfp) argument
[all...]
/linux-master/drivers/scsi/qedi/
H A Dqedi_fw_scsi.h16 struct scsi_sge *sgl; member in struct:scsi_sgl_task_params
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c50 sg = pages->sgl;
64 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) {
68 src = pages->sgl;
/linux-master/drivers/vfio/pci/pds/
H A Dcmds.c206 struct pds_lm_sg_elem *sgl, *sge; local
223 /* alloc sgl */
225 sgl = kzalloc(sgl_size, GFP_KERNEL);
226 if (!sgl) {
231 /* fill sgl */
232 sge = sgl;
240 sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE);
246 lm_file->sgl = sgl;
252 kfree(sgl);
[all...]
/linux-master/drivers/spi/
H A Dspi-bcm2835.c497 if (bs->tx_buf && !sg_is_last(&tfr->tx_sg.sgl[0]))
498 bs->tx_prologue = sg_dma_len(&tfr->tx_sg.sgl[0]) & 3;
500 if (bs->rx_buf && !sg_is_last(&tfr->rx_sg.sgl[0])) {
501 bs->rx_prologue = sg_dma_len(&tfr->rx_sg.sgl[0]) & 3;
504 if (!bs->tx_buf || sg_is_last(&tfr->tx_sg.sgl[0])) {
509 !(sg_dma_len(&tfr->tx_sg.sgl[0]) & ~3);
531 sg_dma_address(&tfr->rx_sg.sgl[0]),
534 sg_dma_address(&tfr->rx_sg.sgl[0]) += bs->rx_prologue;
535 sg_dma_len(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue;
557 sg_dma_address(&tfr->tx_sg.sgl[
678 struct scatterlist *sgl; local
[all...]
/linux-master/include/linux/mtd/
H A Dubi.h242 int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl,
270 struct ubi_sgl *sgl, int offset, int len)
272 return ubi_leb_read_sg(desc, lnum, sgl, offset, len, 0);
269 ubi_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl, int offset, int len) argument

Completed in 494 milliseconds

1234567891011>>