Searched refs:sg (Results 376 - 400 of 900) sorted by relevance

<<11121314151617181920>>

/linux-master/include/linux/mtd/
H A Dubi.h118 * @list_pos: current position in @sg[]
119 * @page_pos: current position in @sg[@list_pos]
120 * @sg: the scatter gather list itself
129 struct scatterlist sg[UBI_MAX_SG_COUNT]; member in struct:ubi_sgl
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_blit.c482 if (!src->ttm->pages && src->ttm->sg) {
487 ret = drm_prime_sg_to_page_array(src->ttm->sg, src_pages,
492 if (!dst->ttm->pages && dst->ttm->sg) {
499 ret = drm_prime_sg_to_page_array(dst->ttm->sg, dst_pages,
/linux-master/net/tls/
H A Dtls_device_fallback.c40 static void chain_to_walk(struct scatterlist *sg, struct scatter_walk *walk) argument
42 struct scatterlist *src = walk->sg;
45 sg_set_page(sg, sg_page(src),
48 scatterwalk_crypto_chain(sg, sg_next(src), 2);
/linux-master/drivers/iommu/
H A Dvirtio-iommu.c224 struct scatterlist *sg[2] = { &top_sg, &bottom_sg }; local
247 ret = virtqueue_add_sgs(vq, sg, 1, 1, req, GFP_ATOMIC);
251 ret = virtqueue_add_sgs(vq, sg, 1, 1, req, GFP_ATOMIC);
616 struct scatterlist sg[1]; local
629 sg_init_one(sg, evt, sizeof(*evt));
630 ret = virtqueue_add_inbuf(vq, sg, 1, evt, GFP_ATOMIC);
1110 struct scatterlist sg[1]; local
1121 sg_init_one(sg, &evts[i], sizeof(*evts));
1122 ret = virtqueue_add_inbuf(vq, sg, 1, &evts[i], GFP_KERNEL);
/linux-master/drivers/mmc/host/
H A Dwmt-sdmmc.c302 dma_unmap_sg(mmc_dev(priv->mmc), req->data->sg,
305 dma_unmap_sg(mmc_dev(priv->mmc), req->data->sg,
568 struct scatterlist *sg; local
619 sg_cnt = dma_map_sg(mmc_dev(mmc), req->data->sg,
625 sg_cnt = dma_map_sg(mmc_dev(mmc), req->data->sg,
635 for_each_sg(req->data->sg, sg, sg_cnt, i) {
637 while (offset < sg_dma_len(sg)) {
639 sg_dma_address(sg)+offset,
H A Dmmc_spi.c776 struct scatterlist *sg; local
799 for_each_sg(data->sg, sg, data->sg_len, n_sg) {
802 unsigned length = sg->length;
805 kmap_addr = kmap(sg_page(sg));
807 t->tx_buf = kmap_addr + sg->offset;
809 t->rx_buf = kmap_addr + sg->offset;
835 flush_dcache_page(sg_page(sg));
836 kunmap(sg_page(sg));
H A Dsdhci.c677 sg_copy_to_buffer(data->sg, data->sg_len,
691 data->sg, data->sg_len,
704 static char *sdhci_kmap_atomic(struct scatterlist *sg) argument
706 return kmap_local_page(sg_page(sg)) + sg->offset;
752 struct scatterlist *sg; local
770 for_each_sg(data->sg, sg, host->sg_count, i) {
771 addr = sg_dma_address(sg);
772 len = sg_dma_len(sg);
843 struct scatterlist *sg; local
1121 struct scatterlist *sg; local
[all...]
H A Dmxs-mmc.c146 dma_unmap_sg(mmc_dev(host->mmc), data->sg,
224 dma_map_sg(mmc_dev(host->mmc), data->sg,
226 sgl = data->sg;
241 dma_unmap_sg(mmc_dev(host->mmc), data->sg,
348 struct scatterlist *sgl = data->sg, *sg; local
391 * take special care of the case that data size from data->sg
394 for_each_sg(sgl, sg, sg_len, i)
395 data_size += sg->length;
439 /* append data sg */
[all...]
H A Dsunplus-mmc.c432 struct scatterlist *sg; local
437 count = dma_map_sg(host->mmc->parent, data->sg, data->sg_len,
444 for_each_sg(data->sg, sg, count, i) {
445 dma_addr = sg_dma_address(sg);
446 dma_size = sg_dma_len(sg) / data->blksz - 1;
637 sg_miter_start(&sg_miter, data->sg, data->sg_len, flags);
699 dma_unmap_sg(host->mmc->parent, data->sg, data->sg_len, mmc_get_dma_dir(data));
/linux-master/drivers/accel/habanalabs/common/
H A Dmemory.c849 struct scatterlist *sg; local
871 for_each_sgtable_dma_sg(userptr->sgt, sg, i) {
872 npages = hl_get_sg_info(sg, &dma_addr);
900 for_each_sgtable_dma_sg(userptr->sgt, sg, i) {
901 npages = hl_get_sg_info(sg, &dma_addr);
1515 static int set_dma_sg(struct scatterlist *sg, u64 bar_address, u64 chunk_size, argument
1527 sg_set_page(sg, NULL, chunk_size, 0);
1528 sg_dma_address(sg) = addr;
1529 sg_dma_len(sg) = chunk_size;
1541 struct scatterlist *sg; local
1768 struct scatterlist *sg; local
[all...]
/linux-master/drivers/crypto/chelsio/
H A Dchcr_algo.c134 static int sg_nents_xlen(struct scatterlist *sg, unsigned int reqlen, argument
142 while (sg && skip) {
143 if (sg_dma_len(sg) <= skip) {
144 skip -= sg_dma_len(sg);
146 sg = sg_next(sg);
153 while (sg && reqlen) {
154 less = min(reqlen, sg_dma_len(sg) - skip_len);
158 sg = sg_next(sg);
422 dsgl_walk_add_sg(struct dsgl_walk *walk, struct scatterlist *sg, unsigned int slen, unsigned int skip) argument
507 ulptx_walk_add_sg(struct ulptx_walk *walk, struct scatterlist *sg, unsigned int len, unsigned int skip) argument
[all...]
/linux-master/drivers/scsi/qla2xxx/
H A Dqla_iocb.c200 struct scatterlist *sg; local
222 scsi_for_each_sg(cmd, sg, tot_dsds, i) {
236 append_dsd32(&cur_dsd, sg);
256 struct scatterlist *sg; local
278 scsi_for_each_sg(cmd, sg, tot_dsds, i) {
292 append_dsd64(&cur_dsd, sg);
367 /* Map the sg table so we have an accurate count of sg entries needed */
707 struct scatterlist *sg; local
741 scsi_for_each_sg(cmd, sg, tot_dsd
796 struct scatterlist *sg; local
956 struct scatterlist *sg, *sgl; local
1033 struct scatterlist *sg, *sgl; local
3226 struct scatterlist *sg; local
3293 struct scatterlist *sg; local
4058 struct scatterlist *sg; local
[all...]
/linux-master/drivers/mailbox/
H A Dbcm-flexrm-mailbox.c566 struct scatterlist *sg; local
570 for (sg = msg->spu.src; sg; sg = sg_next(sg)) {
571 if (sg->length & 0xf) {
572 if (sg->length > SRC_LENGTH_MASK)
575 if (sg->length > (MSRC_LENGTH_MASK * 16))
579 for (sg = msg->spu.dst; sg; s
[all...]
/linux-master/drivers/misc/mei/
H A Dbus.c959 * @total_in_len: total length of data in 'in' sg, can be less than the sum of buffers sizes
980 struct scatterlist *sg; local
1019 for (i = 0, sg = sg_in; i < sg_in_nents; i++, sg++) {
1020 ext_hdr->sgl[i].low = lower_32_bits(sg_dma_address(sg));
1021 ext_hdr->sgl[i].high = upper_32_bits(sg_dma_address(sg));
1022 sg_len = min_t(unsigned int, sg_dma_len(sg), PAGE_SIZE);
1028 for (i = sg_in_nents, sg = sg_out; i < sg_in_nents + sg_out_nents; i++, sg++) {
1029 ext_hdr->sgl[i].low = lower_32_bits(sg_dma_address(sg));
[all...]
/linux-master/drivers/scsi/sym53c8xx_2/
H A Dsym_fw2.h1085 offsetof (struct sym_ccb, phys.pm0.sg),
1105 offsetof (struct sym_ccb, phys.pm0.sg),
1150 offsetof (struct sym_ccb, phys.pm1.sg),
1170 offsetof (struct sym_ccb, phys.pm1.sg),
1682 offsetof(struct sym_ccb, phys.pm0.sg.size),
1684 offsetof(struct sym_ccb, phys.pm0.sg.addr),
1712 offsetof(struct sym_ccb, phys.pm1.sg.size),
1714 offsetof(struct sym_ccb, phys.pm1.sg.addr),
/linux-master/drivers/net/wireless/ti/wl18xx/
H A Dacx.c305 wl->conf.sg.params[WL18XX_CONF_SG_TIME_SYNC],
314 acx->sync_mode = wl->conf.sg.params[WL18XX_CONF_SG_TIME_SYNC];
/linux-master/include/linux/
H A Dmemstick.h243 struct scatterlist sg; member in union:memstick_request::__anon228
320 const struct scatterlist *sg);
/linux-master/arch/arm/mach-footbridge/
H A Ddma-isa.c87 if (!dma->sg) {
92 dma->sg = &dma->buf;
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_mem.c112 if (tt->sg)
113 args.sgl = tt->sg->sgl;
/linux-master/drivers/ata/
H A Dpata_arasan_cf.c431 static int sg_xfer(struct arasan_cf_dev *acdev, struct scatterlist *sg) argument
439 sglen = sg_dma_len(sg);
441 src = sg_dma_address(sg);
444 dest = sg_dma_address(sg);
449 * For each sg:
464 /* continue dma xfers until current sg is completed */
523 struct scatterlist *sg; local
538 for_each_sg(qc->sg, sg, qc->n_elem, temp) {
539 ret = sg_xfer(acdev, sg);
[all...]
/linux-master/drivers/memstick/host/
H A Djmb38x_ms.c309 length = host->req->sg.length - host->block_pos;
310 off = host->req->sg.offset + host->block_pos;
320 pg = nth_page(sg_page(&host->req->sg),
404 data_len = host->req->sg.length;
418 if (1 != dma_map_sg(&host->chip->pdev->dev, &host->req->sg, 1,
425 data_len = sg_dma_len(&host->req->sg);
426 writel(sg_dma_address(&host->req->sg),
486 dma_unmap_sg(&host->chip->pdev->dev, &host->req->sg, 1,
/linux-master/drivers/gpu/drm/i915/gt/
H A Dselftest_migrate.c139 struct scatterlist *sg,
146 struct sgt_dma it = sg_sgt(sg);
211 if (err || !it.sg || !sg_dma_len(it.sg))
225 struct scatterlist *sg,
246 err = intel_context_copy_ccs(ce, deps, sg, pat_index,
844 struct scatterlist *sg,
859 err = intel_context_migrate_clear(ce, NULL, sg, pat_index,
137 intel_context_copy_ccs(struct intel_context *ce, const struct i915_deps *deps, struct scatterlist *sg, unsigned int pat_index, bool write_to_ccs, struct i915_request **out) argument
222 intel_migrate_ccs_copy(struct intel_migrate *m, struct i915_gem_ww_ctx *ww, const struct i915_deps *deps, struct scatterlist *sg, unsigned int pat_index, bool write_to_ccs, struct i915_request **out) argument
843 __perf_clear_blt(struct intel_context *ce, struct scatterlist *sg, unsigned int pat_index, bool is_lmem, size_t sz) argument
/linux-master/drivers/target/iscsi/
H A Discsi_target.c894 struct scatterlist *sg; local
910 sg = &cmd->se_cmd.t_data_sg[ent];
913 cmd->first_data_sg = sg;
919 if (WARN_ON_ONCE(!sg || i >= nvec))
922 cur_len = min_t(u32, data_length, sg->length - page_off);
924 iov[i].iov_base = kmap(sg_page(sg)) + sg->offset + page_off;
929 sg = sg_next(sg);
938 pr_err("offset %d + length %d overflow; %d/%d; sg
951 struct scatterlist *sg; local
1416 struct scatterlist *sg; local
1468 struct scatterlist sg[2]; local
[all...]
/linux-master/drivers/dma/
H A Dmpc512x_dma.c695 struct scatterlist *sg; local
706 for_each_sg(sgl, sg, sg_len, i) {
743 tcd->daddr = sg_dma_address(sg);
745 if (!IS_ALIGNED(sg_dma_address(sg), mchan->dwidth))
751 tcd->saddr = sg_dma_address(sg);
754 if (!IS_ALIGNED(sg_dma_address(sg), mchan->swidth))
765 tcd->nbytes = sg_dma_len(sg);
773 len = sg_dma_len(sg);
/linux-master/net/ceph/
H A Dmessenger_v2.c914 static void init_sgs(struct scatterlist **sg, void *buf, int buf_len, u8 *pad) argument
930 sg_set_page(*sg, page, len, 0);
931 *sg = sg_next(*sg);
935 sg_set_buf(*sg, buf, buf_len);
936 *sg = sg_next(*sg);
940 sg_set_buf(*sg, pad, padding_len(buf_len));
941 *sg = sg_next(*sg);
945 init_sgs_cursor(struct scatterlist **sg, struct ceph_msg_data_cursor *cursor, u8 *pad) argument
979 init_sgs_pages(struct scatterlist **sg, struct page **pages, int dpos, int dlen, u8 *pad) argument
1062 struct scatterlist sg; local
1250 struct scatterlist sg; local
[all...]

Completed in 327 milliseconds

<<11121314151617181920>>