/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_migrate.h | 49 struct scatterlist *sg, 57 struct scatterlist *sg,
|
/linux-master/drivers/infiniband/core/ |
H A D | umem.c | 51 struct scatterlist *sg; local 58 for_each_sgtable_sg(&umem->sgt_append.sgt, sg, i) 59 unpin_user_page_range_dirty_lock(sg_page(sg), 60 DIV_ROUND_UP(sg->length, PAGE_SIZE), make_dirty); 83 struct scatterlist *sg; local 109 for_each_sgtable_dma_sg(&umem->sgt_append.sgt, sg, i) { 113 mask |= (sg_dma_address(sg) + pgoff) ^ va; 114 va += sg_dma_len(sg) - pgoff;
|
H A D | rw.c | 92 struct rdma_rw_reg_ctx *reg, struct scatterlist *sg, 106 ret = ib_map_mr_sg(reg->mr, sg, nents, &offset, PAGE_SIZE); 125 u32 port_num, struct scatterlist *sg, u32 sg_cnt, u32 offset, 144 ret = rdma_rw_init_one_mr(qp, port_num, reg, sg, sg_cnt, 176 sg = sg_next(sg); 196 struct scatterlist *sg, u32 sg_cnt, u32 offset, 227 for (j = 0; j < nr_sge; j++, sg = sg_next(sg)) { 228 sge->addr = sg_dma_address(sg) 91 rdma_rw_init_one_mr(struct ib_qp *qp, u32 port_num, struct rdma_rw_reg_ctx *reg, struct scatterlist *sg, u32 sg_cnt, u32 offset) argument 124 rdma_rw_init_mr_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 195 rdma_rw_init_map_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 sg_cnt, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 251 rdma_rw_init_single_wr(struct rdma_rw_ctx *ctx, struct ib_qp *qp, struct scatterlist *sg, u32 offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 292 rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, u32 sg_offset, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 364 rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, struct ib_sig_attrs *sig_attrs, u64 remote_addr, u32 rkey, enum dma_data_direction dir) argument 574 rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, enum dma_data_direction dir) argument 613 rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num, struct scatterlist *sg, u32 sg_cnt, struct scatterlist *prot_sg, u32 prot_sg_cnt, enum dma_data_direction dir) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_phys.c | 25 struct scatterlist *sg; local 33 if (overflows_type(obj->base.size, sg->length)) 57 sg = st->sgl; 58 sg->offset = 0; 59 sg->length = obj->base.size; 61 sg_assign_page(sg, (struct page *)vaddr); 62 sg_dma_address(sg) = dma; 63 sg_dma_len(sg) = obj->base.size;
|
/linux-master/drivers/dma/hsu/ |
H A D | hsu.c | 87 hsu_chan_writel(hsuc, HSU_CH_DxSAR(i), desc->sg[i].addr); 88 hsu_chan_writel(hsuc, HSU_CH_DxTSR(i), desc->sg[i].len); 252 desc->sg = kcalloc(nents, sizeof(*desc->sg), GFP_NOWAIT); 253 if (!desc->sg) { 265 kfree(desc->sg); 276 struct scatterlist *sg; local 283 for_each_sg(sgl, sg, sg_len, i) { 284 desc->sg[i].addr = sg_dma_address(sg); [all...] |
/linux-master/arch/sparc/mm/ |
H A D | iommu.c | 253 struct scatterlist *sg; local 256 for_each_sg(sgl, sg, nents, j) { 257 sg->dma_address =__sbus_iommu_map_page(dev, sg_page(sg), 258 sg->offset, sg->length, per_page_flush); 259 if (sg->dma_address == DMA_MAPPING_ERROR) 261 sg->dma_length = sg->length; 302 struct scatterlist *sg; local [all...] |
/linux-master/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_reqmgr.c | 61 dma_unmap_sg(dev, sr->in.sg, sg_nents(sr->in.sg), 66 sr->in.sg = NULL; 69 dma_unmap_sg(dev, sr->out.sg, sg_nents(sr->out.sg), 74 sr->out.sg = NULL; 112 struct scatterlist *sg; local 127 sg = sgtbl->sg; 128 /* populate device sg componen 162 struct scatterlist *sg; local [all...] |
/linux-master/drivers/scsi/ |
H A D | libiscsi_tcp.c | 82 * @sg: scatterlist 83 * @offset: byte offset into that sg entry 86 * data is copied to the indicated sg entry, at the given 91 struct scatterlist *sg, unsigned int offset) 93 segment->sg = sg; 95 segment->size = min(sg->length - offset, 111 struct scatterlist *sg; local 113 if (segment->data != NULL || !segment->sg) 116 sg 90 iscsi_tcp_segment_init_sg(struct iscsi_segment *segment, struct scatterlist *sg, unsigned int offset) argument 194 struct scatterlist sg; local 317 struct scatterlist sg; local 376 struct scatterlist *sg; local [all...] |
/linux-master/drivers/media/pci/cx25821/ |
H A D | cx25821-core.c | 993 struct scatterlist *sg; local 1007 sg = sglist; 1009 while (offset && offset >= sg_dma_len(sg)) { 1010 offset -= sg_dma_len(sg); 1011 sg = sg_next(sg); 1013 if (bpl <= sg_dma_len(sg) - offset) { 1017 *(rp++) = cpu_to_le32(sg_dma_address(sg) + offset); 1024 (sg_dma_len(sg) - offset)); 1025 *(rp++) = cpu_to_le32(sg_dma_address(sg) 1104 struct scatterlist *sg; local [all...] |
/linux-master/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_ctrl_mbox.c | 162 struct octep_ctrl_mbox_msg_buf *sg; local 186 sg = &msg->sg_list[s]; 187 w_sz = (sg->sz <= buf_sz) ? sg->sz : buf_sz; 188 octep_write_mbox_data(q, &pi, ci, sg->msg, w_sz); 227 struct octep_ctrl_mbox_msg_buf *sg; local 249 sg = &msg->sg_list[s]; 250 r_sz = (sg->sz <= buf_sz) ? sg->sz : buf_sz; 251 octep_read_mbox_data(q, pi, &ci, sg [all...] |
/linux-master/drivers/dma/ti/ |
H A D | omap-dma.c | 127 struct omap_sg sg[] __counted_by(sglen); 257 if (d->sg[i].t2_desc) 258 dma_pool_free(od->desc_pool, d->sg[i].t2_desc, 259 d->sg[i].t2_desc_paddr); 269 struct omap_sg *sg = &d->sg[idx]; local 270 struct omap_type2_desc *t2_desc = sg->t2_desc; 273 d->sg[idx - 1].t2_desc->next_desc = sg->t2_desc_paddr; 277 t2_desc->en = sg 541 struct omap_sg *sg = d->sg + c->sgidx; local 786 omap_dma_sg_size(struct omap_sg *sg) argument 1293 struct omap_sg *sg; local [all...] |
/linux-master/include/soc/fsl/ |
H A D | qman.h | 101 * 'contig' implies a contiguous buffer, whereas 'sg' implies a 202 static inline dma_addr_t qm_sg_addr(const struct qm_sg_entry *sg) argument 204 return be64_to_cpu(sg->data) & 0xffffffffffLLU; 207 static inline u64 qm_sg_entry_get64(const struct qm_sg_entry *sg) argument 209 return be64_to_cpu(sg->data) & 0xffffffffffLLU; 212 static inline void qm_sg_entry_set64(struct qm_sg_entry *sg, u64 addr) argument 214 sg->addr_hi = upper_32_bits(addr); 215 sg->addr_lo = cpu_to_be32(lower_32_bits(addr)); 218 static inline bool qm_sg_entry_is_final(const struct qm_sg_entry *sg) argument 220 return be32_to_cpu(sg 223 qm_sg_entry_is_ext(const struct qm_sg_entry *sg) argument 228 qm_sg_entry_get_len(const struct qm_sg_entry *sg) argument 233 qm_sg_entry_set_len(struct qm_sg_entry *sg, int len) argument 238 qm_sg_entry_set_f(struct qm_sg_entry *sg, int len) argument 243 qm_sg_entry_get_off(const struct qm_sg_entry *sg) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 67 struct scatterlist *sg; local 69 for (sg = st->sgl; sg; sg = __sg_next(sg)) { 70 if (sg_page(sg)) 71 __free_pages(sg_page(sg), get_order(sg->length)); 83 struct scatterlist *sg; local 101 sg 216 struct scatterlist *sg; local 270 struct scatterlist *sg; local [all...] |
/linux-master/drivers/staging/rts5208/ |
H A D | rtsx_transport.c | 63 struct scatterlist *sg = local 75 struct page *page = sg_page(sg) + 76 ((sg->offset + *offset) >> PAGE_SHIFT); 77 unsigned int poff = (sg->offset + *offset) & 79 unsigned int sglen = sg->length - *offset; 89 ++sg; 312 struct scatterlist *sg, int num_sg, 327 if (!sg || num_sg <= 0 || !offset || !index) 356 sg_cnt = dma_map_sg(&rtsx->pci->dev, sg, num_sg, dma_dir); 359 sg_ptr = sg; 311 rtsx_transfer_sglist_adma_partial(struct rtsx_chip *chip, u8 card, struct scatterlist *sg, int num_sg, unsigned int *index, unsigned int *offset, int size, enum dma_data_direction dma_dir, int timeout) argument 477 rtsx_transfer_sglist_adma(struct rtsx_chip *chip, u8 card, struct scatterlist *sg, int num_sg, enum dma_data_direction dma_dir, int timeout) argument 720 struct scatterlist *sg = buf; local [all...] |
/linux-master/drivers/gpu/drm/virtio/ |
H A D | virtgpu_object.c | 141 struct scatterlist *sg; local 163 for_each_sgtable_dma_sg(pages, sg, si) { 164 (*ents)[si].addr = cpu_to_le64(sg_dma_address(sg)); 165 (*ents)[si].length = cpu_to_le32(sg_dma_len(sg)); 169 for_each_sgtable_sg(pages, sg, si) { 170 (*ents)[si].addr = cpu_to_le64(sg_phys(sg)); 171 (*ents)[si].length = cpu_to_le32(sg->length);
|
/linux-master/drivers/crypto/caam/ |
H A D | caamrng.c | 120 struct scatterlist sg[1]; local 124 sg_init_table(sg, ARRAY_SIZE(sg)); 125 nents = kfifo_dma_in_prepare(&ctx->fifo, sg, ARRAY_SIZE(sg), 130 len = caam_rng_read_one(ctx->jrdev, sg_virt(&sg[0]), 131 sg[0].length,
|
/linux-master/drivers/gpu/drm/mediatek/ |
H A D | mtk_drm_gem.c | 105 if (mtk_gem->sg) 106 drm_prime_gem_destroy(obj, mtk_gem->sg); 213 struct dma_buf_attachment *attach, struct sg_table *sg) 218 if (drm_prime_get_contiguous_size(sg) < attach->dmabuf->size) { 227 mtk_gem->dma_addr = sg_dma_address(sg->sgl); 228 mtk_gem->sg = sg; 212 mtk_gem_prime_import_sg_table(struct drm_device *dev, struct dma_buf_attachment *attach, struct sg_table *sg) argument
|
/linux-master/drivers/crypto/ |
H A D | atmel-sha.c | 101 struct scatterlist *sg; member in struct:atmel_sha_reqctx 102 unsigned int offset; /* offset in current sg */ 125 struct scatterlist *sg; member in struct:atmel_sha_dma 308 count = min(ctx->sg->length - ctx->offset, ctx->total); 314 * because the sg length is 0. In the latest case, 315 * check if there is another sg in the list, a 0 length 316 * sg doesn't necessarily mean the end of the sg list. 318 if ((ctx->sg->length == 0) && !sg_is_last(ctx->sg)) { 642 struct scatterlist sg[2]; local 763 struct scatterlist *sg; local 1387 atmel_sha_dma_check_aligned(struct atmel_sha_dev *dd, struct scatterlist *sg, size_t len) argument 1426 struct scatterlist *sg; local 1548 atmel_sha_cpu_start(struct atmel_sha_dev *dd, struct scatterlist *sg, unsigned int len, bool idatar0_only, bool wait_data_ready, atmel_sha_fn_t resume) argument [all...] |
/linux-master/drivers/spi/ |
H A D | spi-topcliff-pch.c | 901 struct scatterlist *sg; local 991 sg = dma->sg_rx_p; 992 for (i = 0; i < num; i++, sg++) { 994 sg->offset = size * i; 995 sg->offset = sg->offset * (*bpw / 8); 996 sg_set_page(sg, virt_to_page(dma->rx_buf_virt), rem, 997 sg->offset); 998 sg_dma_len(sg) = rem; 1000 sg [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.h | 62 struct scatterlist sg[MLX4_ICM_CHUNK_LEN]; member in union:mlx4_icm_chunk::__anon6427 130 return sg_dma_address(&iter->chunk->sg[iter->page_idx]); 138 return sg_dma_len(&iter->chunk->sg[iter->page_idx]);
|
/linux-master/net/wireless/ |
H A D | lib80211_crypt_ccmp.c | 195 struct scatterlist sg[2]; local 218 sg_init_table(sg, 2); 219 sg_set_buf(&sg[0], aad, aad_len); 220 sg_set_buf(&sg[1], skb->data + hdr_len + CCMP_HDR_LEN, 225 aead_request_set_crypt(req, sg, sg, data_len, iv); 260 struct scatterlist sg[2]; local 319 sg_init_table(sg, 2); 320 sg_set_buf(&sg[0], aad, aad_len); 321 sg_set_buf(&sg[ [all...] |
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-cmac.c | 61 struct scatterlist *sg, *cmac_key_sg = NULL; local 114 sg = NULL; 117 sg = ccp_crypto_sg_table_add(&rctx->data_sg, &rctx->buf_sg); 118 if (!sg) { 125 sg = ccp_crypto_sg_table_add(&rctx->data_sg, req->src); 126 if (!sg) { 140 sg = ccp_crypto_sg_table_add(&rctx->data_sg, &rctx->pad_sg); 141 if (!sg) { 146 if (sg) { 147 sg_mark_end(sg); [all...] |
/linux-master/drivers/nvdimm/ |
H A D | nd_virtio.c | 43 struct scatterlist *sgs[2], sg, ret; local 57 sg_init_one(&sg, &req_data->req, sizeof(req_data->req)); 58 sgs[0] = &sg;
|
/linux-master/drivers/staging/rtl8192e/ |
H A D | rtllib_crypt_ccmp.c | 187 struct scatterlist sg[2]; local 200 sg_init_table(sg, 2); 201 sg_set_buf(&sg[0], aad, aad_len); 202 sg_set_buf(&sg[1], skb->data + hdr_len + CCMP_HDR_LEN, 207 aead_request_set_crypt(req, sg, sg, data_len, iv); 271 struct scatterlist sg[2]; local 282 sg_init_table(sg, 2); 283 sg_set_buf(&sg[0], aad, aad_len); 284 sg_set_buf(&sg[ [all...] |
/linux-master/drivers/infiniband/ulp/iser/ |
H A D | iser_memory.c | 82 data->dma_nents = ib_dma_map_sg(dev, data->sg, data->size, dma_dir); 91 pdata->dma_nents = ib_dma_map_sg(dev, pdata->sg, pdata->size, dma_dir); 101 ib_dma_unmap_sg(dev, data->sg, data->size, dma_dir); 114 ib_dma_unmap_sg(dev, data->sg, data->size, dma_dir); 119 ib_dma_unmap_sg(dev, pdata->sg, pdata->size, dma_dir); 126 struct scatterlist *sg = mem->sg; local 138 reg->sge.addr = sg_dma_address(&sg[0]); 139 reg->sge.length = sg_dma_len(&sg[0]); 272 ret = ib_map_mr_sg_pi(mr, mem->sg, me [all...] |