Lines Matching refs:sg
68 /* copy the unaligned sg the buffer which is used for RDMA */
70 struct scatterlist *sg;
75 for_each_sg(sgl, sg, data->size, i) {
76 from = kmap_atomic(sg_page(sg), KM_USER0);
78 from + sg->offset,
79 sg->length);
81 p += sg->length;
123 struct scatterlist *sgl, *sg;
128 /* copy back read RDMA to unaligned sg */
135 for_each_sg(sgl, sg, sg_size, i) {
136 to = kmap_atomic(sg_page(sg), KM_SOFTIRQ0);
137 memcpy(to + sg->offset,
139 sg->length);
141 p += sg->length;
175 struct scatterlist *sg, *sgl = (struct scatterlist *)data->buf;
186 for_each_sg(sgl, sg, data->dma_nents, i) {
187 start_addr = ib_sg_dma_address(ibdev, sg);
190 dma_len = ib_sg_dma_len(ibdev, sg);
226 struct scatterlist *sgl, *sg, *next_sg = NULL;
236 for_each_sg(sgl, sg, data->dma_nents, i) {
240 next_sg = sg_next(sg);
244 end_addr = start_addr + ib_sg_dma_len(ibdev, sg);
257 iser_dbg("Found %d aligned entries out of %d in sg:0x%p\n",
266 struct scatterlist *sg;
272 for_each_sg(sgl, sg, data->dma_nents, i)
273 iser_warn("sg[%d] dma_addr:0x%lX page:0x%p "
275 i, (unsigned long)ib_sg_dma_address(ibdev, sg),
276 sg_page(sg), sg->offset,
277 sg->length, ib_sg_dma_len(ibdev, sg));
299 iser_dbg("Translating sg sz: %d\n", data->dma_nents);
301 iser_dbg("sg len %d page_vec_len %d\n", data->dma_nents,page_vec_len);
367 struct scatterlist *sg;
390 sg = (struct scatterlist *)mem->buf;
394 regd_buf->reg.len = ib_sg_dma_len(ibdev, &sg[0]);
395 regd_buf->reg.va = ib_sg_dma_address(ibdev, &sg[0]);