Lines Matching refs:sgt
68 if (dobj->sgt)
70 dobj->sgt, DMA_TO_DEVICE);
392 struct sg_table *sgt;
395 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
396 if (!sgt)
404 if (sg_alloc_table(sgt, count, GFP_KERNEL))
409 for_each_sgtable_sg(sgt, sg, i) {
419 if (dma_map_sgtable(attach->dev, sgt, dir, 0))
423 if (sg_alloc_table(sgt, 1, GFP_KERNEL))
426 sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0);
428 if (dma_map_sgtable(attach->dev, sgt, dir, 0))
432 if (sg_alloc_table(sgt, 1, GFP_KERNEL))
434 sg_dma_address(sgt->sgl) = dobj->dev_addr;
435 sg_dma_len(sgt->sgl) = dobj->obj.size;
439 return sgt;
442 for_each_sgtable_sg(sgt, sg, i)
446 sg_free_table(sgt);
448 kfree(sgt);
453 struct sg_table *sgt, enum dma_data_direction dir)
460 dma_unmap_sgtable(attach->dev, sgt, dir, 0);
465 for_each_sgtable_sg(sgt, sg, i)
469 sg_free_table(sgt);
470 kfree(sgt);
542 dobj->sgt = dma_buf_map_attachment_unlocked(dobj->obj.import_attach,
544 if (IS_ERR(dobj->sgt)) {
545 ret = PTR_ERR(dobj->sgt);
546 dobj->sgt = NULL;
550 if (dobj->sgt->nents > 1) {
554 if (sg_dma_len(dobj->sgt->sgl) < dobj->obj.size) {
558 dobj->dev_addr = sg_dma_address(dobj->sgt->sgl);