Lines Matching defs:mem

58 		struct scatterlist mem;
68 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages,
72 __free_pages(sg_page(&chunk->mem[i]),
73 get_order(chunk->mem[i].length));
81 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length,
82 lowmem_page_address(sg_page(&chunk->mem[i])),
83 sg_dma_address(&chunk->mem[i]));
106 static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask)
118 sg_set_page(mem, page, PAGE_SIZE << order, 0);
122 static int mthca_alloc_icm_coherent(struct device *dev, struct scatterlist *mem,
125 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, &sg_dma_address(mem),
130 sg_set_buf(mem, buf, PAGE_SIZE << order);
131 BUG_ON(mem->offset);
132 sg_dma_len(mem) = PAGE_SIZE << order;
163 sg_init_table(chunk->mem, MTHCA_ICM_CHUNK_LEN);
174 &chunk->mem[chunk->npages],
177 ret = mthca_alloc_icm_pages(&chunk->mem[chunk->npages],
186 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem,
206 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem,
299 if (sg_dma_len(&chunk->mem[i]) > dma_offset)
300 *dma_handle = sg_dma_address(&chunk->mem[i]) +
302 dma_offset -= sg_dma_len(&chunk->mem[i]);
307 if (chunk->mem[i].length > offset) {
308 page = sg_page(&chunk->mem[i]);
311 offset -= chunk->mem[i].length;
478 sg_set_page(&db_tab->page[i].mem, pages[0], MTHCA_ICM_PAGE_SIZE,
481 ret = pci_map_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE);
487 ret = mthca_MAP_ICM_page(dev, sg_dma_address(&db_tab->page[i].mem),
490 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE);
491 put_page(sg_page(&db_tab->page[i].mem));
539 sg_init_table(&db_tab->page[i].mem, 1);
556 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE);
557 put_page(sg_page(&db_tab->page[i].mem));