Searched refs:page_list (Results 1 - 25 of 42) sorted by relevance

12

/freebsd-10.0-release/sys/ofed/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c122 if (array->page_list[p].page)
123 return array->page_list[p].page[index & MTHCA_ARRAY_MASK];
133 if (!array->page_list[p].page)
134 array->page_list[p].page = (void **) get_zeroed_page(GFP_ATOMIC);
136 if (!array->page_list[p].page)
139 array->page_list[p].page[index & MTHCA_ARRAY_MASK] = value;
140 ++array->page_list[p].used;
149 if (--array->page_list[p].used == 0) {
150 free_page((unsigned long) array->page_list[p].page);
151 array->page_list[
[all...]
H A Dmthca_eq.c230 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE;
482 eq->page_list = kmalloc(npages * sizeof *eq->page_list,
484 if (!eq->page_list)
488 eq->page_list[i].buf = NULL;
500 eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev,
502 if (!eq->page_list[i].buf)
506 pci_unmap_addr_set(&eq->page_list[i], mapping, t);
508 clear_page(eq->page_list[i].buf);
578 if (eq->page_list[
[all...]
H A Dmthca_provider.h54 struct mthca_buf_list *page_list; member in union:mthca_buf
114 struct mthca_buf_list *page_list; member in struct:mthca_eq
H A Dmthca_mr.c712 static inline int mthca_check_fmr(struct mthca_fmr *fmr, u64 *page_list, argument
726 /* Trust the user not to pass misaligned data in page_list */
729 if (page_list[i] & ~page_mask)
740 int mthca_tavor_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, argument
749 err = mthca_check_fmr(fmr, page_list, list_len, iova);
762 __be64 mtt_entry = cpu_to_be64(page_list[i] |
781 int mthca_arbel_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, argument
789 err = mthca_check_fmr(fmr, page_list, list_len, iova);
807 fmr->mem.arbel.mtts[i] = cpu_to_be64(page_list[i] |
H A Dmthca_provider.c932 u64 *page_list; local
970 page_list = kmalloc(npages * sizeof *page_list, GFP_KERNEL);
971 if (!page_list) {
981 page_list[n++] = buffer_list[i].addr + ((u64) j << shift);
992 page_list, shift, npages,
997 kfree(page_list);
1002 kfree(page_list);
1068 len = sg_dma_len(&chunk->page_list[j]) >> shift;
1070 pages[i++] = sg_dma_address(&chunk->page_list[
[all...]
H A Dmthca_dev.h191 } *page_list; member in struct:mthca_array
483 int mthca_tavor_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list,
486 int mthca_arbel_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list,
/freebsd-10.0-release/sys/ofed/include/rdma/
H A Dib_fmr_pool.h76 u64 page_list[0]; member in struct:ib_pool_fmr
87 u64 *page_list,
H A Dib_umem.h83 struct scatterlist page_list[0]; member in struct:ib_umem_chunk
/freebsd-10.0-release/sys/xen/interface/
H A Dkexec.h98 unsigned long page_list[KEXEC_XEN_NO_PAGES]; member in struct:xen_kexec_image
/freebsd-10.0-release/sys/ofed/drivers/infiniband/core/
H A Dumem.c59 ((PAGE_SIZE - offsetof(struct ib_umem_chunk, page_list)) / \
60 ((void *) &((struct ib_umem_chunk *) 0)->page_list[1] - \
61 (void *) &((struct ib_umem_chunk *) 0)->page_list[0]))
120 ib_dma_unmap_sg_attrs(dev, chunk->page_list,
123 struct page *page = sg_page(&chunk->page_list[i]);
137 ib_dma_unmap_sg_attrs(dev, chunk->page_list,
140 struct page *page = sg_page(&chunk->page_list[i]);
172 struct page **page_list; local
214 page_list = (struct page **) __get_free_page(GFP_KERNEL);
215 if (!page_list) {
[all...]
H A Dfmr_pool.c114 u64 *page_list,
125 bucket = pool->cache_bucket + ib_fmr_hash(*page_list);
130 !memcmp(page_list, fmr->page_list,
131 page_list_len * sizeof *page_list))
428 * @page_list:List of pages to map
429 * @list_len:Number of pages in @page_list
435 u64 *page_list,
449 page_list,
474 result = ib_map_phys_fmr(fmr->fmr, page_list, list_le
113 ib_fmr_cache_lookup(struct ib_fmr_pool *pool, u64 *page_list, int page_list_len, u64 io_virtual_address) argument
434 ib_fmr_pool_map_phys(struct ib_fmr_pool *pool_handle, u64 *page_list, int list_len, u64 io_virtual_address) argument
[all...]
H A Dverbs.c894 struct ib_fast_reg_page_list *page_list; local
899 page_list = device->alloc_fast_reg_page_list(device, max_page_list_len);
901 if (!IS_ERR(page_list)) {
902 page_list->device = device;
903 page_list->max_page_list_len = max_page_list_len;
906 return page_list;
910 void ib_free_fast_reg_page_list(struct ib_fast_reg_page_list *page_list) argument
912 page_list->device->free_fast_reg_page_list(page_list);
/freebsd-10.0-release/sys/ofed/drivers/infiniband/hw/mlx4/
H A Dmr.c226 sg_dma_address(&chunk->page_list[j])) {
228 len += sg_dma_len(&chunk->page_list[j]);
245 sg_dma_address(&chunk->page_list[j]);
246 len = sg_dma_len(&chunk->page_list[j]);
340 sg_dma_address(&chunk->page_list[0]);
369 sg_dma_address(&chunk->page_list[j]);
395 sg_dma_len(&chunk->page_list[j]);
405 sg_dma_len(&chunk->page_list[j]);
630 mfrpl->ibfrpl.page_list = kmalloc(size, GFP_KERNEL);
631 if (!mfrpl->ibfrpl.page_list)
650 mlx4_ib_free_fast_reg_page_list(struct ib_fast_reg_page_list *page_list) argument
696 mlx4_ib_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, int npages, u64 iova) argument
[all...]
H A Ddoorbell.c77 db->dma = sg_dma_address(chunk->page_list) + (virt & ~PAGE_MASK);
/freebsd-10.0-release/sys/ofed/drivers/net/mlx4/
H A Dalloc.c234 buf->page_list = kcalloc(buf->nbufs, sizeof(*buf->page_list),
236 if (!buf->page_list)
240 buf->page_list[i].buf =
243 if (!buf->page_list[i].buf)
246 buf->page_list[i].map = t;
248 memset(buf->page_list[i].buf, 0, PAGE_SIZE);
257 pages[i] = virt_to_page(buf->page_list[i].buf);
286 if (buf->page_list[i].buf)
288 buf->page_list[
[all...]
H A Dmr.c534 int start_index, int npages, u64 *page_list)
551 mtts[i] = cpu_to_be64(page_list[i] | MLX4_MTT_FLAG_PRESENT);
560 int start_index, int npages, u64 *page_list)
575 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list);
580 page_list += chunk;
588 int start_index, int npages, u64 *page_list)
611 inbox[i + 2] = cpu_to_be64(page_list[i] |
621 page_list += chunk;
627 return __mlx4_write_mtt(dev, mtt, start_index, npages, page_list);
634 u64 *page_list; local
533 mlx4_write_mtt_chunk(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument
559 __mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument
587 mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument
717 mlx4_check_fmr(struct mlx4_fmr *fmr, u64 *page_list, int npages, u64 iova) argument
744 mlx4_map_phys_fmr(struct mlx4_dev *dev, struct mlx4_fmr *fmr, u64 *page_list, int npages, u64 iova, u32 *lkey, u32 *rkey) argument
[all...]
H A Den_resources.c80 pages[i] = virt_to_page(buf->page_list[i].buf);
H A Deq.c112 return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE;
885 eq->page_list = kmalloc(npages * sizeof *eq->page_list,
887 if (!eq->page_list)
891 eq->page_list[i].buf = NULL;
903 eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev,
905 if (!eq->page_list[i].buf)
909 eq->page_list[i].map = t;
911 memset(eq->page_list[i].buf, 0, PAGE_SIZE);
964 if (eq->page_list[
[all...]
/freebsd-10.0-release/sys/dev/cxgb/ulp/iw_cxgb/
H A Diw_cxgb_mem.c176 __be64 **page_list)
220 *page_list = kmalloc(sizeof(u64) * *npages, M_NOWAIT);
221 if (!*page_list)
229 (*page_list)[n++] = htobe64(buffer_list[i].addr +
170 build_phys_page_list(struct ib_phys_buf *buffer_list, int num_phys_buf, u64 *iova_start, u64 *total_size, int *npages, int *shift, __be64 **page_list) argument
H A Diw_cxgb_provider.c405 __be64 *page_list; local
437 &total_size, &npages, &shift, &page_list);
443 cxfree(page_list);
447 ret = iwch_write_pbl(mhp, page_list, npages, 0);
448 cxfree(page_list);
487 __be64 *page_list = NULL; local
517 &shift, &page_list);
523 cxfree(page_list);
597 len = sg_dma_len(&chunk->page_list[j]) >> shift;
600 &chunk->page_list[
[all...]
H A Diw_cxgb_provider.h357 __be64 **page_list);
/freebsd-10.0-release/sys/dev/cxgbe/iw_cxgbe/
H A Dmem.c309 int *shift, __be64 **page_list)
353 *page_list = kmalloc(sizeof(u64) * *npages, GFP_KERNEL);
354 if (!*page_list)
362 (*page_list)[n++] = cpu_to_be64(buffer_list[i].addr +
382 __be64 *page_list = NULL; local
415 &shift, &page_list);
421 kfree(page_list);
443 __be64 *page_list; local
476 &page_list);
482 kfree(page_list);
306 build_phys_page_list(struct ib_phys_buf *buffer_list, int num_phys_buf, u64 *iova_start, u64 *total_size, int *npages, int *shift, __be64 **page_list) argument
[all...]
/freebsd-10.0-release/sys/dev/drm2/ttm/
H A Dttm_page_alloc_dma.c125 * @page_list: The link to the 'page_list' in 'struct dma_pool'.
131 struct list_head page_list; member in struct:dma_page
391 list_for_each_entry_safe(d_page, tmp, d_pages, page_list) {
392 list_del(&d_page->page_list);
404 list_del(&d_page->page_list);
449 page_list) {
454 list_move(&dma_p->page_list, &d_pages);
687 list_for_each_entry_safe(d_page, tmp, d_pages, page_list) {
691 list_del(&d_page->page_list);
[all...]
/freebsd-10.0-release/sys/contrib/rdma/krping/
H A Dkrping.c175 struct ib_fast_reg_page_list *page_list; member in struct:krping_cb
520 cb->fastreg_wr.wr.fast_reg.page_list = cb->page_list;
611 cb->page_list = ib_alloc_fast_reg_page_list(
614 if (IS_ERR(cb->page_list)) {
616 ret = PTR_ERR(cb->page_list);
620 cb->page_list->max_page_list_len);
626 DEBUG_LOG(cb, "fastreg rkey 0x%x page_list %p"
628 cb->page_list, cb->page_list_len);
709 if (cb->page_list
[all...]
/freebsd-10.0-release/sys/ofed/include/linux/mlx4/
H A Ddevice.h497 struct mlx4_buf_list *page_list; member in struct:mlx4_buf
897 return buf->page_list[offset >> PAGE_SHIFT].buf +
921 int start_index, int npages, u64 *page_list);
1086 int mlx4_map_phys_fmr(struct mlx4_dev *dev, struct mlx4_fmr *fmr, u64 *page_list,

Completed in 296 milliseconds

12