/freebsd-10.0-release/sys/ofed/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 122 if (array->page_list[p].page) 123 return array->page_list[p].page[index & MTHCA_ARRAY_MASK]; 133 if (!array->page_list[p].page) 134 array->page_list[p].page = (void **) get_zeroed_page(GFP_ATOMIC); 136 if (!array->page_list[p].page) 139 array->page_list[p].page[index & MTHCA_ARRAY_MASK] = value; 140 ++array->page_list[p].used; 149 if (--array->page_list[p].used == 0) { 150 free_page((unsigned long) array->page_list[p].page); 151 array->page_list[ [all...] |
H A D | mthca_eq.c | 230 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; 482 eq->page_list = kmalloc(npages * sizeof *eq->page_list, 484 if (!eq->page_list) 488 eq->page_list[i].buf = NULL; 500 eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev, 502 if (!eq->page_list[i].buf) 506 pci_unmap_addr_set(&eq->page_list[i], mapping, t); 508 clear_page(eq->page_list[i].buf); 578 if (eq->page_list[ [all...] |
H A D | mthca_provider.h | 54 struct mthca_buf_list *page_list; member in union:mthca_buf 114 struct mthca_buf_list *page_list; member in struct:mthca_eq
|
H A D | mthca_mr.c | 712 static inline int mthca_check_fmr(struct mthca_fmr *fmr, u64 *page_list, argument 726 /* Trust the user not to pass misaligned data in page_list */ 729 if (page_list[i] & ~page_mask) 740 int mthca_tavor_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, argument 749 err = mthca_check_fmr(fmr, page_list, list_len, iova); 762 __be64 mtt_entry = cpu_to_be64(page_list[i] | 781 int mthca_arbel_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, argument 789 err = mthca_check_fmr(fmr, page_list, list_len, iova); 807 fmr->mem.arbel.mtts[i] = cpu_to_be64(page_list[i] |
|
H A D | mthca_provider.c | 932 u64 *page_list; local 970 page_list = kmalloc(npages * sizeof *page_list, GFP_KERNEL); 971 if (!page_list) { 981 page_list[n++] = buffer_list[i].addr + ((u64) j << shift); 992 page_list, shift, npages, 997 kfree(page_list); 1002 kfree(page_list); 1068 len = sg_dma_len(&chunk->page_list[j]) >> shift; 1070 pages[i++] = sg_dma_address(&chunk->page_list[ [all...] |
H A D | mthca_dev.h | 191 } *page_list; member in struct:mthca_array 483 int mthca_tavor_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, 486 int mthca_arbel_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list,
|
/freebsd-10.0-release/sys/ofed/include/rdma/ |
H A D | ib_fmr_pool.h | 76 u64 page_list[0]; member in struct:ib_pool_fmr 87 u64 *page_list,
|
H A D | ib_umem.h | 83 struct scatterlist page_list[0]; member in struct:ib_umem_chunk
|
/freebsd-10.0-release/sys/xen/interface/ |
H A D | kexec.h | 98 unsigned long page_list[KEXEC_XEN_NO_PAGES]; member in struct:xen_kexec_image
|
/freebsd-10.0-release/sys/ofed/drivers/infiniband/core/ |
H A D | umem.c | 59 ((PAGE_SIZE - offsetof(struct ib_umem_chunk, page_list)) / \ 60 ((void *) &((struct ib_umem_chunk *) 0)->page_list[1] - \ 61 (void *) &((struct ib_umem_chunk *) 0)->page_list[0])) 120 ib_dma_unmap_sg_attrs(dev, chunk->page_list, 123 struct page *page = sg_page(&chunk->page_list[i]); 137 ib_dma_unmap_sg_attrs(dev, chunk->page_list, 140 struct page *page = sg_page(&chunk->page_list[i]); 172 struct page **page_list; local 214 page_list = (struct page **) __get_free_page(GFP_KERNEL); 215 if (!page_list) { [all...] |
H A D | fmr_pool.c | 114 u64 *page_list, 125 bucket = pool->cache_bucket + ib_fmr_hash(*page_list); 130 !memcmp(page_list, fmr->page_list, 131 page_list_len * sizeof *page_list)) 428 * @page_list:List of pages to map 429 * @list_len:Number of pages in @page_list 435 u64 *page_list, 449 page_list, 474 result = ib_map_phys_fmr(fmr->fmr, page_list, list_le 113 ib_fmr_cache_lookup(struct ib_fmr_pool *pool, u64 *page_list, int page_list_len, u64 io_virtual_address) argument 434 ib_fmr_pool_map_phys(struct ib_fmr_pool *pool_handle, u64 *page_list, int list_len, u64 io_virtual_address) argument [all...] |
H A D | verbs.c | 894 struct ib_fast_reg_page_list *page_list; local 899 page_list = device->alloc_fast_reg_page_list(device, max_page_list_len); 901 if (!IS_ERR(page_list)) { 902 page_list->device = device; 903 page_list->max_page_list_len = max_page_list_len; 906 return page_list; 910 void ib_free_fast_reg_page_list(struct ib_fast_reg_page_list *page_list) argument 912 page_list->device->free_fast_reg_page_list(page_list);
|
/freebsd-10.0-release/sys/ofed/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 226 sg_dma_address(&chunk->page_list[j])) { 228 len += sg_dma_len(&chunk->page_list[j]); 245 sg_dma_address(&chunk->page_list[j]); 246 len = sg_dma_len(&chunk->page_list[j]); 340 sg_dma_address(&chunk->page_list[0]); 369 sg_dma_address(&chunk->page_list[j]); 395 sg_dma_len(&chunk->page_list[j]); 405 sg_dma_len(&chunk->page_list[j]); 630 mfrpl->ibfrpl.page_list = kmalloc(size, GFP_KERNEL); 631 if (!mfrpl->ibfrpl.page_list) 650 mlx4_ib_free_fast_reg_page_list(struct ib_fast_reg_page_list *page_list) argument 696 mlx4_ib_map_phys_fmr(struct ib_fmr *ibfmr, u64 *page_list, int npages, u64 iova) argument [all...] |
H A D | doorbell.c | 77 db->dma = sg_dma_address(chunk->page_list) + (virt & ~PAGE_MASK);
|
/freebsd-10.0-release/sys/ofed/drivers/net/mlx4/ |
H A D | alloc.c | 234 buf->page_list = kcalloc(buf->nbufs, sizeof(*buf->page_list), 236 if (!buf->page_list) 240 buf->page_list[i].buf = 243 if (!buf->page_list[i].buf) 246 buf->page_list[i].map = t; 248 memset(buf->page_list[i].buf, 0, PAGE_SIZE); 257 pages[i] = virt_to_page(buf->page_list[i].buf); 286 if (buf->page_list[i].buf) 288 buf->page_list[ [all...] |
H A D | mr.c | 534 int start_index, int npages, u64 *page_list) 551 mtts[i] = cpu_to_be64(page_list[i] | MLX4_MTT_FLAG_PRESENT); 560 int start_index, int npages, u64 *page_list) 575 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list); 580 page_list += chunk; 588 int start_index, int npages, u64 *page_list) 611 inbox[i + 2] = cpu_to_be64(page_list[i] | 621 page_list += chunk; 627 return __mlx4_write_mtt(dev, mtt, start_index, npages, page_list); 634 u64 *page_list; local 533 mlx4_write_mtt_chunk(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument 559 __mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument 587 mlx4_write_mtt(struct mlx4_dev *dev, struct mlx4_mtt *mtt, int start_index, int npages, u64 *page_list) argument 717 mlx4_check_fmr(struct mlx4_fmr *fmr, u64 *page_list, int npages, u64 iova) argument 744 mlx4_map_phys_fmr(struct mlx4_dev *dev, struct mlx4_fmr *fmr, u64 *page_list, int npages, u64 iova, u32 *lkey, u32 *rkey) argument [all...] |
H A D | en_resources.c | 80 pages[i] = virt_to_page(buf->page_list[i].buf);
|
H A D | eq.c | 112 return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE; 885 eq->page_list = kmalloc(npages * sizeof *eq->page_list, 887 if (!eq->page_list) 891 eq->page_list[i].buf = NULL; 903 eq->page_list[i].buf = dma_alloc_coherent(&dev->pdev->dev, 905 if (!eq->page_list[i].buf) 909 eq->page_list[i].map = t; 911 memset(eq->page_list[i].buf, 0, PAGE_SIZE); 964 if (eq->page_list[ [all...] |
/freebsd-10.0-release/sys/dev/cxgb/ulp/iw_cxgb/ |
H A D | iw_cxgb_mem.c | 176 __be64 **page_list) 220 *page_list = kmalloc(sizeof(u64) * *npages, M_NOWAIT); 221 if (!*page_list) 229 (*page_list)[n++] = htobe64(buffer_list[i].addr + 170 build_phys_page_list(struct ib_phys_buf *buffer_list, int num_phys_buf, u64 *iova_start, u64 *total_size, int *npages, int *shift, __be64 **page_list) argument
|
H A D | iw_cxgb_provider.c | 405 __be64 *page_list; local 437 &total_size, &npages, &shift, &page_list); 443 cxfree(page_list); 447 ret = iwch_write_pbl(mhp, page_list, npages, 0); 448 cxfree(page_list); 487 __be64 *page_list = NULL; local 517 &shift, &page_list); 523 cxfree(page_list); 597 len = sg_dma_len(&chunk->page_list[j]) >> shift; 600 &chunk->page_list[ [all...] |
H A D | iw_cxgb_provider.h | 357 __be64 **page_list);
|
/freebsd-10.0-release/sys/dev/cxgbe/iw_cxgbe/ |
H A D | mem.c | 309 int *shift, __be64 **page_list) 353 *page_list = kmalloc(sizeof(u64) * *npages, GFP_KERNEL); 354 if (!*page_list) 362 (*page_list)[n++] = cpu_to_be64(buffer_list[i].addr + 382 __be64 *page_list = NULL; local 415 &shift, &page_list); 421 kfree(page_list); 443 __be64 *page_list; local 476 &page_list); 482 kfree(page_list); 306 build_phys_page_list(struct ib_phys_buf *buffer_list, int num_phys_buf, u64 *iova_start, u64 *total_size, int *npages, int *shift, __be64 **page_list) argument [all...] |
/freebsd-10.0-release/sys/dev/drm2/ttm/ |
H A D | ttm_page_alloc_dma.c | 125 * @page_list: The link to the 'page_list' in 'struct dma_pool'. 131 struct list_head page_list; member in struct:dma_page 391 list_for_each_entry_safe(d_page, tmp, d_pages, page_list) { 392 list_del(&d_page->page_list); 404 list_del(&d_page->page_list); 449 page_list) { 454 list_move(&dma_p->page_list, &d_pages); 687 list_for_each_entry_safe(d_page, tmp, d_pages, page_list) { 691 list_del(&d_page->page_list); [all...] |
/freebsd-10.0-release/sys/contrib/rdma/krping/ |
H A D | krping.c | 175 struct ib_fast_reg_page_list *page_list; member in struct:krping_cb 520 cb->fastreg_wr.wr.fast_reg.page_list = cb->page_list; 611 cb->page_list = ib_alloc_fast_reg_page_list( 614 if (IS_ERR(cb->page_list)) { 616 ret = PTR_ERR(cb->page_list); 620 cb->page_list->max_page_list_len); 626 DEBUG_LOG(cb, "fastreg rkey 0x%x page_list %p" 628 cb->page_list, cb->page_list_len); 709 if (cb->page_list [all...] |
/freebsd-10.0-release/sys/ofed/include/linux/mlx4/ |
H A D | device.h | 497 struct mlx4_buf_list *page_list; member in struct:mlx4_buf 897 return buf->page_list[offset >> PAGE_SHIFT].buf + 921 int start_index, int npages, u64 *page_list); 1086 int mlx4_map_phys_fmr(struct mlx4_dev *dev, struct mlx4_fmr *fmr, u64 *page_list,
|