/linux-master/drivers/accel/qaic/ |
H A D | qaic_control.c | 196 struct page **page_list; member in struct:dma_xfer 334 put_page(xfer->page_list[i]); 335 kfree(xfer->page_list); 397 struct page **page_list; local 424 page_list = kmalloc_array(nr_pages, sizeof(*page_list), GFP_KERNEL | __GFP_NOWARN); 425 if (!page_list) { 434 ret = get_user_pages_fast(xfer_start_addr, nr_pages, 0, page_list); 449 ret = sg_alloc_table_from_pages(sgt, page_list, nr_pages, 462 xfer->page_list [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 118 return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE; 990 eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list), 992 if (!eq->page_list) 996 eq->page_list[i].buf = NULL; 1008 eq->page_list[i].buf = dma_alloc_coherent(&dev->persist-> 1012 if (!eq->page_list[i].buf) 1016 eq->page_list[i].map = t; 1073 if (eq->page_list[i].buf) 1075 eq->page_list[ [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_db.c | 20 list_for_each_entry(page, &context->page_list, list) 40 list_add(&page->list, &context->page_list);
|
H A D | hns_roce_mr.c | 430 mr->page_list[mr->npages++] = addr; 447 mr->page_list = kvcalloc(mr->pbl_mtr.hem_cfg.buf_pg_count, 449 if (!mr->page_list) 463 ret = hns_roce_mtr_map(hr_dev, mtr, mr->page_list, mr->npages); 472 kvfree(mr->page_list); 473 mr->page_list = NULL;
|
/linux-master/drivers/md/ |
H A D | dm-io.c | 190 struct page_list *pl = dp->context_ptr; 199 struct page_list *pl = dp->context_ptr; 205 static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned int offset)
|
H A D | dm-integrity.c | 178 struct page_list *journal; 179 struct page_list *journal_io; 180 struct page_list *journal_xor; 181 struct page_list *recalc_bitmap; 182 struct page_list *may_write_bitmap; 578 static bool block_bitmap_op(struct dm_integrity_c *ic, struct page_list *bitmap, 684 static void block_bitmap_copy(struct dm_integrity_c *ic, struct page_list *dst, struct page_list *src) 734 static struct journal_sector *access_page_list(struct dm_integrity_c *ic, struct page_list *pl, 889 struct page_list *source_p [all...] |
/linux-master/drivers/virt/vboxguest/ |
H A D | vboxguest_utils.c | 355 dst_parm->u.page_list.size = len; 356 dst_parm->u.page_list.offset = *off_extra; 585 dst_parm->u.page_list.size = src_parm->u.page_list.size;
|
/linux-master/arch/x86/include/asm/ |
H A D | kexec.h | 128 unsigned long page_list,
|
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_misc.c | 206 u64 *page_list, 216 ret = pvrdma_page_dir_insert_dma(pdir, i, page_list[i]); 205 pvrdma_page_dir_insert_page_list(struct pvrdma_page_dir *pdir, u64 *page_list, int num_pages) argument
|
H A D | pvrdma.h | 550 u64 *page_list, int num_pages);
|
/linux-master/drivers/rapidio/devices/ |
H A D | rio_mport_cdev.c | 543 struct page **page_list; member in struct:mport_dma_req 581 if (req->page_list) { 582 unpin_user_pages(req->page_list, req->nr_pages); 583 kfree(req->page_list); 814 struct page **page_list = NULL; local 858 page_list = kmalloc_array(nr_pages, 859 sizeof(*page_list), GFP_KERNEL); 860 if (page_list == NULL) { 869 page_list); 889 ret = sg_alloc_table_from_pages(&req->sgt, page_list, nr_page [all...] |
/linux-master/drivers/vhost/ |
H A D | vdpa.c | 1091 struct page **page_list; local 1101 page_list = (struct page **) __get_free_page(GFP_KERNEL); 1102 if (!page_list) 1129 gup_flags, page_list); 1134 unpin_user_pages(page_list, pinned); 1142 map_pfn = page_to_pfn(page_list[0]); 1145 unsigned long this_pfn = page_to_pfn(page_list[i]); 1158 * page_list. The remaining outstanding 1163 unpin_user_pages(&page_list[i], 1206 free_page((unsigned long)page_list); [all...] |
/linux-master/drivers/net/ethernet/google/gve/ |
H A D | gve_adminq.c | 976 __be64 *page_list; local 981 page_list = dma_alloc_coherent(hdev, size, &page_list_bus, GFP_KERNEL); 982 if (!page_list) 986 page_list[i] = cpu_to_be64(qpl->page_buses[i]); 997 dma_free_coherent(hdev, size, page_list, page_list_bus);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_validation.c | 116 list_add_tail(&page->lru, &ctx->page_list); 139 list_for_each_entry_safe(entry, next, &ctx->page_list, lru) {
|
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.h | 128 u64 *page_list; member in struct:bnxt_re_frpl
|
/linux-master/drivers/vfio/pci/mlx5/ |
H A D | cmd.h | 135 struct page **page_list; member in struct:mlx5_vhca_recv_buf
|
/linux-master/drivers/vdpa/vdpa_user/ |
H A D | vduse_dev.c | 1061 struct page **page_list = NULL; local 1079 page_list = __vmalloc(array_size(npages, sizeof(struct page *)), 1082 if (!page_list || !umem) 1092 page_list); 1099 page_list, pinned); 1105 umem->pages = page_list; 1114 unpin_user_pages(page_list, pinned); 1119 vfree(page_list);
|
/linux-master/drivers/misc/genwqe/ |
H A D | card_base.h | 170 struct page **page_list; /* list of pages used by user buff */ member in struct:dma_mapping
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gpu_error.h | 46 struct list_head page_list; member in struct:i915_vma_coredump
|
H A D | i915_gpu_error.c | 298 list_add_tail(&page->lru, &dst->page_list); 412 list_add_tail(&virt_to_page(ptr)->lru, &dst->page_list); 633 list_for_each_entry(page, &vma->page_list, lru) { 638 if (page == list_last_entry(&vma->page_list, typeof(*page), lru)) 1020 list_for_each_entry_safe(page, n, &vma->page_list, lru) { 1113 INIT_LIST_HEAD(&dst->page_list); 1196 list_for_each_entry_safe_reverse(page, n, &dst->page_list, lru) {
|
/linux-master/include/linux/ |
H A D | gfp.h | 215 struct list_head *page_list,
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_verbs.h | 320 struct xarray page_list; member in struct:rxe_mr
|
/linux-master/drivers/dma/ |
H A D | ste_dma40.c | 3369 unsigned long *page_list; local 3378 page_list = kmalloc_array(MAX_LCLA_ALLOC_ATTEMPTS, 3379 sizeof(*page_list), 3381 if (!page_list) 3388 page_list[i] = __get_free_pages(GFP_KERNEL, 3390 if (!page_list[i]) { 3397 free_pages(page_list[j], base->lcla_pool.pages); 3401 if ((virt_to_phys((void *)page_list[i]) & 3407 free_pages(page_list[j], base->lcla_pool.pages); 3410 base->lcla_pool.base = (void *)page_list[ [all...] |
/linux-master/mm/ |
H A D | hugetlb.c | 2721 LIST_HEAD(page_list); 2751 list_add(&folio->lru, &page_list); 2756 update_and_free_pages_bulk(h, &page_list); 3691 LIST_HEAD(page_list); 3709 list_add(&folio->lru, &page_list); 3715 update_and_free_pages_bulk(h, &page_list); 3765 LIST_HEAD(page_list); 3857 prep_and_add_allocated_folios(h, &page_list); 3862 list_add(&folio->lru, &page_list); 3867 prep_and_add_allocated_folios(h, &page_list); [all...] |
/linux-master/include/linux/mlx4/ |
H A D | device.h | 644 struct mlx4_buf_list *page_list; member in struct:mlx4_buf 1093 return buf->page_list[offset >> PAGE_SHIFT].buf + 1134 int start_index, int npages, u64 *page_list);
|