/linux-master/drivers/misc/bcm-vk/ |
H A D | bcm_vk_sg.c | 57 /* Calculate number of pages */ 62 /* Allocate DMA pages */ 63 dma->pages = kmalloc_array(dma->nr_pages, 66 if (!dma->pages) 69 dev_dbg(dev, "Alloc DMA Pages [0x%llx+0x%x => %d pages]\n", 74 /* Get user pages into memory */ 78 dma->pages); 103 /* Map all pages into DMA */ 107 dma->pages[0], 113 __free_page(dma->pages[ [all...] |
/linux-master/drivers/iommu/iommufd/ |
H A D | iova_bitmap.c | 16 * Main data structure for tracking mapped user pages of bitmap data. 44 /* number of pages pinned */ 47 /* pinned pages representing the bitmap data */ 48 struct page **pages; member in struct:iova_bitmap_map 60 * The bitmap object uses one base page to store all the pinned pages 92 * mapping up to PAGE_SIZE / sizeof(struct page*) maximum of pages. 117 /* length of the IOVA range set ahead the pinned pages */ 160 * Pins the bitmap user pages for the current range window. 175 * mapped but capped at a maximum number of pages. 195 FOLL_WRITE, mapped->pages); [all...] |
/linux-master/drivers/media/common/videobuf2/ |
H A D | videobuf2-dma-sg.c | 37 struct page **pages; member in struct:vb2_dma_sg_buf 67 struct page *pages; local 76 pages = NULL; 77 while (!pages) { 78 pages = alloc_pages(GFP_KERNEL | __GFP_ZERO | 80 if (pages) 85 __free_page(buf->pages[last_page]); 91 split_page(pages, order); 93 buf->pages[last_page++] = &pages[ [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | pin_system.c | 19 struct page **pages; member in struct:sdma_mmu_node 65 static void unpin_vector_pages(struct mm_struct *mm, struct page **pages, argument 68 hfi1_release_user_pages(mm, pages + start, npages, false); 69 kfree(pages); 80 unpin_vector_pages(mm_from_sdma_node(node), node->pages, 0, 120 struct page **pages; local 122 pages = kcalloc(npages, sizeof(*pages), GFP_KERNEL); 123 if (!pages) 136 SDMA_DBG(req, "Acquire user pages start_addres [all...] |
H A D | user_pages.c | 18 * Determine whether the caller can pin pages. 22 * to pin buffer pages in order to determine whether they should do so. 44 * Pinning these pages would exceed this process's locked memory 54 * 'unlimited' (-1), the value of this limit will be > 2^42 pages 62 * per-process tracking of pinned pages. It also assumes that each 71 * Pinning these pages would exceed the size limit for this cache. 81 bool writable, struct page **pages) 86 ret = pin_user_pages_fast(vaddr, npages, gup_flags, pages); 80 hfi1_acquire_user_pages(struct mm_struct *mm, unsigned long vaddr, size_t npages, bool writable, struct page **pages) argument
|
H A D | user_exp_rcv.h | 24 struct page **pages; member in struct:tid_user_buf 39 struct page *pages[] __counted_by(npages);
|
/linux-master/drivers/gpu/drm/rockchip/ |
H A D | rockchip_drm_gem.c | 84 rk_obj->pages = drm_gem_get_pages(&rk_obj->base); 85 if (IS_ERR(rk_obj->pages)) 86 return PTR_ERR(rk_obj->pages); 91 rk_obj->pages, rk_obj->num_pages); 99 * to flush the pages associated with it. 112 drm_gem_put_pages(&rk_obj->base, rk_obj->pages, false, false); 120 drm_gem_put_pages(&rk_obj->base, rk_obj->pages, true, true); 137 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, 209 if (rk_obj->pages) 225 return vm_map_pages(vma, rk_obj->pages, coun [all...] |
/linux-master/fs/erofs/ |
H A D | zutil.c | 11 struct page **pages; member in struct:z_erofs_gbuf 84 tmp_pages[j] = gbuf->pages[j]; 98 kfree(gbuf->pages); 99 gbuf->pages = tmp_pages; 111 if (tmp_pages[j] && tmp_pages[j] != gbuf->pages[j]) 137 z_erofs_rsvbuf->pages = kcalloc(z_erofs_rsv_nrpages, 138 sizeof(*z_erofs_rsvbuf->pages), GFP_KERNEL); 139 if (!z_erofs_rsvbuf->pages) { 161 if (!gbuf->pages) 165 if (gbuf->pages[ [all...] |
/linux-master/drivers/fpga/ |
H A D | dfl-afu-dma-region.c | 27 * afu_dma_pin_pages - pin pages of given dma memory region 31 * Pin all the pages of given dfl_afu_dma_region. 45 region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL); 46 if (!region->pages) { 52 region->pages); 61 dev_dbg(dev, "%d pages pinned\n", pinned); 66 unpin_user_pages(region->pages, pinned); 68 kfree(region->pages); 75 * afu_dma_unpin_pages - unpin pages of given dma memory region 79 * Unpin all the pages o [all...] |
/linux-master/fs/nfs/ |
H A D | nfs3acl.c | 49 struct page *pages[NFSACL_MAXPAGES] = { }; local 52 /* The xdr layer may allocate pages here. */ 53 .pages = pages, 101 /* pages may have been allocated at the xdr layer. */ 102 for (count = 0; count < NFSACL_MAXPAGES && args.pages[count]; count++) 103 __free_page(args.pages[count]); 166 struct page *pages[NFSACL_MAXPAGES]; local 171 .pages = pages, [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_userptr.c | 56 * Block for operations on BOs to finish and mark pages as accessed and 159 struct sg_table *pages) 164 if (!pages) 167 __i915_gem_object_release_shmem(obj, pages, true); 168 i915_gem_gtt_finish_pages(obj, pages); 178 for_each_sgt_page(page, sgt_iter, pages) { 206 sg_free_table(pages); 207 kfree(pages); 214 struct sg_table *pages; local 226 pages 158 i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument [all...] |
/linux-master/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem.c | 42 /** protects pin_cnt, block, pages, dma_addrs and vaddr */ 85 * Array of backing pages, if allocated. Note that pages are never 88 struct page **pages; member in struct:omap_gem_object 90 /** addresses corresponding to pages in above array */ 126 int stride_pfn; /* stride in pages */ 225 * Ensure backing pages are allocated. Must be called with the omap_obj.lock 232 struct page **pages; local 240 * If not using shmem (in which case backing pages don't need to be 241 * allocated) or if pages ar 386 struct page *pages[64]; /* XXX is this too much to have on stack? */ local 717 struct page **pages = omap_obj->pages; local 938 omap_gem_get_pages(struct drm_gem_object *obj, struct page ***pages, bool remap) argument 1413 struct page **pages; local [all...] |
/linux-master/tools/testing/radix-tree/ |
H A D | regression2.c | 82 struct page *pages[1]; local 114 radix_tree_gang_lookup_tag_slot(&mt_tree, (void ***)pages, start, end,
|
/linux-master/tools/perf/Documentation/ |
H A D | cat-texi.perl | 28 * Git Man Pages: (gitman). Manual pages for Git revision control system
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_gtt.c | 29 struct sg_table *pages) 33 pages->sgl, pages->nents, 44 * try again - if there are no more pages to remove from 47 GEM_BUG_ON(obj->mm.pages == pages); 57 struct sg_table *pages) 67 dma_unmap_sg(i915->drm.dev, pages->sgl, pages->nents, 88 * neighbouring nodes if the colors do not match (to ensure guard pages betwee 28 i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument 56 i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument [all...] |
H A D | i915_gem_gtt.h | 24 struct sg_table *pages); 26 struct sg_table *pages);
|
/linux-master/tools/testing/selftests/kvm/ |
H A D | access_tracking_perf_test.c | 10 * however the clear_young notifier can be triggered by marking a pages as idle 26 * 2. page_idle only operates on LRU pages. Newly allocated pages are not 33 * the TLB and the number of pages held in pagevecs are a small fraction of the 131 uint64_t pages = vcpu_args->pages; local 148 for (page = 0; page < pages; page++) { 166 * Assumption: Less than 1% of pages are going to be swapped out from 169 TEST_ASSERT(no_pfn < pages / 100, 170 "vCPU %d: No PFN for %" PRIu64 " out of %" PRIu64 " pages 190 vcpu_idx, still_idle, pages); local [all...] |
/linux-master/drivers/gpu/drm/virtio/ |
H A D | virtgpu_object.c | 142 struct sg_table *pages; local 145 pages = drm_gem_shmem_get_pages_sgt(&bo->base); 146 if (IS_ERR(pages)) 147 return PTR_ERR(pages); 150 *nents = pages->nents; 152 *nents = pages->orig_nents; 163 for_each_sgtable_dma_sg(pages, sg, si) { 169 for_each_sgtable_sg(pages, sg, si) {
|
/linux-master/fs/ramfs/ |
H A D | file-nommu.c | 58 * add a contiguous set of pages into a ramfs inode when it's truncated from 65 struct page *pages; local 82 /* allocate enough contiguous pages to be able to satisfy the 84 pages = alloc_pages(gfp, order); 85 if (!pages) 88 /* split the high-order page into an array of single pages */ 92 split_page(pages, order); 94 /* trim off any pages we don't actually require */ 96 __free_page(pages + loop); 100 data = page_address(pages); [all...] |
/linux-master/include/linux/ |
H A D | cma.h | 49 extern bool cma_pages_valid(struct cma *cma, const struct page *pages, unsigned long count); 50 extern bool cma_release(struct cma *cma, const struct page *pages, unsigned long count);
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_hmm.c | 28 * @write: if write to this range, we mark pages in this range 47 * xe_build_sg() - build a scatter gather table for all the physical pages/pfn 48 * in a hmm_range. dma-map pages if necessary. dma-address is save in sg table 53 * has the pfn numbers of pages that back up this hmm address range. 56 * for system pages. If write we map it bi-diretional; otherwise 67 * FIXME: This function currently only support pages in system 84 struct page **pages; local 89 pages = kvmalloc_array(npages, sizeof(*pages), GFP_KERNEL); 90 if (!pages) [all...] |
/linux-master/drivers/gpu/drm/mediatek/ |
H A D | mtk_gem.c | 251 mtk_gem->pages = kcalloc(npages, sizeof(*mtk_gem->pages), GFP_KERNEL); 252 if (!mtk_gem->pages) { 258 drm_prime_sg_to_page_array(sgt, mtk_gem->pages, npages); 260 mtk_gem->kvaddr = vmap(mtk_gem->pages, npages, VM_MAP, 265 kfree(mtk_gem->pages); 282 if (!mtk_gem->pages) 287 kfree(mtk_gem->pages);
|
/linux-master/net/smc/ |
H A D | smc_rx.c | 146 put_page(spd->pages[i]); 157 struct page **pages; local 164 pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL); 165 if (!pages) 187 pages[0] = smc->conn.rmb_desc->pages; 196 pages[i] = vmalloc_to_page(buf); 207 spd.pages = pages; [all...] |
/linux-master/drivers/xen/ |
H A D | grant-table.c | 863 int gnttab_pages_set_private(int nr_pages, struct page **pages) argument 875 set_page_private(pages[i], (unsigned long)foreign); 877 SetPagePrivate(pages[i]); 885 * gnttab_alloc_pages - alloc pages suitable for grant mapping into 886 * @nr_pages: number of pages to alloc 887 * @pages: returns the pages 889 int gnttab_alloc_pages(int nr_pages, struct page **pages) argument 893 ret = xen_alloc_unpopulated_pages(nr_pages, pages); 897 ret = gnttab_pages_set_private(nr_pages, pages); 1028 gnttab_pages_clear_private(int nr_pages, struct page **pages) argument 1048 gnttab_free_pages(int nr_pages, struct page **pages) argument 1223 gnttab_foreach_grant(struct page **pages, unsigned int nr_grefs, xen_grant_fn_t fn, void *data) argument 1245 gnttab_map_refs(struct gnttab_map_grant_ref *map_ops, struct gnttab_map_grant_ref *kmap_ops, struct page **pages, unsigned int count) argument 1290 gnttab_unmap_refs(struct gnttab_unmap_grant_ref *unmap_ops, struct gnttab_unmap_grant_ref *kunmap_ops, struct page **pages, unsigned int count) argument [all...] |
/linux-master/tools/testing/selftests/powerpc/mm/ |
H A D | subpage_prot.c | 96 long i, j, pages, err; local 98 pages = size / 0x10000; 99 map = malloc(pages * 4); 106 for (i = 0; i < pages; i++) { 120 for (i = 0; i < pages; i++) {
|