Searched refs:pages (Results 101 - 125 of 933) sorted by relevance

1234567891011>>

/linux-master/drivers/misc/bcm-vk/
H A Dbcm_vk_sg.c57 /* Calculate number of pages */
62 /* Allocate DMA pages */
63 dma->pages = kmalloc_array(dma->nr_pages,
66 if (!dma->pages)
69 dev_dbg(dev, "Alloc DMA Pages [0x%llx+0x%x => %d pages]\n",
74 /* Get user pages into memory */
78 dma->pages);
103 /* Map all pages into DMA */
107 dma->pages[0],
113 __free_page(dma->pages[
[all...]
/linux-master/drivers/iommu/iommufd/
H A Diova_bitmap.c16 * Main data structure for tracking mapped user pages of bitmap data.
44 /* number of pages pinned */
47 /* pinned pages representing the bitmap data */
48 struct page **pages; member in struct:iova_bitmap_map
60 * The bitmap object uses one base page to store all the pinned pages
92 * mapping up to PAGE_SIZE / sizeof(struct page*) maximum of pages.
117 /* length of the IOVA range set ahead the pinned pages */
160 * Pins the bitmap user pages for the current range window.
175 * mapped but capped at a maximum number of pages.
195 FOLL_WRITE, mapped->pages);
[all...]
/linux-master/drivers/media/common/videobuf2/
H A Dvideobuf2-dma-sg.c37 struct page **pages; member in struct:vb2_dma_sg_buf
67 struct page *pages; local
76 pages = NULL;
77 while (!pages) {
78 pages = alloc_pages(GFP_KERNEL | __GFP_ZERO |
80 if (pages)
85 __free_page(buf->pages[last_page]);
91 split_page(pages, order);
93 buf->pages[last_page++] = &pages[
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dpin_system.c19 struct page **pages; member in struct:sdma_mmu_node
65 static void unpin_vector_pages(struct mm_struct *mm, struct page **pages, argument
68 hfi1_release_user_pages(mm, pages + start, npages, false);
69 kfree(pages);
80 unpin_vector_pages(mm_from_sdma_node(node), node->pages, 0,
120 struct page **pages; local
122 pages = kcalloc(npages, sizeof(*pages), GFP_KERNEL);
123 if (!pages)
136 SDMA_DBG(req, "Acquire user pages start_addres
[all...]
H A Duser_pages.c18 * Determine whether the caller can pin pages.
22 * to pin buffer pages in order to determine whether they should do so.
44 * Pinning these pages would exceed this process's locked memory
54 * 'unlimited' (-1), the value of this limit will be > 2^42 pages
62 * per-process tracking of pinned pages. It also assumes that each
71 * Pinning these pages would exceed the size limit for this cache.
81 bool writable, struct page **pages)
86 ret = pin_user_pages_fast(vaddr, npages, gup_flags, pages);
80 hfi1_acquire_user_pages(struct mm_struct *mm, unsigned long vaddr, size_t npages, bool writable, struct page **pages) argument
H A Duser_exp_rcv.h24 struct page **pages; member in struct:tid_user_buf
39 struct page *pages[] __counted_by(npages);
/linux-master/drivers/gpu/drm/rockchip/
H A Drockchip_drm_gem.c84 rk_obj->pages = drm_gem_get_pages(&rk_obj->base);
85 if (IS_ERR(rk_obj->pages))
86 return PTR_ERR(rk_obj->pages);
91 rk_obj->pages, rk_obj->num_pages);
99 * to flush the pages associated with it.
112 drm_gem_put_pages(&rk_obj->base, rk_obj->pages, false, false);
120 drm_gem_put_pages(&rk_obj->base, rk_obj->pages, true, true);
137 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP,
209 if (rk_obj->pages)
225 return vm_map_pages(vma, rk_obj->pages, coun
[all...]
/linux-master/fs/erofs/
H A Dzutil.c11 struct page **pages; member in struct:z_erofs_gbuf
84 tmp_pages[j] = gbuf->pages[j];
98 kfree(gbuf->pages);
99 gbuf->pages = tmp_pages;
111 if (tmp_pages[j] && tmp_pages[j] != gbuf->pages[j])
137 z_erofs_rsvbuf->pages = kcalloc(z_erofs_rsv_nrpages,
138 sizeof(*z_erofs_rsvbuf->pages), GFP_KERNEL);
139 if (!z_erofs_rsvbuf->pages) {
161 if (!gbuf->pages)
165 if (gbuf->pages[
[all...]
/linux-master/drivers/fpga/
H A Ddfl-afu-dma-region.c27 * afu_dma_pin_pages - pin pages of given dma memory region
31 * Pin all the pages of given dfl_afu_dma_region.
45 region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL);
46 if (!region->pages) {
52 region->pages);
61 dev_dbg(dev, "%d pages pinned\n", pinned);
66 unpin_user_pages(region->pages, pinned);
68 kfree(region->pages);
75 * afu_dma_unpin_pages - unpin pages of given dma memory region
79 * Unpin all the pages o
[all...]
/linux-master/fs/nfs/
H A Dnfs3acl.c49 struct page *pages[NFSACL_MAXPAGES] = { }; local
52 /* The xdr layer may allocate pages here. */
53 .pages = pages,
101 /* pages may have been allocated at the xdr layer. */
102 for (count = 0; count < NFSACL_MAXPAGES && args.pages[count]; count++)
103 __free_page(args.pages[count]);
166 struct page *pages[NFSACL_MAXPAGES]; local
171 .pages = pages,
[all...]
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_userptr.c56 * Block for operations on BOs to finish and mark pages as accessed and
159 struct sg_table *pages)
164 if (!pages)
167 __i915_gem_object_release_shmem(obj, pages, true);
168 i915_gem_gtt_finish_pages(obj, pages);
178 for_each_sgt_page(page, sgt_iter, pages) {
206 sg_free_table(pages);
207 kfree(pages);
214 struct sg_table *pages; local
226 pages
158 i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument
[all...]
/linux-master/drivers/gpu/drm/omapdrm/
H A Domap_gem.c42 /** protects pin_cnt, block, pages, dma_addrs and vaddr */
85 * Array of backing pages, if allocated. Note that pages are never
88 struct page **pages; member in struct:omap_gem_object
90 /** addresses corresponding to pages in above array */
126 int stride_pfn; /* stride in pages */
225 * Ensure backing pages are allocated. Must be called with the omap_obj.lock
232 struct page **pages; local
240 * If not using shmem (in which case backing pages don't need to be
241 * allocated) or if pages ar
386 struct page *pages[64]; /* XXX is this too much to have on stack? */ local
717 struct page **pages = omap_obj->pages; local
938 omap_gem_get_pages(struct drm_gem_object *obj, struct page ***pages, bool remap) argument
1413 struct page **pages; local
[all...]
/linux-master/tools/testing/radix-tree/
H A Dregression2.c82 struct page *pages[1]; local
114 radix_tree_gang_lookup_tag_slot(&mt_tree, (void ***)pages, start, end,
/linux-master/tools/perf/Documentation/
H A Dcat-texi.perl28 * Git Man Pages: (gitman). Manual pages for Git revision control system
/linux-master/drivers/gpu/drm/i915/
H A Di915_gem_gtt.c29 struct sg_table *pages)
33 pages->sgl, pages->nents,
44 * try again - if there are no more pages to remove from
47 GEM_BUG_ON(obj->mm.pages == pages);
57 struct sg_table *pages)
67 dma_unmap_sg(i915->drm.dev, pages->sgl, pages->nents,
88 * neighbouring nodes if the colors do not match (to ensure guard pages betwee
28 i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument
56 i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument
[all...]
H A Di915_gem_gtt.h24 struct sg_table *pages);
26 struct sg_table *pages);
/linux-master/tools/testing/selftests/kvm/
H A Daccess_tracking_perf_test.c10 * however the clear_young notifier can be triggered by marking a pages as idle
26 * 2. page_idle only operates on LRU pages. Newly allocated pages are not
33 * the TLB and the number of pages held in pagevecs are a small fraction of the
131 uint64_t pages = vcpu_args->pages; local
148 for (page = 0; page < pages; page++) {
166 * Assumption: Less than 1% of pages are going to be swapped out from
169 TEST_ASSERT(no_pfn < pages / 100,
170 "vCPU %d: No PFN for %" PRIu64 " out of %" PRIu64 " pages
190 vcpu_idx, still_idle, pages); local
[all...]
/linux-master/drivers/gpu/drm/virtio/
H A Dvirtgpu_object.c142 struct sg_table *pages; local
145 pages = drm_gem_shmem_get_pages_sgt(&bo->base);
146 if (IS_ERR(pages))
147 return PTR_ERR(pages);
150 *nents = pages->nents;
152 *nents = pages->orig_nents;
163 for_each_sgtable_dma_sg(pages, sg, si) {
169 for_each_sgtable_sg(pages, sg, si) {
/linux-master/fs/ramfs/
H A Dfile-nommu.c58 * add a contiguous set of pages into a ramfs inode when it's truncated from
65 struct page *pages; local
82 /* allocate enough contiguous pages to be able to satisfy the
84 pages = alloc_pages(gfp, order);
85 if (!pages)
88 /* split the high-order page into an array of single pages */
92 split_page(pages, order);
94 /* trim off any pages we don't actually require */
96 __free_page(pages + loop);
100 data = page_address(pages);
[all...]
/linux-master/include/linux/
H A Dcma.h49 extern bool cma_pages_valid(struct cma *cma, const struct page *pages, unsigned long count);
50 extern bool cma_release(struct cma *cma, const struct page *pages, unsigned long count);
/linux-master/drivers/gpu/drm/xe/
H A Dxe_hmm.c28 * @write: if write to this range, we mark pages in this range
47 * xe_build_sg() - build a scatter gather table for all the physical pages/pfn
48 * in a hmm_range. dma-map pages if necessary. dma-address is save in sg table
53 * has the pfn numbers of pages that back up this hmm address range.
56 * for system pages. If write we map it bi-diretional; otherwise
67 * FIXME: This function currently only support pages in system
84 struct page **pages; local
89 pages = kvmalloc_array(npages, sizeof(*pages), GFP_KERNEL);
90 if (!pages)
[all...]
/linux-master/drivers/gpu/drm/mediatek/
H A Dmtk_gem.c251 mtk_gem->pages = kcalloc(npages, sizeof(*mtk_gem->pages), GFP_KERNEL);
252 if (!mtk_gem->pages) {
258 drm_prime_sg_to_page_array(sgt, mtk_gem->pages, npages);
260 mtk_gem->kvaddr = vmap(mtk_gem->pages, npages, VM_MAP,
265 kfree(mtk_gem->pages);
282 if (!mtk_gem->pages)
287 kfree(mtk_gem->pages);
/linux-master/net/smc/
H A Dsmc_rx.c146 put_page(spd->pages[i]);
157 struct page **pages; local
164 pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL);
165 if (!pages)
187 pages[0] = smc->conn.rmb_desc->pages;
196 pages[i] = vmalloc_to_page(buf);
207 spd.pages = pages;
[all...]
/linux-master/drivers/xen/
H A Dgrant-table.c863 int gnttab_pages_set_private(int nr_pages, struct page **pages) argument
875 set_page_private(pages[i], (unsigned long)foreign);
877 SetPagePrivate(pages[i]);
885 * gnttab_alloc_pages - alloc pages suitable for grant mapping into
886 * @nr_pages: number of pages to alloc
887 * @pages: returns the pages
889 int gnttab_alloc_pages(int nr_pages, struct page **pages) argument
893 ret = xen_alloc_unpopulated_pages(nr_pages, pages);
897 ret = gnttab_pages_set_private(nr_pages, pages);
1028 gnttab_pages_clear_private(int nr_pages, struct page **pages) argument
1048 gnttab_free_pages(int nr_pages, struct page **pages) argument
1223 gnttab_foreach_grant(struct page **pages, unsigned int nr_grefs, xen_grant_fn_t fn, void *data) argument
1245 gnttab_map_refs(struct gnttab_map_grant_ref *map_ops, struct gnttab_map_grant_ref *kmap_ops, struct page **pages, unsigned int count) argument
1290 gnttab_unmap_refs(struct gnttab_unmap_grant_ref *unmap_ops, struct gnttab_unmap_grant_ref *kunmap_ops, struct page **pages, unsigned int count) argument
[all...]
/linux-master/tools/testing/selftests/powerpc/mm/
H A Dsubpage_prot.c96 long i, j, pages, err; local
98 pages = size / 0x10000;
99 map = malloc(pages * 4);
106 for (i = 0; i < pages; i++) {
120 for (i = 0; i < pages; i++) {

Completed in 421 milliseconds

1234567891011>>