Searched refs:pages (Results 1 - 25 of 931) sorted by relevance

1234567891011>>

/linux-master/net/9p/
H A Dtrans_common.h7 void p9_release_pages(struct page **pages, int nr_pages);
H A Dtrans_common.c12 * p9_release_pages - Release pages after the transaction.
13 * @pages: array of pages to be put
16 void p9_release_pages(struct page **pages, int nr_pages) argument
21 if (pages[i])
22 put_page(pages[i]);
/linux-master/kernel/dma/
H A Dremap.c15 return area->pages;
19 * Remaps an array of PAGE_SIZE pages into another vm_area.
22 void *dma_common_pages_remap(struct page **pages, size_t size, argument
27 vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT,
30 find_vm_area(vaddr)->pages = pages;
42 struct page **pages; local
46 pages = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL);
47 if (!pages)
50 pages[
[all...]
/linux-master/net/ceph/
H A Dpagevec.c13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) argument
19 set_page_dirty_lock(pages[i]);
20 put_page(pages[i]);
22 kvfree(pages);
26 void ceph_release_page_vector(struct page **pages, int num_pages) argument
31 __free_pages(pages[i], 0);
32 kfree(pages);
37 * allocate a vector new pages
41 struct page **pages; local
44 pages
61 ceph_copy_user_to_page_vector(struct page **pages, const void __user *data, loff_t off, size_t len) argument
87 ceph_copy_to_page_vector(struct page **pages, const void *data, loff_t off, size_t len) argument
110 ceph_copy_from_page_vector(struct page **pages, void *data, loff_t off, size_t len) argument
137 ceph_zero_page_vector_range(int off, int len, struct page **pages) argument
[all...]
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c12 struct sg_table *pages)
18 for_each_sgt_page(page, sgt_iter, pages) {
24 sg_free_table(pages);
25 kfree(pages);
34 struct sg_table *pages; local
41 pages = kmalloc(sizeof(*pages), GFP);
42 if (!pages)
45 if (sg_alloc_table(pages, npages, GFP)) {
46 kfree(pages);
11 huge_free_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument
85 huge_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument
[all...]
H A Dmock_dmabuf.h14 struct page *pages[]; member in struct:mock_dmabuf
/linux-master/fs/erofs/
H A Dpcpubuf.c6 * per-CPU virtual memory (in pages) in advance to store such inplace I/O
15 struct page **pages; member in struct:erofs_pcpubuf
64 struct page **pages, **oldpages; local
67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL);
68 if (!pages) {
74 pages[i] = erofs_allocpage(&pagepool, GFP_KERNEL);
75 if (!pages[i]) {
77 oldpages = pages;
81 ptr = vmap(pages, nrpage
[all...]
/linux-master/include/xen/
H A Dmem-reservation.h30 struct page **pages,
34 struct page **pages);
38 struct page **pages,
43 __xenmem_reservation_va_mapping_update(count, pages, frames);
48 struct page **pages)
52 __xenmem_reservation_va_mapping_reset(count, pages);
37 xenmem_reservation_va_mapping_update(unsigned long count, struct page **pages, xen_pfn_t *frames) argument
47 xenmem_reservation_va_mapping_reset(unsigned long count, struct page **pages) argument
H A Dballoon.h15 /* Number of pages in high- and low-memory balloons. */
29 int xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages);
30 void xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages);
H A Dxen.h65 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages);
66 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages);
72 struct page **pages)
74 return xen_alloc_ballooned_pages(nr_pages, pages);
77 struct page **pages)
79 xen_free_ballooned_pages(nr_pages, pages);
71 xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument
76 xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument
/linux-master/drivers/iommu/iommufd/
H A Dpages.c69 * allocation can hold about 26M of 4k pages and 13G of 2M pages in an
163 static void iopt_pages_add_npinned(struct iopt_pages *pages, size_t npages) argument
167 rc = check_add_overflow(pages->npinned, npages, &pages->npinned);
169 WARN_ON(rc || pages->npinned > pages->npages);
172 static void iopt_pages_sub_npinned(struct iopt_pages *pages, size_t npages) argument
176 rc = check_sub_overflow(pages->npinned, npages, &pages
181 iopt_pages_err_unpin(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **page_list) argument
249 iopt_pages_find_domain_area(struct iopt_pages *pages, unsigned long index) argument
575 pages_to_xarray(struct xarray *xa, unsigned long start_index, unsigned long last_index, struct page **pages) argument
615 batch_from_pages(struct pfn_batch *batch, struct page **pages, size_t npages) argument
625 batch_unpin(struct pfn_batch *batch, struct iopt_pages *pages, unsigned int first_page_off, size_t npages) argument
708 pfn_reader_user_init(struct pfn_reader_user *user, struct iopt_pages *pages) argument
721 pfn_reader_user_destroy(struct pfn_reader_user *user, struct iopt_pages *pages) argument
736 pfn_reader_user_pin(struct pfn_reader_user *user, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument
804 incr_user_locked_vm(struct iopt_pages *pages, unsigned long npages) argument
822 decr_user_locked_vm(struct iopt_pages *pages, unsigned long npages) argument
830 update_mm_locked_vm(struct iopt_pages *pages, unsigned long npages, bool inc, struct pfn_reader_user *user) argument
857 do_update_pinned(struct iopt_pages *pages, unsigned long npages, bool inc, struct pfn_reader_user *user) argument
886 update_unpinned(struct iopt_pages *pages) argument
903 pfn_reader_user_update_pinned(struct pfn_reader_user *user, struct iopt_pages *pages) argument
937 struct iopt_pages *pages; member in struct:pfn_reader
961 struct iopt_pages *pages = pfns->pages; local
1062 pfn_reader_init(struct pfn_reader *pfns, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument
1093 struct iopt_pages *pages = pfns->pages; local
1113 struct iopt_pages *pages = pfns->pages; local
1121 pfn_reader_first(struct pfn_reader *pfns, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument
1144 struct iopt_pages *pages; local
1183 struct iopt_pages *pages = container_of(kref, struct iopt_pages, kref); local
1197 iopt_area_unpin_domain(struct pfn_batch *batch, struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long start_index, unsigned long last_index, unsigned long *unmapped_end_index, unsigned long real_last_index) argument
1257 __iopt_area_unfill_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long last_index) argument
1309 iopt_area_unfill_partial_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long end_index) argument
1342 iopt_area_unfill_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain) argument
1408 iopt_area_fill_domains(struct iopt_area *area, struct iopt_pages *pages) argument
1492 iopt_area_unfill_domains(struct iopt_area *area, struct iopt_pages *pages) argument
1519 iopt_pages_unpin_xarray(struct pfn_batch *batch, struct iopt_pages *pages, unsigned long start_index, unsigned long end_index) argument
1542 iopt_pages_unfill_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument
1590 iopt_pages_fill_from_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument
1610 iopt_pages_fill_from_domain(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument
1632 iopt_pages_fill_from_mm(struct iopt_pages *pages, struct pfn_reader_user *user, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument
1671 iopt_pages_fill_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument
1743 iopt_pages_rw_slow(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, unsigned long offset, void *data, unsigned long length, unsigned int flags) argument
1784 iopt_pages_rw_page(struct iopt_pages *pages, unsigned long index, unsigned long offset, void *data, unsigned long length, unsigned int flags) argument
1831 iopt_pages_rw_access(struct iopt_pages *pages, unsigned long start_byte, void *data, unsigned long length, unsigned int flags) argument
1886 iopt_pages_get_exact_access(struct iopt_pages *pages, unsigned long index, unsigned long last) argument
1919 struct iopt_pages *pages = area->pages; local
1974 struct iopt_pages *pages = area->pages; local
[all...]
/linux-master/arch/s390/hypfs/
H A Dhypfs_diag.h16 void *diag204_get_buffer(enum diag204_format fmt, int *pages);
17 int diag204_store(void *buf, int pages);
H A Dhypfs_diag.c30 static int diag204_buf_pages; /* number of pages for diag204 data */
49 * up to 93 pages!
58 void *diag204_get_buffer(enum diag204_format fmt, int *pages) argument
61 *pages = diag204_buf_pages;
65 *pages = 1;
67 *pages = diag204((unsigned long)DIAG204_SUBC_RSI |
69 if (*pages <= 0)
72 diag204_buf = __vmalloc_node(array_size(*pages, PAGE_SIZE),
77 diag204_buf_pages = *pages;
98 int pages, r local
141 diag204_store(void *buf, int pages) argument
[all...]
/linux-master/sound/soc/sof/
H A Dsof-utils.c29 int i, pages; local
31 pages = snd_sgbuf_aligned_pages(size);
33 dev_dbg(dev, "generating page table for %p size 0x%zx pages %d\n",
34 dmab->area, size, pages);
36 for (i = 0; i < pages; i++) {
71 return pages;
/linux-master/mm/
H A Dpercpu-vm.c23 * pcpu_get_pages - get temp pages array
30 * Pointer to temp pages array on success.
34 static struct page **pages; local
35 size_t pages_size = pcpu_nr_units * pcpu_unit_pages * sizeof(pages[0]);
39 if (!pages)
40 pages = pcpu_mem_zalloc(pages_size, GFP_KERNEL);
41 return pages;
45 * pcpu_free_pages - free pages which were allocated for @chunk
46 * @chunk: chunk pages were allocated for
47 * @pages
54 pcpu_free_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument
82 pcpu_alloc_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end, gfp_t gfp) argument
153 pcpu_unmap_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument
193 __pcpu_map_pages(unsigned long addr, struct page **pages, int nr_pages) argument
214 pcpu_map_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument
279 struct page **pages; local
315 struct page **pages; local
[all...]
H A Dgup_test.c10 static void put_back_pages(unsigned int cmd, struct page **pages, argument
19 put_page(pages[i]);
25 unpin_user_pages(pages, nr_pages);
29 unpin_user_pages(pages, nr_pages);
32 put_page(pages[i]);
39 static void verify_dma_pinned(unsigned int cmd, struct page **pages, argument
50 folio = page_folio(pages[i]);
53 "pages[%lu] is NOT dma-pinned\n", i)) {
59 "pages[%lu] is NOT pinnable but pinned\n",
69 static void dump_pages_test(struct gup_test *gup, struct page **pages, argument
106 struct page **pages; local
227 struct page **pages; local
[all...]
/linux-master/arch/x86/include/asm/
H A Dpgtable_32.h56 * With PAE paging (PTRS_PER_PMD > 1), we allocate PTRS_PER_PGD == 4 pages for
57 * the PMD's in addition to the pages required for the last level pagetables.
60 #define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD)
62 #define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PGD)
66 * Number of possible pages in the lowmem region.
/linux-master/arch/arm64/kvm/hyp/include/nvhe/
H A Dffa.h14 int hyp_ffa_init(void *pages);
/linux-master/drivers/media/common/videobuf2/
H A Dframe_vector.c16 * @nr_frames: number of pages / pfns from start to map
18 * @vec: structure which receives pages / pfns of the addresses mapped.
22 * with page frame numbers or page pointers to corresponding pages (choice
24 * belongs to a normal vma, the function grabs reference to each of the pages
29 * The function returns number of pages mapped which may be less than
72 * put_vaddr_frames() - drop references to pages if get_vaddr_frames() acquired
76 * Drop references to pages if get_vaddr_frames() acquired them. We also
82 struct page **pages; local
86 pages = frame_vector_pages(vec);
89 * get_vaddr_frames() got pages bu
114 struct page **pages; local
140 struct page **pages; local
[all...]
/linux-master/fs/isofs/
H A Dcompress.c37 * to one zisofs block. Store the data in the @pages array with @pcount
42 struct page **pages, unsigned poffset,
68 if (!pages[i])
70 memzero_page(pages[i], 0, PAGE_SIZE);
71 SetPageUptodate(pages[i]);
121 if (pages[curpage]) {
122 stream.next_out = kmap_local_page(pages[curpage])
174 if (pages[curpage]) {
175 flush_dcache_page(pages[curpage]);
176 SetPageUptodate(pages[curpag
40 zisofs_uncompress_block(struct inode *inode, loff_t block_start, loff_t block_end, int pcount, struct page **pages, unsigned poffset, int *errp) argument
206 zisofs_fill_pages(struct inode *inode, int full_page, int pcount, struct page **pages) argument
313 struct page **pages; local
[all...]
/linux-master/include/trace/events/
H A Dtlb.h40 TP_PROTO(int reason, unsigned long pages),
41 TP_ARGS(reason, pages),
45 __field(unsigned long, pages)
50 __entry->pages = pages;
53 TP_printk("pages:%ld reason:%s (%d)",
54 __entry->pages,
/linux-master/drivers/xen/
H A Dxlate_mmu.c47 /* Break down the pages in 4KB chunk and call fn for each gfn */
48 static void xen_for_each_gfn(struct page **pages, unsigned nr_gfn, argument
57 page = pages[i / XEN_PFN_PER_PAGE];
71 struct page **pages; member in struct:remap_data
99 struct page *page = info->pages[info->index++];
148 struct page **pages)
163 data.pages = pages;
184 int nr, struct page **pages)
186 xen_for_each_gfn(pages, n
143 xen_xlate_remap_gfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *gfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) argument
183 xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, int nr, struct page **pages) argument
217 struct page **pages; local
267 struct page **pages; member in struct:remap_pfn
[all...]
/linux-master/include/linux/
H A Dballoon_compaction.h5 * Common interface definitions for making balloon pages movable by compaction.
32 * set of exposed rules are satisfied while we are dealing with balloon pages
50 * procedures to find the proper balloon device holding memory pages they'll
55 unsigned long isolated_pages; /* # of isolated pages for migration */
56 spinlock_t pages_lock; /* Protection to pages list */
57 struct list_head pages; /* Pages enqueued & handled to Host */ member in struct:balloon_dev_info
67 struct list_head *pages);
69 struct list_head *pages, size_t n_req_pages);
75 INIT_LIST_HEAD(&balloon->pages);
89 * pages lis
164 balloon_page_push(struct list_head *pages, struct page *page) argument
176 balloon_page_pop(struct list_head *pages) argument
[all...]
/linux-master/fs/squashfs/
H A Dpage_actor.c32 if (actor->next_page == actor->pages)
44 int pages, int length)
51 actor->length = length ? : pages * PAGE_SIZE;
53 actor->pages = pages;
70 if ((actor->next_page == actor->pages) ||
106 struct page **page, int pages, int length)
123 actor->length = length ? : pages * PAGE_SIZE;
125 actor->pages = pages;
43 squashfs_page_actor_init(void **buffer, int pages, int length) argument
105 squashfs_page_actor_init_special(struct squashfs_sb_info *msblk, struct page **page, int pages, int length) argument
[all...]
/linux-master/arch/arm/mach-omap1/
H A Dsram-init.c63 int pages; local
71 pages = PAGE_ALIGN(size) / PAGE_SIZE;
73 set_memory_rw(base, pages);
77 set_memory_rox(base, pages);
92 int pages; local
124 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE;
126 set_memory_rox(base, pages);

Completed in 219 milliseconds

1234567891011>>