/linux-master/net/9p/ |
H A D | trans_common.h | 7 void p9_release_pages(struct page **pages, int nr_pages);
|
H A D | trans_common.c | 12 * p9_release_pages - Release pages after the transaction. 13 * @pages: array of pages to be put 16 void p9_release_pages(struct page **pages, int nr_pages) argument 21 if (pages[i]) 22 put_page(pages[i]);
|
/linux-master/kernel/dma/ |
H A D | remap.c | 15 return area->pages; 19 * Remaps an array of PAGE_SIZE pages into another vm_area. 22 void *dma_common_pages_remap(struct page **pages, size_t size, argument 27 vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT, 30 find_vm_area(vaddr)->pages = pages; 42 struct page **pages; local 46 pages = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL); 47 if (!pages) 50 pages[ [all...] |
/linux-master/net/ceph/ |
H A D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) argument 19 set_page_dirty_lock(pages[i]); 20 put_page(pages[i]); 22 kvfree(pages); 26 void ceph_release_page_vector(struct page **pages, int num_pages) argument 31 __free_pages(pages[i], 0); 32 kfree(pages); 37 * allocate a vector new pages 41 struct page **pages; local 44 pages 61 ceph_copy_user_to_page_vector(struct page **pages, const void __user *data, loff_t off, size_t len) argument 87 ceph_copy_to_page_vector(struct page **pages, const void *data, loff_t off, size_t len) argument 110 ceph_copy_from_page_vector(struct page **pages, void *data, loff_t off, size_t len) argument 137 ceph_zero_page_vector_range(int off, int len, struct page **pages) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 12 struct sg_table *pages) 18 for_each_sgt_page(page, sgt_iter, pages) { 24 sg_free_table(pages); 25 kfree(pages); 34 struct sg_table *pages; local 41 pages = kmalloc(sizeof(*pages), GFP); 42 if (!pages) 45 if (sg_alloc_table(pages, npages, GFP)) { 46 kfree(pages); 11 huge_free_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument 85 huge_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) argument [all...] |
H A D | mock_dmabuf.h | 14 struct page *pages[]; member in struct:mock_dmabuf
|
/linux-master/fs/erofs/ |
H A D | pcpubuf.c | 6 * per-CPU virtual memory (in pages) in advance to store such inplace I/O 15 struct page **pages; member in struct:erofs_pcpubuf 64 struct page **pages, **oldpages; local 67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL); 68 if (!pages) { 74 pages[i] = erofs_allocpage(&pagepool, GFP_KERNEL); 75 if (!pages[i]) { 77 oldpages = pages; 81 ptr = vmap(pages, nrpage [all...] |
/linux-master/include/xen/ |
H A D | mem-reservation.h | 30 struct page **pages, 34 struct page **pages); 38 struct page **pages, 43 __xenmem_reservation_va_mapping_update(count, pages, frames); 48 struct page **pages) 52 __xenmem_reservation_va_mapping_reset(count, pages); 37 xenmem_reservation_va_mapping_update(unsigned long count, struct page **pages, xen_pfn_t *frames) argument 47 xenmem_reservation_va_mapping_reset(unsigned long count, struct page **pages) argument
|
H A D | balloon.h | 15 /* Number of pages in high- and low-memory balloons. */ 29 int xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages); 30 void xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages);
|
H A D | xen.h | 65 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages); 66 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages); 72 struct page **pages) 74 return xen_alloc_ballooned_pages(nr_pages, pages); 77 struct page **pages) 79 xen_free_ballooned_pages(nr_pages, pages); 71 xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument 76 xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument
|
/linux-master/drivers/iommu/iommufd/ |
H A D | pages.c | 69 * allocation can hold about 26M of 4k pages and 13G of 2M pages in an 163 static void iopt_pages_add_npinned(struct iopt_pages *pages, size_t npages) argument 167 rc = check_add_overflow(pages->npinned, npages, &pages->npinned); 169 WARN_ON(rc || pages->npinned > pages->npages); 172 static void iopt_pages_sub_npinned(struct iopt_pages *pages, size_t npages) argument 176 rc = check_sub_overflow(pages->npinned, npages, &pages 181 iopt_pages_err_unpin(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **page_list) argument 249 iopt_pages_find_domain_area(struct iopt_pages *pages, unsigned long index) argument 575 pages_to_xarray(struct xarray *xa, unsigned long start_index, unsigned long last_index, struct page **pages) argument 615 batch_from_pages(struct pfn_batch *batch, struct page **pages, size_t npages) argument 625 batch_unpin(struct pfn_batch *batch, struct iopt_pages *pages, unsigned int first_page_off, size_t npages) argument 708 pfn_reader_user_init(struct pfn_reader_user *user, struct iopt_pages *pages) argument 721 pfn_reader_user_destroy(struct pfn_reader_user *user, struct iopt_pages *pages) argument 736 pfn_reader_user_pin(struct pfn_reader_user *user, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument 804 incr_user_locked_vm(struct iopt_pages *pages, unsigned long npages) argument 822 decr_user_locked_vm(struct iopt_pages *pages, unsigned long npages) argument 830 update_mm_locked_vm(struct iopt_pages *pages, unsigned long npages, bool inc, struct pfn_reader_user *user) argument 857 do_update_pinned(struct iopt_pages *pages, unsigned long npages, bool inc, struct pfn_reader_user *user) argument 886 update_unpinned(struct iopt_pages *pages) argument 903 pfn_reader_user_update_pinned(struct pfn_reader_user *user, struct iopt_pages *pages) argument 937 struct iopt_pages *pages; member in struct:pfn_reader 961 struct iopt_pages *pages = pfns->pages; local 1062 pfn_reader_init(struct pfn_reader *pfns, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument 1093 struct iopt_pages *pages = pfns->pages; local 1113 struct iopt_pages *pages = pfns->pages; local 1121 pfn_reader_first(struct pfn_reader *pfns, struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument 1144 struct iopt_pages *pages; local 1183 struct iopt_pages *pages = container_of(kref, struct iopt_pages, kref); local 1197 iopt_area_unpin_domain(struct pfn_batch *batch, struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long start_index, unsigned long last_index, unsigned long *unmapped_end_index, unsigned long real_last_index) argument 1257 __iopt_area_unfill_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long last_index) argument 1309 iopt_area_unfill_partial_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain, unsigned long end_index) argument 1342 iopt_area_unfill_domain(struct iopt_area *area, struct iopt_pages *pages, struct iommu_domain *domain) argument 1408 iopt_area_fill_domains(struct iopt_area *area, struct iopt_pages *pages) argument 1492 iopt_area_unfill_domains(struct iopt_area *area, struct iopt_pages *pages) argument 1519 iopt_pages_unpin_xarray(struct pfn_batch *batch, struct iopt_pages *pages, unsigned long start_index, unsigned long end_index) argument 1542 iopt_pages_unfill_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index) argument 1590 iopt_pages_fill_from_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument 1610 iopt_pages_fill_from_domain(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument 1632 iopt_pages_fill_from_mm(struct iopt_pages *pages, struct pfn_reader_user *user, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument 1671 iopt_pages_fill_xarray(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, struct page **out_pages) argument 1743 iopt_pages_rw_slow(struct iopt_pages *pages, unsigned long start_index, unsigned long last_index, unsigned long offset, void *data, unsigned long length, unsigned int flags) argument 1784 iopt_pages_rw_page(struct iopt_pages *pages, unsigned long index, unsigned long offset, void *data, unsigned long length, unsigned int flags) argument 1831 iopt_pages_rw_access(struct iopt_pages *pages, unsigned long start_byte, void *data, unsigned long length, unsigned int flags) argument 1886 iopt_pages_get_exact_access(struct iopt_pages *pages, unsigned long index, unsigned long last) argument 1919 struct iopt_pages *pages = area->pages; local 1974 struct iopt_pages *pages = area->pages; local [all...] |
/linux-master/arch/s390/hypfs/ |
H A D | hypfs_diag.h | 16 void *diag204_get_buffer(enum diag204_format fmt, int *pages); 17 int diag204_store(void *buf, int pages);
|
H A D | hypfs_diag.c | 30 static int diag204_buf_pages; /* number of pages for diag204 data */ 49 * up to 93 pages! 58 void *diag204_get_buffer(enum diag204_format fmt, int *pages) argument 61 *pages = diag204_buf_pages; 65 *pages = 1; 67 *pages = diag204((unsigned long)DIAG204_SUBC_RSI | 69 if (*pages <= 0) 72 diag204_buf = __vmalloc_node(array_size(*pages, PAGE_SIZE), 77 diag204_buf_pages = *pages; 98 int pages, r local 141 diag204_store(void *buf, int pages) argument [all...] |
/linux-master/sound/soc/sof/ |
H A D | sof-utils.c | 29 int i, pages; local 31 pages = snd_sgbuf_aligned_pages(size); 33 dev_dbg(dev, "generating page table for %p size 0x%zx pages %d\n", 34 dmab->area, size, pages); 36 for (i = 0; i < pages; i++) { 71 return pages;
|
/linux-master/mm/ |
H A D | percpu-vm.c | 23 * pcpu_get_pages - get temp pages array 30 * Pointer to temp pages array on success. 34 static struct page **pages; local 35 size_t pages_size = pcpu_nr_units * pcpu_unit_pages * sizeof(pages[0]); 39 if (!pages) 40 pages = pcpu_mem_zalloc(pages_size, GFP_KERNEL); 41 return pages; 45 * pcpu_free_pages - free pages which were allocated for @chunk 46 * @chunk: chunk pages were allocated for 47 * @pages 54 pcpu_free_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument 82 pcpu_alloc_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end, gfp_t gfp) argument 153 pcpu_unmap_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument 193 __pcpu_map_pages(unsigned long addr, struct page **pages, int nr_pages) argument 214 pcpu_map_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end) argument 279 struct page **pages; local 315 struct page **pages; local [all...] |
H A D | gup_test.c | 10 static void put_back_pages(unsigned int cmd, struct page **pages, argument 19 put_page(pages[i]); 25 unpin_user_pages(pages, nr_pages); 29 unpin_user_pages(pages, nr_pages); 32 put_page(pages[i]); 39 static void verify_dma_pinned(unsigned int cmd, struct page **pages, argument 50 folio = page_folio(pages[i]); 53 "pages[%lu] is NOT dma-pinned\n", i)) { 59 "pages[%lu] is NOT pinnable but pinned\n", 69 static void dump_pages_test(struct gup_test *gup, struct page **pages, argument 106 struct page **pages; local 227 struct page **pages; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | pgtable_32.h | 56 * With PAE paging (PTRS_PER_PMD > 1), we allocate PTRS_PER_PGD == 4 pages for 57 * the PMD's in addition to the pages required for the last level pagetables. 60 #define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD) 62 #define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PGD) 66 * Number of possible pages in the lowmem region.
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | ffa.h | 14 int hyp_ffa_init(void *pages);
|
/linux-master/drivers/media/common/videobuf2/ |
H A D | frame_vector.c | 16 * @nr_frames: number of pages / pfns from start to map 18 * @vec: structure which receives pages / pfns of the addresses mapped. 22 * with page frame numbers or page pointers to corresponding pages (choice 24 * belongs to a normal vma, the function grabs reference to each of the pages 29 * The function returns number of pages mapped which may be less than 72 * put_vaddr_frames() - drop references to pages if get_vaddr_frames() acquired 76 * Drop references to pages if get_vaddr_frames() acquired them. We also 82 struct page **pages; local 86 pages = frame_vector_pages(vec); 89 * get_vaddr_frames() got pages bu 114 struct page **pages; local 140 struct page **pages; local [all...] |
/linux-master/fs/isofs/ |
H A D | compress.c | 37 * to one zisofs block. Store the data in the @pages array with @pcount 42 struct page **pages, unsigned poffset, 68 if (!pages[i]) 70 memzero_page(pages[i], 0, PAGE_SIZE); 71 SetPageUptodate(pages[i]); 121 if (pages[curpage]) { 122 stream.next_out = kmap_local_page(pages[curpage]) 174 if (pages[curpage]) { 175 flush_dcache_page(pages[curpage]); 176 SetPageUptodate(pages[curpag 40 zisofs_uncompress_block(struct inode *inode, loff_t block_start, loff_t block_end, int pcount, struct page **pages, unsigned poffset, int *errp) argument 206 zisofs_fill_pages(struct inode *inode, int full_page, int pcount, struct page **pages) argument 313 struct page **pages; local [all...] |
/linux-master/include/trace/events/ |
H A D | tlb.h | 40 TP_PROTO(int reason, unsigned long pages), 41 TP_ARGS(reason, pages), 45 __field(unsigned long, pages) 50 __entry->pages = pages; 53 TP_printk("pages:%ld reason:%s (%d)", 54 __entry->pages,
|
/linux-master/drivers/xen/ |
H A D | xlate_mmu.c | 47 /* Break down the pages in 4KB chunk and call fn for each gfn */ 48 static void xen_for_each_gfn(struct page **pages, unsigned nr_gfn, argument 57 page = pages[i / XEN_PFN_PER_PAGE]; 71 struct page **pages; member in struct:remap_data 99 struct page *page = info->pages[info->index++]; 148 struct page **pages) 163 data.pages = pages; 184 int nr, struct page **pages) 186 xen_for_each_gfn(pages, n 143 xen_xlate_remap_gfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *gfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) argument 183 xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, int nr, struct page **pages) argument 217 struct page **pages; local 267 struct page **pages; member in struct:remap_pfn [all...] |
/linux-master/include/linux/ |
H A D | balloon_compaction.h | 5 * Common interface definitions for making balloon pages movable by compaction. 32 * set of exposed rules are satisfied while we are dealing with balloon pages 50 * procedures to find the proper balloon device holding memory pages they'll 55 unsigned long isolated_pages; /* # of isolated pages for migration */ 56 spinlock_t pages_lock; /* Protection to pages list */ 57 struct list_head pages; /* Pages enqueued & handled to Host */ member in struct:balloon_dev_info 67 struct list_head *pages); 69 struct list_head *pages, size_t n_req_pages); 75 INIT_LIST_HEAD(&balloon->pages); 89 * pages lis 164 balloon_page_push(struct list_head *pages, struct page *page) argument 176 balloon_page_pop(struct list_head *pages) argument [all...] |
/linux-master/fs/squashfs/ |
H A D | page_actor.c | 32 if (actor->next_page == actor->pages) 44 int pages, int length) 51 actor->length = length ? : pages * PAGE_SIZE; 53 actor->pages = pages; 70 if ((actor->next_page == actor->pages) || 106 struct page **page, int pages, int length) 123 actor->length = length ? : pages * PAGE_SIZE; 125 actor->pages = pages; 43 squashfs_page_actor_init(void **buffer, int pages, int length) argument 105 squashfs_page_actor_init_special(struct squashfs_sb_info *msblk, struct page **page, int pages, int length) argument [all...] |
/linux-master/arch/arm/mach-omap1/ |
H A D | sram-init.c | 63 int pages; local 71 pages = PAGE_ALIGN(size) / PAGE_SIZE; 73 set_memory_rw(base, pages); 77 set_memory_rox(base, pages); 92 int pages; local 124 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE; 126 set_memory_rox(base, pages);
|