Searched refs:nr_pages (Results 1 - 25 of 390) sorted by relevance

1234567891011>>

/linux-master/net/9p/
H A Dtrans_common.h7 void p9_release_pages(struct page **pages, int nr_pages);
H A Dtrans_common.c14 * @nr_pages: size of array
16 void p9_release_pages(struct page **pages, int nr_pages) argument
20 for (i = 0; i < nr_pages; i++)
/linux-master/include/linux/
H A Dpage_counter.h60 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
61 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
63 unsigned long nr_pages,
65 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
66 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages);
67 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages);
70 unsigned long nr_pages)
72 WRITE_ONCE(counter->high, nr_pages);
75 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages);
77 unsigned long *nr_pages);
69 page_counter_set_high(struct page_counter *counter, unsigned long nr_pages) argument
[all...]
H A Dhugetlb_cgroup.h127 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
129 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages,
131 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
134 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages,
137 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages,
139 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages,
142 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
144 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages,
152 unsigned long nr_pages,
162 unsigned long nr_pages,
160 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) argument
207 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) argument
213 hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) argument
220 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument
227 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument
233 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument
238 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument
243 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) argument
250 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) argument
[all...]
H A Dmemory_hotplug.h171 long nr_pages);
173 extern int mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages,
175 extern void mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages);
176 extern int online_pages(unsigned long pfn, unsigned long nr_pages,
205 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages,
209 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
214 unsigned long nr_pages, struct mhp_params *params)
216 return __add_pages(nid, start_pfn, nr_pages, params);
219 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
330 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages,
213 add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, struct mhp_params *params) argument
339 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument
[all...]
H A Drmap.h198 struct page *page, int nr_pages, enum rmap_level level)
212 VM_WARN_ON_ONCE(nr_pages <= 0);
214 VM_WARN_ON_FOLIO(page_folio(page + nr_pages - 1) != folio, folio);
226 VM_WARN_ON_FOLIO(nr_pages != HPAGE_PMD_NR, folio);
237 void folio_add_anon_rmap_ptes(struct folio *, struct page *, int nr_pages,
245 void folio_add_file_rmap_ptes(struct folio *, struct page *, int nr_pages,
251 void folio_remove_rmap_ptes(struct folio *, struct page *, int nr_pages,
322 struct page *page, int nr_pages, enum rmap_level level)
324 const int orig_nr_pages = nr_pages;
326 __folio_rmap_sanity_checks(folio, page, nr_pages, leve
197 __folio_rmap_sanity_checks(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
321 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
357 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument
388 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument
472 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument
520 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument
674 unsigned long nr_pages; member in struct:page_vma_mapped_walk
[all...]
/linux-master/mm/
H A Dcma.h47 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages);
48 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages);
49 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages);
52 unsigned long nr_pages) {};
54 unsigned long nr_pages) {};
56 unsigned long nr_pages) {};
51 cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages) argument
53 cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages) argument
55 cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages) argument
H A Dpage_counter.c47 * @nr_pages: number of pages to cancel
49 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) argument
53 new = atomic_long_sub_return(nr_pages, &counter->usage);
55 if (WARN_ONCE(new < 0, "page_counter underflow: %ld nr_pages=%lu\n",
56 new, nr_pages)) {
66 * @nr_pages: number of pages to charge
70 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) argument
77 new = atomic_long_add_return(nr_pages, &c->usage);
91 * @nr_pages: number of pages to charge
98 unsigned long nr_pages,
97 page_counter_try_charge(struct page_counter *counter, unsigned long nr_pages, struct page_counter **fail) argument
153 page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) argument
171 page_counter_set_max(struct page_counter *counter, unsigned long nr_pages) argument
210 page_counter_set_min(struct page_counter *counter, unsigned long nr_pages) argument
227 page_counter_set_low(struct page_counter *counter, unsigned long nr_pages) argument
246 page_counter_memparse(const char *buf, const char *max, unsigned long *nr_pages) argument
[all...]
H A Dpercpu-km.c55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; local
65 pages = alloc_pages(gfp, order_base_2(nr_pages));
71 for (i = 0; i < nr_pages; i++)
78 pcpu_chunk_populated(chunk, 0, nr_pages);
89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; local
98 __free_pages(chunk->data, order_base_2(nr_pages));
109 size_t nr_pages, alloc_pages; local
117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT;
118 alloc_pages = roundup_pow_of_two(nr_pages);
120 if (alloc_pages > nr_pages)
[all...]
H A Dgup_test.c11 unsigned long nr_pages, unsigned int gup_test_flags)
18 for (i = 0; i < nr_pages; i++)
25 unpin_user_pages(pages, nr_pages);
29 unpin_user_pages(pages, nr_pages);
31 for (i = 0; i < nr_pages; i++)
40 unsigned long nr_pages)
49 for (i = 0; i < nr_pages; i++) {
70 unsigned long nr_pages)
80 if (gup->which_pages[i] > nr_pages) {
104 unsigned long i, nr_pages, add local
10 put_back_pages(unsigned int cmd, struct page **pages, unsigned long nr_pages, unsigned int gup_test_flags) argument
39 verify_dma_pinned(unsigned int cmd, struct page **pages, unsigned long nr_pages) argument
69 dump_pages_test(struct gup_test *gup, struct page **pages, unsigned long nr_pages) argument
224 long nr_pages, cur_pages, addr, remaining_pages; local
[all...]
H A Dsparse.c185 unsigned long nr_pages)
188 int end = subsection_map_index(pfn + nr_pages - 1);
193 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) argument
195 int end_sec = pfn_to_section_nr(pfn + nr_pages - 1);
198 if (!nr_pages)
205 pfns = min(nr_pages, PAGES_PER_SECTION
215 nr_pages -= pfns;
219 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) argument
429 unsigned long nr_pages, int nid, struct vmem_altmap *altmap,
634 unsigned long nr_pages, in
184 subsection_mask_set(unsigned long *map, unsigned long pfn, unsigned long nr_pages) argument
428 __populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument
633 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument
640 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument
656 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) argument
683 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) argument
705 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument
713 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument
722 unsigned long magic, nr_pages; local
749 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) argument
759 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) argument
781 section_deactivate(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument
830 section_activate(int nid, unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument
892 sparse_add_section(int nid, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument
928 sparse_remove_section(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument
[all...]
H A Dmemory_hotplug.c59 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); local
68 return pageblock_align(nr_pages);
69 return nr_pages;
299 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) argument
316 if (!IS_ALIGNED(pfn | nr_pages, min_align))
369 int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, argument
372 const unsigned long end_pfn = pfn + nr_pages;
380 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false));
387 || vmem_altmap_offset(altmap) > nr_pages) {
394 if (check_pfn_span(pfn, nr_pages)) {
527 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument
573 __remove_pages(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument
644 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) argument
682 node_states_check_changes_online(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) argument
705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument
716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) argument
750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) argument
843 auto_movable_can_online_movable(int nid, struct memory_group *group, unsigned long nr_pages) argument
897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument
963 auto_movable_zone_for_pfn(int nid, struct memory_group *group, unsigned long pfn, unsigned long nr_pages) argument
1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument
1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) argument
1068 adjust_present_page_count(struct page *page, struct memory_group *group, long nr_pages) argument
1089 mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, struct zone *zone, bool mhp_off_inaccessible) argument
1124 mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages) argument
1147 online_pages(unsigned long pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument
1199 adjust_present_page_count(pfn_to_page(pfn), group, nr_pages); local
1885 node_states_check_changes_offline(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) argument
1932 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) argument
1944 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument
[all...]
H A Dcma_sysfs.c17 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages) argument
19 atomic64_add(nr_pages, &cma->nr_pages_succeeded);
22 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages) argument
24 atomic64_add(nr_pages, &cma->nr_pages_failed);
27 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages) argument
29 atomic64_add(nr_pages, &cma->nr_pages_released);
/linux-master/arch/arm64/kvm/hyp/include/nvhe/
H A Dearly_alloc.h10 void *hyp_early_alloc_contig(unsigned int nr_pages);
/linux-master/include/xen/
H A Dballoon.h29 int xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages);
30 void xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages);
H A Dxen.h65 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages);
66 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages);
71 static inline int xen_alloc_unpopulated_pages(unsigned int nr_pages, argument
74 return xen_alloc_ballooned_pages(nr_pages, pages);
76 static inline void xen_free_unpopulated_pages(unsigned int nr_pages, argument
79 xen_free_ballooned_pages(nr_pages, pages);
/linux-master/fs/iomap/
H A Dswapfile.c19 unsigned long nr_pages; /* number of pages collected */ member in struct:iomap_swapfile_info
33 unsigned long nr_pages; local
40 if (unlikely(isi->nr_pages >= isi->sis->max))
42 max_pages = isi->sis->max - isi->nr_pages;
55 nr_pages = next_ppage - first_ppage;
56 nr_pages = min(nr_pages, max_pages);
72 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage);
76 isi->nr_pages
[all...]
/linux-master/drivers/firmware/efi/libstub/
H A Drelocate.c28 unsigned long nr_pages; local
45 nr_pages = size / EFI_PAGE_SIZE;
60 if (desc->num_pages < nr_pages)
74 EFI_LOADER_DATA, nr_pages, &start);
120 unsigned long nr_pages; local
138 nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE;
140 EFI_LOADER_DATA, nr_pages, &efi_addr);
/linux-master/io_uring/
H A Dmemmap.c16 static void *io_mem_alloc_compound(struct page **pages, int nr_pages, argument
32 for (i = 0; i < nr_pages; i++)
38 static void *io_mem_alloc_single(struct page **pages, int nr_pages, size_t size, argument
44 for (i = 0; i < nr_pages; i++) {
50 ret = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL);
64 int nr_pages; local
67 nr_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
68 pages = kvmalloc_array(nr_pages, sizeof(struct page *), gfp);
72 ret = io_mem_alloc_compound(pages, nr_pages, size, gfp);
76 ret = io_mem_alloc_single(pages, nr_pages, siz
134 unsigned long start, end, nr_pages; local
171 unsigned int nr_pages; local
234 unsigned long nr_pages = npages; local
[all...]
/linux-master/arch/arm64/kvm/hyp/nvhe/
H A Dearly_alloc.c24 void *hyp_early_alloc_contig(unsigned int nr_pages) argument
26 unsigned long size = (nr_pages << PAGE_SHIFT);
29 if (!nr_pages)
H A Dsetup.c38 unsigned long nr_pages; local
42 nr_pages = hyp_vmemmap_pages(sizeof(struct hyp_page));
43 vmemmap_base = hyp_early_alloc_contig(nr_pages);
47 nr_pages = hyp_vm_table_pages();
48 vm_table_base = hyp_early_alloc_contig(nr_pages);
52 nr_pages = hyp_s1_pgtable_pages();
53 hyp_pgt_base = hyp_early_alloc_contig(nr_pages);
57 nr_pages = host_s2_pgtable_pages();
58 host_s2_pgt_base = hyp_early_alloc_contig(nr_pages);
62 nr_pages
261 unsigned long nr_pages, reserved_pages, pfn; local
[all...]
/linux-master/arch/arm64/include/asm/
H A Dkvm_pkvm.h29 unsigned long nr_pages = reg->size >> PAGE_SHIFT; local
33 end = start + nr_pages * vmemmap_entry_size;
57 static inline unsigned long __hyp_pgtable_max_pages(unsigned long nr_pages) argument
64 nr_pages = DIV_ROUND_UP(nr_pages, PTRS_PER_PTE);
65 total += nr_pages;
/linux-master/drivers/xen/
H A Dunpopulated-alloc.c34 static int fill_list(unsigned int nr_pages) argument
39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION);
154 * @nr_pages: Number of pages
158 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument
169 return xen_alloc_ballooned_pages(nr_pages, pages);
172 if (list_count < nr_pages) {
173 ret = fill_list(nr_pages - list_count);
178 for (i = 0; i < nr_pages; i++) {
211 * @nr_pages: Number of pages
214 void xen_free_unpopulated_pages(unsigned int nr_pages, struc argument
[all...]
H A Dballoon.c385 static enum bp_state increase_reservation(unsigned long nr_pages) argument
391 if (nr_pages > ARRAY_SIZE(frame_list))
392 nr_pages = ARRAY_SIZE(frame_list);
395 for (i = 0; i < nr_pages; i++) {
397 nr_pages = i;
405 rc = xenmem_reservation_increase(nr_pages, frame_list);
424 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) argument
432 if (nr_pages > ARRAY_SIZE(frame_list))
433 nr_pages = ARRAY_SIZE(frame_list);
435 for (i = 0; i < nr_pages;
566 add_ballooned_pages(unsigned int nr_pages) argument
599 xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages) argument
652 xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages) argument
[all...]
/linux-master/net/rds/
H A Dinfo.c163 unsigned long nr_pages = 0; local
187 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK))
190 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL);
195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages);
196 if (ret != nr_pages) {
198 nr_pages = ret;
200 nr_pages = 0;
205 rdsdebug("len %d nr_pages %lu\n", len, nr_pages);
238 unpin_user_pages(pages, nr_pages);
[all...]

Completed in 221 milliseconds

1234567891011>>