/linux-master/net/9p/ |
H A D | trans_common.h | 7 void p9_release_pages(struct page **pages, int nr_pages);
|
H A D | trans_common.c | 14 * @nr_pages: size of array 16 void p9_release_pages(struct page **pages, int nr_pages) argument 20 for (i = 0; i < nr_pages; i++)
|
/linux-master/include/linux/ |
H A D | page_counter.h | 60 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); 61 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); 63 unsigned long nr_pages, 65 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); 66 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); 67 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); 70 unsigned long nr_pages) 72 WRITE_ONCE(counter->high, nr_pages); 75 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); 77 unsigned long *nr_pages); 69 page_counter_set_high(struct page_counter *counter, unsigned long nr_pages) argument [all...] |
H A D | hugetlb_cgroup.h | 127 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, 129 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, 131 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, 134 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, 137 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, 139 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, 142 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, 144 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, 152 unsigned long nr_pages, 162 unsigned long nr_pages, 160 hugetlb_cgroup_uncharge_file_region(struct resv_map *resv, struct file_region *rg, unsigned long nr_pages, bool region_del) argument 207 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) argument 213 hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup **ptr) argument 220 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 227 hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg, struct folio *folio) argument 233 hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, struct folio *folio) argument 238 hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, struct folio *folio) argument 243 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) argument 250 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, struct hugetlb_cgroup *h_cg) argument [all...] |
H A D | memory_hotplug.h | 171 long nr_pages); 173 extern int mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, 175 extern void mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages); 176 extern int online_pages(unsigned long pfn, unsigned long nr_pages, 205 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, 209 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 214 unsigned long nr_pages, struct mhp_params *params) 216 return __add_pages(nid, start_pfn, nr_pages, params); 219 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 330 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages, 213 add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, struct mhp_params *params) argument 339 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument [all...] |
H A D | rmap.h | 198 struct page *page, int nr_pages, enum rmap_level level) 212 VM_WARN_ON_ONCE(nr_pages <= 0); 214 VM_WARN_ON_FOLIO(page_folio(page + nr_pages - 1) != folio, folio); 226 VM_WARN_ON_FOLIO(nr_pages != HPAGE_PMD_NR, folio); 237 void folio_add_anon_rmap_ptes(struct folio *, struct page *, int nr_pages, 245 void folio_add_file_rmap_ptes(struct folio *, struct page *, int nr_pages, 251 void folio_remove_rmap_ptes(struct folio *, struct page *, int nr_pages, 322 struct page *page, int nr_pages, enum rmap_level level) 324 const int orig_nr_pages = nr_pages; 326 __folio_rmap_sanity_checks(folio, page, nr_pages, leve 197 __folio_rmap_sanity_checks(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 321 __folio_dup_file_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 357 folio_dup_file_rmap_ptes(struct folio *folio, struct page *page, int nr_pages) argument 388 __folio_try_dup_anon_rmap(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma, enum rmap_level level) argument 472 folio_try_dup_anon_rmap_ptes(struct folio *folio, struct page *page, int nr_pages, struct vm_area_struct *src_vma) argument 520 __folio_try_share_anon_rmap(struct folio *folio, struct page *page, int nr_pages, enum rmap_level level) argument 674 unsigned long nr_pages; member in struct:page_vma_mapped_walk [all...] |
/linux-master/mm/ |
H A D | cma.h | 47 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages); 48 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages); 49 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages); 52 unsigned long nr_pages) {}; 54 unsigned long nr_pages) {}; 56 unsigned long nr_pages) {}; 51 cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages) argument 53 cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages) argument 55 cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages) argument
|
H A D | page_counter.c | 47 * @nr_pages: number of pages to cancel 49 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) argument 53 new = atomic_long_sub_return(nr_pages, &counter->usage); 55 if (WARN_ONCE(new < 0, "page_counter underflow: %ld nr_pages=%lu\n", 56 new, nr_pages)) { 66 * @nr_pages: number of pages to charge 70 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) argument 77 new = atomic_long_add_return(nr_pages, &c->usage); 91 * @nr_pages: number of pages to charge 98 unsigned long nr_pages, 97 page_counter_try_charge(struct page_counter *counter, unsigned long nr_pages, struct page_counter **fail) argument 153 page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) argument 171 page_counter_set_max(struct page_counter *counter, unsigned long nr_pages) argument 210 page_counter_set_min(struct page_counter *counter, unsigned long nr_pages) argument 227 page_counter_set_low(struct page_counter *counter, unsigned long nr_pages) argument 246 page_counter_memparse(const char *buf, const char *max, unsigned long *nr_pages) argument [all...] |
H A D | percpu-km.c | 55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; local 65 pages = alloc_pages(gfp, order_base_2(nr_pages)); 71 for (i = 0; i < nr_pages; i++) 78 pcpu_chunk_populated(chunk, 0, nr_pages); 89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; local 98 __free_pages(chunk->data, order_base_2(nr_pages)); 109 size_t nr_pages, alloc_pages; local 117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; 118 alloc_pages = roundup_pow_of_two(nr_pages); 120 if (alloc_pages > nr_pages) [all...] |
H A D | gup_test.c | 11 unsigned long nr_pages, unsigned int gup_test_flags) 18 for (i = 0; i < nr_pages; i++) 25 unpin_user_pages(pages, nr_pages); 29 unpin_user_pages(pages, nr_pages); 31 for (i = 0; i < nr_pages; i++) 40 unsigned long nr_pages) 49 for (i = 0; i < nr_pages; i++) { 70 unsigned long nr_pages) 80 if (gup->which_pages[i] > nr_pages) { 104 unsigned long i, nr_pages, add local 10 put_back_pages(unsigned int cmd, struct page **pages, unsigned long nr_pages, unsigned int gup_test_flags) argument 39 verify_dma_pinned(unsigned int cmd, struct page **pages, unsigned long nr_pages) argument 69 dump_pages_test(struct gup_test *gup, struct page **pages, unsigned long nr_pages) argument 224 long nr_pages, cur_pages, addr, remaining_pages; local [all...] |
H A D | sparse.c | 185 unsigned long nr_pages) 188 int end = subsection_map_index(pfn + nr_pages - 1); 193 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) argument 195 int end_sec = pfn_to_section_nr(pfn + nr_pages - 1); 198 if (!nr_pages) 205 pfns = min(nr_pages, PAGES_PER_SECTION 215 nr_pages -= pfns; 219 void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) argument 429 unsigned long nr_pages, int nid, struct vmem_altmap *altmap, 634 unsigned long nr_pages, in 184 subsection_mask_set(unsigned long *map, unsigned long pfn, unsigned long nr_pages) argument 428 __populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument 633 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument 640 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument 656 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) argument 683 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) argument 705 populate_section_memmap(unsigned long pfn, unsigned long nr_pages, int nid, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument 713 depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument 722 unsigned long magic, nr_pages; local 749 clear_subsection_map(unsigned long pfn, unsigned long nr_pages) argument 759 fill_subsection_map(unsigned long pfn, unsigned long nr_pages) argument 781 section_deactivate(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument 830 section_activate(int nid, unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument 892 sparse_add_section(int nid, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument 928 sparse_remove_section(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument [all...] |
H A D | memory_hotplug.c | 59 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); local 68 return pageblock_align(nr_pages); 69 return nr_pages; 299 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) argument 316 if (!IS_ALIGNED(pfn | nr_pages, min_align)) 369 int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, argument 372 const unsigned long end_pfn = pfn + nr_pages; 380 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false)); 387 || vmem_altmap_offset(altmap) > nr_pages) { 394 if (check_pfn_span(pfn, nr_pages)) { 527 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument 573 __remove_pages(unsigned long pfn, unsigned long nr_pages, struct vmem_altmap *altmap) argument 644 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) argument 682 node_states_check_changes_online(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) argument 705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument 716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) argument 750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) argument 843 auto_movable_can_online_movable(int nid, struct memory_group *group, unsigned long nr_pages) argument 897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument 963 auto_movable_zone_for_pfn(int nid, struct memory_group *group, unsigned long pfn, unsigned long nr_pages) argument 1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument 1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) argument 1068 adjust_present_page_count(struct page *page, struct memory_group *group, long nr_pages) argument 1089 mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, struct zone *zone, bool mhp_off_inaccessible) argument 1124 mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages) argument 1147 online_pages(unsigned long pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument 1199 adjust_present_page_count(pfn_to_page(pfn), group, nr_pages); local 1885 node_states_check_changes_offline(unsigned long nr_pages, struct zone *zone, struct memory_notify *arg) argument 1932 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) argument 1944 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument [all...] |
H A D | cma_sysfs.c | 17 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages) argument 19 atomic64_add(nr_pages, &cma->nr_pages_succeeded); 22 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages) argument 24 atomic64_add(nr_pages, &cma->nr_pages_failed); 27 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages) argument 29 atomic64_add(nr_pages, &cma->nr_pages_released);
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | early_alloc.h | 10 void *hyp_early_alloc_contig(unsigned int nr_pages);
|
/linux-master/include/xen/ |
H A D | balloon.h | 29 int xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages); 30 void xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages);
|
H A D | xen.h | 65 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages); 66 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages); 71 static inline int xen_alloc_unpopulated_pages(unsigned int nr_pages, argument 74 return xen_alloc_ballooned_pages(nr_pages, pages); 76 static inline void xen_free_unpopulated_pages(unsigned int nr_pages, argument 79 xen_free_ballooned_pages(nr_pages, pages);
|
/linux-master/fs/iomap/ |
H A D | swapfile.c | 19 unsigned long nr_pages; /* number of pages collected */ member in struct:iomap_swapfile_info 33 unsigned long nr_pages; local 40 if (unlikely(isi->nr_pages >= isi->sis->max)) 42 max_pages = isi->sis->max - isi->nr_pages; 55 nr_pages = next_ppage - first_ppage; 56 nr_pages = min(nr_pages, max_pages); 72 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); 76 isi->nr_pages [all...] |
/linux-master/drivers/firmware/efi/libstub/ |
H A D | relocate.c | 28 unsigned long nr_pages; local 45 nr_pages = size / EFI_PAGE_SIZE; 60 if (desc->num_pages < nr_pages) 74 EFI_LOADER_DATA, nr_pages, &start); 120 unsigned long nr_pages; local 138 nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; 140 EFI_LOADER_DATA, nr_pages, &efi_addr);
|
/linux-master/io_uring/ |
H A D | memmap.c | 16 static void *io_mem_alloc_compound(struct page **pages, int nr_pages, argument 32 for (i = 0; i < nr_pages; i++) 38 static void *io_mem_alloc_single(struct page **pages, int nr_pages, size_t size, argument 44 for (i = 0; i < nr_pages; i++) { 50 ret = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL); 64 int nr_pages; local 67 nr_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; 68 pages = kvmalloc_array(nr_pages, sizeof(struct page *), gfp); 72 ret = io_mem_alloc_compound(pages, nr_pages, size, gfp); 76 ret = io_mem_alloc_single(pages, nr_pages, siz 134 unsigned long start, end, nr_pages; local 171 unsigned int nr_pages; local 234 unsigned long nr_pages = npages; local [all...] |
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | early_alloc.c | 24 void *hyp_early_alloc_contig(unsigned int nr_pages) argument 26 unsigned long size = (nr_pages << PAGE_SHIFT); 29 if (!nr_pages)
|
H A D | setup.c | 38 unsigned long nr_pages; local 42 nr_pages = hyp_vmemmap_pages(sizeof(struct hyp_page)); 43 vmemmap_base = hyp_early_alloc_contig(nr_pages); 47 nr_pages = hyp_vm_table_pages(); 48 vm_table_base = hyp_early_alloc_contig(nr_pages); 52 nr_pages = hyp_s1_pgtable_pages(); 53 hyp_pgt_base = hyp_early_alloc_contig(nr_pages); 57 nr_pages = host_s2_pgtable_pages(); 58 host_s2_pgt_base = hyp_early_alloc_contig(nr_pages); 62 nr_pages 261 unsigned long nr_pages, reserved_pages, pfn; local [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_pkvm.h | 29 unsigned long nr_pages = reg->size >> PAGE_SHIFT; local 33 end = start + nr_pages * vmemmap_entry_size; 57 static inline unsigned long __hyp_pgtable_max_pages(unsigned long nr_pages) argument 64 nr_pages = DIV_ROUND_UP(nr_pages, PTRS_PER_PTE); 65 total += nr_pages;
|
/linux-master/drivers/xen/ |
H A D | unpopulated-alloc.c | 34 static int fill_list(unsigned int nr_pages) argument 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); 154 * @nr_pages: Number of pages 158 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) argument 169 return xen_alloc_ballooned_pages(nr_pages, pages); 172 if (list_count < nr_pages) { 173 ret = fill_list(nr_pages - list_count); 178 for (i = 0; i < nr_pages; i++) { 211 * @nr_pages: Number of pages 214 void xen_free_unpopulated_pages(unsigned int nr_pages, struc argument [all...] |
H A D | balloon.c | 385 static enum bp_state increase_reservation(unsigned long nr_pages) argument 391 if (nr_pages > ARRAY_SIZE(frame_list)) 392 nr_pages = ARRAY_SIZE(frame_list); 395 for (i = 0; i < nr_pages; i++) { 397 nr_pages = i; 405 rc = xenmem_reservation_increase(nr_pages, frame_list); 424 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) argument 432 if (nr_pages > ARRAY_SIZE(frame_list)) 433 nr_pages = ARRAY_SIZE(frame_list); 435 for (i = 0; i < nr_pages; 566 add_ballooned_pages(unsigned int nr_pages) argument 599 xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages) argument 652 xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages) argument [all...] |
/linux-master/net/rds/ |
H A D | info.c | 163 unsigned long nr_pages = 0; local 187 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) 190 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); 195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); 196 if (ret != nr_pages) { 198 nr_pages = ret; 200 nr_pages = 0; 205 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); 238 unpin_user_pages(pages, nr_pages); [all...] |