/linux-master/drivers/edac/ |
H A D | i82975x_edac.c | 366 u32 cumul_size, nr_pages; local 396 nr_pages = cumul_size - last_cumul_size; 397 if (!nr_pages) 409 dimm->nr_pages = nr_pages / csrow->nr_channels;
|
H A D | armada_xp_edac.c | 253 dimm->nr_pages = 524288; 256 dimm->nr_pages = 65536; 259 dimm->nr_pages = 131072; 262 dimm->nr_pages = 262144; 265 dimm->nr_pages = 1048576; 268 dimm->nr_pages = 2097152;
|
H A D | i82875p_edac.c | 349 u32 cumul_size, nr_pages; local 372 nr_pages = cumul_size - last_cumul_size; 378 dimm->nr_pages = nr_pages / nr_chans;
|
H A D | e7xxx_edac.c | 359 u32 dra, cumul_size, nr_pages; local 390 nr_pages = cumul_size - last_cumul_size; 411 dimm->nr_pages = nr_pages / (drc_chan + 1);
|
H A D | edac_mc.c | 94 if (!dimm->nr_pages) 104 edac_dbg(4, " dimm->nr_pages = 0x%x\n", dimm->nr_pages); 617 u32 nr_pages = 0; local 621 nr_pages += csrow->channels[j]->dimm->nr_pages; 622 if (!nr_pages) 626 if (csrow->channels[j]->dimm->nr_pages) 767 n += dimm->nr_pages; 1039 if (!dimm->nr_pages) [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_builtin.c | 58 struct page *kvm_alloc_hpt_cma(unsigned long nr_pages) argument 60 VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT); 62 return cma_alloc(kvm_cma, nr_pages, order_base_2(HPT_ALIGN_PAGES), 67 void kvm_free_hpt_cma(struct page *page, unsigned long nr_pages) argument 69 cma_release(kvm_cma, page, nr_pages);
|
/linux-master/drivers/video/fbdev/ |
H A D | pvr2fb.c | 646 unsigned int nr_pages; local 653 nr_pages = (count + PAGE_SIZE - 1) >> PAGE_SHIFT; 655 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); 659 ret = pin_user_pages_fast((unsigned long)buf, nr_pages, FOLL_WRITE, pages); 660 if (ret < nr_pages) { 663 * Clamp the unsigned nr_pages to zero so that the 667 nr_pages = 0; 669 nr_pages = ret; 684 end = (unsigned long)page_address(pages[nr_pages]); 685 len = nr_pages << PAGE_SHIF [all...] |
/linux-master/drivers/xen/ |
H A D | gntdev-dmabuf.c | 53 int nr_pages; member in struct:gntdev_dmabuf 196 dmabuf_pages_to_sgt(struct page **pages, unsigned int nr_pages) argument 207 ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0, 208 nr_pages << PAGE_SHIFT, 265 pr_debug("Mapping %d pages for dev %p\n", gntdev_dmabuf->nr_pages, 283 gntdev_dmabuf->nr_pages); 370 gntdev_dmabuf->nr_pages = args->count; 552 gntdev_dmabuf->nr_pages = count; 613 if (attach->dmabuf->size != gntdev_dmabuf->nr_pages << PAGE_SHIFT) { 616 attach->dmabuf->size, gntdev_dmabuf->nr_pages); [all...] |
/linux-master/drivers/hv/ |
H A D | hv_balloon.c | 601 unsigned long nr_pages) 607 while (pfn < start_pfn + nr_pages) { 616 (pfn < start_pfn + nr_pages)) { 651 mem->nr_pages); 1068 unsigned long nr_pages = totalram_pages(); local 1081 if (nr_pages < MB2PAGES(128)) 1082 min_pages = MB2PAGES(8) + (nr_pages >> 1); 1083 else if (nr_pages < MB2PAGES(512)) 1084 min_pages = MB2PAGES(40) + (nr_pages >> 2); 1085 else if (nr_pages < MB2PAGE 600 hv_page_offline_check(unsigned long start_pfn, unsigned long nr_pages) argument [all...] |
/linux-master/mm/ |
H A D | page-writeback.c | 274 unsigned long nr_pages = 0; local 283 nr_pages += zone_page_state(zone, NR_FREE_PAGES); 291 nr_pages -= min(nr_pages, pgdat->totalreserve_pages); 293 nr_pages += node_page_state(pgdat, NR_INACTIVE_FILE); 294 nr_pages += node_page_state(pgdat, NR_ACTIVE_FILE); 296 return nr_pages; 309 unsigned long nr_pages; local 318 nr_pages = zone_page_state(z, NR_FREE_PAGES); 320 nr_pages 487 unsigned long nr_pages = 0; local [all...] |
H A D | page_alloc.c | 559 int nr_pages = 1 << order; local 562 for (i = 1; i < nr_pages; i++) 626 static inline void account_freepages(struct zone *zone, int nr_pages, argument 632 __mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages); 635 __mod_zone_page_state(zone, NR_FREE_CMA_PAGES, nr_pages); 1157 int nr_pages; local 1167 nr_pages = 1 << order; 1178 count -= nr_pages; 1179 pcp->count -= nr_pages; 1218 unsigned int nr_pages local 4490 alloc_pages_bulk_noprof(gfp_t gfp, int preferred_nid, nodemask_t *nodemask, int nr_pages, struct list_head *page_list, struct page **page_array) argument 6530 __alloc_contig_pages(unsigned long start_pfn, unsigned long nr_pages, gfp_t gfp_mask) argument 6539 pfn_range_valid_contig(struct zone *z, unsigned long start_pfn, unsigned long nr_pages) argument 6562 zone_spans_last_pfn(const struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument 6591 alloc_contig_pages_noprof(unsigned long nr_pages, gfp_t gfp_mask, int nid, nodemask_t *nodemask) argument 6629 free_contig_range(unsigned long pfn, unsigned long nr_pages) argument [all...] |
/linux-master/drivers/accel/qaic/ |
H A D | qaic_control.c | 198 unsigned long nr_pages; member in struct:dma_xfer 333 for (i = 0; i < xfer->nr_pages; ++i) 398 unsigned long nr_pages; local 421 nr_pages = need_pages; 424 page_list = kmalloc_array(nr_pages, sizeof(*page_list), GFP_KERNEL | __GFP_NOWARN); 426 nr_pages = nr_pages / 2; 427 if (!nr_pages) 434 ret = get_user_pages_fast(xfer_start_addr, nr_pages, 0, page_list); 437 if (ret != nr_pages) { [all...] |
/linux-master/arch/s390/kvm/ |
H A D | pci.c | 194 static inline void unaccount_mem(unsigned long nr_pages) argument 199 atomic_long_sub(nr_pages, &user->locked_vm); 201 atomic64_sub(nr_pages, ¤t->mm->pinned_vm); 204 static inline int account_mem(unsigned long nr_pages) argument 213 new_pages = cur_pages + nr_pages; 219 atomic64_add(nr_pages, ¤t->mm->pinned_vm);
|
/linux-master/fs/iomap/ |
H A D | direct-io.c | 287 int nr_pages, ret = 0; local 370 nr_pages = bio_iov_vecs_to_alloc(dio->submit.iter, BIO_MAX_VECS); 379 bio = iomap_dio_alloc_bio(iter, dio, nr_pages, bio_opf); 411 nr_pages = bio_iov_vecs_to_alloc(dio->submit.iter, 416 if (nr_pages) 420 } while (nr_pages);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 103 unsigned long nr_pages; local 154 nr_pages = min_t(unsigned long, 163 sg_set_folio(sg, folio, nr_pages * PAGE_SIZE, 0); 166 sg->length += nr_pages * PAGE_SIZE; 168 next_pfn = folio_pfn(folio) + nr_pages; 169 i += nr_pages - 1;
|
/linux-master/arch/x86/mm/ |
H A D | init_64.c | 950 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, argument 955 ret = __add_pages(nid, start_pfn, nr_pages, params); 960 nr_pages << PAGE_SHIFT); 969 unsigned long nr_pages = size >> PAGE_SHIFT; local 973 return add_pages(nid, start_pfn, nr_pages, params); 979 unsigned int nr_pages = 1 << order; local 987 while (nr_pages--) 990 while (nr_pages--) 1262 unsigned long nr_pages = size >> PAGE_SHIFT; local 1264 __remove_pages(start_pfn, nr_pages, altma 1557 register_page_bootmem_memmap(unsigned long section_nr, struct page *start_page, unsigned long nr_pages) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | init_64.c | 336 unsigned long nr_pages, addr; local 353 nr_pages = 1 << page_order; 357 vmem_altmap_free(altmap, nr_pages); 367 while (nr_pages--)
|
/linux-master/drivers/md/ |
H A D | dm-stripe.c | 317 long nr_pages, enum dax_access_mode mode, void **kaddr, 322 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); 326 size_t nr_pages) 330 return dax_zero_page_range(dax_dev, pgoff, nr_pages); 316 stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument 325 stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, size_t nr_pages) argument
|
/linux-master/include/linux/ |
H A D | splice.h | 60 int nr_pages; /* number of populated pages in map */ member in struct:splice_pipe_desc
|
H A D | memremap.h | 102 unsigned long nr_pages, int mf_flags);
|
/linux-master/fs/ |
H A D | file_table.c | 528 unsigned long nr_pages = totalram_pages(); local 529 unsigned long memreserve = (nr_pages - nr_free_pages()) * 3/2; 531 memreserve = min(memreserve, nr_pages - 1); 532 n = ((nr_pages - memreserve) * (PAGE_SIZE / 1024)) / 10;
|
/linux-master/drivers/nvdimm/ |
H A D | pmem.c | 244 long nr_pages, enum dax_access_mode mode, void **kaddr, 249 unsigned int num = PFN_PHYS(nr_pages) >> SECTOR_SHIFT; 273 dev_dbg(pmem->bb.dev, "start sector(%llu), nr_pages(%ld), first_bad(%llu), actual_nr(%ld)\n", 274 sector, nr_pages, first_bad, actual_nr); 285 return nr_pages; 295 size_t nr_pages) 305 pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, 310 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); 435 unsigned long pfn, unsigned long nr_pages, int mf_flags) 440 u64 len = nr_pages << PAGE_SHIF 243 __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument 294 pmem_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, size_t nr_pages) argument 304 pmem_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument 434 pmem_pagemap_memory_failure(struct dev_pagemap *pgmap, unsigned long pfn, unsigned long nr_pages, int mf_flags) argument [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | uffd-common.h | 101 extern unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size;
|
/linux-master/arch/x86/kernel/ |
H A D | machine_kexec_64.c | 524 unsigned int nr_pages; local 534 nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1; 536 return set_pages_ro(page, nr_pages); 538 return set_pages_rw(page, nr_pages);
|
/linux-master/drivers/net/ethernet/chelsio/libcxgb/ |
H A D | libcxgb_ppm.h | 94 int nr_pages; member in struct:cxgbi_task_tag_info 323 int cxgbi_ppm_ppods_reserve(struct cxgbi_ppm *, unsigned short nr_pages,
|