Searched refs:PAGES_PER_SECTION (Results 1 - 20 of 20) sorted by relevance
/linux-master/mm/ |
H A D | bootmem_info.c | 61 mapsize = sizeof(struct page) * PAGES_PER_SECTION; 90 register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION); 118 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
|
H A D | page_ext.c | 279 table_size = page_ext_size * PAGES_PER_SECTION; 312 table_size = page_ext_size * PAGES_PER_SECTION; 374 for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) 380 end = pfn - PAGES_PER_SECTION; 381 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) 404 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) 409 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) 461 pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) {
|
H A D | sparse.c | 205 pfns = min(nr_pages, PAGES_PER_SECTION 244 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { 419 return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); 425 return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); 526 map = __populate_section_memmap(pfn, PAGES_PER_SECTION, 596 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 649 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); 708 PAGES_PER_SECTION), GFP_KERNEL, nid); 723 nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeo [all...] |
H A D | memory_hotplug.c | 315 min_align = PAGES_PER_SECTION; 771 if (!IS_ALIGNED(start_pfn, PAGES_PER_SECTION)) 773 if (!IS_ALIGNED(start_pfn + nr_pages, PAGES_PER_SECTION)) 993 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 1000 online_pages += PAGES_PER_SECTION; 1118 if (nr_pages >= PAGES_PER_SECTION) 1119 online_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION)); 1133 if (nr_pages >= PAGES_PER_SECTION) 1134 offline_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION)); 1164 !IS_ALIGNED(pfn + nr_pages, PAGES_PER_SECTION))) [all...] |
H A D | mm_init.c | 698 if ((nr_initialised > PAGES_PER_SECTION) && 699 (pfn & (PAGES_PER_SECTION - 1)) == 0) { 965 end_pfn = round_up(end_pfn, PAGES_PER_SECTION); 2225 unsigned long epfn_align = ALIGN(epfn, PAGES_PER_SECTION); 2231 .align = PAGES_PER_SECTION, 2232 .min_chunk = PAGES_PER_SECTION, 2255 * the nearest PAGES_PER_SECTION boundary. So we're adding memory in increments 2257 * PAGES_PER_SECTION * sizeof(struct page) bytes. 2269 unsigned long nr_pages_needed = ALIGN(1 << order, PAGES_PER_SECTION); 2313 if ((first_deferred_pfn ^ spfn) < PAGES_PER_SECTION) [all...] |
H A D | memblock.c | 2076 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); 2101 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) { 2103 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION));
|
H A D | compaction.c | 291 return section_nr_to_pfn(start_nr) + PAGES_PER_SECTION;
|
H A D | page_alloc.c | 358 pfn &= (PAGES_PER_SECTION-1);
|
/linux-master/drivers/xen/ |
H A D | unpopulated-alloc.c | 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); 54 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL);
|
H A D | balloon.c | 249 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); 276 balloon_hotplug = round_up(credit, PAGES_PER_SECTION);
|
/linux-master/drivers/base/ |
H A D | memory.c | 189 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; 260 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; 454 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; 555 unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; 706 const unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
|
H A D | node.c | 801 pfn = round_down(pfn + PAGES_PER_SECTION, 802 PAGES_PER_SECTION) - 1;
|
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | memtrace.c | 98 if (IS_ALIGNED(pfn, PAGES_PER_SECTION))
|
/linux-master/arch/x86/mm/ |
H A D | numa.c | 550 if (pfn_align && pfn_align < PAGES_PER_SECTION) { 553 PFN_PHYS(PAGES_PER_SECTION) >> 20);
|
H A D | init_64.c | 1541 if (end - start < PAGES_PER_SECTION * sizeof(struct page))
|
/linux-master/scripts/gdb/linux/ |
H A D | mm.py | 67 self.PAGES_PER_SECTION = 1 << self.PFN_SECTION_SHIFT 68 self.PAGE_SECTION_MASK = (~(self.PAGES_PER_SECTION - 1)) & ((1 << 64) - 1)
|
/linux-master/fs/proc/ |
H A D | page.c | 34 return round_up(max_pfn, PAGES_PER_SECTION);
|
/linux-master/include/linux/ |
H A D | mmzone.h | 1748 #define PAGES_PER_SECTION (1UL << PFN_SECTION_SHIFT) macro 1749 #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1)) 1767 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
|
/linux-master/drivers/virtio/ |
H A D | virtio_mem.c | 2156 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 2161 rc = virtio_mem_fake_offline(vm, pfn, PAGES_PER_SECTION); 2185 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 2189 virtio_mem_fake_online(pfn, PAGES_PER_SECTION); 2207 pfn += PAGES_PER_SECTION) { 2227 pfn += PAGES_PER_SECTION) {
|
/linux-master/arch/arm64/mm/ |
H A D | mmu.c | 1381 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
|
Completed in 280 milliseconds