/linux-master/tools/testing/memblock/linux/ |
H A D | mmzone.h | 20 #define MAX_PAGE_ORDER 10 macro 21 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 23 #define pageblock_order MAX_PAGE_ORDER
|
/linux-master/mm/ |
H A D | shuffle.h | 7 #define SHUFFLE_ORDER MAX_PAGE_ORDER
|
H A D | debug_page_alloc.c | 25 if (kstrtoul(buf, 10, &res) < 0 || res > MAX_PAGE_ORDER / 2) {
|
H A D | page_reporting.c | 23 return param_set_uint_minmax(val, kp, 0, MAX_PAGE_ORDER); 373 if (prdev->order > 0 && prdev->order <= MAX_PAGE_ORDER)
|
H A D | page_isolation.c | 229 if (order >= pageblock_order && order < MAX_PAGE_ORDER) { 293 * Free and in-use pages can be as big as MAX_PAGE_ORDER and contain more than one 296 * MAX_PAGE_ORDER = pageblock_order + 1, a MAX_PAGE_ORDER page has two 298 * [ MAX_PAGE_ORDER ] 455 * can be bigger than MAX_PAGE_ORDER, but after it is 463 if (++order > MAX_PAGE_ORDER) 664 * Note: pageblock_nr_pages != MAX_PAGE_ORDER. Then, chunks of free
|
H A D | page_owner.c | 461 if (freepage_order <= MAX_PAGE_ORDER) 696 if (freepage_order <= MAX_PAGE_ORDER) 804 if (order > 0 && order <= MAX_PAGE_ORDER)
|
H A D | page_alloc.c | 731 if (order >= MAX_PAGE_ORDER - 1) 786 while (order < MAX_PAGE_ORDER) { 1308 if (order == MAX_PAGE_ORDER && __free_unaccepted(page)) 1338 * of the pfn range). For example, if the pageblock order is MAX_PAGE_ORDER, which 2029 for (current_order = MAX_PAGE_ORDER; current_order >= min_order; 2067 VM_BUG_ON(current_order > MAX_PAGE_ORDER); 4551 if (WARN_ON_ONCE_GFP(order > MAX_PAGE_ORDER, gfp)) 4843 * This function is also limited by MAX_PAGE_ORDER. 6420 if (++order > MAX_PAGE_ORDER) { 6682 return order <= MAX_PAGE_ORDER; [all...] |
H A D | mm_init.c | 1464 unsigned int order = MAX_PAGE_ORDER; 1647 * The zone's endpoints aren't required to be MAX_PAGE_ORDER 1973 __free_pages_core(page, MAX_PAGE_ORDER); 1977 /* Accept chunks smaller than MAX_PAGE_ORDER upfront */ 2000 * We check if a current MAX_PAGE_ORDER block is valid by only checking the 2158 * Initialize and free pages in MAX_PAGE_ORDER sized increments so that 2301 * Initialize and free pages in MAX_PAGE_ORDER sized increments so 2519 } else if (get_order(size) > MAX_PAGE_ORDER || hashdist) { 2766 * bigger than MAX_PAGE_ORDER unless SPARSEMEM.
|
/linux-master/include/linux/ |
H A D | pageblock-flags.h | 44 #define pageblock_order min_t(unsigned int, HUGETLB_PAGE_ORDER, MAX_PAGE_ORDER) 51 #define pageblock_order MAX_PAGE_ORDER
|
H A D | mmzone.h | 30 #define MAX_PAGE_ORDER 10 macro 32 #define MAX_PAGE_ORDER CONFIG_ARCH_FORCE_MAX_ORDER macro 34 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 38 #define NR_PAGE_ORDERS (MAX_PAGE_ORDER + 1) 944 /* Pages to be accepted. All pages on the list are MAX_PAGE_ORDER */ 1754 #if (MAX_PAGE_ORDER + PAGE_SHIFT) > SECTION_SIZE_BITS 1755 #error Allocator MAX_PAGE_ORDER exceeds SECTION_SIZE
|
H A D | slab.h | 348 #define KMALLOC_SHIFT_MAX (MAX_PAGE_ORDER + PAGE_SHIFT)
|
/linux-master/mm/kmsan/ |
H A D | init.c | 144 .order = MAX_PAGE_ORDER, 214 collect.order = MAX_PAGE_ORDER; 215 for (int i = MAX_PAGE_ORDER; i >= 0; i--) {
|
/linux-master/drivers/gpu/drm/ttm/tests/ |
H A D | ttm_pool_test.c | 111 .order = MAX_PAGE_ORDER + 1, 120 .order = MAX_PAGE_ORDER + 1, 167 if (params->order <= MAX_PAGE_ORDER) { 184 min_t(unsigned int, MAX_PAGE_ORDER,
|
/linux-master/drivers/crypto/hisilicon/ |
H A D | sgl.c | 73 * the pool may allocate a block of memory of size PAGE_SIZE * 2^MAX_PAGE_ORDER, 76 block_size = 1 << (PAGE_SHIFT + MAX_PAGE_ORDER < 32 ? 77 PAGE_SHIFT + MAX_PAGE_ORDER : 31);
|
/linux-master/kernel/events/ |
H A D | ring_buffer.c | 613 if (order > MAX_PAGE_ORDER) 614 order = MAX_PAGE_ORDER; 705 * than: PAGE_SIZE << MAX_PAGE_ORDER; directly bail out in this case. 707 if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER) 824 if (order_base_2(size) > PAGE_SHIFT+MAX_PAGE_ORDER)
|
/linux-master/drivers/base/regmap/ |
H A D | regmap-debugfs.c | 229 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) 230 count = PAGE_SIZE << MAX_PAGE_ORDER; 376 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) 377 count = PAGE_SIZE << MAX_PAGE_ORDER;
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_internal.c | 39 int max_order = MAX_PAGE_ORDER;
|
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | page_alloc.c | 231 pool->max_order = min(MAX_PAGE_ORDER,
|
/linux-master/kernel/dma/ |
H A D | pool.c | 87 /* Cannot allocate larger than MAX_PAGE_ORDER */ 88 order = min(get_order(pool_size), MAX_PAGE_ORDER); 193 * sizes to 128KB per 1GB of memory, min 128KB, max MAX_PAGE_ORDER.
|
/linux-master/fs/ramfs/ |
H A D | file-nommu.c | 73 if (unlikely(order > MAX_PAGE_ORDER))
|
/linux-master/arch/sparc/mm/ |
H A D | tsb.c | 405 if (max_tsb_size > PAGE_SIZE << MAX_PAGE_ORDER) 406 max_tsb_size = PAGE_SIZE << MAX_PAGE_ORDER;
|
/linux-master/arch/um/kernel/ |
H A D | um_arch.c | 376 * Zones have to begin on a 1 << MAX_PAGE_ORDER page boundary, 379 max_physmem &= ~((1 << (PAGE_SHIFT + MAX_PAGE_ORDER)) - 1);
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | iommu_api.c | 100 chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) /
|
/linux-master/drivers/misc/genwqe/ |
H A D | card_utils.c | 213 if (get_order(size) > MAX_PAGE_ORDER) 311 if (get_order(sgl->sgl_size) > MAX_PAGE_ORDER) {
|
/linux-master/tools/perf/ |
H A D | builtin-kmem.c | 286 #define MAX_PAGE_ORDER 11 macro 288 static int order_stats[MAX_PAGE_ORDER][MAX_MIGRATE_TYPES]; 1225 for (o = 0; o < MAX_PAGE_ORDER; o++) {
|