/linux-master/mm/ |
H A D | fail_page_alloc.c | 24 bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) argument 28 if (order < fail_page_alloc.min_order) 42 return should_fail_ex(&fail_page_alloc.attr, 1 << order, flags); 59 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order);
|
H A D | show_mem.c | 354 unsigned int order; local 366 for (order = 0; order < NR_PAGE_ORDERS; order++) { 367 struct free_area *area = &zone->free_area[order]; 370 nr[order] = area->nr_free; 371 total += nr[order] << order; 373 types[order] = 0; 376 types[order] | [all...] |
H A D | vmstat.c | 1056 unsigned int order; local 1062 for (order = 0; order < NR_PAGE_ORDERS; order++) { 1071 blocks = data_race(zone->free_area[order].nr_free); 1075 info->free_pages += blocks << order; 1078 if (order >= suitable_order) 1080 (order - suitable_order); 1091 static int __fragmentation_index(unsigned int order, struct contig_page_info *info) argument 1093 unsigned long requested = 1UL << order; 1119 extfrag_for_order(struct zone *zone, unsigned int order) argument 1133 fragmentation_index(struct zone *zone, unsigned int order) argument 1479 int order; local 1504 int order, mtype; local 1546 int order; local 2159 unusable_free_index(unsigned int order, struct contig_page_info *info) argument 2180 unsigned int order; local 2230 unsigned int order; local [all...] |
/linux-master/scripts/atomic/ |
H A D | gen-atomic-long.sh | 35 #gen_proto_order_variant(meta, pfx, name, sfx, order, arg...) 42 local order="$1"; shift 44 local atomicname="${pfx}${name}${sfx}${order}" 52 gen_kerneldoc "raw_" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "atomic_long" "long" "$@"
|
/linux-master/arch/powerpc/platforms/cell/ |
H A D | ras.c | 99 int order; member in struct:ptcal_area 107 static int __init cbe_ptcal_enable_on_node(int nid, int order) argument 121 area->order = order; 124 area->order); 134 * page, in order to avoid prefetches in memcpy and similar 155 __free_pages(area->pages, area->order); 166 int order, found_mic = 0; local 179 order = get_order(*size); 184 cbe_ptcal_enable_on_node(of_node_to_nid(np), order); local [all...] |
/linux-master/include/trace/events/ |
H A D | oom.h | 37 int order, 44 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 49 __field( int, order) 60 __entry->order = order; 68 TP_printk("node=%d zone=%-8s order=%d reclaimable=%lu available=%lu min_wmark=%lu no_progress_loops=%d wmark_check=%d", 70 __entry->order, 185 TP_PROTO(int order, 192 TP_ARGS(order, priority, result, retries, max_retries, ret), 195 __field( int, order) [all...] |
/linux-master/arch/arm/lib/ |
H A D | lib1funcs.S | 106 .macro ARM_DIV2_ORDER divisor, order 110 clz \order, \divisor 111 rsb \order, \order, #31 117 movhs \order, #16 118 movlo \order, #0 122 addhs \order, \order, #8 126 addhs \order, \order, # [all...] |
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_cmdbuf.c | 88 int granule_offs, order, ret; local 93 order = order_base_2(ALIGN(size, SUBALLOC_GRANULE) / SUBALLOC_GRANULE); 97 SUBALLOC_GRANULES, order); 121 int order = order_base_2(ALIGN(cmdbuf->size, SUBALLOC_GRANULE) / local 130 order);
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_db.c | 94 struct hns_roce_db *db, int order) 99 for (o = order; o <= 1; ++o) { 112 if (o > order) 113 set_bit(i ^ 1, pgdir->bits[order]); 119 db->order = order; 125 int order) 133 if (!hns_roce_alloc_db_from_pgdir(pgdir, db, order)) 145 WARN_ON(hns_roce_alloc_db_from_pgdir(pgdir, db, order)); 160 o = db->order; 93 hns_roce_alloc_db_from_pgdir(struct hns_roce_db_pgdir *pgdir, struct hns_roce_db *db, int order) argument 124 hns_roce_alloc_db(struct hns_roce_dev *hr_dev, struct hns_roce_db *db, int order) argument [all...] |
/linux-master/drivers/pci/endpoint/ |
H A D | pci-epc-mem.c | 16 * pci_epc_mem_get_order() - determine the allocation order of a memory size 18 * @size: the size for which to get the order 21 * always gets order with a constant PAGE_SIZE. 25 int order; local 31 order = fls(size); 33 order = fls64(size); 35 return order; 186 int order; local 193 order = pci_epc_mem_get_order(mem, align_size); 196 order); 251 int order; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_internal.c | 71 int order = min(fls(npages) - 1, max_order); local 75 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL), 76 order); 79 if (!order--) 83 max_order = order; 86 sg_set_page(sg, page, PAGE_SIZE << order, 0); 89 npages -= 1 << order;
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_proc.c | 29 int order; local 57 /* Find highest order we can actually allocate */ 58 order = 31 - __builtin_clz(num_pages); 61 pages[i] = alloc_pages_node(node, GFP_KERNEL, order); 64 if (!order) { 69 --order; 72 split_page(pages[i], order); 73 counts[i] = 1 << order;
|
/linux-master/kernel/locking/ |
H A D | test-ww_mutex.c | 404 int *order; local 407 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); 408 if (!order) 409 return order; 412 order[n] = n; 417 tmp = order[n]; 418 order[n] = order[r]; 419 order[ 437 int *order; local 496 int *order; local [all...] |
/linux-master/drivers/staging/media/ipu3/ |
H A D | ipu3-dmamap.c | 55 unsigned int order = __fls(order_mask); local 57 order_size = 1U << order; 59 gfp | high_order_gfp : gfp, order); 62 if (!order) 65 split_page(page, order); 69 __free_pages(page, order); 233 unsigned long order, base_pfn; local 239 order = __ffs(IPU3_PAGE_SIZE); 240 base_pfn = max_t(unsigned long, 1, imgu->mmu->aperture_start >> order); 241 init_iova_domain(&imgu->iova_domain, 1UL << order, base_pf [all...] |
/linux-master/kernel/dma/ |
H A D | pool.c | 82 unsigned int order; local 88 order = min(get_order(pool_size), MAX_PAGE_ORDER); 91 pool_size = 1 << (PAGE_SHIFT + order); 93 page = dma_alloc_from_contiguous(NULL, 1 << order, 94 order, false); 96 page = alloc_pages(gfp, order); 97 } while (!page && order-- > 0); 117 1 << order); 130 1 << order); 139 __free_pages(page, order); [all...] |
/linux-master/drivers/gpu/drm/tests/ |
H A D | drm_buddy_test.c | 19 static inline u64 get_size(int order, u64 chunk_size) argument 21 return (1 << order) * chunk_size; 28 unsigned int i, count, *order; local 45 order = drm_random_order(count, &prng); 46 KUNIT_EXPECT_TRUE(test, order); 50 * in some random order allocate within each bias, using various 59 bias_start = order[i] * bias_size; 164 kfree(order); 270 unsigned int order; local 344 * repeat the whole thing, increment the order unti 495 int order, top; local 577 unsigned int order; local 674 int order; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mr.c | 46 static u32 mlx4_buddy_alloc(struct mlx4_buddy *buddy, int order) argument 54 for (o = order; o <= buddy->max_order; ++o) 69 while (o > order) { 78 seg <<= order; local 83 static void mlx4_buddy_free(struct mlx4_buddy *buddy, u32 seg, int order) argument 85 seg >>= order; local 89 while (test_bit(seg ^ 1, buddy->bits[order])) { 90 clear_bit(seg ^ 1, buddy->bits[order]); 91 --buddy->num_free[order]; 93 ++order; 150 __mlx4_alloc_mtt_range(struct mlx4_dev *dev, int order) argument 174 mlx4_alloc_mtt_range(struct mlx4_dev *dev, int order) argument 217 __mlx4_free_mtt_range(struct mlx4_dev *dev, u32 offset, int order) argument 231 mlx4_free_mtt_range(struct mlx4_dev *dev, u32 offset, int order) argument [all...] |
/linux-master/tools/testing/radix-tree/ |
H A D | test.h | 10 unsigned int order; member in struct:item 13 struct item *item_create(unsigned long index, unsigned int order); 36 void iteration_test(unsigned order, unsigned duration);
|
/linux-master/include/linux/ |
H A D | fault-inject.h | 94 bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order); 97 bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order); 99 static inline bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) argument
|
H A D | huge_mm.h | 69 * and including PMD_ORDER, except order-0 (which is not "huge") and order-1 88 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ 89 (!!thp_vma_allowable_orders(vma, vm_flags, tva_flags, BIT(order))) 155 * order-aligned within the file. The hugepage is 156 * guaranteed to be order-aligned within the file, but we must 157 * check that the order-aligned addresses in the VMA map to 158 * order-aligned offsets within the file, else the hugepage will 164 unsigned long addr, int order) 166 unsigned long hpage_size = PAGE_SIZE << order; 163 thp_vma_suitable_order(struct vm_area_struct *vma, unsigned long addr, int order) argument 191 int order; local 284 count_mthp_stat(int order, enum mthp_stat_item item) argument 292 count_mthp_stat(int order, enum mthp_stat_item item) argument 425 thp_vma_suitable_order(struct vm_area_struct *vma, unsigned long addr, int order) argument [all...] |
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | gfp.h | 26 void *hyp_alloc_pages(struct hyp_pool *pool, unsigned short order);
|
/linux-master/arch/x86/kernel/cpu/mce/ |
H A D | genpool.c | 51 * reverse time order, so we need to reverse it. While doing that we can 65 /* squeeze out duplicates while reversing order */ 121 int mce_numrecords, mce_poolsz, order; local 126 order = order_base_2(sizeof(struct mce_evt_llist)); 127 gpool = gen_pool_create(order, -1); 132 mce_poolsz = mce_numrecords * (1 << order);
|
/linux-master/arch/s390/mm/ |
H A D | init.c | 65 unsigned int order; local 70 order = 7; 73 while (order > 2 && (totalram_pages() >> 10) < (1UL << order)) 74 order--; 76 empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); 81 split_page(page, order); 82 for (i = 1 << order; i > 0; i--) { 87 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK;
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_mr.c | 43 int order; member in struct:mthca_mtt 84 static u32 mthca_buddy_alloc(struct mthca_buddy *buddy, int order) argument 92 for (o = order; o <= buddy->max_order; ++o) 107 while (o > order) { 116 seg <<= order; local 121 static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order) argument 123 seg >>= order; local 127 while (test_bit(seg ^ 1, buddy->bits[order])) { 128 __clear_bit(seg ^ 1, buddy->bits[order]); 129 --buddy->num_free[order]; 188 mthca_alloc_mtt_range(struct mthca_dev *dev, int order, struct mthca_buddy *buddy) argument [all...] |
/linux-master/lib/ |
H A D | genalloc.c | 246 int order = pool->min_alloc_order; local 253 end_bit = chunk_size(chunk) >> order; 282 int order = pool->min_alloc_order; local 295 nbits = (size + (1UL << order) - 1) >> order; 302 end_bit = chunk_size(chunk) >> order; 316 addr = chunk->start_addr + ((unsigned long)start_bit << order); 317 size = nbits << order; 491 int order = pool->min_alloc_order; local 501 nbits = (size + (1UL << order) 672 int order; local 699 int order; local [all...] |