/linux-master/mm/ |
H A D | page_isolation.c | 119 * in MEM_GOING_OFFLINE in order to indicate that these pages 207 unsigned int order; local 220 * approach in order to merge them. Isolation and free will make 224 order = buddy_order(page); 225 if (order >= pageblock_order && order < MAX_PAGE_ORDER) { 227 order, NULL); 229 isolated_page = !!__isolate_free_page(page, order); 258 __putback_isolated_page(page, order, migratetype); 371 int order local [all...] |
H A D | internal.h | 315 unsigned int order); 454 * This function returns the order of a free page in the buddy system. In 456 * page from being allocated in parallel and returning garbage as the order. 485 * (c) a page and its buddy have the same order && 491 * For recording page's order, we use page_private(page). 494 unsigned int order) 499 if (buddy_order(buddy) != order) 518 * 1) Any buddy B1 will have an order O twin B2 which satisfies 521 * For example, if the starting buddy (buddy2) is #8 its order 525 * 2) Any buddy B will have an order [all...] |
H A D | page_owner.c | 25 unsigned short order; member in struct:page_owner 63 * sure to signal it in order to avoid recursion. 234 unsigned short order, 242 for (i = 0; i < (1 << order); i++) { 245 page_owner->order = order; 261 unsigned short order, 268 for (i = 0; i < (1 << order); i++) { 282 void __reset_page_owner(struct page *page, unsigned short order) argument 298 __update_page_owner_free_handle(page_ext, handle, order, curren 232 __update_page_owner_handle(struct page_ext *page_ext, depot_stack_handle_t handle, unsigned short order, gfp_t gfp_mask, short last_migrate_reason, u64 ts_nsec, pid_t pid, pid_t tgid, char *comm) argument 259 __update_page_owner_free_handle(struct page_ext *page_ext, depot_stack_handle_t handle, unsigned short order, pid_t pid, pid_t tgid, u64 free_ts_nsec) argument 313 __set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument 797 unsigned long order = buddy_order_unsafe(page); local [all...] |
H A D | readahead.c | 469 pgoff_t mark, unsigned int order, gfp_t gfp) 472 struct folio *folio = filemap_alloc_folio(gfp, order); 476 mark = round_down(mark, 1UL << order); 485 ractl->_nr_pages += 1UL << order; 516 unsigned int order = new_order; local 519 if (index & ((1UL << order) - 1)) 520 order = __ffs(index); 522 while (index + (1UL << order) - 1 > limit) 523 order--; 524 err = ra_alloc_folio(ractl, index, mark, order, gf 468 ra_alloc_folio(struct readahead_control *ractl, pgoff_t index, pgoff_t mark, unsigned int order, gfp_t gfp) argument 562 unsigned int order = folio ? folio_order(folio) : 0; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | gart.h | 61 static inline void gart_set_size_and_enable(struct pci_dev *dev, u32 order) argument 69 ctl = order << 1;
|
/linux-master/arch/riscv/include/asm/ |
H A D | kvm_host.h | 68 unsigned long order; member in struct:kvm_riscv_hfence 295 unsigned long order); 298 unsigned long order); 304 unsigned long order); 309 unsigned long order); 324 unsigned long order); 330 unsigned long order, unsigned long asid); 337 unsigned long order);
|
/linux-master/arch/um/include/shared/ |
H A D | kern_util.h | 22 extern unsigned long alloc_stack(int order, int atomic); 23 extern void free_stack(unsigned long stack, int order);
|
/linux-master/arch/x86/kernel/ |
H A D | aperture_64.c | 166 static u32 __init read_agp(int bus, int slot, int func, int cap, u32 *order) argument 184 old_order = *order; 191 *order = 7 - nbits; 192 if ((int)*order < 0) /* < 32MB */ 193 *order = 0; 201 * so let double check that order, and lets trust AMD NB settings: 206 if (aper + (32ULL<<(20 + *order)) > 0x100000000ULL) { 208 bus, slot, func, 32 << *order, apsizereg); 209 *order = old_order; 213 bus, slot, func, aper, aper + (32ULL << (*order 234 search_agp_bridge(u32 *order, int *valid_agp) argument [all...] |
H A D | kvmclock.c | 205 unsigned int order; local 213 order = get_order(ncpus * sizeof(*hvclock_mem)); 215 p = alloc_pages(GFP_KERNEL, order); 217 pr_warn("%s: failed to alloc %d pages", __func__, (1U << order)); 229 1UL << order); 231 __free_pages(p, order); 238 memset(hvclock_mem, 0, PAGE_SIZE << order);
|
/linux-master/drivers/s390/char/ |
H A D | vmcp.c | 63 int nr_pages, order; local 65 order = get_order(session->bufsize); 68 * For anything below order 3 allocations rely on the buddy 69 * allocator. If such low-order allocations can't be handled 72 if (order > 2) 79 session->response = (char *)__get_free_pages(GFP_KERNEL | __GFP_RETRY_MAYFAIL, order); 84 int nr_pages, order; local 89 order = get_order(session->bufsize); 96 free_pages((unsigned long)session->response, order);
|
/linux-master/arch/sh/lib/ |
H A D | memmove.S | 17 ! if dest > src, call memcpy (it copies in decreasing order)
|
/linux-master/tools/power/cpupower/debug/kernel/ |
H A D | Makefile | 16 - rm -rf Module.symvers modules.order
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_random.h | 54 void i915_random_reorder(unsigned int *order,
|
/linux-master/drivers/block/ |
H A D | rbd_types.h | 86 __u8 order; member in struct:rbd_image_header_ondisk::__anon108
|
/linux-master/include/drm/ttm/ |
H A D | ttm_pool.h | 44 * @order: the allocation order our pages have 52 unsigned int order; member in struct:ttm_pool_type 68 * @caching: pools for each caching/order
|
/linux-master/arch/mips/kernel/ |
H A D | irq.c | 56 unsigned int order = get_order(IRQ_STACK_SIZE); local 67 void *s = (void *)__get_free_pages(GFP_KERNEL, order);
|
/linux-master/kernel/dma/ |
H A D | coherent.c | 145 int order = get_order(size); local 155 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order); 200 int order, void *vaddr) 208 bitmap_release_region(mem->bitmap, page, order); 218 * @order: the order of pages allocated 227 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr) argument 231 return __dma_release_from_coherent(mem, order, vaddr); 292 int dma_release_from_global_coherent(int order, void *vaddr) argument 297 return __dma_release_from_coherent(dma_coherent_default_memory, order, 199 __dma_release_from_coherent(struct dma_coherent_mem *mem, int order, void *vaddr) argument [all...] |
/linux-master/arch/riscv/lib/ |
H A D | clear_page.S | 14 #define CBOZ_ALT(order, old, new) \ 16 ((order) << 16) | RISCV_ISA_EXT_ZICBOZ, \
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 117 unsigned int order = get_order(adev->gart.table_size); local 130 p = alloc_pages(gfp_flags, order); 135 for (x = 0; x < (1UL << order); x++) 146 __free_pages(p, order); 214 __free_pages(p, order); 228 unsigned int order = get_order(adev->gart.table_size); local 243 for (x = 0; x < (1UL << order); x++) 245 __free_pages(p, order);
|
/linux-master/tools/mm/ |
H A D | thpmaps | 51 def odkb(order): 52 # Convert page order to KB. 53 return (PAGE_SIZE << order) >> 10 273 for index, order in zip(folios, folio_orders): 275 order = int(order) 276 nr = 1 << order 280 stats[anon][align][order] += nr 312 def cont_parse(vma, order, ranges, anons, heads): 317 nr_cont = 1 << order [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/fw/ |
H A D | paging.c | 45 int blk_idx, order, num_of_pages, size; local 71 order = get_order(size); 72 block = alloc_pages(GFP_KERNEL, order); 83 PAGE_SIZE << order, 97 "Paging: allocated 4K(CSS) bytes (order %d) for firmware paging.\n", 98 order); 101 "Paging: allocated 32K bytes (order %d) for firmware paging.\n", 102 order);
|
/linux-master/fs/ramfs/ |
H A D | file-nommu.c | 66 unsigned order; local 72 order = get_order(newsize); 73 if (unlikely(order > MAX_PAGE_ORDER)) 84 pages = alloc_pages(gfp, order); 88 /* split the high-order page into an array of single pages */ 89 xpages = 1UL << order; 92 split_page(pages, order);
|
/linux-master/mm/kasan/ |
H A D | common.c | 118 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init) argument 126 if (!kasan_sample_page_alloc(order)) 131 PAGE_SIZE << order, init); local 132 for (i = 0; i < (1 << order); i++) 138 void __kasan_poison_pages(struct page *page, unsigned int order, bool init) argument 141 kasan_poison(page_address(page), PAGE_SIZE << order, local 462 bool __kasan_mempool_poison_pages(struct page *page, unsigned int order, argument 480 kasan_poison(ptr, PAGE_SIZE << order, KASAN_PAGE_FREE, false); 485 void __kasan_mempool_unpoison_pages(struct page *page, unsigned int order, argument 488 __kasan_unpoison_pages(page, order, fals [all...] |
/linux-master/drivers/media/test-drivers/vimc/ |
H A D | vimc-debayer.c | 26 enum vimc_debayer_rgb_colors order[2][2]; member in struct:vimc_debayer_pix_map 71 .order = { { VIMC_DEBAYER_BLUE, VIMC_DEBAYER_GREEN }, 76 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_BLUE }, 81 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_RED }, 86 .order = { { VIMC_DEBAYER_RED, VIMC_DEBAYER_GREEN }, 91 .order = { { VIMC_DEBAYER_BLUE, VIMC_DEBAYER_GREEN }, 96 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_BLUE }, 101 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_RED }, 106 .order = { { VIMC_DEBAYER_RED, VIMC_DEBAYER_GREEN }, 111 .order [all...] |
/linux-master/security/integrity/ima/ |
H A D | ima_crypto.c | 37 int order; local 40 order = get_order(size); 41 if (order > MAX_PAGE_ORDER) 43 ima_maxorder = order; 44 ima_bufsize = PAGE_SIZE << order; 230 * max_size amount of memory and then splitting that until zero order is 232 * last_warn is set. Last_warn set affects only last allocation of zero order. 242 int order = ima_maxorder; local 245 if (order) 246 order [all...] |