Searched refs:order (Results 76 - 100 of 526) sorted by relevance

1234567891011>>

/linux-master/mm/
H A Dpage_isolation.c119 * in MEM_GOING_OFFLINE in order to indicate that these pages
207 unsigned int order; local
220 * approach in order to merge them. Isolation and free will make
224 order = buddy_order(page);
225 if (order >= pageblock_order && order < MAX_PAGE_ORDER) {
227 order, NULL);
229 isolated_page = !!__isolate_free_page(page, order);
258 __putback_isolated_page(page, order, migratetype);
371 int order local
[all...]
H A Dinternal.h315 unsigned int order);
454 * This function returns the order of a free page in the buddy system. In
456 * page from being allocated in parallel and returning garbage as the order.
485 * (c) a page and its buddy have the same order &&
491 * For recording page's order, we use page_private(page).
494 unsigned int order)
499 if (buddy_order(buddy) != order)
518 * 1) Any buddy B1 will have an order O twin B2 which satisfies
521 * For example, if the starting buddy (buddy2) is #8 its order
525 * 2) Any buddy B will have an order
[all...]
H A Dpage_owner.c25 unsigned short order; member in struct:page_owner
63 * sure to signal it in order to avoid recursion.
234 unsigned short order,
242 for (i = 0; i < (1 << order); i++) {
245 page_owner->order = order;
261 unsigned short order,
268 for (i = 0; i < (1 << order); i++) {
282 void __reset_page_owner(struct page *page, unsigned short order) argument
298 __update_page_owner_free_handle(page_ext, handle, order, curren
232 __update_page_owner_handle(struct page_ext *page_ext, depot_stack_handle_t handle, unsigned short order, gfp_t gfp_mask, short last_migrate_reason, u64 ts_nsec, pid_t pid, pid_t tgid, char *comm) argument
259 __update_page_owner_free_handle(struct page_ext *page_ext, depot_stack_handle_t handle, unsigned short order, pid_t pid, pid_t tgid, u64 free_ts_nsec) argument
313 __set_page_owner(struct page *page, unsigned short order, gfp_t gfp_mask) argument
797 unsigned long order = buddy_order_unsafe(page); local
[all...]
H A Dreadahead.c469 pgoff_t mark, unsigned int order, gfp_t gfp)
472 struct folio *folio = filemap_alloc_folio(gfp, order);
476 mark = round_down(mark, 1UL << order);
485 ractl->_nr_pages += 1UL << order;
516 unsigned int order = new_order; local
519 if (index & ((1UL << order) - 1))
520 order = __ffs(index);
522 while (index + (1UL << order) - 1 > limit)
523 order--;
524 err = ra_alloc_folio(ractl, index, mark, order, gf
468 ra_alloc_folio(struct readahead_control *ractl, pgoff_t index, pgoff_t mark, unsigned int order, gfp_t gfp) argument
562 unsigned int order = folio ? folio_order(folio) : 0; local
[all...]
/linux-master/arch/x86/include/asm/
H A Dgart.h61 static inline void gart_set_size_and_enable(struct pci_dev *dev, u32 order) argument
69 ctl = order << 1;
/linux-master/arch/riscv/include/asm/
H A Dkvm_host.h68 unsigned long order; member in struct:kvm_riscv_hfence
295 unsigned long order);
298 unsigned long order);
304 unsigned long order);
309 unsigned long order);
324 unsigned long order);
330 unsigned long order, unsigned long asid);
337 unsigned long order);
/linux-master/arch/um/include/shared/
H A Dkern_util.h22 extern unsigned long alloc_stack(int order, int atomic);
23 extern void free_stack(unsigned long stack, int order);
/linux-master/arch/x86/kernel/
H A Daperture_64.c166 static u32 __init read_agp(int bus, int slot, int func, int cap, u32 *order) argument
184 old_order = *order;
191 *order = 7 - nbits;
192 if ((int)*order < 0) /* < 32MB */
193 *order = 0;
201 * so let double check that order, and lets trust AMD NB settings:
206 if (aper + (32ULL<<(20 + *order)) > 0x100000000ULL) {
208 bus, slot, func, 32 << *order, apsizereg);
209 *order = old_order;
213 bus, slot, func, aper, aper + (32ULL << (*order
234 search_agp_bridge(u32 *order, int *valid_agp) argument
[all...]
H A Dkvmclock.c205 unsigned int order; local
213 order = get_order(ncpus * sizeof(*hvclock_mem));
215 p = alloc_pages(GFP_KERNEL, order);
217 pr_warn("%s: failed to alloc %d pages", __func__, (1U << order));
229 1UL << order);
231 __free_pages(p, order);
238 memset(hvclock_mem, 0, PAGE_SIZE << order);
/linux-master/drivers/s390/char/
H A Dvmcp.c63 int nr_pages, order; local
65 order = get_order(session->bufsize);
68 * For anything below order 3 allocations rely on the buddy
69 * allocator. If such low-order allocations can't be handled
72 if (order > 2)
79 session->response = (char *)__get_free_pages(GFP_KERNEL | __GFP_RETRY_MAYFAIL, order);
84 int nr_pages, order; local
89 order = get_order(session->bufsize);
96 free_pages((unsigned long)session->response, order);
/linux-master/arch/sh/lib/
H A Dmemmove.S17 ! if dest > src, call memcpy (it copies in decreasing order)
/linux-master/tools/power/cpupower/debug/kernel/
H A DMakefile16 - rm -rf Module.symvers modules.order
/linux-master/drivers/gpu/drm/i915/selftests/
H A Di915_random.h54 void i915_random_reorder(unsigned int *order,
/linux-master/drivers/block/
H A Drbd_types.h86 __u8 order; member in struct:rbd_image_header_ondisk::__anon108
/linux-master/include/drm/ttm/
H A Dttm_pool.h44 * @order: the allocation order our pages have
52 unsigned int order; member in struct:ttm_pool_type
68 * @caching: pools for each caching/order
/linux-master/arch/mips/kernel/
H A Dirq.c56 unsigned int order = get_order(IRQ_STACK_SIZE); local
67 void *s = (void *)__get_free_pages(GFP_KERNEL, order);
/linux-master/kernel/dma/
H A Dcoherent.c145 int order = get_order(size); local
155 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order);
200 int order, void *vaddr)
208 bitmap_release_region(mem->bitmap, page, order);
218 * @order: the order of pages allocated
227 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr) argument
231 return __dma_release_from_coherent(mem, order, vaddr);
292 int dma_release_from_global_coherent(int order, void *vaddr) argument
297 return __dma_release_from_coherent(dma_coherent_default_memory, order,
199 __dma_release_from_coherent(struct dma_coherent_mem *mem, int order, void *vaddr) argument
[all...]
/linux-master/arch/riscv/lib/
H A Dclear_page.S14 #define CBOZ_ALT(order, old, new) \
16 ((order) << 16) | RISCV_ISA_EXT_ZICBOZ, \
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_gart.c117 unsigned int order = get_order(adev->gart.table_size); local
130 p = alloc_pages(gfp_flags, order);
135 for (x = 0; x < (1UL << order); x++)
146 __free_pages(p, order);
214 __free_pages(p, order);
228 unsigned int order = get_order(adev->gart.table_size); local
243 for (x = 0; x < (1UL << order); x++)
245 __free_pages(p, order);
/linux-master/tools/mm/
H A Dthpmaps51 def odkb(order):
52 # Convert page order to KB.
53 return (PAGE_SIZE << order) >> 10
273 for index, order in zip(folios, folio_orders):
275 order = int(order)
276 nr = 1 << order
280 stats[anon][align][order] += nr
312 def cont_parse(vma, order, ranges, anons, heads):
317 nr_cont = 1 << order
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/fw/
H A Dpaging.c45 int blk_idx, order, num_of_pages, size; local
71 order = get_order(size);
72 block = alloc_pages(GFP_KERNEL, order);
83 PAGE_SIZE << order,
97 "Paging: allocated 4K(CSS) bytes (order %d) for firmware paging.\n",
98 order);
101 "Paging: allocated 32K bytes (order %d) for firmware paging.\n",
102 order);
/linux-master/fs/ramfs/
H A Dfile-nommu.c66 unsigned order; local
72 order = get_order(newsize);
73 if (unlikely(order > MAX_PAGE_ORDER))
84 pages = alloc_pages(gfp, order);
88 /* split the high-order page into an array of single pages */
89 xpages = 1UL << order;
92 split_page(pages, order);
/linux-master/mm/kasan/
H A Dcommon.c118 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init) argument
126 if (!kasan_sample_page_alloc(order))
131 PAGE_SIZE << order, init); local
132 for (i = 0; i < (1 << order); i++)
138 void __kasan_poison_pages(struct page *page, unsigned int order, bool init) argument
141 kasan_poison(page_address(page), PAGE_SIZE << order, local
462 bool __kasan_mempool_poison_pages(struct page *page, unsigned int order, argument
480 kasan_poison(ptr, PAGE_SIZE << order, KASAN_PAGE_FREE, false);
485 void __kasan_mempool_unpoison_pages(struct page *page, unsigned int order, argument
488 __kasan_unpoison_pages(page, order, fals
[all...]
/linux-master/drivers/media/test-drivers/vimc/
H A Dvimc-debayer.c26 enum vimc_debayer_rgb_colors order[2][2]; member in struct:vimc_debayer_pix_map
71 .order = { { VIMC_DEBAYER_BLUE, VIMC_DEBAYER_GREEN },
76 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_BLUE },
81 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_RED },
86 .order = { { VIMC_DEBAYER_RED, VIMC_DEBAYER_GREEN },
91 .order = { { VIMC_DEBAYER_BLUE, VIMC_DEBAYER_GREEN },
96 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_BLUE },
101 .order = { { VIMC_DEBAYER_GREEN, VIMC_DEBAYER_RED },
106 .order = { { VIMC_DEBAYER_RED, VIMC_DEBAYER_GREEN },
111 .order
[all...]
/linux-master/security/integrity/ima/
H A Dima_crypto.c37 int order; local
40 order = get_order(size);
41 if (order > MAX_PAGE_ORDER)
43 ima_maxorder = order;
44 ima_bufsize = PAGE_SIZE << order;
230 * max_size amount of memory and then splitting that until zero order is
232 * last_warn is set. Last_warn set affects only last allocation of zero order.
242 int order = ima_maxorder; local
245 if (order)
246 order
[all...]

Completed in 186 milliseconds

1234567891011>>