/linux-master/include/linux/ |
H A D | compaction.h | 88 unsigned int order, unsigned int alloc_flags, 99 int alloc_flags);
|
H A D | mmzone.h | 1434 int highest_zoneidx, unsigned int alloc_flags, 1438 unsigned int alloc_flags);
|
/linux-master/mm/ |
H A D | page_alloc.c | 1539 unsigned int alloc_flags) 1552 if (alloc_flags & ALLOC_NO_WATERMARKS) 1766 unsigned int alloc_flags, int start_type, bool whole_block) 1792 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) 2007 unsigned int alloc_flags) 2021 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT) 2072 steal_suitable_fallback(zone, page, alloc_flags, start_migratetype, 2088 unsigned int alloc_flags) 2098 if (alloc_flags & ALLOC_CMA && 2109 if (alloc_flags 1538 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) argument 1765 steal_suitable_fallback(struct zone *zone, struct page *page, unsigned int alloc_flags, int start_type, bool whole_block) argument 2006 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) argument 2087 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) argument 2124 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) argument 2721 rmqueue_buddy(struct zone *preferred_zone, struct zone *zone, unsigned int order, unsigned int alloc_flags, int migratetype) argument 2813 __rmqueue_pcplist(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags, struct per_cpu_pages *pcp, struct list_head *list) argument 2844 rmqueue_pcplist(struct zone *preferred_zone, struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) argument 2891 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) argument 2932 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) argument 2963 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) argument 3041 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) argument 3048 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) argument 3131 unsigned int alloc_flags; local 3161 gfp_to_alloc_flags_cma(gfp_t gfp_mask, unsigned int alloc_flags) argument 3176 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) argument 3400 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) argument 3515 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) argument 3574 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) argument 3640 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) argument 3649 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) argument 3787 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) argument 3842 unsigned int alloc_flags = ALLOC_WMARK_MIN | ALLOC_CPUSET; local 3941 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) argument 4053 unsigned int alloc_flags; local 4324 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_gfp, unsigned int *alloc_flags) argument 4400 unsigned int alloc_flags = ALLOC_WMARK_LOW; local 4543 unsigned int alloc_flags = ALLOC_WMARK_LOW; local [all...] |
H A D | compaction.c | 2406 * watermark and alloc_flags have to match, or be more pessimistic than 2408 * compactor's alloc_flags, as they are not relevant for freepage 2471 int alloc_flags) 2509 int highest_zoneidx, unsigned int alloc_flags) 2513 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); 2515 alloc_flags)) 2553 cc->alloc_flags); 2763 unsigned int alloc_flags, int highest_zoneidx, 2774 .alloc_flags = alloc_flags, 2470 compaction_zonelist_suitable(struct alloc_context *ac, int order, int alloc_flags) argument 2508 compaction_suit_allocation_order(struct zone *zone, unsigned int order, int highest_zoneidx, unsigned int alloc_flags) argument 2761 compact_zone_order(struct zone *zone, int order, gfp_t gfp_mask, enum compact_priority prio, unsigned int alloc_flags, int highest_zoneidx, struct page **capture) argument 2826 try_to_compact_pages(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, struct page **capture) argument [all...] |
H A D | internal.h | 600 const unsigned int alloc_flags; /* alloc flags of a direct compactor */ member in struct:compact_control
|
/linux-master/drivers/base/regmap/ |
H A D | regcache-maple.c | 77 map->alloc_flags); 95 ret = mas_store_gfp(&mas, entry, map->alloc_flags); 137 map->alloc_flags); 151 map->alloc_flags); 165 ret = mas_store_gfp(&mas, lower, map->alloc_flags); 173 ret = mas_store_gfp(&mas, upper, map->alloc_flags); 207 buf = kmalloc(val_bytes * (max - min), map->alloc_flags); 323 entry = kcalloc(last - first + 1, sizeof(unsigned long), map->alloc_flags); 334 ret = mas_store_gfp(&mas, entry, map->alloc_flags);
|
H A D | regcache-rbtree.c | 280 map->alloc_flags); 289 map->alloc_flags); 323 rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags); 349 map->alloc_flags); 355 map->alloc_flags);
|
H A D | internal.h | 65 gfp_t alloc_flags; member in struct:regmap
|
H A D | regmap.c | 755 map->alloc_flags = GFP_ATOMIC; 757 map->alloc_flags = GFP_KERNEL; 2350 wval = kmemdup(val, val_count * val_bytes, map->alloc_flags);
|
/linux-master/fs/xfs/ |
H A D | xfs_extent_busy.h | 74 unsigned busy_gen, uint32_t alloc_flags);
|
H A D | xfs_extent_busy.c | 610 uint32_t alloc_flags) 621 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) 627 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) 606 xfs_extent_busy_flush( struct xfs_trans *tp, struct xfs_perag *pag, unsigned busy_gen, uint32_t alloc_flags) argument
|
/linux-master/lib/ |
H A D | stackdepot.c | 586 gfp_t alloc_flags, 632 alloc_flags &= ~GFP_ZONEMASK; 633 alloc_flags &= (GFP_ATOMIC | GFP_KERNEL | __GFP_NOLOCKDEP); 634 alloc_flags |= __GFP_NOWARN; 635 page = alloc_pages(alloc_flags, DEPOT_POOL_ORDER); 683 gfp_t alloc_flags) 685 return stack_depot_save_flags(entries, nr_entries, alloc_flags, 584 stack_depot_save_flags(unsigned long *entries, unsigned int nr_entries, gfp_t alloc_flags, depot_flags_t depot_flags) argument 681 stack_depot_save(unsigned long *entries, unsigned int nr_entries, gfp_t alloc_flags) argument
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_amdkfd_gpuvm.c | 294 u32 alloc_flags = bo->kfd_bo->alloc_flags; local 297 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags, 324 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) 504 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) 506 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) 549 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? 651 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP); 657 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? 716 mem->alloc_flags 1698 u64 alloc_flags; local [all...] |
H A D | amdgpu_amdkfd.h | 80 uint32_t alloc_flags; member in struct:kgd_mem
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_alloc.c | 1578 uint32_t alloc_flags) 1598 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; 1663 acur.busy_gen, alloc_flags); 1667 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; 1693 uint32_t alloc_flags) 1708 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; 1779 busy_gen, alloc_flags); 1783 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; 1878 busy_gen, alloc_flags); 1882 alloc_flags 1576 xfs_alloc_ag_vextent_near( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument 1691 xfs_alloc_ag_vextent_size( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument 2788 xfs_alloc_fix_freelist( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument 3483 xfs_alloc_vextent_prepare_ag( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument 3616 uint32_t alloc_flags = 0; local 3661 xfs_alloc_vextent_iterate_ags( struct xfs_alloc_arg *args, xfs_agnumber_t minimum_agno, xfs_agnumber_t start_agno, xfs_agblock_t target_agbno, uint32_t alloc_flags) argument 3742 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; local 3796 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; local 3868 uint32_t alloc_flags = 0; local [all...] |
H A D | xfs_alloc.h | 199 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
|
/linux-master/drivers/md/ |
H A D | dm-zoned-reclaim.c | 286 int alloc_flags = DMZ_ALLOC_SEQ; local 292 alloc_flags | DMZ_ALLOC_RECLAIM); 293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { 294 alloc_flags = DMZ_ALLOC_RND;
|
H A D | dm-zoned-metadata.c | 2056 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; local 2071 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); 2168 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; local 2177 bzone = dmz_alloc_zone(zmd, 0, alloc_flags);
|
/linux-master/fs/xfs/scrub/ |
H A D | repair.h | 54 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
|
H A D | repair.c | 406 int alloc_flags) 416 return xfs_alloc_fix_freelist(&args, alloc_flags); 404 xrep_fix_freelist( struct xfs_scrub *sc, int alloc_flags) argument
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_chardev.c | 1932 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags; 1935 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { 1943 if (bo_bucket->alloc_flags 1946 bo_bucket->alloc_flags & 1955 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) 1958 else if (bo_bucket->alloc_flags & 1971 "gpu_id = 0x%x alloc_flags = 0x%x idr_handle = 0x%x", 1976 bo_bucket->alloc_flags, 2000 if (bo_buckets[bo_index].alloc_flags [all...] |
/linux-master/fs/btrfs/ |
H A D | block-group.c | 2901 u64 alloc_flags; local 2949 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); 2950 if (alloc_flags != cache->flags) { 2951 ret = btrfs_chunk_alloc(trans, alloc_flags, 2979 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); 2980 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); 2996 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); 2998 check_system_chunk(trans, alloc_flags); 3827 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); local 3829 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORC [all...] |
/linux-master/drivers/iommu/ |
H A D | dma-iommu.c | 916 gfp_t alloc_flags = gfp; local 920 alloc_flags |= __GFP_NORETRY; 921 page = alloc_pages_node(nid, alloc_flags, order);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_object.h | 49 unsigned alloc_flags);
|
/linux-master/include/uapi/linux/ |
H A D | kfd_ioctl.h | 635 __u32 alloc_flags; member in struct:kfd_criu_bo_bucket
|