Searched refs:alloc_flags (Results 1 - 25 of 27) sorted by relevance

12

/linux-master/include/linux/
H A Dcompaction.h88 unsigned int order, unsigned int alloc_flags,
99 int alloc_flags);
H A Dmmzone.h1437 int highest_zoneidx, unsigned int alloc_flags,
1441 unsigned int alloc_flags);
/linux-master/mm/
H A Dpage_alloc.c1474 unsigned int alloc_flags)
1487 if (alloc_flags & ALLOC_NO_WATERMARKS)
1846 unsigned int alloc_flags, bool whole_block)
1874 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD))
2095 unsigned int alloc_flags)
2109 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT)
2162 start_migratetype, alloc_flags, can_steal);
2176 unsigned int alloc_flags)
2186 if (alloc_flags & ALLOC_CMA &&
2197 if (alloc_flags
1473 prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags) argument
1844 steal_suitable_fallback(struct zone *zone, struct page *page, int current_order, int order, int start_type, unsigned int alloc_flags, bool whole_block) argument
2094 __rmqueue_fallback(struct zone *zone, int order, int start_migratetype, unsigned int alloc_flags) argument
2175 __rmqueue(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) argument
2212 rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long count, struct list_head *list, int migratetype, unsigned int alloc_flags) argument
2808 rmqueue_buddy(struct zone *preferred_zone, struct zone *zone, unsigned int order, unsigned int alloc_flags, int migratetype) argument
2898 __rmqueue_pcplist(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags, struct per_cpu_pages *pcp, struct list_head *list) argument
2929 rmqueue_pcplist(struct zone *preferred_zone, struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags) argument
2976 rmqueue(struct zone *preferred_zone, struct zone *zone, unsigned int order, gfp_t gfp_flags, unsigned int alloc_flags, int migratetype) argument
3017 __zone_watermark_unusable_free(struct zone *z, unsigned int order, unsigned int alloc_flags) argument
3048 __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, long free_pages) argument
3126 zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags) argument
3133 zone_watermark_fast(struct zone *z, unsigned int order, unsigned long mark, int highest_zoneidx, unsigned int alloc_flags, gfp_t gfp_mask) argument
3216 unsigned int alloc_flags; local
3246 gfp_to_alloc_flags_cma(gfp_t gfp_mask, unsigned int alloc_flags) argument
3261 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, const struct alloc_context *ac) argument
3485 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac) argument
3600 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) argument
3659 should_compact_retry(struct alloc_context *ac, int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) argument
3725 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, enum compact_result *compact_result) argument
3734 should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_flags, enum compact_result compact_result, enum compact_priority *compact_priority, int *compaction_retries) argument
3872 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, unsigned long *did_some_progress) argument
3927 unsigned int alloc_flags = ALLOC_WMARK_MIN | ALLOC_CPUSET; local
4026 should_reclaim_retry(gfp_t gfp_mask, unsigned order, struct alloc_context *ac, int alloc_flags, bool did_some_progress, int *no_progress_loops) argument
4138 unsigned int alloc_flags; local
4409 prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid, nodemask_t *nodemask, struct alloc_context *ac, gfp_t *alloc_gfp, unsigned int *alloc_flags) argument
4485 unsigned int alloc_flags = ALLOC_WMARK_LOW; local
4628 unsigned int alloc_flags = ALLOC_WMARK_LOW; local
[all...]
H A Dcompaction.c2411 * watermark and alloc_flags have to match, or be more pessimistic than
2413 * compactor's alloc_flags, as they are not relevant for freepage
2476 int alloc_flags)
2514 int highest_zoneidx, unsigned int alloc_flags)
2518 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK);
2520 alloc_flags))
2558 cc->alloc_flags);
2768 unsigned int alloc_flags, int highest_zoneidx,
2779 .alloc_flags = alloc_flags,
2475 compaction_zonelist_suitable(struct alloc_context *ac, int order, int alloc_flags) argument
2513 compaction_suit_allocation_order(struct zone *zone, unsigned int order, int highest_zoneidx, unsigned int alloc_flags) argument
2766 compact_zone_order(struct zone *zone, int order, gfp_t gfp_mask, enum compact_priority prio, unsigned int alloc_flags, int highest_zoneidx, struct page **capture) argument
2831 try_to_compact_pages(gfp_t gfp_mask, unsigned int order, unsigned int alloc_flags, const struct alloc_context *ac, enum compact_priority prio, struct page **capture) argument
[all...]
H A Dinternal.h698 const unsigned int alloc_flags; /* alloc flags of a direct compactor */
/linux-master/drivers/base/regmap/
H A Dregcache-maple.c77 map->alloc_flags);
95 ret = mas_store_gfp(&mas, entry, map->alloc_flags);
137 map->alloc_flags);
151 map->alloc_flags);
165 ret = mas_store_gfp(&mas, lower, map->alloc_flags);
173 ret = mas_store_gfp(&mas, upper, map->alloc_flags);
207 buf = kmalloc(val_bytes * (max - min), map->alloc_flags);
323 entry = kcalloc(last - first + 1, sizeof(unsigned long), map->alloc_flags);
334 ret = mas_store_gfp(&mas, entry, map->alloc_flags);
H A Dregcache-rbtree.c280 map->alloc_flags);
289 map->alloc_flags);
323 rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags);
349 map->alloc_flags);
355 map->alloc_flags);
H A Dinternal.h65 gfp_t alloc_flags; member in struct:regmap
H A Dregmap.c755 map->alloc_flags = GFP_ATOMIC;
757 map->alloc_flags = GFP_KERNEL;
2350 wval = kmemdup(val, val_count * val_bytes, map->alloc_flags);
/linux-master/fs/xfs/
H A Dxfs_extent_busy.h74 unsigned busy_gen, uint32_t alloc_flags);
H A Dxfs_extent_busy.c602 uint32_t alloc_flags)
613 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH)
619 if (alloc_flags & XFS_ALLOC_FLAG_FREEING)
598 xfs_extent_busy_flush( struct xfs_trans *tp, struct xfs_perag *pag, unsigned busy_gen, uint32_t alloc_flags) argument
/linux-master/lib/
H A Dstackdepot.c586 gfp_t alloc_flags,
632 alloc_flags &= ~GFP_ZONEMASK;
633 alloc_flags &= (GFP_ATOMIC | GFP_KERNEL | __GFP_NOLOCKDEP);
634 alloc_flags |= __GFP_NOWARN;
635 page = alloc_pages(alloc_flags, DEPOT_POOL_ORDER);
683 gfp_t alloc_flags)
685 return stack_depot_save_flags(entries, nr_entries, alloc_flags,
584 stack_depot_save_flags(unsigned long *entries, unsigned int nr_entries, gfp_t alloc_flags, depot_flags_t depot_flags) argument
681 stack_depot_save(unsigned long *entries, unsigned int nr_entries, gfp_t alloc_flags) argument
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gpuvm.c294 u32 alloc_flags = bo->kfd_bo->alloc_flags; local
297 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags,
324 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR)
504 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE)
506 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE)
549 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
651 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP);
657 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
716 mem->alloc_flags
1698 u64 alloc_flags; local
[all...]
H A Damdgpu_amdkfd.h80 uint32_t alloc_flags; member in struct:kgd_mem
/linux-master/fs/xfs/libxfs/
H A Dxfs_alloc.c1578 uint32_t alloc_flags)
1598 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH;
1663 acur.busy_gen, alloc_flags);
1667 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH;
1693 uint32_t alloc_flags)
1708 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH;
1779 busy_gen, alloc_flags);
1783 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH;
1878 busy_gen, alloc_flags);
1882 alloc_flags
1576 xfs_alloc_ag_vextent_near( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument
1691 xfs_alloc_ag_vextent_size( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument
2788 xfs_alloc_fix_freelist( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument
3483 xfs_alloc_vextent_prepare_ag( struct xfs_alloc_arg *args, uint32_t alloc_flags) argument
3616 uint32_t alloc_flags = 0; local
3661 xfs_alloc_vextent_iterate_ags( struct xfs_alloc_arg *args, xfs_agnumber_t minimum_agno, xfs_agnumber_t start_agno, xfs_agblock_t target_agbno, uint32_t alloc_flags) argument
3742 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; local
3796 uint32_t alloc_flags = XFS_ALLOC_FLAG_TRYLOCK; local
3868 uint32_t alloc_flags = 0; local
[all...]
H A Dxfs_alloc.h199 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
/linux-master/drivers/md/
H A Ddm-zoned-reclaim.c286 int alloc_flags = DMZ_ALLOC_SEQ; local
292 alloc_flags | DMZ_ALLOC_RECLAIM);
293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) {
294 alloc_flags = DMZ_ALLOC_RND;
H A Ddm-zoned-metadata.c2056 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; local
2071 dzone = dmz_alloc_zone(zmd, 0, alloc_flags);
2168 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; local
2177 bzone = dmz_alloc_zone(zmd, 0, alloc_flags);
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_chardev.c1937 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags;
1940 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) {
1948 if (bo_bucket->alloc_flags
1951 bo_bucket->alloc_flags &
1960 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL)
1963 else if (bo_bucket->alloc_flags &
1976 "gpu_id = 0x%x alloc_flags = 0x%x idr_handle = 0x%x",
1981 bo_bucket->alloc_flags,
2005 if (bo_buckets[bo_index].alloc_flags
[all...]
/linux-master/fs/xfs/scrub/
H A Drepair.h54 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
H A Drepair.c411 int alloc_flags)
421 return xfs_alloc_fix_freelist(&args, alloc_flags);
409 xrep_fix_freelist( struct xfs_scrub *sc, int alloc_flags) argument
/linux-master/fs/btrfs/
H A Dblock-group.c2901 u64 alloc_flags; local
2949 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags);
2950 if (alloc_flags != cache->flags) {
2951 ret = btrfs_chunk_alloc(trans, alloc_flags,
2979 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags);
2980 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE);
2996 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags);
2998 check_system_chunk(trans, alloc_flags);
3827 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); local
3829 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORC
[all...]
/linux-master/drivers/iommu/
H A Ddma-iommu.c914 gfp_t alloc_flags = gfp; local
918 alloc_flags |= __GFP_NORETRY;
919 page = alloc_pages_node(nid, alloc_flags, order);
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_object.h49 unsigned alloc_flags);
/linux-master/include/uapi/linux/
H A Dkfd_ioctl.h635 __u32 alloc_flags; member in struct:kfd_criu_bo_bucket

Completed in 496 milliseconds

12