/linux-master/lib/ |
H A D | ts_kmp.c | 92 gfp_t gfp_mask, int flags) 100 conf = alloc_ts_config(priv_size, gfp_mask); 91 kmp_init(const void *pattern, unsigned int len, gfp_t gfp_mask, int flags) argument
|
H A D | sg_split.c | 135 * @gfp_mask: the allocation flag 152 gfp_t gfp_mask) 157 splitters = kcalloc(nb_splits, sizeof(*splitters), gfp_mask); 170 gfp_mask); 148 sg_split(struct scatterlist *in, const int in_mapped_nents, const off_t skip, const int nb_splits, const size_t *split_sizes, struct scatterlist **out, int *out_mapped_nents, gfp_t gfp_mask) argument
|
H A D | ref_tracker.c | 192 gfp_t gfp_mask = gfp | __GFP_NOWARN; local 202 gfp_mask |= __GFP_NOFAIL; 203 *trackerp = tracker = kzalloc(sizeof(*tracker), gfp_mask);
|
/linux-master/mm/ |
H A D | shrinker.c | 467 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, argument 527 .gfp_mask = gfp_mask, 585 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, argument 594 * @gfp_mask: allocation context 612 unsigned long shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, argument 626 return shrink_slab_memcg(gfp_mask, nid, memcg, priority); 652 .gfp_mask = gfp_mask,
|
H A D | readahead.c | 210 gfp_t gfp_mask = readahead_gfp_mask(mapping); local 221 * gfp_mask, but let's be explicit here. 247 folio = filemap_alloc_folio(gfp_mask, 0); 251 gfp_mask) < 0) { 777 gfp_t gfp_mask = readahead_gfp_mask(mapping); local 789 folio = filemap_alloc_folio(gfp_mask, 0); 792 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) { 816 folio = filemap_alloc_folio(gfp_mask, 0); 819 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) {
|
H A D | vmalloc.c | 1809 preload_this_cpu_lock(spinlock_t *lock, gfp_t gfp_mask, int node) argument 1823 va = kmem_cache_alloc_node(vmap_area_cachep, gfp_mask, node); 1936 int node, gfp_t gfp_mask, 1965 gfp_mask = gfp_mask & GFP_RECLAIM_MASK; 1967 va = kmem_cache_alloc_node(vmap_area_cachep, gfp_mask, node); 1975 kmemleak_scan_area(&va->rb_node, SIZE_MAX, gfp_mask); 1980 preload_this_cpu_lock(&free_vmap_area_lock, gfp_mask, node); 2033 if (!(gfp_mask & __GFP_NOWARN) && printk_ratelimit()) 2553 * @gfp_mask 1933 alloc_vmap_area(unsigned long size, unsigned long align, unsigned long vstart, unsigned long vend, int node, gfp_t gfp_mask, unsigned long va_flags) argument 2557 new_vmap_block(unsigned int order, gfp_t gfp_mask) argument 2698 vb_alloc(unsigned long size, gfp_t gfp_mask) argument 3068 __get_vm_area_node(unsigned long size, unsigned long align, unsigned long shift, unsigned long flags, unsigned long start, unsigned long end, int node, gfp_t gfp_mask, const void *caller) argument 3599 __vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask, pgprot_t prot, unsigned int page_shift, int node) argument 3733 __vmalloc_node_range(unsigned long size, unsigned long align, unsigned long start, unsigned long end, gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, int node, const void *caller) argument 3880 __vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask, int node, const void *caller) argument 3895 __vmalloc(unsigned long size, gfp_t gfp_mask) argument 3933 vmalloc_huge(unsigned long size, gfp_t gfp_mask) argument [all...] |
H A D | hugetlb.c | 1349 static struct folio *dequeue_hugetlb_folio_nodemask(struct hstate *h, gfp_t gfp_mask, argument 1358 zonelist = node_zonelist(nid, gfp_mask); 1362 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { 1365 if (!cpuset_zone_allowed(zone, gfp_mask)) 1397 gfp_t gfp_mask; local 1413 gfp_mask = htlb_alloc_mask(h); 1414 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); 1417 folio = dequeue_hugetlb_folio_nodemask(h, gfp_mask, 1425 folio = dequeue_hugetlb_folio_nodemask(h, gfp_mask, 1565 static struct folio *alloc_gigantic_folio(struct hstate *h, gfp_t gfp_mask, argument 1603 alloc_gigantic_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nodemask) argument 1611 alloc_gigantic_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nodemask) argument 2182 alloc_buddy_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry) argument 2245 __alloc_fresh_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry) argument 2279 only_alloc_fresh_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry) argument 2299 alloc_fresh_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry) argument 2341 gfp_t gfp_mask = htlb_alloc_mask(h) | __GFP_THISNODE; local 2512 alloc_surplus_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask) argument 2553 alloc_migrate_hugetlb_folio(struct hstate *h, gfp_t gfp_mask, int nid, nodemask_t *nmask) argument 2585 gfp_t gfp_mask = htlb_alloc_mask(h); local 2607 alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid, nodemask_t *nmask, gfp_t gfp_mask) argument 3016 gfp_t gfp_mask = htlb_alloc_mask(h) | __GFP_THISNODE; local 3479 gfp_t gfp_mask = htlb_alloc_mask(h) | __GFP_THISNODE; local 4907 gfp_t gfp_mask = htlb_alloc_mask(h); local 6641 gfp_t gfp_mask; local [all...] |
H A D | hugetlb_vmemmap.c | 326 gfp_t gfp_mask = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN; local 336 walk.reuse_page = alloc_pages_node(nid, gfp_mask, 0); 383 gfp_t gfp_mask = GFP_KERNEL | __GFP_RETRY_MAYFAIL; local 389 page = alloc_pages_node(nid, gfp_mask, 0);
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gem_shrinker.c | 29 if (!(sc->gfp_mask & __GFP_DIRECT_RECLAIM)) 31 return current_is_kswapd() || (sc->gfp_mask & __GFP_RECLAIM);
|
/linux-master/fs/nfs/blocklayout/ |
H A D | blocklayout.h | 176 struct pnfs_device *pdev, gfp_t gfp_mask); 193 struct pnfs_block_volume *b, gfp_t gfp_mask);
|
/linux-master/fs/nilfs2/ |
H A D | mdt.h | 78 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz);
|
/linux-master/kernel/ |
H A D | umh.c | 338 * @gfp_mask: gfp mask for memory allocation 357 char **envp, gfp_t gfp_mask, 363 sub_info = kzalloc(sizeof(struct subprocess_info), gfp_mask); 486 gfp_t gfp_mask = (wait == UMH_NO_WAIT) ? GFP_ATOMIC : GFP_KERNEL; local 488 info = call_usermodehelper_setup(path, argv, envp, gfp_mask, 356 call_usermodehelper_setup(const char *path, char **argv, char **envp, gfp_t gfp_mask, int (*init)(struct subprocess_info *info, struct cred *new), void (*cleanup)(struct subprocess_info *info), void *data) argument
|
/linux-master/fs/crypto/ |
H A D | inline_crypt.c | 256 * @gfp_mask: memory allocation flags - these must be a waiting mask so that 268 u64 first_lblk, gfp_t gfp_mask) 278 bio_crypt_set_ctx(bio, ci->ci_enc_key.blk_key, dun, gfp_mask); 311 * @gfp_mask: memory allocation flags 318 gfp_t gfp_mask) 324 fscrypt_set_bio_crypt_ctx(bio, inode, first_lblk, gfp_mask); 267 fscrypt_set_bio_crypt_ctx(struct bio *bio, const struct inode *inode, u64 first_lblk, gfp_t gfp_mask) argument 316 fscrypt_set_bio_crypt_ctx_bh(struct bio *bio, const struct buffer_head *first_bh, gfp_t gfp_mask) argument
|
/linux-master/net/core/ |
H A D | skbuff.c | 274 static void *page_frag_alloc_1k(struct page_frag_1k *nc, gfp_t gfp_mask) argument 625 * @gfp_mask: allocation mask 636 * Buffers may only be allocated from interrupts using a @gfp_mask of 639 struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask, argument 651 gfp_mask |= __GFP_MEMALLOC; 658 skb = kmem_cache_alloc_node(cache, gfp_mask & ~GFP_DMA, node); 668 data = kmalloc_reserve(&size, gfp_mask, node, &pfmemalloc); 707 * @gfp_mask: get_free_pages mask, passed to alloc_skb 717 gfp_t gfp_mask) 731 (gfp_mask 716 __netdev_alloc_skb(struct net_device *dev, unsigned int len, gfp_t gfp_mask) argument 790 __napi_alloc_skb(struct napi_struct *napi, unsigned int len, gfp_t gfp_mask) argument 1912 skb_zerocopy_clone(struct sk_buff *nskb, struct sk_buff *orig, gfp_t gfp_mask) argument 1947 skb_copy_ubufs(struct sk_buff *skb, gfp_t gfp_mask) argument 2044 skb_clone(struct sk_buff *skb, gfp_t gfp_mask) argument 2124 skb_copy(const struct sk_buff *skb, gfp_t gfp_mask) argument 2169 __pskb_copy_fclone(struct sk_buff *skb, int headroom, gfp_t gfp_mask, bool fclone) argument 2233 pskb_expand_head(struct sk_buff *skb, int nhead, int ntail, gfp_t gfp_mask) argument 2457 skb_copy_expand(const struct sk_buff *skb, int newheadroom, int newtailroom, gfp_t gfp_mask) argument 6499 alloc_skb_with_frags(unsigned long header_len, unsigned long data_len, int order, int *errcode, gfp_t gfp_mask) argument 6555 pskb_carve_inside_header(struct sk_buff *skb, const u32 off, const int headlen, gfp_t gfp_mask) argument 6616 pskb_carve_frag_list(struct sk_buff *skb, struct skb_shared_info *shinfo, int eat, gfp_t gfp_mask) argument 6670 pskb_carve_inside_nonlinear(struct sk_buff *skb, const u32 off, int pos, gfp_t gfp_mask) argument [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.c | 107 static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask) argument 115 page = alloc_pages(gfp_mask | __GFP_ZERO, order); 124 int order, gfp_t gfp_mask) 127 gfp_mask); 138 gfp_t gfp_mask, int coherent) 146 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM)); 148 icm = kmalloc(sizeof *icm, gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); 160 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); 176 cur_order, gfp_mask); 179 cur_order, gfp_mask); 123 mthca_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, int order, gfp_t gfp_mask) argument 137 mthca_alloc_icm(struct mthca_dev *dev, int npages, gfp_t gfp_mask, int coherent) argument [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | sa_query.c | 783 static int ib_nl_make_request(struct ib_sa_query *query, gfp_t gfp_mask) argument 803 skb = nlmsg_new(len, gfp_mask); 821 gfp_flag = ((gfp_mask & GFP_ATOMIC) == GFP_ATOMIC) ? GFP_ATOMIC : 1267 static int alloc_mad(struct ib_sa_query *query, gfp_t gfp_mask) argument 1293 gfp_mask, 1335 gfp_t gfp_mask) 1342 ret = __xa_alloc(&queries, &id, query, xa_limit_32b, gfp_mask); 1360 if (!ib_nl_make_request(query, gfp_mask)) 1499 * @gfp_mask:GFP mask to use for internal allocations 1520 unsigned long timeout_ms, gfp_t gfp_mask, 1334 send_mad(struct ib_sa_query *query, unsigned long timeout_ms, gfp_t gfp_mask) argument 1516 ib_sa_path_rec_get(struct ib_sa_client *client, struct ib_device *device, u32 port_num, struct sa_path_rec *rec, ib_sa_comp_mask comp_mask, unsigned long timeout_ms, gfp_t gfp_mask, void (*callback)(int status, struct sa_path_rec *resp, unsigned int num_paths, void *context), void *context, struct ib_sa_query **sa_query) argument 1642 ib_sa_mcmember_rec_query(struct ib_sa_client *client, struct ib_device *device, u32 port_num, u8 method, struct ib_sa_mcmember_rec *rec, ib_sa_comp_mask comp_mask, unsigned long timeout_ms, gfp_t gfp_mask, void (*callback)(int status, struct ib_sa_mcmember_rec *resp, void *context), void *context, struct ib_sa_query **sa_query) argument 1733 ib_sa_guid_info_rec_query(struct ib_sa_client *client, struct ib_device *device, u32 port_num, struct ib_sa_guidinfo_rec *rec, ib_sa_comp_mask comp_mask, u8 method, unsigned long timeout_ms, gfp_t gfp_mask, void (*callback)(int status, struct ib_sa_guidinfo_rec *resp, void *context), void *context, struct ib_sa_query **sa_query) argument 1887 gfp_t gfp_mask = GFP_KERNEL; local [all...] |
/linux-master/block/ |
H A D | bio-integrity.c | 41 * @gfp_mask: Memory allocation mask 49 gfp_t gfp_mask, 60 bip = kmalloc(struct_size(bip, bip_inline_vecs, nr_vecs), gfp_mask); 63 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask); 76 &bip->bip_max_vcnt, gfp_mask); 611 * @gfp_mask: Memory allocation mask 616 gfp_t gfp_mask) 623 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt); 48 bio_integrity_alloc(struct bio *bio, gfp_t gfp_mask, unsigned int nr_vecs) argument 615 bio_integrity_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp_mask) argument
|
/linux-master/net/sunrpc/auth_gss/ |
H A D | gss_mech_switch.c | 357 gfp_t gfp_mask) 359 if (!(*ctx_id = kzalloc(sizeof(**ctx_id), gfp_mask))) 364 *ctx_id, endtime, gfp_mask); 353 gss_import_sec_context(const void *input_token, size_t bufsize, struct gss_api_mech *mech, struct gss_ctx **ctx_id, time64_t *endtime, gfp_t gfp_mask) argument
|
/linux-master/arch/x86/kernel/ |
H A D | module.c | 70 gfp_t gfp_mask = GFP_KERNEL; local 78 MODULES_END, gfp_mask, PAGE_KERNEL, 82 if (p && (kasan_alloc_module_shadow(p, size, gfp_mask) < 0)) {
|
/linux-master/mm/kmsan/ |
H A D | hooks.c | 155 gfp_t gfp_mask = GFP_KERNEL | __GFP_ZERO; local 166 shadow = alloc_pages(gfp_mask, 1); 167 origin = alloc_pages(gfp_mask, 1);
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.h | 81 gfp_t gfp_mask, int coherent);
|
/linux-master/include/linux/ |
H A D | btree.h | 46 * @gfp_mask: gfp mask for the allocation 49 void *btree_alloc(gfp_t gfp_mask, void *pool_data);
|
H A D | swap.h | 393 gfp_t gfp_mask, nodemask_t *mask); 399 gfp_t gfp_mask, 402 gfp_t gfp_mask, bool noswap, 528 static inline int add_swap_count_continuation(swp_entry_t swp, gfp_t gfp_mask) argument
|
/linux-master/include/linux/greybus/ |
H A D | hd.h | 36 struct gb_message *message, gfp_t gfp_mask);
|
/linux-master/security/selinux/ss/ |
H A D | mls.h | 38 gfp_t gfp_mask);
|