/linux-master/net/handshake/ |
H A D | handshake-test.c | 36 gfp_t gfp; member in struct:handshake_req_alloc_test_param 75 .gfp = GFP_KERNEL, 81 .gfp = GFP_KERNEL, 87 .gfp = GFP_KERNEL, 93 .gfp = GFP_KERNEL, 99 .gfp = GFP_KERNEL, 105 .gfp = GFP_KERNEL | __GFP_NOWARN, 111 .gfp = GFP_KERNEL, 135 result = handshake_req_alloc(param->proto, param->gfp);
|
/linux-master/security/apparmor/include/ |
H A D | policy.h | 97 struct aa_policydb *aa_alloc_pdb(gfp_t gfp); 256 struct aa_ruleset *aa_alloc_ruleset(gfp_t gfp); 258 gfp_t gfp); 260 gfp_t gfp); 262 const char *base, gfp_t gfp);
|
/linux-master/include/linux/ |
H A D | virtio.h | 11 #include <linux/gfp.h> 45 gfp_t gfp); 50 gfp_t gfp); 56 gfp_t gfp); 63 gfp_t gfp);
|
H A D | firmware.h | 8 #include <linux/gfp.h> 107 const char *name, struct device *device, gfp_t gfp, void *context, 142 const char *name, struct device *device, gfp_t gfp, void *context, 140 request_firmware_nowait( struct module *module, bool uevent, const char *name, struct device *device, gfp_t gfp, void *context, void (*cont)(const struct firmware *fw, void *context)) argument
|
H A D | btree.h | 46 * @gfp_mask: gfp mask for the allocation 110 * @gfp: allocation flags for node allocations 116 unsigned long *key, void *val, gfp_t gfp); 149 * @gfp: allocation flags 159 struct btree_geo *geo, gfp_t gfp);
|
/linux-master/mm/ |
H A D | kmemleak.c | 118 #define gfp_kmemleak_mask(gfp) (((gfp) & (GFP_KERNEL | GFP_ATOMIC | \ 459 static struct kmemleak_object *mem_pool_alloc(gfp_t gfp) argument 466 object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp)); 641 static struct kmemleak_object *__alloc_object(gfp_t gfp) argument 645 object = mem_pool_alloc(gfp); 743 int min_count, gfp_t gfp, unsigned int objflags) 749 object = __alloc_object(gfp); 762 int min_count, gfp_t gfp) 764 __create_object(ptr, size, min_count, gfp, 742 __create_object(unsigned long ptr, size_t size, int min_count, gfp_t gfp, unsigned int objflags) argument 761 create_object(unsigned long ptr, size_t size, int min_count, gfp_t gfp) argument 768 create_object_phys(unsigned long ptr, size_t size, int min_count, gfp_t gfp) argument 775 create_object_percpu(unsigned long ptr, size_t size, int min_count, gfp_t gfp) argument 931 add_scan_area(unsigned long ptr, size_t size, gfp_t gfp) argument 1038 kmemleak_alloc(const void *ptr, size_t size, int min_count, gfp_t gfp) argument 1057 kmemleak_alloc_percpu(const void __percpu *ptr, size_t size, gfp_t gfp) argument 1080 kmemleak_vmalloc(const struct vm_struct *area, size_t size, gfp_t gfp) argument 1227 kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) argument 1261 kmemleak_alloc_phys(phys_addr_t phys, size_t size, gfp_t gfp) argument [all...] |
H A D | mempolicy.c | 459 static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *pol, 1218 gfp_t gfp; local 1228 gfp = htlb_alloc_mask(h); 1229 nodemask = policy_nodemask(gfp, pol, ilx, &nid); 1230 return alloc_hugetlb_folio_nodemask(h, nid, nodemask, gfp); 1234 gfp = GFP_TRANSHUGE; 1236 gfp = GFP_HIGHUSER_MOVABLE | __GFP_RETRY_MAYFAIL | __GFP_COMP; 1238 page = alloc_pages_mpol(gfp, order, pol, ilx, nid); 1864 * we apply policy when gfp_zone(gfp) = ZONE_MOVABLE only. 2044 static nodemask_t *policy_nodemask(gfp_t gfp, struc argument 2189 alloc_pages_preferred_many(gfp_t gfp, unsigned int order, int nid, nodemask_t *nodemask) argument 2220 alloc_pages_mpol(gfp_t gfp, unsigned int order, struct mempolicy *pol, pgoff_t ilx, int nid) argument 2295 vma_alloc_folio(gfp_t gfp, int order, struct vm_area_struct *vma, unsigned long addr, bool hugepage) argument 2324 alloc_pages(gfp_t gfp, unsigned int order) argument 2340 folio_alloc(gfp_t gfp, unsigned int order) argument 2346 alloc_pages_bulk_array_interleave(gfp_t gfp, struct mempolicy *pol, unsigned long nr_pages, struct page **page_array) argument 2381 alloc_pages_bulk_array_weighted_interleave(gfp_t gfp, struct mempolicy *pol, unsigned long nr_pages, struct page **page_array) argument 2496 alloc_pages_bulk_array_preferred_many(gfp_t gfp, int nid, struct mempolicy *pol, unsigned long nr_pages, struct page **page_array) argument 2522 alloc_pages_bulk_array_mempolicy(gfp_t gfp, unsigned long nr_pages, struct page **page_array) argument [all...] |
/linux-master/kernel/dma/ |
H A D | mapping.c | 12 #include <linux/gfp.h> 80 * @gfp: Allocation flags 90 gfp_t gfp, unsigned long attrs) 95 dr = devres_alloc(dmam_release, sizeof(*dr), gfp); 99 vaddr = dma_alloc_attrs(dev, size, dma_handle, gfp, attrs); 559 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) 565 if (WARN_ON_ONCE(gfp & (__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM))) 567 if (WARN_ON_ONCE(gfp & __GFP_COMP)) 572 return dma_direct_alloc_pages(dev, size, dma_handle, dir, gfp); 575 return ops->alloc_pages(dev, size, dma_handle, dir, gfp); 89 dmam_alloc_attrs(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) argument 558 __dma_alloc_pages(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument 578 dma_alloc_pages(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument 622 alloc_single_sgt(struct device *dev, size_t size, enum dma_data_direction dir, gfp_t gfp) argument 646 dma_alloc_noncontiguous(struct device *dev, size_t size, enum dma_data_direction dir, gfp_t gfp, unsigned long attrs) argument [all...] |
/linux-master/security/apparmor/ |
H A D | policy.c | 125 struct aa_policydb *aa_alloc_pdb(gfp_t gfp) argument 127 struct aa_policydb *pdb = kzalloc(sizeof(struct aa_policydb), gfp); 258 struct aa_ruleset *aa_alloc_ruleset(gfp_t gfp) argument 262 rules = kzalloc(sizeof(*rules), gfp); 327 * @gfp: allocation type 332 gfp_t gfp) 338 profile = kzalloc(struct_size(profile, label.vec, 2), gfp); 342 if (!aa_policy_init(&profile->base, NULL, hname, gfp)) 344 if (!aa_label_init(&profile->label, 1, gfp)) 350 rules = aa_alloc_ruleset(gfp); 331 aa_alloc_profile(const char *hname, struct aa_proxy *proxy, gfp_t gfp) argument 477 __create_missing_ancestors(struct aa_ns *ns, const char *hname, gfp_t gfp) argument 617 aa_alloc_null(struct aa_profile *parent, const char *name, gfp_t gfp) argument 662 aa_new_learning_profile(struct aa_profile *parent, bool hat, const char *base, gfp_t gfp) argument [all...] |
/linux-master/fs/notify/fanotify/ |
H A D | fanotify.c | 401 gfp_t gfp) 429 if (gfp && fh_len > FANOTIFY_INLINE_FH_LEN) { 432 ext_buf = kmalloc(fh_len, gfp); 536 gfp_t gfp) 540 pevent = kmem_cache_alloc(fanotify_path_event_cachep, gfp); 553 gfp_t gfp) 557 pevent = kmem_cache_alloc(fanotify_perm_event_cachep, gfp); 576 gfp_t gfp) 580 ffe = kmem_cache_alloc(fanotify_fid_event_cachep, gfp); 588 hash, gfp); 399 fanotify_encode_fh(struct fanotify_fh *fh, struct inode *inode, unsigned int fh_len, unsigned int *hash, gfp_t gfp) argument 534 fanotify_alloc_path_event(const struct path *path, unsigned int *hash, gfp_t gfp) argument 552 fanotify_alloc_perm_event(const struct path *path, gfp_t gfp) argument 573 fanotify_alloc_fid_event(struct inode *id, __kernel_fsid_t *fsid, unsigned int *hash, gfp_t gfp) argument 593 fanotify_alloc_name_event(struct inode *dir, __kernel_fsid_t *fsid, const struct qstr *name, struct inode *child, struct dentry *moved, unsigned int *hash, gfp_t gfp) argument 713 gfp_t gfp = GFP_KERNEL_ACCOUNT; local [all...] |
/linux-master/fs/nfs/ |
H A D | fscache.h | 102 static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument 105 if (current_is_kswapd() || !(gfp & __GFP_FS)) 186 static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) argument
|
/linux-master/arch/arm64/include/asm/ |
H A D | pgalloc.h | 90 gfp_t gfp = GFP_PGTABLE_USER; local 93 gfp = GFP_PGTABLE_KERNEL; 94 return (p4d_t *)get_zeroed_page(gfp);
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgalloc.h | 116 gfp_t gfp = GFP_PGTABLE_USER; local 119 gfp = GFP_PGTABLE_KERNEL; 120 return (p4d_t *)get_zeroed_page(gfp);
|
/linux-master/drivers/base/ |
H A D | devcoredump.c | 250 * @gfp: allocation flags 256 gfp_t gfp) 258 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev); 313 * @gfp: allocation flags 323 void *data, size_t datalen, gfp_t gfp, 345 devcd = kzalloc(sizeof(*devcd), gfp); 402 * @gfp: allocation flags 410 size_t datalen, gfp_t gfp) 412 dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable, 255 dev_coredumpv(struct device *dev, void *data, size_t datalen, gfp_t gfp) argument 322 dev_coredumpm(struct device *dev, struct module *owner, void *data, size_t datalen, gfp_t gfp, ssize_t (*read)(char *buffer, loff_t offset, size_t count, void *data, size_t datalen), void (*free)(void *data)) argument 409 dev_coredumpsg(struct device *dev, struct scatterlist *table, size_t datalen, gfp_t gfp) argument
|
/linux-master/include/net/netfilter/ |
H A D | nf_conntrack_ecache.h | 76 bool nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp); 91 static inline bool nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp) argument
|
/linux-master/drivers/block/ |
H A D | brd.c | 68 static int brd_insert_page(struct brd_device *brd, sector_t sector, gfp_t gfp) argument 78 page = alloc_page(gfp | __GFP_ZERO | __GFP_HIGHMEM); 87 cur = __xa_cmpxchg(&brd->brd_pages, idx, NULL, page, gfp); 124 gfp_t gfp) 131 ret = brd_insert_page(brd, sector, gfp); 136 ret = brd_insert_page(brd, sector, gfp); 222 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO; local 224 err = copy_to_brd_setup(brd, sector, len, gfp); 123 copy_to_brd_setup(struct brd_device *brd, sector_t sector, size_t n, gfp_t gfp) argument
|
/linux-master/include/net/page_pool/ |
H A D | types.h | 204 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 206 unsigned int size, gfp_t gfp);
|
/linux-master/net/sctp/ |
H A D | sm_sideeffect.c | 34 #include <linux/gfp.h> 48 gfp_t gfp); 57 gfp_t gfp); 697 gfp_t gfp) 706 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) 1108 struct sctp_datamsg *msg, gfp_t gfp) 1113 sctp_outq_tail(&asoc->outqueue, chunk, gfp); 1146 void *event_arg, gfp_t gfp) 1171 &commands, gfp); 1188 gfp_t gfp) 693 sctp_cmd_process_init(struct sctp_cmd_seq *commands, struct sctp_association *asoc, struct sctp_chunk *chunk, struct sctp_init_chunk *peer_init, gfp_t gfp) argument 1107 sctp_cmd_send_msg(struct sctp_association *asoc, struct sctp_datamsg *msg, gfp_t gfp) argument 1143 sctp_do_sm(struct net *net, enum sctp_event_type event_type, union sctp_subtype subtype, enum sctp_state state, struct sctp_endpoint *ep, struct sctp_association *asoc, void *event_arg, gfp_t gfp) argument 1180 sctp_side_effects(enum sctp_event_type event_type, union sctp_subtype subtype, enum sctp_state state, struct sctp_endpoint *ep, struct sctp_association **asoc, void *event_arg, enum sctp_disposition status, struct sctp_cmd_seq *commands, gfp_t gfp) argument 1269 sctp_cmd_interpreter(enum sctp_event_type event_type, union sctp_subtype subtype, enum sctp_state state, struct sctp_endpoint *ep, struct sctp_association *asoc, void *event_arg, enum sctp_disposition status, struct sctp_cmd_seq *commands, gfp_t gfp) argument [all...] |
H A D | outqueue.c | 56 static void sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp); 281 void sctp_outq_tail(struct sctp_outq *q, struct sctp_chunk *chunk, gfp_t gfp) argument 313 sctp_outq_flush(q, 0, gfp); 599 int rtx_timeout, int *start_timer, gfp_t gfp) 675 sctp_packet_transmit(pkt, gfp); 680 error = sctp_packet_transmit(pkt, gfp); 696 error = sctp_packet_transmit(pkt, gfp); 706 error = sctp_packet_transmit(pkt, gfp); 759 void sctp_outq_uncork(struct sctp_outq *q, gfp_t gfp) argument 764 sctp_outq_flush(q, 0, gfp); 598 __sctp_outq_flush_rtx(struct sctp_outq *q, struct sctp_packet *pkt, int rtx_timeout, int *start_timer, gfp_t gfp) argument 767 sctp_packet_singleton(struct sctp_transport *transport, struct sctp_chunk *chunk, gfp_t gfp) argument 796 gfp_t gfp; member in struct:sctp_flush_ctx 1192 sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp) argument [all...] |
/linux-master/include/net/ |
H A D | cfg80211.h | 7135 * @gfp: context flags 7147 gfp_t gfp); 7153 s32 signal, gfp_t gfp) 7160 return cfg80211_inform_bss_frame_data(wiphy, &data, mgmt, len, gfp); 7261 * @gfp: context flags 7275 gfp_t gfp); 7283 s32 signal, gfp_t gfp) 7292 gfp); 7535 * @gfp: allocation flags 7543 const u8 *tsc, gfp_t gfp); 7150 cfg80211_inform_bss_frame(struct wiphy *wiphy, struct ieee80211_channel *rx_channel, struct ieee80211_mgmt *mgmt, size_t len, s32 signal, gfp_t gfp) argument 7278 cfg80211_inform_bss(struct wiphy *wiphy, struct ieee80211_channel *rx_channel, enum cfg80211_bss_frame_type ftype, const u8 *bssid, u64 tsf, u16 capability, u16 beacon_interval, const u8 *ie, size_t ielen, s32 signal, gfp_t gfp) argument 7736 cfg80211_vendor_event_alloc(struct wiphy *wiphy, struct wireless_dev *wdev, int approxlen, int event_idx, gfp_t gfp) argument 7769 cfg80211_vendor_event_alloc_ucast(struct wiphy *wiphy, struct wireless_dev *wdev, unsigned int portid, int approxlen, int event_idx, gfp_t gfp) argument 7787 cfg80211_vendor_event(struct sk_buff *skb, gfp_t gfp) argument 7873 cfg80211_testmode_alloc_event_skb(struct wiphy *wiphy, int approxlen, gfp_t gfp) argument 7890 cfg80211_testmode_event(struct sk_buff *skb, gfp_t gfp) argument 8045 cfg80211_connect_bss(struct net_device *dev, const u8 *bssid, struct cfg80211_bss *bss, const u8 *req_ie, size_t req_ie_len, const u8 *resp_ie, size_t resp_ie_len, int status, gfp_t gfp, enum nl80211_timeout_reason timeout_reason) argument 8087 cfg80211_connect_result(struct net_device *dev, const u8 *bssid, const u8 *req_ie, size_t req_ie_len, const u8 *resp_ie, size_t resp_ie_len, u16 status, gfp_t gfp) argument 8116 cfg80211_connect_timeout(struct net_device *dev, const u8 *bssid, const u8 *req_ie, size_t req_ie_len, gfp_t gfp, enum nl80211_timeout_reason timeout_reason) argument 8307 cfg80211_del_sta(struct net_device *dev, const u8 *mac_addr, gfp_t gfp) argument 8483 cfg80211_mgmt_tx_status(struct wireless_dev *wdev, u64 cookie, const u8 *buf, size_t len, bool ack, gfp_t gfp) argument 8602 cfg80211_radar_event(struct wiphy *wiphy, struct cfg80211_chan_def *chandef, gfp_t gfp) argument 8610 cfg80211_background_radar_event(struct wiphy *wiphy, struct cfg80211_chan_def *chandef, gfp_t gfp) argument [all...] |
/linux-master/drivers/crypto/hisilicon/sec/ |
H A D | sec_algs.c | 179 gfp_t gfp) 194 gfp, &sgl_next_dma); 549 int *steps, gfp_t gfp) 556 sizes = kcalloc(*steps, sizeof(*sizes), gfp); 572 struct device *dev, gfp_t gfp) 580 *splits = kcalloc(steps, sizeof(struct scatterlist *), gfp); 585 *splits_nents = kcalloc(steps, sizeof(int), gfp); 593 *splits, *splits_nents, gfp); 634 struct sec_dev_info *info, gfp_t gfp) 640 el = kzalloc(sizeof(*el), gfp); 174 sec_alloc_and_fill_hw_sgl(struct sec_hw_sgl **sec_sgl, dma_addr_t *psec_sgl, struct scatterlist *sgl, int count, struct sec_dev_info *info, gfp_t gfp) argument 548 sec_alg_alloc_and_calc_split_sizes(int length, size_t **split_sizes, int *steps, gfp_t gfp) argument 568 sec_map_and_split_sg(struct scatterlist *sgl, size_t *split_sizes, int steps, struct scatterlist ***splits, int **splits_nents, int sgl_len_in, struct device *dev, gfp_t gfp) argument 630 sec_alg_alloc_and_fill_el(struct sec_bd_info *template, int encrypt, int el_size, bool different_dest, struct scatterlist *sgl_in, int n_ents_in, struct scatterlist *sgl_out, int n_ents_out, struct sec_dev_info *info, gfp_t gfp) argument 724 gfp_t gfp = skreq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC; local [all...] |
/linux-master/block/ |
H A D | bfq-cgroup.c | 19 static int bfq_stat_init(struct bfq_stat *stat, gfp_t gfp) argument 23 ret = percpu_counter_init(&stat->cpu_cnt, 0, gfp); 455 static int bfqg_stats_init(struct bfqg_stats *stats, gfp_t gfp) argument 457 if (blkg_rwstat_init(&stats->bytes, gfp) || 458 blkg_rwstat_init(&stats->ios, gfp)) 462 if (blkg_rwstat_init(&stats->merged, gfp) || 463 blkg_rwstat_init(&stats->service_time, gfp) || 464 blkg_rwstat_init(&stats->wait_time, gfp) || 465 blkg_rwstat_init(&stats->queued, gfp) || 466 bfq_stat_init(&stats->time, gfp) || 493 bfq_cpd_alloc(gfp_t gfp) argument 510 bfq_pd_alloc(struct gendisk *disk, struct blkcg *blkcg, gfp_t gfp) argument [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | mem.c | 155 gfp_t gfp = GFP_USER | __GFP_ZERO; local 200 gfp |= GFP_HIGHUSER; 202 gfp |= GFP_DMA32; 205 struct page *p = alloc_page(gfp);
|
/linux-master/tools/testing/radix-tree/ |
H A D | regression4.c | 3 #include <linux/gfp.h>
|
H A D | regression2.c | 47 #include <linux/gfp.h>
|