/linux-master/drivers/xen/ |
H A D | grant-dma-ops.c | 79 dma_addr_t *dma_handle, gfp_t gfp, 95 ret = alloc_pages_exact(n_pages * XEN_PAGE_SIZE, gfp); 148 gfp_t gfp) 152 vaddr = xen_grant_dma_alloc(dev, size, dma_handle, gfp, 0); 78 xen_grant_dma_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) argument 145 xen_grant_dma_alloc_pages(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument
|
/linux-master/arch/powerpc/mm/ |
H A D | init_32.c | 28 #include <linux/gfp.h>
|
/linux-master/security/selinux/include/ |
H A D | xfrm.h | 17 struct xfrm_user_sec_ctx *uctx, gfp_t gfp);
|
/linux-master/drivers/iommu/amd/ |
H A D | io_pgtable_v2.c | 136 unsigned long pg_size, gfp_t gfp, bool *updated) 159 page = alloc_pgtable_page(nid, gfp); 238 int prot, gfp_t gfp, size_t *mapped) 259 iova, map_size, gfp, &updated); 135 v2_alloc_pte(int nid, u64 *pgd, unsigned long iova, unsigned long pg_size, gfp_t gfp, bool *updated) argument 236 iommu_v2_map_pages(struct io_pgtable_ops *ops, unsigned long iova, phys_addr_t paddr, size_t pgsize, size_t pgcount, int prot, gfp_t gfp, size_t *mapped) argument
|
/linux-master/lib/ |
H A D | percpu-refcount.c | 53 * @gfp: allocation mask to use 64 unsigned int flags, gfp_t gfp) 72 __alloc_percpu_gfp(sizeof(unsigned long), align, gfp); 76 data = kzalloc(sizeof(*ref->data), gfp); 63 percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release, unsigned int flags, gfp_t gfp) argument
|
H A D | btree.c | 93 static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp) argument 97 node = mempool_alloc(head->mempool, gfp); 410 gfp_t gfp) 415 node = btree_node_alloc(head, gfp); 446 gfp_t gfp) 453 err = btree_grow(head, geo, gfp); 469 new = btree_node_alloc(head, gfp); 474 new, level + 1, gfp); 507 unsigned long *key, void *val, gfp_t gfp) 510 return btree_insert_level(head, geo, key, val, 1, gfp); 409 btree_grow(struct btree_head *head, struct btree_geo *geo, gfp_t gfp) argument 444 btree_insert_level(struct btree_head *head, struct btree_geo *geo, unsigned long *key, void *val, int level, gfp_t gfp) argument 506 btree_insert(struct btree_head *head, struct btree_geo *geo, unsigned long *key, void *val, gfp_t gfp) argument 635 btree_merge(struct btree_head *target, struct btree_head *victim, struct btree_geo *geo, gfp_t gfp) argument [all...] |
/linux-master/net/rds/ |
H A D | ib_recv.c | 102 static int rds_ib_recv_alloc_cache(struct rds_ib_refill_cache *cache, gfp_t gfp) argument 107 cache->percpu = alloc_percpu_gfp(struct rds_ib_cache_head, gfp); 122 int rds_ib_recv_alloc_caches(struct rds_ib_connection *ic, gfp_t gfp) argument 126 ret = rds_ib_recv_alloc_cache(&ic->i_cache_incs, gfp); 128 ret = rds_ib_recv_alloc_cache(&ic->i_cache_frags, gfp); 309 struct rds_ib_recv_work *recv, gfp_t gfp) 314 gfp_t slab_mask = gfp; 315 gfp_t page_mask = gfp; 317 if (gfp & __GFP_DIRECT_RECLAIM) { 383 void rds_ib_recv_refill(struct rds_connection *conn, int prefill, gfp_t gfp) argument 308 rds_ib_recv_refill_one(struct rds_connection *conn, struct rds_ib_recv_work *recv, gfp_t gfp) argument [all...] |
/linux-master/drivers/accel/habanalabs/common/ |
H A D | memory_mgr.c | 141 * @gfp: gfp flags to use for the memory allocations 149 struct hl_mmap_mem_buf_behavior *behavior, gfp_t gfp, 155 buf = kzalloc(sizeof(*buf), gfp); 174 rc = buf->behavior->alloc(buf, gfp, args); 148 hl_mmap_mem_buf_alloc(struct hl_mem_mgr *mmg, struct hl_mmap_mem_buf_behavior *behavior, gfp_t gfp, void *args) argument
|
/linux-master/net/sctp/ |
H A D | stream_sched_fc.c | 28 __u16 weight, gfp_t gfp) 49 __u16 weight, gfp_t gfp) 68 gfp_t gfp) 27 sctp_sched_wfq_set(struct sctp_stream *stream, __u16 sid, __u16 weight, gfp_t gfp) argument 48 sctp_sched_fc_set(struct sctp_stream *stream, __u16 sid, __u16 weight, gfp_t gfp) argument 67 sctp_sched_fc_init_sid(struct sctp_stream *stream, __u16 sid, gfp_t gfp) argument
|
H A D | associola.c | 55 enum sctp_scope scope, gfp_t gfp) 231 if (sctp_stream_init(&asoc->stream, asoc->c.sinit_num_ostreams, 0, gfp)) 255 if (sctp_auth_asoc_copy_shkeys(ep, asoc, gfp)) 287 enum sctp_scope scope, gfp_t gfp) 291 asoc = kzalloc(sizeof(*asoc), gfp); 295 if (!sctp_association_init(asoc, ep, sk, scope, gfp)) 584 const gfp_t gfp, 616 peer = sctp_transport_new(asoc->base.net, addr, gfp); 1572 enum sctp_scope scope, gfp_t gfp) 1591 scope, gfp, flag 51 sctp_association_init( struct sctp_association *asoc, const struct sctp_endpoint *ep, const struct sock *sk, enum sctp_scope scope, gfp_t gfp) argument 285 sctp_association_new(const struct sctp_endpoint *ep, const struct sock *sk, enum sctp_scope scope, gfp_t gfp) argument 582 sctp_assoc_add_peer(struct sctp_association *asoc, const union sctp_addr *addr, const gfp_t gfp, const int peer_state) argument 1571 sctp_assoc_set_bind_addr_from_ep(struct sctp_association *asoc, enum sctp_scope scope, gfp_t gfp) argument 1595 sctp_assoc_set_bind_addr_from_cookie(struct sctp_association *asoc, struct sctp_cookie *cookie, gfp_t gfp) argument 1623 sctp_assoc_set_id(struct sctp_association *asoc, gfp_t gfp) argument [all...] |
/linux-master/mm/ |
H A D | secretmem.c | 55 gfp_t gfp = vmf->gfp_mask; local 70 folio = folio_alloc(gfp | __GFP_ZERO, 0); 85 err = filemap_add_folio(mapping, folio, offset, gfp);
|
H A D | percpu.c | 494 * @gfp: allocation flags 504 static void *pcpu_mem_zalloc(size_t size, gfp_t gfp) argument 510 return kzalloc(size, gfp); 512 return __vmalloc(size, gfp | __GFP_ZERO); 1438 static struct pcpu_chunk *pcpu_alloc_chunk(gfp_t gfp) argument 1443 chunk = pcpu_mem_zalloc(pcpu_chunk_struct_size, gfp); 1452 sizeof(chunk->alloc_map[0]), gfp); 1457 sizeof(chunk->bound_map[0]), gfp); 1462 sizeof(chunk->md_blocks[0]), gfp); 1470 sizeof(struct obj_cgroup *), gfp); 1623 pcpu_memcg_pre_alloc_hook(size_t size, gfp_t gfp, struct obj_cgroup **objcgp) argument 1686 pcpu_memcg_pre_alloc_hook(size_t size, gfp_t gfp, struct obj_cgroup **objcgp) argument 1717 pcpu_alloc(size_t size, size_t align, bool reserved, gfp_t gfp) argument 1928 __alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp) argument 2036 const gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN; local [all...] |
H A D | z3fold.c | 195 gfp_t gfp) 198 gfp); 320 struct z3fold_pool *pool, gfp_t gfp) 334 slots = alloc_slots(pool, gfp); 904 * @gfp: gfp flags when allocating the z3fold pool structure 909 static struct z3fold_pool *z3fold_create_pool(const char *name, gfp_t gfp) argument 914 pool = kzalloc(sizeof(struct z3fold_pool), gfp); 989 * @gfp: gfp flag 194 alloc_slots(struct z3fold_pool *pool, gfp_t gfp) argument 319 init_z3fold_page(struct page *page, bool headless, struct z3fold_pool *pool, gfp_t gfp) argument 1001 z3fold_alloc(struct z3fold_pool *pool, size_t size, gfp_t gfp, unsigned long *handle) argument 1375 z3fold_zpool_create(const char *name, gfp_t gfp) argument 1385 z3fold_zpool_malloc(void *pool, size_t size, gfp_t gfp, unsigned long *handle) argument [all...] |
H A D | backing-dev.c | 389 gfp_t gfp) 415 err = fprop_local_init_percpu(&wb->completions, gfp); 420 err = percpu_counter_init(&wb->stat[i], 0, gfp); 551 struct cgroup_subsys_state *memcg_css, gfp_t gfp) 577 wb = kmalloc(sizeof(*wb), gfp); 583 ret = wb_init(wb, bdi, gfp); 587 ret = percpu_ref_init(&wb->refcnt, cgwb_release, 0, gfp); 591 ret = fprop_local_init_percpu(&wb->memcg_completions, gfp); 696 * @gfp: allocation mask to use 703 gfp_t gfp) 388 wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi, gfp_t gfp) argument 550 cgwb_create(struct backing_dev_info *bdi, struct cgroup_subsys_state *memcg_css, gfp_t gfp) argument 701 wb_get_create(struct backing_dev_info *bdi, struct cgroup_subsys_state *memcg_css, gfp_t gfp) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma_resource.c | 363 * @gfp: Allocation mode for memory allocations. 370 * wait interruptible or not. Note that @gfp should ideally be selected so 373 * function with means that @gfp can't call into direct reclaim. In reality 384 gfp_t gfp) 390 might_alloc(gfp); 400 0, gfp); 379 i915_vma_resource_bind_dep_await(struct i915_address_space *vm, struct i915_sw_fence *sw_fence, u64 offset, u64 size, bool intr, gfp_t gfp) argument
|
/linux-master/fs/ramfs/ |
H A D | file-nommu.c | 69 gfp_t gfp = mapping_gfp_mask(inode->i_mapping); local 84 pages = alloc_pages(gfp, order); 108 gfp);
|
/linux-master/include/linux/ |
H A D | pagemap.h | 14 #include <linux/gfp.h> 545 struct folio *filemap_alloc_folio(gfp_t gfp, unsigned int order); 547 static inline struct folio *filemap_alloc_folio(gfp_t gfp, unsigned int order) argument 549 return folio_alloc(gfp, order); 553 static inline struct page *__page_cache_alloc(gfp_t gfp) argument 555 return &filemap_alloc_folio(gfp, 0)->page; 593 * * %FGP_NOFS - __GFP_FS will get cleared in gfp. 635 fgf_t fgp_flags, gfp_t gfp); 637 fgf_t fgp_flags, gfp_t gfp); 1190 pgoff_t index, gfp_t gfp); [all...] |
H A D | bio.h | 389 gfp_t gfp, struct bio_set *bs); 397 * @gfp: gfp mask 404 gfp_t gfp, struct bio_set *bs) 409 return bio_split(bio, sectors, gfp, bs); 428 gfp_t gfp, struct bio_set *bs); 430 struct bio *bio_src, gfp_t gfp); 786 static inline void *bio_integrity_alloc(struct bio * bio, gfp_t gfp, argument 826 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp); 403 bio_next_split(struct bio *bio, int sectors, gfp_t gfp, struct bio_set *bs) argument
|
/linux-master/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | fweh.h | 358 u32 packet_len, gfp_t gfp); 363 gfp_t gfp) 394 brcmf_fweh_process_event(drvr, event_packet, skb->len + ETH_HLEN, gfp); 361 brcmf_fweh_process_skb(struct brcmf_pub *drvr, struct sk_buff *skb, u16 stype, gfp_t gfp) argument
|
/linux-master/arch/x86/include/asm/ |
H A D | pgalloc.h | 152 gfp_t gfp = GFP_KERNEL_ACCOUNT; local 155 gfp &= ~__GFP_ACCOUNT; 156 return (p4d_t *)get_zeroed_page(gfp);
|
/linux-master/block/ |
H A D | bio.c | 176 static inline gfp_t bvec_alloc_gfp(gfp_t gfp) argument 178 return (gfp & ~(__GFP_DIRECT_RECLAIM | __GFP_IO)) | 349 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) 351 struct bio *new = bio_alloc(bdev, nr_pages, opf, gfp); 433 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, 814 static int __bio_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp) argument 828 if (bio_crypt_clone(bio, bio_src, gfp) < 0) 831 bio_integrity_clone(bio, bio_src, gfp) < 0) 840 * @gfp: allocation priority 849 gfp_t gfp, struc 348 blk_next_bio(struct bio *bio, struct block_device *bdev, unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) argument 432 bio_alloc_percpu_cache(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, struct bio_set *bs) argument 848 bio_alloc_clone(struct block_device *bdev, struct bio *bio_src, gfp_t gfp, struct bio_set *bs) argument 879 bio_init_clone(struct block_device *bdev, struct bio *bio, struct bio *bio_src, gfp_t gfp) argument 1621 bio_split(struct bio *bio, int sectors, gfp_t gfp, struct bio_set *bs) argument [all...] |
/linux-master/drivers/iommu/ |
H A D | io-pgtable-dart.c | 109 static void *__dart_alloc_pages(size_t size, gfp_t gfp, argument 115 VM_BUG_ON((gfp & __GFP_HIGHMEM)); 116 p = alloc_pages(gfp | __GFP_ZERO, order); 238 int iommu_prot, gfp_t gfp, size_t *mapped) 265 cptep = __dart_alloc_pages(tblsz, gfp, cfg); 236 dart_map_pages(struct io_pgtable_ops *ops, unsigned long iova, phys_addr_t paddr, size_t pgsize, size_t pgcount, int iommu_prot, gfp_t gfp, size_t *mapped) argument
|
/linux-master/drivers/net/wireless/ath/ath12k/ |
H A D | dbring.c | 13 gfp_t gfp) 39 buf_id = idr_alloc(&ring->bufs_idr, buff, 0, ring->bufs_max, gfp); 77 gfp_t gfp) 97 buff = kzalloc(size, gfp); 101 ret = ath12k_dbring_bufs_replenish(ar, ring, buff, gfp); 10 ath12k_dbring_bufs_replenish(struct ath12k *ar, struct ath12k_dbring *ring, struct ath12k_dbring_element *buff, gfp_t gfp) argument 75 ath12k_dbring_fill_bufs(struct ath12k *ar, struct ath12k_dbring *ring, gfp_t gfp) argument
|
/linux-master/net/rxrpc/ |
H A D | call_object.c | 122 struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *rx, gfp_t gfp, argument 128 call = kmem_cache_zalloc(rxrpc_call_jar, gfp); 197 gfp_t gfp, 206 call = rxrpc_alloc_call(rx, gfp, debug_id); 267 static struct semaphore *rxrpc_get_call_slot(struct rxrpc_call_params *p, gfp_t gfp) argument 297 static int rxrpc_connect_call(struct rxrpc_call *call, gfp_t gfp) argument 304 ret = rxrpc_look_up_bundle(call, gfp); 329 gfp_t gfp, 347 limiter = rxrpc_get_call_slot(p, gfp); 353 call = rxrpc_alloc_client_call(rx, cp, p, gfp, debug_i 194 rxrpc_alloc_client_call(struct rxrpc_sock *rx, struct rxrpc_conn_parameters *cp, struct rxrpc_call_params *p, gfp_t gfp, unsigned int debug_id) argument [all...] |
/linux-master/net/core/ |
H A D | xdp.c | 240 static int __mem_id_cyclic_get(gfp_t gfp) argument 246 id = ida_alloc_range(&mem_id_pool, mem_id_next, MEM_ID_MAX - 1, gfp); 278 gfp_t gfp = GFP_KERNEL; local 304 xdp_alloc = kzalloc(sizeof(*xdp_alloc), gfp); 309 id = __mem_id_cyclic_get(gfp); 591 int xdp_alloc_skb_bulk(void **skbs, int n_skb, gfp_t gfp) argument 593 n_skb = kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, gfp, n_skb, skbs);
|