/linux-master/tools/testing/radix-tree/ |
H A D | linux.c | 9 #include <linux/gfp.h> 57 int gfp) 61 if (!(gfp & __GFP_DIRECT_RECLAIM)) { 84 else if (gfp & __GFP_ZERO) 141 int kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size, argument 154 if (!(gfp & __GFP_DIRECT_RECLAIM)) { 170 if (!(gfp & __GFP_DIRECT_RECLAIM)) { 186 else if (gfp & __GFP_ZERO) 56 kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru, int gfp) argument
|
/linux-master/net/rxrpc/ |
H A D | insecure.c | 20 static struct rxrpc_txbuf *none_alloc_txbuf(struct rxrpc_call *call, size_t remain, gfp_t gfp) argument 22 return rxrpc_alloc_data_txbuf(call, min_t(size_t, remain, RXRPC_JUMBO_DATALEN), 0, gfp);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_managed.c | 87 size_t size, gfp_t gfp, int nid) 96 dr = kmalloc_node_track_caller(tot_size, gfp, nid); 222 * @gfp: GFP allocation flags 228 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) argument 232 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev)); 235 size, gfp); 238 dr->node.name = kstrdup_const("kmalloc", gfp); 250 * @gfp: GFP allocation flags 256 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp) argument 265 buf = drmm_kmalloc(dev, size, gfp); 86 alloc_dr(drmres_release_t release, size_t size, gfp_t gfp, int nid) argument [all...] |
/linux-master/mm/ |
H A D | zpool.c | 135 * @gfp: The GFP flags to use when allocating the pool. 137 * This creates a new zpool of the specified type. The gfp flags will be 147 struct zpool *zpool_create_pool(const char *type, const char *name, gfp_t gfp) argument 166 zpool = kmalloc(sizeof(*zpool), gfp); 174 zpool->pool = driver->create(name, gfp); 243 * @gfp: The GFP flags to use when allocating memory. 247 * The gfp flags will be used when allocating memory, if the 255 int zpool_malloc(struct zpool *zpool, size_t size, gfp_t gfp, argument 258 return zpool->driver->malloc(zpool->pool, size, gfp, handle);
|
H A D | percpu-vm.c | 76 * @gfp: allocation flags passed to the underlying allocator 84 gfp_t gfp) 89 gfp |= __GFP_HIGHMEM; 95 *pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0); 268 * @gfp: allocation flags passed to the underlying memory allocator 277 int page_start, int page_end, gfp_t gfp) 285 if (pcpu_alloc_pages(chunk, pages, page_start, page_end, gfp)) 333 static struct pcpu_chunk *pcpu_create_chunk(gfp_t gfp) argument 338 chunk = pcpu_alloc_chunk(gfp); 82 pcpu_alloc_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end, gfp_t gfp) argument 276 pcpu_populate_chunk(struct pcpu_chunk *chunk, int page_start, int page_end, gfp_t gfp) argument
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_internal.c | 41 gfp_t gfp; local 50 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE; 53 gfp &= ~__GFP_HIGHMEM; 54 gfp |= __GFP_DMA32; 75 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL),
|
/linux-master/net/sctp/ |
H A D | ulpevent.c | 49 gfp_t gfp) 54 skb = alloc_skb(size, gfp); 117 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) 131 sizeof(struct sctp_assoc_change), 0, gfp); 149 MSG_NOTIFICATION, gfp); 244 int flags, int state, int error, gfp_t gfp) 251 MSG_NOTIFICATION, gfp); 376 gfp_t gfp) 395 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); 432 __u16 flags, __u32 error, gfp_t gfp) 48 sctp_ulpevent_new(int size, __u16 msg_flags, gfp_t gfp) argument 114 sctp_ulpevent_make_assoc_change( const struct sctp_association *asoc, __u16 flags, __u16 state, __u16 error, __u16 outbound, __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) argument 241 sctp_ulpevent_make_peer_addr_change( const struct sctp_association *asoc, const struct sockaddr_storage *aaddr, int flags, int state, int error, gfp_t gfp) argument 374 sctp_ulpevent_make_remote_error(const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, gfp_t gfp) argument 430 sctp_ulpevent_make_send_failed( const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, __u32 error, gfp_t gfp) argument 533 sctp_ulpevent_make_send_failed_event( const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, __u32 error, gfp_t gfp) argument 577 sctp_ulpevent_make_shutdown_event( const struct sctp_association *asoc, __u16 flags, gfp_t gfp) argument 640 sctp_ulpevent_make_adaptation_indication( const struct sctp_association *asoc, gfp_t gfp) argument 675 sctp_ulpevent_make_rcvmsg(struct sctp_association *asoc, struct sctp_chunk *chunk, gfp_t gfp) argument 771 sctp_ulpevent_make_pdapi( const struct sctp_association *asoc, __u32 indication, __u32 sid, __u32 seq, __u32 flags, gfp_t gfp) argument 825 sctp_ulpevent_make_authkey( const struct sctp_association *asoc, __u16 key_id, __u32 indication, gfp_t gfp) argument 864 sctp_ulpevent_make_sender_dry_event( const struct sctp_association *asoc, gfp_t gfp) argument 888 sctp_ulpevent_make_stream_reset_event( const struct sctp_association *asoc, __u16 flags, __u16 stream_num, __be16 *stream_list, gfp_t gfp) argument 917 sctp_ulpevent_make_assoc_reset_event( const struct sctp_association *asoc, __u16 flags, __u32 local_tsn, __u32 remote_tsn, gfp_t gfp) argument 944 sctp_ulpevent_make_stream_change_event( const struct sctp_association *asoc, __u16 flags, __u32 strchange_instrms, __u32 strchange_outstrms, gfp_t gfp) argument [all...] |
H A D | auth.c | 58 static struct sctp_auth_bytes *sctp_auth_create_key(__u32 key_len, gfp_t gfp) argument 67 key = kmalloc(sizeof(struct sctp_auth_bytes) + key_len, gfp); 79 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp) argument 84 new = kzalloc(sizeof(struct sctp_shared_key), gfp); 188 gfp_t gfp) 202 new = sctp_auth_create_key(len, gfp); 223 gfp_t gfp) 228 (struct sctp_hmac_algo_param *)asoc->c.auth_hmacs, gfp); 234 gfp_t gfp) 239 gfp); 184 sctp_auth_make_key_vector( struct sctp_random_param *random, struct sctp_chunks_param *chunks, struct sctp_hmac_algo_param *hmacs, gfp_t gfp) argument 221 sctp_auth_make_local_vector( const struct sctp_association *asoc, gfp_t gfp) argument 232 sctp_auth_make_peer_vector( const struct sctp_association *asoc, gfp_t gfp) argument 252 sctp_auth_asoc_set_secret( struct sctp_shared_key *ep_key, struct sctp_auth_bytes *first_vector, struct sctp_auth_bytes *last_vector, gfp_t gfp) argument 286 sctp_auth_asoc_create_secret( const struct sctp_association *asoc, struct sctp_shared_key *ep_key, gfp_t gfp) argument 353 sctp_auth_asoc_copy_shkeys(const struct sctp_endpoint *ep, struct sctp_association *asoc, gfp_t gfp) argument 383 sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp) argument 454 sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp) argument 707 sctp_auth_calculate_hmac(const struct sctp_association *asoc, struct sk_buff *skb, struct sctp_auth_chunk *auth, struct sctp_shared_key *ep_key, gfp_t gfp) argument 1022 sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_sw_fence.c | 342 wait_queue_entry_t *wq, gfp_t gfp) 348 might_sleep_if(gfpflags_allow_blocking(gfp)); 363 wq = kmalloc(sizeof(*wq), gfp); 365 if (!gfpflags_allow_blocking(gfp)) 405 gfp_t gfp) 407 return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp); 477 gfp_t gfp) 484 might_sleep_if(gfpflags_allow_blocking(gfp)); 494 gfp); 496 if (!gfpflags_allow_blocking(gfp)) 340 __i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence, struct i915_sw_fence *signaler, wait_queue_entry_t *wq, gfp_t gfp) argument 403 i915_sw_fence_await_sw_fence_gfp(struct i915_sw_fence *fence, struct i915_sw_fence *signaler, gfp_t gfp) argument 474 i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence, struct dma_fence *dma, unsigned long timeout, gfp_t gfp) argument 572 i915_sw_fence_await_reservation(struct i915_sw_fence *fence, struct dma_resv *resv, bool write, unsigned long timeout, gfp_t gfp) argument [all...] |
H A D | i915_deps.c | 48 * @gfp: The allocation mode for subsequenst allocations. 50 void i915_deps_init(struct i915_deps *deps, gfp_t gfp) argument 53 deps->gfp = gfp; 85 new_fences = kmalloc_array(new_size, sizeof(*new_fences), deps->gfp);
|
/linux-master/include/linux/ |
H A D | buffer_head.h | 201 gfp_t gfp); 228 unsigned size, gfp_t gfp); 233 sector_t block, unsigned size, gfp_t gfp); 339 gfp_t gfp; local 341 gfp = mapping_gfp_constraint(bdev->bd_inode->i_mapping, ~__GFP_FS); 342 gfp |= __GFP_NOFAIL; 344 return bdev_getblk(bdev, block, size, gfp); 350 gfp_t gfp; local 352 gfp = mapping_gfp_constraint(bdev->bd_inode->i_mapping, ~__GFP_FS); 353 gfp | 364 sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) argument [all...] |
H A D | ptr_ring.h | 467 static inline void **__ptr_ring_init_queue_alloc(unsigned int size, gfp_t gfp) argument 471 return kvmalloc_array(size, sizeof(void *), gfp | __GFP_ZERO); 487 static inline int ptr_ring_init(struct ptr_ring *r, int size, gfp_t gfp) argument 489 r->queue = __ptr_ring_init_queue_alloc(size, gfp); 559 int size, gfp_t gfp, 590 static inline int ptr_ring_resize(struct ptr_ring *r, int size, gfp_t gfp, argument 594 void **queue = __ptr_ring_init_queue_alloc(size, gfp); 603 old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy); 622 gfp_t gfp, void (*destroy)(void *)) 628 queues = kmalloc_array(nrings, sizeof(*queues), gfp); 558 __ptr_ring_swap_queue(struct ptr_ring *r, void **queue, int size, gfp_t gfp, void (*destroy)(void *)) argument 619 ptr_ring_resize_multiple(struct ptr_ring **rings, unsigned int nrings, int size, gfp_t gfp, void (*destroy)(void *)) argument [all...] |
H A D | percpu.h | 132 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp) __alloc_size(1); 141 #define alloc_percpu_gfp(type, gfp) \ 143 __alignof__(type), gfp)
|
H A D | dma-mapping.h | 133 gfp_t gfp, unsigned long attrs); 153 enum dma_data_direction dir, gfp_t gfp, unsigned long attrs); 228 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) 289 size_t size, enum dma_data_direction dir, gfp_t gfp, 314 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp); 321 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) 323 struct page *page = dma_alloc_pages(dev, size, dma_handle, dir, gfp); 429 dma_addr_t *dma_handle, gfp_t gfp) 431 return dma_alloc_attrs(dev, size, dma_handle, gfp, 432 (gfp 227 dmam_alloc_attrs(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) argument 288 dma_alloc_noncontiguous(struct device *dev, size_t size, enum dma_data_direction dir, gfp_t gfp, unsigned long attrs) argument 320 dma_alloc_noncoherent(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument 428 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) argument 550 dmam_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) argument 557 dma_alloc_wc(struct device *dev, size_t size, dma_addr_t *dma_addr, gfp_t gfp) argument [all...] |
/linux-master/lib/ |
H A D | objpool.c | 79 if (pool->gfp & GFP_ATOMIC) 80 slot = kmalloc_node(size, pool->gfp, cpu_to_node(i)); 82 slot = __vmalloc_node(size, sizeof(void *), pool->gfp, 113 gfp_t gfp, void *context, objpool_init_obj_cb objinit, 136 pool->gfp = gfp & ~__GFP_ZERO; 140 pool->cpu_slots = kzalloc(slot_size, pool->gfp); 112 objpool_init(struct objpool_head *pool, int nr_objs, int object_size, gfp_t gfp, void *context, objpool_init_obj_cb objinit, objpool_fini_cb release) argument
|
/linux-master/kernel/dma/ |
H A D | contiguous.c | 329 static struct page *cma_alloc_aligned(struct cma *cma, size_t size, gfp_t gfp) argument 333 return cma_alloc(cma, size >> PAGE_SHIFT, align, gfp & __GFP_NOWARN); 340 * @gfp: Allocation flags. 351 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp) argument 358 if (!gfpflags_allow_blocking(gfp)) 361 return cma_alloc_aligned(dev->cma_area, size, gfp); 366 if (nid != NUMA_NO_NODE && !(gfp & (GFP_DMA | GFP_DMA32))) { 371 page = cma_alloc_aligned(cma, size, gfp); 378 page = cma_alloc_aligned(cma, size, gfp); 387 return cma_alloc_aligned(dma_contiguous_default_area, size, gfp); [all...] |
H A D | ops_helpers.c | 62 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) 67 page = dma_alloc_contiguous(dev, size, gfp); 69 page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size)); 61 dma_common_alloc_pages(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument
|
/linux-master/fs/btrfs/ |
H A D | lru_cache.c | 108 gfp_t gfp) 114 head = kmalloc(sizeof(*head), gfp); 118 ret = mtree_insert(&cache->entries, key, head, gfp); 106 btrfs_lru_cache_store(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *new_entry, gfp_t gfp) argument
|
/linux-master/arch/microblaze/kernel/ |
H A D | dma.c | 12 #include <linux/gfp.h>
|
/linux-master/sound/virtio/ |
H A D | virtio_ctl_msg.h | 24 size_t response_size, gfp_t gfp);
|
/linux-master/include/net/ |
H A D | dst_cache.h | 96 * @gfp: allocation flags 98 int dst_cache_init(struct dst_cache *dst_cache, gfp_t gfp);
|
/linux-master/drivers/net/wireless/intel/iwlwifi/ |
H A D | iwl-modparams.h | 10 #include <linux/gfp.h>
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | pkvm.h | 12 #include <nvhe/gfp.h>
|
/linux-master/block/ |
H A D | blk-ioprio.c | 122 ioprio_alloc_pd(struct gendisk *disk, struct blkcg *blkcg, gfp_t gfp) argument 126 ioprio_blkg = kzalloc(sizeof(*ioprio_blkg), gfp); 140 static struct blkcg_policy_data *ioprio_alloc_cpd(gfp_t gfp) argument 144 blkcg = kzalloc(sizeof(*blkcg), gfp);
|
/linux-master/drivers/iommu/ |
H A D | s390-iommu.c | 157 static unsigned long *dma_alloc_cpu_table(gfp_t gfp) argument 161 table = kmem_cache_alloc(dma_region_table_cache, gfp); 206 static unsigned long *dma_alloc_page_table(gfp_t gfp) argument 210 table = kmem_cache_alloc(dma_page_table_cache, gfp); 219 static unsigned long *dma_get_seg_table_origin(unsigned long *rtep, gfp_t gfp) argument 228 sto = dma_alloc_cpu_table(gfp); 246 static unsigned long *dma_get_page_table_origin(unsigned long *step, gfp_t gfp) argument 255 pto = dma_alloc_page_table(gfp); 272 static unsigned long *dma_walk_cpu_trans(unsigned long *rto, dma_addr_t dma_addr, gfp_t gfp) argument 278 sto = dma_get_seg_table_origin(&rto[rtx], gfp); 552 s390_iommu_validate_trans(struct s390_domain *s390_domain, phys_addr_t pa, dma_addr_t dma_addr, unsigned long nr_pages, int flags, gfp_t gfp) argument 610 s390_iommu_map_pages(struct iommu_domain *domain, unsigned long iova, phys_addr_t paddr, size_t pgsize, size_t pgcount, int prot, gfp_t gfp, size_t *mapped) argument [all...] |