Searched refs:gfp (Results 76 - 100 of 1100) sorted by relevance

1234567891011>>

/linux-master/tools/testing/radix-tree/
H A Dlinux.c9 #include <linux/gfp.h>
57 int gfp)
61 if (!(gfp & __GFP_DIRECT_RECLAIM)) {
84 else if (gfp & __GFP_ZERO)
141 int kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size, argument
154 if (!(gfp & __GFP_DIRECT_RECLAIM)) {
170 if (!(gfp & __GFP_DIRECT_RECLAIM)) {
186 else if (gfp & __GFP_ZERO)
56 kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru, int gfp) argument
/linux-master/net/rxrpc/
H A Dinsecure.c20 static struct rxrpc_txbuf *none_alloc_txbuf(struct rxrpc_call *call, size_t remain, gfp_t gfp) argument
22 return rxrpc_alloc_data_txbuf(call, min_t(size_t, remain, RXRPC_JUMBO_DATALEN), 0, gfp);
/linux-master/drivers/gpu/drm/
H A Ddrm_managed.c87 size_t size, gfp_t gfp, int nid)
96 dr = kmalloc_node_track_caller(tot_size, gfp, nid);
222 * @gfp: GFP allocation flags
228 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) argument
232 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev));
235 size, gfp);
238 dr->node.name = kstrdup_const("kmalloc", gfp);
250 * @gfp: GFP allocation flags
256 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp) argument
265 buf = drmm_kmalloc(dev, size, gfp);
86 alloc_dr(drmres_release_t release, size_t size, gfp_t gfp, int nid) argument
[all...]
/linux-master/mm/
H A Dzpool.c135 * @gfp: The GFP flags to use when allocating the pool.
137 * This creates a new zpool of the specified type. The gfp flags will be
147 struct zpool *zpool_create_pool(const char *type, const char *name, gfp_t gfp) argument
166 zpool = kmalloc(sizeof(*zpool), gfp);
174 zpool->pool = driver->create(name, gfp);
243 * @gfp: The GFP flags to use when allocating memory.
247 * The gfp flags will be used when allocating memory, if the
255 int zpool_malloc(struct zpool *zpool, size_t size, gfp_t gfp, argument
258 return zpool->driver->malloc(zpool->pool, size, gfp, handle);
H A Dpercpu-vm.c76 * @gfp: allocation flags passed to the underlying allocator
84 gfp_t gfp)
89 gfp |= __GFP_HIGHMEM;
95 *pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0);
268 * @gfp: allocation flags passed to the underlying memory allocator
277 int page_start, int page_end, gfp_t gfp)
285 if (pcpu_alloc_pages(chunk, pages, page_start, page_end, gfp))
333 static struct pcpu_chunk *pcpu_create_chunk(gfp_t gfp) argument
338 chunk = pcpu_alloc_chunk(gfp);
82 pcpu_alloc_pages(struct pcpu_chunk *chunk, struct page **pages, int page_start, int page_end, gfp_t gfp) argument
276 pcpu_populate_chunk(struct pcpu_chunk *chunk, int page_start, int page_end, gfp_t gfp) argument
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_internal.c41 gfp_t gfp; local
50 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE;
53 gfp &= ~__GFP_HIGHMEM;
54 gfp |= __GFP_DMA32;
75 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL),
/linux-master/net/sctp/
H A Dulpevent.c49 gfp_t gfp)
54 skb = alloc_skb(size, gfp);
117 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp)
131 sizeof(struct sctp_assoc_change), 0, gfp);
149 MSG_NOTIFICATION, gfp);
244 int flags, int state, int error, gfp_t gfp)
251 MSG_NOTIFICATION, gfp);
376 gfp_t gfp)
395 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp);
432 __u16 flags, __u32 error, gfp_t gfp)
48 sctp_ulpevent_new(int size, __u16 msg_flags, gfp_t gfp) argument
114 sctp_ulpevent_make_assoc_change( const struct sctp_association *asoc, __u16 flags, __u16 state, __u16 error, __u16 outbound, __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) argument
241 sctp_ulpevent_make_peer_addr_change( const struct sctp_association *asoc, const struct sockaddr_storage *aaddr, int flags, int state, int error, gfp_t gfp) argument
374 sctp_ulpevent_make_remote_error(const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, gfp_t gfp) argument
430 sctp_ulpevent_make_send_failed( const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, __u32 error, gfp_t gfp) argument
533 sctp_ulpevent_make_send_failed_event( const struct sctp_association *asoc, struct sctp_chunk *chunk, __u16 flags, __u32 error, gfp_t gfp) argument
577 sctp_ulpevent_make_shutdown_event( const struct sctp_association *asoc, __u16 flags, gfp_t gfp) argument
640 sctp_ulpevent_make_adaptation_indication( const struct sctp_association *asoc, gfp_t gfp) argument
675 sctp_ulpevent_make_rcvmsg(struct sctp_association *asoc, struct sctp_chunk *chunk, gfp_t gfp) argument
771 sctp_ulpevent_make_pdapi( const struct sctp_association *asoc, __u32 indication, __u32 sid, __u32 seq, __u32 flags, gfp_t gfp) argument
825 sctp_ulpevent_make_authkey( const struct sctp_association *asoc, __u16 key_id, __u32 indication, gfp_t gfp) argument
864 sctp_ulpevent_make_sender_dry_event( const struct sctp_association *asoc, gfp_t gfp) argument
888 sctp_ulpevent_make_stream_reset_event( const struct sctp_association *asoc, __u16 flags, __u16 stream_num, __be16 *stream_list, gfp_t gfp) argument
917 sctp_ulpevent_make_assoc_reset_event( const struct sctp_association *asoc, __u16 flags, __u32 local_tsn, __u32 remote_tsn, gfp_t gfp) argument
944 sctp_ulpevent_make_stream_change_event( const struct sctp_association *asoc, __u16 flags, __u32 strchange_instrms, __u32 strchange_outstrms, gfp_t gfp) argument
[all...]
H A Dauth.c58 static struct sctp_auth_bytes *sctp_auth_create_key(__u32 key_len, gfp_t gfp) argument
67 key = kmalloc(sizeof(struct sctp_auth_bytes) + key_len, gfp);
79 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp) argument
84 new = kzalloc(sizeof(struct sctp_shared_key), gfp);
188 gfp_t gfp)
202 new = sctp_auth_create_key(len, gfp);
223 gfp_t gfp)
228 (struct sctp_hmac_algo_param *)asoc->c.auth_hmacs, gfp);
234 gfp_t gfp)
239 gfp);
184 sctp_auth_make_key_vector( struct sctp_random_param *random, struct sctp_chunks_param *chunks, struct sctp_hmac_algo_param *hmacs, gfp_t gfp) argument
221 sctp_auth_make_local_vector( const struct sctp_association *asoc, gfp_t gfp) argument
232 sctp_auth_make_peer_vector( const struct sctp_association *asoc, gfp_t gfp) argument
252 sctp_auth_asoc_set_secret( struct sctp_shared_key *ep_key, struct sctp_auth_bytes *first_vector, struct sctp_auth_bytes *last_vector, gfp_t gfp) argument
286 sctp_auth_asoc_create_secret( const struct sctp_association *asoc, struct sctp_shared_key *ep_key, gfp_t gfp) argument
353 sctp_auth_asoc_copy_shkeys(const struct sctp_endpoint *ep, struct sctp_association *asoc, gfp_t gfp) argument
383 sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp) argument
454 sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp) argument
707 sctp_auth_calculate_hmac(const struct sctp_association *asoc, struct sk_buff *skb, struct sctp_auth_chunk *auth, struct sctp_shared_key *ep_key, gfp_t gfp) argument
1022 sctp_auth_init(struct sctp_endpoint *ep, gfp_t gfp) argument
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_sw_fence.c342 wait_queue_entry_t *wq, gfp_t gfp)
348 might_sleep_if(gfpflags_allow_blocking(gfp));
363 wq = kmalloc(sizeof(*wq), gfp);
365 if (!gfpflags_allow_blocking(gfp))
405 gfp_t gfp)
407 return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp);
477 gfp_t gfp)
484 might_sleep_if(gfpflags_allow_blocking(gfp));
494 gfp);
496 if (!gfpflags_allow_blocking(gfp))
340 __i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence, struct i915_sw_fence *signaler, wait_queue_entry_t *wq, gfp_t gfp) argument
403 i915_sw_fence_await_sw_fence_gfp(struct i915_sw_fence *fence, struct i915_sw_fence *signaler, gfp_t gfp) argument
474 i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence, struct dma_fence *dma, unsigned long timeout, gfp_t gfp) argument
572 i915_sw_fence_await_reservation(struct i915_sw_fence *fence, struct dma_resv *resv, bool write, unsigned long timeout, gfp_t gfp) argument
[all...]
H A Di915_deps.c48 * @gfp: The allocation mode for subsequenst allocations.
50 void i915_deps_init(struct i915_deps *deps, gfp_t gfp) argument
53 deps->gfp = gfp;
85 new_fences = kmalloc_array(new_size, sizeof(*new_fences), deps->gfp);
/linux-master/include/linux/
H A Dbuffer_head.h201 gfp_t gfp);
228 unsigned size, gfp_t gfp);
233 sector_t block, unsigned size, gfp_t gfp);
339 gfp_t gfp; local
341 gfp = mapping_gfp_constraint(bdev->bd_inode->i_mapping, ~__GFP_FS);
342 gfp |= __GFP_NOFAIL;
344 return bdev_getblk(bdev, block, size, gfp);
350 gfp_t gfp; local
352 gfp = mapping_gfp_constraint(bdev->bd_inode->i_mapping, ~__GFP_FS);
353 gfp |
364 sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) argument
[all...]
H A Dptr_ring.h467 static inline void **__ptr_ring_init_queue_alloc(unsigned int size, gfp_t gfp) argument
471 return kvmalloc_array(size, sizeof(void *), gfp | __GFP_ZERO);
487 static inline int ptr_ring_init(struct ptr_ring *r, int size, gfp_t gfp) argument
489 r->queue = __ptr_ring_init_queue_alloc(size, gfp);
559 int size, gfp_t gfp,
590 static inline int ptr_ring_resize(struct ptr_ring *r, int size, gfp_t gfp, argument
594 void **queue = __ptr_ring_init_queue_alloc(size, gfp);
603 old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy);
622 gfp_t gfp, void (*destroy)(void *))
628 queues = kmalloc_array(nrings, sizeof(*queues), gfp);
558 __ptr_ring_swap_queue(struct ptr_ring *r, void **queue, int size, gfp_t gfp, void (*destroy)(void *)) argument
619 ptr_ring_resize_multiple(struct ptr_ring **rings, unsigned int nrings, int size, gfp_t gfp, void (*destroy)(void *)) argument
[all...]
H A Dpercpu.h132 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp) __alloc_size(1);
141 #define alloc_percpu_gfp(type, gfp) \
143 __alignof__(type), gfp)
H A Ddma-mapping.h133 gfp_t gfp, unsigned long attrs);
153 enum dma_data_direction dir, gfp_t gfp, unsigned long attrs);
228 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
289 size_t size, enum dma_data_direction dir, gfp_t gfp,
314 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
321 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
323 struct page *page = dma_alloc_pages(dev, size, dma_handle, dir, gfp);
429 dma_addr_t *dma_handle, gfp_t gfp)
431 return dma_alloc_attrs(dev, size, dma_handle, gfp,
432 (gfp
227 dmam_alloc_attrs(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) argument
288 dma_alloc_noncontiguous(struct device *dev, size_t size, enum dma_data_direction dir, gfp_t gfp, unsigned long attrs) argument
320 dma_alloc_noncoherent(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument
428 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) argument
550 dmam_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) argument
557 dma_alloc_wc(struct device *dev, size_t size, dma_addr_t *dma_addr, gfp_t gfp) argument
[all...]
/linux-master/lib/
H A Dobjpool.c79 if (pool->gfp & GFP_ATOMIC)
80 slot = kmalloc_node(size, pool->gfp, cpu_to_node(i));
82 slot = __vmalloc_node(size, sizeof(void *), pool->gfp,
113 gfp_t gfp, void *context, objpool_init_obj_cb objinit,
136 pool->gfp = gfp & ~__GFP_ZERO;
140 pool->cpu_slots = kzalloc(slot_size, pool->gfp);
112 objpool_init(struct objpool_head *pool, int nr_objs, int object_size, gfp_t gfp, void *context, objpool_init_obj_cb objinit, objpool_fini_cb release) argument
/linux-master/kernel/dma/
H A Dcontiguous.c329 static struct page *cma_alloc_aligned(struct cma *cma, size_t size, gfp_t gfp) argument
333 return cma_alloc(cma, size >> PAGE_SHIFT, align, gfp & __GFP_NOWARN);
340 * @gfp: Allocation flags.
351 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp) argument
358 if (!gfpflags_allow_blocking(gfp))
361 return cma_alloc_aligned(dev->cma_area, size, gfp);
366 if (nid != NUMA_NO_NODE && !(gfp & (GFP_DMA | GFP_DMA32))) {
371 page = cma_alloc_aligned(cma, size, gfp);
378 page = cma_alloc_aligned(cma, size, gfp);
387 return cma_alloc_aligned(dma_contiguous_default_area, size, gfp);
[all...]
H A Dops_helpers.c62 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp)
67 page = dma_alloc_contiguous(dev, size, gfp);
69 page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size));
61 dma_common_alloc_pages(struct device *dev, size_t size, dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) argument
/linux-master/fs/btrfs/
H A Dlru_cache.c108 gfp_t gfp)
114 head = kmalloc(sizeof(*head), gfp);
118 ret = mtree_insert(&cache->entries, key, head, gfp);
106 btrfs_lru_cache_store(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *new_entry, gfp_t gfp) argument
/linux-master/arch/microblaze/kernel/
H A Ddma.c12 #include <linux/gfp.h>
/linux-master/sound/virtio/
H A Dvirtio_ctl_msg.h24 size_t response_size, gfp_t gfp);
/linux-master/include/net/
H A Ddst_cache.h96 * @gfp: allocation flags
98 int dst_cache_init(struct dst_cache *dst_cache, gfp_t gfp);
/linux-master/drivers/net/wireless/intel/iwlwifi/
H A Diwl-modparams.h10 #include <linux/gfp.h>
/linux-master/arch/arm64/kvm/hyp/include/nvhe/
H A Dpkvm.h12 #include <nvhe/gfp.h>
/linux-master/block/
H A Dblk-ioprio.c122 ioprio_alloc_pd(struct gendisk *disk, struct blkcg *blkcg, gfp_t gfp) argument
126 ioprio_blkg = kzalloc(sizeof(*ioprio_blkg), gfp);
140 static struct blkcg_policy_data *ioprio_alloc_cpd(gfp_t gfp) argument
144 blkcg = kzalloc(sizeof(*blkcg), gfp);
/linux-master/drivers/iommu/
H A Ds390-iommu.c157 static unsigned long *dma_alloc_cpu_table(gfp_t gfp) argument
161 table = kmem_cache_alloc(dma_region_table_cache, gfp);
206 static unsigned long *dma_alloc_page_table(gfp_t gfp) argument
210 table = kmem_cache_alloc(dma_page_table_cache, gfp);
219 static unsigned long *dma_get_seg_table_origin(unsigned long *rtep, gfp_t gfp) argument
228 sto = dma_alloc_cpu_table(gfp);
246 static unsigned long *dma_get_page_table_origin(unsigned long *step, gfp_t gfp) argument
255 pto = dma_alloc_page_table(gfp);
272 static unsigned long *dma_walk_cpu_trans(unsigned long *rto, dma_addr_t dma_addr, gfp_t gfp) argument
278 sto = dma_get_seg_table_origin(&rto[rtx], gfp);
552 s390_iommu_validate_trans(struct s390_domain *s390_domain, phys_addr_t pa, dma_addr_t dma_addr, unsigned long nr_pages, int flags, gfp_t gfp) argument
610 s390_iommu_map_pages(struct iommu_domain *domain, unsigned long iova, phys_addr_t paddr, size_t pgsize, size_t pgcount, int prot, gfp_t gfp, size_t *mapped) argument
[all...]

Completed in 471 milliseconds

1234567891011>>