Searched refs:gfp_mask (Results 1 - 25 of 225) sorted by path

123456789

/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_cmd.h252 gfp_t gfp_mask);
/linux-master/drivers/net/ethernet/mellanox/mlx4/
H A Dicm.h81 gfp_t gfp_mask, int coherent);
/linux-master/drivers/net/wireless/ath/
H A Dmain.c31 gfp_t gfp_mask)
49 skb = __dev_alloc_skb(len + common->cachelsz - 1, gfp_mask);
29 ath_rxbuf_alloc(struct ath_common *common, u32 len, gfp_t gfp_mask) argument
/linux-master/drivers/scsi/aic94xx/
H A Daic94xx_hwi.h363 gfp_t gfp_mask);
/linux-master/net/ceph/
H A Dmsgpool.c12 static void *msgpool_alloc(gfp_t gfp_mask, void *arg) argument
18 gfp_mask, true);
/linux-master/arch/arm/kernel/
H A Dmodule.c39 gfp_t gfp_mask = GFP_KERNEL; local
44 gfp_mask |= __GFP_NOWARN;
47 gfp_mask, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE,
/linux-master/arch/s390/kernel/
H A Dmodule.c59 gfp_t gfp_mask = GFP_KERNEL; local
66 MODULES_END, gfp_mask, PAGE_KERNEL,
69 if (p && (kasan_alloc_module_shadow(p, size, gfp_mask) < 0)) {
/linux-master/arch/s390/pci/
H A Dpci_clp.c90 static void *clp_alloc_block(gfp_t gfp_mask) argument
92 return (void *) __get_free_pages(gfp_mask, get_order(CLP_BLK_SIZE));
/linux-master/arch/um/drivers/
H A Dnet_kern.c370 struct transport *transport, gfp_t gfp_mask)
379 device = kzalloc(sizeof(*device), gfp_mask);
548 void **init_out, char **mac_out, gfp_t gfp_mask)
562 *init_out = kmalloc(transport->setup_size, gfp_mask);
369 eth_configure(int n, void *init, char *mac, struct transport *transport, gfp_t gfp_mask) argument
547 check_transport(struct transport *transport, char *eth, int n, void **init_out, char **mac_out, gfp_t gfp_mask) argument
/linux-master/arch/x86/kernel/
H A Dmodule.c70 gfp_t gfp_mask = GFP_KERNEL; local
78 MODULES_END, gfp_mask, PAGE_KERNEL,
82 if (p && (kasan_alloc_module_shadow(p, size, gfp_mask) < 0)) {
/linux-master/arch/x86/platform/efi/
H A Defi_64.c73 gfp_t gfp_mask; local
75 gfp_mask = GFP_KERNEL | __GFP_ZERO;
76 efi_pgd = (pgd_t *)__get_free_pages(gfp_mask, PGD_ALLOCATION_ORDER);
/linux-master/block/
H A Dbio-integrity.c41 * @gfp_mask: Memory allocation mask
49 gfp_t gfp_mask,
60 bip = kmalloc(struct_size(bip, bip_inline_vecs, nr_vecs), gfp_mask);
63 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask);
76 &bip->bip_max_vcnt, gfp_mask);
611 * @gfp_mask: Memory allocation mask
616 gfp_t gfp_mask)
623 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt);
48 bio_integrity_alloc(struct bio *bio, gfp_t gfp_mask, unsigned int nr_vecs) argument
615 bio_integrity_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp_mask) argument
H A Dbio.c183 gfp_t gfp_mask)
204 bvl = kmem_cache_alloc(bvs->slab, bvec_alloc_gfp(gfp_mask));
205 if (likely(bvl) || !(gfp_mask & __GFP_DIRECT_RECLAIM))
210 return mempool_alloc(pool, gfp_mask);
463 * @gfp_mask: the GFP_* mask given to the slab allocator
493 blk_opf_t opf, gfp_t gfp_mask,
496 gfp_t saved_gfp = gfp_mask;
507 gfp_mask, bs);
541 gfp_mask &= ~__GFP_DIRECT_RECLAIM;
543 p = mempool_alloc(&bs->bio_pool, gfp_mask);
182 bvec_alloc(mempool_t *pool, unsigned short *nr_vecs, gfp_t gfp_mask) argument
492 bio_alloc_bioset(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask, struct bio_set *bs) argument
599 bio_kmalloc(unsigned short nr_vecs, gfp_t gfp_mask) argument
[all...]
H A Dblk-cgroup.c301 * @gfp_mask: allocation mask to use
306 gfp_t gfp_mask)
312 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node);
315 if (percpu_ref_init(&blkg->refcnt, blkg_release, 0, gfp_mask))
317 blkg->iostat_cpu = alloc_percpu_gfp(struct blkg_iostat_set, gfp_mask);
346 pd = pol->pd_alloc_fn(disk, blkcg, gfp_mask);
305 blkg_alloc(struct blkcg *blkcg, struct gendisk *disk, gfp_t gfp_mask) argument
H A Dblk-crypto-internal.h189 gfp_t gfp_mask);
195 * @gfp_mask: Memory allocation flags
198 * @gfp_mask doesn't include %__GFP_DIRECT_RECLAIM.
201 gfp_t gfp_mask)
204 return __blk_crypto_rq_bio_prep(rq, bio, gfp_mask);
200 blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument
H A Dblk-crypto.c92 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask)
97 * The caller must use a gfp_mask that contains __GFP_DIRECT_RECLAIM so
100 WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM));
102 bc = mempool_alloc(bio_crypt_ctx_pool, gfp_mask);
116 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) argument
118 dst->bi_crypt_context = mempool_alloc(bio_crypt_ctx_pool, gfp_mask);
304 gfp_t gfp_mask)
307 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask);
91 bio_crypt_set_ctx(struct bio *bio, const struct blk_crypto_key *key, const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) argument
303 __blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument
H A Dblk-lib.c39 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop)
67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask);
92 * @gfp_mask: memory allocation flags (for bio_alloc)
98 sector_t nr_sects, gfp_t gfp_mask)
105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio);
119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
137 bio = blk_next_bio(bio, bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask);
166 sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
178 REQ_OP_WRITE, gfp_mask);
201 * @gfp_mask
38 __blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) argument
97 blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask) argument
118 __blkdev_issue_write_zeroes(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument
165 __blkdev_issue_zero_pages(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) argument
215 __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument
249 blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, unsigned flags) argument
[all...]
H A Dblk-map.c22 gfp_t gfp_mask)
29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask);
132 struct iov_iter *iter, gfp_t gfp_mask)
142 bmd = bio_alloc_map_data(iter, gfp_mask);
157 bio = bio_kmalloc(nr_pages, gfp_mask);
185 page = alloc_page(GFP_NOIO | gfp_mask);
254 unsigned int nr_vecs, gfp_t gfp_mask)
259 bio = bio_alloc_bioset(NULL, nr_vecs, rq->cmd_flags, gfp_mask,
264 bio = bio_kmalloc(nr_vecs, gfp_mask);
273 gfp_t gfp_mask)
21 bio_alloc_map_data(struct iov_iter *data, gfp_t gfp_mask) argument
131 bio_copy_user_iov(struct request *rq, struct rq_map_data *map_data, struct iov_iter *iter, gfp_t gfp_mask) argument
253 blk_rq_map_bio_alloc(struct request *rq, unsigned int nr_vecs, gfp_t gfp_mask) argument
272 bio_map_user_iov(struct request *rq, struct iov_iter *iter, gfp_t gfp_mask) argument
389 bio_map_kern(struct request_queue *q, void *data, unsigned int len, gfp_t gfp_mask) argument
474 bio_copy_kern(struct request_queue *q, void *data, unsigned int len, gfp_t gfp_mask, int reading) argument
632 blk_rq_map_user_iov(struct request_queue *q, struct request *rq, struct rq_map_data *map_data, const struct iov_iter *iter, gfp_t gfp_mask) argument
687 blk_rq_map_user(struct request_queue *q, struct request *rq, struct rq_map_data *map_data, void __user *ubuf, unsigned long len, gfp_t gfp_mask) argument
701 blk_rq_map_user_io(struct request *req, struct rq_map_data *map_data, void __user *ubuf, unsigned long buf_len, gfp_t gfp_mask, bool vec, int iov_count, bool check_iter_count, int rw) argument
782 blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf, unsigned int len, gfp_t gfp_mask) argument
[all...]
H A Dblk-mq.c3120 * @gfp_mask: memory allocation mask for bio
3133 struct bio_set *bs, gfp_t gfp_mask,
3143 bio = bio_alloc_clone(rq->q->disk->part0, bio_src, gfp_mask,
3171 if (rq->bio && blk_crypto_rq_bio_prep(rq, rq->bio, gfp_mask) < 0)
3132 blk_rq_prep_clone(struct request *rq, struct request *rq_src, struct bio_set *bs, gfp_t gfp_mask, int (*bio_ctr)(struct bio *, struct bio *, void *), void *data) argument
H A Dblk.h89 gfp_t gfp_mask);
/linux-master/drivers/base/
H A Ddevres.c1112 * @gfp_mask: Allocation gfp flags
1123 gfp_t gfp_mask, unsigned int order)
1128 addr = __get_free_pages(gfp_mask, order);
1122 devm_get_free_pages(struct device *dev, gfp_t gfp_mask, unsigned int order) argument
/linux-master/drivers/block/drbd/
H A Ddrbd_receiver.c361 unsigned int request_size, unsigned int payload_size, gfp_t gfp_mask) __must_hold(local)
371 peer_req = mempool_alloc(&drbd_ee_mempool, gfp_mask & ~__GFP_HIGHMEM);
373 if (!(gfp_mask & __GFP_NOWARN))
380 gfpflags_allow_blocking(gfp_mask));
/linux-master/drivers/connector/
H A Dconnector.c62 gfp_t gfp_mask, netlink_filter_fn filter,
97 skb = nlmsg_new(size, gfp_mask);
115 gfp_mask, filter,
118 !gfpflags_allow_blocking(gfp_mask));
124 gfp_t gfp_mask)
126 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask,
61 cn_netlink_send_mult(struct cn_msg *msg, u16 len, u32 portid, u32 __group, gfp_t gfp_mask, netlink_filter_fn filter, void *filter_data) argument
123 cn_netlink_send(struct cn_msg *msg, u32 portid, u32 __group, gfp_t gfp_mask) argument
/linux-master/drivers/crypto/ccp/
H A Dsev-dev.c427 static struct page *__snp_alloc_firmware_pages(gfp_t gfp_mask, int order) argument
436 page = alloc_pages(gfp_mask, order);
452 void *snp_alloc_firmware_page(gfp_t gfp_mask) argument
456 page = __snp_alloc_firmware_pages(gfp_mask, 0);
/linux-master/drivers/firewire/
H A Dcore-cdev.c487 struct client_resource *resource, gfp_t gfp_mask)
489 bool preload = gfpflags_allow_blocking(gfp_mask);
494 idr_preload(gfp_mask);
486 add_client_resource(struct client *client, struct client_resource *resource, gfp_t gfp_mask) argument

Completed in 301 milliseconds

123456789