/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_cmd.h | 252 gfp_t gfp_mask);
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.h | 81 gfp_t gfp_mask, int coherent);
|
/linux-master/drivers/net/wireless/ath/ |
H A D | main.c | 31 gfp_t gfp_mask) 49 skb = __dev_alloc_skb(len + common->cachelsz - 1, gfp_mask); 29 ath_rxbuf_alloc(struct ath_common *common, u32 len, gfp_t gfp_mask) argument
|
/linux-master/drivers/scsi/aic94xx/ |
H A D | aic94xx_hwi.h | 363 gfp_t gfp_mask);
|
/linux-master/net/ceph/ |
H A D | msgpool.c | 12 static void *msgpool_alloc(gfp_t gfp_mask, void *arg) argument 18 gfp_mask, true);
|
/linux-master/arch/arm/kernel/ |
H A D | module.c | 39 gfp_t gfp_mask = GFP_KERNEL; local 44 gfp_mask |= __GFP_NOWARN; 47 gfp_mask, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE,
|
/linux-master/arch/s390/kernel/ |
H A D | module.c | 59 gfp_t gfp_mask = GFP_KERNEL; local 66 MODULES_END, gfp_mask, PAGE_KERNEL, 69 if (p && (kasan_alloc_module_shadow(p, size, gfp_mask) < 0)) {
|
/linux-master/arch/s390/pci/ |
H A D | pci_clp.c | 90 static void *clp_alloc_block(gfp_t gfp_mask) argument 92 return (void *) __get_free_pages(gfp_mask, get_order(CLP_BLK_SIZE));
|
/linux-master/arch/um/drivers/ |
H A D | net_kern.c | 370 struct transport *transport, gfp_t gfp_mask) 379 device = kzalloc(sizeof(*device), gfp_mask); 548 void **init_out, char **mac_out, gfp_t gfp_mask) 562 *init_out = kmalloc(transport->setup_size, gfp_mask); 369 eth_configure(int n, void *init, char *mac, struct transport *transport, gfp_t gfp_mask) argument 547 check_transport(struct transport *transport, char *eth, int n, void **init_out, char **mac_out, gfp_t gfp_mask) argument
|
/linux-master/arch/x86/kernel/ |
H A D | module.c | 70 gfp_t gfp_mask = GFP_KERNEL; local 78 MODULES_END, gfp_mask, PAGE_KERNEL, 82 if (p && (kasan_alloc_module_shadow(p, size, gfp_mask) < 0)) {
|
/linux-master/arch/x86/platform/efi/ |
H A D | efi_64.c | 73 gfp_t gfp_mask; local 75 gfp_mask = GFP_KERNEL | __GFP_ZERO; 76 efi_pgd = (pgd_t *)__get_free_pages(gfp_mask, PGD_ALLOCATION_ORDER);
|
/linux-master/block/ |
H A D | bio-integrity.c | 41 * @gfp_mask: Memory allocation mask 49 gfp_t gfp_mask, 60 bip = kmalloc(struct_size(bip, bip_inline_vecs, nr_vecs), gfp_mask); 63 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask); 76 &bip->bip_max_vcnt, gfp_mask); 611 * @gfp_mask: Memory allocation mask 616 gfp_t gfp_mask) 623 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt); 48 bio_integrity_alloc(struct bio *bio, gfp_t gfp_mask, unsigned int nr_vecs) argument 615 bio_integrity_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp_mask) argument
|
H A D | bio.c | 183 gfp_t gfp_mask) 204 bvl = kmem_cache_alloc(bvs->slab, bvec_alloc_gfp(gfp_mask)); 205 if (likely(bvl) || !(gfp_mask & __GFP_DIRECT_RECLAIM)) 210 return mempool_alloc(pool, gfp_mask); 463 * @gfp_mask: the GFP_* mask given to the slab allocator 493 blk_opf_t opf, gfp_t gfp_mask, 496 gfp_t saved_gfp = gfp_mask; 507 gfp_mask, bs); 541 gfp_mask &= ~__GFP_DIRECT_RECLAIM; 543 p = mempool_alloc(&bs->bio_pool, gfp_mask); 182 bvec_alloc(mempool_t *pool, unsigned short *nr_vecs, gfp_t gfp_mask) argument 492 bio_alloc_bioset(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask, struct bio_set *bs) argument 599 bio_kmalloc(unsigned short nr_vecs, gfp_t gfp_mask) argument [all...] |
H A D | blk-cgroup.c | 301 * @gfp_mask: allocation mask to use 306 gfp_t gfp_mask) 312 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node); 315 if (percpu_ref_init(&blkg->refcnt, blkg_release, 0, gfp_mask)) 317 blkg->iostat_cpu = alloc_percpu_gfp(struct blkg_iostat_set, gfp_mask); 346 pd = pol->pd_alloc_fn(disk, blkcg, gfp_mask); 305 blkg_alloc(struct blkcg *blkcg, struct gendisk *disk, gfp_t gfp_mask) argument
|
H A D | blk-crypto-internal.h | 189 gfp_t gfp_mask); 195 * @gfp_mask: Memory allocation flags 198 * @gfp_mask doesn't include %__GFP_DIRECT_RECLAIM. 201 gfp_t gfp_mask) 204 return __blk_crypto_rq_bio_prep(rq, bio, gfp_mask); 200 blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument
|
H A D | blk-crypto.c | 92 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) 97 * The caller must use a gfp_mask that contains __GFP_DIRECT_RECLAIM so 100 WARN_ON_ONCE(!(gfp_mask & __GFP_DIRECT_RECLAIM)); 102 bc = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); 116 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) argument 118 dst->bi_crypt_context = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); 304 gfp_t gfp_mask) 307 rq->crypt_ctx = mempool_alloc(bio_crypt_ctx_pool, gfp_mask); 91 bio_crypt_set_ctx(struct bio *bio, const struct blk_crypto_key *key, const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) argument 303 __blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument
|
H A D | blk-lib.c | 39 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) 67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask); 92 * @gfp_mask: memory allocation flags (for bio_alloc) 98 sector_t nr_sects, gfp_t gfp_mask) 105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio); 119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, 137 bio = blk_next_bio(bio, bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask); 166 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, 178 REQ_OP_WRITE, gfp_mask); 201 * @gfp_mask 38 __blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) argument 97 blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask) argument 118 __blkdev_issue_write_zeroes(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument 165 __blkdev_issue_zero_pages(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) argument 215 __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument 249 blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, unsigned flags) argument [all...] |
H A D | blk-map.c | 22 gfp_t gfp_mask) 29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask); 132 struct iov_iter *iter, gfp_t gfp_mask) 142 bmd = bio_alloc_map_data(iter, gfp_mask); 157 bio = bio_kmalloc(nr_pages, gfp_mask); 185 page = alloc_page(GFP_NOIO | gfp_mask); 254 unsigned int nr_vecs, gfp_t gfp_mask) 259 bio = bio_alloc_bioset(NULL, nr_vecs, rq->cmd_flags, gfp_mask, 264 bio = bio_kmalloc(nr_vecs, gfp_mask); 273 gfp_t gfp_mask) 21 bio_alloc_map_data(struct iov_iter *data, gfp_t gfp_mask) argument 131 bio_copy_user_iov(struct request *rq, struct rq_map_data *map_data, struct iov_iter *iter, gfp_t gfp_mask) argument 253 blk_rq_map_bio_alloc(struct request *rq, unsigned int nr_vecs, gfp_t gfp_mask) argument 272 bio_map_user_iov(struct request *rq, struct iov_iter *iter, gfp_t gfp_mask) argument 389 bio_map_kern(struct request_queue *q, void *data, unsigned int len, gfp_t gfp_mask) argument 474 bio_copy_kern(struct request_queue *q, void *data, unsigned int len, gfp_t gfp_mask, int reading) argument 632 blk_rq_map_user_iov(struct request_queue *q, struct request *rq, struct rq_map_data *map_data, const struct iov_iter *iter, gfp_t gfp_mask) argument 687 blk_rq_map_user(struct request_queue *q, struct request *rq, struct rq_map_data *map_data, void __user *ubuf, unsigned long len, gfp_t gfp_mask) argument 701 blk_rq_map_user_io(struct request *req, struct rq_map_data *map_data, void __user *ubuf, unsigned long buf_len, gfp_t gfp_mask, bool vec, int iov_count, bool check_iter_count, int rw) argument 782 blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf, unsigned int len, gfp_t gfp_mask) argument [all...] |
H A D | blk-mq.c | 3120 * @gfp_mask: memory allocation mask for bio 3133 struct bio_set *bs, gfp_t gfp_mask, 3143 bio = bio_alloc_clone(rq->q->disk->part0, bio_src, gfp_mask, 3171 if (rq->bio && blk_crypto_rq_bio_prep(rq, rq->bio, gfp_mask) < 0) 3132 blk_rq_prep_clone(struct request *rq, struct request *rq_src, struct bio_set *bs, gfp_t gfp_mask, int (*bio_ctr)(struct bio *, struct bio *, void *), void *data) argument
|
H A D | blk.h | 89 gfp_t gfp_mask);
|
/linux-master/drivers/base/ |
H A D | devres.c | 1112 * @gfp_mask: Allocation gfp flags 1123 gfp_t gfp_mask, unsigned int order) 1128 addr = __get_free_pages(gfp_mask, order); 1122 devm_get_free_pages(struct device *dev, gfp_t gfp_mask, unsigned int order) argument
|
/linux-master/drivers/block/drbd/ |
H A D | drbd_receiver.c | 361 unsigned int request_size, unsigned int payload_size, gfp_t gfp_mask) __must_hold(local) 371 peer_req = mempool_alloc(&drbd_ee_mempool, gfp_mask & ~__GFP_HIGHMEM); 373 if (!(gfp_mask & __GFP_NOWARN)) 380 gfpflags_allow_blocking(gfp_mask));
|
/linux-master/drivers/connector/ |
H A D | connector.c | 62 gfp_t gfp_mask, netlink_filter_fn filter, 97 skb = nlmsg_new(size, gfp_mask); 115 gfp_mask, filter, 118 !gfpflags_allow_blocking(gfp_mask)); 124 gfp_t gfp_mask) 126 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask, 61 cn_netlink_send_mult(struct cn_msg *msg, u16 len, u32 portid, u32 __group, gfp_t gfp_mask, netlink_filter_fn filter, void *filter_data) argument 123 cn_netlink_send(struct cn_msg *msg, u32 portid, u32 __group, gfp_t gfp_mask) argument
|
/linux-master/drivers/crypto/ccp/ |
H A D | sev-dev.c | 427 static struct page *__snp_alloc_firmware_pages(gfp_t gfp_mask, int order) argument 436 page = alloc_pages(gfp_mask, order); 452 void *snp_alloc_firmware_page(gfp_t gfp_mask) argument 456 page = __snp_alloc_firmware_pages(gfp_mask, 0);
|
/linux-master/drivers/firewire/ |
H A D | core-cdev.c | 487 struct client_resource *resource, gfp_t gfp_mask) 489 bool preload = gfpflags_allow_blocking(gfp_mask); 494 idr_preload(gfp_mask); 486 add_client_resource(struct client *client, struct client_resource *resource, gfp_t gfp_mask) argument
|