/linux-master/drivers/md/bcache/ |
H A D | debug.h | 5 struct bio; 12 void bch_data_verify(struct cached_dev *dc, struct bio *bio); 21 static inline void bch_data_verify(struct cached_dev *dc, struct bio *bio) {} argument
|
/linux-master/drivers/nvdimm/ |
H A D | virtio_pmem.h | 54 int async_pmem_flush(struct nd_region *nd_region, struct bio *bio);
|
/linux-master/drivers/xen/ |
H A D | biomerge.c | 2 #include <linux/bio.h>
|
/linux-master/arch/m68k/emu/ |
H A D | nfblock.c | 60 static void nfhd_submit_bio(struct bio *bio) argument 62 struct nfhd_device *dev = bio->bi_bdev->bd_disk->private_data; 66 sector_t sec = bio->bi_iter.bi_sector; 68 dir = bio_data_dir(bio); 70 bio_for_each_segment(bvec, bio, iter) { 77 bio_endio(bio);
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | papr_scm.c | 130 struct bio *bio __maybe_unused)
|
/linux-master/arch/xtensa/platforms/iss/ |
H A D | simdisk.c | 18 #include <linux/bio.h> 103 static void simdisk_submit_bio(struct bio *bio) argument 105 struct simdisk *dev = bio->bi_bdev->bd_disk->private_data; 108 sector_t sector = bio->bi_iter.bi_sector; 110 bio_for_each_segment(bvec, bio, iter) { 115 bio_data_dir(bio) == WRITE); 120 bio_endio(bio);
|
/linux-master/block/ |
H A D | Makefile | 6 obj-y := bdev.o fops.o bio.o elevator.o blk-core.o blk-sysfs.o \ 29 obj-$(CONFIG_BLK_DEV_INTEGRITY) += bio-integrity.o blk-integrity.o
|
H A D | bfq-cgroup.c | 345 struct bfq_group *bfqg = blkg_to_bfqg(rq->bio->bi_blkg); 599 struct bfq_group *bfq_bio_bfqg(struct bfq_data *bfqd, struct bio *bio) argument 601 struct blkcg_gq *blkg = bio->bi_blkg; 611 bio_associate_blkg_from_css(bio, &blkg->blkcg->css); 616 bio_associate_blkg_from_css(bio, 739 * so that we cannot merge bio to a request from the 779 void bfq_bic_update_cgroup(struct bfq_io_cq *bic, struct bio *bio) argument 782 struct bfq_group *bfqg = bfq_bio_bfqg(bfqd, bio); 1463 bfq_bic_update_cgroup(struct bfq_io_cq *bic, struct bio *bio) argument 1470 bfq_bio_bfqg(struct bfq_data *bfqd, struct bio *bio) argument [all...] |
H A D | bfq-iosched.c | 1790 * Get the index of the actuator that will serve bio. 1792 static unsigned int bfq_actuator_index(struct bfq_data *bfqd, struct bio *bio) argument 1801 /* bio_end_sector(bio) gives the sector after the last one */ 1802 end = bio_end_sector(bio) - 1; 1811 "bfq_actuator_index: bio sector out of ranges: end=%llu\n", 1835 unsigned int act_idx = bfq_actuator_index(bfqd, rq->bio); 2360 struct bio *bio, 2367 return elv_rb_find(&bfqq->sort_list, bio_end_sector(bio)); 2359 bfq_find_rq_fmerge(struct bfq_data *bfqd, struct bio *bio, struct request_queue *q) argument 2445 bfq_bio_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs) argument 2485 bfq_request_merge(struct request_queue *q, struct request **req, struct bio *bio) argument 3227 bfq_allow_bio_merge(struct request_queue *q, struct request *rq, struct bio *bio) argument 5561 bfq_check_ioprio_change(struct bfq_io_cq *bic, struct bio *bio) argument 5832 bfq_get_queue(struct bfq_data *bfqd, struct bio *bio, bool is_sync, struct bfq_io_cq *bic, bool respawn) argument 6739 bfq_get_bfqq_handle_split(struct bfq_data *bfqd, struct bfq_io_cq *bic, struct bio *bio, bool split, bool is_sync, bool *new_queue) argument 6848 struct bio *bio = rq->bio; local [all...] |
H A D | bfq-iosched.h | 801 * bic associated with the task issuing current bio for 803 * be able to perform the bic lookup, needed by bio-merge 809 /* bfqq associated with the task issuing current bio for merging */ 1089 void bfq_bic_update_cgroup(struct bfq_io_cq *bic, struct bio *bio); 1091 struct bfq_group *bfq_bio_bfqg(struct bfq_data *bfqd, struct bio *bio);
|
H A D | bio-integrity.c | 3 * bio-integrity.c - bio data integrity extensions 12 #include <linux/bio.h> 39 * bio_integrity_alloc - Allocate integrity payload and attach it to bio 40 * @bio: bio to attach integrity metadata to 44 * Description: This function prepares a bio for attaching integrity 48 struct bio_integrity_payload *bio_integrity_alloc(struct bio *bio, argument 53 struct bio_set *bs = bio 142 bio_integrity_free(struct bio *bio) argument 166 bio_integrity_add_page(struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 206 bio_integrity_copy_user(struct bio *bio, struct bio_vec *bvec, int nr_vecs, unsigned int len, unsigned int direction, u32 seed) argument 265 bio_integrity_init_user(struct bio *bio, struct bio_vec *bvec, int nr_vecs, unsigned int len, u32 seed) argument 311 bio_integrity_map_user(struct bio *bio, void __user *ubuf, ssize_t bytes, u32 seed) argument 383 bio_integrity_process(struct bio *bio, struct bvec_iter *proc_iter, integrity_processing_fn *proc_fn) argument 427 bio_integrity_prep(struct bio *bio) argument 532 struct bio *bio = bip->bip_bio; local 557 __bio_integrity_endio(struct bio *bio) argument 582 bio_integrity_advance(struct bio *bio, unsigned int bytes_done) argument 598 bio_integrity_trim(struct bio *bio) argument 615 bio_integrity_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp_mask) argument [all...] |
H A D | bio.c | 7 #include <linux/bio.h> 31 struct bio *free_list; 32 struct bio *free_list_irq; 67 * fs_bio_set is the bio_set containing bio and iovec memory pools used by 92 snprintf(bslab->name, sizeof(bslab->name), "bio-%d", size); 114 return bs->front_pad + sizeof(struct bio) + bs->back_pad; 143 if (WARN(!bslab, KERN_ERR "bio: unable to find slab!\n")) 213 void bio_uninit(struct bio *bio) argument 216 if (bio 228 bio_free(struct bio *bio) argument 245 bio_init(struct bio *bio, struct block_device *bdev, struct bio_vec *table, unsigned short max_vecs, blk_opf_t opf) argument 300 bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) argument 312 __bio_chain_endio(struct bio *bio) argument 322 bio_chain_endio(struct bio *bio) argument 338 bio_chain(struct bio *bio, struct bio *parent) argument 348 blk_next_bio(struct bio *bio, struct block_device *bdev, unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) argument 365 struct bio *bio; local 382 struct bio *bio; local 437 struct bio *bio; local 497 struct bio *bio; local 601 struct bio *bio; local 609 zero_fill_bio_iter(struct bio *bio, struct bvec_iter start) argument 629 bio_truncate(struct bio *bio, unsigned new_size) argument 681 guard_bio_eod(struct bio *bio) argument 707 struct bio *bio; local 760 bio_put_percpu_cache(struct bio *bio) argument 800 bio_put(struct bio *bio) argument 814 __bio_clone(struct bio *bio, struct bio *bio_src, gfp_t gfp) argument 851 struct bio *bio; local 879 bio_init_clone(struct block_device *bdev, struct bio *bio, struct bio *bio_src, gfp_t gfp) argument 900 bio_full(struct bio *bio, unsigned len) argument 968 bio_add_hw_page(struct request_queue *q, struct bio *bio, struct page *page, unsigned int len, unsigned int offset, unsigned int max_sectors, bool *same_page) argument 1023 bio_add_pc_page(struct request_queue *q, struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 1048 bio_add_zone_append_page(struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 1075 __bio_add_page(struct bio *bio, struct page *page, unsigned int len, unsigned int off) argument 1097 bio_add_page(struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 1121 bio_add_folio_nofail(struct bio *bio, struct folio *folio, size_t len, size_t off) argument 1143 bio_add_folio(struct bio *bio, struct folio *folio, size_t len, size_t off) argument 1152 __bio_release_pages(struct bio *bio, bool mark_dirty) argument 1175 bio_iov_bvec_set(struct bio *bio, struct iov_iter *iter) argument 1195 bio_iov_add_page(struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 1215 bio_iov_add_zone_append_page(struct bio *bio, struct page *page, unsigned int len, unsigned int offset) argument 1241 __bio_iov_iter_get_pages(struct bio *bio, struct iov_iter *iter) argument 1333 bio_iov_iter_get_pages(struct bio *bio, struct iov_iter *iter) argument 1356 submit_bio_wait_endio(struct bio *bio) argument 1372 submit_bio_wait(struct bio *bio) argument 1387 __bio_advance(struct bio *bio, unsigned bytes) argument 1435 bio_free_pages(struct bio *bio) argument 1468 bio_set_pages_dirty(struct bio *bio) argument 1502 struct bio *bio, *next; local 1517 bio_check_pages_dirty(struct bio *bio) argument 1539 bio_remaining_done(struct bio *bio) argument 1571 bio_endio(struct bio *bio) argument 1621 bio_split(struct bio *bio, int sectors, gfp_t gfp, struct bio_set *bs) argument 1660 bio_trim(struct bio *bio, sector_t offset, sector_t size) argument [all...] |
H A D | blk-cgroup-fc-appid.c | 46 * blkcg_get_fc_appid - get the fc app identifier associated with a bio 47 * @bio: target bio 51 char *blkcg_get_fc_appid(struct bio *bio) argument 53 if (!bio->bi_blkg || bio->bi_blkg->blkcg->fc_app_id[0] == '\0') 55 return bio->bi_blkg->blkcg->fc_app_id;
|
H A D | blk-cgroup.c | 215 struct bio *bio; local 230 while ((bio = bio_list_pop(&bios))) 231 submit_bio(bio); 237 * When a shared kthread issues a bio for a cgroup, doing so synchronously can 242 void blkcg_punt_bio_submit(struct bio *bio) argument 244 struct blkcg_gq *blkg = bio->bi_blkg; 248 bio_list_add(&blkg->async_bios, bio); 253 submit_bio(bio); 278 bio_blkcg_css(struct bio *bio) argument 2009 blkg_tryget_closest(struct bio *bio, struct cgroup_subsys_state *css) argument 2042 bio_associate_blkg_from_css(struct bio *bio, struct cgroup_subsys_state *css) argument 2066 bio_associate_blkg(struct bio *bio) argument 2098 blk_cgroup_io_type(struct bio *bio) argument 2107 blk_cgroup_bio_start(struct bio *bio) argument [all...] |
H A D | blk-cgroup.h | 225 * bio_issue_as_root_blkg - see if this bio needs to be issued as root blkg 226 * @return: true if this bio needs to be submitted with the root blkg context. 228 * In order to avoid priority inversions we sometimes need to issue a bio as if 231 * the bio and attach the appropriate blkg to the bio. Then we call this helper 235 static inline bool bio_issue_as_root_blkg(struct bio *bio) argument 237 return (bio->bi_opf & (REQ_META | REQ_SWAP)) != 0; 380 static inline void blkcg_bio_issue_init(struct bio *bio) argument 463 blk_cgroup_mergeable(struct request *rq, struct bio *bio) argument 502 blkcg_bio_issue_init(struct bio *bio) argument 503 blk_cgroup_bio_start(struct bio *bio) argument 504 blk_cgroup_mergeable(struct request *rq, struct bio *bio) argument [all...] |
H A D | blk-core.c | 9 * bio rewrite, highmem i/o, etc, Jens Axboe <axboe@suse.de> - may 2001 17 #include <linux/bio.h> 135 * string format. Useful in the debugging and tracing bio or request. For 339 int __bio_queue_enter(struct request_queue *q, struct bio *bio) argument 342 struct gendisk *disk = bio->bi_bdev->bd_disk; 344 if (bio->bi_opf & REQ_NOWAIT) { 347 bio_wouldblock_error(bio); 369 bio_io_error(bio); 513 static inline void bio_check_ro(struct bio *bi argument 532 should_fail_bio(struct bio *bio) argument 545 bio_check_eod(struct bio *bio) argument 565 blk_partition_remap(struct bio *bio) argument 584 blk_check_zone_append(struct request_queue *q, struct bio *bio) argument 615 __submit_bio(struct bio *bio) argument 649 __submit_bio_noacct(struct bio *bio) argument 693 __submit_bio_noacct_mq(struct bio *bio) argument 706 submit_bio_noacct_nocheck(struct bio *bio) argument 743 submit_bio_noacct(struct bio *bio) argument 850 bio_set_ioprio(struct bio *bio) argument 871 submit_bio(struct bio *bio) argument 897 bio_poll(struct bio *bio, struct io_comp_batch *iob, unsigned int flags) argument 952 struct bio *bio; local 1018 bio_start_io_acct(struct bio *bio) argument 1041 bio_end_io_acct_remapped(struct bio *bio, unsigned long start_time, struct block_device *orig_bdev) argument [all...] |
H A D | blk-crypto-fallback.c | 38 "Number of preallocated bio fallback crypto contexts for blk-crypto to use during crypto API fallback"); 43 * Copy of the bvec_iter when this bio was submitted. 44 * We only want to en/decrypt the part of the bio as described by the 45 * bvec_iter upon submission because bio might be split before being 52 struct bio *bio; member in struct:bio_fallback_crypt_ctx::__anon291::__anon292 144 static void blk_crypto_fallback_encrypt_endio(struct bio *enc_bio) 146 struct bio *src_bio = enc_bio->bi_private; 160 static struct bio *blk_crypto_fallback_clone_bio(struct bio *bio_sr 165 struct bio *bio; local 213 struct bio *bio = *bio_ptr; local 381 struct bio *bio = f_ctx->bio; local 452 blk_crypto_fallback_decrypt_endio(struct bio *bio) argument 493 struct bio *bio = *bio_ptr; local [all...] |
H A D | blk-crypto-internal.h | 9 #include <linux/bio.h> 31 bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio); 37 struct bio *bio) 40 bio->bi_crypt_context); 44 struct bio *bio) 46 return bio_crypt_ctx_mergeable(bio->bi_crypt_context, 47 bio 36 bio_crypt_ctx_back_mergeable(struct request *req, struct bio *bio) argument 43 bio_crypt_ctx_front_mergeable(struct request *req, struct bio *bio) argument 96 bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio) argument 102 bio_crypt_ctx_front_mergeable(struct request *req, struct bio *bio) argument 108 bio_crypt_ctx_back_mergeable(struct request *req, struct bio *bio) argument 135 bio_crypt_advance(struct bio *bio, unsigned int bytes) argument 142 bio_crypt_free_ctx(struct bio *bio) argument 148 bio_crypt_do_front_merge(struct request *rq, struct bio *bio) argument 200 blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument [all...] |
H A D | blk-crypto.c | 12 #include <linux/bio.h> 50 * concurrently) * (maximum recursive depth of a bio), so that we don't 58 "Number of bio crypto contexts to preallocate"); 87 panic("Failed to allocate mem for bio crypt ctxs\n"); 91 void bio_crypt_set_ctx(struct bio *bio, const struct blk_crypto_key *key, argument 107 bio->bi_crypt_context = bc; 110 void __bio_crypt_free_ctx(struct bio *bio) argument 112 mempool_free(bio 144 __bio_crypt_advance(struct bio *bio, unsigned int bytes) argument 193 bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio) argument 213 bio_crypt_check_alignment(struct bio *bio) argument 275 struct bio *bio = *bio_ptr; local 303 __blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio, gfp_t gfp_mask) argument [all...] |
H A D | blk-flush.c | 56 * complete. The first completion updates the contained bio but doesn't 57 * finish it so that the bio submitter is notified only after the whole 62 * bio attached to it, which is guaranteed as they aren't allowed to be 68 #include <linux/bio.h> 128 * After flush data completion, @rq->bio is %NULL but we need to 129 * complete the bio again. @rq->biotail is guaranteed to equal the 130 * original @rq->bio. Restore it. 132 rq->bio = rq->biotail; 404 WARN_ON_ONCE(rq->bio != rq->biotail); 471 struct bio bi local [all...] |
H A D | blk-integrity.c | 12 #include <linux/bio.h> 22 * @bio: bio with integrity metadata attached 25 * scatterlist corresponding to the integrity metadata in a bio. 27 int blk_rq_count_integrity_sg(struct request_queue *q, struct bio *bio) argument 35 bio_for_each_integrity_vec(iv, bio, iter) { 61 * @bio: bio with integrity metadata attached 68 int blk_rq_map_integrity_sg(struct request_queue *q, struct bio *bi argument 187 blk_integrity_merge_bio(struct request_queue *q, struct request *req, struct bio *bio) argument [all...] |
H A D | blk-ioc.c | 8 #include <linux/bio.h>
|
H A D | blk-iocost.c | 106 * fill up and a bio which is ready to be issued has to wait for a request 567 struct bio *bio; member in struct:iocg_wait 717 static void iocg_commit_bio(struct ioc_gq *iocg, struct bio *bio, argument 722 bio->bi_iocost_cost = cost; 1473 iocg_commit_bio(ctx->iocg, wait->bio, wait->abs_cost, cost); 2521 static void calc_vtime_cost_builtin(struct bio *bio, struct ioc_gq *iocg, argument 2526 u64 pages = max_t(u64, bio_sectors(bio) >> IOC_SECT_TO_PAGE_SHIF 2566 calc_vtime_cost(struct bio *bio, struct ioc_gq *iocg, bool is_merge) argument 2599 ioc_rqos_throttle(struct rq_qos *rqos, struct bio *bio) argument 2734 ioc_rqos_merge(struct rq_qos *rqos, struct request *rq, struct bio *bio) argument 2793 ioc_rqos_done_bio(struct rq_qos *rqos, struct bio *bio) argument [all...] |
H A D | blk-iolatency.c | 7 * - It's bio based, so the latency covers the whole block layer in addition to 14 * that we can have as many outstanding bio's as we're allowed to. Only at 463 static void blkcg_iolatency_throttle(struct rq_qos *rqos, struct bio *bio) argument 466 struct blkcg_gq *blkg = bio->bi_blkg; 467 bool issue_as_root = bio_issue_as_root_blkg(bio); 481 (bio->bi_opf & REQ_SWAP) == REQ_SWAP); 507 * We don't want to count issue_as_root bio's in the cgroups latency 591 static void blkcg_iolatency_done_bio(struct rq_qos *rqos, struct bio *bio) argument [all...] |
H A D | blk-ioprio.c | 87 static struct ioprio_blkcg *ioprio_blkcg_from_bio(struct bio *bio) argument 89 struct blkg_policy_data *pd = blkg_to_pd(bio->bi_blkg, &ioprio_policy); 187 void blkcg_set_ioprio(struct bio *bio) argument 189 struct ioprio_blkcg *blkcg = ioprio_blkcg_from_bio(bio); 204 if (IOPRIO_PRIO_CLASS(bio->bi_ioprio) != IOPRIO_CLASS_RT) 205 bio->bi_ioprio = IOPRIO_PRIO_VALUE(IOPRIO_CLASS_RT, 4); 213 * If the bio I/O priority equals IOPRIO_CLASS_NONE, the cgroup I/O 214 * priority is assigned to the bio [all...] |