Searched refs:bi_opf (Results 1 - 25 of 105) sorted by path

12345

/linux-master/block/
H A Dbfq-iosched.c2469 bfqd->bio_bfqq = bic_to_bfqq(bic, op_is_sync(bio->bi_opf),
3231 bool is_sync = op_is_sync(bio->bi_opf);
H A Dbio-integrity.c85 bio->bi_opf |= REQ_INTEGRITY;
154 bio->bi_opf &= ~REQ_INTEGRITY;
H A Dbio.c250 bio->bi_opf = opf;
308 bio->bi_opf = opf;
818 if (bio->bi_opf & REQ_ALLOC_CACHE)
864 bio = bio_alloc_bioset(bdev, 0, bio_src->bi_opf, gfp, bs);
895 bio_init(bio, bdev, bio_src->bi_io_vec, 0, bio_src->bi_opf);
1391 bio->bi_opf |= REQ_SYNC;
H A Dblk-cgroup.c2063 if (blk_op_is_passthrough(bio->bi_opf))
2093 if (op_is_discard(bio->bi_opf))
2095 if (op_is_write(bio->bi_opf))
H A Dblk-cgroup.h237 return (bio->bi_opf & (REQ_META | REQ_SWAP)) != 0;
H A Dblk-core.c344 if (bio->bi_opf & REQ_NOWAIT) {
516 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio))
555 current->comm, bio->bi_bdev, bio->bi_opf,
609 bio->bi_opf |= REQ_NOMERGE;
760 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev))
777 if (op_is_flush(bio->bi_opf)) {
782 bio->bi_opf &= ~(REQ_PREFLUSH | REQ_FUA);
1241 sizeof_field(struct bio, bi_opf));
H A Dblk-crypto-fallback.c171 bio_src->bi_opf);
H A Dblk-iocost.c2683 (bio->bi_opf & REQ_SWAP) == REQ_SWAP);
2784 (bio->bi_opf & REQ_SWAP) == REQ_SWAP);
H A Dblk-iolatency.c481 (bio->bi_opf & REQ_SWAP) == REQ_SWAP);
H A Dblk-lib.c128 bio->bi_opf |= REQ_NOUNMAP;
H A Dblk-map.c245 if (bio->bi_opf & REQ_ALLOC_CACHE) {
804 bio->bi_opf &= ~REQ_OP_MASK;
805 bio->bi_opf |= req_op(rq);
H A Dblk-merge.c312 if (bio->bi_opf & REQ_NOWAIT) {
375 split->bi_opf |= REQ_NOMERGE;
744 WARN_ON_ONCE((bio->bi_opf & REQ_FAILFAST_MASK) &&
745 (bio->bi_opf & REQ_FAILFAST_MASK) != ff);
746 bio->bi_opf |= ff;
753 if (bio->bi_opf & REQ_RAHEAD)
756 return bio->bi_opf & REQ_FAILFAST_MASK;
768 if (bio->bi_opf & REQ_RAHEAD)
769 bio->bi_opf |= REQ_FAILFAST_MASK;
773 req->cmd_flags |= bio->bi_opf
[all...]
H A Dblk-mq-sched.c352 hctx = blk_mq_map_queue(q, bio->bi_opf, ctx);
H A Dblk-mq-sched.h33 return !(bio->bi_opf & REQ_NOMERGE_FLAGS);
H A Dblk-mq.c961 req->cmd_flags |= req->bio->bi_opf & REQ_FAILFAST_MASK;
1271 if (rq->bio && rq->bio->bi_opf & REQ_POLLED)
2548 if (bio->bi_opf & REQ_RAHEAD)
2856 .cmd_flags = bio->bi_opf,
2872 if (bio->bi_opf & REQ_NOWAIT)
2913 rq->cmd_flags = bio->bi_opf;
2934 const int is_sync = op_is_sync(bio->bi_opf);
2943 rq = blk_mq_peek_cached_request(plug, q, bio->bi_opf);
3009 if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq))
H A Dblk-throttle.h183 blkg_rwstat_add(&tg->stat_bytes, bio->bi_opf,
186 blkg_rwstat_add(&tg->stat_ios, bio->bi_opf, 1);
H A Dblk-wbt.c604 if ((bio->bi_opf & (REQ_SYNC | REQ_IDLE)) ==
658 __wbt_wait(rwb, flags, bio->bi_opf);
H A Dblk-zoned.c991 bio->bi_opf &= ~REQ_OP_MASK;
992 bio->bi_opf |= REQ_OP_WRITE | REQ_NOMERGE;
1053 if (bio->bi_opf & REQ_NOWAIT)
1128 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio))
1238 bio->bi_opf &= ~REQ_OP_MASK;
1239 bio->bi_opf |= REQ_OP_ZONE_APPEND;
H A Dbounce.c168 bio_src->bi_opf, GFP_NOIO, &bounce_bio_set);
H A Dfops.c85 bio.bi_opf |= REQ_NOWAIT;
227 bio->bi_opf |= REQ_NOWAIT;
347 bio->bi_opf |= REQ_NOWAIT;
350 bio->bi_opf |= REQ_POLLED;
H A Dkyber-iosched.c571 struct blk_mq_hw_ctx *hctx = blk_mq_map_queue(q, bio->bi_opf, ctx);
574 unsigned int sched_domain = kyber_sched_domain(bio->bi_opf);
/linux-master/drivers/block/
H A Dbrd.c241 bio->bi_opf, sector);
243 if (err == -ENOMEM && bio->bi_opf & REQ_NOWAIT) {
/linux-master/drivers/block/drbd/
H A Ddrbd_main.c1649 return (bio->bi_opf & REQ_SYNC ? DP_RW_SYNC : 0) |
1650 (bio->bi_opf & REQ_FUA ? DP_FUA : 0) |
1651 (bio->bi_opf & REQ_PREFLUSH ? DP_FLUSH : 0) |
1655 (DP_ZEROES |(!(bio->bi_opf & REQ_NOUNMAP) ? DP_DISCARD : 0))
1659 return bio->bi_opf & REQ_SYNC ? DP_RW_SYNC : 0;
H A Ddrbd_req.c266 !(req->master_bio->bi_opf & REQ_RAHEAD) &&
1121 D_ASSERT(device, req->master_bio->bi_opf & REQ_PREFLUSH);
1159 else if (bio->bi_opf & REQ_RAHEAD)
1174 ((bio->bi_opf & REQ_NOUNMAP) ? 0 : EE_TRIM));
H A Ddrbd_worker.c261 if (bio->bi_opf & REQ_RAHEAD)

Completed in 1035 milliseconds

12345