/linux-master/block/ |
H A D | blk-cgroup-rwstat.h | 62 blk_opf_t opf, uint64_t val) 66 if (op_is_discard(opf)) 68 else if (op_is_write(opf)) 75 if (op_is_sync(opf)) 61 blkg_rwstat_add(struct blkg_rwstat *rwstat, blk_opf_t opf, uint64_t val) argument
|
H A D | blk-mq.h | 85 static inline enum hctx_type blk_mq_get_hctx_type(blk_opf_t opf) argument 92 if (opf & REQ_POLLED) 94 else if ((opf & REQ_OP_MASK) == REQ_OP_READ) 102 * @opf: operation type (REQ_OP_*) and flags (e.g. REQ_POLLED). 106 blk_opf_t opf, 109 return ctx->hctxs[blk_mq_get_hctx_type(opf)]; 105 blk_mq_map_queue(struct request_queue *q, blk_opf_t opf, struct blk_mq_ctx *ctx) argument
|
H A D | blk-wbt.c | 533 static inline unsigned int get_limit(struct rq_wb *rwb, blk_opf_t opf) argument 537 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD) 548 if ((opf & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd()) 550 else if ((opf & REQ_BACKGROUND) || close_io(rwb)) { 565 blk_opf_t opf; member in struct:wbt_wait_data 571 return rq_wait_inc_below(rqw, get_limit(data->rwb, data->opf)); 585 blk_opf_t opf) 591 .opf = opf, 584 __wbt_wait(struct rq_wb *rwb, enum wbt_flags wb_acct, blk_opf_t opf) argument
|
H A D | bio.c | 246 unsigned short max_vecs, blk_opf_t opf) 250 bio->bi_opf = opf; 292 * @opf: operation and flags for bio 300 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) argument 308 bio->bi_opf = opf; 349 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) 351 struct bio *new = bio_alloc(bdev, nr_pages, opf, gfp); 433 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, 453 bio_init(bio, bdev, nr_vecs ? bio->bi_inline_vecs : NULL, nr_vecs, opf); 462 * @opf 245 bio_init(struct bio *bio, struct block_device *bdev, struct bio_vec *table, unsigned short max_vecs, blk_opf_t opf) argument 348 blk_next_bio(struct bio *bio, struct block_device *bdev, unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) argument 432 bio_alloc_percpu_cache(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, struct bio_set *bs) argument 492 bio_alloc_bioset(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask, struct bio_set *bs) argument [all...] |
H A D | fops.c | 29 blk_opf_t opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; local 33 opf |= REQ_FUA; 34 return opf; 171 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); local 179 opf |= REQ_ALLOC_CACHE; 180 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, 253 bio = bio_alloc(bdev, nr_pages, opf, GFP_KERNEL); 309 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); local 319 opf |= REQ_ALLOC_CACHE; 320 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNE [all...] |
H A D | bfq-cgroup.c | 223 blk_opf_t opf) 225 blkg_rwstat_add(&bfqg->stats.queued, opf, 1); 231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) argument 233 blkg_rwstat_add(&bfqg->stats.queued, opf, -1); 236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) argument 238 blkg_rwstat_add(&bfqg->stats.merged, opf, 1); 242 u64 io_start_time_ns, blk_opf_t opf) 248 blkg_rwstat_add(&stats->service_time, opf, 251 blkg_rwstat_add(&stats->wait_time, opf, 257 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { } argument 222 bfqg_stats_update_io_add(struct bfq_group *bfqg, struct bfq_queue *bfqq, blk_opf_t opf) argument 241 bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns, u64 io_start_time_ns, blk_opf_t opf) argument 258 bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) argument 259 bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns, u64 io_start_time_ns, blk_opf_t opf) argument [all...] |
H A D | bfq-iosched.h | 1071 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf); 1072 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf); 1074 u64 io_start_time_ns, blk_opf_t opf); 1082 blk_opf_t opf);
|
/linux-master/drivers/md/ |
H A D | dm-io.c | 306 static void do_region(const blk_opf_t opf, unsigned int region, argument 319 const enum req_op op = opf & REQ_OP_MASK; 353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, 385 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, argument 395 opf |= REQ_SYNC; 403 if (where[i].count || (opf & REQ_PREFLUSH)) 404 do_region(opf, i, where + i, dp, io, ioprio); 428 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, 434 if (num_regions > 1 && !op_is_write(opf)) { 451 dispatch_io(opf, num_region 427 sync_io(struct dm_io_client *client, unsigned int num_regions, struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, unsigned long *error_bits, unsigned short ioprio) argument 461 async_io(struct dm_io_client *client, unsigned int num_regions, struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, io_notify_fn fn, void *context, unsigned short ioprio) argument [all...] |
H A D | dm-snap-persistent.c | 232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, argument 241 .bi_opf = opf, 288 static int area_io(struct pstore *ps, blk_opf_t opf) argument 292 return chunk_io(ps, ps->area, chunk, opf, 0);
|
/linux-master/arch/sparc/kernel/ |
H A D | visemul.c | 294 static void edge(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument 307 switch (opf) { 352 switch (opf) { 372 static void array(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument 394 switch (opf) { 477 static void pformat(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument 483 scale = (gsr >> 3) & (opf == FPACK16_OPF ? 0xf : 0x1f); 484 switch (opf) { 589 static void pmul(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument 594 switch (opf) { 705 pcmp(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument 807 unsigned int opf; local [all...] |
/linux-master/drivers/nvme/target/ |
H A D | io-cmd-bdev.c | 249 blk_opf_t opf; local 264 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; 266 opf |= REQ_FUA; 269 opf = REQ_OP_READ; 274 opf |= REQ_NOMERGE; 281 ARRAY_SIZE(req->inline_bvec), opf); 283 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf, 310 opf, GFP_KERNEL);
|
/linux-master/fs/btrfs/ |
H A D | bio.h | 99 struct btrfs_bio *btrfs_bio_alloc(unsigned int nr_vecs, blk_opf_t opf,
|
/linux-master/drivers/scsi/device_handler/ |
H A D | scsi_dh_hp_sw.c | 83 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | local 103 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, 132 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | local 157 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT,
|
H A D | scsi_dh_emc.c | 242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | local 269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len,
|
/linux-master/fs/ |
H A D | direct-io.c | 119 blk_opf_t opf; /* request operation type and flags */ member in struct:dio 170 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 247 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 337 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 405 bio = bio_alloc(bdev, nr_vecs, dio->opf, GFP_KERNEL); 428 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 503 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 607 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 790 const enum req_op dio_op = dio->opf & REQ_OP_MASK; 907 const enum req_op dio_op = dio->opf [all...] |
H A D | mpage.c | 175 blk_opf_t opf = REQ_OP_READ; local 184 opf |= REQ_RAHEAD; 297 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf,
|
/linux-master/fs/gfs2/ |
H A D | lops.h | 20 void gfs2_log_submit_bio(struct bio **biop, blk_opf_t opf);
|
/linux-master/fs/nilfs2/ |
H A D | btnode.c | 73 sector_t pblocknr, blk_opf_t opf, 106 if (opf & REQ_RAHEAD) { 112 } else { /* opf == REQ_OP_READ */ 125 submit_bh(opf, bh); 72 nilfs_btnode_submit_block(struct address_space *btnc, __u64 blocknr, sector_t pblocknr, blk_opf_t opf, struct buffer_head **pbh, sector_t *submit_ptr) argument
|
H A D | mdt.c | 114 nilfs_mdt_submit_block(struct inode *inode, unsigned long blkoff, blk_opf_t opf, argument 129 if (opf & REQ_RAHEAD) { 134 } else /* opf == REQ_OP_READ */ 151 submit_bh(opf, bh); 155 opf & REQ_OP_MASK);
|
/linux-master/fs/hfsplus/ |
H A D | wrapper.c | 33 * @opf: I/O operation type and flags 49 void *buf, void **data, blk_opf_t opf) 51 const enum req_op op = opf & REQ_OP_MASK; 68 bio = bio_alloc(sb->s_bdev, 1, opf, GFP_NOIO); 48 hfsplus_submit_bio(struct super_block *sb, sector_t sector, void *buf, void **data, blk_opf_t opf) argument
|
/linux-master/include/linux/ |
H A D | bio.h | 422 blk_opf_t opf, gfp_t gfp_mask, 435 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) 437 return bio_alloc_bioset(bdev, nr_vecs, opf, gfp_mask, &fs_bio_set); 473 unsigned short max_vecs, blk_opf_t opf); 475 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf); 826 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp); 434 bio_alloc(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) argument
|
H A D | blktrace_api.h | 113 void blk_fill_rwbs(char *rwbs, blk_opf_t opf);
|
/linux-master/drivers/target/ |
H A D | target_core_iblock.c | 358 blk_opf_t opf) 367 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, 737 blk_opf_t opf; local 750 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; 758 opf |= REQ_FUA; 760 opf |= REQ_FUA; 763 opf = REQ_OP_READ; 778 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); 811 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); 357 iblock_get_bio(struct se_cmd *cmd, sector_t lba, u32 sg_num, blk_opf_t opf) argument
|
/linux-master/drivers/block/ |
H A D | brd.c | 211 unsigned int len, unsigned int off, blk_opf_t opf, 217 if (op_is_write(opf)) { 222 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO; 230 if (!op_is_write(opf)) { 210 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, blk_opf_t opf, sector_t sector) argument
|
/linux-master/kernel/trace/ |
H A D | blktrace.c | 216 const blk_opf_t opf, u32 what, int error, 230 const enum req_op op = opf & REQ_OP_MASK; 236 what |= MASK_TC_BIT(opf, SYNC); 237 what |= MASK_TC_BIT(opf, RAHEAD); 238 what |= MASK_TC_BIT(opf, META); 239 what |= MASK_TC_BIT(opf, PREFLUSH); 240 what |= MASK_TC_BIT(opf, FUA); 1869 * @opf: request operation type (REQ_OP_XXX) and flags for the tracepoint 1876 void blk_fill_rwbs(char *rwbs, blk_opf_t opf) argument 1880 if (opf 215 __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes, const blk_opf_t opf, u32 what, int error, int pdu_len, void *pdu_data, u64 cgid) argument [all...] |