Searched refs:opf (Results 1 - 25 of 56) sorted by relevance

123

/linux-master/block/
H A Dblk-cgroup-rwstat.h62 blk_opf_t opf, uint64_t val)
66 if (op_is_discard(opf))
68 else if (op_is_write(opf))
75 if (op_is_sync(opf))
61 blkg_rwstat_add(struct blkg_rwstat *rwstat, blk_opf_t opf, uint64_t val) argument
H A Dblk-mq.h85 static inline enum hctx_type blk_mq_get_hctx_type(blk_opf_t opf) argument
92 if (opf & REQ_POLLED)
94 else if ((opf & REQ_OP_MASK) == REQ_OP_READ)
102 * @opf: operation type (REQ_OP_*) and flags (e.g. REQ_POLLED).
106 blk_opf_t opf,
109 return ctx->hctxs[blk_mq_get_hctx_type(opf)];
105 blk_mq_map_queue(struct request_queue *q, blk_opf_t opf, struct blk_mq_ctx *ctx) argument
H A Dblk-wbt.c533 static inline unsigned int get_limit(struct rq_wb *rwb, blk_opf_t opf) argument
537 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD)
548 if ((opf & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd())
550 else if ((opf & REQ_BACKGROUND) || close_io(rwb)) {
565 blk_opf_t opf; member in struct:wbt_wait_data
571 return rq_wait_inc_below(rqw, get_limit(data->rwb, data->opf));
585 blk_opf_t opf)
591 .opf = opf,
584 __wbt_wait(struct rq_wb *rwb, enum wbt_flags wb_acct, blk_opf_t opf) argument
H A Dbio.c246 unsigned short max_vecs, blk_opf_t opf)
250 bio->bi_opf = opf;
292 * @opf: operation and flags for bio
300 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) argument
308 bio->bi_opf = opf;
349 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp)
351 struct bio *new = bio_alloc(bdev, nr_pages, opf, gfp);
433 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp,
453 bio_init(bio, bdev, nr_vecs ? bio->bi_inline_vecs : NULL, nr_vecs, opf);
462 * @opf
245 bio_init(struct bio *bio, struct block_device *bdev, struct bio_vec *table, unsigned short max_vecs, blk_opf_t opf) argument
348 blk_next_bio(struct bio *bio, struct block_device *bdev, unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) argument
432 bio_alloc_percpu_cache(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, struct bio_set *bs) argument
492 bio_alloc_bioset(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask, struct bio_set *bs) argument
[all...]
H A Dfops.c29 blk_opf_t opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; local
33 opf |= REQ_FUA;
34 return opf;
171 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); local
179 opf |= REQ_ALLOC_CACHE;
180 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL,
253 bio = bio_alloc(bdev, nr_pages, opf, GFP_KERNEL);
309 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); local
319 opf |= REQ_ALLOC_CACHE;
320 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNE
[all...]
H A Dbfq-cgroup.c223 blk_opf_t opf)
225 blkg_rwstat_add(&bfqg->stats.queued, opf, 1);
231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) argument
233 blkg_rwstat_add(&bfqg->stats.queued, opf, -1);
236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) argument
238 blkg_rwstat_add(&bfqg->stats.merged, opf, 1);
242 u64 io_start_time_ns, blk_opf_t opf)
248 blkg_rwstat_add(&stats->service_time, opf,
251 blkg_rwstat_add(&stats->wait_time, opf,
257 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { } argument
222 bfqg_stats_update_io_add(struct bfq_group *bfqg, struct bfq_queue *bfqq, blk_opf_t opf) argument
241 bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns, u64 io_start_time_ns, blk_opf_t opf) argument
258 bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) argument
259 bfqg_stats_update_completion(struct bfq_group *bfqg, u64 start_time_ns, u64 io_start_time_ns, blk_opf_t opf) argument
[all...]
H A Dbfq-iosched.h1071 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf);
1072 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf);
1074 u64 io_start_time_ns, blk_opf_t opf);
1082 blk_opf_t opf);
/linux-master/drivers/md/
H A Ddm-io.c306 static void do_region(const blk_opf_t opf, unsigned int region, argument
319 const enum req_op op = opf & REQ_OP_MASK;
353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO,
385 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, argument
395 opf |= REQ_SYNC;
403 if (where[i].count || (opf & REQ_PREFLUSH))
404 do_region(opf, i, where + i, dp, io, ioprio);
428 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp,
434 if (num_regions > 1 && !op_is_write(opf)) {
451 dispatch_io(opf, num_region
427 sync_io(struct dm_io_client *client, unsigned int num_regions, struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, unsigned long *error_bits, unsigned short ioprio) argument
461 async_io(struct dm_io_client *client, unsigned int num_regions, struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, io_notify_fn fn, void *context, unsigned short ioprio) argument
[all...]
H A Ddm-snap-persistent.c232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, argument
241 .bi_opf = opf,
288 static int area_io(struct pstore *ps, blk_opf_t opf) argument
292 return chunk_io(ps, ps->area, chunk, opf, 0);
/linux-master/arch/sparc/kernel/
H A Dvisemul.c294 static void edge(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument
307 switch (opf) {
352 switch (opf) {
372 static void array(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument
394 switch (opf) {
477 static void pformat(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument
483 scale = (gsr >> 3) & (opf == FPACK16_OPF ? 0xf : 0x1f);
484 switch (opf) {
589 static void pmul(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument
594 switch (opf) {
705 pcmp(struct pt_regs *regs, unsigned int insn, unsigned int opf) argument
807 unsigned int opf; local
[all...]
/linux-master/drivers/nvme/target/
H A Dio-cmd-bdev.c249 blk_opf_t opf; local
264 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE;
266 opf |= REQ_FUA;
269 opf = REQ_OP_READ;
274 opf |= REQ_NOMERGE;
281 ARRAY_SIZE(req->inline_bvec), opf);
283 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf,
310 opf, GFP_KERNEL);
/linux-master/fs/btrfs/
H A Dbio.h99 struct btrfs_bio *btrfs_bio_alloc(unsigned int nr_vecs, blk_opf_t opf,
/linux-master/drivers/scsi/device_handler/
H A Dscsi_dh_hp_sw.c83 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | local
103 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT,
132 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | local
157 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT,
H A Dscsi_dh_emc.c242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | local
269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len,
/linux-master/fs/
H A Ddirect-io.c119 blk_opf_t opf; /* request operation type and flags */ member in struct:dio
170 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
247 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
337 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
405 bio = bio_alloc(bdev, nr_vecs, dio->opf, GFP_KERNEL);
428 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
503 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
607 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
790 const enum req_op dio_op = dio->opf & REQ_OP_MASK;
907 const enum req_op dio_op = dio->opf
[all...]
H A Dmpage.c175 blk_opf_t opf = REQ_OP_READ; local
184 opf |= REQ_RAHEAD;
297 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf,
/linux-master/fs/gfs2/
H A Dlops.h20 void gfs2_log_submit_bio(struct bio **biop, blk_opf_t opf);
/linux-master/fs/nilfs2/
H A Dbtnode.c73 sector_t pblocknr, blk_opf_t opf,
106 if (opf & REQ_RAHEAD) {
112 } else { /* opf == REQ_OP_READ */
125 submit_bh(opf, bh);
72 nilfs_btnode_submit_block(struct address_space *btnc, __u64 blocknr, sector_t pblocknr, blk_opf_t opf, struct buffer_head **pbh, sector_t *submit_ptr) argument
H A Dmdt.c114 nilfs_mdt_submit_block(struct inode *inode, unsigned long blkoff, blk_opf_t opf, argument
129 if (opf & REQ_RAHEAD) {
134 } else /* opf == REQ_OP_READ */
151 submit_bh(opf, bh);
155 opf & REQ_OP_MASK);
/linux-master/fs/hfsplus/
H A Dwrapper.c33 * @opf: I/O operation type and flags
49 void *buf, void **data, blk_opf_t opf)
51 const enum req_op op = opf & REQ_OP_MASK;
68 bio = bio_alloc(sb->s_bdev, 1, opf, GFP_NOIO);
48 hfsplus_submit_bio(struct super_block *sb, sector_t sector, void *buf, void **data, blk_opf_t opf) argument
/linux-master/include/linux/
H A Dbio.h422 blk_opf_t opf, gfp_t gfp_mask,
435 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask)
437 return bio_alloc_bioset(bdev, nr_vecs, opf, gfp_mask, &fs_bio_set);
473 unsigned short max_vecs, blk_opf_t opf);
475 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf);
826 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp);
434 bio_alloc(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) argument
H A Dblktrace_api.h113 void blk_fill_rwbs(char *rwbs, blk_opf_t opf);
/linux-master/drivers/target/
H A Dtarget_core_iblock.c358 blk_opf_t opf)
367 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf,
737 blk_opf_t opf; local
750 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE;
758 opf |= REQ_FUA;
760 opf |= REQ_FUA;
763 opf = REQ_OP_READ;
778 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf);
811 bio = iblock_get_bio(cmd, block_lba, sg_num, opf);
357 iblock_get_bio(struct se_cmd *cmd, sector_t lba, u32 sg_num, blk_opf_t opf) argument
/linux-master/drivers/block/
H A Dbrd.c211 unsigned int len, unsigned int off, blk_opf_t opf,
217 if (op_is_write(opf)) {
222 gfp_t gfp = opf & REQ_NOWAIT ? GFP_NOWAIT : GFP_NOIO;
230 if (!op_is_write(opf)) {
210 brd_do_bvec(struct brd_device *brd, struct page *page, unsigned int len, unsigned int off, blk_opf_t opf, sector_t sector) argument
/linux-master/kernel/trace/
H A Dblktrace.c216 const blk_opf_t opf, u32 what, int error,
230 const enum req_op op = opf & REQ_OP_MASK;
236 what |= MASK_TC_BIT(opf, SYNC);
237 what |= MASK_TC_BIT(opf, RAHEAD);
238 what |= MASK_TC_BIT(opf, META);
239 what |= MASK_TC_BIT(opf, PREFLUSH);
240 what |= MASK_TC_BIT(opf, FUA);
1869 * @opf: request operation type (REQ_OP_XXX) and flags for the tracepoint
1876 void blk_fill_rwbs(char *rwbs, blk_opf_t opf) argument
1880 if (opf
215 __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes, const blk_opf_t opf, u32 what, int error, int pdu_len, void *pdu_data, u64 cgid) argument
[all...]

Completed in 317 milliseconds

123