Searched refs:blk_mq_hw_ctx (Results 1 - 25 of 52) sorted by relevance

123

/linux-master/block/
H A Dblk-mq-debugfs.h9 struct blk_mq_hw_ctx;
25 struct blk_mq_hw_ctx *hctx);
26 void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx);
33 struct blk_mq_hw_ctx *hctx);
34 void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx);
44 struct blk_mq_hw_ctx *hctx)
48 static inline void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx)
69 struct blk_mq_hw_ctx *hctx)
73 static inline void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx)
H A Dblk-mq.h26 struct blk_mq_hw_ctx *hctxs[HCTX_MAX_TYPES];
48 bool blk_mq_dispatch_rq_list(struct blk_mq_hw_ctx *hctx, struct list_head *,
50 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list);
51 struct request *blk_mq_dequeue_from_ctx(struct blk_mq_hw_ctx *hctx,
78 static inline struct blk_mq_hw_ctx *blk_mq_map_queue_type(struct request_queue *q,
105 static inline struct blk_mq_hw_ctx *blk_mq_map_queue(struct request_queue *q,
121 extern void blk_mq_hctx_kobj_init(struct blk_mq_hw_ctx *hctx);
160 struct blk_mq_hw_ctx *hctx;
176 int blk_mq_tag_update_depth(struct blk_mq_hw_ctx *hctx,
189 struct blk_mq_hw_ctx *hct
[all...]
H A Dblk-mq-sched.h16 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx);
17 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx);
19 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx);
25 static inline void blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx)
70 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx *hctx)
80 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx *hctx)
H A Dblk-mq-sysfs.c34 struct blk_mq_hw_ctx *hctx = container_of(kobj, struct blk_mq_hw_ctx,
46 ssize_t (*show)(struct blk_mq_hw_ctx *, char *);
53 struct blk_mq_hw_ctx *hctx;
58 hctx = container_of(kobj, struct blk_mq_hw_ctx, kobj);
70 static ssize_t blk_mq_hw_sysfs_nr_tags_show(struct blk_mq_hw_ctx *hctx,
76 static ssize_t blk_mq_hw_sysfs_nr_reserved_tags_show(struct blk_mq_hw_ctx *hctx,
82 static ssize_t blk_mq_hw_sysfs_cpus_show(struct blk_mq_hw_ctx *hctx, char *page)
144 static void blk_mq_unregister_hctx(struct blk_mq_hw_ctx *hctx)
158 static int blk_mq_register_hctx(struct blk_mq_hw_ctx *hct
[all...]
H A Delevator.h24 struct blk_mq_hw_ctx;
29 int (*init_hctx)(struct blk_mq_hw_ctx *, unsigned int);
30 void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int);
31 void (*depth_updated)(struct blk_mq_hw_ctx *);
41 void (*insert_requests)(struct blk_mq_hw_ctx *hctx, struct list_head *list,
43 struct request *(*dispatch_request)(struct blk_mq_hw_ctx *);
44 bool (*has_work)(struct blk_mq_hw_ctx *);
H A Dblk-mq-sched.c22 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx)
31 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx)
58 struct blk_mq_hw_ctx *hctx =
87 static int __blk_mq_do_dispatch_sched(struct blk_mq_hw_ctx *hctx)
178 static int blk_mq_do_dispatch_sched(struct blk_mq_hw_ctx *hctx)
196 static struct blk_mq_ctx *blk_mq_next_ctx(struct blk_mq_hw_ctx *hctx,
215 static int blk_mq_do_dispatch_ctx(struct blk_mq_hw_ctx *hctx)
270 static int __blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx)
319 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx)
342 struct blk_mq_hw_ctx *hct
[all...]
H A Dblk-mq-debugfs.c178 struct blk_mq_hw_ctx *hctx = data;
206 struct blk_mq_hw_ctx *hctx = data;
315 struct blk_mq_hw_ctx *hctx = m->private;
323 struct blk_mq_hw_ctx *hctx = m->private;
331 struct blk_mq_hw_ctx *hctx = m->private;
345 struct blk_mq_hw_ctx *hctx;
365 struct blk_mq_hw_ctx *hctx = data;
382 struct blk_mq_hw_ctx *hctx = data;
391 struct blk_mq_hw_ctx *hctx = data;
416 struct blk_mq_hw_ctx *hct
[all...]
H A Dblk-mq.c49 static void blk_mq_try_issue_list_directly(struct blk_mq_hw_ctx *hctx,
51 static int blk_hctx_poll(struct request_queue *q, struct blk_mq_hw_ctx *hctx,
58 static bool blk_mq_hctx_has_pending(struct blk_mq_hw_ctx *hctx)
68 static void blk_mq_hctx_mark_pending(struct blk_mq_hw_ctx *hctx,
77 static void blk_mq_hctx_clear_pending(struct blk_mq_hw_ctx *hctx,
305 struct blk_mq_hw_ctx *hctx;
350 struct blk_mq_hw_ctx *hctx = data->hctx;
708 struct blk_mq_hw_ctx *hctx = rq->mq_hctx;
1062 static inline void blk_mq_flush_tag_batch(struct blk_mq_hw_ctx *hctx,
1076 struct blk_mq_hw_ctx *cur_hct
[all...]
H A Dkyber-iosched.c453 static void kyber_depth_updated(struct blk_mq_hw_ctx *hctx)
464 static int kyber_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx)
519 static void kyber_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx)
571 struct blk_mq_hw_ctx *hctx = blk_mq_map_queue(q, bio->bi_opf, ctx);
590 static void kyber_insert_requests(struct blk_mq_hw_ctx *hctx,
698 struct blk_mq_hw_ctx *hctx = READ_ONCE(wqe->private);
708 struct blk_mq_hw_ctx *hctx)
756 struct blk_mq_hw_ctx *hctx)
803 static struct request *kyber_dispatch_request(struct blk_mq_hw_ctx *hctx)
849 static bool kyber_has_work(struct blk_mq_hw_ctx *hct
[all...]
H A Dblk-mq-tag.c38 void __blk_mq_tag_busy(struct blk_mq_hw_ctx *hctx)
80 void __blk_mq_tag_idle(struct blk_mq_hw_ctx *hctx)
245 struct blk_mq_hw_ctx *hctx;
269 struct blk_mq_hw_ctx *hctx = iter_data->hctx;
312 static void bt_for_each(struct blk_mq_hw_ctx *hctx, struct request_queue *q,
516 struct blk_mq_hw_ctx *hctx;
601 int blk_mq_tag_update_depth(struct blk_mq_hw_ctx *hctx,
H A Dblk-flush.c355 struct blk_mq_hw_ctx *hctx = rq->mq_hctx;
534 void blk_mq_hctx_set_fq_lock_class(struct blk_mq_hw_ctx *hctx,
H A Dmq-deadline.c596 static struct request *dd_dispatch_request(struct blk_mq_hw_ctx *hctx)
644 static void dd_depth_updated(struct blk_mq_hw_ctx *hctx)
656 static int dd_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx)
794 static void dd_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
863 static void dd_insert_requests(struct blk_mq_hw_ctx *hctx,
890 static bool dd_has_write_work(struct blk_mq_hw_ctx *hctx)
955 static bool dd_has_work(struct blk_mq_hw_ctx *hctx)
H A Dbsg-lib.c272 static blk_status_t bsg_queue_rq(struct blk_mq_hw_ctx *hctx,
/linux-master/include/linux/
H A Dblk-mq.h82 struct blk_mq_hw_ctx *mq_hctx;
285 * struct blk_mq_hw_ctx - State for a hardware queue facing the hardware
288 struct blk_mq_hw_ctx { struct
538 blk_status_t (*queue_rq)(struct blk_mq_hw_ctx *,
548 void (*commit_rqs)(struct blk_mq_hw_ctx *);
588 int (*poll)(struct blk_mq_hw_ctx *, struct io_comp_batch *);
600 int (*init_hctx)(struct blk_mq_hw_ctx *, void *, unsigned int);
604 void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int);
878 void blk_mq_stop_hw_queue(struct blk_mq_hw_ctx *hctx);
879 void blk_mq_start_hw_queue(struct blk_mq_hw_ctx *hct
[all...]
/linux-master/drivers/block/rnbd/
H A Drnbd-clt.h105 struct blk_mq_hw_ctx *hctx;
H A Drnbd-clt.c1096 struct blk_mq_hw_ctx *hctx,
1111 static blk_status_t rnbd_queue_rq(struct blk_mq_hw_ctx *hctx,
1158 static int rnbd_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
1312 struct blk_mq_hw_ctx *hctx)
1322 struct blk_mq_hw_ctx *hctx;
/linux-master/drivers/nvme/target/
H A Dloop.c131 static blk_status_t nvme_loop_queue_rq(struct blk_mq_hw_ctx *hctx,
218 static int nvme_loop_init_hctx(struct blk_mq_hw_ctx *hctx, void *data,
238 static int nvme_loop_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data,
/linux-master/drivers/s390/block/
H A Dscm_blk.c283 static blk_status_t scm_blk_request(struct blk_mq_hw_ctx *hctx,
332 static int scm_blk_init_hctx(struct blk_mq_hw_ctx *hctx, void *data,
346 static void scm_blk_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int idx)
/linux-master/drivers/block/
H A Dz2ram.c68 static blk_status_t z2_queue_rq(struct blk_mq_hw_ctx *hctx,
H A Dvirtio_blk.c132 static inline struct virtio_blk_vq *get_virtio_blk_vq(struct blk_mq_hw_ctx *hctx)
214 static int virtblk_map_data(struct blk_mq_hw_ctx *hctx, struct request *req,
378 static void virtio_commit_rqs(struct blk_mq_hw_ctx *hctx)
405 static blk_status_t virtblk_prep_rq(struct blk_mq_hw_ctx *hctx,
427 static blk_status_t virtio_queue_rq(struct blk_mq_hw_ctx *hctx,
1205 static int virtblk_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
/linux-master/drivers/md/
H A Ddm-rq.c477 static blk_status_t dm_mq_queue_rq(struct blk_mq_hw_ctx *hctx,
/linux-master/drivers/block/aoe/
H A Daoeblk.c250 static blk_status_t aoeblk_queue_rq(struct blk_mq_hw_ctx *hctx,
/linux-master/drivers/mmc/core/
H A Dqueue.c229 static blk_status_t mmc_mq_queue_rq(struct blk_mq_hw_ctx *hctx,
/linux-master/drivers/block/null_blk/
H A Dmain.c1485 static int null_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
1518 struct blk_mq_hw_ctx *hctx = rq->mq_hctx;
1549 static blk_status_t null_queue_rq(struct blk_mq_hw_ctx *hctx,
1630 static int null_init_hctx(struct blk_mq_hw_ctx *hctx, void *driver_data,
/linux-master/drivers/nvme/host/
H A Dtcp.c537 static int nvme_tcp_init_hctx(struct blk_mq_hw_ctx *hctx, void *data,
547 static int nvme_tcp_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data,
2551 static void nvme_tcp_commit_rqs(struct blk_mq_hw_ctx *hctx)
2559 static blk_status_t nvme_tcp_queue_rq(struct blk_mq_hw_ctx *hctx,
2590 static int nvme_tcp_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)

Completed in 367 milliseconds

123