/linux-master/block/ |
H A D | blk-mq-debugfs.h | 9 struct blk_mq_hw_ctx; 25 struct blk_mq_hw_ctx *hctx); 26 void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx); 33 struct blk_mq_hw_ctx *hctx); 34 void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx); 44 struct blk_mq_hw_ctx *hctx) 48 static inline void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx) 69 struct blk_mq_hw_ctx *hctx) 73 static inline void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx)
|
H A D | blk-mq.h | 26 struct blk_mq_hw_ctx *hctxs[HCTX_MAX_TYPES]; 48 bool blk_mq_dispatch_rq_list(struct blk_mq_hw_ctx *hctx, struct list_head *, 50 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list); 51 struct request *blk_mq_dequeue_from_ctx(struct blk_mq_hw_ctx *hctx, 78 static inline struct blk_mq_hw_ctx *blk_mq_map_queue_type(struct request_queue *q, 105 static inline struct blk_mq_hw_ctx *blk_mq_map_queue(struct request_queue *q, 121 extern void blk_mq_hctx_kobj_init(struct blk_mq_hw_ctx *hctx); 160 struct blk_mq_hw_ctx *hctx; 176 int blk_mq_tag_update_depth(struct blk_mq_hw_ctx *hctx, 189 struct blk_mq_hw_ctx *hct [all...] |
H A D | blk-mq-sched.h | 16 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx); 17 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx); 19 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx); 25 static inline void blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx) 70 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx *hctx) 80 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx *hctx)
|
H A D | blk-mq-sysfs.c | 34 struct blk_mq_hw_ctx *hctx = container_of(kobj, struct blk_mq_hw_ctx, 46 ssize_t (*show)(struct blk_mq_hw_ctx *, char *); 53 struct blk_mq_hw_ctx *hctx; 58 hctx = container_of(kobj, struct blk_mq_hw_ctx, kobj); 70 static ssize_t blk_mq_hw_sysfs_nr_tags_show(struct blk_mq_hw_ctx *hctx, 76 static ssize_t blk_mq_hw_sysfs_nr_reserved_tags_show(struct blk_mq_hw_ctx *hctx, 82 static ssize_t blk_mq_hw_sysfs_cpus_show(struct blk_mq_hw_ctx *hctx, char *page) 144 static void blk_mq_unregister_hctx(struct blk_mq_hw_ctx *hctx) 158 static int blk_mq_register_hctx(struct blk_mq_hw_ctx *hct [all...] |
H A D | elevator.h | 24 struct blk_mq_hw_ctx; 29 int (*init_hctx)(struct blk_mq_hw_ctx *, unsigned int); 30 void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int); 31 void (*depth_updated)(struct blk_mq_hw_ctx *); 41 void (*insert_requests)(struct blk_mq_hw_ctx *hctx, struct list_head *list, 43 struct request *(*dispatch_request)(struct blk_mq_hw_ctx *); 44 bool (*has_work)(struct blk_mq_hw_ctx *);
|
H A D | blk-mq-sched.c | 22 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx) 31 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx) 58 struct blk_mq_hw_ctx *hctx = 87 static int __blk_mq_do_dispatch_sched(struct blk_mq_hw_ctx *hctx) 178 static int blk_mq_do_dispatch_sched(struct blk_mq_hw_ctx *hctx) 196 static struct blk_mq_ctx *blk_mq_next_ctx(struct blk_mq_hw_ctx *hctx, 215 static int blk_mq_do_dispatch_ctx(struct blk_mq_hw_ctx *hctx) 270 static int __blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx) 319 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx) 342 struct blk_mq_hw_ctx *hct [all...] |
H A D | blk-mq-debugfs.c | 178 struct blk_mq_hw_ctx *hctx = data; 206 struct blk_mq_hw_ctx *hctx = data; 315 struct blk_mq_hw_ctx *hctx = m->private; 323 struct blk_mq_hw_ctx *hctx = m->private; 331 struct blk_mq_hw_ctx *hctx = m->private; 345 struct blk_mq_hw_ctx *hctx; 365 struct blk_mq_hw_ctx *hctx = data; 382 struct blk_mq_hw_ctx *hctx = data; 391 struct blk_mq_hw_ctx *hctx = data; 416 struct blk_mq_hw_ctx *hct [all...] |
H A D | blk-mq.c | 49 static void blk_mq_try_issue_list_directly(struct blk_mq_hw_ctx *hctx, 51 static int blk_hctx_poll(struct request_queue *q, struct blk_mq_hw_ctx *hctx, 58 static bool blk_mq_hctx_has_pending(struct blk_mq_hw_ctx *hctx) 68 static void blk_mq_hctx_mark_pending(struct blk_mq_hw_ctx *hctx, 77 static void blk_mq_hctx_clear_pending(struct blk_mq_hw_ctx *hctx, 305 struct blk_mq_hw_ctx *hctx; 350 struct blk_mq_hw_ctx *hctx = data->hctx; 708 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; 1062 static inline void blk_mq_flush_tag_batch(struct blk_mq_hw_ctx *hctx, 1076 struct blk_mq_hw_ctx *cur_hct [all...] |
H A D | kyber-iosched.c | 453 static void kyber_depth_updated(struct blk_mq_hw_ctx *hctx) 464 static int kyber_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) 519 static void kyber_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) 571 struct blk_mq_hw_ctx *hctx = blk_mq_map_queue(q, bio->bi_opf, ctx); 590 static void kyber_insert_requests(struct blk_mq_hw_ctx *hctx, 698 struct blk_mq_hw_ctx *hctx = READ_ONCE(wqe->private); 708 struct blk_mq_hw_ctx *hctx) 756 struct blk_mq_hw_ctx *hctx) 803 static struct request *kyber_dispatch_request(struct blk_mq_hw_ctx *hctx) 849 static bool kyber_has_work(struct blk_mq_hw_ctx *hct [all...] |
H A D | blk-mq-tag.c | 38 void __blk_mq_tag_busy(struct blk_mq_hw_ctx *hctx) 80 void __blk_mq_tag_idle(struct blk_mq_hw_ctx *hctx) 245 struct blk_mq_hw_ctx *hctx; 269 struct blk_mq_hw_ctx *hctx = iter_data->hctx; 312 static void bt_for_each(struct blk_mq_hw_ctx *hctx, struct request_queue *q, 516 struct blk_mq_hw_ctx *hctx; 601 int blk_mq_tag_update_depth(struct blk_mq_hw_ctx *hctx,
|
H A D | blk-flush.c | 355 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; 534 void blk_mq_hctx_set_fq_lock_class(struct blk_mq_hw_ctx *hctx,
|
H A D | mq-deadline.c | 596 static struct request *dd_dispatch_request(struct blk_mq_hw_ctx *hctx) 644 static void dd_depth_updated(struct blk_mq_hw_ctx *hctx) 656 static int dd_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) 794 static void dd_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, 863 static void dd_insert_requests(struct blk_mq_hw_ctx *hctx, 890 static bool dd_has_write_work(struct blk_mq_hw_ctx *hctx) 955 static bool dd_has_work(struct blk_mq_hw_ctx *hctx)
|
H A D | bsg-lib.c | 272 static blk_status_t bsg_queue_rq(struct blk_mq_hw_ctx *hctx,
|
/linux-master/include/linux/ |
H A D | blk-mq.h | 82 struct blk_mq_hw_ctx *mq_hctx; 285 * struct blk_mq_hw_ctx - State for a hardware queue facing the hardware 288 struct blk_mq_hw_ctx { struct 538 blk_status_t (*queue_rq)(struct blk_mq_hw_ctx *, 548 void (*commit_rqs)(struct blk_mq_hw_ctx *); 588 int (*poll)(struct blk_mq_hw_ctx *, struct io_comp_batch *); 600 int (*init_hctx)(struct blk_mq_hw_ctx *, void *, unsigned int); 604 void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int); 878 void blk_mq_stop_hw_queue(struct blk_mq_hw_ctx *hctx); 879 void blk_mq_start_hw_queue(struct blk_mq_hw_ctx *hct [all...] |
/linux-master/drivers/block/rnbd/ |
H A D | rnbd-clt.h | 105 struct blk_mq_hw_ctx *hctx;
|
H A D | rnbd-clt.c | 1096 struct blk_mq_hw_ctx *hctx, 1111 static blk_status_t rnbd_queue_rq(struct blk_mq_hw_ctx *hctx, 1158 static int rnbd_rdma_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) 1312 struct blk_mq_hw_ctx *hctx) 1322 struct blk_mq_hw_ctx *hctx;
|
/linux-master/drivers/nvme/target/ |
H A D | loop.c | 131 static blk_status_t nvme_loop_queue_rq(struct blk_mq_hw_ctx *hctx, 218 static int nvme_loop_init_hctx(struct blk_mq_hw_ctx *hctx, void *data, 238 static int nvme_loop_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data,
|
/linux-master/drivers/s390/block/ |
H A D | scm_blk.c | 283 static blk_status_t scm_blk_request(struct blk_mq_hw_ctx *hctx, 332 static int scm_blk_init_hctx(struct blk_mq_hw_ctx *hctx, void *data, 346 static void scm_blk_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int idx)
|
/linux-master/drivers/block/ |
H A D | z2ram.c | 68 static blk_status_t z2_queue_rq(struct blk_mq_hw_ctx *hctx,
|
H A D | virtio_blk.c | 132 static inline struct virtio_blk_vq *get_virtio_blk_vq(struct blk_mq_hw_ctx *hctx) 214 static int virtblk_map_data(struct blk_mq_hw_ctx *hctx, struct request *req, 378 static void virtio_commit_rqs(struct blk_mq_hw_ctx *hctx) 405 static blk_status_t virtblk_prep_rq(struct blk_mq_hw_ctx *hctx, 427 static blk_status_t virtio_queue_rq(struct blk_mq_hw_ctx *hctx, 1205 static int virtblk_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
|
/linux-master/drivers/md/ |
H A D | dm-rq.c | 477 static blk_status_t dm_mq_queue_rq(struct blk_mq_hw_ctx *hctx,
|
/linux-master/drivers/block/aoe/ |
H A D | aoeblk.c | 250 static blk_status_t aoeblk_queue_rq(struct blk_mq_hw_ctx *hctx,
|
/linux-master/drivers/mmc/core/ |
H A D | queue.c | 229 static blk_status_t mmc_mq_queue_rq(struct blk_mq_hw_ctx *hctx,
|
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 1485 static int null_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob) 1518 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; 1549 static blk_status_t null_queue_rq(struct blk_mq_hw_ctx *hctx, 1630 static int null_init_hctx(struct blk_mq_hw_ctx *hctx, void *driver_data,
|
/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 537 static int nvme_tcp_init_hctx(struct blk_mq_hw_ctx *hctx, void *data, 547 static int nvme_tcp_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data, 2551 static void nvme_tcp_commit_rqs(struct blk_mq_hw_ctx *hctx) 2559 static blk_status_t nvme_tcp_queue_rq(struct blk_mq_hw_ctx *hctx, 2590 static int nvme_tcp_poll(struct blk_mq_hw_ctx *hctx, struct io_comp_batch *iob)
|