/linux-master/include/scsi/ |
H A D | scsi_tcq.h | 32 if (hwq < shost->tag_set.nr_hw_queues) { 33 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq],
|
/linux-master/drivers/s390/block/ |
H A D | dasd_genhd.c | 56 block->tag_set.ops = &dasd_mq_ops; 57 block->tag_set.cmd_size = sizeof(struct dasd_ccw_req); 58 block->tag_set.nr_hw_queues = nr_hw_queues; 59 block->tag_set.queue_depth = queue_depth; 60 block->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 61 block->tag_set.numa_node = NUMA_NO_NODE; 62 rc = blk_mq_alloc_tag_set(&block->tag_set); 66 gdp = blk_mq_alloc_disk(&block->tag_set, &lim, block); 68 blk_mq_free_tag_set(&block->tag_set); 127 blk_mq_free_tag_set(&block->tag_set); [all...] |
H A D | scm_blk.h | 20 struct blk_mq_tag_set tag_set; member in struct:scm_blk_dev
|
H A D | scm_blk.c | 461 bdev->tag_set.ops = &scm_mq_ops; 462 bdev->tag_set.cmd_size = sizeof(blk_status_t); 463 bdev->tag_set.nr_hw_queues = nr_requests; 464 bdev->tag_set.queue_depth = nr_requests_per_io * nr_requests; 465 bdev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 466 bdev->tag_set.numa_node = NUMA_NO_NODE; 468 ret = blk_mq_alloc_tag_set(&bdev->tag_set); 472 bdev->gendisk = blk_mq_alloc_disk(&bdev->tag_set, &lim, scmdev); 508 blk_mq_free_tag_set(&bdev->tag_set); 518 blk_mq_free_tag_set(&bdev->tag_set); [all...] |
/linux-master/drivers/block/ |
H A D | z2ram.c | 310 static struct blk_mq_tag_set tag_set; variable in typeref:struct:blk_mq_tag_set 321 disk = blk_mq_alloc_disk(&tag_set, NULL, NULL); 352 tag_set.ops = &z2_mq_ops; 353 tag_set.nr_hw_queues = 1; 354 tag_set.nr_maps = 1; 355 tag_set.queue_depth = 16; 356 tag_set.numa_node = NUMA_NO_NODE; 357 tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 358 ret = blk_mq_alloc_tag_set(&tag_set); 371 blk_mq_free_tag_set(&tag_set); [all...] |
H A D | ataflop.c | 305 struct blk_mq_tag_set tag_set; member in struct:atari_floppy_struct 1997 disk = blk_mq_alloc_disk(&unit[drive].tag_set, NULL, NULL); 2051 blk_mq_free_tag_set(&unit[i].tag_set); 2069 blk_mq_free_tag_set(&fs->tag_set); 2082 memset(&unit[i].tag_set, 0, sizeof(unit[i].tag_set)); 2083 unit[i].tag_set.ops = &ataflop_mq_ops; 2084 unit[i].tag_set.nr_hw_queues = 1; 2085 unit[i].tag_set.nr_maps = 1; 2086 unit[i].tag_set [all...] |
H A D | nbd.c | 116 struct blk_mq_tag_set tag_set; member in struct:nbd_device 254 blk_mq_free_tag_set(&nbd->tag_set); 457 (config->num_connections == 1 && nbd->tag_set.timeout)) { 490 if (!nbd->tag_set.timeout) { 779 if (hwq < nbd->tag_set.nr_hw_queues) 780 req = blk_mq_tag_to_rq(nbd->tag_set.tags[hwq], 946 blk_mq_tagset_busy_iter(&nbd->tag_set, nbd_clear_req, NULL); 1245 if (nbd->tag_set.timeout) 1246 sock->sk->sk_sndtimeo = nbd->tag_set.timeout; 1374 nbd->tag_set [all...] |
H A D | virtio_blk.c | 72 struct blk_mq_tag_set tag_set; member in struct:virtio_blk 1493 memset(&vblk->tag_set, 0, sizeof(vblk->tag_set)); 1494 vblk->tag_set.ops = &virtio_mq_ops; 1495 vblk->tag_set.queue_depth = queue_depth; 1496 vblk->tag_set.numa_node = NUMA_NO_NODE; 1497 vblk->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 1498 vblk->tag_set.cmd_size = 1501 vblk->tag_set.driver_data = vblk; 1502 vblk->tag_set [all...] |
/linux-master/drivers/md/ |
H A D | dm-rq.c | 541 md->tag_set = kzalloc_node(sizeof(struct blk_mq_tag_set), GFP_KERNEL, md->numa_node_id); 542 if (!md->tag_set) 545 md->tag_set->ops = &dm_mq_ops; 546 md->tag_set->queue_depth = dm_get_blk_mq_queue_depth(); 547 md->tag_set->numa_node = md->numa_node_id; 548 md->tag_set->flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_STACKING; 549 md->tag_set->nr_hw_queues = dm_get_blk_mq_nr_hw_queues(); 550 md->tag_set->driver_data = md; 552 md->tag_set->cmd_size = sizeof(struct dm_rq_target_io); 556 md->tag_set [all...] |
/linux-master/drivers/mmc/core/ |
H A D | queue.c | 379 disk = blk_mq_alloc_disk(&mq->tag_set, &lim, mq); 429 memset(&mq->tag_set, 0, sizeof(mq->tag_set)); 430 mq->tag_set.ops = &mmc_mq_ops; 436 mq->tag_set.queue_depth = 439 mq->tag_set.queue_depth = MMC_QUEUE_DEPTH; 440 mq->tag_set.numa_node = NUMA_NO_NODE; 441 mq->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_BLOCKING; 442 mq->tag_set.nr_hw_queues = 1; 443 mq->tag_set [all...] |
H A D | queue.h | 77 struct blk_mq_tag_set tag_set; member in struct:mmc_queue
|
/linux-master/block/ |
H A D | bsg-lib.c | 22 struct blk_mq_tag_set tag_set; member in struct:bsg_set 279 container_of(q->tag_set, struct bsg_set, tag_set); 324 container_of(q->tag_set, struct bsg_set, tag_set); 329 blk_mq_free_tag_set(&bset->tag_set); 338 container_of(rq->q->tag_set, struct bsg_set, tag_set); 376 set = &bset->tag_set;
|
H A D | blk-mq-sched.c | 384 if (blk_mq_is_shared_tags(q->tag_set->flags)) { 389 hctx->sched_tags = blk_mq_alloc_map_and_rqs(q->tag_set, hctx_idx, 423 struct blk_mq_tag_set *set = queue->tag_set; 443 unsigned int flags = q->tag_set->flags; 454 q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth, 512 if (blk_mq_is_shared_tags(q->tag_set->flags)) { 513 blk_mq_free_rqs(q->tag_set, q->sched_shared_tags, 518 blk_mq_free_rqs(q->tag_set,
|
H A D | elevator.c | 561 (q->tag_set && (q->tag_set->flags & BLK_MQ_F_NO_SCHED))) 572 if (q->tag_set && q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT) 576 !blk_mq_is_shared_tags(q->tag_set->flags)) 708 q->nr_requests = q->tag_set->queue_depth;
|
H A D | blk-mq.h | 82 return xa_load(&q->hctx_table, q->tag_set->map[type].mq_map[cpu]); 452 if ((q)->tag_set->flags & BLK_MQ_F_BLOCKING) { \ 453 struct blk_mq_tag_set *__tag_set = (q)->tag_set; \
|
H A D | blk-mq-tag.c | 271 struct blk_mq_tag_set *set = q->tag_set; 492 * Note: if @q->tag_set is shared with other request queues then @fn will be 507 if (blk_mq_is_shared_tags(q->tag_set->flags)) { 508 struct blk_mq_tags *tags = q->tag_set->shared_tags; 615 struct blk_mq_tag_set *set = hctx->queue->tag_set; 663 q->nr_requests - q->tag_set->reserved_tags);
|
/linux-master/drivers/mtd/ubi/ |
H A D | block.c | 85 struct blk_mq_tag_set tag_set; member in struct:ubiblock 383 dev->tag_set.ops = &ubiblock_mq_ops; 384 dev->tag_set.queue_depth = 64; 385 dev->tag_set.numa_node = NUMA_NO_NODE; 386 dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_BLOCKING; 387 dev->tag_set.cmd_size = sizeof(struct ubiblock_pdu); 388 dev->tag_set.driver_data = dev; 389 dev->tag_set.nr_hw_queues = 1; 391 ret = blk_mq_alloc_tag_set(&dev->tag_set); 399 gd = blk_mq_alloc_disk(&dev->tag_set, [all...] |
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 72 static struct blk_mq_tag_set tag_set; variable in typeref:struct:blk_mq_tag_set 390 set = dev->nullb->tag_set; 1678 if (nullb->tag_set == &nullb->__tag_set) 1679 blk_mq_free_tag_set(nullb->tag_set); 1744 if (tag_set.ops) 1747 tag_set.nr_hw_queues = g_submit_queues; 1748 tag_set.queue_depth = g_hw_queue_depth; 1749 tag_set.numa_node = g_home_node; 1750 tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 1752 tag_set [all...] |
/linux-master/include/linux/mtd/ |
H A D | blktrans.h | 34 struct blk_mq_tag_set *tag_set; member in struct:mtd_blktrans_dev
|
/linux-master/drivers/mtd/ |
H A D | mtd_blkdevs.c | 33 blk_mq_free_tag_set(dev->tag_set); 34 kfree(dev->tag_set); 327 new->tag_set = kzalloc(sizeof(*new->tag_set), GFP_KERNEL); 328 if (!new->tag_set) 331 ret = blk_mq_alloc_sq_tag_set(new->tag_set, &mtd_mq_ops, 2, 341 gd = blk_mq_alloc_disk(new->tag_set, &lim, new); 401 blk_mq_free_tag_set(new->tag_set); 403 kfree(new->tag_set);
|
/linux-master/drivers/block/rnbd/ |
H A D | rnbd-clt.h | 89 struct blk_mq_tag_set tag_set; member in struct:rnbd_clt_session
|
H A D | rnbd-clt.c | 737 if (sess->tag_set.tags) 738 blk_mq_free_tag_set(&sess->tag_set); 1206 struct blk_mq_tag_set *tag_set = &sess->tag_set; local 1208 memset(tag_set, 0, sizeof(*tag_set)); 1209 tag_set->ops = &rnbd_mq_ops; 1210 tag_set->queue_depth = sess->queue_depth; 1211 tag_set->numa_node = NUMA_NO_NODE; 1212 tag_set [all...] |
/linux-master/arch/um/drivers/ |
H A D | ubd_kern.c | 161 struct blk_mq_tag_set tag_set; member in struct:ubd 799 blk_mq_free_tag_set(&ubd_dev->tag_set); 879 ubd_dev->tag_set.ops = &ubd_mq_ops; 880 ubd_dev->tag_set.queue_depth = 64; 881 ubd_dev->tag_set.numa_node = NUMA_NO_NODE; 882 ubd_dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; 883 ubd_dev->tag_set.driver_data = ubd_dev; 884 ubd_dev->tag_set.nr_hw_queues = 1; 886 err = blk_mq_alloc_tag_set(&ubd_dev->tag_set); 890 disk = blk_mq_alloc_disk(&ubd_dev->tag_set, [all...] |
/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 1960 struct Scsi_Host *shost = container_of(set, struct Scsi_Host, tag_set); 2056 struct blk_mq_tag_set *tag_set = &shost->tag_set; local 2065 memset(tag_set, 0, sizeof(*tag_set)); 2067 tag_set->ops = &scsi_mq_ops; 2069 tag_set->ops = &scsi_mq_ops_no_commit; 2070 tag_set->nr_hw_queues = shost->nr_hw_queues ? : 1; 2071 tag_set->nr_maps = shost->nr_maps ? : 1; 2072 tag_set [all...] |
/linux-master/drivers/nvme/target/ |
H A D | loop.c | 35 struct blk_mq_tag_set tag_set; member in struct:nvme_loop_ctrl 88 return queue->ctrl->tag_set.tags[queue_idx - 1]; 213 (set == &ctrl->tag_set) ? hctx_idx + 1 : 0); 458 blk_mq_update_nr_hw_queues(&ctrl->tag_set, 500 ret = nvme_alloc_io_tag_set(&ctrl->ctrl, &ctrl->tag_set,
|