Lines Matching refs:vdev

141  * Maximum number of parallelly executing I/Os per top-level vdev.
151 * Maximum number of parallelly executed bytes per leaf vdev. We attempt
152 * to strike a balance here between keeping the vdev queues full of I/Os
190 &zfs_top_maxinflight, 0, "Maximum I/Os per top-level vdev");
301 vdev_t *q_vd; /* top-level vdev that this queue represents */
400 /* we discard the vdev id, since we can deduce it from the queue */
446 * phase are done per top-level vdev and are handled separately.
736 * Purge all vdev caches and probe all devices. We do this here
2366 * that can add scan_io_t's to the vdev scan queues from
2854 * punches out one thread per top-level vdev, each of which processes
2855 * only that vdev's scan queue. We can parallelize the I/O here because
2856 * we know that each queue's io's only affect its own top-level vdev.
3031 * The indirect vdev can point to multiple
3047 * gang members reside on the same vdev.
3060 * Check if the top-level vdev must resilver this offset.
3360 * phase are done per top-level vdev and are handled separately.
3431 /* need to issue scrubbing IOs from per-vdev queues */
3594 vdev_t *vdev;
3597 vdev = vdev_lookup_top(spa, DVA_GET_VDEV(&dva));
3598 ASSERT(vdev != NULL);
3600 mutex_enter(&vdev->vdev_scan_io_queue_lock);
3601 if (vdev->vdev_scan_io_queue == NULL)
3602 vdev->vdev_scan_io_queue = scan_io_queue_create(vdev);
3604 scan_io_queue_insert(vdev->vdev_scan_io_queue, bp,
3606 mutex_exit(&vdev->vdev_scan_io_queue_lock);
3704 * caller to specify that they want per top level vdev IO rate limiting
3876 * a mirror vdev due to zpool attach/detach.
3915 vdev_t *vdev;
3922 vdev = vdev_lookup_top(spa, DVA_GET_VDEV(&bp->blk_dva[dva_i]));
3923 ASSERT(vdev != NULL);
3924 q_lock = &vdev->vdev_scan_io_queue_lock;
3925 queue = vdev->vdev_scan_io_queue;