Lines Matching defs:limits

29  * blk_set_stacking_limits - set default limits for stacking devices
32 * Prepare queue limits for applying limits from underlying devices using
45 /* Inherit limits from component devices */
101 * Check that the limits in lim are valid, initialize defaults for unset
187 * might need to move out of stacking limits since we have immutable
209 * to set the limits. Once the transition is finished the check
226 * Set the default limits for a newly allocated queue. @lim contains the
227 * initial limits set by the driver, which could be no limit in which case
242 * queue_limits_commit_update - commit an atomic update of queue limits
244 * @lim: limits to apply
246 * Apply the limits in @lim that were obtained from queue_limits_start_update()
258 q->limits = *lim;
268 * queue_limits_set - apply queue limits to queue
270 * @lim: limits to apply
272 * Apply the limits in @lim that were freshly initialized to @q.
273 * To update existing limits use queue_limits_start_update() and
299 q->limits.chunk_sectors = chunk_sectors;
311 struct queue_limits *lim = &q->limits;
327 q->limits.max_secure_erase_sectors = max_sectors;
340 q->limits.max_write_zeroes_sectors = max_write_zeroes_sectors;
363 max_sectors = min(q->limits.max_hw_sectors,
365 max_sectors = min(q->limits.chunk_sectors, max_sectors);
375 q->limits.max_zone_append_sectors = max_sectors;
391 struct queue_limits *limits = &q->limits;
393 limits->logical_block_size = size;
395 if (limits->discard_granularity < limits->logical_block_size)
396 limits->discard_granularity = limits->logical_block_size;
398 if (limits->physical_block_size < size)
399 limits->physical_block_size = size;
401 if (limits->io_min < limits->physical_block_size)
402 limits->io_min = limits->physical_block_size;
404 limits->max_hw_sectors =
405 round_down(limits->max_hw_sectors, size >> SECTOR_SHIFT);
406 limits->max_sectors =
407 round_down(limits->max_sectors, size >> SECTOR_SHIFT);
423 q->limits.physical_block_size = size;
425 if (q->limits.physical_block_size < q->limits.logical_block_size)
426 q->limits.physical_block_size = q->limits.logical_block_size;
428 if (q->limits.discard_granularity < q->limits.physical_block_size)
429 q->limits.discard_granularity = q->limits.physical_block_size;
431 if (q->limits.io_min < q->limits.physical_block_size)
432 q->limits.io_min = q->limits.physical_block_size;
451 q->limits.zone_write_granularity = size;
453 if (q->limits.zone_write_granularity < q->limits.logical_block_size)
454 q->limits.zone_write_granularity = q->limits.logical_block_size;
471 q->limits.alignment_offset =
472 offset & (q->limits.physical_block_size - 1);
473 q->limits.misaligned = 0;
479 blk_apply_bdi_limits(disk->bdi, &disk->queue->limits);
485 * @limits: the queue limits
494 void blk_limits_io_min(struct queue_limits *limits, unsigned int min)
496 limits->io_min = min;
498 if (limits->io_min < limits->logical_block_size)
499 limits->io_min = limits->logical_block_size;
501 if (limits->io_min < limits->physical_block_size)
502 limits->io_min = limits->physical_block_size;
522 blk_limits_io_min(&q->limits, min);
528 * @limits: the queue limits
539 void blk_limits_io_opt(struct queue_limits *limits, unsigned int opt)
541 limits->io_opt = opt;
589 * @t: the stacking driver limits (top device)
590 * @b: the underlying queue limits (bottom, component device)
754 * @t: the stacking driver limits (top device)
770 if (blk_stack_limits(t, &bdev_get_queue(bdev)->limits,
845 q->limits.zoned = true;
854 if (q->limits.misaligned)
857 return queue_limit_alignment_offset(&q->limits,
859 return q->limits.alignment_offset;
868 return queue_limit_discard_alignment(&q->limits,
870 return q->limits.discard_alignment;