Lines Matching refs:sectors

345 		r10_bio->devs[slot].addr + (r10_bio->sectors);
431 r10_bio->sectors,
532 r10_bio->sectors) &&
729 int sectors = r10_bio->sectors;
749 if (raid1_should_read_first(conf->mddev, this_sector, sectors))
764 r10_bio->devs[slot].addr + sectors >
771 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset)
775 if (is_badblock(rdev, dev_sector, sectors,
786 if (!do_balance && sectors > bad_sectors)
787 sectors = bad_sectors;
788 if (best_good_sectors > sectors)
789 best_good_sectors = sectors;
804 best_good_sectors = sectors;
1123 struct bio *bio, sector_t sectors)
1132 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) {
1142 sectors);
1188 if (!regular_request_wait(mddev, conf, bio, r10_bio->sectors))
1214 r10_bio->sectors = max_sectors;
1316 if (!r10_bio->sectors)
1320 r10_bio->sectors) < 0) {
1350 sector_t sectors;
1374 sectors = r10_bio->sectors;
1375 if (!regular_request_wait(mddev, conf, bio, sectors))
1380 bio->bi_iter.bi_sector + sectors > conf->reshape_progress)
1381 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe &&
1416 max_sectors = r10_bio->sectors;
1478 if (max_sectors < r10_bio->sectors)
1479 r10_bio->sectors = max_sectors;
1481 if (r10_bio->sectors < bio_sectors(bio)) {
1482 struct bio *split = bio_split(bio, r10_bio->sectors,
1495 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0);
1506 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors)
1514 r10_bio->sectors = sectors;
1693 r10_bio->sectors = 0;
1833 int sectors = bio_sectors(bio);
1851 sectors > chunk_sects
1855 sectors = chunk_sects -
1858 __make_request(mddev, bio, sectors);
2200 atomic_add(r10_bio->sectors,
2240 sector_t s = r10_bio->sectors;
2287 r10_bio->sectors)) {
2332 fbio->bi_iter.bi_size = r10_bio->sectors << 9;
2336 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9);
2358 int sectors = r10_bio->sectors;
2361 if (sectors < (len / 512))
2362 len = sectors * 512;
2367 sectors -= len/512;
2371 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches);
2428 md_done_sync(mddev, r10_bio->sectors, 1);
2435 * Recovery happens across physical sectors.
2456 int sectors = r10_bio->sectors;
2462 while (sectors) {
2463 int s = sectors;
2520 sectors -= s;
2568 int sectors, struct page *page, enum req_op op)
2570 if (rdev_has_badblock(rdev, sector, sectors) &&
2573 if (sync_page_io(rdev, sector, sectors << 9, page, op, false))
2583 if (!rdev_set_badblocks(rdev, sector, sectors, 0))
2599 int sectors = r10_bio->sectors, slot = r10_bio->read_slot;
2618 while(sectors) {
2619 int s = sectors;
2692 pr_notice("md/raid10:%s: read correction write failed (%d sectors at %llu on %pg)\n",
2724 pr_notice("md/raid10:%s: unable to read back corrected sectors (%d sectors at %llu on %pg)\n",
2735 pr_info("md/raid10:%s: read error corrected (%d sectors at %llu on %pg)\n",
2747 sectors -= s;
2771 int sectors;
2772 int sect_to_write = r10_bio->sectors;
2781 sectors = ((r10_bio->sector + block_sectors)
2788 if (sectors > sect_to_write)
2789 sectors = sect_to_write;
2790 /* Write at 'sector' for 'sectors' */
2793 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors);
2802 sectors, 0)
2806 sect_to_write -= sectors;
2807 sector += sectors;
2808 sectors = block_sectors;
2874 r10_bio->sectors, 0);
2879 r10_bio->sectors, 0))
2891 r10_bio->sectors, 0);
2896 r10_bio->sectors, 0))
2911 r10_bio->sectors, 0);
2928 r10_bio->sectors, 0);
3575 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1;
3678 r10_bio->sectors = nr_sectors;
3732 r10_bio->sectors = nr_sectors;
3763 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks)
3771 if (!sectors)
3772 sectors = conf->dev_sectors;
3774 size = sectors >> conf->geo.chunk_shift;
3784 /* Calculate the number of sectors-per-device that will
4180 static int raid10_resize(struct mddev *mddev, sector_t sectors)
4184 * number of sectors used to be an appropriate multiple
4204 size = raid10_size(mddev, sectors, 0);
4214 if (sectors > mddev->dev_sectors &&
4219 calc_sectors(conf, sectors);
4253 rdev->sectors = size;
4651 /* If restarting in the middle, skip the initial sectors */
4750 r10_bio->sectors = last - sector_nr + 1;
4847 r10_bio->sectors = nr_sectors;
4850 md_sync_acct_bio(read_bio, r10_bio->sectors);
4887 md_done_sync(mddev, r10_bio->sectors, 0);
4910 md_sync_acct_bio(b, r10_bio->sectors);
4952 int sectors = r10_bio->sectors;
4971 while (sectors) {
4972 int s = sectors;
5012 sectors -= s;
5046 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1);