Lines Matching defs:sector

13 static sector_t bio_discard_limit(struct block_device *bdev, sector_t sector)
19 sector += bdev->bd_start_sect;
22 round_up(sector, discard_granularity >> SECTOR_SHIFT);
28 if (granularity_aligned_sector != sector)
29 return granularity_aligned_sector - sector;
38 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
57 if ((sector | nr_sects) & bs_mask)
65 min(nr_sects, bio_discard_limit(bdev, sector));
68 bio->bi_iter.bi_sector = sector;
70 sector += req_sects;
90 * @sector: start sector
97 int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio);
119 sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
138 bio->bi_iter.bi_sector = sector;
144 sector += len;
166 sector_t sector, sector_t nr_sects, gfp_t gfp_mask,
179 bio->bi_iter.bi_sector = sector;
185 sector += bi_size >> 9;
199 * @sector: start sector
215 int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
223 if ((sector | nr_sects) & bs_mask)
226 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp_mask,
231 return __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask,
239 * @sector: start sector
249 int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
259 if ((sector | nr_sects) & bs_mask)
266 ret = __blkdev_issue_write_zeroes(bdev, sector, nr_sects,
269 ret = __blkdev_issue_zero_pages(bdev, sector, nr_sects,
300 int blkdev_issue_secure_erase(struct block_device *bdev, sector_t sector,
316 if ((sector | nr_sects) & bs_mask)
326 bio->bi_iter.bi_sector = sector;
329 sector += len;