/linux-master/drivers/md/ |
H A D | dm-zoned-reclaim.c | 79 ret = blkdev_issue_zeroout(dev->bdev, 165 src.bdev = src_zone->dev->bdev; 169 dst.bdev = dst_zone->dev->bdev;
|
H A D | dm-switch.c | 326 bio_set_dev(bio, sctx->path_list[path_nr].dmdev->bdev); 520 static int switch_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument 527 *bdev = sctx->path_list[path_nr].dmdev->bdev; 533 bdev_nr_sectors((*bdev)))
|
H A D | dm-flakey.c | 337 bio_set_dev(bio, fc->dev->bdev); 429 bio_init(clone, fc->dev->bdev, bio->bi_inline_vecs, nr_iovecs, bio->bi_opf); 641 static int flakey_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument 645 *bdev = fc->dev->bdev; 650 if (fc->start || ti->len != bdev_nr_sectors((*bdev))) 661 return dm_report_zones(fc->dev->bdev, fc->start,
|
H A D | md.c | 207 rdev->bdev->bd_disk->queue->nr_hw_queues != 1 && 522 /* sync bdev before setting device to readonly or stopping raid*/ 580 bi = bio_alloc_bioset(rdev->bdev, 0, 948 if (rdev->bdev->bd_dev == dev) 959 if (rdev->bdev->bd_dev == dev) 981 return MD_NEW_SIZE_SECTORS(bdev_nr_sectors(rdev->bdev)); 1051 bio = bio_alloc_bioset(rdev->meta_bdev ? rdev->meta_bdev : rdev->bdev, 1091 bio_init(&bio, rdev->bdev, &bvec, 1, opf); 1121 rdev->bdev); 1290 rdev->bdev); 7617 md_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 7702 md_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg) argument 7903 md_compat_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg) argument 7922 md_set_read_only(struct block_device *bdev, bool ro) argument 8759 struct block_device *bdev = (*bio)->bi_bdev; local [all...] |
H A D | dm-clone-target.c | 261 bio_set_dev(bio, clone->source_dev->bdev); 266 bio_set_dev(bio, clone->dest_dev->bdev); 823 from.bdev = clone->source_dev->bdev; 827 to.bdev = clone->dest_dev->bdev; 1152 r = blkdev_issue_flush(clone->dest_dev->bdev); 1491 format_dev_t(buf, clone->metadata_dev->bdev->bd_dev); 1494 format_dev_t(buf, clone->dest_dev->bdev->bd_dev); 1497 format_dev_t(buf, clone->source_dev->bdev [all...] |
H A D | dm-cache-metadata.c | 109 struct block_device *bdev; member in struct:dm_cache_metadata 339 sector_t bdev_size = bdev_nr_sectors(cmd->bdev); 454 if (bdev_read_only(cmd->bdev)) 539 cmd->bm = dm_block_manager_create(cmd->bdev, DM_CACHE_METADATA_BLOCK_SIZE << SECTOR_SHIFT, 754 static struct dm_cache_metadata *metadata_open(struct block_device *bdev, argument 772 cmd->bdev = bdev; 802 static struct dm_cache_metadata *lookup(struct block_device *bdev) argument 807 if (cmd->bdev == bdev) { 815 lookup_or_open(struct block_device *bdev, sector_t data_block_size, bool may_format_device, size_t policy_hint_size, unsigned int metadata_version) argument 860 dm_cache_metadata_open(struct block_device *bdev, sector_t data_block_size, bool may_format_device, size_t policy_hint_size, unsigned int metadata_version) argument [all...] |
H A D | dm-snap-persistent.c | 236 .bdev = dm_snap_cow(ps->store->snap)->bdev, 321 bdev) >> 9); 497 client = dm_bufio_client_create(dm_snap_cow(ps->store->snap)->bdev, 581 *total_sectors = get_dev_size(dm_snap_cow(store->snap)->bdev); 678 sector_t size = get_dev_size(dm_snap_cow(store->snap)->bdev);
|
H A D | dm-log.c | 309 .bdev = lc->header_location.bdev, 442 lc->header_location.bdev = lc->log_dev->bdev; 450 bdev_logical_block_size(lc->header_location.bdev)); 452 if (buf_size > bdev_nr_bytes(dev->bdev)) {
|
H A D | dm-dust.c | 229 bio_set_dev(bio, dd->dev->bdev); 418 sector_t size = bdev_nr_sectors(dd->dev->bdev); 537 static int dust_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument 542 *bdev = dev->bdev; 547 if (dd->start || ti->len != bdev_nr_sectors(dev->bdev))
|
H A D | dm-mpath.c | 514 struct block_device *bdev; local 538 bdev = pgpath->path.dev->bdev; 539 q = bdev_get_queue(bdev); 661 bio_set_dev(bio, pgpath->path.dev->bdev); 882 static int setup_scsi_dh(struct block_device *bdev, struct multipath *m, argument 885 struct request_queue *q = bdev_get_queue(bdev); 915 DMINFO("retaining handler on device %pg", bdev); 961 q = bdev_get_queue(p->path.dev->bdev); 965 r = setup_scsi_dh(p->path.dev->bdev, 2028 multipath_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument [all...] |
/linux-master/block/partitions/ |
H A D | ibm.c | 329 struct block_device *bdev = disk->part0; local 344 blocksize = bdev_logical_block_size(bdev); 347 nr_sectors = bdev_nr_sectors(bdev); 360 geo->start = get_start_sect(bdev); 361 if (disk->fops->getgeo(bdev, geo))
|
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_object.c | 142 r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, type, 403 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_PRIV); 404 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man); 411 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_VRAM); 412 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man);
|
H A D | qxl_release.c | 448 struct ttm_device *bdev; local 458 bdev = bo->bdev; 459 qdev = container_of(bdev, struct qxl_device, mman.bdev);
|
/linux-master/drivers/mtd/devices/ |
H A D | block2mtd.c | 266 struct block_device *bdev; local 287 bdev = file_bdev(bdev_file); 289 if (MAJOR(bdev->bd_dev) == MTD_BLOCK_MAJOR) { 294 if ((long)bdev->bd_inode->i_size % erase_size) { 312 dev->mtd.size = bdev->bd_inode->i_size & PAGE_MASK;
|
/linux-master/fs/btrfs/ |
H A D | dev-replace.c | 248 struct block_device *bdev; local 264 bdev = file_bdev(bdev_file); 266 if (!btrfs_check_device_zone_type(fs_info, bdev)) { 273 sync_blockdev(bdev); 276 if (device->bdev == bdev) { 285 if (bdev_nr_bytes(bdev) < btrfs_device_get_total_bytes(srcdev)) { 314 device->bdev = bdev; 319 set_blocksize(device->bdev, BTRFS_BDEV_BLOCKSIZ [all...] |
/linux-master/drivers/block/ |
H A D | loop.c | 528 static int loop_validate_file(struct file *file, struct block_device *bdev) argument 538 if (f->f_mapping->host->i_rdev == bdev->bd_dev) 561 static int loop_change_fd(struct loop_device *lo, struct block_device *bdev, argument 589 error = loop_validate_file(file, bdev); 993 struct block_device *bdev, 1017 error = bd_prepare_to_claim(bdev, loop_configure, NULL); 1030 error = loop_validate_file(file, bdev); 1073 lo->lo_device = bdev; 1118 bd_abort_claiming(bdev, loop_configure); 1126 bd_abort_claiming(bdev, loop_configur 992 loop_configure(struct loop_device *lo, blk_mode_t mode, struct block_device *bdev, const struct loop_config *config) argument 1523 lo_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg) argument 1684 lo_compat_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_bo.c | 129 struct xe_device *xe = ttm_to_xe_device(res->bo->bdev); 410 static int xe_ttm_io_mem_reserve(struct ttm_device *bdev, argument 413 struct xe_device *xe = ttm_to_xe_device(bdev); 526 struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev); 575 struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev); 617 struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev); 842 ret = ttm_tt_populate(bo->ttm.bdev, bo->ttm.ttm, &ctx); 899 ret = ttm_tt_populate(bo->ttm.bdev, bo->ttm.ttm, &ctx); 944 struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev); 955 spin_lock(&ttm_bo->bdev 2303 struct ttm_device *bdev = tbo->bdev; local [all...] |
/linux-master/fs/iomap/ |
H A D | direct-io.c | 59 return bio_alloc_bioset(iter->iomap.bdev, nr_vecs, opf, 61 return bio_alloc(iter->iomap.bdev, nr_vecs, opf, GFP_KERNEL); 291 if ((pos | length) & (bdev_logical_block_size(iomap->bdev) - 1) || 292 !bdev_iter_is_aligned(iomap->bdev, dio->submit.iter)) 319 (bdev_fua(iomap->bdev) || !bdev_write_cache(iomap->bdev)))
|
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 203 r = ttm_bo_init_validate(&rdev->mman.bdev, &bo->tbo, type, 351 struct ttm_device *bdev = &rdev->mman.bdev; local 362 man = ttm_manager_type(bdev, TTM_PL_VRAM); 365 return ttm_resource_manager_evict_all(bdev, man); 420 ttm_manager_type(&rdev->mman.bdev, TTM_PL_VRAM);
|
/linux-master/fs/nfsd/ |
H A D | blocklayout.c | 334 struct block_device *bdev = file->nf_file->f_path.mnt->mnt_sb->s_bdev; local 336 bdev->bd_disk->fops->pr_ops->pr_preempt(bdev, NFSD_MDS_PR_KEY,
|
/linux-master/fs/jbd2/ |
H A D | revoke.c | 331 struct block_device *bdev; local 344 bdev = journal->j_fs_dev; 348 bh = __find_get_block(bdev, blocknr, journal->j_blocksize); 358 bh2 = __find_get_block(bdev, blocknr, journal->j_blocksize);
|
/linux-master/drivers/message/fusion/ |
H A D | mptscsih.h | 126 extern int mptscsih_bios_param(struct scsi_device * sdev, struct block_device *bdev, sector_t capacity, int geom[]);
|
/linux-master/drivers/block/aoe/ |
H A D | aoeblk.c | 272 aoeblk_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 274 struct aoedev *d = bdev->bd_disk->private_data; 288 aoeblk_ioctl(struct block_device *bdev, blk_mode_t mode, uint cmd, ulong arg) argument 295 d = bdev->bd_disk->private_data;
|
/linux-master/fs/erofs/ |
H A D | data.c | 281 iomap->bdev = mdev.m_bdev; 385 struct block_device *bdev = inode->i_sb->s_bdev; local 388 if (bdev) 389 blksize_mask = bdev_logical_block_size(bdev) - 1;
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | volume.h | 133 struct block_device *bdev);
|