Searched refs:bdev (Results 1 - 25 of 414) sorted by last modified time

1234567891011>>

/linux-master/fs/ntfs3/
H A Dsuper.c1158 struct block_device *bdev = sb->s_bdev; local
1194 if (bdev_max_discard_sectors(bdev) && bdev_discard_granularity(bdev)) {
1195 sbi->discard_granularity = bdev_discard_granularity(bdev);
1201 err = ntfs_init_from_boot(sb, bdev_logical_block_size(bdev),
1202 bdev_nr_bytes(bdev), &boot2);
1638 struct block_device *bdev = sb->s_bdev; local
1653 clean_bdev_aliases(bdev, devblock++, 1);
1655 sync_blockdev(bdev);
/linux-master/drivers/md/
H A Ddm.c405 static int dm_blk_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument
407 struct mapped_device *md = bdev->bd_disk->private_data;
413 struct block_device **bdev)
436 r = ti->type->prepare_ioctl(ti, bdev);
451 static int dm_blk_ioctl(struct block_device *bdev, blk_mode_t mode, argument
454 struct mapped_device *md = bdev->bd_disk->private_data;
457 r = dm_prepare_ioctl(md, &srcu_idx, &bdev);
464 * subset of the parent bdev; require extra privileges.
475 if (!bdev->bd_disk->fops->ioctl)
478 r = bdev
412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument
730 struct block_device *bdev; local
3188 dm_call_pr(struct block_device *bdev, iterate_devices_callout_fn fn, struct dm_pr *pr) argument
3249 dm_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, u32 flags) argument
3302 dm_pr_reserve(struct block_device *bdev, u64 key, enum pr_type type, u32 flags) argument
3345 dm_pr_release(struct block_device *bdev, u64 key, enum pr_type type) argument
3380 dm_pr_preempt(struct block_device *bdev, u64 old_key, u64 new_key, enum pr_type type, bool abort) argument
3398 dm_pr_clear(struct block_device *bdev, u64 key) argument
3436 dm_pr_read_keys(struct block_device *bdev, struct pr_keys *keys) argument
3468 dm_pr_read_reservation(struct block_device *bdev, struct pr_held_reservation *rsv) argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_object.c81 struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev);
127 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev);
441 WARN_ON(amdgpu_ttm_adev((*bo)->tbo.bdev)->in_suspend);
470 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT);
472 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM);
601 bo->tbo.bdev = &adev->mman.bdev;
615 r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, bp->type,
728 struct amdgpu_device *adev = amdgpu_ttm_adev(vmbo->bo.tbo.bdev);
753 struct amdgpu_device *adev = amdgpu_ttm_adev(shadow->tbo.bdev);
[all...]
H A Damdgpu_amdkfd_gpuvm.c293 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
471 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
489 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
816 bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev);
870 struct amdgpu_device *bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev);
1606 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
2973 peer_vm->root.bo->tbo.bdev);
H A Damdgpu_vm.c388 spin_lock(&adev->mman.bdev.lru_lock);
390 spin_unlock(&adev->mman.bdev.lru_lock);
1163 bo_adev = amdgpu_ttm_adev(bo->tbo.bdev);
H A Damdgpu_ttm.h55 struct ttm_device bdev; member in struct:amdgpu_mman
H A Damdgpu_ttm.c68 static int amdgpu_ttm_backend_bind(struct ttm_device *bdev,
71 static void amdgpu_ttm_backend_unbind(struct ttm_device *bdev,
78 return ttm_range_man_init(&adev->mman.bdev, type,
93 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev);
356 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev);
474 r = amdgpu_ttm_backend_bind(bo->bdev, bo->ttm, new_mem);
480 adev = amdgpu_ttm_adev(bo->bdev);
500 amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm);
571 static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, argument
574 struct amdgpu_device *adev = amdgpu_ttm_adev(bdev);
768 amdgpu_ttm_tt_pin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument
805 amdgpu_ttm_tt_unpin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument
887 amdgpu_ttm_backend_bind(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_resource *bo_mem) argument
1022 amdgpu_ttm_backend_unbind(struct ttm_device *bdev, struct ttm_tt *ttm) argument
1050 amdgpu_ttm_backend_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument
1107 amdgpu_ttm_tt_populate(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) argument
1148 amdgpu_ttm_tt_unpopulate(struct ttm_device *bdev, struct ttm_tt *ttm) argument
[all...]
H A Damdgpu_object.h210 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
/linux-master/block/
H A Dbdev.c37 struct block_device bdev; member in struct:bdev_inode
48 return &BDEV_I(inode)->bdev;
58 static void bdev_write_inode(struct block_device *bdev) argument
60 struct inode *inode = bdev->bd_inode;
70 bdev, ret);
77 static void kill_bdev(struct block_device *bdev) argument
79 struct address_space *mapping = bdev->bd_inode->i_mapping;
89 void invalidate_bdev(struct block_device *bdev) argument
91 struct address_space *mapping = bdev->bd_inode->i_mapping;
102 * Drop all buffers & page cache for given bdev rang
105 truncate_bdev_range(struct block_device *bdev, blk_mode_t mode, loff_t lstart, loff_t lend) argument
134 set_init_blocksize(struct block_device *bdev) argument
147 set_blocksize(struct block_device *bdev, int size) argument
191 sync_blockdev_nowait(struct block_device *bdev) argument
203 sync_blockdev(struct block_device *bdev) argument
211 sync_blockdev_range(struct block_device *bdev, loff_t lstart, loff_t lend) argument
232 bdev_freeze(struct block_device *bdev) argument
268 bdev_thaw(struct block_device *bdev) argument
321 struct block_device *bdev = I_BDEV(inode); local
399 struct block_device *bdev; local
430 bdev_set_nr_sectors(struct block_device *bdev, sector_t sectors) argument
438 bdev_add(struct block_device *bdev, dev_t dev) argument
472 bd_may_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument
514 bd_prepare_to_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument
566 bd_finish_claiming(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument
597 bd_abort_claiming(struct block_device *bdev, void *holder) argument
605 bd_end_claim(struct block_device *bdev, void *holder) argument
640 blkdev_flush_mapping(struct block_device *bdev) argument
648 blkdev_put_whole(struct block_device *bdev) argument
656 blkdev_get_whole(struct block_device *bdev, blk_mode_t mode) argument
752 struct block_device *bdev; local
774 blkdev_put_no_open(struct block_device *bdev) argument
779 bdev_writes_blocked(struct block_device *bdev) argument
784 bdev_block_writes(struct block_device *bdev) argument
789 bdev_unblock_writes(struct block_device *bdev) argument
794 bdev_may_open(struct block_device *bdev, blk_mode_t mode) argument
806 bdev_claim_write_access(struct block_device *bdev, blk_mode_t mode) argument
825 struct block_device *bdev; local
858 bdev_open(struct block_device *bdev, blk_mode_t mode, void *holder, const struct blk_holder_ops *hops, struct file *bdev_file) argument
969 struct block_device *bdev; local
1027 struct block_device *bdev = file_bdev(bdev_file); local
1041 struct block_device *bdev = file_bdev(bdev_file); local
1097 struct block_device *bdev = file_bdev(bdev_file); local
1169 bdev_mark_dead(struct block_device *bdev, bool surprise) argument
1196 struct block_device *bdev; local
1250 struct block_device *bdev; local
[all...]
H A Dioctl.c16 static int blkpg_do_ioctl(struct block_device *bdev, argument
19 struct gendisk *disk = bdev->bd_disk;
27 if (bdev_is_partition(bdev))
39 if (!IS_ALIGNED(p.start | p.length, bdev_logical_block_size(bdev)))
62 static int blkpg_ioctl(struct block_device *bdev, argument
71 return blkpg_do_ioctl(bdev, udata, op);
82 static int compat_blkpg_ioctl(struct block_device *bdev, argument
91 return blkpg_do_ioctl(bdev, compat_ptr(udata), op);
95 static int blk_ioctl_discard(struct block_device *bdev, blk_mode_t mode, argument
100 struct inode *inode = bdev
134 blk_ioctl_secure_erase(struct block_device *bdev, blk_mode_t mode, void __user *argp) argument
165 blk_ioctl_zeroout(struct block_device *bdev, blk_mode_t mode, unsigned long arg) argument
254 blkdev_compat_ptr_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned cmd, unsigned long arg) argument
268 blkdev_pr_allowed(struct block_device *bdev, blk_mode_t mode) argument
283 blkdev_pr_register(struct block_device *bdev, blk_mode_t mode, struct pr_registration __user *arg) argument
301 blkdev_pr_reserve(struct block_device *bdev, blk_mode_t mode, struct pr_reservation __user *arg) argument
319 blkdev_pr_release(struct block_device *bdev, blk_mode_t mode, struct pr_reservation __user *arg) argument
337 blkdev_pr_preempt(struct block_device *bdev, blk_mode_t mode, struct pr_preempt __user *arg, bool abort) argument
355 blkdev_pr_clear(struct block_device *bdev, blk_mode_t mode, struct pr_clear __user *arg) argument
373 blkdev_flushbuf(struct block_device *bdev, unsigned cmd, unsigned long arg) argument
391 blkdev_roset(struct block_device *bdev, unsigned cmd, unsigned long arg) argument
410 blkdev_getgeo(struct block_device *bdev, struct hd_geometry __user *argp) argument
444 compat_hdio_getgeo(struct block_device *bdev, struct compat_hd_geometry __user *ugeo) argument
476 blkdev_bszset(struct block_device *bdev, blk_mode_t mode, int __user *argp) argument
505 blkdev_common_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg, void __user *argp) argument
597 struct block_device *bdev = I_BDEV(file->f_mapping->host); local
658 struct block_device *bdev = I_BDEV(file->f_mapping->host); local
[all...]
H A Dblk-iocost.c3227 disk = ctx.bdev->bd_disk;
3421 q = bdev_get_queue(ctx.bdev);
3429 ret = blk_iocost_init(ctx.bdev->bd_disk);
/linux-master/fs/btrfs/
H A Dscrub.c1765 !stripe->dev->bdev)) {
2423 if (map->stripes[i].dev->bdev == scrub_dev->bdev &&
2761 bio_init(&bio, dev->bdev, &bvec, 1, REQ_OP_READ);
H A Dinode.c7722 iomap->bdev = fs_info->fs_devices->latest_dev->bdev;
10932 sis->bdev = device->bdev;
/linux-master/include/linux/
H A Dblkdev.h330 int blkdev_report_zones(struct block_device *bdev, sector_t sector,
332 int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
622 unsigned int bdev_nr_zones(struct block_device *bdev);
657 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) argument
659 return bdev->bd_disk->queue->limits.max_open_zones;
662 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) argument
664 return bdev->bd_disk->queue->limits.max_active_zones;
668 static inline unsigned int bdev_nr_zones(struct block_device *bdev) argument
685 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) argument
690 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) argument
731 bdev_read_only(struct block_device *bdev) argument
743 get_start_sect(struct block_device *bdev) argument
748 bdev_nr_sectors(struct block_device *bdev) argument
753 bdev_nr_bytes(struct block_device *bdev) argument
807 bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk) argument
812 bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk) argument
845 bdev_get_queue(struct block_device *bdev) argument
1048 blkdev_issue_flush(struct block_device *bdev) argument
1098 bdev_is_partition(struct block_device *bdev) argument
1168 bdev_max_zone_append_sectors(struct block_device *bdev) argument
1173 bdev_max_segments(struct block_device *bdev) argument
1188 bdev_logical_block_size(struct block_device *bdev) argument
1198 bdev_physical_block_size(struct block_device *bdev) argument
1208 bdev_io_min(struct block_device *bdev) argument
1218 bdev_io_opt(struct block_device *bdev) argument
1230 bdev_zone_write_granularity(struct block_device *bdev) argument
1238 bdev_max_discard_sectors(struct block_device *bdev) argument
1243 bdev_discard_granularity(struct block_device *bdev) argument
1249 bdev_max_secure_erase_sectors(struct block_device *bdev) argument
1254 bdev_write_zeroes_sectors(struct block_device *bdev) argument
1264 bdev_nonrot(struct block_device *bdev) argument
1269 bdev_synchronous(struct block_device *bdev) argument
1275 bdev_stable_writes(struct block_device *bdev) argument
1281 bdev_write_cache(struct block_device *bdev) argument
1286 bdev_fua(struct block_device *bdev) argument
1291 bdev_nowait(struct block_device *bdev) argument
1296 bdev_is_zoned(struct block_device *bdev) argument
1301 bdev_zone_no(struct block_device *bdev, sector_t sec) argument
1312 bdev_op_is_zoned_write(struct block_device *bdev, enum req_op op) argument
1318 bdev_zone_sectors(struct block_device *bdev) argument
1327 bdev_offset_from_zone_start(struct block_device *bdev, sector_t sector) argument
1333 bdev_is_zone_start(struct block_device *bdev, sector_t sector) argument
1344 bdev_dma_alignment(struct block_device *bdev) argument
1349 bdev_iter_is_aligned(struct block_device *bdev, struct iov_iter *iter) argument
1369 block_size(struct block_device *bdev) argument
1552 invalidate_bdev(struct block_device *bdev) argument
1555 sync_blockdev(struct block_device *bdev) argument
1559 sync_blockdev_nowait(struct block_device *bdev) argument
[all...]
/linux-master/fs/bcachefs/
H A Dsuper.c187 if (ca->disk_sb.bdev && ca->disk_sb.bdev->bd_dev == dev) {
627 ca->disk_sb.bdev)
628 sysfs_remove_link(bdev_kobj(ca->disk_sb.bdev), "bcachefs");
1119 prt_bdevname(&buf, sb->bdev);
1121 prt_bdevname(&buf, fs->bdev);
1127 prt_bdevname(&buf, fs->bdev);
1132 prt_bdevname(&buf, sb->bdev);
1155 prt_bdevname(&buf, sb->bdev);
1157 prt_bdevname(&buf, fs->bdev);
[all...]
H A Djournal_io.c1008 bio_init(bio, ca->disk_sb.bdev, bio->bi_inline_vecs, nr_bvecs, REQ_OP_READ);
1747 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_WRITE|REQ_SYNC|REQ_META);
1792 bio_reset(bio, ca->disk_sb.bdev,
H A Dfs.c1967 struct block_device *bdev = ca->disk_sb.bdev; local
1970 sb->s_bdev = bdev;
1971 sb->s_dev = bdev->bd_dev;
H A Dbtree_node_scan.c123 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_READ);
H A Dbtree_io.c1302 bio_set_dev(bio, ca->disk_sb.bdev);
1622 bio_set_dev(&rb->bio, ca->disk_sb.bdev);
1704 bio_set_dev(bio, ca->disk_sb.bdev);
H A Dbackpointers.c473 bio = bio_alloc(ca->disk_sb.bdev, buf_pages(data_buf, bytes), REQ_OP_READ, GFP_KERNEL);
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_ttm_buffer.c273 static int vmw_ttm_bind(struct ttm_device *bdev, argument
320 static void vmw_ttm_unbind(struct ttm_device *bdev, argument
348 static void vmw_ttm_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument
362 static int vmw_ttm_populate(struct ttm_device *bdev, argument
375 return ttm_pool_alloc(&bdev->pool, ttm, ctx);
378 static void vmw_ttm_unpopulate(struct ttm_device *bdev, argument
388 vmw_ttm_unbind(bdev, ttm);
397 ttm_pool_free(&bdev->pool, ttm);
411 vmw_be->dev_priv = vmw_priv_from_ttm(bo->bdev);
438 static int vmw_ttm_io_mem_reserve(struct ttm_device *bdev, struc argument
[all...]
H A Dvmwgfx_drv.h502 struct ttm_device bdev; member in struct:vmw_private
668 static inline struct vmw_private *vmw_priv_from_ttm(struct ttm_device *bdev) argument
670 return container_of(bdev, struct vmw_private, bdev);
H A Dvmwgfx_drv.c556 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB);
557 ttm_resource_manager_evict_all(&dev_priv->bdev, man);
592 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB);
593 ttm_resource_manager_evict_all(&dev_priv->bdev, man);
712 ret = ttm_range_man_init(&dev_priv->bdev, TTM_PL_VRAM, false,
714 ttm_resource_manager_set_used(ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM), false);
720 ttm_range_man_fini(&dev_priv->bdev, TTM_PL_VRAM);
1050 ret = ttm_device_init(&dev_priv->bdev, &vmw_bo_driver,
1167 ttm_device_fini(&dev_priv->bdev);
1212 ttm_device_fini(&dev_priv->bdev);
[all...]
H A Dvmwgfx_bo.c383 struct ttm_device *bdev = &dev_priv->bdev; local
397 ret = ttm_bo_init_reserved(bdev, &vmw_bo->tbo, params->bo_type,
637 struct ttm_device *bdev = bo->bdev; local
638 struct vmw_private *dev_priv = vmw_priv_from_ttm(bdev);
820 struct ttm_device *bdev = bo->tbo.bdev; local
821 struct vmw_private *vmw = vmw_priv_from_ttm(bdev);
846 struct ttm_device *bdev local
[all...]
H A Dvmwgfx_blit.c471 ret = dst->bdev->funcs->ttm_tt_populate(dst->bdev, dst->ttm, &ctx);
477 ret = src->bdev->funcs->ttm_tt_populate(src->bdev, src->ttm, &ctx);

Completed in 642 milliseconds

1234567891011>>