/linux-master/fs/ntfs3/ |
H A D | super.c | 1158 struct block_device *bdev = sb->s_bdev; local 1194 if (bdev_max_discard_sectors(bdev) && bdev_discard_granularity(bdev)) { 1195 sbi->discard_granularity = bdev_discard_granularity(bdev); 1201 err = ntfs_init_from_boot(sb, bdev_logical_block_size(bdev), 1202 bdev_nr_bytes(bdev), &boot2); 1638 struct block_device *bdev = sb->s_bdev; local 1653 clean_bdev_aliases(bdev, devblock++, 1); 1655 sync_blockdev(bdev);
|
/linux-master/drivers/md/ |
H A D | dm.c | 405 static int dm_blk_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 407 struct mapped_device *md = bdev->bd_disk->private_data; 413 struct block_device **bdev) 436 r = ti->type->prepare_ioctl(ti, bdev); 451 static int dm_blk_ioctl(struct block_device *bdev, blk_mode_t mode, argument 454 struct mapped_device *md = bdev->bd_disk->private_data; 457 r = dm_prepare_ioctl(md, &srcu_idx, &bdev); 464 * subset of the parent bdev; require extra privileges. 475 if (!bdev->bd_disk->fops->ioctl) 478 r = bdev 412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument 730 struct block_device *bdev; local 3188 dm_call_pr(struct block_device *bdev, iterate_devices_callout_fn fn, struct dm_pr *pr) argument 3249 dm_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, u32 flags) argument 3302 dm_pr_reserve(struct block_device *bdev, u64 key, enum pr_type type, u32 flags) argument 3345 dm_pr_release(struct block_device *bdev, u64 key, enum pr_type type) argument 3380 dm_pr_preempt(struct block_device *bdev, u64 old_key, u64 new_key, enum pr_type type, bool abort) argument 3398 dm_pr_clear(struct block_device *bdev, u64 key) argument 3436 dm_pr_read_keys(struct block_device *bdev, struct pr_keys *keys) argument 3468 dm_pr_read_reservation(struct block_device *bdev, struct pr_held_reservation *rsv) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.c | 81 struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); 127 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); 441 WARN_ON(amdgpu_ttm_adev((*bo)->tbo.bdev)->in_suspend); 470 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); 472 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM); 601 bo->tbo.bdev = &adev->mman.bdev; 615 r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, bp->type, 728 struct amdgpu_device *adev = amdgpu_ttm_adev(vmbo->bo.tbo.bdev); 753 struct amdgpu_device *adev = amdgpu_ttm_adev(shadow->tbo.bdev); [all...] |
H A D | amdgpu_amdkfd_gpuvm.c | 293 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 471 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); 489 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); 816 bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev); 870 struct amdgpu_device *bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev); 1606 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); 2973 peer_vm->root.bo->tbo.bdev);
|
H A D | amdgpu_vm.c | 388 spin_lock(&adev->mman.bdev.lru_lock); 390 spin_unlock(&adev->mman.bdev.lru_lock); 1163 bo_adev = amdgpu_ttm_adev(bo->tbo.bdev);
|
H A D | amdgpu_ttm.h | 55 struct ttm_device bdev; member in struct:amdgpu_mman
|
H A D | amdgpu_ttm.c | 68 static int amdgpu_ttm_backend_bind(struct ttm_device *bdev, 71 static void amdgpu_ttm_backend_unbind(struct ttm_device *bdev, 78 return ttm_range_man_init(&adev->mman.bdev, type, 93 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev); 356 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev); 474 r = amdgpu_ttm_backend_bind(bo->bdev, bo->ttm, new_mem); 480 adev = amdgpu_ttm_adev(bo->bdev); 500 amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm); 571 static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, argument 574 struct amdgpu_device *adev = amdgpu_ttm_adev(bdev); 768 amdgpu_ttm_tt_pin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument 805 amdgpu_ttm_tt_unpin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument 887 amdgpu_ttm_backend_bind(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_resource *bo_mem) argument 1022 amdgpu_ttm_backend_unbind(struct ttm_device *bdev, struct ttm_tt *ttm) argument 1050 amdgpu_ttm_backend_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument 1107 amdgpu_ttm_tt_populate(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) argument 1148 amdgpu_ttm_tt_unpopulate(struct ttm_device *bdev, struct ttm_tt *ttm) argument [all...] |
H A D | amdgpu_object.h | 210 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
|
/linux-master/block/ |
H A D | bdev.c | 37 struct block_device bdev; member in struct:bdev_inode 48 return &BDEV_I(inode)->bdev; 58 static void bdev_write_inode(struct block_device *bdev) argument 60 struct inode *inode = bdev->bd_inode; 70 bdev, ret); 77 static void kill_bdev(struct block_device *bdev) argument 79 struct address_space *mapping = bdev->bd_inode->i_mapping; 89 void invalidate_bdev(struct block_device *bdev) argument 91 struct address_space *mapping = bdev->bd_inode->i_mapping; 102 * Drop all buffers & page cache for given bdev rang 105 truncate_bdev_range(struct block_device *bdev, blk_mode_t mode, loff_t lstart, loff_t lend) argument 134 set_init_blocksize(struct block_device *bdev) argument 147 set_blocksize(struct block_device *bdev, int size) argument 191 sync_blockdev_nowait(struct block_device *bdev) argument 203 sync_blockdev(struct block_device *bdev) argument 211 sync_blockdev_range(struct block_device *bdev, loff_t lstart, loff_t lend) argument 232 bdev_freeze(struct block_device *bdev) argument 268 bdev_thaw(struct block_device *bdev) argument 321 struct block_device *bdev = I_BDEV(inode); local 399 struct block_device *bdev; local 430 bdev_set_nr_sectors(struct block_device *bdev, sector_t sectors) argument 438 bdev_add(struct block_device *bdev, dev_t dev) argument 472 bd_may_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 514 bd_prepare_to_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 566 bd_finish_claiming(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 597 bd_abort_claiming(struct block_device *bdev, void *holder) argument 605 bd_end_claim(struct block_device *bdev, void *holder) argument 640 blkdev_flush_mapping(struct block_device *bdev) argument 648 blkdev_put_whole(struct block_device *bdev) argument 656 blkdev_get_whole(struct block_device *bdev, blk_mode_t mode) argument 752 struct block_device *bdev; local 774 blkdev_put_no_open(struct block_device *bdev) argument 779 bdev_writes_blocked(struct block_device *bdev) argument 784 bdev_block_writes(struct block_device *bdev) argument 789 bdev_unblock_writes(struct block_device *bdev) argument 794 bdev_may_open(struct block_device *bdev, blk_mode_t mode) argument 806 bdev_claim_write_access(struct block_device *bdev, blk_mode_t mode) argument 825 struct block_device *bdev; local 858 bdev_open(struct block_device *bdev, blk_mode_t mode, void *holder, const struct blk_holder_ops *hops, struct file *bdev_file) argument 969 struct block_device *bdev; local 1027 struct block_device *bdev = file_bdev(bdev_file); local 1041 struct block_device *bdev = file_bdev(bdev_file); local 1097 struct block_device *bdev = file_bdev(bdev_file); local 1169 bdev_mark_dead(struct block_device *bdev, bool surprise) argument 1196 struct block_device *bdev; local 1250 struct block_device *bdev; local [all...] |
H A D | ioctl.c | 16 static int blkpg_do_ioctl(struct block_device *bdev, argument 19 struct gendisk *disk = bdev->bd_disk; 27 if (bdev_is_partition(bdev)) 39 if (!IS_ALIGNED(p.start | p.length, bdev_logical_block_size(bdev))) 62 static int blkpg_ioctl(struct block_device *bdev, argument 71 return blkpg_do_ioctl(bdev, udata, op); 82 static int compat_blkpg_ioctl(struct block_device *bdev, argument 91 return blkpg_do_ioctl(bdev, compat_ptr(udata), op); 95 static int blk_ioctl_discard(struct block_device *bdev, blk_mode_t mode, argument 100 struct inode *inode = bdev 134 blk_ioctl_secure_erase(struct block_device *bdev, blk_mode_t mode, void __user *argp) argument 165 blk_ioctl_zeroout(struct block_device *bdev, blk_mode_t mode, unsigned long arg) argument 254 blkdev_compat_ptr_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned cmd, unsigned long arg) argument 268 blkdev_pr_allowed(struct block_device *bdev, blk_mode_t mode) argument 283 blkdev_pr_register(struct block_device *bdev, blk_mode_t mode, struct pr_registration __user *arg) argument 301 blkdev_pr_reserve(struct block_device *bdev, blk_mode_t mode, struct pr_reservation __user *arg) argument 319 blkdev_pr_release(struct block_device *bdev, blk_mode_t mode, struct pr_reservation __user *arg) argument 337 blkdev_pr_preempt(struct block_device *bdev, blk_mode_t mode, struct pr_preempt __user *arg, bool abort) argument 355 blkdev_pr_clear(struct block_device *bdev, blk_mode_t mode, struct pr_clear __user *arg) argument 373 blkdev_flushbuf(struct block_device *bdev, unsigned cmd, unsigned long arg) argument 391 blkdev_roset(struct block_device *bdev, unsigned cmd, unsigned long arg) argument 410 blkdev_getgeo(struct block_device *bdev, struct hd_geometry __user *argp) argument 444 compat_hdio_getgeo(struct block_device *bdev, struct compat_hd_geometry __user *ugeo) argument 476 blkdev_bszset(struct block_device *bdev, blk_mode_t mode, int __user *argp) argument 505 blkdev_common_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg, void __user *argp) argument 597 struct block_device *bdev = I_BDEV(file->f_mapping->host); local 658 struct block_device *bdev = I_BDEV(file->f_mapping->host); local [all...] |
H A D | blk-iocost.c | 3227 disk = ctx.bdev->bd_disk; 3421 q = bdev_get_queue(ctx.bdev); 3429 ret = blk_iocost_init(ctx.bdev->bd_disk);
|
/linux-master/fs/btrfs/ |
H A D | scrub.c | 1765 !stripe->dev->bdev)) { 2423 if (map->stripes[i].dev->bdev == scrub_dev->bdev && 2761 bio_init(&bio, dev->bdev, &bvec, 1, REQ_OP_READ);
|
H A D | inode.c | 7722 iomap->bdev = fs_info->fs_devices->latest_dev->bdev; 10932 sis->bdev = device->bdev;
|
/linux-master/include/linux/ |
H A D | blkdev.h | 330 int blkdev_report_zones(struct block_device *bdev, sector_t sector, 332 int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op, 622 unsigned int bdev_nr_zones(struct block_device *bdev); 657 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) argument 659 return bdev->bd_disk->queue->limits.max_open_zones; 662 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) argument 664 return bdev->bd_disk->queue->limits.max_active_zones; 668 static inline unsigned int bdev_nr_zones(struct block_device *bdev) argument 685 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) argument 690 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) argument 731 bdev_read_only(struct block_device *bdev) argument 743 get_start_sect(struct block_device *bdev) argument 748 bdev_nr_sectors(struct block_device *bdev) argument 753 bdev_nr_bytes(struct block_device *bdev) argument 807 bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk) argument 812 bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk) argument 845 bdev_get_queue(struct block_device *bdev) argument 1048 blkdev_issue_flush(struct block_device *bdev) argument 1098 bdev_is_partition(struct block_device *bdev) argument 1168 bdev_max_zone_append_sectors(struct block_device *bdev) argument 1173 bdev_max_segments(struct block_device *bdev) argument 1188 bdev_logical_block_size(struct block_device *bdev) argument 1198 bdev_physical_block_size(struct block_device *bdev) argument 1208 bdev_io_min(struct block_device *bdev) argument 1218 bdev_io_opt(struct block_device *bdev) argument 1230 bdev_zone_write_granularity(struct block_device *bdev) argument 1238 bdev_max_discard_sectors(struct block_device *bdev) argument 1243 bdev_discard_granularity(struct block_device *bdev) argument 1249 bdev_max_secure_erase_sectors(struct block_device *bdev) argument 1254 bdev_write_zeroes_sectors(struct block_device *bdev) argument 1264 bdev_nonrot(struct block_device *bdev) argument 1269 bdev_synchronous(struct block_device *bdev) argument 1275 bdev_stable_writes(struct block_device *bdev) argument 1281 bdev_write_cache(struct block_device *bdev) argument 1286 bdev_fua(struct block_device *bdev) argument 1291 bdev_nowait(struct block_device *bdev) argument 1296 bdev_is_zoned(struct block_device *bdev) argument 1301 bdev_zone_no(struct block_device *bdev, sector_t sec) argument 1312 bdev_op_is_zoned_write(struct block_device *bdev, enum req_op op) argument 1318 bdev_zone_sectors(struct block_device *bdev) argument 1327 bdev_offset_from_zone_start(struct block_device *bdev, sector_t sector) argument 1333 bdev_is_zone_start(struct block_device *bdev, sector_t sector) argument 1344 bdev_dma_alignment(struct block_device *bdev) argument 1349 bdev_iter_is_aligned(struct block_device *bdev, struct iov_iter *iter) argument 1369 block_size(struct block_device *bdev) argument 1552 invalidate_bdev(struct block_device *bdev) argument 1555 sync_blockdev(struct block_device *bdev) argument 1559 sync_blockdev_nowait(struct block_device *bdev) argument [all...] |
/linux-master/fs/bcachefs/ |
H A D | super.c | 187 if (ca->disk_sb.bdev && ca->disk_sb.bdev->bd_dev == dev) { 627 ca->disk_sb.bdev) 628 sysfs_remove_link(bdev_kobj(ca->disk_sb.bdev), "bcachefs"); 1119 prt_bdevname(&buf, sb->bdev); 1121 prt_bdevname(&buf, fs->bdev); 1127 prt_bdevname(&buf, fs->bdev); 1132 prt_bdevname(&buf, sb->bdev); 1155 prt_bdevname(&buf, sb->bdev); 1157 prt_bdevname(&buf, fs->bdev); [all...] |
H A D | journal_io.c | 1008 bio_init(bio, ca->disk_sb.bdev, bio->bi_inline_vecs, nr_bvecs, REQ_OP_READ); 1747 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_WRITE|REQ_SYNC|REQ_META); 1792 bio_reset(bio, ca->disk_sb.bdev,
|
H A D | fs.c | 1967 struct block_device *bdev = ca->disk_sb.bdev; local 1970 sb->s_bdev = bdev; 1971 sb->s_dev = bdev->bd_dev;
|
H A D | btree_node_scan.c | 123 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_READ);
|
H A D | btree_io.c | 1302 bio_set_dev(bio, ca->disk_sb.bdev); 1622 bio_set_dev(&rb->bio, ca->disk_sb.bdev); 1704 bio_set_dev(bio, ca->disk_sb.bdev);
|
H A D | backpointers.c | 473 bio = bio_alloc(ca->disk_sb.bdev, buf_pages(data_buf, bytes), REQ_OP_READ, GFP_KERNEL);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_ttm_buffer.c | 273 static int vmw_ttm_bind(struct ttm_device *bdev, argument 320 static void vmw_ttm_unbind(struct ttm_device *bdev, argument 348 static void vmw_ttm_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument 362 static int vmw_ttm_populate(struct ttm_device *bdev, argument 375 return ttm_pool_alloc(&bdev->pool, ttm, ctx); 378 static void vmw_ttm_unpopulate(struct ttm_device *bdev, argument 388 vmw_ttm_unbind(bdev, ttm); 397 ttm_pool_free(&bdev->pool, ttm); 411 vmw_be->dev_priv = vmw_priv_from_ttm(bo->bdev); 438 static int vmw_ttm_io_mem_reserve(struct ttm_device *bdev, struc argument [all...] |
H A D | vmwgfx_drv.h | 502 struct ttm_device bdev; member in struct:vmw_private 668 static inline struct vmw_private *vmw_priv_from_ttm(struct ttm_device *bdev) argument 670 return container_of(bdev, struct vmw_private, bdev);
|
H A D | vmwgfx_drv.c | 556 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); 557 ttm_resource_manager_evict_all(&dev_priv->bdev, man); 592 man = ttm_manager_type(&dev_priv->bdev, VMW_PL_MOB); 593 ttm_resource_manager_evict_all(&dev_priv->bdev, man); 712 ret = ttm_range_man_init(&dev_priv->bdev, TTM_PL_VRAM, false, 714 ttm_resource_manager_set_used(ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM), false); 720 ttm_range_man_fini(&dev_priv->bdev, TTM_PL_VRAM); 1050 ret = ttm_device_init(&dev_priv->bdev, &vmw_bo_driver, 1167 ttm_device_fini(&dev_priv->bdev); 1212 ttm_device_fini(&dev_priv->bdev); [all...] |
H A D | vmwgfx_bo.c | 383 struct ttm_device *bdev = &dev_priv->bdev; local 397 ret = ttm_bo_init_reserved(bdev, &vmw_bo->tbo, params->bo_type, 637 struct ttm_device *bdev = bo->bdev; local 638 struct vmw_private *dev_priv = vmw_priv_from_ttm(bdev); 820 struct ttm_device *bdev = bo->tbo.bdev; local 821 struct vmw_private *vmw = vmw_priv_from_ttm(bdev); 846 struct ttm_device *bdev local [all...] |
H A D | vmwgfx_blit.c | 471 ret = dst->bdev->funcs->ttm_tt_populate(dst->bdev, dst->ttm, &ctx); 477 ret = src->bdev->funcs->ttm_tt_populate(src->bdev, src->ttm, &ctx);
|