/linux-master/include/linux/ |
H A D | dm-io.h | 20 struct block_device *bdev; member in struct:dm_io_region
|
/linux-master/fs/bcachefs/ |
H A D | super_types.h | 8 struct block_device *bdev; member in struct:bch_sb_handle
|
/linux-master/drivers/net/wireless/broadcom/b43/ |
H A D | bus.h | 17 struct bcma_device *bdev; member in union:b43_bus_dev::__anon7168 71 return (dev->bdev->bus->hosttype == BCMA_HOSTTYPE_PCI);
|
/linux-master/fs/btrfs/ |
H A D | zoned.h | 33 * Number of zones, zone size and types of zones if bdev is a 64 int btrfs_sb_log_location_bdev(struct block_device *bdev, int mirror, int rw, 69 int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror); 145 static inline int btrfs_sb_log_location_bdev(struct block_device *bdev, argument 164 static inline int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror) argument 323 struct block_device *bdev) 330 if (!bdev_is_zoned(bdev)) 334 (bdev_zone_sectors(bdev) << SECTOR_SHIFT); 338 return !bdev_is_zoned(bdev); 322 btrfs_check_device_zone_type(const struct btrfs_fs_info *fs_info, struct block_device *bdev) argument
|
H A D | zoned.c | 84 static int sb_write_pointer(struct block_device *bdev, struct blk_zone *zones, argument 121 struct address_space *mapping = bdev->bd_inode->i_mapping; 180 struct block_device *bdev) 182 return (sector_t)zone_number << ilog2(bdev_zone_sectors(bdev)); 200 sector_t bdev_size = bdev_nr_sectors(device->bdev); 230 if (!bdev_is_zoned(device->bdev)) { 265 ret = blkdev_report_zones(device->bdev, pos >> SECTOR_SHIFT, *nr_zones, 346 if (!device->bdev) 362 struct block_device *bdev = device->bdev; local 179 zone_start_sector(u32 zone_number, struct block_device *bdev) argument 801 sb_log_location(struct block_device *bdev, struct blk_zone *zones, int rw, u64 *bytenr_ret) argument 862 btrfs_sb_log_location_bdev(struct block_device *bdev, int mirror, int rw, u64 *bytenr_ret) argument 1001 btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror) argument [all...] |
/linux-master/block/ |
H A D | early-lookup.c | 23 struct block_device *bdev = dev_to_bdev(dev); local 26 if (!bdev->bd_meta_info || 27 strncasecmp(cmp->uuid, bdev->bd_meta_info->uuid, cmp->len)) 104 struct block_device *bdev = dev_to_bdev(dev); local 107 if (!bdev->bd_meta_info || strcmp(label, bdev->bd_meta_info->volname))
|
H A D | blk-core.c | 745 struct block_device *bdev = bio->bi_bdev; local 746 struct request_queue *q = bdev_get_queue(bdev); 755 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) 764 if (bdev->bd_partno && unlikely(blk_partition_remap(bio))) 799 if (!bdev_max_discard_sectors(bdev)) 803 if (!bdev_max_secure_erase_sectors(bdev)) 900 struct block_device *bdev; local 904 bdev = READ_ONCE(bio->bi_bdev); 905 if (!bdev) 908 q = bdev_get_queue(bdev); 1000 bdev_start_io_acct(struct block_device *bdev, enum req_op op, unsigned long start_time) argument 1024 bdev_end_io_acct(struct block_device *bdev, enum req_op op, unsigned int sectors, unsigned long start_time) argument [all...] |
/linux-master/drivers/md/ |
H A D | dm-zoned-target.c | 128 clone = bio_alloc_clone(dev->bdev, bio, GFP_NOIO, &dmz->bio_set); 589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) { 612 disk = dmz_dev->bdev->bd_disk; 695 struct block_device *bdev; local 704 bdev = ddev->bdev; 705 if (!bdev_is_zoned(bdev)) { 731 dev->bdev = bdev; 734 dev->capacity = bdev_nr_sectors(bdev); 782 struct block_device *bdev = zoned_dev->bdev; local 799 struct block_device *bdev = zoned_dev->bdev; local 1019 dmz_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument [all...] |
H A D | dm-table.c | 222 if (dd->dm_dev->bdev->bd_dev == dev) 235 struct block_device *bdev = dev->bdev; local 236 sector_t dev_size = bdev_nr_sectors(bdev); 245 dm_device_name(ti->table->md), bdev, 256 if (bdev_is_zoned(bdev)) { 257 unsigned int zone_sectors = bdev_zone_sectors(bdev); 263 zone_sectors, bdev); 280 zone_sectors, bdev); 292 limits->logical_block_size, bdev); 409 struct block_device *bdev = dev->bdev; local 864 struct block_device *bdev = dev->bdev; local [all...] |
H A D | dm-snap.c | 316 struct block_device *bdev; member in struct:origin 383 static unsigned int origin_hash(struct block_device *bdev) argument 385 return bdev->bd_dev & ORIGIN_MASK; 395 if (bdev_equal(o->bdev, origin)) 403 struct list_head *sl = &_origins[origin_hash(o->bdev)]; 415 if (bdev_equal(o->dev->bdev, origin)) 423 struct list_head *sl = &_dm_origins[origin_hash(o->dev->bdev)]; 458 o = __lookup_origin(snap->origin->bdev); 465 if (!bdev_equal(s->cow->bdev, snap->cow->bdev)) 554 struct block_device *bdev = snap->origin->bdev; local 596 struct block_device *bdev = s->origin->bdev; local 1798 struct block_device *bdev = s->origin->bdev; local 2287 get_origin_minimum_chunksize(struct block_device *bdev) argument [all...] |
H A D | dm.c | 405 static int dm_blk_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 407 struct mapped_device *md = bdev->bd_disk->private_data; 413 struct block_device **bdev) 436 r = ti->type->prepare_ioctl(ti, bdev); 451 static int dm_blk_ioctl(struct block_device *bdev, blk_mode_t mode, argument 454 struct mapped_device *md = bdev->bd_disk->private_data; 457 r = dm_prepare_ioctl(md, &srcu_idx, &bdev); 464 * subset of the parent bdev; require extra privileges. 475 if (!bdev->bd_disk->fops->ioctl) 478 r = bdev 412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument 730 struct block_device *bdev; local 3182 dm_call_pr(struct block_device *bdev, iterate_devices_callout_fn fn, struct dm_pr *pr) argument 3243 dm_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, u32 flags) argument 3296 dm_pr_reserve(struct block_device *bdev, u64 key, enum pr_type type, u32 flags) argument 3339 dm_pr_release(struct block_device *bdev, u64 key, enum pr_type type) argument 3374 dm_pr_preempt(struct block_device *bdev, u64 old_key, u64 new_key, enum pr_type type, bool abort) argument 3392 dm_pr_clear(struct block_device *bdev, u64 key) argument 3430 dm_pr_read_keys(struct block_device *bdev, struct pr_keys *keys) argument 3462 dm_pr_read_reservation(struct block_device *bdev, struct pr_held_reservation *rsv) argument [all...] |
H A D | dm-exception-store.h | 170 static inline sector_t get_dev_size(struct block_device *bdev) argument 172 return bdev_nr_sectors(bdev);
|
H A D | dm-snap-transient.c | 46 sector_t size = get_dev_size(dm_snap_cow(store->snap)->bdev); 72 *total_sectors = get_dev_size(dm_snap_cow(store->snap)->bdev);
|
/linux-master/include/drm/ttm/ |
H A D | ttm_resource.h | 143 * @bdev: ttm device this manager belongs to 157 struct ttm_device *bdev; member in struct:ttm_resource_manager 168 * Protected by the bdev->lru_lock. 174 * bdev->lru_lock. 364 bool ttm_resource_intersects(struct ttm_device *bdev, 374 struct ttm_device *bdev, 377 int ttm_resource_manager_evict_all(struct ttm_device *bdev, 414 struct ttm_device *bdev, 418 struct ttm_device *bdev,
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_gem_vram_helper.c | 192 struct ttm_device *bdev; local 219 bdev = &vmm->bdev; 221 gbo->bo.bdev = bdev; 228 ret = ttm_bo_init_validate(bdev, &gbo->bo, ttm_bo_type_device, 847 static void bo_driver_ttm_tt_destroy(struct ttm_device *bdev, struct ttm_tt *tt) argument 929 static int bo_driver_io_mem_reserve(struct ttm_device *bdev, argument 932 struct drm_vram_mm *vmm = drm_vram_mm_of_bdev(bdev); 967 struct ttm_resource_manager *man = ttm_manager_type(&vmm->bdev, TTM_PL_VRA [all...] |
/linux-master/drivers/net/fddi/ |
H A D | defxx.c | 377 struct device __maybe_unused *bdev = bp->bus_dev; local 399 struct device __maybe_unused *bdev = bp->bus_dev; local 434 struct device *bdev = bp->bus_dev; local 435 int dfx_bus_pci = dev_is_pci(bdev); 436 int dfx_bus_eisa = DFX_BUS_EISA(bdev); 437 int dfx_bus_tc = DFX_BUS_TC(bdev); 442 bar_start[0] = pci_resource_start(to_pci_dev(bdev), num); 443 bar_len[0] = pci_resource_len(to_pci_dev(bdev), num); 448 unsigned long base_addr = to_eisa_device(bdev)->base_addr; 478 bar_start[0] = to_tc_dev(bdev) 530 dfx_register(struct device *bdev) argument 724 struct device *bdev = bp->bus_dev; local 883 struct device *bdev = bp->bus_dev; local 945 struct device __maybe_unused *bdev = bp->bus_dev; local 1043 struct device *bdev = bp->bus_dev; local 1951 struct device *bdev = bp->bus_dev; local 3695 dfx_unregister(struct device *bdev) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 68 static int amdgpu_ttm_backend_bind(struct ttm_device *bdev, 71 static void amdgpu_ttm_backend_unbind(struct ttm_device *bdev, 78 return ttm_range_man_init(&adev->mman.bdev, type, 93 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev); 356 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev); 474 r = amdgpu_ttm_backend_bind(bo->bdev, bo->ttm, new_mem); 480 adev = amdgpu_ttm_adev(bo->bdev); 500 amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm); 571 static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, argument 574 struct amdgpu_device *adev = amdgpu_ttm_adev(bdev); 768 amdgpu_ttm_tt_pin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument 805 amdgpu_ttm_tt_unpin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument 887 amdgpu_ttm_backend_bind(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_resource *bo_mem) argument 1022 amdgpu_ttm_backend_unbind(struct ttm_device *bdev, struct ttm_tt *ttm) argument 1050 amdgpu_ttm_backend_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument 1107 amdgpu_ttm_tt_populate(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) argument 1148 amdgpu_ttm_tt_unpopulate(struct ttm_device *bdev, struct ttm_tt *ttm) argument [all...] |
H A D | amdgpu_gtt_mgr.c | 51 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); 286 ttm_resource_manager_init(man, &adev->mman.bdev, gtt_size); 293 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, &mgr->manager); 314 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); 323 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, NULL);
|
H A D | amdgpu_dma_buf.c | 60 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 92 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 149 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 243 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 374 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); 505 if (amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) &&
|
/linux-master/drivers/target/ |
H A D | target_core_iblock.c | 442 struct block_device *bdev = IBLOCK_DEV(cmd->se_dev)->ibd_bd; local 446 ret = blkdev_issue_discard(bdev, 459 iblock_execute_zero_out(struct block_device *bdev, struct se_cmd *cmd) argument 479 ret = blkdev_issue_zeroout(bdev, 494 struct block_device *bdev = IBLOCK_DEV(cmd->se_dev)->ibd_bd; local 523 if (bdev_write_zeroes_sectors(bdev)) { 524 if (!iblock_execute_zero_out(bdev, cmd)) 849 struct block_device *bdev = ib_dev->ibd_bd; local 850 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops; 870 ret = ops->pr_register(bdev, ke 968 struct block_device *bdev = ib_dev->ibd_bd; local 1037 struct block_device *bdev = ib_dev->ibd_bd; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | core_linecard_dev.c | 84 linecard->bdev = linecard_bdev; 90 struct mlxsw_linecard_bdev *linecard_bdev = linecard->bdev; 97 linecard->bdev = NULL;
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_bo_evict.c | 29 struct ttm_device *bdev = &xe->ttm; local 43 ttm_manager_type(bdev, mem_type); 46 ret = ttm_resource_manager_evict_all(bdev, man);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_bo.c | 383 struct ttm_device *bdev = &dev_priv->bdev; local 397 ret = ttm_bo_init_reserved(bdev, &vmw_bo->tbo, params->bo_type, 637 struct ttm_device *bdev = bo->bdev; local 638 struct vmw_private *dev_priv = vmw_priv_from_ttm(bdev); 820 struct ttm_device *bdev = bo->tbo.bdev; local 821 struct vmw_private *vmw = vmw_priv_from_ttm(bdev); 846 struct ttm_device *bdev local [all...] |
/linux-master/drivers/nvme/target/ |
H A D | Makefile | 13 discovery.o io-cmd-file.o io-cmd-bdev.o
|
/linux-master/kernel/trace/ |
H A D | blktrace.c | 500 struct block_device *bdev) 502 if (bdev) { 503 bt->start_lba = bdev->bd_start_sect; 504 bt->end_lba = bdev->bd_start_sect + bdev_nr_sectors(bdev); 515 struct block_device *bdev, 537 * bdev can be NULL, as with scsi-generic, this is a helpful as 566 if (bdev && !bdev_is_partition(bdev)) 600 blk_trace_setup_lba(bt, bdev); 499 blk_trace_setup_lba(struct blk_trace *bt, struct block_device *bdev) argument 514 do_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, struct blk_user_trace_setup *buts) argument 621 __blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument 642 blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument 657 compat_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument 730 blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) argument 1627 blk_trace_setup_queue(struct request_queue *q, struct block_device *bdev) argument 1767 struct block_device *bdev = dev_to_bdev(dev); local 1801 struct block_device *bdev = dev_to_bdev(dev); local [all...] |