Searched refs:bdev (Results 76 - 100 of 414) sorted by relevance

1234567891011>>

/linux-master/include/linux/
H A Ddm-io.h20 struct block_device *bdev; member in struct:dm_io_region
/linux-master/fs/bcachefs/
H A Dsuper_types.h8 struct block_device *bdev; member in struct:bch_sb_handle
/linux-master/drivers/net/wireless/broadcom/b43/
H A Dbus.h17 struct bcma_device *bdev; member in union:b43_bus_dev::__anon7168
71 return (dev->bdev->bus->hosttype == BCMA_HOSTTYPE_PCI);
/linux-master/fs/btrfs/
H A Dzoned.h33 * Number of zones, zone size and types of zones if bdev is a
64 int btrfs_sb_log_location_bdev(struct block_device *bdev, int mirror, int rw,
69 int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror);
145 static inline int btrfs_sb_log_location_bdev(struct block_device *bdev, argument
164 static inline int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror) argument
323 struct block_device *bdev)
330 if (!bdev_is_zoned(bdev))
334 (bdev_zone_sectors(bdev) << SECTOR_SHIFT);
338 return !bdev_is_zoned(bdev);
322 btrfs_check_device_zone_type(const struct btrfs_fs_info *fs_info, struct block_device *bdev) argument
H A Dzoned.c84 static int sb_write_pointer(struct block_device *bdev, struct blk_zone *zones, argument
121 struct address_space *mapping = bdev->bd_inode->i_mapping;
180 struct block_device *bdev)
182 return (sector_t)zone_number << ilog2(bdev_zone_sectors(bdev));
200 sector_t bdev_size = bdev_nr_sectors(device->bdev);
230 if (!bdev_is_zoned(device->bdev)) {
265 ret = blkdev_report_zones(device->bdev, pos >> SECTOR_SHIFT, *nr_zones,
346 if (!device->bdev)
362 struct block_device *bdev = device->bdev; local
179 zone_start_sector(u32 zone_number, struct block_device *bdev) argument
801 sb_log_location(struct block_device *bdev, struct blk_zone *zones, int rw, u64 *bytenr_ret) argument
862 btrfs_sb_log_location_bdev(struct block_device *bdev, int mirror, int rw, u64 *bytenr_ret) argument
1001 btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror) argument
[all...]
/linux-master/block/
H A Dearly-lookup.c23 struct block_device *bdev = dev_to_bdev(dev); local
26 if (!bdev->bd_meta_info ||
27 strncasecmp(cmp->uuid, bdev->bd_meta_info->uuid, cmp->len))
104 struct block_device *bdev = dev_to_bdev(dev); local
107 if (!bdev->bd_meta_info || strcmp(label, bdev->bd_meta_info->volname))
H A Dblk-core.c745 struct block_device *bdev = bio->bi_bdev; local
746 struct request_queue *q = bdev_get_queue(bdev);
755 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev))
764 if (bdev->bd_partno && unlikely(blk_partition_remap(bio)))
799 if (!bdev_max_discard_sectors(bdev))
803 if (!bdev_max_secure_erase_sectors(bdev))
900 struct block_device *bdev; local
904 bdev = READ_ONCE(bio->bi_bdev);
905 if (!bdev)
908 q = bdev_get_queue(bdev);
1000 bdev_start_io_acct(struct block_device *bdev, enum req_op op, unsigned long start_time) argument
1024 bdev_end_io_acct(struct block_device *bdev, enum req_op op, unsigned int sectors, unsigned long start_time) argument
[all...]
/linux-master/drivers/md/
H A Ddm-zoned-target.c128 clone = bio_alloc_clone(dev->bdev, bio, GFP_NOIO, &dmz->bio_set);
589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) {
612 disk = dmz_dev->bdev->bd_disk;
695 struct block_device *bdev; local
704 bdev = ddev->bdev;
705 if (!bdev_is_zoned(bdev)) {
731 dev->bdev = bdev;
734 dev->capacity = bdev_nr_sectors(bdev);
782 struct block_device *bdev = zoned_dev->bdev; local
799 struct block_device *bdev = zoned_dev->bdev; local
1019 dmz_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) argument
[all...]
H A Ddm-table.c222 if (dd->dm_dev->bdev->bd_dev == dev)
235 struct block_device *bdev = dev->bdev; local
236 sector_t dev_size = bdev_nr_sectors(bdev);
245 dm_device_name(ti->table->md), bdev,
256 if (bdev_is_zoned(bdev)) {
257 unsigned int zone_sectors = bdev_zone_sectors(bdev);
263 zone_sectors, bdev);
280 zone_sectors, bdev);
292 limits->logical_block_size, bdev);
409 struct block_device *bdev = dev->bdev; local
864 struct block_device *bdev = dev->bdev; local
[all...]
H A Ddm-snap.c316 struct block_device *bdev; member in struct:origin
383 static unsigned int origin_hash(struct block_device *bdev) argument
385 return bdev->bd_dev & ORIGIN_MASK;
395 if (bdev_equal(o->bdev, origin))
403 struct list_head *sl = &_origins[origin_hash(o->bdev)];
415 if (bdev_equal(o->dev->bdev, origin))
423 struct list_head *sl = &_dm_origins[origin_hash(o->dev->bdev)];
458 o = __lookup_origin(snap->origin->bdev);
465 if (!bdev_equal(s->cow->bdev, snap->cow->bdev))
554 struct block_device *bdev = snap->origin->bdev; local
596 struct block_device *bdev = s->origin->bdev; local
1798 struct block_device *bdev = s->origin->bdev; local
2287 get_origin_minimum_chunksize(struct block_device *bdev) argument
[all...]
H A Ddm.c405 static int dm_blk_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument
407 struct mapped_device *md = bdev->bd_disk->private_data;
413 struct block_device **bdev)
436 r = ti->type->prepare_ioctl(ti, bdev);
451 static int dm_blk_ioctl(struct block_device *bdev, blk_mode_t mode, argument
454 struct mapped_device *md = bdev->bd_disk->private_data;
457 r = dm_prepare_ioctl(md, &srcu_idx, &bdev);
464 * subset of the parent bdev; require extra privileges.
475 if (!bdev->bd_disk->fops->ioctl)
478 r = bdev
412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument
730 struct block_device *bdev; local
3182 dm_call_pr(struct block_device *bdev, iterate_devices_callout_fn fn, struct dm_pr *pr) argument
3243 dm_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, u32 flags) argument
3296 dm_pr_reserve(struct block_device *bdev, u64 key, enum pr_type type, u32 flags) argument
3339 dm_pr_release(struct block_device *bdev, u64 key, enum pr_type type) argument
3374 dm_pr_preempt(struct block_device *bdev, u64 old_key, u64 new_key, enum pr_type type, bool abort) argument
3392 dm_pr_clear(struct block_device *bdev, u64 key) argument
3430 dm_pr_read_keys(struct block_device *bdev, struct pr_keys *keys) argument
3462 dm_pr_read_reservation(struct block_device *bdev, struct pr_held_reservation *rsv) argument
[all...]
H A Ddm-exception-store.h170 static inline sector_t get_dev_size(struct block_device *bdev) argument
172 return bdev_nr_sectors(bdev);
H A Ddm-snap-transient.c46 sector_t size = get_dev_size(dm_snap_cow(store->snap)->bdev);
72 *total_sectors = get_dev_size(dm_snap_cow(store->snap)->bdev);
/linux-master/include/drm/ttm/
H A Dttm_resource.h143 * @bdev: ttm device this manager belongs to
157 struct ttm_device *bdev; member in struct:ttm_resource_manager
168 * Protected by the bdev->lru_lock.
174 * bdev->lru_lock.
364 bool ttm_resource_intersects(struct ttm_device *bdev,
374 struct ttm_device *bdev,
377 int ttm_resource_manager_evict_all(struct ttm_device *bdev,
414 struct ttm_device *bdev,
418 struct ttm_device *bdev,
/linux-master/drivers/gpu/drm/
H A Ddrm_gem_vram_helper.c192 struct ttm_device *bdev; local
219 bdev = &vmm->bdev;
221 gbo->bo.bdev = bdev;
228 ret = ttm_bo_init_validate(bdev, &gbo->bo, ttm_bo_type_device,
847 static void bo_driver_ttm_tt_destroy(struct ttm_device *bdev, struct ttm_tt *tt) argument
929 static int bo_driver_io_mem_reserve(struct ttm_device *bdev, argument
932 struct drm_vram_mm *vmm = drm_vram_mm_of_bdev(bdev);
967 struct ttm_resource_manager *man = ttm_manager_type(&vmm->bdev, TTM_PL_VRA
[all...]
/linux-master/drivers/net/fddi/
H A Ddefxx.c377 struct device __maybe_unused *bdev = bp->bus_dev; local
399 struct device __maybe_unused *bdev = bp->bus_dev; local
434 struct device *bdev = bp->bus_dev; local
435 int dfx_bus_pci = dev_is_pci(bdev);
436 int dfx_bus_eisa = DFX_BUS_EISA(bdev);
437 int dfx_bus_tc = DFX_BUS_TC(bdev);
442 bar_start[0] = pci_resource_start(to_pci_dev(bdev), num);
443 bar_len[0] = pci_resource_len(to_pci_dev(bdev), num);
448 unsigned long base_addr = to_eisa_device(bdev)->base_addr;
478 bar_start[0] = to_tc_dev(bdev)
530 dfx_register(struct device *bdev) argument
724 struct device *bdev = bp->bus_dev; local
883 struct device *bdev = bp->bus_dev; local
945 struct device __maybe_unused *bdev = bp->bus_dev; local
1043 struct device *bdev = bp->bus_dev; local
1951 struct device *bdev = bp->bus_dev; local
3695 dfx_unregister(struct device *bdev) argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_ttm.c68 static int amdgpu_ttm_backend_bind(struct ttm_device *bdev,
71 static void amdgpu_ttm_backend_unbind(struct ttm_device *bdev,
78 return ttm_range_man_init(&adev->mman.bdev, type,
93 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev);
356 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev);
474 r = amdgpu_ttm_backend_bind(bo->bdev, bo->ttm, new_mem);
480 adev = amdgpu_ttm_adev(bo->bdev);
500 amdgpu_ttm_backend_unbind(bo->bdev, bo->ttm);
571 static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, argument
574 struct amdgpu_device *adev = amdgpu_ttm_adev(bdev);
768 amdgpu_ttm_tt_pin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument
805 amdgpu_ttm_tt_unpin_userptr(struct ttm_device *bdev, struct ttm_tt *ttm) argument
887 amdgpu_ttm_backend_bind(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_resource *bo_mem) argument
1022 amdgpu_ttm_backend_unbind(struct ttm_device *bdev, struct ttm_tt *ttm) argument
1050 amdgpu_ttm_backend_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) argument
1107 amdgpu_ttm_tt_populate(struct ttm_device *bdev, struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) argument
1148 amdgpu_ttm_tt_unpopulate(struct ttm_device *bdev, struct ttm_tt *ttm) argument
[all...]
H A Damdgpu_gtt_mgr.c51 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT);
286 ttm_resource_manager_init(man, &adev->mman.bdev, gtt_size);
293 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, &mgr->manager);
314 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man);
323 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, NULL);
H A Damdgpu_dma_buf.c60 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
92 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
149 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
243 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
374 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
505 if (amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) &&
/linux-master/drivers/target/
H A Dtarget_core_iblock.c442 struct block_device *bdev = IBLOCK_DEV(cmd->se_dev)->ibd_bd; local
446 ret = blkdev_issue_discard(bdev,
459 iblock_execute_zero_out(struct block_device *bdev, struct se_cmd *cmd) argument
479 ret = blkdev_issue_zeroout(bdev,
494 struct block_device *bdev = IBLOCK_DEV(cmd->se_dev)->ibd_bd; local
523 if (bdev_write_zeroes_sectors(bdev)) {
524 if (!iblock_execute_zero_out(bdev, cmd))
849 struct block_device *bdev = ib_dev->ibd_bd; local
850 const struct pr_ops *ops = bdev->bd_disk->fops->pr_ops;
870 ret = ops->pr_register(bdev, ke
968 struct block_device *bdev = ib_dev->ibd_bd; local
1037 struct block_device *bdev = ib_dev->ibd_bd; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlxsw/
H A Dcore_linecard_dev.c84 linecard->bdev = linecard_bdev;
90 struct mlxsw_linecard_bdev *linecard_bdev = linecard->bdev;
97 linecard->bdev = NULL;
/linux-master/drivers/gpu/drm/xe/
H A Dxe_bo_evict.c29 struct ttm_device *bdev = &xe->ttm; local
43 ttm_manager_type(bdev, mem_type);
46 ret = ttm_resource_manager_evict_all(bdev, man);
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_bo.c383 struct ttm_device *bdev = &dev_priv->bdev; local
397 ret = ttm_bo_init_reserved(bdev, &vmw_bo->tbo, params->bo_type,
637 struct ttm_device *bdev = bo->bdev; local
638 struct vmw_private *dev_priv = vmw_priv_from_ttm(bdev);
820 struct ttm_device *bdev = bo->tbo.bdev; local
821 struct vmw_private *vmw = vmw_priv_from_ttm(bdev);
846 struct ttm_device *bdev local
[all...]
/linux-master/drivers/nvme/target/
H A DMakefile13 discovery.o io-cmd-file.o io-cmd-bdev.o
/linux-master/kernel/trace/
H A Dblktrace.c500 struct block_device *bdev)
502 if (bdev) {
503 bt->start_lba = bdev->bd_start_sect;
504 bt->end_lba = bdev->bd_start_sect + bdev_nr_sectors(bdev);
515 struct block_device *bdev,
537 * bdev can be NULL, as with scsi-generic, this is a helpful as
566 if (bdev && !bdev_is_partition(bdev))
600 blk_trace_setup_lba(bt, bdev);
499 blk_trace_setup_lba(struct blk_trace *bt, struct block_device *bdev) argument
514 do_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, struct blk_user_trace_setup *buts) argument
621 __blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument
642 blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument
657 compat_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, struct block_device *bdev, char __user *arg) argument
730 blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) argument
1627 blk_trace_setup_queue(struct request_queue *q, struct block_device *bdev) argument
1767 struct block_device *bdev = dev_to_bdev(dev); local
1801 struct block_device *bdev = dev_to_bdev(dev); local
[all...]

Completed in 308 milliseconds

1234567891011>>