/linux-master/drivers/hid/ |
H A D | hid-picolcd_backlight.c | 17 static int picolcd_get_brightness(struct backlight_device *bdev) argument 19 struct picolcd_data *data = bl_get_data(bdev); 23 static int picolcd_set_brightness(struct backlight_device *bdev) argument 25 struct picolcd_data *data = bl_get_data(bdev); 32 data->lcd_brightness = bdev->props.brightness & 0x0ff; 33 data->lcd_power = bdev->props.power; 42 static int picolcd_check_bl_fb(struct backlight_device *bdev, struct fb_info *fb) argument 44 return fb && fb == picolcd_fbinfo((struct picolcd_data *)bl_get_data(bdev)); 56 struct backlight_device *bdev; local 69 bdev 84 struct backlight_device *bdev = data->backlight; local [all...] |
/linux-master/drivers/net/fddi/ |
H A D | defza.h | 665 struct device *bdev; /* pointer to the bus device */ member in struct:fza_private
|
/linux-master/drivers/net/wireless/broadcom/b43/ |
H A D | bus.c | 23 return 0; /* bcma_bus_may_powerdown(dev->bdev->bus); */ 32 return bcma_core_is_enabled(dev->bdev); 37 bcma_core_enable(dev->bdev, core_specific_flags); 42 bcma_core_disable(dev->bdev, core_specific_flags); 46 return bcma_read16(dev->bdev, offset); 50 return bcma_read32(dev->bdev, offset); 55 bcma_write16(dev->bdev, offset, value); 60 bcma_write32(dev->bdev, offset, value); 66 bcma_block_read(dev->bdev, buffer, count, offset, reg_width); 72 bcma_block_write(dev->bdev, buffe [all...] |
H A D | bus.h | 17 struct bcma_device *bdev; member in union:b43_bus_dev::__anon7168 71 return (dev->bdev->bus->hosttype == BCMA_HOSTTYPE_PCI);
|
H A D | phy_lcn.c | 589 struct bcma_drv_cc *cc = &dev->dev->bdev->bus->drv_cc; 700 struct bcma_drv_cc *cc = &dev->dev->bdev->bus->drv_cc;
|
/linux-master/arch/m68k/emu/ |
H A D | nfblock.c | 80 static int nfhd_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 82 struct nfhd_device *dev = bdev->bd_disk->private_data;
|
/linux-master/arch/um/drivers/ |
H A D | ubd_kern.c | 111 static int ubd_ioctl(struct block_device *bdev, blk_mode_t mode, 113 static int ubd_getgeo(struct block_device *bdev, struct hd_geometry *geo); 1344 static int ubd_getgeo(struct block_device *bdev, struct hd_geometry *geo) argument 1346 struct ubd *ubd_dev = bdev->bd_disk->private_data; 1354 static int ubd_ioctl(struct block_device *bdev, blk_mode_t mode, argument 1357 struct ubd *ubd_dev = bdev->bd_disk->private_data;
|
/linux-master/block/ |
H A D | Makefile | 6 obj-y := bdev.o fops.o bio.o elevator.o blk-core.o blk-sysfs.o \
|
H A D | bdev.c | 37 struct block_device bdev; member in struct:bdev_inode 48 return &BDEV_I(inode)->bdev; 58 static void bdev_write_inode(struct block_device *bdev) argument 60 struct inode *inode = bdev->bd_inode; 70 bdev, ret); 77 static void kill_bdev(struct block_device *bdev) argument 79 struct address_space *mapping = bdev->bd_inode->i_mapping; 89 void invalidate_bdev(struct block_device *bdev) argument 91 struct address_space *mapping = bdev->bd_inode->i_mapping; 102 * Drop all buffers & page cache for given bdev rang 105 truncate_bdev_range(struct block_device *bdev, blk_mode_t mode, loff_t lstart, loff_t lend) argument 134 set_init_blocksize(struct block_device *bdev) argument 147 set_blocksize(struct block_device *bdev, int size) argument 191 sync_blockdev_nowait(struct block_device *bdev) argument 203 sync_blockdev(struct block_device *bdev) argument 211 sync_blockdev_range(struct block_device *bdev, loff_t lstart, loff_t lend) argument 232 bdev_freeze(struct block_device *bdev) argument 268 bdev_thaw(struct block_device *bdev) argument 321 struct block_device *bdev = I_BDEV(inode); local 399 struct block_device *bdev; local 430 bdev_set_nr_sectors(struct block_device *bdev, sector_t sectors) argument 438 bdev_add(struct block_device *bdev, dev_t dev) argument 472 bd_may_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 514 bd_prepare_to_claim(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 566 bd_finish_claiming(struct block_device *bdev, void *holder, const struct blk_holder_ops *hops) argument 597 bd_abort_claiming(struct block_device *bdev, void *holder) argument 605 bd_end_claim(struct block_device *bdev, void *holder) argument 640 blkdev_flush_mapping(struct block_device *bdev) argument 648 blkdev_put_whole(struct block_device *bdev) argument 656 blkdev_get_whole(struct block_device *bdev, blk_mode_t mode) argument 752 struct block_device *bdev; local 774 blkdev_put_no_open(struct block_device *bdev) argument 779 bdev_writes_blocked(struct block_device *bdev) argument 784 bdev_block_writes(struct block_device *bdev) argument 789 bdev_unblock_writes(struct block_device *bdev) argument 794 bdev_may_open(struct block_device *bdev, blk_mode_t mode) argument 806 bdev_claim_write_access(struct block_device *bdev, blk_mode_t mode) argument 825 struct block_device *bdev; local 858 bdev_open(struct block_device *bdev, blk_mode_t mode, void *holder, const struct blk_holder_ops *hops, struct file *bdev_file) argument 969 struct block_device *bdev; local 1027 struct block_device *bdev = file_bdev(bdev_file); local 1041 struct block_device *bdev = file_bdev(bdev_file); local 1097 struct block_device *bdev = file_bdev(bdev_file); local 1169 bdev_mark_dead(struct block_device *bdev, bool surprise) argument 1196 struct block_device *bdev; local 1250 struct block_device *bdev; local [all...] |
H A D | bio.c | 245 void bio_init(struct bio *bio, struct block_device *bdev, struct bio_vec *table, argument 249 bio->bi_bdev = bdev; 264 if (bdev) 291 * @bdev: block device to use the bio for 300 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) argument 305 bio->bi_bdev = bdev; 348 struct bio *blk_next_bio(struct bio *bio, struct block_device *bdev, argument 351 struct bio *new = bio_alloc(bdev, nr_pages, opf, gfp); 432 static struct bio *bio_alloc_percpu_cache(struct block_device *bdev, argument 453 bio_init(bio, bdev, nr_vec 492 bio_alloc_bioset(struct block_device *bdev, unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask, struct bio_set *bs) argument 848 bio_alloc_clone(struct block_device *bdev, struct bio *bio_src, gfp_t gfp, struct bio_set *bs) argument 879 bio_init_clone(struct block_device *bdev, struct bio *bio, struct bio *bio_src, gfp_t gfp) argument [all...] |
H A D | blk-cgroup.c | 747 * blkg_conf_open_bdev - parse and open bdev for per-blkg config update 751 * @ctx->input and get and store the matching bdev in @ctx->bdev. @ctx->body is 756 * explicitly if bdev access is needed without resolving the blkcg / policy part 763 struct block_device *bdev; local 766 if (ctx->bdev) 777 bdev = blkdev_get_no_open(MKDEV(major, minor)); 778 if (!bdev) 780 if (bdev_is_partition(bdev)) { 781 blkdev_put_no_open(bdev); 1073 struct block_device *bdev = dev_to_bdev(dev); local [all...] |
H A D | blk-cgroup.h | 214 struct block_device *bdev; member in struct:blkg_conf_ctx
|
H A D | blk-core.c | 745 struct block_device *bdev = bio->bi_bdev; local 746 struct request_queue *q = bdev_get_queue(bdev); 755 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) 764 if (bdev->bd_partno && unlikely(blk_partition_remap(bio))) 799 if (!bdev_max_discard_sectors(bdev)) 803 if (!bdev_max_secure_erase_sectors(bdev)) 900 struct block_device *bdev; local 904 bdev = READ_ONCE(bio->bi_bdev); 905 if (!bdev) 908 q = bdev_get_queue(bdev); 1000 bdev_start_io_acct(struct block_device *bdev, enum req_op op, unsigned long start_time) argument 1024 bdev_end_io_acct(struct block_device *bdev, enum req_op op, unsigned int sectors, unsigned long start_time) argument [all...] |
H A D | blk-crypto.c | 360 bool blk_crypto_config_supported_natively(struct block_device *bdev, argument 363 return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile, 372 bool blk_crypto_config_supported(struct block_device *bdev, argument 376 blk_crypto_config_supported_natively(bdev, cfg); 381 * @bdev: block device to operate on 394 int blk_crypto_start_using_key(struct block_device *bdev, argument 397 if (blk_crypto_config_supported_natively(bdev, &key->crypto_cfg)) 404 * @bdev: a block_device on which I/O using the key may have been done 417 void blk_crypto_evict_key(struct block_device *bdev, argument 420 struct request_queue *q = bdev_get_queue(bdev); [all...] |
H A D | blk-flush.c | 464 * @bdev: blockdev to issue flush for 469 int blkdev_issue_flush(struct block_device *bdev) argument 473 bio_init(&bio, bdev, NULL, 0, REQ_OP_WRITE | REQ_PREFLUSH);
|
H A D | blk-iocost.c | 3227 disk = ctx.bdev->bd_disk; 3421 q = bdev_get_queue(ctx.bdev); 3429 ret = blk_iocost_init(ctx.bdev->bd_disk);
|
H A D | blk-iolatency.c | 849 lockdep_assert_held(&ctx.bdev->bd_queue->rq_qos_mutex); 850 if (!iolat_rq_qos(ctx.bdev->bd_queue)) 851 ret = blk_iolatency_init(ctx.bdev->bd_disk);
|
H A D | blk-lib.c | 13 static sector_t bio_discard_limit(struct block_device *bdev, sector_t sector) argument 15 unsigned int discard_granularity = bdev_discard_granularity(bdev); 18 if (bdev_is_partition(bdev)) 19 sector += bdev->bd_start_sect; 38 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector, argument 44 if (bdev_read_only(bdev)) 46 if (!bdev_max_discard_sectors(bdev)) 50 if (WARN_ON_ONCE(!bdev_discard_granularity(bdev))) { 52 bdev); 56 bs_mask = (bdev_logical_block_size(bdev) >> 97 blkdev_issue_discard(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask) argument 118 __blkdev_issue_write_zeroes(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument 165 __blkdev_issue_zero_pages(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop) argument 215 __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, unsigned flags) argument 249 blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp_mask, unsigned flags) argument 300 blkdev_issue_secure_erase(struct block_device *bdev, sector_t sector, sector_t nr_sects, gfp_t gfp) argument [all...] |
H A D | blk-settings.c | 892 * @bdev: the underlying block device (bottom) 904 void queue_limits_stack_bdev(struct queue_limits *t, struct block_device *bdev, argument 907 if (blk_stack_limits(t, &bdev_get_queue(bdev)->limits, 908 get_start_sect(bdev) + offset)) 910 pfx, bdev); 1100 int bdev_alignment_offset(struct block_device *bdev) argument 1102 struct request_queue *q = bdev_get_queue(bdev); 1106 if (bdev_is_partition(bdev)) 1108 bdev->bd_start_sect); 1113 unsigned int bdev_discard_alignment(struct block_device *bdev) argument [all...] |
H A D | blk-zoned.c | 102 * @bdev: Target device 107 unsigned int bdev_nr_zones(struct block_device *bdev) argument 109 sector_t zone_sectors = bdev_zone_sectors(bdev); 111 if (!bdev_is_zoned(bdev)) 113 return (bdev_nr_sectors(bdev) + zone_sectors - 1) >> 120 * @bdev: Target block device 137 int blkdev_report_zones(struct block_device *bdev, sector_t sector, argument 140 struct gendisk *disk = bdev->bd_disk; 143 if (!bdev_is_zoned(bdev) || WARN_ON_ONCE(!disk->fops->report_zones)) 179 static int blkdev_zone_reset_all_emulated(struct block_device *bdev) argument 224 blkdev_zone_reset_all(struct block_device *bdev) argument 247 blkdev_zone_mgmt(struct block_device *bdev, enum req_op op, sector_t sector, sector_t nr_sectors) argument 323 blkdev_report_zones_ioctl(struct block_device *bdev, unsigned int cmd, unsigned long arg) argument 356 blkdev_truncate_zone_range(struct block_device *bdev, blk_mode_t mode, const struct blk_zone_range *zrange) argument 376 blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg) argument [all...] |
H A D | blk.h | 411 int blkdev_report_zones_ioctl(struct block_device *bdev, unsigned int cmd, 413 int blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, 417 static inline int blkdev_report_zones_ioctl(struct block_device *bdev, argument 422 static inline int blkdev_zone_mgmt_ioctl(struct block_device *bdev, argument 430 void bdev_add(struct block_device *bdev, dev_t dev); 444 void bdev_set_nr_sectors(struct block_device *bdev, sector_t sectors); 481 int truncate_bdev_range(struct block_device *bdev, blk_mode_t mode, 599 int bdev_open(struct block_device *bdev, blk_mode_t mode, void *holder,
|
H A D | early-lookup.c | 23 struct block_device *bdev = dev_to_bdev(dev); local 26 if (!bdev->bd_meta_info || 27 strncasecmp(cmp->uuid, bdev->bd_meta_info->uuid, cmp->len)) 104 struct block_device *bdev = dev_to_bdev(dev); local 107 if (!bdev->bd_meta_info || strcmp(label, bdev->bd_meta_info->volname))
|
H A D | fops.c | 37 static bool blkdev_dio_unaligned(struct block_device *bdev, loff_t pos, argument 40 return pos & (bdev_logical_block_size(bdev) - 1) || 41 !bdev_iter_is_aligned(bdev, iter); 49 struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host); local 56 if (blkdev_dio_unaligned(bdev, pos, iter)) 69 bio_init(&bio, bdev, vecs, nr_pages, REQ_OP_READ); 73 bio_init(&bio, bdev, vecs, nr_pages, dio_bio_write_op(iocb)); 166 struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host); local 175 if (blkdev_dio_unaligned(bdev, pos, iter)) 180 bio = bio_alloc_bioset(bdev, nr_page 307 struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host); local 388 struct block_device *bdev = I_BDEV(inode); local 544 struct block_device *bdev = I_BDEV(filp->f_mapping->host); local 604 struct block_device *bdev; local 671 struct block_device *bdev = I_BDEV(file->f_mapping->host); local 719 struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host); local 770 struct block_device *bdev = I_BDEV(inode); local [all...] |
H A D | genhd.c | 415 /* Mark bdev as having a submit_bio, if needed */ 562 struct block_device *bdev; local 573 xa_for_each(&disk->part_tbl, idx, bdev) { 574 if (!kobject_get_unless_zero(&bdev->bd_device.kobj)) 578 bdev_mark_dead(bdev, surprise); 580 put_device(&bdev->bd_device); 655 * Prevent new openers by unlinked the bdev inode. 745 struct block_device *bdev = disk->part0; local 747 invalidate_bdev(bdev); 748 bdev 956 struct block_device *bdev = dev_to_bdev(dev); local 1005 struct block_device *bdev = dev_to_bdev(dev); local [all...] |
H A D | holder.c | 13 static struct bd_holder_disk *bd_find_holder_disk(struct block_device *bdev, argument 19 if (holder->holder_dir == bdev->bd_holder_dir) 35 * bd_link_disk_holder - create symlinks between holding disk and slave bdev 36 * @bdev: the claimed slave bdev 43 * - from "slaves" directory of the holder @disk to the claimed @bdev 44 * - from "holders" directory of the @bdev to the holder @disk 52 * The caller must have claimed @bdev before calling this function and 53 * ensure that both @bdev and @disk are valid during the creation and 62 int bd_link_disk_holder(struct block_device *bdev, struc argument 138 bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk) argument [all...] |