Searched refs:md (Results 1 - 25 of 238) sorted by relevance

12345678910

/linux-master/drivers/md/
H A Ddm-ima.h59 void dm_ima_reset_data(struct mapped_device *md);
61 void dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap);
62 void dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all);
63 void dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map);
64 void dm_ima_measure_on_device_rename(struct mapped_device *md);
68 static inline void dm_ima_reset_data(struct mapped_device *md) {} argument
70 static inline void dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap) {} argument
71 static inline void dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all) {} argument
72 static inline void dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map) {} argument
73 static inline void dm_ima_measure_on_device_rename(struct mapped_device *md) {} argument
[all...]
H A Ddm-sysfs.c15 ssize_t (*show)(struct mapped_device *md, char *p);
16 ssize_t (*store)(struct mapped_device *md, const char *p, size_t count);
27 struct mapped_device *md; local
34 md = dm_get_from_kobject(kobj);
35 if (!md)
38 ret = dm_attr->show(md, page);
39 dm_put(md);
52 struct mapped_device *md; local
59 md = dm_get_from_kobject(kobj);
60 if (!md)
69 dm_attr_name_show(struct mapped_device *md, char *buf) argument
78 dm_attr_uuid_show(struct mapped_device *md, char *buf) argument
87 dm_attr_suspended_show(struct mapped_device *md, char *buf) argument
94 dm_attr_use_blk_mq_show(struct mapped_device *md, char *buf) argument
133 dm_sysfs_init(struct mapped_device *md) argument
143 dm_sysfs_exit(struct mapped_device *md) argument
[all...]
H A Ddm.c308 int dm_deleting_md(struct mapped_device *md) argument
310 return test_bit(DMF_DELETING, &md->flags);
315 struct mapped_device *md; local
319 md = disk->private_data;
320 if (!md)
323 if (test_bit(DMF_FREEING, &md->flags) ||
324 dm_deleting_md(md)) {
325 md = NULL;
329 dm_get(md);
330 atomic_inc(&md
339 struct mapped_device *md; local
356 dm_open_count(struct mapped_device *md) argument
364 dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred) argument
384 dm_cancel_deferred_remove(struct mapped_device *md) argument
407 struct mapped_device *md = bdev->bd_disk->private_data; local
412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument
446 dm_unprepare_ioctl(struct mapped_device *md, int srcu_idx) argument
454 struct mapped_device *md = bdev->bd_disk->private_data; local
573 alloc_io(struct mapped_device *md, struct bio *bio, gfp_t gfp_mask) argument
617 struct mapped_device *md = ci->io->md; local
670 queue_io(struct mapped_device *md, struct bio *bio) argument
699 dm_sync_table(struct mapped_device *md) argument
725 open_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode) argument
777 close_table_device(struct table_device *td, struct mapped_device *md) argument
799 dm_get_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode, struct dm_dev **result) argument
821 dm_put_table_device(struct mapped_device *md, struct dm_dev *d) argument
834 dm_get_geometry(struct mapped_device *md, struct hd_geometry *geo) argument
844 dm_set_geometry(struct mapped_device *md, struct hd_geometry *geo) argument
858 __noflush_suspending(struct mapped_device *md) argument
865 struct mapped_device *md = io->md; local
877 dm_kick_requeue(struct mapped_device *md, bool first_stage) argument
895 struct mapped_device *md = io->md; local
939 struct mapped_device *md = io->md; local
987 struct mapped_device *md = container_of(work, struct mapped_device, local
1073 dm_get_queue_limits(struct mapped_device *md) argument
1078 disable_discard(struct mapped_device *md) argument
1086 disable_write_zeroes(struct mapped_device *md) argument
1106 struct mapped_device *md = io->md; local
1230 struct mapped_device *md = dax_get_private(dax_dev); local
1257 struct mapped_device *md = dax_get_private(dax_dev); local
1284 struct mapped_device *md = dax_get_private(dax_dev); local
1385 __set_swap_bios_limit(struct mapped_device *md, int latch) argument
1406 struct mapped_device *md = io->md; local
1774 dm_split_and_process_bio(struct mapped_device *md, struct dm_table *map, struct bio *bio) argument
1845 struct mapped_device *md = bio->bi_bdev->bd_disk->private_data; local
2000 cleanup_mapped_device(struct mapped_device *md) argument
2061 struct mapped_device *md; local
2177 free_dev(struct mapped_device *md) argument
2200 struct mapped_device *md = context; local
2216 __bind(struct mapped_device *md, struct dm_table *t, struct queue_limits *limits) argument
2286 __unbind(struct mapped_device *md) argument
2305 struct mapped_device *md; local
2321 dm_lock_md_type(struct mapped_device *md) argument
2326 dm_unlock_md_type(struct mapped_device *md) argument
2331 dm_set_md_type(struct mapped_device *md, enum dm_queue_mode type) argument
2337 dm_get_md_type(struct mapped_device *md) argument
2342 dm_get_immutable_target_type(struct mapped_device *md) argument
2350 dm_setup_md_queue(struct mapped_device *md, struct dm_table *t) argument
2422 struct mapped_device *md; local
2444 dm_get_mdptr(struct mapped_device *md) argument
2449 dm_set_mdptr(struct mapped_device *md, void *ptr) argument
2454 dm_get(struct mapped_device *md) argument
2460 dm_hold(struct mapped_device *md) argument
2473 dm_device_name(struct mapped_device *md) argument
2479 __dm_destroy(struct mapped_device *md, bool wait) argument
2526 dm_destroy(struct mapped_device *md) argument
2531 dm_destroy_immediate(struct mapped_device *md) argument
2536 dm_put(struct mapped_device *md) argument
2542 dm_in_flight_bios(struct mapped_device *md) argument
2553 dm_wait_for_bios_completion(struct mapped_device *md, unsigned int task_state) argument
2578 dm_wait_for_completion(struct mapped_device *md, unsigned int task_state) argument
2605 struct mapped_device *md = container_of(work, struct mapped_device, work); local
2621 dm_queue_flush(struct mapped_device *md) argument
2631 dm_swap_table(struct mapped_device *md, struct dm_table *table) argument
2676 lock_fs(struct mapped_device *md) argument
2688 unlock_fs(struct mapped_device *md) argument
2705 __dm_suspend(struct mapped_device *md, struct dm_table *map, unsigned int suspend_flags, unsigned int task_state, int dmf_suspended_flag) argument
2813 dm_suspend(struct mapped_device *md, unsigned int suspend_flags) argument
2854 __dm_resume(struct mapped_device *md, struct dm_table *map) argument
2878 dm_resume(struct mapped_device *md) argument
2920 __dm_internal_suspend(struct mapped_device *md, unsigned int suspend_flags) argument
2950 __dm_internal_resume(struct mapped_device *md) argument
2986 dm_internal_suspend_noflush(struct mapped_device *md) argument
2994 dm_internal_resume(struct mapped_device *md) argument
3007 dm_internal_suspend_fast(struct mapped_device *md) argument
3020 dm_internal_resume_fast(struct mapped_device *md) argument
3037 dm_kobject_uevent(struct mapped_device *md, enum kobject_action action, unsigned int cookie, bool need_resize_uevent) argument
3063 dm_next_uevent_seq(struct mapped_device *md) argument
3068 dm_get_event_nr(struct mapped_device *md) argument
3073 dm_wait_event(struct mapped_device *md, int event_nr) argument
3079 dm_uevent_add(struct mapped_device *md, struct list_head *elist) argument
3092 dm_disk(struct mapped_device *md) argument
3098 dm_kobject(struct mapped_device *md) argument
3105 struct mapped_device *md; local
3121 dm_suspended_md(struct mapped_device *md) argument
3126 dm_post_suspending_md(struct mapped_device *md) argument
3131 dm_suspended_internally_md(struct mapped_device *md) argument
3136 dm_test_deferred_remove_flag(struct mapped_device *md) argument
3185 struct mapped_device *md = bdev->bd_disk->private_data; local
3394 struct mapped_device *md = bdev->bd_disk->private_data; local
[all...]
H A Ddm-era-target.c36 struct writeset_metadata md; member in struct:writeset
96 ws->md.nr_bits = nr_blocks;
97 r = setup_on_disk_bitset(info, ws->md.nr_bits, &ws->md.root);
142 r = dm_bitset_set_bit(info, ws->md.root, block, &ws->md.root);
311 static int superblock_read_lock(struct era_metadata *md, argument
314 return dm_bm_read_lock(md->bm, SUPERBLOCK_LOCATION,
318 static int superblock_lock_zero(struct era_metadata *md, argument
321 return dm_bm_write_lock_zero(md
325 superblock_lock(struct era_metadata *md, struct dm_block **sblock) argument
378 struct era_metadata *md = context; local
392 struct era_metadata *md = context; local
411 setup_writeset_tree_info(struct era_metadata *md) argument
424 setup_era_array_info(struct era_metadata *md) argument
437 setup_infos(struct era_metadata *md) argument
446 create_fresh_metadata(struct era_metadata *md) argument
480 save_sm_root(struct era_metadata *md) argument
493 copy_sm_root(struct era_metadata *md, struct superblock_disk *disk) argument
505 prepare_superblock(struct era_metadata *md, struct superblock_disk *disk) argument
527 write_superblock(struct era_metadata *md) argument
552 format_metadata(struct era_metadata *md) argument
570 open_metadata(struct era_metadata *md) argument
621 open_or_format_metadata(struct era_metadata *md, bool may_format) argument
637 create_persistent_data_objects(struct era_metadata *md, bool may_format) argument
656 destroy_persistent_data_objects(struct era_metadata *md) argument
666 swap_writeset(struct era_metadata *md, struct writeset *new_writeset) argument
694 metadata_digest_remove_writeset(struct era_metadata *md, struct digest *d) argument
713 metadata_digest_transcribe_writeset(struct era_metadata *md, struct digest *d) argument
747 metadata_digest_lookup_writeset(struct era_metadata *md, struct digest *d) argument
789 metadata_digest_start(struct era_metadata *md, struct digest *d) argument
811 struct era_metadata *md = kzalloc(sizeof(*md), GFP_KERNEL); local
832 metadata_close(struct era_metadata *md) argument
849 metadata_resize(struct era_metadata *md, void *arg) argument
893 metadata_era_archive(struct era_metadata *md) argument
924 next_writeset(struct era_metadata *md) argument
930 metadata_new_era(struct era_metadata *md) argument
947 metadata_era_rollover(struct era_metadata *md) argument
970 metadata_current_marked(struct era_metadata *md, dm_block_t block) argument
983 metadata_commit(struct era_metadata *md) argument
1020 metadata_checkpoint(struct era_metadata *md) argument
1032 metadata_take_snap(struct era_metadata *md) argument
1091 metadata_drop_snap(struct era_metadata *md) argument
1144 metadata_get_stats(struct era_metadata *md, void *ptr) argument
1181 struct era_metadata *md; member in struct:era
1394 in_worker0(struct era *era, int (*fn)(struct era_metadata *md)) argument
1404 in_worker1(struct era *era, int (*fn)(struct era_metadata *md, void *ref), void *arg) argument
1470 struct era_metadata *md; local
[all...]
H A Ddm-rq.h34 int dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t);
35 void dm_mq_cleanup_mapped_device(struct mapped_device *md);
40 void dm_mq_kick_requeue_list(struct mapped_device *md);
44 ssize_t dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf);
45 ssize_t dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md,
H A Ddm-rq.c19 struct mapped_device *md; member in struct:dm_rq_target_io
59 int dm_request_based(struct mapped_device *md) argument
61 return queue_is_mq(md->queue);
127 static void rq_end_stats(struct mapped_device *md, struct request *orig) argument
129 if (unlikely(dm_stats_used(&md->stats))) {
133 dm_stats_account_io(&md->stats, rq_data_dir(orig),
140 * Don't touch any member of the md after calling this function because
141 * the md may be freed in dm_put() at the end of this function.
144 static void rq_completed(struct mapped_device *md) argument
149 dm_put(md);
160 struct mapped_device *md = tio->md; local
176 dm_mq_kick_requeue_list(struct mapped_device *md) argument
190 struct mapped_device *md = tio->md; local
258 struct mapped_device *md = tio->md; local
338 init_tio(struct dm_rq_target_io *tio, struct request *rq, struct mapped_device *md) argument
366 struct mapped_device *md = tio->md; local
422 dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf) argument
427 dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md, const char *buf, size_t count) argument
433 dm_start_request(struct mapped_device *md, struct request *orig) argument
460 struct mapped_device *md = set->driver_data; local
482 struct mapped_device *md = tio->md; local
536 dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t) argument
578 dm_mq_cleanup_mapped_device(struct mapped_device *md) argument
[all...]
H A Ddm-zone.c21 static int dm_blk_do_report_zones(struct mapped_device *md, struct dm_table *t, argument
25 struct gendisk *disk = md->disk;
59 struct mapped_device *md = disk->private_data; local
63 if (dm_suspended_md(md))
66 map = dm_get_live_table(md, &srcu_idx);
70 ret = dm_blk_do_report_zones(md, map, sector, nr_zones, cb, data);
72 dm_put_live_table(md, srcu_idx);
125 bool dm_is_zone_write(struct mapped_device *md, struct bio *bio) argument
127 struct request_queue *q = md->queue;
141 void dm_cleanup_zoned_dev(struct mapped_device *md) argument
180 struct mapped_device *md = data; local
226 dm_revalidate_zones(struct mapped_device *md, struct dm_table *t) argument
291 struct mapped_device *md = t->md; local
328 dm_update_zone_wp_offset(struct mapped_device *md, unsigned int zno, unsigned int *wp_ofst) argument
367 dm_zone_map_bio_begin(struct mapped_device *md, unsigned int zno, struct bio *clone) argument
423 dm_zone_map_bio_end(struct mapped_device *md, unsigned int zno, struct orig_bio_details *orig_bio_details, unsigned int nr_sectors) argument
516 struct mapped_device *md = io->md; local
587 struct mapped_device *md = io->md; local
[all...]
H A Ddm-ima.c67 static int dm_ima_alloc_and_copy_name_uuid(struct mapped_device *md, char **dev_name, argument
83 r = dm_copy_name_and_uuid(md, *dev_name, *dev_uuid);
102 static int dm_ima_alloc_and_copy_device_data(struct mapped_device *md, char **device_data, argument
108 r = dm_ima_alloc_and_copy_name_uuid(md, &dev_name, &dev_uuid, noio);
120 dev_name, dev_uuid, md->disk->major, md->disk->first_minor,
121 md->disk->minors, num_targets);
149 static int dm_ima_alloc_and_copy_capacity_str(struct mapped_device *md, char **capacity_str, argument
154 capacity = get_capacity(md->disk);
169 void dm_ima_reset_data(struct mapped_device *md) argument
369 dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap) argument
474 dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all) argument
605 dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map) argument
704 dm_ima_measure_on_device_rename(struct mapped_device *md) argument
[all...]
H A Ddm.h77 void dm_lock_md_type(struct mapped_device *md);
78 void dm_unlock_md_type(struct mapped_device *md);
79 void dm_set_md_type(struct mapped_device *md, enum dm_queue_mode type);
80 enum dm_queue_mode dm_get_md_type(struct mapped_device *md);
81 struct target_type *dm_get_immutable_target_type(struct mapped_device *md);
83 int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t);
107 void dm_cleanup_zoned_dev(struct mapped_device *md);
110 bool dm_is_zone_write(struct mapped_device *md, struct bio *bio);
113 static inline void dm_cleanup_zoned_dev(struct mapped_device *md) {} argument
115 static inline bool dm_is_zone_write(struct mapped_device *md, struc argument
[all...]
/linux-master/drivers/video/fbdev/omap/
H A Dlcd_mipid.c56 static void mipid_transfer(struct mipid_device *md, int cmd, const u8 *wbuf, argument
64 BUG_ON(md->spi == NULL);
105 r = spi_sync(md->spi, &m);
107 dev_dbg(&md->spi->dev, "spi_sync %d\n", r);
113 static inline void mipid_cmd(struct mipid_device *md, int cmd) argument
115 mipid_transfer(md, cmd, NULL, 0, NULL, 0);
118 static inline void mipid_write(struct mipid_device *md, argument
121 mipid_transfer(md, reg, buf, len, NULL, 0);
124 static inline void mipid_read(struct mipid_device *md, argument
127 mipid_transfer(md, re
130 set_data_lines(struct mipid_device *md, int data_lines) argument
148 send_init_string(struct mipid_device *md) argument
156 hw_guard_start(struct mipid_device *md, int guard_msec) argument
162 hw_guard_wait(struct mipid_device *md) argument
172 set_sleep_mode(struct mipid_device *md, int on) argument
194 set_display_state(struct mipid_device *md, int enabled) argument
203 struct mipid_device *md = to_mipid_device(panel); local
221 struct mipid_device *md = to_mipid_device(panel); local
231 struct mipid_device *md = to_mipid_device(panel); local
245 read_first_pixel(struct mipid_device *md) argument
275 struct mipid_device *md = to_mipid_device(panel); local
312 ls041y3_esd_recover(struct mipid_device *md) argument
319 ls041y3_esd_check_mode1(struct mipid_device *md) argument
335 ls041y3_esd_check_mode2(struct mipid_device *md) argument
371 ls041y3_esd_check(struct mipid_device *md) argument
378 mipid_esd_start_check(struct mipid_device *md) argument
385 mipid_esd_stop_check(struct mipid_device *md) argument
393 struct mipid_device *md = container_of(work, struct mipid_device, local
404 struct mipid_device *md = to_mipid_device(panel); local
425 struct mipid_device *md = to_mipid_device(panel); local
447 panel_enabled(struct mipid_device *md) argument
464 struct mipid_device *md = to_mipid_device(panel); local
482 struct mipid_device *md = to_mipid_device(panel); local
513 mipid_detect(struct mipid_device *md) argument
552 struct mipid_device *md; local
587 struct mipid_device *md = dev_get_drvdata(&spi->dev); local
[all...]
/linux-master/tools/thermal/lib/
H A Dmainloop.c26 struct mainloop_data *md; local
45 md = events[i].data.ptr;
47 if (md->cb(md->fd, md->data) > 0)
59 struct mainloop_data *md; local
61 md = malloc(sizeof(*md));
62 if (!md)
65 md
[all...]
/linux-master/drivers/soundwire/
H A Dmaster.c42 struct sdw_master_device *md = dev_to_sdw_master_device(dev); \
43 return sprintf(buf, format_string, md->bus->prop.field); \
59 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local
63 for (i = 0; i < md->bus->prop.num_clk_freq; i++)
65 md->bus->prop.clk_freq[i]);
75 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local
79 for (i = 0; i < md->bus->prop.num_clk_gears; i++)
81 md->bus->prop.clk_gears[i]);
105 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local
107 kfree(md);
130 struct sdw_master_device *md; local
[all...]
/linux-master/arch/riscv/kernel/
H A Defi.c19 static __init pgprot_t efimem_to_pgprot_map(efi_memory_desc_t *md) argument
21 u64 attr = md->attribute;
22 u32 type = md->type;
46 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument
48 pgprot_t prot = __pgprot(pgprot_val(efimem_to_pgprot_map(md)) &
53 for (i = 0; i < md->num_pages; i++)
54 create_pgd_mapping(mm->pgd, md->virt_addr + i * PAGE_SIZE,
55 md->phys_addr + i * PAGE_SIZE,
62 efi_memory_desc_t *md = data; local
66 if (md
80 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool ignored) argument
[all...]
/linux-master/drivers/net/wwan/t7xx/
H A Dt7xx_modem_ops.c82 struct t7xx_modem *md = t7xx_dev->md; local
88 ctl = md->fsm_ctl;
95 spin_lock_bh(&md->exp_lock);
97 md->exp_id |= int_sta;
98 if (md->exp_id & D2H_INT_EXCEPTION_INIT) {
103 md->exp_id &= ~D2H_INT_EXCEPTION_INIT;
106 } else if (md->exp_id & D2H_INT_PORT_ENUM) {
107 md->exp_id &= ~D2H_INT_PORT_ENUM;
114 if ((md
297 t7xx_md_exception(struct t7xx_modem *md, enum hif_ex_stage stage) argument
316 t7xx_wait_hif_ex_hk_event(struct t7xx_modem *md, int event_id) argument
444 t7xx_core_reset(struct t7xx_modem *md) argument
467 t7xx_core_hk_handler(struct t7xx_modem *md, struct t7xx_sys_info *core_info, struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id, enum t7xx_fsm_event_state err_detect) argument
533 struct t7xx_modem *md = container_of(work, struct t7xx_modem, handshake_work); local
547 struct t7xx_modem *md = container_of(work, struct t7xx_modem, ap_handshake_work); local
559 t7xx_md_event_notify(struct t7xx_modem *md, enum md_event_id evt_id) argument
621 t7xx_md_exception_handshake(struct t7xx_modem *md) argument
647 struct t7xx_modem *md; local
676 struct t7xx_modem *md = t7xx_dev->md; local
701 struct t7xx_modem *md; local
768 struct t7xx_modem *md = t7xx_dev->md; local
[all...]
/linux-master/drivers/net/mdio/
H A Dmdio-mux-bcm6368.c39 struct bcm6368_mdiomux_desc *md = bus->priv; local
43 __raw_writel(0, md->base + MDIOC_REG);
48 if (md->ext_phy)
51 __raw_writel(reg, md->base + MDIOC_REG);
53 ret = __raw_readw(md->base + MDIOD_REG);
61 struct bcm6368_mdiomux_desc *md = bus->priv; local
64 __raw_writel(0, md->base + MDIOC_REG);
69 if (md->ext_phy)
73 __raw_writel(reg, md->base + MDIOC_REG);
82 struct bcm6368_mdiomux_desc *md local
91 struct bcm6368_mdiomux_desc *md; local
158 struct bcm6368_mdiomux_desc *md = platform_get_drvdata(pdev); local
[all...]
H A Dmdio-mux-bcm-iproc.c59 static void mdio_mux_iproc_config(struct iproc_mdiomux_desc *md) argument
65 val = readl(md->base + MDIO_SCAN_CTRL_OFFSET);
67 writel(val, md->base + MDIO_SCAN_CTRL_OFFSET);
69 if (md->core_clk) {
73 divisor = clk_get_rate(md->core_clk) / MDIO_OPERATING_FREQUENCY;
77 writel(val, md->base + MDIO_RATE_ADJ_EXT_OFFSET);
78 writel(val, md->base + MDIO_RATE_ADJ_INT_OFFSET);
138 struct iproc_mdiomux_desc *md = bus->priv; local
141 ret = start_miim_ops(md->base, false, phyid, reg, 0, MDIO_CTRL_READ_OP);
151 struct iproc_mdiomux_desc *md local
165 struct iproc_mdiomux_desc *md = bus->priv; local
180 struct iproc_mdiomux_desc *md = bus->priv; local
195 struct iproc_mdiomux_desc *md = data; local
212 struct iproc_mdiomux_desc *md; local
294 struct iproc_mdiomux_desc *md = platform_get_drvdata(pdev); local
304 struct iproc_mdiomux_desc *md = dev_get_drvdata(dev); local
313 struct iproc_mdiomux_desc *md = dev_get_drvdata(dev); local
[all...]
/linux-master/drivers/clk/qcom/
H A Dclk-regmap-mux-div.c23 int mux_div_set_src_div(struct clk_regmap_mux_div *md, u32 src, u32 div) argument
27 const char *name = clk_hw_get_name(&md->clkr.hw);
29 val = (div << md->hid_shift) | (src << md->src_shift);
30 mask = ((BIT(md->hid_width) - 1) << md->hid_shift) |
31 ((BIT(md->src_width) - 1) << md->src_shift);
33 ret = regmap_update_bits(md->clkr.regmap, CFG_RCGR + md
59 mux_div_get_src_div(struct clk_regmap_mux_div *md, u32 *src, u32 *div) argument
91 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
127 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
165 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
181 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
189 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
197 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
205 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local
[all...]
/linux-master/arch/arm64/kernel/
H A Defi.c17 static bool region_is_misaligned(const efi_memory_desc_t *md) argument
21 return !PAGE_ALIGNED(md->phys_addr) ||
22 !PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT);
30 static __init pteval_t create_mapping_protection(efi_memory_desc_t *md) argument
32 u64 attr = md->attribute;
33 u32 type = md->type;
38 if (region_is_misaligned(md)) {
74 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument
76 pteval_t prot_val = create_mapping_protection(md);
77 bool page_mappings_only = (md
98 const efi_memory_desc_t *md; member in struct:set_perm_data
105 const efi_memory_desc_t *md = spd->md; local
118 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool has_bti) argument
[all...]
/linux-master/drivers/firmware/efi/libstub/
H A Drandomalloc.c17 static unsigned long get_entry_num_slots(efi_memory_desc_t *md, argument
25 if (md->type != EFI_CONVENTIONAL_MEMORY)
29 (md->attribute & EFI_MEMORY_SP))
32 region_end = min(md->phys_addr + md->num_pages * EFI_PAGE_SIZE - 1,
37 first_slot = round_up(max(md->phys_addr, alloc_min), align);
52 #define MD_NUM_SLOTS(md) ((md)->virt_addr)
79 efi_memory_desc_t *md = (void *)map->map + map_offset; local
82 slots = get_entry_num_slots(md, siz
110 efi_memory_desc_t *md = (void *)map->map + map_offset; local
[all...]
/linux-master/tools/build/feature/
H A Dtest-libcrypto.c9 unsigned char md[MD5_DIGEST_LENGTH + SHA_DIGEST_LENGTH]; local
19 EVP_DigestFinal_ex(mdctx, &md[0], &digest_len);
22 SHA1(&dat[0], sizeof(dat), &md[0]);
/linux-master/drivers/firmware/efi/
H A Dmemattr.c60 efi_memory_desc_t *md; local
84 for_each_efi_memory_desc(md) {
85 u64 md_paddr = md->phys_addr;
86 u64 md_size = md->num_pages << EFI_PAGE_SHIFT;
88 if (!(md->attribute & EFI_MEMORY_RUNTIME))
90 if (md->virt_addr == 0 && md->phys_addr != 0) {
107 if (md->type != in->type) {
112 out->virt_addr = in_paddr + (md->virt_addr - md_paddr);
162 efi_memory_desc_t md; local
[all...]
H A Defi-init.c27 static int __init is_memory(efi_memory_desc_t *md) argument
29 if (md->attribute & (EFI_MEMORY_WB|EFI_MEMORY_WT|EFI_MEMORY_WC))
41 efi_memory_desc_t *md; local
43 for_each_efi_memory_desc(md) {
44 if (!(md->attribute & EFI_MEMORY_RUNTIME))
46 if (md->virt_addr == 0)
49 if (md->virt_addr <= addr &&
50 (addr - md->virt_addr) < (md->num_pages << EFI_PAGE_SHIFT))
51 return md
136 is_usable_memory(efi_memory_desc_t *md) argument
160 efi_memory_desc_t *md; local
[all...]
/linux-master/arch/x86/platform/efi/
H A Defi_32.c36 void __init efi_map_region(efi_memory_desc_t *md) argument
42 start_pfn = PFN_DOWN(md->phys_addr);
43 size = md->num_pages << PAGE_SHIFT;
44 end = md->phys_addr + size;
48 va = __va(md->phys_addr);
50 if (!(md->attribute & EFI_MEMORY_WB))
51 set_memory_uc((unsigned long)va, md->num_pages);
53 va = ioremap_cache(md->phys_addr, size);
56 md->virt_addr = (unsigned long)va;
58 pr_err("ioremap of 0x%llX failed!\n", md
87 efi_map_region_fixed(efi_memory_desc_t *md) argument
132 efi_memory_desc_t *md; local
[all...]
H A Dmemmap.c105 * @md: EFI memory descriptor to split
111 int __init efi_memmap_split_count(efi_memory_desc_t *md, struct range *range) argument
117 start = md->phys_addr;
118 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1;
155 efi_memory_desc_t *md; local
181 md = new;
182 start = md->phys_addr;
183 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1;
186 md
[all...]
/linux-master/arch/arm/kernel/
H A Defi.c16 efi_memory_desc_t *md = data; local
19 if (md->attribute & EFI_MEMORY_RO)
21 if (md->attribute & EFI_MEMORY_XP)
28 efi_memory_desc_t *md,
33 base = md->virt_addr;
34 size = md->num_pages << EFI_PAGE_SHIFT;
44 return apply_to_page_range(mm, base, size, set_permissions, md);
49 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument
52 .virtual = md->virt_addr,
53 .pfn = __phys_to_pfn(md
27 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool ignored) argument
[all...]

Completed in 328 milliseconds

12345678910