/linux-master/drivers/md/bcache/ |
H A D | debug.h | 7 struct cache_set; 30 void bch_debug_init_cache_set(struct cache_set *c); 32 static inline void bch_debug_init_cache_set(struct cache_set *c) {}
|
H A D | extents.h | 9 struct cache_set; 12 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k); 13 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k);
|
H A D | alloc.c | 86 void bch_rescale_priorities(struct cache_set *c, int sectors) 480 void bch_bucket_free(struct cache_set *c, struct bkey *k) 488 int __bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, 520 int bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, 565 static struct open_bucket *pick_data_bucket(struct cache_set *c, 606 bool bch_alloc_sectors(struct cache_set *c, 697 void bch_open_buckets_free(struct cache_set *c) 709 int bch_open_buckets_alloc(struct cache_set *c)
|
H A D | bcache.h | 254 struct cache_set *c; 418 struct cache_set *set; 516 struct cache_set { struct 797 static inline size_t sector_to_bucket(struct cache_set *c, sector_t s) 802 static inline sector_t bucket_to_sector(struct cache_set *c, size_t b) 807 static inline sector_t bucket_remainder(struct cache_set *c, sector_t s) 812 static inline size_t PTR_BUCKET_NR(struct cache_set *c, 819 static inline struct bucket *PTR_BUCKET(struct cache_set *c, 833 static inline uint8_t ptr_stale(struct cache_set *c, const struct bkey *k, 839 static inline bool ptr_available(struct cache_set * [all...] |
H A D | btree.c | 128 void bkey_put(struct cache_set *c, struct bkey *k) 157 * See the comment arount cache_set->fill_iter. 583 static struct btree *mca_bucket_alloc(struct cache_set *c, 670 struct cache_set *c = shrink->private_data; 737 struct cache_set *c = shrink->private_data; 748 void bch_btree_cache_free(struct cache_set *c) 796 int bch_btree_cache_alloc(struct cache_set *c) 850 static struct hlist_head *mca_hash(struct cache_set *c, struct bkey *k) 855 static struct btree *mca_find(struct cache_set *c, struct bkey *k) 869 static int mca_cannibalize_lock(struct cache_set * [all...] |
H A D | btree.h | 126 struct cache_set *c; 195 static inline void set_gc_sectors(struct cache_set *c) 200 void bkey_put(struct cache_set *c, struct bkey *k); 231 struct cache_set *c; 268 struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op, 271 struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op, 277 int bch_btree_insert(struct cache_set *c, struct keylist *keys, 280 int bch_gc_thread_start(struct cache_set *c); 281 void bch_initial_gc_finish(struct cache_set *c); 282 void bch_moving_gc(struct cache_set * [all...] |
H A D | debug.c | 163 struct cache_set *c; 209 struct cache_set *c = inode->i_private; 237 void bch_debug_init_cache_set(struct cache_set *c)
|
H A D | extents.c | 47 static bool __ptr_invalid(struct cache_set *c, const struct bkey *k) 68 static const char *bch_ptr_status(struct cache_set *c, const struct bkey *k) 149 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k) 314 struct cache_set *c, 328 struct cache_set *c = container_of(b, struct btree, keys)->c; 480 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k)
|
H A D | features.c | 56 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size) 63 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size) 70 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size)
|
H A D | features.h | 109 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size); 110 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size); 111 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size);
|
H A D | io.c | 17 void bch_bbio_free(struct bio *bio, struct cache_set *c) 24 struct bio *bch_bbio_alloc(struct cache_set *c) 35 void __bch_submit_bbio(struct bio *bio, struct cache_set *c) 46 void bch_submit_bbio(struct bio *bio, struct cache_set *c, 137 void bch_bbio_count_io_errors(struct cache_set *c, struct bio *bio, 167 void bch_bbio_endio(struct cache_set *c, struct bio *bio,
|
H A D | journal.c | 170 int bch_journal_read(struct cache_set *c, struct list_head *list) 293 void bch_journal_mark(struct cache_set *c, struct list_head *list) 339 static bool is_discard_enabled(struct cache_set *s) 349 int bch_journal_replay(struct cache_set *s, struct list_head *list) 416 static void btree_flush_write(struct cache_set *c) 630 static unsigned int free_journal_buckets(struct cache_set *c) 649 static void journal_reclaim(struct cache_set *c) 742 closure_type(c, struct cache_set, journal.io); 751 closure_type(c, struct cache_set, journal.io); 828 closure_type(c, struct cache_set, journa [all...] |
H A D | journal.h | 97 struct cache_set *c; 103 /* Embedded in struct cache_set */ 171 struct cache_set; 175 atomic_t *bch_journal(struct cache_set *c, 179 void bch_journal_mark(struct cache_set *c, struct list_head *list); 180 void bch_journal_meta(struct cache_set *c, struct closure *cl); 181 int bch_journal_read(struct cache_set *c, struct list_head *list); 182 int bch_journal_replay(struct cache_set *c, struct list_head *list); 184 void bch_journal_free(struct cache_set *c); 185 int bch_journal_alloc(struct cache_set * [all...] |
H A D | movinggc.c | 24 struct cache_set *c = container_of(buf, struct cache_set, 126 static void read_moving(struct cache_set *c) 197 void bch_moving_gc(struct cache_set *c) 248 void bch_moving_init_cache_set(struct cache_set *c)
|
H A D | request.c | 91 struct cache_set *c) 324 unsigned int bch_get_congested(const struct cache_set *c) 365 struct cache_set *c = dc->disk.c; 1130 static void quit_max_writeback_rate(struct cache_set *c,
|
H A D | request.h | 7 struct cache_set *c; 36 unsigned int bch_get_congested(const struct cache_set *c);
|
H A D | stats.c | 198 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, 207 void bch_mark_cache_miss_collision(struct cache_set *c, struct bcache_device *d) 215 void bch_mark_sectors_bypassed(struct cache_set *c, struct cached_dev *dc,
|
H A D | stats.h | 40 struct cache_set; 54 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, 56 void bch_mark_cache_miss_collision(struct cache_set *c, 58 void bch_mark_sectors_bypassed(struct cache_set *c,
|
H A D | super.c | 368 closure_type(c, struct cache_set, sb_write); 373 void bcache_write_super(struct cache_set *c) 403 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); 412 closure_type(c, struct cache_set, uuid_write); 417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k, 458 static char *uuid_read(struct cache_set *c, struct jset *j, struct closure *cl) 499 static int __uuid_write(struct cache_set *c) 525 int bch_uuid_write(struct cache_set *c) 535 static struct uuid_entry *uuid_find(struct cache_set * [all...] |
H A D | sysfs.c | 296 struct cache_set *c; 632 static int bch_bset_print_stats(struct cache_set *c, char *buf) 658 static unsigned int bch_root_usage(struct cache_set *c) 682 static size_t bch_cache_size(struct cache_set *c) 695 static unsigned int bch_cache_max_chain(struct cache_set *c) 712 static unsigned int bch_btree_used(struct cache_set *c) 718 static unsigned int bch_average_key_size(struct cache_set *c) 727 struct cache_set *c = container_of(kobj, struct cache_set, kobj); 814 struct cache_set * [all...] |
H A D | writeback.c | 20 static void update_gc_after_writeback(struct cache_set *c) 32 struct cache_set *c = dc->disk.c; 101 struct cache_set *c = dc->disk.c; 160 static bool idle_counter_exceeded(struct cache_set *c) 207 static bool set_at_max_writeback_rate(struct cache_set *c, 247 struct cache_set *c = dc->disk.c; 596 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode, 742 struct cache_set *c = dc->disk.c; 872 static int bch_root_node_dirty_init(struct cache_set *c, 894 * The op may be added to cache_set' [all...] |
H A D | writeback.h | 38 struct cache_set *c; 148 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode,
|
/linux-master/include/trace/events/ |
H A D | bcache.h | 152 TP_PROTO(struct cache_set *c, u64 inode, struct bio *bio, 194 DECLARE_EVENT_CLASS(cache_set, 195 TP_PROTO(struct cache_set *c), 214 DEFINE_EVENT(cache_set, bcache_journal_full, 215 TP_PROTO(struct cache_set *c), 219 DEFINE_EVENT(cache_set, bcache_journal_entry_full, 220 TP_PROTO(struct cache_set *c), 252 DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize, 253 TP_PROTO(struct cache_set *c), 287 DEFINE_EVENT(cache_set, bcache_btree_node_alloc_fai [all...] |