Searched refs:cache_set (Results 1 - 23 of 23) sorted by path

/linux-master/drivers/md/bcache/
H A Ddebug.h7 struct cache_set;
30 void bch_debug_init_cache_set(struct cache_set *c);
32 static inline void bch_debug_init_cache_set(struct cache_set *c) {}
H A Dextents.h9 struct cache_set;
12 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k);
13 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k);
H A Dalloc.c86 void bch_rescale_priorities(struct cache_set *c, int sectors)
480 void bch_bucket_free(struct cache_set *c, struct bkey *k)
488 int __bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve,
520 int bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve,
565 static struct open_bucket *pick_data_bucket(struct cache_set *c,
606 bool bch_alloc_sectors(struct cache_set *c,
697 void bch_open_buckets_free(struct cache_set *c)
709 int bch_open_buckets_alloc(struct cache_set *c)
H A Dbcache.h254 struct cache_set *c;
418 struct cache_set *set;
516 struct cache_set { struct
797 static inline size_t sector_to_bucket(struct cache_set *c, sector_t s)
802 static inline sector_t bucket_to_sector(struct cache_set *c, size_t b)
807 static inline sector_t bucket_remainder(struct cache_set *c, sector_t s)
812 static inline size_t PTR_BUCKET_NR(struct cache_set *c,
819 static inline struct bucket *PTR_BUCKET(struct cache_set *c,
833 static inline uint8_t ptr_stale(struct cache_set *c, const struct bkey *k,
839 static inline bool ptr_available(struct cache_set *
[all...]
H A Dbtree.c128 void bkey_put(struct cache_set *c, struct bkey *k)
157 * See the comment arount cache_set->fill_iter.
583 static struct btree *mca_bucket_alloc(struct cache_set *c,
670 struct cache_set *c = shrink->private_data;
737 struct cache_set *c = shrink->private_data;
748 void bch_btree_cache_free(struct cache_set *c)
796 int bch_btree_cache_alloc(struct cache_set *c)
850 static struct hlist_head *mca_hash(struct cache_set *c, struct bkey *k)
855 static struct btree *mca_find(struct cache_set *c, struct bkey *k)
869 static int mca_cannibalize_lock(struct cache_set *
[all...]
H A Dbtree.h126 struct cache_set *c;
195 static inline void set_gc_sectors(struct cache_set *c)
200 void bkey_put(struct cache_set *c, struct bkey *k);
231 struct cache_set *c;
268 struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op,
271 struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op,
277 int bch_btree_insert(struct cache_set *c, struct keylist *keys,
280 int bch_gc_thread_start(struct cache_set *c);
281 void bch_initial_gc_finish(struct cache_set *c);
282 void bch_moving_gc(struct cache_set *
[all...]
H A Ddebug.c163 struct cache_set *c;
209 struct cache_set *c = inode->i_private;
237 void bch_debug_init_cache_set(struct cache_set *c)
H A Dextents.c47 static bool __ptr_invalid(struct cache_set *c, const struct bkey *k)
68 static const char *bch_ptr_status(struct cache_set *c, const struct bkey *k)
149 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k)
314 struct cache_set *c,
328 struct cache_set *c = container_of(b, struct btree, keys)->c;
480 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k)
H A Dfeatures.c56 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size)
63 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size)
70 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size)
H A Dfeatures.h109 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size);
110 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size);
111 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size);
H A Dio.c17 void bch_bbio_free(struct bio *bio, struct cache_set *c)
24 struct bio *bch_bbio_alloc(struct cache_set *c)
35 void __bch_submit_bbio(struct bio *bio, struct cache_set *c)
46 void bch_submit_bbio(struct bio *bio, struct cache_set *c,
137 void bch_bbio_count_io_errors(struct cache_set *c, struct bio *bio,
167 void bch_bbio_endio(struct cache_set *c, struct bio *bio,
H A Djournal.c170 int bch_journal_read(struct cache_set *c, struct list_head *list)
293 void bch_journal_mark(struct cache_set *c, struct list_head *list)
339 static bool is_discard_enabled(struct cache_set *s)
349 int bch_journal_replay(struct cache_set *s, struct list_head *list)
416 static void btree_flush_write(struct cache_set *c)
630 static unsigned int free_journal_buckets(struct cache_set *c)
649 static void journal_reclaim(struct cache_set *c)
742 closure_type(c, struct cache_set, journal.io);
751 closure_type(c, struct cache_set, journal.io);
828 closure_type(c, struct cache_set, journa
[all...]
H A Djournal.h97 struct cache_set *c;
103 /* Embedded in struct cache_set */
171 struct cache_set;
175 atomic_t *bch_journal(struct cache_set *c,
179 void bch_journal_mark(struct cache_set *c, struct list_head *list);
180 void bch_journal_meta(struct cache_set *c, struct closure *cl);
181 int bch_journal_read(struct cache_set *c, struct list_head *list);
182 int bch_journal_replay(struct cache_set *c, struct list_head *list);
184 void bch_journal_free(struct cache_set *c);
185 int bch_journal_alloc(struct cache_set *
[all...]
H A Dmovinggc.c24 struct cache_set *c = container_of(buf, struct cache_set,
126 static void read_moving(struct cache_set *c)
197 void bch_moving_gc(struct cache_set *c)
248 void bch_moving_init_cache_set(struct cache_set *c)
H A Drequest.c91 struct cache_set *c)
324 unsigned int bch_get_congested(const struct cache_set *c)
365 struct cache_set *c = dc->disk.c;
1130 static void quit_max_writeback_rate(struct cache_set *c,
H A Drequest.h7 struct cache_set *c;
36 unsigned int bch_get_congested(const struct cache_set *c);
H A Dstats.c198 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d,
207 void bch_mark_cache_miss_collision(struct cache_set *c, struct bcache_device *d)
215 void bch_mark_sectors_bypassed(struct cache_set *c, struct cached_dev *dc,
H A Dstats.h40 struct cache_set;
54 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d,
56 void bch_mark_cache_miss_collision(struct cache_set *c,
58 void bch_mark_sectors_bypassed(struct cache_set *c,
H A Dsuper.c368 closure_type(c, struct cache_set, sb_write);
373 void bcache_write_super(struct cache_set *c)
403 struct cache_set *c = container_of(cl, struct cache_set, uuid_write);
412 closure_type(c, struct cache_set, uuid_write);
417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k,
458 static char *uuid_read(struct cache_set *c, struct jset *j, struct closure *cl)
499 static int __uuid_write(struct cache_set *c)
525 int bch_uuid_write(struct cache_set *c)
535 static struct uuid_entry *uuid_find(struct cache_set *
[all...]
H A Dsysfs.c296 struct cache_set *c;
632 static int bch_bset_print_stats(struct cache_set *c, char *buf)
658 static unsigned int bch_root_usage(struct cache_set *c)
682 static size_t bch_cache_size(struct cache_set *c)
695 static unsigned int bch_cache_max_chain(struct cache_set *c)
712 static unsigned int bch_btree_used(struct cache_set *c)
718 static unsigned int bch_average_key_size(struct cache_set *c)
727 struct cache_set *c = container_of(kobj, struct cache_set, kobj);
814 struct cache_set *
[all...]
H A Dwriteback.c20 static void update_gc_after_writeback(struct cache_set *c)
32 struct cache_set *c = dc->disk.c;
101 struct cache_set *c = dc->disk.c;
160 static bool idle_counter_exceeded(struct cache_set *c)
207 static bool set_at_max_writeback_rate(struct cache_set *c,
247 struct cache_set *c = dc->disk.c;
596 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode,
742 struct cache_set *c = dc->disk.c;
872 static int bch_root_node_dirty_init(struct cache_set *c,
894 * The op may be added to cache_set'
[all...]
H A Dwriteback.h38 struct cache_set *c;
148 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode,
/linux-master/include/trace/events/
H A Dbcache.h152 TP_PROTO(struct cache_set *c, u64 inode, struct bio *bio,
194 DECLARE_EVENT_CLASS(cache_set,
195 TP_PROTO(struct cache_set *c),
214 DEFINE_EVENT(cache_set, bcache_journal_full,
215 TP_PROTO(struct cache_set *c),
219 DEFINE_EVENT(cache_set, bcache_journal_entry_full,
220 TP_PROTO(struct cache_set *c),
252 DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize,
253 TP_PROTO(struct cache_set *c),
287 DEFINE_EVENT(cache_set, bcache_btree_node_alloc_fai
[all...]

Completed in 478 milliseconds