/kernel/linux/linux-5.10/drivers/md/bcache/ |
H A D | btree.h | 126 struct cache_set *c; 195 static inline void set_gc_sectors(struct cache_set *c) in set_gc_sectors() 200 void bkey_put(struct cache_set *c, struct bkey *k); 231 struct cache_set *c; 268 struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op, 271 struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op, 277 int bch_btree_insert(struct cache_set *c, struct keylist *keys, 280 int bch_gc_thread_start(struct cache_set *c); 281 void bch_initial_gc_finish(struct cache_set *c); 282 void bch_moving_gc(struct cache_set * [all...] |
H A D | bcache.h | 254 struct cache_set *c; 409 struct cache_set *set; 508 struct cache_set { struct 789 static inline size_t sector_to_bucket(struct cache_set *c, sector_t s) in sector_to_bucket() 794 static inline sector_t bucket_to_sector(struct cache_set *c, size_t b) in bucket_to_sector() 799 static inline sector_t bucket_remainder(struct cache_set *c, sector_t s) in bucket_remainder() 804 static inline struct cache *PTR_CACHE(struct cache_set *c, in PTR_CACHE() 811 static inline size_t PTR_BUCKET_NR(struct cache_set *c, in PTR_BUCKET_NR() 818 static inline struct bucket *PTR_BUCKET(struct cache_set *c, in PTR_BUCKET() 832 static inline uint8_t ptr_stale(struct cache_set * [all...] |
H A D | journal.h | 97 struct cache_set *c; 103 /* Embedded in struct cache_set */ 171 struct cache_set; 175 atomic_t *bch_journal(struct cache_set *c, 179 void bch_journal_mark(struct cache_set *c, struct list_head *list); 180 void bch_journal_meta(struct cache_set *c, struct closure *cl); 181 int bch_journal_read(struct cache_set *c, struct list_head *list); 182 int bch_journal_replay(struct cache_set *c, struct list_head *list); 184 void bch_journal_free(struct cache_set *c); 185 int bch_journal_alloc(struct cache_set * [all...] |
H A D | stats.h | 42 struct cache_set; 56 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, 58 void bch_mark_cache_readahead(struct cache_set *c, struct bcache_device *d); 59 void bch_mark_cache_miss_collision(struct cache_set *c, 61 void bch_mark_sectors_bypassed(struct cache_set *c,
|
H A D | extents.h | 9 struct cache_set; 12 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k); 13 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k);
|
H A D | io.c | 17 void bch_bbio_free(struct bio *bio, struct cache_set *c) in bch_bbio_free() 24 struct bio *bch_bbio_alloc(struct cache_set *c) in bch_bbio_alloc() 34 void __bch_submit_bbio(struct bio *bio, struct cache_set *c) in __bch_submit_bbio() 45 void bch_submit_bbio(struct bio *bio, struct cache_set *c, in bch_submit_bbio() 136 void bch_bbio_count_io_errors(struct cache_set *c, struct bio *bio, in bch_bbio_count_io_errors() 166 void bch_bbio_endio(struct cache_set *c, struct bio *bio, in bch_bbio_endio()
|
H A D | journal.c | 171 int bch_journal_read(struct cache_set *c, struct list_head *list) in bch_journal_read() 294 void bch_journal_mark(struct cache_set *c, struct list_head *list) in bch_journal_mark() 340 static bool is_discard_enabled(struct cache_set *s) in is_discard_enabled() 350 int bch_journal_replay(struct cache_set *s, struct list_head *list) in bch_journal_replay() 417 static void btree_flush_write(struct cache_set *c) in btree_flush_write() 633 static unsigned int free_journal_buckets(struct cache_set *c) in free_journal_buckets() 652 static void journal_reclaim(struct cache_set *c) in journal_reclaim() 745 struct cache_set *c = container_of(cl, struct cache_set, journal.io); 754 struct cache_set * [all...] |
H A D | debug.h | 7 struct cache_set; 30 void bch_debug_init_cache_set(struct cache_set *c); 32 static inline void bch_debug_init_cache_set(struct cache_set *c) {} in bch_debug_init_cache_set()
|
H A D | super.c | 369 struct cache_set *c = container_of(cl, struct cache_set, sb_write); in bcache_write_super_unlock() 374 void bcache_write_super(struct cache_set *c) in bcache_write_super() 405 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); in uuid_endio() 414 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); in uuid_io_unlock() 419 static void uuid_io(struct cache_set *c, int op, unsigned long op_flags, in uuid_io() 460 static char *uuid_read(struct cache_set *c, struct jset *j, struct closure *cl) in uuid_read() 501 static int __uuid_write(struct cache_set * [all...] |
H A D | features.c | 56 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_compat() 63 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_ro_compat() 70 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_incompat()
|
H A D | btree.c | 128 void bkey_put(struct cache_set *c, struct bkey *k) in bkey_put() 157 * See the comment arount cache_set->fill_iter. in bch_btree_node_read_done() 562 static struct btree *mca_bucket_alloc(struct cache_set *c, in mca_bucket_alloc() 649 struct cache_set *c = container_of(shrink, struct cache_set, shrink); in bch_mca_scan() 716 struct cache_set *c = container_of(shrink, struct cache_set, shrink); in bch_mca_count() 727 void bch_btree_cache_free(struct cache_set *c) in bch_btree_cache_free() 775 int bch_btree_cache_alloc(struct cache_set *c) in bch_btree_cache_alloc() 824 static struct hlist_head *mca_hash(struct cache_set * [all...] |
H A D | request.h | 7 struct cache_set *c; 36 unsigned int bch_get_congested(const struct cache_set *c);
|
/kernel/linux/linux-6.6/drivers/md/bcache/ |
H A D | btree.h | 126 struct cache_set *c; 195 static inline void set_gc_sectors(struct cache_set *c) in set_gc_sectors() 200 void bkey_put(struct cache_set *c, struct bkey *k); 231 struct cache_set *c; 268 struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op, 271 struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op, 277 int bch_btree_insert(struct cache_set *c, struct keylist *keys, 280 int bch_gc_thread_start(struct cache_set *c); 281 void bch_initial_gc_finish(struct cache_set *c); 282 void bch_moving_gc(struct cache_set * [all...] |
H A D | bcache.h | 254 struct cache_set *c; 417 struct cache_set *set; 514 struct cache_set { struct 795 static inline size_t sector_to_bucket(struct cache_set *c, sector_t s) in sector_to_bucket() 800 static inline sector_t bucket_to_sector(struct cache_set *c, size_t b) in bucket_to_sector() 805 static inline sector_t bucket_remainder(struct cache_set *c, sector_t s) in bucket_remainder() 810 static inline size_t PTR_BUCKET_NR(struct cache_set *c, in PTR_BUCKET_NR() 817 static inline struct bucket *PTR_BUCKET(struct cache_set *c, in PTR_BUCKET() 831 static inline uint8_t ptr_stale(struct cache_set *c, const struct bkey *k, in ptr_stale() 837 static inline bool ptr_available(struct cache_set * [all...] |
H A D | journal.h | 97 struct cache_set *c; 103 /* Embedded in struct cache_set */ 171 struct cache_set; 175 atomic_t *bch_journal(struct cache_set *c, 179 void bch_journal_mark(struct cache_set *c, struct list_head *list); 180 void bch_journal_meta(struct cache_set *c, struct closure *cl); 181 int bch_journal_read(struct cache_set *c, struct list_head *list); 182 int bch_journal_replay(struct cache_set *c, struct list_head *list); 184 void bch_journal_free(struct cache_set *c); 185 int bch_journal_alloc(struct cache_set * [all...] |
H A D | stats.h | 40 struct cache_set; 54 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, 56 void bch_mark_cache_miss_collision(struct cache_set *c, 58 void bch_mark_sectors_bypassed(struct cache_set *c,
|
H A D | extents.h | 9 struct cache_set; 12 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k); 13 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k);
|
H A D | io.c | 17 void bch_bbio_free(struct bio *bio, struct cache_set *c) in bch_bbio_free() 24 struct bio *bch_bbio_alloc(struct cache_set *c) in bch_bbio_alloc() 35 void __bch_submit_bbio(struct bio *bio, struct cache_set *c) in __bch_submit_bbio() 46 void bch_submit_bbio(struct bio *bio, struct cache_set *c, in bch_submit_bbio() 137 void bch_bbio_count_io_errors(struct cache_set *c, struct bio *bio, in bch_bbio_count_io_errors() 167 void bch_bbio_endio(struct cache_set *c, struct bio *bio, in bch_bbio_endio()
|
H A D | journal.c | 170 int bch_journal_read(struct cache_set *c, struct list_head *list) in bch_journal_read() 293 void bch_journal_mark(struct cache_set *c, struct list_head *list) in bch_journal_mark() 339 static bool is_discard_enabled(struct cache_set *s) in is_discard_enabled() 349 int bch_journal_replay(struct cache_set *s, struct list_head *list) in bch_journal_replay() 416 static void btree_flush_write(struct cache_set *c) in btree_flush_write() 630 static unsigned int free_journal_buckets(struct cache_set *c) in free_journal_buckets() 649 static void journal_reclaim(struct cache_set *c) in journal_reclaim() 742 struct cache_set *c = container_of(cl, struct cache_set, journal.io); 751 struct cache_set * [all...] |
H A D | debug.h | 7 struct cache_set; 30 void bch_debug_init_cache_set(struct cache_set *c); 32 static inline void bch_debug_init_cache_set(struct cache_set *c) {} in bch_debug_init_cache_set()
|
H A D | super.c | 368 struct cache_set *c = container_of(cl, struct cache_set, sb_write); in bcache_write_super_unlock() 373 void bcache_write_super(struct cache_set *c) in bcache_write_super() 403 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); in uuid_endio() 412 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); in uuid_io_unlock() 417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k, in uuid_io() 458 static char *uuid_read(struct cache_set *c, struct jset *j, struct closure *cl) in uuid_read() 499 static int __uuid_write(struct cache_set * [all...] |
H A D | features.c | 56 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_compat() 63 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_ro_compat() 70 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_incompat()
|
H A D | btree.c | 128 void bkey_put(struct cache_set *c, struct bkey *k) in bkey_put() 157 * See the comment arount cache_set->fill_iter. in bch_btree_node_read_done() 583 static struct btree *mca_bucket_alloc(struct cache_set *c, in mca_bucket_alloc() 670 struct cache_set *c = container_of(shrink, struct cache_set, shrink); in bch_mca_scan() 737 struct cache_set *c = container_of(shrink, struct cache_set, shrink); in bch_mca_count() 748 void bch_btree_cache_free(struct cache_set *c) in bch_btree_cache_free() 796 int bch_btree_cache_alloc(struct cache_set *c) in bch_btree_cache_alloc() 845 static struct hlist_head *mca_hash(struct cache_set * [all...] |
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | bcache.h | 152 TP_PROTO(struct cache_set *c, u64 inode, struct bio *bio, 194 DECLARE_EVENT_CLASS(cache_set, 195 TP_PROTO(struct cache_set *c), 214 DEFINE_EVENT(cache_set, bcache_journal_full, 215 TP_PROTO(struct cache_set *c), 219 DEFINE_EVENT(cache_set, bcache_journal_entry_full, 220 TP_PROTO(struct cache_set *c), 252 DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize, 253 TP_PROTO(struct cache_set *c), 287 DEFINE_EVENT(cache_set, bcache_btree_node_alloc_fai [all...] |
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | bcache.h | 152 TP_PROTO(struct cache_set *c, u64 inode, struct bio *bio, 194 DECLARE_EVENT_CLASS(cache_set, 195 TP_PROTO(struct cache_set *c), 214 DEFINE_EVENT(cache_set, bcache_journal_full, 215 TP_PROTO(struct cache_set *c), 219 DEFINE_EVENT(cache_set, bcache_journal_entry_full, 220 TP_PROTO(struct cache_set *c), 252 DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize, 253 TP_PROTO(struct cache_set *c), 287 DEFINE_EVENT(cache_set, bcache_btree_node_alloc_fai [all...] |