/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-cache-target.c | 11 #include "dm-cache-metadata.h" 24 #define DM_MSG_PREFIX "cache" 27 "A percentage of time allocated for copying to and/or from cache"); 35 * cblock: index of a cache block 36 * promotion: movement of a block from origin to cache 37 * demotion: movement of a block from cache to origin 38 * migration: movement of a block between the origin and cache device, 241 * The block size of the device holding cache data must be 256 * dirty. If you lose the cache device you will lose data. 262 * Data is written to both cache an 297 struct cache { global() struct 421 struct cache *cache; global() member 433 writethrough_mode(struct cache *cache) writethrough_mode() argument 438 writeback_mode(struct cache *cache) writeback_mode() argument 443 passthrough_mode(struct cache *cache) passthrough_mode() argument 450 wake_deferred_bio_worker(struct cache *cache) wake_deferred_bio_worker() argument 455 wake_migration_worker(struct cache *cache) wake_migration_worker() argument 465 alloc_prison_cell(struct cache *cache) alloc_prison_cell() argument 470 free_prison_cell(struct cache *cache, struct dm_bio_prison_cell_v2 *cell) free_prison_cell() argument 475 alloc_migration(struct cache *cache) alloc_migration() argument 491 struct cache *cache = mg->cache; free_migration() local 556 defer_bio(struct cache *cache, struct bio *bio) defer_bio() argument 565 defer_bios(struct cache *cache, struct bio_list *bios) defer_bios() argument 577 bio_detain_shared(struct cache *cache, dm_oblock_t oblock, struct bio *bio) bio_detain_shared() argument 608 is_dirty(struct cache *cache, dm_cblock_t b) is_dirty() argument 613 set_dirty(struct cache *cache, dm_cblock_t cblock) set_dirty() argument 625 force_set_dirty(struct cache *cache, dm_cblock_t cblock) force_set_dirty() argument 632 force_clear_dirty(struct cache *cache, dm_cblock_t cblock) force_clear_dirty() argument 644 block_size_is_power_of_two(struct cache *cache) block_size_is_power_of_two() argument 656 oblocks_per_dblock(struct cache *cache) oblocks_per_dblock() argument 668 oblock_to_dblock(struct cache *cache, dm_oblock_t oblock) oblock_to_dblock() argument 674 set_discard(struct cache *cache, dm_dblock_t b) set_discard() argument 684 clear_discard(struct cache *cache, dm_dblock_t b) clear_discard() argument 691 is_discarded(struct cache *cache, dm_dblock_t b) is_discarded() argument 702 is_discarded_oblock(struct cache *cache, dm_oblock_t b) is_discarded_oblock() argument 719 remap_to_origin(struct cache *cache, struct bio *bio) remap_to_origin() argument 724 remap_to_cache(struct cache *cache, struct bio *bio, dm_cblock_t cblock) remap_to_cache() argument 741 check_if_tick_bio_needed(struct cache *cache, struct bio *bio) check_if_tick_bio_needed() argument 755 remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock) remap_to_origin_clear_discard() argument 765 remap_to_cache_dirty(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_cache_dirty() argument 776 get_bio_block(struct cache *cache, struct bio *bio) get_bio_block() argument 788 accountable_bio(struct cache *cache, struct bio *bio) accountable_bio() argument 793 accounted_begin(struct cache *cache, struct bio *bio) accounted_begin() argument 804 accounted_complete(struct cache *cache, struct bio *bio) accounted_complete() argument 811 accounted_request(struct cache *cache, struct bio *bio) accounted_request() argument 819 struct cache *cache = context; issue_op() local 828 remap_to_origin_and_cache(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_origin_and_cache() argument 850 get_cache_mode(struct cache *cache) get_cache_mode() argument 855 cache_device_name(struct cache *cache) cache_device_name() argument 860 notify_mode_switch(struct cache *cache, enum cache_metadata_mode mode) notify_mode_switch() argument 873 set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mode) set_cache_mode() argument 914 abort_transaction(struct cache *cache) abort_transaction() argument 933 metadata_operation_failed(struct cache *cache, const char *op, int r) metadata_operation_failed() argument 943 load_stats(struct cache *cache) load_stats() argument 954 save_stats(struct cache *cache) save_stats() argument 994 inc_io_migrations(struct cache *cache) inc_io_migrations() argument 999 dec_io_migrations(struct cache *cache) dec_io_migrations() argument 1009 calc_discard_block_range(struct cache *cache, struct bio *bio, dm_dblock_t *b, dm_dblock_t *e) calc_discard_block_range() argument 1025 prevent_background_work(struct cache *cache) prevent_background_work() argument 1032 allow_background_work(struct cache *cache) allow_background_work() argument 1039 background_work_begin(struct cache *cache) background_work_begin() argument 1050 background_work_end(struct cache *cache) background_work_end() argument 1059 bio_writes_complete_block(struct cache *cache, struct bio *bio) bio_writes_complete_block() argument 1065 optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block) optimisable_bio() argument 1098 struct cache *cache = mg->cache; copy() local 1114 bio_drop_shared_lock(struct cache *cache, struct bio *bio) bio_drop_shared_lock() argument 1126 struct cache *cache = mg->cache; overwrite_endio() local 1172 struct cache *cache = mg->cache; mg_complete() local 1241 struct cache *cache = mg->cache; mg_update_metadata() local 1343 struct cache *cache = mg->cache; mg_full_copy() local 1397 struct cache *cache = mg->cache; mg_lock_writes() local 1428 mg_start(struct cache *cache, struct policy_work *op, struct bio *bio) mg_start() argument 1457 struct cache *cache = mg->cache; invalidate_complete() local 1479 invalidate_cblock(struct cache *cache, dm_cblock_t cblock) invalidate_cblock() argument 1508 struct cache *cache = mg->cache; invalidate_remove() local 1527 struct cache *cache = mg->cache; invalidate_lock() local 1559 invalidate_start(struct cache *cache, dm_cblock_t cblock, dm_oblock_t oblock, struct bio *bio) invalidate_start() argument 1587 spare_migration_bandwidth(struct cache *cache) spare_migration_bandwidth() argument 1599 inc_hit_counter(struct cache *cache, struct bio *bio) inc_hit_counter() argument 1605 inc_miss_counter(struct cache *cache, struct bio *bio) inc_miss_counter() argument 1613 map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block, bool *commit_needed) map_bio() argument 1728 process_bio(struct cache *cache, struct bio *bio) process_bio() argument 1741 commit(struct cache *cache, bool clean_shutdown) commit() argument 1761 struct cache *cache = context; commit_op() local 1771 process_flush_bio(struct cache *cache, struct bio *bio) process_flush_bio() argument 1784 process_discard_bio(struct cache *cache, struct bio *bio) process_discard_bio() argument 1810 struct cache *cache = container_of(ws, struct cache, deferred_bio_worker); process_deferred_bios() local 1844 requeue_deferred_bios(struct cache *cache) requeue_deferred_bios() argument 1866 struct cache *cache = container_of(to_delayed_work(ws), struct cache, waker); do_waker() local 1878 struct cache *cache = container_of(ws, struct cache, migration_worker); check_migrations() local 1912 destroy(struct cache *cache) destroy() argument 1960 struct cache *cache = ti->private; cache_dtr() local 2268 process_config_option(struct cache *cache, const char *key, const char *value) process_config_option() argument 2283 set_config_value(struct cache *cache, const char *key, const char *value) set_config_value() argument 2296 set_config_values(struct cache *cache, int argc, const char **argv) set_config_values() argument 2317 create_cache_policy(struct cache *cache, struct cache_args *ca, char **error) create_cache_policy() argument 2360 set_cache_size(struct cache *cache, dm_cblock_t size) set_cache_size() argument 2379 struct cache *cache; cache_create() local 2572 copy_ctr_args(struct cache *cache, int argc, const char **argv) copy_ctr_args() argument 2600 struct cache *cache = NULL; cache_ctr() local 2633 struct cache *cache = ti->private; cache_map() local 2665 struct cache *cache = ti->private; cache_end_io() local 2683 write_dirty_bitset(struct cache *cache) write_dirty_bitset() argument 2697 write_discard_bitset(struct cache *cache) write_discard_bitset() argument 2724 write_hints(struct cache *cache) write_hints() argument 2743 sync_metadata(struct cache *cache) sync_metadata() argument 2775 struct cache *cache = ti->private; cache_postsuspend() local 2797 struct cache *cache = context; load_mapping() local 2815 struct cache *cache; global() member 2825 discard_load_info_init(struct cache *cache, struct discard_load_info *li) discard_load_info_init() argument 2892 get_cache_dev_size(struct cache *cache) get_cache_dev_size() argument 2899 can_resize(struct cache *cache, dm_cblock_t new_size) can_resize() argument 2925 resize_cache_dev(struct cache *cache, dm_cblock_t new_size) resize_cache_dev() argument 2944 struct cache *cache = ti->private; cache_preresume() local 3005 struct cache *cache = ti->private; cache_resume() local 3012 emit_flags(struct cache *cache, char *result, unsigned int maxlen, ssize_t *sz_ptr) emit_flags() argument 3065 struct cache *cache = ti->private; cache_status() local 3196 parse_cblock_range(struct cache *cache, const char *str, struct cblock_range *result) parse_cblock_range() argument 3233 validate_cblock_range(struct cache *cache, struct cblock_range *range) validate_cblock_range() argument 3265 request_invalidation(struct cache *cache, struct cblock_range *range) request_invalidation() argument 3287 process_invalidate_cblocks_message(struct cache *cache, unsigned int count, const char **cblock_ranges) process_invalidate_cblocks_message() argument 3331 struct cache *cache = ti->private; cache_message() local 3355 struct cache *cache = ti->private; cache_iterate_devices() local 3368 disable_passdown_if_not_supported(struct cache *cache) disable_passdown_if_not_supported() argument 3390 set_discard_limits(struct cache *cache, struct queue_limits *limits) set_discard_limits() argument 3416 struct cache *cache = ti->private; cache_io_hints() local [all...] |
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm-cache-target.c | 10 #include "dm-cache-metadata.h" 22 #define DM_MSG_PREFIX "cache" 25 "A percentage of time allocated for copying to and/or from cache"); 33 * cblock: index of a cache block 34 * promotion: movement of a block from origin to cache 35 * demotion: movement of a block from cache to origin 36 * migration: movement of a block between the origin and cache device, 310 * The block size of the device holding cache data must be 325 * dirty. If you lose the cache device you will lose data. 331 * Data is written to both cache an 366 struct cache { global() struct 490 struct cache *cache; global() member 502 writethrough_mode(struct cache *cache) writethrough_mode() argument 507 writeback_mode(struct cache *cache) writeback_mode() argument 512 passthrough_mode(struct cache *cache) passthrough_mode() argument 519 wake_deferred_bio_worker(struct cache *cache) wake_deferred_bio_worker() argument 524 wake_migration_worker(struct cache *cache) wake_migration_worker() argument 534 alloc_prison_cell(struct cache *cache) alloc_prison_cell() argument 539 free_prison_cell(struct cache *cache, struct dm_bio_prison_cell_v2 *cell) free_prison_cell() argument 544 alloc_migration(struct cache *cache) alloc_migration() argument 560 struct cache *cache = mg->cache; free_migration() local 622 defer_bio(struct cache *cache, struct bio *bio) defer_bio() argument 631 defer_bios(struct cache *cache, struct bio_list *bios) defer_bios() argument 643 bio_detain_shared(struct cache *cache, dm_oblock_t oblock, struct bio *bio) bio_detain_shared() argument 674 is_dirty(struct cache *cache, dm_cblock_t b) is_dirty() argument 679 set_dirty(struct cache *cache, dm_cblock_t cblock) set_dirty() argument 691 force_set_dirty(struct cache *cache, dm_cblock_t cblock) force_set_dirty() argument 698 force_clear_dirty(struct cache *cache, dm_cblock_t cblock) force_clear_dirty() argument 710 block_size_is_power_of_two(struct cache *cache) block_size_is_power_of_two() argument 722 oblocks_per_dblock(struct cache *cache) oblocks_per_dblock() argument 734 oblock_to_dblock(struct cache *cache, dm_oblock_t oblock) oblock_to_dblock() argument 740 set_discard(struct cache *cache, dm_dblock_t b) set_discard() argument 750 clear_discard(struct cache *cache, dm_dblock_t b) clear_discard() argument 757 is_discarded(struct cache *cache, dm_dblock_t b) is_discarded() argument 767 is_discarded_oblock(struct cache *cache, dm_oblock_t b) is_discarded_oblock() argument 781 remap_to_origin(struct cache *cache, struct bio *bio) remap_to_origin() argument 786 remap_to_cache(struct cache *cache, struct bio *bio, dm_cblock_t cblock) remap_to_cache() argument 803 check_if_tick_bio_needed(struct cache *cache, struct bio *bio) check_if_tick_bio_needed() argument 817 __remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock, bool bio_has_pbd) __remap_to_origin_clear_discard() argument 827 remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock) remap_to_origin_clear_discard() argument 834 remap_to_cache_dirty(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_cache_dirty() argument 845 get_bio_block(struct cache *cache, struct bio *bio) get_bio_block() argument 857 accountable_bio(struct cache *cache, struct bio *bio) accountable_bio() argument 862 accounted_begin(struct cache *cache, struct bio *bio) accounted_begin() argument 873 accounted_complete(struct cache *cache, struct bio *bio) accounted_complete() argument 880 accounted_request(struct cache *cache, struct bio *bio) accounted_request() argument 888 struct cache *cache = context; issue_op() local 896 remap_to_origin_and_cache(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_origin_and_cache() argument 917 get_cache_mode(struct cache *cache) get_cache_mode() argument 922 cache_device_name(struct cache *cache) cache_device_name() argument 927 notify_mode_switch(struct cache *cache, enum cache_metadata_mode mode) notify_mode_switch() argument 940 set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mode) set_cache_mode() argument 981 abort_transaction(struct cache *cache) abort_transaction() argument 1000 metadata_operation_failed(struct cache *cache, const char *op, int r) metadata_operation_failed() argument 1010 load_stats(struct cache *cache) load_stats() argument 1021 save_stats(struct cache *cache) save_stats() argument 1060 inc_io_migrations(struct cache *cache) inc_io_migrations() argument 1065 dec_io_migrations(struct cache *cache) dec_io_migrations() argument 1075 calc_discard_block_range(struct cache *cache, struct bio *bio, dm_dblock_t *b, dm_dblock_t *e) calc_discard_block_range() argument 1091 prevent_background_work(struct cache *cache) prevent_background_work() argument 1098 allow_background_work(struct cache *cache) allow_background_work() argument 1105 background_work_begin(struct cache *cache) background_work_begin() argument 1116 background_work_end(struct cache *cache) background_work_end() argument 1125 bio_writes_complete_block(struct cache *cache, struct bio *bio) bio_writes_complete_block() argument 1131 optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block) optimisable_bio() argument 1163 struct cache *cache = mg->cache; copy() local 1179 bio_drop_shared_lock(struct cache *cache, struct bio *bio) bio_drop_shared_lock() argument 1191 struct cache *cache = mg->cache; overwrite_endio() local 1237 struct cache *cache = mg->cache; mg_complete() local 1305 struct cache *cache = mg->cache; mg_update_metadata() local 1407 struct cache *cache = mg->cache; mg_full_copy() local 1460 struct cache *cache = mg->cache; mg_lock_writes() local 1491 mg_start(struct cache *cache, struct policy_work *op, struct bio *bio) mg_start() argument 1518 struct cache *cache = mg->cache; invalidate_complete() local 1539 invalidate_cblock(struct cache *cache, dm_cblock_t cblock) invalidate_cblock() argument 1566 struct cache *cache = mg->cache; invalidate_remove() local 1585 struct cache *cache = mg->cache; invalidate_lock() local 1617 invalidate_start(struct cache *cache, dm_cblock_t cblock, dm_oblock_t oblock, struct bio *bio) invalidate_start() argument 1643 spare_migration_bandwidth(struct cache *cache) spare_migration_bandwidth() argument 1655 inc_hit_counter(struct cache *cache, struct bio *bio) inc_hit_counter() argument 1661 inc_miss_counter(struct cache *cache, struct bio *bio) inc_miss_counter() argument 1669 map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block, bool *commit_needed) map_bio() argument 1784 process_bio(struct cache *cache, struct bio *bio) process_bio() argument 1797 commit(struct cache *cache, bool clean_shutdown) commit() argument 1817 struct cache *cache = context; commit_op() local 1827 process_flush_bio(struct cache *cache, struct bio *bio) process_flush_bio() argument 1840 process_discard_bio(struct cache *cache, struct bio *bio) process_discard_bio() argument 1864 struct cache *cache = container_of(ws, struct cache, deferred_bio_worker); process_deferred_bios() local 1897 requeue_deferred_bios(struct cache *cache) requeue_deferred_bios() argument 1919 struct cache *cache = container_of(to_delayed_work(ws), struct cache, waker); do_waker() local 1931 struct cache *cache = container_of(ws, struct cache, migration_worker); check_migrations() local 1963 destroy(struct cache *cache) destroy() argument 2011 struct cache *cache = ti->private; cache_dtr() local 2320 process_config_option(struct cache *cache, const char *key, const char *value) process_config_option() argument 2335 set_config_value(struct cache *cache, const char *key, const char *value) set_config_value() argument 2348 set_config_values(struct cache *cache, int argc, const char **argv) set_config_values() argument 2369 create_cache_policy(struct cache *cache, struct cache_args *ca, char **error) create_cache_policy() argument 2412 set_cache_size(struct cache *cache, dm_cblock_t size) set_cache_size() argument 2431 struct cache *cache; cache_create() local 2623 copy_ctr_args(struct cache *cache, int argc, const char **argv) copy_ctr_args() argument 2651 struct cache *cache = NULL; cache_ctr() local 2684 struct cache *cache = ti->private; cache_map() local 2716 struct cache *cache = ti->private; cache_end_io() local 2734 write_dirty_bitset(struct cache *cache) write_dirty_bitset() argument 2748 write_discard_bitset(struct cache *cache) write_discard_bitset() argument 2775 write_hints(struct cache *cache) write_hints() argument 2794 sync_metadata(struct cache *cache) sync_metadata() argument 2826 struct cache *cache = ti->private; cache_postsuspend() local 2849 struct cache *cache = context; load_mapping() local 2871 struct cache *cache; global() member 2881 discard_load_info_init(struct cache *cache, struct discard_load_info *li) discard_load_info_init() argument 2948 get_cache_dev_size(struct cache *cache) get_cache_dev_size() argument 2955 can_resize(struct cache *cache, dm_cblock_t new_size) can_resize() argument 2981 resize_cache_dev(struct cache *cache, dm_cblock_t new_size) resize_cache_dev() argument 3000 struct cache *cache = ti->private; cache_preresume() local 3061 struct cache *cache = ti->private; cache_resume() local 3068 emit_flags(struct cache *cache, char *result, unsigned maxlen, ssize_t *sz_ptr) emit_flags() argument 3121 struct cache *cache = ti->private; cache_status() local 3228 parse_cblock_range(struct cache *cache, const char *str, struct cblock_range *result) parse_cblock_range() argument 3265 validate_cblock_range(struct cache *cache, struct cblock_range *range) validate_cblock_range() argument 3297 request_invalidation(struct cache *cache, struct cblock_range *range) request_invalidation() argument 3319 process_invalidate_cblocks_message(struct cache *cache, unsigned count, const char **cblock_ranges) process_invalidate_cblocks_message() argument 3363 struct cache *cache = ti->private; cache_message() local 3387 struct cache *cache = ti->private; cache_iterate_devices() local 3407 disable_passdown_if_not_supported(struct cache *cache) disable_passdown_if_not_supported() argument 3430 set_discard_limits(struct cache *cache, struct queue_limits *limits) set_discard_limits() argument 3456 struct cache *cache = ti->private; cache_io_hints() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
H A D | cacheinfo.c | 3 * Processor cache information made available to userspace via sysfs; 28 * - a "cache" kobject for the top-level directory 29 * - a list of "index" objects representing the cpu's local cache hierarchy 32 struct kobject *kobj; /* bare (not embedded) kobject for cache 37 /* "index" object: each cpu's cache directory has an index 38 * subdirectory corresponding to a cache object associated with the 44 struct cache *cache; member 48 * cache type */ 53 /* Allow for both [di]-cache 118 struct cache { global() struct 139 cache_type_string(const struct cache *cache) cache_type_string() argument 144 cache_init(struct cache *cache, int type, int level, struct device_node *ofnode) cache_init() argument 156 struct cache *cache; new_cache() local 165 release_cache_debugcheck(struct cache *cache) release_cache_debugcheck() argument 178 release_cache(struct cache *cache) release_cache() argument 192 cache_cpu_set(struct cache *cache, int cpu) cache_cpu_set() argument 206 cache_size(const struct cache *cache, unsigned int *ret) cache_size() argument 221 cache_size_kb(const struct cache *cache, unsigned int *ret) cache_size_kb() argument 233 cache_get_line_size(const struct cache *cache, unsigned int *ret) cache_get_line_size() argument 256 cache_nr_sets(const struct cache *cache, unsigned int *ret) cache_nr_sets() argument 271 cache_associativity(const struct cache *cache, unsigned int *ret) cache_associativity() argument 303 cache_find_first_sibling(struct cache *cache) cache_find_first_sibling() argument 321 struct cache *cache = NULL; cache_lookup_by_node() local 387 struct cache *cache; cache_do_one_devnode() local 400 struct cache *cache; cache_lookup_or_instantiate() local 434 do_subsidiary_caches_debugcheck(struct cache *cache) do_subsidiary_caches_debugcheck() argument 446 do_subsidiary_caches(struct cache *cache) do_subsidiary_caches() argument 556 struct cache *cache; size_show() local 573 struct cache *cache; line_size_show() local 589 struct cache *cache; nr_sets_show() local 605 struct cache *cache; associativity_show() local 620 struct cache *cache; type_show() local 633 struct cache *cache; level_show() local 659 get_big_core_shared_cpu_map(int cpu, struct cache *cache) get_big_core_shared_cpu_map() argument 671 struct cache *cache; show_shared_cpumap() local 739 struct cache *cache; cacheinfo_create_index_opt_attrs() local 777 cacheinfo_create_index_dir(struct cache *cache, int index, struct cache_dir *cache_dir) cacheinfo_create_index_dir() argument 806 struct cache *cache; cacheinfo_sysfs_populate() local 823 struct cache *cache; cacheinfo_cpu_online() local 840 struct cache *cache; cache_lookup_by_cpu() local 880 cache_cpu_clear(struct cache *cache, int cpu) cache_cpu_clear() argument 904 struct cache *cache; cacheinfo_cpu_offline() local [all...] |
/kernel/linux/linux-6.6/arch/powerpc/kernel/ |
H A D | cacheinfo.c | 3 * Processor cache information made available to userspace via sysfs; 27 * - a "cache" kobject for the top-level directory 28 * - a list of "index" objects representing the cpu's local cache hierarchy 31 struct kobject *kobj; /* bare (not embedded) kobject for cache 36 /* "index" object: each cpu's cache directory has an index 37 * subdirectory corresponding to a cache object associated with the 43 struct cache *cache; member 47 * cache type */ 52 /* Allow for both [di]-cache 117 struct cache { global() struct 139 cache_type_string(const struct cache *cache) cache_type_string() argument 144 cache_init(struct cache *cache, int type, int level, struct device_node *ofnode, int group_id) cache_init() argument 158 struct cache *cache; new_cache() local 167 release_cache_debugcheck(struct cache *cache) release_cache_debugcheck() argument 180 release_cache(struct cache *cache) release_cache() argument 194 cache_cpu_set(struct cache *cache, int cpu) cache_cpu_set() argument 208 cache_size(const struct cache *cache, unsigned int *ret) cache_size() argument 223 cache_size_kb(const struct cache *cache, unsigned int *ret) cache_size_kb() argument 235 cache_get_line_size(const struct cache *cache, unsigned int *ret) cache_get_line_size() argument 258 cache_nr_sets(const struct cache *cache, unsigned int *ret) cache_nr_sets() argument 273 cache_associativity(const struct cache *cache, unsigned int *ret) cache_associativity() argument 305 cache_find_first_sibling(struct cache *cache) cache_find_first_sibling() argument 326 struct cache *cache = NULL; cache_lookup_by_node_group() local 394 struct cache *cache; cache_do_one_devnode() local 408 struct cache *cache; cache_lookup_or_instantiate() local 442 do_subsidiary_caches_debugcheck(struct cache *cache) do_subsidiary_caches_debugcheck() argument 477 do_subsidiary_caches(struct cache *cache, unsigned int cpu_id) do_subsidiary_caches() argument 592 struct cache *cache; size_show() local 609 struct cache *cache; line_size_show() local 625 struct cache *cache; nr_sets_show() local 641 struct cache *cache; associativity_show() local 656 struct cache *cache; type_show() local 669 struct cache *cache; level_show() local 684 struct cache *cache; show_shared_cpumap() local 747 struct cache *cache; cacheinfo_create_index_opt_attrs() local 785 cacheinfo_create_index_dir(struct cache *cache, int index, struct cache_dir *cache_dir) cacheinfo_create_index_dir() argument 814 struct cache *cache; cacheinfo_sysfs_populate() local 831 struct cache *cache; cacheinfo_cpu_online() local 848 struct cache *cache; cache_lookup_by_cpu() local 890 cache_cpu_clear(struct cache *cache, int cpu) cache_cpu_clear() argument 914 struct cache *cache; cacheinfo_cpu_offline() local [all...] |
/kernel/linux/linux-5.10/fs/cachefiles/ |
H A D | bind.c | 2 /* Bind and unbind a cache from the filesystem backing it 25 * bind a directory as a cache 27 int cachefiles_daemon_bind(struct cachefiles_cache *cache, char *args) in cachefiles_daemon_bind() argument 30 cache->frun_percent, in cachefiles_daemon_bind() 31 cache->fcull_percent, in cachefiles_daemon_bind() 32 cache->fstop_percent, in cachefiles_daemon_bind() 33 cache->brun_percent, in cachefiles_daemon_bind() 34 cache->bcull_percent, in cachefiles_daemon_bind() 35 cache->bstop_percent, in cachefiles_daemon_bind() 39 ASSERT(cache in cachefiles_daemon_bind() 81 cachefiles_daemon_add_cache(struct cachefiles_cache *cache) cachefiles_daemon_add_cache() argument 255 cachefiles_daemon_unbind(struct cachefiles_cache *cache) cachefiles_daemon_unbind() argument [all...] |
H A D | daemon.c | 59 int (*handler)(struct cachefiles_cache *cache, char *args); 85 struct cachefiles_cache *cache; in cachefiles_daemon_open() local 97 /* allocate a cache record */ in cachefiles_daemon_open() 98 cache = kzalloc(sizeof(struct cachefiles_cache), GFP_KERNEL); in cachefiles_daemon_open() 99 if (!cache) { in cachefiles_daemon_open() 104 mutex_init(&cache->daemon_mutex); in cachefiles_daemon_open() 105 cache->active_nodes = RB_ROOT; in cachefiles_daemon_open() 106 rwlock_init(&cache->active_lock); in cachefiles_daemon_open() 107 init_waitqueue_head(&cache->daemon_pollwq); in cachefiles_daemon_open() 114 cache in cachefiles_daemon_open() 131 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_release() local 160 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_read() local 217 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_write() local 293 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_poll() local 312 cachefiles_daemon_range_error(struct cachefiles_cache *cache, char *args) cachefiles_daemon_range_error() argument 324 cachefiles_daemon_frun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_frun() argument 348 cachefiles_daemon_fcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fcull() argument 372 cachefiles_daemon_fstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fstop() argument 396 cachefiles_daemon_brun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_brun() argument 420 cachefiles_daemon_bcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bcull() argument 444 cachefiles_daemon_bstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bstop() argument 468 cachefiles_daemon_dir(struct cachefiles_cache *cache, char *args) cachefiles_daemon_dir() argument 496 cachefiles_daemon_secctx(struct cachefiles_cache *cache, char *args) cachefiles_daemon_secctx() argument 524 cachefiles_daemon_tag(struct cachefiles_cache *cache, char *args) cachefiles_daemon_tag() argument 550 cachefiles_daemon_cull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_cull() argument 599 cachefiles_daemon_debug(struct cachefiles_cache *cache, char *args) cachefiles_daemon_debug() argument 622 cachefiles_daemon_inuse(struct cachefiles_cache *cache, char *args) cachefiles_daemon_inuse() argument 671 cachefiles_has_space(struct cachefiles_cache *cache, unsigned fnr, unsigned bnr) cachefiles_has_space() argument [all...] |
/kernel/linux/linux-6.6/fs/cachefiles/ |
H A D | daemon.c | 61 int (*handler)(struct cachefiles_cache *cache, char *args); 86 * Prepare a cache for caching. 90 struct cachefiles_cache *cache; in cachefiles_daemon_open() local 102 /* allocate a cache record */ in cachefiles_daemon_open() 103 cache = kzalloc(sizeof(struct cachefiles_cache), GFP_KERNEL); in cachefiles_daemon_open() 104 if (!cache) { in cachefiles_daemon_open() 109 mutex_init(&cache->daemon_mutex); in cachefiles_daemon_open() 110 init_waitqueue_head(&cache->daemon_pollwq); in cachefiles_daemon_open() 111 INIT_LIST_HEAD(&cache->volumes); in cachefiles_daemon_open() 112 INIT_LIST_HEAD(&cache in cachefiles_daemon_open() 135 cachefiles_flush_reqs(struct cachefiles_cache *cache) cachefiles_flush_reqs() argument 168 cachefiles_put_unbind_pincount(struct cachefiles_cache *cache) cachefiles_put_unbind_pincount() argument 177 cachefiles_get_unbind_pincount(struct cachefiles_cache *cache) cachefiles_get_unbind_pincount() argument 187 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_release() local 208 cachefiles_do_daemon_read(struct cachefiles_cache *cache, char __user *_buffer, size_t buflen) cachefiles_do_daemon_read() argument 259 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_read() local 281 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_write() local 357 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_poll() local 381 cachefiles_daemon_range_error(struct cachefiles_cache *cache, char *args) cachefiles_daemon_range_error() argument 393 cachefiles_daemon_frun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_frun() argument 417 cachefiles_daemon_fcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fcull() argument 441 cachefiles_daemon_fstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fstop() argument 465 cachefiles_daemon_brun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_brun() argument 489 cachefiles_daemon_bcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bcull() argument 513 cachefiles_daemon_bstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bstop() argument 537 cachefiles_daemon_dir(struct cachefiles_cache *cache, char *args) cachefiles_daemon_dir() argument 565 cachefiles_daemon_secctx(struct cachefiles_cache *cache, char *args) cachefiles_daemon_secctx() argument 593 cachefiles_daemon_tag(struct cachefiles_cache *cache, char *args) cachefiles_daemon_tag() argument 619 cachefiles_daemon_cull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_cull() argument 667 cachefiles_daemon_debug(struct cachefiles_cache *cache, char *args) cachefiles_daemon_debug() argument 690 cachefiles_daemon_inuse(struct cachefiles_cache *cache, char *args) cachefiles_daemon_inuse() argument 737 cachefiles_daemon_bind(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bind() argument 798 cachefiles_daemon_unbind(struct cachefiles_cache *cache) cachefiles_daemon_unbind() argument [all...] |
H A D | cache.c | 2 /* Manage high-level VFS aspects of a cache. 14 * Bring a cache online. 16 int cachefiles_add_cache(struct cachefiles_cache *cache) in cachefiles_add_cache() argument 27 cache_cookie = fscache_acquire_cache(cache->tag); in cachefiles_add_cache() 32 ret = cachefiles_get_security_ID(cache); in cachefiles_add_cache() 36 cachefiles_begin_secure(cache, &saved_cred); in cachefiles_add_cache() 38 /* look up the directory at the root of the cache */ in cachefiles_add_cache() 39 ret = kern_path(cache->rootdirname, LOOKUP_DIRECTORY, &path); in cachefiles_add_cache() 43 cache->mnt = path.mnt; in cachefiles_add_cache() 48 pr_warn("File cache o in cachefiles_add_cache() 184 cachefiles_has_space(struct cachefiles_cache *cache, unsigned fnr, unsigned bnr, enum cachefiles_has_space_for reason) cachefiles_has_space() argument 287 cachefiles_withdraw_objects(struct cachefiles_cache *cache) cachefiles_withdraw_objects() argument 317 cachefiles_withdraw_volumes(struct cachefiles_cache *cache) cachefiles_withdraw_volumes() argument 343 cachefiles_sync_cache(struct cachefiles_cache *cache) cachefiles_sync_cache() argument 367 cachefiles_withdraw_cache(struct cachefiles_cache *cache) cachefiles_withdraw_cache() argument [all...] |
/kernel/linux/linux-6.6/fs/fscache/ |
H A D | cache.c | 2 /* FS-Cache cache handling 22 * Allocate a cache cookie. 26 struct fscache_cache *cache; in fscache_alloc_cache() local 28 cache = kzalloc(sizeof(*cache), GFP_KERNEL); in fscache_alloc_cache() 29 if (cache) { in fscache_alloc_cache() 31 cache->name = kstrdup(name, GFP_KERNEL); in fscache_alloc_cache() 32 if (!cache->name) { in fscache_alloc_cache() 33 kfree(cache); in fscache_alloc_cache() 37 refcount_set(&cache in fscache_alloc_cache() 44 fscache_get_cache_maybe(struct fscache_cache *cache, enum fscache_cache_trace where) fscache_get_cache_maybe() argument 61 struct fscache_cache *candidate, *cache, *unnamed = NULL; fscache_lookup_cache() local 151 struct fscache_cache *cache; fscache_acquire_cache() local 179 fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where) fscache_put_cache() argument 209 fscache_relinquish_cache(struct fscache_cache *cache) fscache_relinquish_cache() argument 234 fscache_add_cache(struct fscache_cache *cache, const struct fscache_cache_ops *ops, void *cache_priv) fscache_add_cache() argument 289 fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) fscache_begin_cache_access() argument 315 fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) fscache_end_cache_access() argument 337 fscache_io_error(struct fscache_cache *cache) fscache_io_error() argument 355 fscache_withdraw_cache(struct fscache_cache *cache) fscache_withdraw_cache() argument 382 struct fscache_cache *cache; fscache_caches_seq_show() local [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | swap_slots.c | 3 * Manage cache of swap slots to be used for and returned from 25 * The swap slots cache is protected by a mutex instead of 42 /* Serialize swap slots cache enable/disable operations */ 107 /* if global pool of slot caches too low, deactivate cache */ in check_cache_active() 116 struct swap_slots_cache *cache; in alloc_swap_slot_cache() local 137 cache = &per_cpu(swp_slots, cpu); in alloc_swap_slot_cache() 138 if (cache->slots || cache->slots_ret) { in alloc_swap_slot_cache() 139 /* cache already allocated */ in alloc_swap_slot_cache() 148 if (!cache in alloc_swap_slot_cache() 172 struct swap_slots_cache *cache; drain_slots_cache_cpu() local 261 refill_swap_slots_cache(struct swap_slots_cache *cache) refill_swap_slots_cache() argument 276 struct swap_slots_cache *cache; free_swap_slot() local 309 struct swap_slots_cache *cache; get_swap_page() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | swap_slots.c | 3 * Manage cache of swap slots to be used for and returned from 25 * The swap slots cache is protected by a mutex instead of 43 /* Serialize swap slots cache enable/disable operations */ 106 /* if global pool of slot caches too low, deactivate cache */ in check_cache_active() 115 struct swap_slots_cache *cache; in alloc_swap_slot_cache() local 136 cache = &per_cpu(swp_slots, cpu); in alloc_swap_slot_cache() 137 if (cache->slots || cache->slots_ret) { in alloc_swap_slot_cache() 138 /* cache already allocated */ in alloc_swap_slot_cache() 147 if (!cache in alloc_swap_slot_cache() 171 struct swap_slots_cache *cache; drain_slots_cache_cpu() local 259 refill_swap_slots_cache(struct swap_slots_cache *cache) refill_swap_slots_cache() argument 274 struct swap_slots_cache *cache; free_swap_slot() local 305 struct swap_slots_cache *cache; folio_alloc_swap() local [all...] |
/kernel/linux/linux-5.10/fs/btrfs/tests/ |
H A D | free-space-tests.c | 10 #include "../free-space-cache.h" 20 static int test_extents(struct btrfs_block_group *cache) in test_extents() argument 27 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents() 33 ret = btrfs_remove_free_space(cache, 0, SZ_4M); in test_extents() 39 if (test_check_exists(cache, 0, SZ_4M)) { in test_extents() 45 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents() 51 ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); in test_extents() 57 ret = btrfs_remove_free_space(cache, 0, SZ_1M); in test_extents() 63 ret = btrfs_remove_free_space(cache, SZ_2M, 4096); in test_extents() 69 if (test_check_exists(cache, in test_extents() 90 test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps() argument 158 test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps_and_extents() argument 333 check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) check_num_extents_and_bitmaps() argument 353 check_cache_empty(struct btrfs_block_group *cache) check_cache_empty() argument 395 test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) test_steal_space_from_bitmap_to_extent() argument 831 struct btrfs_block_group *cache; btrfs_test_free_space_cache() local [all...] |
H A D | free-space-tree-tests.c | 21 struct btrfs_block_group *cache, in __check_free_space_extents() 34 info = search_free_space_info(trans, cache, path, 0); in __check_free_space_extents() 51 end = cache->start + cache->length; in __check_free_space_extents() 59 bit = free_space_test_bit(cache, path, offset); in __check_free_space_extents() 108 struct btrfs_block_group *cache, in check_free_space_extents() 117 info = search_free_space_info(trans, cache, path, 0); in check_free_space_extents() 126 ret = __check_free_space_extents(trans, fs_info, cache, path, extents, in check_free_space_extents() 133 ret = convert_free_space_to_extents(trans, cache, path); in check_free_space_extents() 139 ret = convert_free_space_to_bitmaps(trans, cache, pat in check_free_space_extents() 19 __check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) __check_free_space_extents() argument 106 check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) check_free_space_extents() argument 149 test_empty_block_group(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_empty_block_group() argument 163 test_remove_all(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_all() argument 184 test_remove_beginning(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_beginning() argument 207 test_remove_end(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_end() argument 230 test_remove_middle(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_middle() argument 254 test_merge_left(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_left() argument 291 test_merge_right(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_right() argument 329 test_merge_both(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_both() argument 372 test_merge_none(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_none() argument 428 struct btrfs_block_group *cache = NULL; run_test() local [all...] |
/kernel/linux/linux-5.10/drivers/acpi/acpica/ |
H A D | utcache.c | 4 * Module Name: utcache - local cache allocation routines 21 * PARAMETERS: cache_name - Ascii name for the cache 23 * max_depth - Maximum depth of the cache (in objects) 24 * return_cache - Where the new cache object is returned 28 * DESCRIPTION: Create a cache object 36 struct acpi_memory_list *cache; in acpi_os_create_cache() local 44 /* Create the cache object */ in acpi_os_create_cache() 46 cache = acpi_os_allocate(sizeof(struct acpi_memory_list)); in acpi_os_create_cache() 47 if (!cache) { in acpi_os_create_cache() 51 /* Populate the cache objec in acpi_os_create_cache() 74 acpi_os_purge_cache(struct acpi_memory_list *cache) acpi_os_purge_cache() argument 120 acpi_os_delete_cache(struct acpi_memory_list *cache) acpi_os_delete_cache() argument 153 acpi_os_release_object(struct acpi_memory_list *cache, void *object) acpi_os_release_object() argument 208 acpi_os_acquire_object(struct acpi_memory_list *cache) acpi_os_acquire_object() argument [all...] |
/kernel/linux/linux-6.6/drivers/acpi/acpica/ |
H A D | utcache.c | 4 * Module Name: utcache - local cache allocation routines 21 * PARAMETERS: cache_name - Ascii name for the cache 23 * max_depth - Maximum depth of the cache (in objects) 24 * return_cache - Where the new cache object is returned 28 * DESCRIPTION: Create a cache object 36 struct acpi_memory_list *cache; in acpi_os_create_cache() local 44 /* Create the cache object */ in acpi_os_create_cache() 46 cache = acpi_os_allocate(sizeof(struct acpi_memory_list)); in acpi_os_create_cache() 47 if (!cache) { in acpi_os_create_cache() 51 /* Populate the cache objec in acpi_os_create_cache() 74 acpi_os_purge_cache(struct acpi_memory_list *cache) acpi_os_purge_cache() argument 120 acpi_os_delete_cache(struct acpi_memory_list *cache) acpi_os_delete_cache() argument 153 acpi_os_release_object(struct acpi_memory_list *cache, void *object) acpi_os_release_object() argument 208 acpi_os_acquire_object(struct acpi_memory_list *cache) acpi_os_acquire_object() argument [all...] |
/kernel/linux/linux-5.10/fs/squashfs/ |
H A D | cache.c | 8 * cache.c 15 * This file implements a generic cache implementation used for both caches, 16 * plus functions layered ontop of the generic cache implementation to 19 * To avoid out of memory and fragmentation issues with vmalloc the cache 22 * It should be noted that the cache is not used for file datablocks, these 23 * are decompressed and cached in the page-cache in the normal way. The 24 * cache is only used to temporarily cache fragment and metadata blocks 49 * Look-up block in cache, and increment usage count. If not in cache, rea 52 squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length) squashfs_cache_get() argument 175 struct squashfs_cache *cache = entry->cache; squashfs_cache_put() local 197 squashfs_cache_delete(struct squashfs_cache *cache) squashfs_cache_delete() argument 227 struct squashfs_cache *cache = kzalloc(sizeof(*cache), GFP_KERNEL); squashfs_cache_init() local [all...] |
/kernel/linux/linux-6.6/fs/squashfs/ |
H A D | cache.c | 8 * cache.c 15 * This file implements a generic cache implementation used for both caches, 16 * plus functions layered ontop of the generic cache implementation to 19 * To avoid out of memory and fragmentation issues with vmalloc the cache 22 * It should be noted that the cache is not used for file datablocks, these 23 * are decompressed and cached in the page-cache in the normal way. The 24 * cache is only used to temporarily cache fragment and metadata blocks 49 * Look-up block in cache, and increment usage count. If not in cache, rea 52 squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length) squashfs_cache_get() argument 175 struct squashfs_cache *cache = entry->cache; squashfs_cache_put() local 197 squashfs_cache_delete(struct squashfs_cache *cache) squashfs_cache_delete() argument 227 struct squashfs_cache *cache = kzalloc(sizeof(*cache), GFP_KERNEL); squashfs_cache_init() local [all...] |
/kernel/linux/linux-6.6/fs/btrfs/tests/ |
H A D | free-space-tests.c | 10 #include "../free-space-cache.h" 20 static int test_extents(struct btrfs_block_group *cache) in test_extents() argument 27 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents() 33 ret = btrfs_remove_free_space(cache, 0, SZ_4M); in test_extents() 39 if (test_check_exists(cache, 0, SZ_4M)) { in test_extents() 45 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents() 51 ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); in test_extents() 57 ret = btrfs_remove_free_space(cache, 0, SZ_1M); in test_extents() 63 ret = btrfs_remove_free_space(cache, SZ_2M, 4096); in test_extents() 69 if (test_check_exists(cache, in test_extents() 90 test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps() argument 158 test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps_and_extents() argument 333 check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) check_num_extents_and_bitmaps() argument 353 check_cache_empty(struct btrfs_block_group *cache) check_cache_empty() argument 395 test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) test_steal_space_from_bitmap_to_extent() argument 833 test_bytes_index(struct btrfs_block_group *cache, u32 sectorsize) test_bytes_index() argument 1008 struct btrfs_block_group *cache; btrfs_test_free_space_cache() local [all...] |
H A D | free-space-tree-tests.c | 22 struct btrfs_block_group *cache, in __check_free_space_extents() 35 info = search_free_space_info(trans, cache, path, 0); in __check_free_space_extents() 52 end = cache->start + cache->length; in __check_free_space_extents() 60 bit = free_space_test_bit(cache, path, offset); in __check_free_space_extents() 109 struct btrfs_block_group *cache, in check_free_space_extents() 118 info = search_free_space_info(trans, cache, path, 0); in check_free_space_extents() 127 ret = __check_free_space_extents(trans, fs_info, cache, path, extents, in check_free_space_extents() 134 ret = convert_free_space_to_extents(trans, cache, path); in check_free_space_extents() 140 ret = convert_free_space_to_bitmaps(trans, cache, pat in check_free_space_extents() 20 __check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) __check_free_space_extents() argument 107 check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) check_free_space_extents() argument 150 test_empty_block_group(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_empty_block_group() argument 164 test_remove_all(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_all() argument 185 test_remove_beginning(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_beginning() argument 208 test_remove_end(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_end() argument 231 test_remove_middle(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_middle() argument 255 test_merge_left(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_left() argument 292 test_merge_right(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_right() argument 330 test_merge_both(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_both() argument 373 test_merge_none(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_none() argument 429 struct btrfs_block_group *cache = NULL; run_test() local [all...] |
/kernel/linux/linux-6.6/fs/ |
H A D | mbcache.c | 16 * Ext2 and ext4 use this cache for deduplication of extended attribute blocks. 21 * identifies a cache entry. 33 /* Maximum entries in cache to avoid degrading hash too much */ 38 /* Number of entries in cache */ 41 /* Work for shrinking when the cache has too many entries */ 47 static unsigned long mb_cache_shrink(struct mb_cache *cache, 50 static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, in mb_cache_entry_head() argument 53 return &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb_cache_entry_head() 58 * in cache 74 mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, u64 value, bool reusable) mb_cache_entry_create() argument 127 __mb_cache_entry_free(struct mb_cache *cache, struct mb_cache_entry *entry) __mb_cache_entry_free() argument 152 __entry_find(struct mb_cache *cache, struct mb_cache_entry *entry, u32 key) __entry_find() argument 192 mb_cache_entry_find_first(struct mb_cache *cache, u32 key) mb_cache_entry_find_first() argument 209 mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_find_next() argument 222 mb_cache_entry_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_get() argument 254 mb_cache_entry_delete_or_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete_or_get() argument 286 mb_cache_entry_touch(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_touch() argument 296 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_count() local 303 mb_cache_shrink(struct mb_cache *cache, unsigned long nr_to_scan) mb_cache_shrink() argument 336 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_scan() local 346 struct mb_cache *cache = container_of(work, struct mb_cache, mb_cache_shrink_worker() local 359 struct mb_cache *cache; mb_cache_create() local 405 mb_cache_destroy(struct mb_cache *cache) mb_cache_destroy() argument [all...] |
/kernel/linux/linux-5.10/fs/fscache/ |
H A D | cache.c | 2 /* FS-Cache cache handling 21 * look up a cache tag 67 * release a reference to a cache tag 86 * select a cache in which to store an object 87 * - the cache addremove semaphore must be at least read-locked by the caller 95 struct fscache_cache *cache; in fscache_select_cache_for_object() local 100 _leave(" = NULL [no cache]"); in fscache_select_cache_for_object() 104 /* we check the parent to determine the cache to use */ in fscache_select_cache_for_object() 108 * cache */ in fscache_select_cache_for_object() 113 cache in fscache_select_cache_for_object() 178 fscache_init_cache(struct fscache_cache *cache, const struct fscache_cache_ops *ops, const char *idfmt, ...) fscache_init_cache() argument 213 fscache_add_cache(struct fscache_cache *cache, struct fscache_object *ifsdef, const char *tagname) fscache_add_cache() argument 313 fscache_io_error(struct fscache_cache *cache) fscache_io_error() argument 325 fscache_withdraw_all_objects(struct fscache_cache *cache, struct list_head *dying_objects) fscache_withdraw_all_objects() argument 361 fscache_withdraw_cache(struct fscache_cache *cache) fscache_withdraw_cache() argument [all...] |
/kernel/linux/linux-5.10/fs/ |
H A D | mbcache.c | 16 * Ext2 and ext4 use this cache for deduplication of extended attribute blocks. 21 * identifies a cache entry. 33 /* Maximum entries in cache to avoid degrading hash too much */ 38 /* Number of entries in cache */ 41 /* Work for shrinking when the cache has too many entries */ 47 static unsigned long mb_cache_shrink(struct mb_cache *cache, 50 static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, in mb_cache_entry_head() argument 53 return &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb_cache_entry_head() 58 * in cache 74 mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, u64 value, bool reusable) mb_cache_entry_create() argument 127 __mb_cache_entry_free(struct mb_cache *cache, struct mb_cache_entry *entry) __mb_cache_entry_free() argument 152 __entry_find(struct mb_cache *cache, struct mb_cache_entry *entry, u32 key) __entry_find() argument 192 mb_cache_entry_find_first(struct mb_cache *cache, u32 key) mb_cache_entry_find_first() argument 209 mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_find_next() argument 222 mb_cache_entry_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_get() argument 250 mb_cache_entry_delete(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete() argument 291 mb_cache_entry_delete_or_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete_or_get() argument 323 mb_cache_entry_touch(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_touch() argument 333 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_count() local 340 mb_cache_shrink(struct mb_cache *cache, unsigned long nr_to_scan) mb_cache_shrink() argument 373 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_scan() local 383 struct mb_cache *cache = container_of(work, struct mb_cache, mb_cache_shrink_worker() local 396 struct mb_cache *cache; mb_cache_create() local 442 mb_cache_destroy(struct mb_cache *cache) mb_cache_destroy() argument [all...] |
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | block-group.c | 8 #include "free-space-cache.h" 123 void btrfs_get_block_group(struct btrfs_block_group *cache) in btrfs_get_block_group() argument 125 refcount_inc(&cache->refs); in btrfs_get_block_group() 128 void btrfs_put_block_group(struct btrfs_block_group *cache) in btrfs_put_block_group() argument 130 if (refcount_dec_and_test(&cache->refs)) { in btrfs_put_block_group() 131 WARN_ON(cache->pinned > 0); in btrfs_put_block_group() 132 WARN_ON(cache->reserved > 0); in btrfs_put_block_group() 139 if (WARN_ON(!list_empty(&cache->discard_list))) in btrfs_put_block_group() 140 btrfs_discard_cancel_work(&cache->fs_info->discard_ctl, in btrfs_put_block_group() 141 cache); in btrfs_put_block_group() 165 struct btrfs_block_group *cache; btrfs_add_block_group_cache() local 204 struct btrfs_block_group *cache, *ret = NULL; block_group_cache_tree_search() local 259 btrfs_next_block_group( struct btrfs_block_group *cache) btrfs_next_block_group() argument 367 btrfs_get_caching_control( struct btrfs_block_group *cache) btrfs_get_caching_control() argument 403 btrfs_wait_block_group_cache_progress(struct btrfs_block_group *cache, u64 num_bytes) btrfs_wait_block_group_cache_progress() argument 418 btrfs_wait_block_group_cache_done(struct btrfs_block_group *cache) btrfs_wait_block_group_cache_done() argument 693 btrfs_cache_block_group(struct btrfs_block_group *cache, int load_cache_only) btrfs_cache_block_group() argument 1230 inc_block_group_ro(struct btrfs_block_group *cache, int force) inc_block_group_ro() argument 1741 exclude_super_stripes(struct btrfs_block_group *cache) exclude_super_stripes() argument 1783 link_block_group(struct btrfs_block_group *cache) link_block_group() argument 1796 struct btrfs_block_group *cache; btrfs_create_block_group_cache() local 1888 read_block_group_item(struct btrfs_block_group *cache, struct btrfs_path *path, const struct btrfs_key *key) read_block_group_item() argument 1909 struct btrfs_block_group *cache; read_one_block_group() local 2010 struct btrfs_block_group *cache; btrfs_read_block_groups() local 2158 struct btrfs_block_group *cache; btrfs_make_block_group() local 2237 btrfs_inc_block_group_ro(struct btrfs_block_group *cache, bool do_chunk_alloc) btrfs_inc_block_group_ro() argument 2324 btrfs_dec_block_group_ro(struct btrfs_block_group *cache) btrfs_dec_block_group_ro() argument 2343 update_block_group_item(struct btrfs_trans_handle *trans, struct btrfs_path *path, struct btrfs_block_group *cache) update_block_group_item() argument 2545 struct btrfs_block_group *cache, *tmp; btrfs_setup_space_cache() local 2583 struct btrfs_block_group *cache; btrfs_start_dirty_block_groups() local 2749 struct btrfs_block_group *cache; btrfs_write_dirty_block_groups() local 2874 struct btrfs_block_group *cache = NULL; btrfs_update_block_group() local 2989 btrfs_add_reserved_bytes(struct btrfs_block_group *cache, u64 ram_bytes, u64 num_bytes, int delalloc) btrfs_add_reserved_bytes() argument 3032 btrfs_free_reserved_bytes(struct btrfs_block_group *cache, u64 num_bytes, int delalloc) btrfs_free_reserved_bytes() argument 3416 btrfs_freeze_block_group(struct btrfs_block_group *cache) btrfs_freeze_block_group() argument [all...] |
/kernel/linux/linux-6.6/io_uring/ |
H A D | alloc_cache.h | 5 * Don't allow the cache to grow beyond this size. 13 static inline bool io_alloc_cache_put(struct io_alloc_cache *cache, in io_alloc_cache_put() argument 16 if (cache->nr_cached < cache->max_cached) { in io_alloc_cache_put() 17 cache->nr_cached++; in io_alloc_cache_put() 18 wq_stack_add_head(&entry->node, &cache->list); in io_alloc_cache_put() 26 static inline bool io_alloc_cache_empty(struct io_alloc_cache *cache) in io_alloc_cache_empty() argument 28 return !cache->list.next; in io_alloc_cache_empty() 31 static inline struct io_cache_entry *io_alloc_cache_get(struct io_alloc_cache *cache) in io_alloc_cache_get() argument 33 if (cache in io_alloc_cache_get() 46 io_alloc_cache_init(struct io_alloc_cache *cache, unsigned max_nr, size_t size) io_alloc_cache_init() argument 55 io_alloc_cache_free(struct io_alloc_cache *cache, void (*free)(struct io_cache_entry *)) io_alloc_cache_free() argument [all...] |
/kernel/linux/linux-5.10/fs/nfs/ |
H A D | nfs42xattr.c | 6 * User extended attribute client side cache functions. 21 * a cache structure attached to NFS inodes. This structure is allocated 22 * when needed, and freed when the cache is zapped. 24 * The cache structure contains as hash table of entries, and a pointer 25 * to a special-cased entry for the listxattr cache. 28 * counting. The cache entries use a similar refcounting scheme. 30 * This makes freeing a cache, both from the shrinker and from the 31 * zap cache path, easy. It also means that, in current use cases, 40 * Two shrinkers deal with the cache entries themselves: one for 45 * The other shrinker frees the cache structure 64 struct nfs4_xattr_cache *cache; global() member 106 nfs4_xattr_hash_init(struct nfs4_xattr_cache *cache) nfs4_xattr_hash_init() argument 270 struct nfs4_xattr_cache *cache; nfs4_xattr_free_cache_cb() local 290 struct nfs4_xattr_cache *cache; nfs4_xattr_alloc_cache() local 310 nfs4_xattr_set_listcache(struct nfs4_xattr_cache *cache, struct nfs4_xattr_entry *new) nfs4_xattr_set_listcache() argument 385 nfs4_xattr_discard_cache(struct nfs4_xattr_cache *cache) nfs4_xattr_discard_cache() argument 430 struct nfs4_xattr_cache *cache, *oldcache, *newcache; nfs4_xattr_get_cache() local 505 nfs4_xattr_hash_bucket(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_bucket() argument 527 nfs4_xattr_hash_add(struct nfs4_xattr_cache *cache, struct nfs4_xattr_entry *entry) nfs4_xattr_hash_add() argument 565 nfs4_xattr_hash_remove(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_remove() argument 588 nfs4_xattr_hash_find(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_find() argument 612 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_get() local 651 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_list() local 692 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_add() local 723 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_remove() local 743 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_set_list() local 829 struct nfs4_xattr_cache *cache = container_of(item, cache_lru_isolate() local 863 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_scan() local 894 struct nfs4_xattr_cache *cache; entry_lru_isolate() local 985 struct nfs4_xattr_cache *cache = (struct nfs4_xattr_cache *)p; nfs4_xattr_cache_init_once() local [all...] |