Home
last modified time | relevance | path

Searched refs:cache (Results 1 - 25 of 1745) sorted by relevance

12345678910>>...70

/kernel/linux/linux-6.6/drivers/md/
H A Ddm-cache-target.c11 #include "dm-cache-metadata.h"
24 #define DM_MSG_PREFIX "cache"
27 "A percentage of time allocated for copying to and/or from cache");
35 * cblock: index of a cache block
36 * promotion: movement of a block from origin to cache
37 * demotion: movement of a block from cache to origin
38 * migration: movement of a block between the origin and cache device,
241 * The block size of the device holding cache data must be
256 * dirty. If you lose the cache device you will lose data.
262 * Data is written to both cache an
297 struct cache { global() struct
421 struct cache *cache; global() member
433 writethrough_mode(struct cache *cache) writethrough_mode() argument
438 writeback_mode(struct cache *cache) writeback_mode() argument
443 passthrough_mode(struct cache *cache) passthrough_mode() argument
450 wake_deferred_bio_worker(struct cache *cache) wake_deferred_bio_worker() argument
455 wake_migration_worker(struct cache *cache) wake_migration_worker() argument
465 alloc_prison_cell(struct cache *cache) alloc_prison_cell() argument
470 free_prison_cell(struct cache *cache, struct dm_bio_prison_cell_v2 *cell) free_prison_cell() argument
475 alloc_migration(struct cache *cache) alloc_migration() argument
491 struct cache *cache = mg->cache; free_migration() local
556 defer_bio(struct cache *cache, struct bio *bio) defer_bio() argument
565 defer_bios(struct cache *cache, struct bio_list *bios) defer_bios() argument
577 bio_detain_shared(struct cache *cache, dm_oblock_t oblock, struct bio *bio) bio_detain_shared() argument
608 is_dirty(struct cache *cache, dm_cblock_t b) is_dirty() argument
613 set_dirty(struct cache *cache, dm_cblock_t cblock) set_dirty() argument
625 force_set_dirty(struct cache *cache, dm_cblock_t cblock) force_set_dirty() argument
632 force_clear_dirty(struct cache *cache, dm_cblock_t cblock) force_clear_dirty() argument
644 block_size_is_power_of_two(struct cache *cache) block_size_is_power_of_two() argument
656 oblocks_per_dblock(struct cache *cache) oblocks_per_dblock() argument
668 oblock_to_dblock(struct cache *cache, dm_oblock_t oblock) oblock_to_dblock() argument
674 set_discard(struct cache *cache, dm_dblock_t b) set_discard() argument
684 clear_discard(struct cache *cache, dm_dblock_t b) clear_discard() argument
691 is_discarded(struct cache *cache, dm_dblock_t b) is_discarded() argument
702 is_discarded_oblock(struct cache *cache, dm_oblock_t b) is_discarded_oblock() argument
719 remap_to_origin(struct cache *cache, struct bio *bio) remap_to_origin() argument
724 remap_to_cache(struct cache *cache, struct bio *bio, dm_cblock_t cblock) remap_to_cache() argument
741 check_if_tick_bio_needed(struct cache *cache, struct bio *bio) check_if_tick_bio_needed() argument
755 remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock) remap_to_origin_clear_discard() argument
765 remap_to_cache_dirty(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_cache_dirty() argument
776 get_bio_block(struct cache *cache, struct bio *bio) get_bio_block() argument
788 accountable_bio(struct cache *cache, struct bio *bio) accountable_bio() argument
793 accounted_begin(struct cache *cache, struct bio *bio) accounted_begin() argument
804 accounted_complete(struct cache *cache, struct bio *bio) accounted_complete() argument
811 accounted_request(struct cache *cache, struct bio *bio) accounted_request() argument
819 struct cache *cache = context; issue_op() local
828 remap_to_origin_and_cache(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_origin_and_cache() argument
850 get_cache_mode(struct cache *cache) get_cache_mode() argument
855 cache_device_name(struct cache *cache) cache_device_name() argument
860 notify_mode_switch(struct cache *cache, enum cache_metadata_mode mode) notify_mode_switch() argument
873 set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mode) set_cache_mode() argument
914 abort_transaction(struct cache *cache) abort_transaction() argument
933 metadata_operation_failed(struct cache *cache, const char *op, int r) metadata_operation_failed() argument
943 load_stats(struct cache *cache) load_stats() argument
954 save_stats(struct cache *cache) save_stats() argument
994 inc_io_migrations(struct cache *cache) inc_io_migrations() argument
999 dec_io_migrations(struct cache *cache) dec_io_migrations() argument
1009 calc_discard_block_range(struct cache *cache, struct bio *bio, dm_dblock_t *b, dm_dblock_t *e) calc_discard_block_range() argument
1025 prevent_background_work(struct cache *cache) prevent_background_work() argument
1032 allow_background_work(struct cache *cache) allow_background_work() argument
1039 background_work_begin(struct cache *cache) background_work_begin() argument
1050 background_work_end(struct cache *cache) background_work_end() argument
1059 bio_writes_complete_block(struct cache *cache, struct bio *bio) bio_writes_complete_block() argument
1065 optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block) optimisable_bio() argument
1098 struct cache *cache = mg->cache; copy() local
1114 bio_drop_shared_lock(struct cache *cache, struct bio *bio) bio_drop_shared_lock() argument
1126 struct cache *cache = mg->cache; overwrite_endio() local
1172 struct cache *cache = mg->cache; mg_complete() local
1241 struct cache *cache = mg->cache; mg_update_metadata() local
1343 struct cache *cache = mg->cache; mg_full_copy() local
1397 struct cache *cache = mg->cache; mg_lock_writes() local
1428 mg_start(struct cache *cache, struct policy_work *op, struct bio *bio) mg_start() argument
1457 struct cache *cache = mg->cache; invalidate_complete() local
1479 invalidate_cblock(struct cache *cache, dm_cblock_t cblock) invalidate_cblock() argument
1508 struct cache *cache = mg->cache; invalidate_remove() local
1527 struct cache *cache = mg->cache; invalidate_lock() local
1559 invalidate_start(struct cache *cache, dm_cblock_t cblock, dm_oblock_t oblock, struct bio *bio) invalidate_start() argument
1587 spare_migration_bandwidth(struct cache *cache) spare_migration_bandwidth() argument
1599 inc_hit_counter(struct cache *cache, struct bio *bio) inc_hit_counter() argument
1605 inc_miss_counter(struct cache *cache, struct bio *bio) inc_miss_counter() argument
1613 map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block, bool *commit_needed) map_bio() argument
1728 process_bio(struct cache *cache, struct bio *bio) process_bio() argument
1741 commit(struct cache *cache, bool clean_shutdown) commit() argument
1761 struct cache *cache = context; commit_op() local
1771 process_flush_bio(struct cache *cache, struct bio *bio) process_flush_bio() argument
1784 process_discard_bio(struct cache *cache, struct bio *bio) process_discard_bio() argument
1810 struct cache *cache = container_of(ws, struct cache, deferred_bio_worker); process_deferred_bios() local
1844 requeue_deferred_bios(struct cache *cache) requeue_deferred_bios() argument
1866 struct cache *cache = container_of(to_delayed_work(ws), struct cache, waker); do_waker() local
1878 struct cache *cache = container_of(ws, struct cache, migration_worker); check_migrations() local
1912 destroy(struct cache *cache) destroy() argument
1960 struct cache *cache = ti->private; cache_dtr() local
2268 process_config_option(struct cache *cache, const char *key, const char *value) process_config_option() argument
2283 set_config_value(struct cache *cache, const char *key, const char *value) set_config_value() argument
2296 set_config_values(struct cache *cache, int argc, const char **argv) set_config_values() argument
2317 create_cache_policy(struct cache *cache, struct cache_args *ca, char **error) create_cache_policy() argument
2360 set_cache_size(struct cache *cache, dm_cblock_t size) set_cache_size() argument
2379 struct cache *cache; cache_create() local
2572 copy_ctr_args(struct cache *cache, int argc, const char **argv) copy_ctr_args() argument
2600 struct cache *cache = NULL; cache_ctr() local
2633 struct cache *cache = ti->private; cache_map() local
2665 struct cache *cache = ti->private; cache_end_io() local
2683 write_dirty_bitset(struct cache *cache) write_dirty_bitset() argument
2697 write_discard_bitset(struct cache *cache) write_discard_bitset() argument
2724 write_hints(struct cache *cache) write_hints() argument
2743 sync_metadata(struct cache *cache) sync_metadata() argument
2775 struct cache *cache = ti->private; cache_postsuspend() local
2797 struct cache *cache = context; load_mapping() local
2815 struct cache *cache; global() member
2825 discard_load_info_init(struct cache *cache, struct discard_load_info *li) discard_load_info_init() argument
2892 get_cache_dev_size(struct cache *cache) get_cache_dev_size() argument
2899 can_resize(struct cache *cache, dm_cblock_t new_size) can_resize() argument
2925 resize_cache_dev(struct cache *cache, dm_cblock_t new_size) resize_cache_dev() argument
2944 struct cache *cache = ti->private; cache_preresume() local
3005 struct cache *cache = ti->private; cache_resume() local
3012 emit_flags(struct cache *cache, char *result, unsigned int maxlen, ssize_t *sz_ptr) emit_flags() argument
3065 struct cache *cache = ti->private; cache_status() local
3196 parse_cblock_range(struct cache *cache, const char *str, struct cblock_range *result) parse_cblock_range() argument
3233 validate_cblock_range(struct cache *cache, struct cblock_range *range) validate_cblock_range() argument
3265 request_invalidation(struct cache *cache, struct cblock_range *range) request_invalidation() argument
3287 process_invalidate_cblocks_message(struct cache *cache, unsigned int count, const char **cblock_ranges) process_invalidate_cblocks_message() argument
3331 struct cache *cache = ti->private; cache_message() local
3355 struct cache *cache = ti->private; cache_iterate_devices() local
3368 disable_passdown_if_not_supported(struct cache *cache) disable_passdown_if_not_supported() argument
3390 set_discard_limits(struct cache *cache, struct queue_limits *limits) set_discard_limits() argument
3416 struct cache *cache = ti->private; cache_io_hints() local
[all...]
/kernel/linux/linux-5.10/drivers/md/
H A Ddm-cache-target.c10 #include "dm-cache-metadata.h"
22 #define DM_MSG_PREFIX "cache"
25 "A percentage of time allocated for copying to and/or from cache");
33 * cblock: index of a cache block
34 * promotion: movement of a block from origin to cache
35 * demotion: movement of a block from cache to origin
36 * migration: movement of a block between the origin and cache device,
310 * The block size of the device holding cache data must be
325 * dirty. If you lose the cache device you will lose data.
331 * Data is written to both cache an
366 struct cache { global() struct
490 struct cache *cache; global() member
502 writethrough_mode(struct cache *cache) writethrough_mode() argument
507 writeback_mode(struct cache *cache) writeback_mode() argument
512 passthrough_mode(struct cache *cache) passthrough_mode() argument
519 wake_deferred_bio_worker(struct cache *cache) wake_deferred_bio_worker() argument
524 wake_migration_worker(struct cache *cache) wake_migration_worker() argument
534 alloc_prison_cell(struct cache *cache) alloc_prison_cell() argument
539 free_prison_cell(struct cache *cache, struct dm_bio_prison_cell_v2 *cell) free_prison_cell() argument
544 alloc_migration(struct cache *cache) alloc_migration() argument
560 struct cache *cache = mg->cache; free_migration() local
622 defer_bio(struct cache *cache, struct bio *bio) defer_bio() argument
631 defer_bios(struct cache *cache, struct bio_list *bios) defer_bios() argument
643 bio_detain_shared(struct cache *cache, dm_oblock_t oblock, struct bio *bio) bio_detain_shared() argument
674 is_dirty(struct cache *cache, dm_cblock_t b) is_dirty() argument
679 set_dirty(struct cache *cache, dm_cblock_t cblock) set_dirty() argument
691 force_set_dirty(struct cache *cache, dm_cblock_t cblock) force_set_dirty() argument
698 force_clear_dirty(struct cache *cache, dm_cblock_t cblock) force_clear_dirty() argument
710 block_size_is_power_of_two(struct cache *cache) block_size_is_power_of_two() argument
722 oblocks_per_dblock(struct cache *cache) oblocks_per_dblock() argument
734 oblock_to_dblock(struct cache *cache, dm_oblock_t oblock) oblock_to_dblock() argument
740 set_discard(struct cache *cache, dm_dblock_t b) set_discard() argument
750 clear_discard(struct cache *cache, dm_dblock_t b) clear_discard() argument
757 is_discarded(struct cache *cache, dm_dblock_t b) is_discarded() argument
767 is_discarded_oblock(struct cache *cache, dm_oblock_t b) is_discarded_oblock() argument
781 remap_to_origin(struct cache *cache, struct bio *bio) remap_to_origin() argument
786 remap_to_cache(struct cache *cache, struct bio *bio, dm_cblock_t cblock) remap_to_cache() argument
803 check_if_tick_bio_needed(struct cache *cache, struct bio *bio) check_if_tick_bio_needed() argument
817 __remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock, bool bio_has_pbd) __remap_to_origin_clear_discard() argument
827 remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock) remap_to_origin_clear_discard() argument
834 remap_to_cache_dirty(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_cache_dirty() argument
845 get_bio_block(struct cache *cache, struct bio *bio) get_bio_block() argument
857 accountable_bio(struct cache *cache, struct bio *bio) accountable_bio() argument
862 accounted_begin(struct cache *cache, struct bio *bio) accounted_begin() argument
873 accounted_complete(struct cache *cache, struct bio *bio) accounted_complete() argument
880 accounted_request(struct cache *cache, struct bio *bio) accounted_request() argument
888 struct cache *cache = context; issue_op() local
896 remap_to_origin_and_cache(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) remap_to_origin_and_cache() argument
917 get_cache_mode(struct cache *cache) get_cache_mode() argument
922 cache_device_name(struct cache *cache) cache_device_name() argument
927 notify_mode_switch(struct cache *cache, enum cache_metadata_mode mode) notify_mode_switch() argument
940 set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mode) set_cache_mode() argument
981 abort_transaction(struct cache *cache) abort_transaction() argument
1000 metadata_operation_failed(struct cache *cache, const char *op, int r) metadata_operation_failed() argument
1010 load_stats(struct cache *cache) load_stats() argument
1021 save_stats(struct cache *cache) save_stats() argument
1060 inc_io_migrations(struct cache *cache) inc_io_migrations() argument
1065 dec_io_migrations(struct cache *cache) dec_io_migrations() argument
1075 calc_discard_block_range(struct cache *cache, struct bio *bio, dm_dblock_t *b, dm_dblock_t *e) calc_discard_block_range() argument
1091 prevent_background_work(struct cache *cache) prevent_background_work() argument
1098 allow_background_work(struct cache *cache) allow_background_work() argument
1105 background_work_begin(struct cache *cache) background_work_begin() argument
1116 background_work_end(struct cache *cache) background_work_end() argument
1125 bio_writes_complete_block(struct cache *cache, struct bio *bio) bio_writes_complete_block() argument
1131 optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block) optimisable_bio() argument
1163 struct cache *cache = mg->cache; copy() local
1179 bio_drop_shared_lock(struct cache *cache, struct bio *bio) bio_drop_shared_lock() argument
1191 struct cache *cache = mg->cache; overwrite_endio() local
1237 struct cache *cache = mg->cache; mg_complete() local
1305 struct cache *cache = mg->cache; mg_update_metadata() local
1407 struct cache *cache = mg->cache; mg_full_copy() local
1460 struct cache *cache = mg->cache; mg_lock_writes() local
1491 mg_start(struct cache *cache, struct policy_work *op, struct bio *bio) mg_start() argument
1518 struct cache *cache = mg->cache; invalidate_complete() local
1539 invalidate_cblock(struct cache *cache, dm_cblock_t cblock) invalidate_cblock() argument
1566 struct cache *cache = mg->cache; invalidate_remove() local
1585 struct cache *cache = mg->cache; invalidate_lock() local
1617 invalidate_start(struct cache *cache, dm_cblock_t cblock, dm_oblock_t oblock, struct bio *bio) invalidate_start() argument
1643 spare_migration_bandwidth(struct cache *cache) spare_migration_bandwidth() argument
1655 inc_hit_counter(struct cache *cache, struct bio *bio) inc_hit_counter() argument
1661 inc_miss_counter(struct cache *cache, struct bio *bio) inc_miss_counter() argument
1669 map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block, bool *commit_needed) map_bio() argument
1784 process_bio(struct cache *cache, struct bio *bio) process_bio() argument
1797 commit(struct cache *cache, bool clean_shutdown) commit() argument
1817 struct cache *cache = context; commit_op() local
1827 process_flush_bio(struct cache *cache, struct bio *bio) process_flush_bio() argument
1840 process_discard_bio(struct cache *cache, struct bio *bio) process_discard_bio() argument
1864 struct cache *cache = container_of(ws, struct cache, deferred_bio_worker); process_deferred_bios() local
1897 requeue_deferred_bios(struct cache *cache) requeue_deferred_bios() argument
1919 struct cache *cache = container_of(to_delayed_work(ws), struct cache, waker); do_waker() local
1931 struct cache *cache = container_of(ws, struct cache, migration_worker); check_migrations() local
1963 destroy(struct cache *cache) destroy() argument
2011 struct cache *cache = ti->private; cache_dtr() local
2320 process_config_option(struct cache *cache, const char *key, const char *value) process_config_option() argument
2335 set_config_value(struct cache *cache, const char *key, const char *value) set_config_value() argument
2348 set_config_values(struct cache *cache, int argc, const char **argv) set_config_values() argument
2369 create_cache_policy(struct cache *cache, struct cache_args *ca, char **error) create_cache_policy() argument
2412 set_cache_size(struct cache *cache, dm_cblock_t size) set_cache_size() argument
2431 struct cache *cache; cache_create() local
2623 copy_ctr_args(struct cache *cache, int argc, const char **argv) copy_ctr_args() argument
2651 struct cache *cache = NULL; cache_ctr() local
2684 struct cache *cache = ti->private; cache_map() local
2716 struct cache *cache = ti->private; cache_end_io() local
2734 write_dirty_bitset(struct cache *cache) write_dirty_bitset() argument
2748 write_discard_bitset(struct cache *cache) write_discard_bitset() argument
2775 write_hints(struct cache *cache) write_hints() argument
2794 sync_metadata(struct cache *cache) sync_metadata() argument
2826 struct cache *cache = ti->private; cache_postsuspend() local
2849 struct cache *cache = context; load_mapping() local
2871 struct cache *cache; global() member
2881 discard_load_info_init(struct cache *cache, struct discard_load_info *li) discard_load_info_init() argument
2948 get_cache_dev_size(struct cache *cache) get_cache_dev_size() argument
2955 can_resize(struct cache *cache, dm_cblock_t new_size) can_resize() argument
2981 resize_cache_dev(struct cache *cache, dm_cblock_t new_size) resize_cache_dev() argument
3000 struct cache *cache = ti->private; cache_preresume() local
3061 struct cache *cache = ti->private; cache_resume() local
3068 emit_flags(struct cache *cache, char *result, unsigned maxlen, ssize_t *sz_ptr) emit_flags() argument
3121 struct cache *cache = ti->private; cache_status() local
3228 parse_cblock_range(struct cache *cache, const char *str, struct cblock_range *result) parse_cblock_range() argument
3265 validate_cblock_range(struct cache *cache, struct cblock_range *range) validate_cblock_range() argument
3297 request_invalidation(struct cache *cache, struct cblock_range *range) request_invalidation() argument
3319 process_invalidate_cblocks_message(struct cache *cache, unsigned count, const char **cblock_ranges) process_invalidate_cblocks_message() argument
3363 struct cache *cache = ti->private; cache_message() local
3387 struct cache *cache = ti->private; cache_iterate_devices() local
3407 disable_passdown_if_not_supported(struct cache *cache) disable_passdown_if_not_supported() argument
3430 set_discard_limits(struct cache *cache, struct queue_limits *limits) set_discard_limits() argument
3456 struct cache *cache = ti->private; cache_io_hints() local
[all...]
/kernel/linux/linux-5.10/arch/powerpc/kernel/
H A Dcacheinfo.c3 * Processor cache information made available to userspace via sysfs;
28 * - a "cache" kobject for the top-level directory
29 * - a list of "index" objects representing the cpu's local cache hierarchy
32 struct kobject *kobj; /* bare (not embedded) kobject for cache
37 /* "index" object: each cpu's cache directory has an index
38 * subdirectory corresponding to a cache object associated with the
44 struct cache *cache; member
48 * cache type */
53 /* Allow for both [di]-cache
118 struct cache { global() struct
139 cache_type_string(const struct cache *cache) cache_type_string() argument
144 cache_init(struct cache *cache, int type, int level, struct device_node *ofnode) cache_init() argument
156 struct cache *cache; new_cache() local
165 release_cache_debugcheck(struct cache *cache) release_cache_debugcheck() argument
178 release_cache(struct cache *cache) release_cache() argument
192 cache_cpu_set(struct cache *cache, int cpu) cache_cpu_set() argument
206 cache_size(const struct cache *cache, unsigned int *ret) cache_size() argument
221 cache_size_kb(const struct cache *cache, unsigned int *ret) cache_size_kb() argument
233 cache_get_line_size(const struct cache *cache, unsigned int *ret) cache_get_line_size() argument
256 cache_nr_sets(const struct cache *cache, unsigned int *ret) cache_nr_sets() argument
271 cache_associativity(const struct cache *cache, unsigned int *ret) cache_associativity() argument
303 cache_find_first_sibling(struct cache *cache) cache_find_first_sibling() argument
321 struct cache *cache = NULL; cache_lookup_by_node() local
387 struct cache *cache; cache_do_one_devnode() local
400 struct cache *cache; cache_lookup_or_instantiate() local
434 do_subsidiary_caches_debugcheck(struct cache *cache) do_subsidiary_caches_debugcheck() argument
446 do_subsidiary_caches(struct cache *cache) do_subsidiary_caches() argument
556 struct cache *cache; size_show() local
573 struct cache *cache; line_size_show() local
589 struct cache *cache; nr_sets_show() local
605 struct cache *cache; associativity_show() local
620 struct cache *cache; type_show() local
633 struct cache *cache; level_show() local
659 get_big_core_shared_cpu_map(int cpu, struct cache *cache) get_big_core_shared_cpu_map() argument
671 struct cache *cache; show_shared_cpumap() local
739 struct cache *cache; cacheinfo_create_index_opt_attrs() local
777 cacheinfo_create_index_dir(struct cache *cache, int index, struct cache_dir *cache_dir) cacheinfo_create_index_dir() argument
806 struct cache *cache; cacheinfo_sysfs_populate() local
823 struct cache *cache; cacheinfo_cpu_online() local
840 struct cache *cache; cache_lookup_by_cpu() local
880 cache_cpu_clear(struct cache *cache, int cpu) cache_cpu_clear() argument
904 struct cache *cache; cacheinfo_cpu_offline() local
[all...]
/kernel/linux/linux-6.6/arch/powerpc/kernel/
H A Dcacheinfo.c3 * Processor cache information made available to userspace via sysfs;
27 * - a "cache" kobject for the top-level directory
28 * - a list of "index" objects representing the cpu's local cache hierarchy
31 struct kobject *kobj; /* bare (not embedded) kobject for cache
36 /* "index" object: each cpu's cache directory has an index
37 * subdirectory corresponding to a cache object associated with the
43 struct cache *cache; member
47 * cache type */
52 /* Allow for both [di]-cache
117 struct cache { global() struct
139 cache_type_string(const struct cache *cache) cache_type_string() argument
144 cache_init(struct cache *cache, int type, int level, struct device_node *ofnode, int group_id) cache_init() argument
158 struct cache *cache; new_cache() local
167 release_cache_debugcheck(struct cache *cache) release_cache_debugcheck() argument
180 release_cache(struct cache *cache) release_cache() argument
194 cache_cpu_set(struct cache *cache, int cpu) cache_cpu_set() argument
208 cache_size(const struct cache *cache, unsigned int *ret) cache_size() argument
223 cache_size_kb(const struct cache *cache, unsigned int *ret) cache_size_kb() argument
235 cache_get_line_size(const struct cache *cache, unsigned int *ret) cache_get_line_size() argument
258 cache_nr_sets(const struct cache *cache, unsigned int *ret) cache_nr_sets() argument
273 cache_associativity(const struct cache *cache, unsigned int *ret) cache_associativity() argument
305 cache_find_first_sibling(struct cache *cache) cache_find_first_sibling() argument
326 struct cache *cache = NULL; cache_lookup_by_node_group() local
394 struct cache *cache; cache_do_one_devnode() local
408 struct cache *cache; cache_lookup_or_instantiate() local
442 do_subsidiary_caches_debugcheck(struct cache *cache) do_subsidiary_caches_debugcheck() argument
477 do_subsidiary_caches(struct cache *cache, unsigned int cpu_id) do_subsidiary_caches() argument
592 struct cache *cache; size_show() local
609 struct cache *cache; line_size_show() local
625 struct cache *cache; nr_sets_show() local
641 struct cache *cache; associativity_show() local
656 struct cache *cache; type_show() local
669 struct cache *cache; level_show() local
684 struct cache *cache; show_shared_cpumap() local
747 struct cache *cache; cacheinfo_create_index_opt_attrs() local
785 cacheinfo_create_index_dir(struct cache *cache, int index, struct cache_dir *cache_dir) cacheinfo_create_index_dir() argument
814 struct cache *cache; cacheinfo_sysfs_populate() local
831 struct cache *cache; cacheinfo_cpu_online() local
848 struct cache *cache; cache_lookup_by_cpu() local
890 cache_cpu_clear(struct cache *cache, int cpu) cache_cpu_clear() argument
914 struct cache *cache; cacheinfo_cpu_offline() local
[all...]
/kernel/linux/linux-5.10/fs/cachefiles/
H A Dbind.c2 /* Bind and unbind a cache from the filesystem backing it
25 * bind a directory as a cache
27 int cachefiles_daemon_bind(struct cachefiles_cache *cache, char *args) in cachefiles_daemon_bind() argument
30 cache->frun_percent, in cachefiles_daemon_bind()
31 cache->fcull_percent, in cachefiles_daemon_bind()
32 cache->fstop_percent, in cachefiles_daemon_bind()
33 cache->brun_percent, in cachefiles_daemon_bind()
34 cache->bcull_percent, in cachefiles_daemon_bind()
35 cache->bstop_percent, in cachefiles_daemon_bind()
39 ASSERT(cache in cachefiles_daemon_bind()
81 cachefiles_daemon_add_cache(struct cachefiles_cache *cache) cachefiles_daemon_add_cache() argument
255 cachefiles_daemon_unbind(struct cachefiles_cache *cache) cachefiles_daemon_unbind() argument
[all...]
H A Ddaemon.c59 int (*handler)(struct cachefiles_cache *cache, char *args);
85 struct cachefiles_cache *cache; in cachefiles_daemon_open() local
97 /* allocate a cache record */ in cachefiles_daemon_open()
98 cache = kzalloc(sizeof(struct cachefiles_cache), GFP_KERNEL); in cachefiles_daemon_open()
99 if (!cache) { in cachefiles_daemon_open()
104 mutex_init(&cache->daemon_mutex); in cachefiles_daemon_open()
105 cache->active_nodes = RB_ROOT; in cachefiles_daemon_open()
106 rwlock_init(&cache->active_lock); in cachefiles_daemon_open()
107 init_waitqueue_head(&cache->daemon_pollwq); in cachefiles_daemon_open()
114 cache in cachefiles_daemon_open()
131 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_release() local
160 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_read() local
217 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_write() local
293 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_poll() local
312 cachefiles_daemon_range_error(struct cachefiles_cache *cache, char *args) cachefiles_daemon_range_error() argument
324 cachefiles_daemon_frun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_frun() argument
348 cachefiles_daemon_fcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fcull() argument
372 cachefiles_daemon_fstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fstop() argument
396 cachefiles_daemon_brun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_brun() argument
420 cachefiles_daemon_bcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bcull() argument
444 cachefiles_daemon_bstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bstop() argument
468 cachefiles_daemon_dir(struct cachefiles_cache *cache, char *args) cachefiles_daemon_dir() argument
496 cachefiles_daemon_secctx(struct cachefiles_cache *cache, char *args) cachefiles_daemon_secctx() argument
524 cachefiles_daemon_tag(struct cachefiles_cache *cache, char *args) cachefiles_daemon_tag() argument
550 cachefiles_daemon_cull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_cull() argument
599 cachefiles_daemon_debug(struct cachefiles_cache *cache, char *args) cachefiles_daemon_debug() argument
622 cachefiles_daemon_inuse(struct cachefiles_cache *cache, char *args) cachefiles_daemon_inuse() argument
671 cachefiles_has_space(struct cachefiles_cache *cache, unsigned fnr, unsigned bnr) cachefiles_has_space() argument
[all...]
/kernel/linux/linux-6.6/fs/cachefiles/
H A Ddaemon.c61 int (*handler)(struct cachefiles_cache *cache, char *args);
86 * Prepare a cache for caching.
90 struct cachefiles_cache *cache; in cachefiles_daemon_open() local
102 /* allocate a cache record */ in cachefiles_daemon_open()
103 cache = kzalloc(sizeof(struct cachefiles_cache), GFP_KERNEL); in cachefiles_daemon_open()
104 if (!cache) { in cachefiles_daemon_open()
109 mutex_init(&cache->daemon_mutex); in cachefiles_daemon_open()
110 init_waitqueue_head(&cache->daemon_pollwq); in cachefiles_daemon_open()
111 INIT_LIST_HEAD(&cache->volumes); in cachefiles_daemon_open()
112 INIT_LIST_HEAD(&cache in cachefiles_daemon_open()
135 cachefiles_flush_reqs(struct cachefiles_cache *cache) cachefiles_flush_reqs() argument
168 cachefiles_put_unbind_pincount(struct cachefiles_cache *cache) cachefiles_put_unbind_pincount() argument
177 cachefiles_get_unbind_pincount(struct cachefiles_cache *cache) cachefiles_get_unbind_pincount() argument
187 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_release() local
208 cachefiles_do_daemon_read(struct cachefiles_cache *cache, char __user *_buffer, size_t buflen) cachefiles_do_daemon_read() argument
259 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_read() local
281 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_write() local
357 struct cachefiles_cache *cache = file->private_data; cachefiles_daemon_poll() local
381 cachefiles_daemon_range_error(struct cachefiles_cache *cache, char *args) cachefiles_daemon_range_error() argument
393 cachefiles_daemon_frun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_frun() argument
417 cachefiles_daemon_fcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fcull() argument
441 cachefiles_daemon_fstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_fstop() argument
465 cachefiles_daemon_brun(struct cachefiles_cache *cache, char *args) cachefiles_daemon_brun() argument
489 cachefiles_daemon_bcull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bcull() argument
513 cachefiles_daemon_bstop(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bstop() argument
537 cachefiles_daemon_dir(struct cachefiles_cache *cache, char *args) cachefiles_daemon_dir() argument
565 cachefiles_daemon_secctx(struct cachefiles_cache *cache, char *args) cachefiles_daemon_secctx() argument
593 cachefiles_daemon_tag(struct cachefiles_cache *cache, char *args) cachefiles_daemon_tag() argument
619 cachefiles_daemon_cull(struct cachefiles_cache *cache, char *args) cachefiles_daemon_cull() argument
667 cachefiles_daemon_debug(struct cachefiles_cache *cache, char *args) cachefiles_daemon_debug() argument
690 cachefiles_daemon_inuse(struct cachefiles_cache *cache, char *args) cachefiles_daemon_inuse() argument
737 cachefiles_daemon_bind(struct cachefiles_cache *cache, char *args) cachefiles_daemon_bind() argument
798 cachefiles_daemon_unbind(struct cachefiles_cache *cache) cachefiles_daemon_unbind() argument
[all...]
H A Dcache.c2 /* Manage high-level VFS aspects of a cache.
14 * Bring a cache online.
16 int cachefiles_add_cache(struct cachefiles_cache *cache) in cachefiles_add_cache() argument
27 cache_cookie = fscache_acquire_cache(cache->tag); in cachefiles_add_cache()
32 ret = cachefiles_get_security_ID(cache); in cachefiles_add_cache()
36 cachefiles_begin_secure(cache, &saved_cred); in cachefiles_add_cache()
38 /* look up the directory at the root of the cache */ in cachefiles_add_cache()
39 ret = kern_path(cache->rootdirname, LOOKUP_DIRECTORY, &path); in cachefiles_add_cache()
43 cache->mnt = path.mnt; in cachefiles_add_cache()
48 pr_warn("File cache o in cachefiles_add_cache()
184 cachefiles_has_space(struct cachefiles_cache *cache, unsigned fnr, unsigned bnr, enum cachefiles_has_space_for reason) cachefiles_has_space() argument
287 cachefiles_withdraw_objects(struct cachefiles_cache *cache) cachefiles_withdraw_objects() argument
317 cachefiles_withdraw_volumes(struct cachefiles_cache *cache) cachefiles_withdraw_volumes() argument
343 cachefiles_sync_cache(struct cachefiles_cache *cache) cachefiles_sync_cache() argument
367 cachefiles_withdraw_cache(struct cachefiles_cache *cache) cachefiles_withdraw_cache() argument
[all...]
/kernel/linux/linux-6.6/fs/fscache/
H A Dcache.c2 /* FS-Cache cache handling
22 * Allocate a cache cookie.
26 struct fscache_cache *cache; in fscache_alloc_cache() local
28 cache = kzalloc(sizeof(*cache), GFP_KERNEL); in fscache_alloc_cache()
29 if (cache) { in fscache_alloc_cache()
31 cache->name = kstrdup(name, GFP_KERNEL); in fscache_alloc_cache()
32 if (!cache->name) { in fscache_alloc_cache()
33 kfree(cache); in fscache_alloc_cache()
37 refcount_set(&cache in fscache_alloc_cache()
44 fscache_get_cache_maybe(struct fscache_cache *cache, enum fscache_cache_trace where) fscache_get_cache_maybe() argument
61 struct fscache_cache *candidate, *cache, *unnamed = NULL; fscache_lookup_cache() local
151 struct fscache_cache *cache; fscache_acquire_cache() local
179 fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where) fscache_put_cache() argument
209 fscache_relinquish_cache(struct fscache_cache *cache) fscache_relinquish_cache() argument
234 fscache_add_cache(struct fscache_cache *cache, const struct fscache_cache_ops *ops, void *cache_priv) fscache_add_cache() argument
289 fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) fscache_begin_cache_access() argument
315 fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) fscache_end_cache_access() argument
337 fscache_io_error(struct fscache_cache *cache) fscache_io_error() argument
355 fscache_withdraw_cache(struct fscache_cache *cache) fscache_withdraw_cache() argument
382 struct fscache_cache *cache; fscache_caches_seq_show() local
[all...]
/kernel/linux/linux-5.10/mm/
H A Dswap_slots.c3 * Manage cache of swap slots to be used for and returned from
25 * The swap slots cache is protected by a mutex instead of
42 /* Serialize swap slots cache enable/disable operations */
107 /* if global pool of slot caches too low, deactivate cache */ in check_cache_active()
116 struct swap_slots_cache *cache; in alloc_swap_slot_cache() local
137 cache = &per_cpu(swp_slots, cpu); in alloc_swap_slot_cache()
138 if (cache->slots || cache->slots_ret) { in alloc_swap_slot_cache()
139 /* cache already allocated */ in alloc_swap_slot_cache()
148 if (!cache in alloc_swap_slot_cache()
172 struct swap_slots_cache *cache; drain_slots_cache_cpu() local
261 refill_swap_slots_cache(struct swap_slots_cache *cache) refill_swap_slots_cache() argument
276 struct swap_slots_cache *cache; free_swap_slot() local
309 struct swap_slots_cache *cache; get_swap_page() local
[all...]
/kernel/linux/linux-6.6/mm/
H A Dswap_slots.c3 * Manage cache of swap slots to be used for and returned from
25 * The swap slots cache is protected by a mutex instead of
43 /* Serialize swap slots cache enable/disable operations */
106 /* if global pool of slot caches too low, deactivate cache */ in check_cache_active()
115 struct swap_slots_cache *cache; in alloc_swap_slot_cache() local
136 cache = &per_cpu(swp_slots, cpu); in alloc_swap_slot_cache()
137 if (cache->slots || cache->slots_ret) { in alloc_swap_slot_cache()
138 /* cache already allocated */ in alloc_swap_slot_cache()
147 if (!cache in alloc_swap_slot_cache()
171 struct swap_slots_cache *cache; drain_slots_cache_cpu() local
259 refill_swap_slots_cache(struct swap_slots_cache *cache) refill_swap_slots_cache() argument
274 struct swap_slots_cache *cache; free_swap_slot() local
305 struct swap_slots_cache *cache; folio_alloc_swap() local
[all...]
/kernel/linux/linux-5.10/fs/btrfs/tests/
H A Dfree-space-tests.c10 #include "../free-space-cache.h"
20 static int test_extents(struct btrfs_block_group *cache) in test_extents() argument
27 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents()
33 ret = btrfs_remove_free_space(cache, 0, SZ_4M); in test_extents()
39 if (test_check_exists(cache, 0, SZ_4M)) { in test_extents()
45 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents()
51 ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); in test_extents()
57 ret = btrfs_remove_free_space(cache, 0, SZ_1M); in test_extents()
63 ret = btrfs_remove_free_space(cache, SZ_2M, 4096); in test_extents()
69 if (test_check_exists(cache, in test_extents()
90 test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps() argument
158 test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps_and_extents() argument
333 check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) check_num_extents_and_bitmaps() argument
353 check_cache_empty(struct btrfs_block_group *cache) check_cache_empty() argument
395 test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) test_steal_space_from_bitmap_to_extent() argument
831 struct btrfs_block_group *cache; btrfs_test_free_space_cache() local
[all...]
H A Dfree-space-tree-tests.c21 struct btrfs_block_group *cache, in __check_free_space_extents()
34 info = search_free_space_info(trans, cache, path, 0); in __check_free_space_extents()
51 end = cache->start + cache->length; in __check_free_space_extents()
59 bit = free_space_test_bit(cache, path, offset); in __check_free_space_extents()
108 struct btrfs_block_group *cache, in check_free_space_extents()
117 info = search_free_space_info(trans, cache, path, 0); in check_free_space_extents()
126 ret = __check_free_space_extents(trans, fs_info, cache, path, extents, in check_free_space_extents()
133 ret = convert_free_space_to_extents(trans, cache, path); in check_free_space_extents()
139 ret = convert_free_space_to_bitmaps(trans, cache, pat in check_free_space_extents()
19 __check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) __check_free_space_extents() argument
106 check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) check_free_space_extents() argument
149 test_empty_block_group(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_empty_block_group() argument
163 test_remove_all(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_all() argument
184 test_remove_beginning(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_beginning() argument
207 test_remove_end(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_end() argument
230 test_remove_middle(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_middle() argument
254 test_merge_left(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_left() argument
291 test_merge_right(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_right() argument
329 test_merge_both(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_both() argument
372 test_merge_none(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_none() argument
428 struct btrfs_block_group *cache = NULL; run_test() local
[all...]
/kernel/linux/linux-5.10/drivers/acpi/acpica/
H A Dutcache.c4 * Module Name: utcache - local cache allocation routines
21 * PARAMETERS: cache_name - Ascii name for the cache
23 * max_depth - Maximum depth of the cache (in objects)
24 * return_cache - Where the new cache object is returned
28 * DESCRIPTION: Create a cache object
36 struct acpi_memory_list *cache; in acpi_os_create_cache() local
44 /* Create the cache object */ in acpi_os_create_cache()
46 cache = acpi_os_allocate(sizeof(struct acpi_memory_list)); in acpi_os_create_cache()
47 if (!cache) { in acpi_os_create_cache()
51 /* Populate the cache objec in acpi_os_create_cache()
74 acpi_os_purge_cache(struct acpi_memory_list *cache) acpi_os_purge_cache() argument
120 acpi_os_delete_cache(struct acpi_memory_list *cache) acpi_os_delete_cache() argument
153 acpi_os_release_object(struct acpi_memory_list *cache, void *object) acpi_os_release_object() argument
208 acpi_os_acquire_object(struct acpi_memory_list *cache) acpi_os_acquire_object() argument
[all...]
/kernel/linux/linux-6.6/drivers/acpi/acpica/
H A Dutcache.c4 * Module Name: utcache - local cache allocation routines
21 * PARAMETERS: cache_name - Ascii name for the cache
23 * max_depth - Maximum depth of the cache (in objects)
24 * return_cache - Where the new cache object is returned
28 * DESCRIPTION: Create a cache object
36 struct acpi_memory_list *cache; in acpi_os_create_cache() local
44 /* Create the cache object */ in acpi_os_create_cache()
46 cache = acpi_os_allocate(sizeof(struct acpi_memory_list)); in acpi_os_create_cache()
47 if (!cache) { in acpi_os_create_cache()
51 /* Populate the cache objec in acpi_os_create_cache()
74 acpi_os_purge_cache(struct acpi_memory_list *cache) acpi_os_purge_cache() argument
120 acpi_os_delete_cache(struct acpi_memory_list *cache) acpi_os_delete_cache() argument
153 acpi_os_release_object(struct acpi_memory_list *cache, void *object) acpi_os_release_object() argument
208 acpi_os_acquire_object(struct acpi_memory_list *cache) acpi_os_acquire_object() argument
[all...]
/kernel/linux/linux-5.10/fs/squashfs/
H A Dcache.c8 * cache.c
15 * This file implements a generic cache implementation used for both caches,
16 * plus functions layered ontop of the generic cache implementation to
19 * To avoid out of memory and fragmentation issues with vmalloc the cache
22 * It should be noted that the cache is not used for file datablocks, these
23 * are decompressed and cached in the page-cache in the normal way. The
24 * cache is only used to temporarily cache fragment and metadata blocks
49 * Look-up block in cache, and increment usage count. If not in cache, rea
52 squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length) squashfs_cache_get() argument
175 struct squashfs_cache *cache = entry->cache; squashfs_cache_put() local
197 squashfs_cache_delete(struct squashfs_cache *cache) squashfs_cache_delete() argument
227 struct squashfs_cache *cache = kzalloc(sizeof(*cache), GFP_KERNEL); squashfs_cache_init() local
[all...]
/kernel/linux/linux-6.6/fs/squashfs/
H A Dcache.c8 * cache.c
15 * This file implements a generic cache implementation used for both caches,
16 * plus functions layered ontop of the generic cache implementation to
19 * To avoid out of memory and fragmentation issues with vmalloc the cache
22 * It should be noted that the cache is not used for file datablocks, these
23 * are decompressed and cached in the page-cache in the normal way. The
24 * cache is only used to temporarily cache fragment and metadata blocks
49 * Look-up block in cache, and increment usage count. If not in cache, rea
52 squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length) squashfs_cache_get() argument
175 struct squashfs_cache *cache = entry->cache; squashfs_cache_put() local
197 squashfs_cache_delete(struct squashfs_cache *cache) squashfs_cache_delete() argument
227 struct squashfs_cache *cache = kzalloc(sizeof(*cache), GFP_KERNEL); squashfs_cache_init() local
[all...]
/kernel/linux/linux-6.6/fs/btrfs/tests/
H A Dfree-space-tests.c10 #include "../free-space-cache.h"
20 static int test_extents(struct btrfs_block_group *cache) in test_extents() argument
27 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents()
33 ret = btrfs_remove_free_space(cache, 0, SZ_4M); in test_extents()
39 if (test_check_exists(cache, 0, SZ_4M)) { in test_extents()
45 ret = btrfs_add_free_space(cache, 0, SZ_4M); in test_extents()
51 ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); in test_extents()
57 ret = btrfs_remove_free_space(cache, 0, SZ_1M); in test_extents()
63 ret = btrfs_remove_free_space(cache, SZ_2M, 4096); in test_extents()
69 if (test_check_exists(cache, in test_extents()
90 test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps() argument
158 test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) test_bitmaps_and_extents() argument
333 check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) check_num_extents_and_bitmaps() argument
353 check_cache_empty(struct btrfs_block_group *cache) check_cache_empty() argument
395 test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) test_steal_space_from_bitmap_to_extent() argument
833 test_bytes_index(struct btrfs_block_group *cache, u32 sectorsize) test_bytes_index() argument
1008 struct btrfs_block_group *cache; btrfs_test_free_space_cache() local
[all...]
H A Dfree-space-tree-tests.c22 struct btrfs_block_group *cache, in __check_free_space_extents()
35 info = search_free_space_info(trans, cache, path, 0); in __check_free_space_extents()
52 end = cache->start + cache->length; in __check_free_space_extents()
60 bit = free_space_test_bit(cache, path, offset); in __check_free_space_extents()
109 struct btrfs_block_group *cache, in check_free_space_extents()
118 info = search_free_space_info(trans, cache, path, 0); in check_free_space_extents()
127 ret = __check_free_space_extents(trans, fs_info, cache, path, extents, in check_free_space_extents()
134 ret = convert_free_space_to_extents(trans, cache, path); in check_free_space_extents()
140 ret = convert_free_space_to_bitmaps(trans, cache, pat in check_free_space_extents()
20 __check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) __check_free_space_extents() argument
107 check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) check_free_space_extents() argument
150 test_empty_block_group(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_empty_block_group() argument
164 test_remove_all(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_all() argument
185 test_remove_beginning(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_beginning() argument
208 test_remove_end(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_end() argument
231 test_remove_middle(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_remove_middle() argument
255 test_merge_left(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_left() argument
292 test_merge_right(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_right() argument
330 test_merge_both(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_both() argument
373 test_merge_none(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) test_merge_none() argument
429 struct btrfs_block_group *cache = NULL; run_test() local
[all...]
/kernel/linux/linux-6.6/fs/
H A Dmbcache.c16 * Ext2 and ext4 use this cache for deduplication of extended attribute blocks.
21 * identifies a cache entry.
33 /* Maximum entries in cache to avoid degrading hash too much */
38 /* Number of entries in cache */
41 /* Work for shrinking when the cache has too many entries */
47 static unsigned long mb_cache_shrink(struct mb_cache *cache,
50 static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, in mb_cache_entry_head() argument
53 return &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb_cache_entry_head()
58 * in cache
74 mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, u64 value, bool reusable) mb_cache_entry_create() argument
127 __mb_cache_entry_free(struct mb_cache *cache, struct mb_cache_entry *entry) __mb_cache_entry_free() argument
152 __entry_find(struct mb_cache *cache, struct mb_cache_entry *entry, u32 key) __entry_find() argument
192 mb_cache_entry_find_first(struct mb_cache *cache, u32 key) mb_cache_entry_find_first() argument
209 mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_find_next() argument
222 mb_cache_entry_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_get() argument
254 mb_cache_entry_delete_or_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete_or_get() argument
286 mb_cache_entry_touch(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_touch() argument
296 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_count() local
303 mb_cache_shrink(struct mb_cache *cache, unsigned long nr_to_scan) mb_cache_shrink() argument
336 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_scan() local
346 struct mb_cache *cache = container_of(work, struct mb_cache, mb_cache_shrink_worker() local
359 struct mb_cache *cache; mb_cache_create() local
405 mb_cache_destroy(struct mb_cache *cache) mb_cache_destroy() argument
[all...]
/kernel/linux/linux-5.10/fs/fscache/
H A Dcache.c2 /* FS-Cache cache handling
21 * look up a cache tag
67 * release a reference to a cache tag
86 * select a cache in which to store an object
87 * - the cache addremove semaphore must be at least read-locked by the caller
95 struct fscache_cache *cache; in fscache_select_cache_for_object() local
100 _leave(" = NULL [no cache]"); in fscache_select_cache_for_object()
104 /* we check the parent to determine the cache to use */ in fscache_select_cache_for_object()
108 * cache */ in fscache_select_cache_for_object()
113 cache in fscache_select_cache_for_object()
178 fscache_init_cache(struct fscache_cache *cache, const struct fscache_cache_ops *ops, const char *idfmt, ...) fscache_init_cache() argument
213 fscache_add_cache(struct fscache_cache *cache, struct fscache_object *ifsdef, const char *tagname) fscache_add_cache() argument
313 fscache_io_error(struct fscache_cache *cache) fscache_io_error() argument
325 fscache_withdraw_all_objects(struct fscache_cache *cache, struct list_head *dying_objects) fscache_withdraw_all_objects() argument
361 fscache_withdraw_cache(struct fscache_cache *cache) fscache_withdraw_cache() argument
[all...]
/kernel/linux/linux-5.10/fs/
H A Dmbcache.c16 * Ext2 and ext4 use this cache for deduplication of extended attribute blocks.
21 * identifies a cache entry.
33 /* Maximum entries in cache to avoid degrading hash too much */
38 /* Number of entries in cache */
41 /* Work for shrinking when the cache has too many entries */
47 static unsigned long mb_cache_shrink(struct mb_cache *cache,
50 static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, in mb_cache_entry_head() argument
53 return &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; in mb_cache_entry_head()
58 * in cache
74 mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, u64 value, bool reusable) mb_cache_entry_create() argument
127 __mb_cache_entry_free(struct mb_cache *cache, struct mb_cache_entry *entry) __mb_cache_entry_free() argument
152 __entry_find(struct mb_cache *cache, struct mb_cache_entry *entry, u32 key) __entry_find() argument
192 mb_cache_entry_find_first(struct mb_cache *cache, u32 key) mb_cache_entry_find_first() argument
209 mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_find_next() argument
222 mb_cache_entry_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_get() argument
250 mb_cache_entry_delete(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete() argument
291 mb_cache_entry_delete_or_get(struct mb_cache *cache, u32 key, u64 value) mb_cache_entry_delete_or_get() argument
323 mb_cache_entry_touch(struct mb_cache *cache, struct mb_cache_entry *entry) mb_cache_entry_touch() argument
333 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_count() local
340 mb_cache_shrink(struct mb_cache *cache, unsigned long nr_to_scan) mb_cache_shrink() argument
373 struct mb_cache *cache = container_of(shrink, struct mb_cache, mb_cache_scan() local
383 struct mb_cache *cache = container_of(work, struct mb_cache, mb_cache_shrink_worker() local
396 struct mb_cache *cache; mb_cache_create() local
442 mb_cache_destroy(struct mb_cache *cache) mb_cache_destroy() argument
[all...]
/kernel/linux/linux-5.10/fs/btrfs/
H A Dblock-group.c8 #include "free-space-cache.h"
123 void btrfs_get_block_group(struct btrfs_block_group *cache) in btrfs_get_block_group() argument
125 refcount_inc(&cache->refs); in btrfs_get_block_group()
128 void btrfs_put_block_group(struct btrfs_block_group *cache) in btrfs_put_block_group() argument
130 if (refcount_dec_and_test(&cache->refs)) { in btrfs_put_block_group()
131 WARN_ON(cache->pinned > 0); in btrfs_put_block_group()
132 WARN_ON(cache->reserved > 0); in btrfs_put_block_group()
139 if (WARN_ON(!list_empty(&cache->discard_list))) in btrfs_put_block_group()
140 btrfs_discard_cancel_work(&cache->fs_info->discard_ctl, in btrfs_put_block_group()
141 cache); in btrfs_put_block_group()
165 struct btrfs_block_group *cache; btrfs_add_block_group_cache() local
204 struct btrfs_block_group *cache, *ret = NULL; block_group_cache_tree_search() local
259 btrfs_next_block_group( struct btrfs_block_group *cache) btrfs_next_block_group() argument
367 btrfs_get_caching_control( struct btrfs_block_group *cache) btrfs_get_caching_control() argument
403 btrfs_wait_block_group_cache_progress(struct btrfs_block_group *cache, u64 num_bytes) btrfs_wait_block_group_cache_progress() argument
418 btrfs_wait_block_group_cache_done(struct btrfs_block_group *cache) btrfs_wait_block_group_cache_done() argument
693 btrfs_cache_block_group(struct btrfs_block_group *cache, int load_cache_only) btrfs_cache_block_group() argument
1230 inc_block_group_ro(struct btrfs_block_group *cache, int force) inc_block_group_ro() argument
1741 exclude_super_stripes(struct btrfs_block_group *cache) exclude_super_stripes() argument
1783 link_block_group(struct btrfs_block_group *cache) link_block_group() argument
1796 struct btrfs_block_group *cache; btrfs_create_block_group_cache() local
1888 read_block_group_item(struct btrfs_block_group *cache, struct btrfs_path *path, const struct btrfs_key *key) read_block_group_item() argument
1909 struct btrfs_block_group *cache; read_one_block_group() local
2010 struct btrfs_block_group *cache; btrfs_read_block_groups() local
2158 struct btrfs_block_group *cache; btrfs_make_block_group() local
2237 btrfs_inc_block_group_ro(struct btrfs_block_group *cache, bool do_chunk_alloc) btrfs_inc_block_group_ro() argument
2324 btrfs_dec_block_group_ro(struct btrfs_block_group *cache) btrfs_dec_block_group_ro() argument
2343 update_block_group_item(struct btrfs_trans_handle *trans, struct btrfs_path *path, struct btrfs_block_group *cache) update_block_group_item() argument
2545 struct btrfs_block_group *cache, *tmp; btrfs_setup_space_cache() local
2583 struct btrfs_block_group *cache; btrfs_start_dirty_block_groups() local
2749 struct btrfs_block_group *cache; btrfs_write_dirty_block_groups() local
2874 struct btrfs_block_group *cache = NULL; btrfs_update_block_group() local
2989 btrfs_add_reserved_bytes(struct btrfs_block_group *cache, u64 ram_bytes, u64 num_bytes, int delalloc) btrfs_add_reserved_bytes() argument
3032 btrfs_free_reserved_bytes(struct btrfs_block_group *cache, u64 num_bytes, int delalloc) btrfs_free_reserved_bytes() argument
3416 btrfs_freeze_block_group(struct btrfs_block_group *cache) btrfs_freeze_block_group() argument
[all...]
/kernel/linux/linux-6.6/io_uring/
H A Dalloc_cache.h5 * Don't allow the cache to grow beyond this size.
13 static inline bool io_alloc_cache_put(struct io_alloc_cache *cache, in io_alloc_cache_put() argument
16 if (cache->nr_cached < cache->max_cached) { in io_alloc_cache_put()
17 cache->nr_cached++; in io_alloc_cache_put()
18 wq_stack_add_head(&entry->node, &cache->list); in io_alloc_cache_put()
26 static inline bool io_alloc_cache_empty(struct io_alloc_cache *cache) in io_alloc_cache_empty() argument
28 return !cache->list.next; in io_alloc_cache_empty()
31 static inline struct io_cache_entry *io_alloc_cache_get(struct io_alloc_cache *cache) in io_alloc_cache_get() argument
33 if (cache in io_alloc_cache_get()
46 io_alloc_cache_init(struct io_alloc_cache *cache, unsigned max_nr, size_t size) io_alloc_cache_init() argument
55 io_alloc_cache_free(struct io_alloc_cache *cache, void (*free)(struct io_cache_entry *)) io_alloc_cache_free() argument
[all...]
/kernel/linux/linux-5.10/fs/nfs/
H A Dnfs42xattr.c6 * User extended attribute client side cache functions.
21 * a cache structure attached to NFS inodes. This structure is allocated
22 * when needed, and freed when the cache is zapped.
24 * The cache structure contains as hash table of entries, and a pointer
25 * to a special-cased entry for the listxattr cache.
28 * counting. The cache entries use a similar refcounting scheme.
30 * This makes freeing a cache, both from the shrinker and from the
31 * zap cache path, easy. It also means that, in current use cases,
40 * Two shrinkers deal with the cache entries themselves: one for
45 * The other shrinker frees the cache structure
64 struct nfs4_xattr_cache *cache; global() member
106 nfs4_xattr_hash_init(struct nfs4_xattr_cache *cache) nfs4_xattr_hash_init() argument
270 struct nfs4_xattr_cache *cache; nfs4_xattr_free_cache_cb() local
290 struct nfs4_xattr_cache *cache; nfs4_xattr_alloc_cache() local
310 nfs4_xattr_set_listcache(struct nfs4_xattr_cache *cache, struct nfs4_xattr_entry *new) nfs4_xattr_set_listcache() argument
385 nfs4_xattr_discard_cache(struct nfs4_xattr_cache *cache) nfs4_xattr_discard_cache() argument
430 struct nfs4_xattr_cache *cache, *oldcache, *newcache; nfs4_xattr_get_cache() local
505 nfs4_xattr_hash_bucket(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_bucket() argument
527 nfs4_xattr_hash_add(struct nfs4_xattr_cache *cache, struct nfs4_xattr_entry *entry) nfs4_xattr_hash_add() argument
565 nfs4_xattr_hash_remove(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_remove() argument
588 nfs4_xattr_hash_find(struct nfs4_xattr_cache *cache, const char *name) nfs4_xattr_hash_find() argument
612 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_get() local
651 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_list() local
692 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_add() local
723 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_remove() local
743 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_set_list() local
829 struct nfs4_xattr_cache *cache = container_of(item, cache_lru_isolate() local
863 struct nfs4_xattr_cache *cache; nfs4_xattr_cache_scan() local
894 struct nfs4_xattr_cache *cache; entry_lru_isolate() local
985 struct nfs4_xattr_cache *cache = (struct nfs4_xattr_cache *)p; nfs4_xattr_cache_init_once() local
[all...]

Completed in 17 milliseconds

12345678910>>...70