/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_mru_cache.c | 10 * The MRU Cache data structure consists of a data store, an array of lists and 16 * The group count controls how many lists are created, and thereby how finely 18 * all the lists whose time has expired are deleted. 22 * five. Five internal lists will be created, each representing a two second 62 * current list. Doubly linked lists are used to make the list maintenance 64 * inserts and lookups can occur when there are no free lists available. When 78 * both the data store and to one of the lists. It must also be possible to 88 * likely result in a loop in one of the lists. That's a sure-fire recipe for 93 struct list_head *lists; /* Array of lists, on member [all...] |
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_mru_cache.c | 10 * The MRU Cache data structure consists of a data store, an array of lists and 16 * The group count controls how many lists are created, and thereby how finely 18 * all the lists whose time has expired are deleted. 22 * five. Five internal lists will be created, each representing a two second 62 * current list. Doubly linked lists are used to make the list maintenance 64 * inserts and lookups can occur when there are no free lists available. When 78 * both the data store and to one of the lists. It must also be possible to 88 * likely result in a loop in one of the lists. That's a sure-fire recipe for 93 struct list_head *lists; /* Array of lists, on member [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/ |
H A D | selftest_workarounds.c | 62 reference_lists_init(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_init() argument 67 memset(lists, 0, sizeof(*lists)); in reference_lists_init() 69 wa_init_start(&lists->gt_wa_list, gt, "GT_REF", "global"); in reference_lists_init() 70 gt_init_workarounds(gt, &lists->gt_wa_list); in reference_lists_init() 71 wa_init_finish(&lists->gt_wa_list); in reference_lists_init() 74 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init() 81 &lists->engine[id].ctx_wa_list, in reference_lists_init() 87 reference_lists_fini(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_fini() argument 93 intel_wa_list_free(&lists in reference_lists_fini() 1166 verify_wa_lists(struct intel_gt *gt, struct wa_lists *lists, const char *str) verify_wa_lists() argument 1201 struct wa_lists *lists; live_gpu_reset_workarounds() local 1245 struct wa_lists *lists; live_engine_reset_workarounds() local [all...] |
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | bpf_lru_list.c | 33 return &loc_l->lists[LOCAL_FREE_LIST_IDX]; in local_free_list() 38 return &loc_l->lists[LOCAL_PENDING_LIST_IDX]; in local_pending_list() 98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in() 126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move() 147 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE]; in __bpf_lru_list_rotate_active() 174 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_rotate_inactive() 217 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink_inactive() 277 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE])) in __bpf_lru_list_shrink() 278 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink() 280 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIV in __bpf_lru_list_shrink() [all...] |
H A D | bpf_lru_list.h | 31 struct list_head lists[NR_BPF_LRU_LIST_T]; member 40 struct list_head lists[NR_BPF_LRU_LOCAL_LIST_T]; member
|
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | bpf_lru_list.c | 33 return &loc_l->lists[LOCAL_FREE_LIST_IDX]; in local_free_list() 38 return &loc_l->lists[LOCAL_PENDING_LIST_IDX]; in local_pending_list() 98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in() 126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move() 147 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE]; in __bpf_lru_list_rotate_active() 174 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_rotate_inactive() 217 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink_inactive() 277 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE])) in __bpf_lru_list_shrink() 278 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink() 280 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIV in __bpf_lru_list_shrink() [all...] |
H A D | bpf_lru_list.h | 32 struct list_head lists[NR_BPF_LRU_LIST_T]; member 41 struct list_head lists[NR_BPF_LRU_LOCAL_LIST_T]; member
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
H A D | selftest_workarounds.c | 61 reference_lists_init(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_init() argument 66 memset(lists, 0, sizeof(*lists)); in reference_lists_init() 68 wa_init_start(&lists->gt_wa_list, "GT_REF", "global"); in reference_lists_init() 69 gt_init_workarounds(gt->i915, &lists->gt_wa_list); in reference_lists_init() 70 wa_init_finish(&lists->gt_wa_list); in reference_lists_init() 73 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init() 80 &lists->engine[id].ctx_wa_list, in reference_lists_init() 86 reference_lists_fini(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_fini() argument 92 intel_wa_list_free(&lists in reference_lists_fini() 1131 verify_wa_lists(struct i915_gem_context *ctx, struct wa_lists *lists, const char *str) verify_wa_lists() argument 1162 struct wa_lists lists; live_gpu_reset_workarounds() local 1209 struct wa_lists lists; live_engine_reset_workarounds() local [all...] |
/kernel/linux/linux-5.10/scripts/gdb/linux/ |
H A D | modules.py | 16 from linux import cpus, utils, lists namespace 30 for module in lists.list_for_each_entry(modules, module_ptr_type, "list"): 86 for use in lists.list_for_each_entry(sources, t, "source_list"):
|
H A D | clk.py | 8 from linux import utils, lists, constants namespace 14 return lists.hlist_for_each_entry(hlist_head,
|
H A D | proc.py | 18 from linux import lists namespace 173 for vfs in lists.list_for_each_entry(namespace['list'],
|
H A D | genpd.py | 9 from linux.lists import list_for_each_entry
|
/kernel/linux/linux-5.10/scripts/ |
H A D | dev-needs.sh | 14 /sys/devices and then lists the probe dependency chain (suppliers and 19 By default it lists the full path to the devices under /sys/devices. 25 -c lists the compatible string of the dependencies 26 -d lists the driver name of the dependencies that have probed 27 -m lists the module name of the dependencies that have a module
|
/kernel/linux/linux-6.6/scripts/ |
H A D | dev-needs.sh | 14 /sys/devices and then lists the probe dependency chain (suppliers and 19 By default it lists the full path to the devices under /sys/devices. 25 -c lists the compatible string of the dependencies 26 -d lists the driver name of the dependencies that have probed 27 -m lists the module name of the dependencies that have a module
|
/kernel/linux/linux-6.6/scripts/gdb/linux/ |
H A D | modules.py | 16 from linux import cpus, utils, lists, constants namespace 30 for module in lists.list_for_each_entry(modules, module_ptr_type, "list"): 92 for use in lists.list_for_each_entry(sources, t, "source_list"):
|
H A D | clk.py | 8 from linux import utils, lists, constants namespace 14 return lists.hlist_for_each_entry(hlist_head,
|
H A D | vmalloc.py | 11 from linux import lists, utils, stackdepot, constants, mm namespace 33 for vmap_area in lists.list_for_each_entry(vmap_area_list, vmap_area_ptr_type, "list"):
|
H A D | slab.py | 12 from linux import lists, utils, stackdepot, constants, mm namespace 165 for slab in lists.list_for_each_entry(slab_list, slab_ptr_type, "slab_list"): 190 for cache in lists.list_for_each_entry(slab_caches, kmem_cache_ptr_type, 'list'): 286 for slab in lists.list_for_each_entry(slab, slab_ptr_type, 'slab_list'): 299 for cache in lists.list_for_each_entry(slab_caches, kmem_cache_ptr_type, 'list'):
|
H A D | proc.py | 19 from linux import lists namespace 175 for mnt in lists.list_for_each_entry(namespace['list'],
|
H A D | genpd.py | 9 from linux.lists import list_for_each_entry
|
/kernel/linux/linux-6.6/mm/ |
H A D | mmzone.c | 83 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init() 90 list_del(&lruvec->lists[LRU_UNEVICTABLE]); in lruvec_init()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | mm_inline.h | 16 * functions that manipulate the LRU lists, to sort a page onto the right LRU 53 list_add(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list() 60 list_add_tail(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list_tail() 95 * lists; and clears its Unevictable or Active flags, ready for freeing. 119 * into the array of LRU lists.
|
/kernel/linux/linux-5.10/scripts/gdb/ |
H A D | vmlinux-gdb.py | 32 import linux.lists namespace
|
/kernel/linux/linux-6.6/scripts/gdb/ |
H A D | vmlinux-gdb.py | 36 import linux.lists namespace
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_capture.c | 26 * Define all device tables of GuC error capture register lists 174 /* List of lists */ 297 const struct __guc_mmio_reg_descr_group *lists) in guc_capture_alloc_steered_lists() 307 list = guc_capture_get_one_list(lists, GUC_CAPTURE_LIST_INDEX_PF, in guc_capture_alloc_steered_lists() 359 const struct __guc_mmio_reg_descr_group *lists; in guc_capture_get_device_reglist() local 362 lists = xe_lp_lists; in guc_capture_get_device_reglist() 364 lists = gen8_lists; in guc_capture_get_device_reglist() 372 guc_capture_alloc_steered_lists(guc, lists); in guc_capture_get_device_reglist() 374 return lists; in guc_capture_get_device_reglist() 506 /* intentional empty lists ca in guc_capture_getlistsize() 296 guc_capture_alloc_steered_lists(struct intel_guc *guc, const struct __guc_mmio_reg_descr_group *lists) guc_capture_alloc_steered_lists() argument [all...] |