Home
last modified time | relevance | path

Searched refs:lists (Results 1 - 25 of 64) sorted by relevance

123

/kernel/linux/linux-5.10/fs/xfs/
H A Dxfs_mru_cache.c10 * The MRU Cache data structure consists of a data store, an array of lists and
16 * The group count controls how many lists are created, and thereby how finely
18 * all the lists whose time has expired are deleted.
22 * five. Five internal lists will be created, each representing a two second
62 * current list. Doubly linked lists are used to make the list maintenance
64 * inserts and lookups can occur when there are no free lists available. When
78 * both the data store and to one of the lists. It must also be possible to
88 * likely result in a loop in one of the lists. That's a sure-fire recipe for
93 struct list_head *lists; /* Array of lists, on member
[all...]
/kernel/linux/linux-6.6/fs/xfs/
H A Dxfs_mru_cache.c10 * The MRU Cache data structure consists of a data store, an array of lists and
16 * The group count controls how many lists are created, and thereby how finely
18 * all the lists whose time has expired are deleted.
22 * five. Five internal lists will be created, each representing a two second
62 * current list. Doubly linked lists are used to make the list maintenance
64 * inserts and lookups can occur when there are no free lists available. When
78 * both the data store and to one of the lists. It must also be possible to
88 * likely result in a loop in one of the lists. That's a sure-fire recipe for
93 struct list_head *lists; /* Array of lists, on member
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/
H A Dselftest_workarounds.c62 reference_lists_init(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_init() argument
67 memset(lists, 0, sizeof(*lists)); in reference_lists_init()
69 wa_init_start(&lists->gt_wa_list, gt, "GT_REF", "global"); in reference_lists_init()
70 gt_init_workarounds(gt, &lists->gt_wa_list); in reference_lists_init()
71 wa_init_finish(&lists->gt_wa_list); in reference_lists_init()
74 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init()
81 &lists->engine[id].ctx_wa_list, in reference_lists_init()
87 reference_lists_fini(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_fini() argument
93 intel_wa_list_free(&lists in reference_lists_fini()
1166 verify_wa_lists(struct intel_gt *gt, struct wa_lists *lists, const char *str) verify_wa_lists() argument
1201 struct wa_lists *lists; live_gpu_reset_workarounds() local
1245 struct wa_lists *lists; live_engine_reset_workarounds() local
[all...]
/kernel/linux/linux-5.10/kernel/bpf/
H A Dbpf_lru_list.c33 return &loc_l->lists[LOCAL_FREE_LIST_IDX]; in local_free_list()
38 return &loc_l->lists[LOCAL_PENDING_LIST_IDX]; in local_pending_list()
98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in()
126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move()
147 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE]; in __bpf_lru_list_rotate_active()
174 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_rotate_inactive()
217 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink_inactive()
277 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE])) in __bpf_lru_list_shrink()
278 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink()
280 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIV in __bpf_lru_list_shrink()
[all...]
H A Dbpf_lru_list.h31 struct list_head lists[NR_BPF_LRU_LIST_T]; member
40 struct list_head lists[NR_BPF_LRU_LOCAL_LIST_T]; member
/kernel/linux/linux-6.6/kernel/bpf/
H A Dbpf_lru_list.c33 return &loc_l->lists[LOCAL_FREE_LIST_IDX]; in local_free_list()
38 return &loc_l->lists[LOCAL_PENDING_LIST_IDX]; in local_pending_list()
98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in()
126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move()
147 struct list_head *active = &l->lists[BPF_LRU_LIST_T_ACTIVE]; in __bpf_lru_list_rotate_active()
174 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_rotate_inactive()
217 struct list_head *inactive = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink_inactive()
277 if (!list_empty(&l->lists[BPF_LRU_LIST_T_INACTIVE])) in __bpf_lru_list_shrink()
278 force_shrink_list = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in __bpf_lru_list_shrink()
280 force_shrink_list = &l->lists[BPF_LRU_LIST_T_ACTIV in __bpf_lru_list_shrink()
[all...]
H A Dbpf_lru_list.h32 struct list_head lists[NR_BPF_LRU_LIST_T]; member
41 struct list_head lists[NR_BPF_LRU_LOCAL_LIST_T]; member
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
H A Dselftest_workarounds.c61 reference_lists_init(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_init() argument
66 memset(lists, 0, sizeof(*lists)); in reference_lists_init()
68 wa_init_start(&lists->gt_wa_list, "GT_REF", "global"); in reference_lists_init()
69 gt_init_workarounds(gt->i915, &lists->gt_wa_list); in reference_lists_init()
70 wa_init_finish(&lists->gt_wa_list); in reference_lists_init()
73 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init()
80 &lists->engine[id].ctx_wa_list, in reference_lists_init()
86 reference_lists_fini(struct intel_gt *gt, struct wa_lists *lists) in reference_lists_fini() argument
92 intel_wa_list_free(&lists in reference_lists_fini()
1131 verify_wa_lists(struct i915_gem_context *ctx, struct wa_lists *lists, const char *str) verify_wa_lists() argument
1162 struct wa_lists lists; live_gpu_reset_workarounds() local
1209 struct wa_lists lists; live_engine_reset_workarounds() local
[all...]
/kernel/linux/linux-5.10/scripts/gdb/linux/
H A Dmodules.py16 from linux import cpus, utils, lists namespace
30 for module in lists.list_for_each_entry(modules, module_ptr_type, "list"):
86 for use in lists.list_for_each_entry(sources, t, "source_list"):
H A Dclk.py8 from linux import utils, lists, constants namespace
14 return lists.hlist_for_each_entry(hlist_head,
H A Dproc.py18 from linux import lists namespace
173 for vfs in lists.list_for_each_entry(namespace['list'],
H A Dgenpd.py9 from linux.lists import list_for_each_entry
/kernel/linux/linux-5.10/scripts/
H A Ddev-needs.sh14 /sys/devices and then lists the probe dependency chain (suppliers and
19 By default it lists the full path to the devices under /sys/devices.
25 -c lists the compatible string of the dependencies
26 -d lists the driver name of the dependencies that have probed
27 -m lists the module name of the dependencies that have a module
/kernel/linux/linux-6.6/scripts/
H A Ddev-needs.sh14 /sys/devices and then lists the probe dependency chain (suppliers and
19 By default it lists the full path to the devices under /sys/devices.
25 -c lists the compatible string of the dependencies
26 -d lists the driver name of the dependencies that have probed
27 -m lists the module name of the dependencies that have a module
/kernel/linux/linux-6.6/scripts/gdb/linux/
H A Dmodules.py16 from linux import cpus, utils, lists, constants namespace
30 for module in lists.list_for_each_entry(modules, module_ptr_type, "list"):
92 for use in lists.list_for_each_entry(sources, t, "source_list"):
H A Dclk.py8 from linux import utils, lists, constants namespace
14 return lists.hlist_for_each_entry(hlist_head,
H A Dvmalloc.py11 from linux import lists, utils, stackdepot, constants, mm namespace
33 for vmap_area in lists.list_for_each_entry(vmap_area_list, vmap_area_ptr_type, "list"):
H A Dslab.py12 from linux import lists, utils, stackdepot, constants, mm namespace
165 for slab in lists.list_for_each_entry(slab_list, slab_ptr_type, "slab_list"):
190 for cache in lists.list_for_each_entry(slab_caches, kmem_cache_ptr_type, 'list'):
286 for slab in lists.list_for_each_entry(slab, slab_ptr_type, 'slab_list'):
299 for cache in lists.list_for_each_entry(slab_caches, kmem_cache_ptr_type, 'list'):
H A Dproc.py19 from linux import lists namespace
175 for mnt in lists.list_for_each_entry(namespace['list'],
H A Dgenpd.py9 from linux.lists import list_for_each_entry
/kernel/linux/linux-6.6/mm/
H A Dmmzone.c83 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
90 list_del(&lruvec->lists[LRU_UNEVICTABLE]); in lruvec_init()
/kernel/linux/linux-5.10/include/linux/
H A Dmm_inline.h16 * functions that manipulate the LRU lists, to sort a page onto the right LRU
53 list_add(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list()
60 list_add_tail(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list_tail()
95 * lists; and clears its Unevictable or Active flags, ready for freeing.
119 * into the array of LRU lists.
/kernel/linux/linux-5.10/scripts/gdb/
H A Dvmlinux-gdb.py32 import linux.lists namespace
/kernel/linux/linux-6.6/scripts/gdb/
H A Dvmlinux-gdb.py36 import linux.lists namespace
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_capture.c26 * Define all device tables of GuC error capture register lists
174 /* List of lists */
297 const struct __guc_mmio_reg_descr_group *lists) in guc_capture_alloc_steered_lists()
307 list = guc_capture_get_one_list(lists, GUC_CAPTURE_LIST_INDEX_PF, in guc_capture_alloc_steered_lists()
359 const struct __guc_mmio_reg_descr_group *lists; in guc_capture_get_device_reglist() local
362 lists = xe_lp_lists; in guc_capture_get_device_reglist()
364 lists = gen8_lists; in guc_capture_get_device_reglist()
372 guc_capture_alloc_steered_lists(guc, lists); in guc_capture_get_device_reglist()
374 return lists; in guc_capture_get_device_reglist()
506 /* intentional empty lists ca in guc_capture_getlistsize()
296 guc_capture_alloc_steered_lists(struct intel_guc *guc, const struct __guc_mmio_reg_descr_group *lists) guc_capture_alloc_steered_lists() argument
[all...]

Completed in 11 milliseconds

123