Lines Matching refs:list
203 list_for_each_entry_reverse(s, &slab_caches, list) {
260 list_add(&s->list, &slab_caches);
420 * @slab_caches_to_rcu_destroy list. The slab pages are freed
436 list_for_each_entry_safe(s, s2, &to_destroy, list) {
453 list_del(&s->list);
459 list_add_tail(&s->list, &slab_caches_to_rcu_destroy);
580 list_add(&s->list, &slab_caches);
860 static void freelist_randomize(struct rnd_state *state, unsigned int *list,
867 list[i] = i;
873 swap(list[i], list[rand]);
970 struct kmem_cache *s = list_entry(p, struct kmem_cache, list);
987 * Use mutex_trylock to protect the list traverse, dump nothing
998 list_for_each_entry_safe(s, s2, &slab_caches, list) {