Searched refs:nr_levels (Results 1 - 6 of 6) sorted by relevance
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm-cache-policy-smq.c | 256 unsigned nr_levels; member 269 static void q_init(struct queue *q, struct entry_space *es, unsigned nr_levels) in q_init() argument 275 q->nr_levels = nr_levels; in q_init() 277 for (i = 0; i < q->nr_levels; i++) { in q_init() 340 max_level = min(max_level, q->nr_levels); in q_peek() 359 struct entry *e = q_peek(q, q->nr_levels, true); in q_pop() 376 for (; level < q->nr_levels; level++) in __redist_pop_from() 388 unsigned level, nr_levels, entries_per_level, remainder; in q_set_targets_subrange_() local 391 BUG_ON(lend > q->nr_levels); in q_set_targets_subrange_() 526 stats_init(struct stats *s, unsigned nr_levels) stats_init() argument [all...] |
/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-cache-policy-smq.c | 257 unsigned int nr_levels; member 270 static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels) in q_init() argument 276 q->nr_levels = nr_levels; in q_init() 278 for (i = 0; i < q->nr_levels; i++) { in q_init() 341 max_level = min(max_level, q->nr_levels); in q_peek() 360 struct entry *e = q_peek(q, q->nr_levels, true); in q_pop() 377 for (; level < q->nr_levels; level++) in __redist_pop_from() 390 unsigned int level, nr_levels, entries_per_level, remainder; in q_set_targets_subrange_() local 393 BUG_ON(lend > q->nr_levels); in q_set_targets_subrange_() 528 stats_init(struct stats *s, unsigned int nr_levels) stats_init() argument [all...] |
/kernel/linux/linux-5.10/drivers/thermal/ |
H A D | cpufreq_cooling.c | 324 unsigned int nr_levels; in em_is_sane() local 337 nr_levels = cpufreq_cdev->max_level + 1; in em_is_sane() 338 if (em_pd_nr_perf_states(em) != nr_levels) { in em_is_sane() 341 em_pd_nr_perf_states(em), nr_levels); in em_is_sane()
|
/kernel/linux/linux-6.6/drivers/thermal/ |
H A D | cpufreq_cooling.c | 317 unsigned int nr_levels; in em_is_sane() local 330 nr_levels = cpufreq_cdev->max_level + 1; in em_is_sane() 331 if (em_pd_nr_perf_states(em) != nr_levels) { in em_is_sane() 334 em_pd_nr_perf_states(em), nr_levels); in em_is_sane()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | topology.c | 1838 int nr_levels = 0; in sched_init_numa() local 1869 nr_levels = bitmap_weight(distance_map, NR_DISTANCE_VALUES); in sched_init_numa() 1871 distances = kcalloc(nr_levels, sizeof(int), GFP_KERNEL); in sched_init_numa() 1877 for (i = 0, j = 0; i < nr_levels; i++, j++) { in sched_init_numa() 1886 * 'nr_levels' contains the number of unique distances in sched_init_numa() 1895 * the array will contain less then 'nr_levels' members. This could be in sched_init_numa() 1899 * We reset it to 'nr_levels' at the end of this function. in sched_init_numa() 1903 masks = kzalloc(sizeof(void *) * nr_levels, GFP_KERNEL); in sched_init_numa() 1911 for (i = 0; i < nr_levels; i++) { in sched_init_numa() 1941 tl = kzalloc((i + nr_levels in sched_init_numa() 1986 int nr_levels, *distances; sched_reset_numa() local [all...] |
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | topology.c | 1635 int nr_levels = 0; in sched_init_numa() local 1663 nr_levels = bitmap_weight(distance_map, NR_DISTANCE_VALUES); in sched_init_numa() 1665 sched_domains_numa_distance = kcalloc(nr_levels, sizeof(int), GFP_KERNEL); in sched_init_numa() 1671 for (i = 0, j = 0; i < nr_levels; i++, j++) { in sched_init_numa() 1679 * 'nr_levels' contains the number of unique distances in sched_init_numa() 1688 * the array will contain less then 'nr_levels' members. This could be in sched_init_numa() 1692 * We reset it to 'nr_levels' at the end of this function. in sched_init_numa() 1696 sched_domains_numa_masks = kzalloc(sizeof(void *) * nr_levels, GFP_KERNEL); in sched_init_numa() 1704 for (i = 0; i < nr_levels; i++) { in sched_init_numa() 1734 tl = kzalloc((i + nr_levels in sched_init_numa() [all...] |
Completed in 12 milliseconds