Lines Matching defs:size_class

185 struct size_class {
223 struct size_class *size_class[ZS_SIZE_CLASSES];
489 static struct size_class *zspage_class(struct zs_pool *pool,
492 return pool->size_class[zspage->class];
521 static inline void class_stat_inc(struct size_class *class,
527 static inline void class_stat_dec(struct size_class *class,
533 static inline unsigned long zs_stat_get(struct size_class *class, int type)
555 static unsigned long zs_can_compact(struct size_class *class);
561 struct size_class *class;
576 class = pool->size_class[i];
664 static int get_fullness_group(struct size_class *class, struct zspage *zspage)
691 static void insert_zspage(struct size_class *class,
703 static void remove_zspage(struct size_class *class,
722 static int fix_fullness_group(struct size_class *class, struct zspage *zspage)
848 static void __free_zspage(struct zs_pool *pool, struct size_class *class,
879 static void free_zspage(struct zs_pool *pool, struct size_class *class,
900 static void init_zspage(struct size_class *class, struct zspage *zspage)
944 static void create_page_chain(struct size_class *class, struct zspage *zspage,
981 struct size_class *class,
1018 static struct zspage *find_get_zspage(struct size_class *class)
1130 static bool can_merge(struct size_class *prev, int pages_per_zspage,
1140 static bool zspage_full(struct size_class *class, struct zspage *zspage)
1151 * zs_lookup_class_index() - Returns index of the zsmalloc &size_class
1158 * Return: the index of the zsmalloc &size_class that hold objects of the
1163 struct size_class *class;
1165 class = pool->size_class[get_size_class_index(size)];
1200 struct size_class *class;
1261 struct size_class *class;
1290 * zsmalloc &size_class.
1299 * Return: the size (in bytes) of the first huge zsmalloc &size_class.
1313 struct size_class *class;
1319 class = pool->size_class[zspage->class];
1363 struct size_class *class;
1376 class = pool->size_class[get_size_class_index(size)];
1450 struct size_class *class;
1478 static void zs_object_copy(struct size_class *class, unsigned long dst,
1552 static unsigned long find_alloced_obj(struct size_class *class,
1585 struct size_class *class = pool->size_class[src_zspage->class];
1614 static struct zspage *isolate_src_zspage(struct size_class *class)
1631 static struct zspage *isolate_dst_zspage(struct size_class *class)
1655 static int putback_zspage(struct size_class *class, struct zspage *zspage)
1755 static void replace_sub_page(struct size_class *class, struct zspage *zspage,
1802 struct size_class *class;
1906 struct size_class *class;
1915 class = pool->size_class[i];
1931 class = pool->size_class[class_idx];
1972 static unsigned long zs_can_compact(struct size_class *class)
1988 struct size_class *class)
2054 struct size_class *class;
2067 class = pool->size_class[i];
2106 struct size_class *class;
2112 class = pool->size_class[i];
2173 struct size_class *prev_class = NULL;
2191 * Iterate reversely, because, size of size_class that we want to use
2198 struct size_class *class;
2229 * size_class is used for normal zsmalloc operation such
2231 * have one size_class for each size, there is a chance that we
2232 * can get more memory utilization if we use one size_class for
2233 * many different sizes whose size_class have same
2234 * characteristics. So, we makes size_class point to
2235 * previous size_class if possible.
2239 pool->size_class[i] = prev_class;
2244 class = kzalloc(sizeof(struct size_class), GFP_KERNEL);
2252 pool->size_class[i] = class;
2292 struct size_class *class = pool->size_class[i];