Lines Matching defs:new
69 * the partial slab counter. If taken then no new slabs may be added or
1580 /* Move object to the new freelist */
1916 struct page new;
1927 new.counters = counters;
1928 *objects = new.objects - new.inuse;
1930 new.inuse = page->objects;
1931 new.freelist = NULL;
1933 new.freelist = freelist;
1936 VM_BUG_ON(new.frozen);
1937 new.frozen = 1;
1941 new.freelist, new.counters,
2166 struct page new;
2198 new.counters = counters;
2199 new.inuse--;
2200 VM_BUG_ON(!new.frozen);
2204 freelist, new.counters,
2231 new.counters = old.counters;
2233 new.inuse--;
2235 new.freelist = freelist;
2237 new.freelist = old.freelist;
2239 new.frozen = 0;
2241 if (!new.inuse && n->nr_partial >= s->min_partial)
2243 else if (new.freelist) {
2284 new.freelist, new.counters,
2321 struct page new;
2341 new.counters = old.counters;
2342 new.freelist = old.freelist;
2344 new.frozen = 0;
2348 new.freelist, new.counters,
2351 if (unlikely(!new.inuse && n->nr_partial >= s->min_partial)) {
2626 struct page new;
2634 new.counters = counters;
2635 VM_BUG_ON(!new.frozen);
2637 new.inuse = page->objects;
2638 new.frozen = freelist != NULL;
2642 NULL, new.counters,
2652 * Processing is still very fast if new objects have been freed to the
2660 * And if we were unable to get a new slab from the partial slab lists then
2661 * we need to allocate a new slab. This is the slowest path since it involves
2662 * a call to the page allocator and the setup of a new slab.
2976 struct page new;
2995 new.counters = counters;
2996 was_frozen = new.frozen;
2997 new.inuse -= cnt;
2998 if ((!new.inuse || !prior) && !was_frozen) {
3008 new.frozen = 1;
3028 head, new.counters,
3039 } else if (new.frozen) {
3051 if (unlikely(!new.inuse && n->nr_partial >= s->min_partial))
3229 /* Start new detached freelist */
4694 * Not found. Insert new tracking element.