Lines Matching defs:cache
3 * SLUB: A slab allocator that limits cache line use instead of queuing
52 * and to synchronize major metadata changes to slab cache structures.
188 * disabled when slub_debug=O is used and a cache's min order increases with
240 * Core slab cache functions
245 * with an XOR of the address where the pointer is held and a per-cache
257 * example, when __free_slub() iterates over objects in a cache, it
1400 * kmem_cache_flags - apply debugging options to the cache
1403 * @name: name of the cache
1634 /* Pre-initialize the random sequence cache */
1661 /* Initialize each random sequence freelist per cache */
1684 * page might be smaller than the usual size defined by the cache.
2548 pr_warn(" cache: %s, object size: %u, buffer size: %u, default order: %u, min order: %u\n",
2965 * So we still attempt to reduce cache line usage. Just take the slab
3157 void ___cache_free(struct kmem_cache *cache, void *x, unsigned long addr)
3159 do_slab_free(cache, virt_to_head_page(x), x, NULL, 1, addr);
3908 * Release all resources used by a slab cache.
4033 * cache's usercopy region.
4035 * Returns NULL if check passes, otherwise const char * to name of cache
4394 /* Setup random freelists for each cache */
5556 /* Create a unique string id for a slab cache: