Lines Matching refs:gfp_mask
180 gfp_t gfp_mask, int node_id)
190 gfp_mask, node_id);
200 element = pool->alloc(gfp_mask, pool->pool_data);
261 gfp_t gfp_mask, int node_id)
265 pool = kzalloc_node(sizeof(*pool), gfp_mask, node_id);
270 gfp_mask, node_id)) {
363 * @gfp_mask: the usual allocation bitmask.
373 void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask)
380 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO);
381 might_sleep_if(gfp_mask & __GFP_DIRECT_RECLAIM);
383 gfp_mask |= __GFP_NOMEMALLOC; /* don't allocate emergency reserves */
384 gfp_mask |= __GFP_NORETRY; /* don't loop in __alloc_pages */
385 gfp_mask |= __GFP_NOWARN; /* failures are OK */
387 gfp_temp = gfp_mask & ~(__GFP_DIRECT_RECLAIM|__GFP_IO);
413 if (gfp_temp != gfp_mask) {
415 gfp_temp = gfp_mask;
420 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) {
507 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data)
511 return kmem_cache_alloc(mem, gfp_mask);
526 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data)
529 return kmalloc(size, gfp_mask);
543 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data)
546 return alloc_pages(gfp_mask, order);