Lines Matching refs:gfp
191 static void *slob_new_pages(gfp_t gfp, int order, int node)
197 page = __alloc_pages_node(node, gfp, order);
200 page = alloc_pages(gfp, order);
301 static void *slob_alloc(size_t size, gfp_t gfp, int align, int node,
358 b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node);
374 if (unlikely(gfp & __GFP_ZERO))
469 __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller)
475 gfp &= gfp_allowed_mask;
477 fs_reclaim_acquire(gfp);
478 fs_reclaim_release(gfp);
493 m = slob_alloc(size + minalign, gfp, align, node, minalign);
501 size, size + minalign, gfp, node);
506 gfp |= __GFP_COMP;
507 ret = slob_new_pages(gfp, order, node);
510 size, PAGE_SIZE << order, gfp, node);
513 kmemleak_alloc(ret, size, 1, gfp);
517 void *__kmalloc(size_t size, gfp_t gfp)
519 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, _RET_IP_);
523 void *__kmalloc_track_caller(size_t size, gfp_t gfp, unsigned long caller)
525 return __do_kmalloc_node(size, gfp, NUMA_NO_NODE, caller);
530 void *__kmalloc_node_track_caller(size_t size, gfp_t gfp,
533 return __do_kmalloc_node(size, gfp, node, caller);
631 void *__kmalloc_node(size_t size, gfp_t gfp, int node)
633 return __do_kmalloc_node(size, gfp, node, _RET_IP_);
637 void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node)
639 return slob_alloc_node(cachep, gfp, node);