/kernel/linux/linux-6.6/include/linux/ |
H A D | gfp_types.h | 8 * typedef gfp_t - Memory allocation flags. 16 typedef unsigned int __bitwise gfp_t; 71 #define __GFP_DMA ((__force gfp_t)___GFP_DMA) 72 #define __GFP_HIGHMEM ((__force gfp_t)___GFP_HIGHMEM) 73 #define __GFP_DMA32 ((__force gfp_t)___GFP_DMA32) 74 #define __GFP_MOVABLE ((__force gfp_t)___GFP_MOVABLE) /* ZONE_MOVABLE allowed */ 76 #define __GFP_CMA ((__force gfp_t)___GFP_CMA) 105 #define __GFP_RECLAIMABLE ((__force gfp_t)___GFP_RECLAIMABLE) 106 #define __GFP_WRITE ((__force gfp_t)___GFP_WRITE) 107 #define __GFP_HARDWALL ((__force gfp_t)___GFP_HARDWAL [all...] |
H A D | gfp.h | 16 static inline int gfp_migratetype(const gfp_t gfp_flags) in gfp_migratetype() 40 static inline bool gfpflags_allow_blocking(const gfp_t gfp_flags) in gfpflags_allow_blocking() 65 * zone to use given the lowest 4 bits of gfp_t. Entries are GFP_ZONES_SHIFT 135 static inline enum zone_type gfp_zone(gfp_t flags) in gfp_zone() 153 static inline int gfp_zonelist(gfp_t flags) in gfp_zonelist() 171 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist() 183 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, 185 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, 188 unsigned long __alloc_pages_bulk(gfp_t gfp, int preferred_nid, 193 unsigned long alloc_pages_bulk_array_mempolicy(gfp_t gf [all...] |
H A D | slab.h | 226 void * __must_check krealloc(const void *objp, size_t new_size, gfp_t flags) __realloc_size(2); 400 static __always_inline enum kmalloc_cache_type kmalloc_type(gfp_t flags, unsigned long caller) in kmalloc_type() 487 void *__kmalloc(size_t size, gfp_t flags) __assume_kmalloc_alignment __alloc_size(1); 499 void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags) __assume_slab_alignment __malloc; 501 gfp_t gfpflags) __assume_slab_alignment __malloc; 512 int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size, void **p); 519 void *__kmalloc_node(size_t size, gfp_t flags, int node) __assume_kmalloc_alignment 521 void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) __assume_slab_alignment 524 void *kmalloc_trace(struct kmem_cache *s, gfp_t flags, size_t size) 527 void *kmalloc_node_trace(struct kmem_cache *s, gfp_t gfpflag [all...] |
H A D | kmemleak.h | 19 gfp_t gfp) __ref; 21 gfp_t gfp) __ref; 23 gfp_t gfp) __ref; 30 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref; 33 gfp_t gfp) __ref; 39 gfp_t gfp) in kmemleak_alloc_recursive() 62 gfp_t gfp) in kmemleak_alloc() 67 gfp_t gfp) in kmemleak_alloc_recursive() 71 gfp_t gfp) in kmemleak_alloc_percpu() 75 gfp_t gf in kmemleak_vmalloc() [all...] |
H A D | xarray.h | 267 #define XA_FLAGS_LOCK_IRQ ((__force gfp_t)XA_LOCK_IRQ) 268 #define XA_FLAGS_LOCK_BH ((__force gfp_t)XA_LOCK_BH) 269 #define XA_FLAGS_TRACK_FREE ((__force gfp_t)4U) 270 #define XA_FLAGS_ZERO_BUSY ((__force gfp_t)8U) 271 #define XA_FLAGS_ALLOC_WRAPPED ((__force gfp_t)16U) 272 #define XA_FLAGS_ACCOUNT ((__force gfp_t)32U) 273 #define XA_FLAGS_MARK(mark) ((__force gfp_t)((1U << __GFP_BITS_SHIFT) << \ 299 gfp_t xa_flags; 352 void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t); 355 void *entry, gfp_t); [all...] |
H A D | kasan.h | 183 void *object, gfp_t flags, bool init); 185 struct kmem_cache *s, void *object, gfp_t flags, bool init) in kasan_slab_alloc() 193 size_t size, gfp_t flags); 195 const void *object, size_t size, gfp_t flags) in kasan_kmalloc() 203 size_t size, gfp_t flags); 205 size_t size, gfp_t flags) in kasan_kmalloc_large() 213 size_t new_size, gfp_t flags); 215 size_t new_size, gfp_t flags) in kasan_krealloc() 261 gfp_t flags, bool init) in kasan_slab_alloc() 266 size_t size, gfp_t flag in kasan_kmalloc() [all...] |
H A D | fault-inject.h | 94 bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order); 97 bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order); 99 static inline bool __should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in __should_fail_alloc_page() 105 int should_failslab(struct kmem_cache *s, gfp_t gfpflags); 107 extern bool __should_failslab(struct kmem_cache *s, gfp_t gfpflags); 109 static inline bool __should_failslab(struct kmem_cache *s, gfp_t gfpflags) in __should_failslab()
|
H A D | devcoredump.h | 56 gfp_t gfp); 59 void *data, size_t datalen, gfp_t gfp, 65 size_t datalen, gfp_t gfp); 68 size_t datalen, gfp_t gfp) in dev_coredumpv() 75 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() 84 size_t datalen, gfp_t gfp) in dev_coredumpsg()
|
H A D | mempool.h | 13 typedef void * (mempool_alloc_t)(gfp_t gfp_mask, void *pool_data); 41 gfp_t gfp_mask, int node_id); 49 gfp_t gfp_mask, int nid); 53 extern void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) __malloc; 61 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data); 82 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data); 101 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data);
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | gfp.h | 57 #define __GFP_DMA ((__force gfp_t)___GFP_DMA) 58 #define __GFP_HIGHMEM ((__force gfp_t)___GFP_HIGHMEM) 59 #define __GFP_DMA32 ((__force gfp_t)___GFP_DMA32) 60 #define __GFP_MOVABLE ((__force gfp_t)___GFP_MOVABLE) /* ZONE_MOVABLE allowed */ 61 #define __GFP_CMA ((__force gfp_t)___GFP_CMA) 91 #define __GFP_RECLAIMABLE ((__force gfp_t)___GFP_RECLAIMABLE) 92 #define __GFP_WRITE ((__force gfp_t)___GFP_WRITE) 93 #define __GFP_HARDWALL ((__force gfp_t)___GFP_HARDWALL) 94 #define __GFP_THISNODE ((__force gfp_t)___GFP_THISNODE) 95 #define __GFP_ACCOUNT ((__force gfp_t)___GFP_ACCOUN [all...] |
H A D | kmemleak.h | 19 gfp_t gfp) __ref; 21 gfp_t gfp) __ref; 23 gfp_t gfp) __ref; 30 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref; 33 gfp_t gfp) __ref; 40 gfp_t gfp) in kmemleak_alloc_recursive() 63 gfp_t gfp) in kmemleak_alloc() 68 gfp_t gfp) in kmemleak_alloc_recursive() 72 gfp_t gfp) in kmemleak_alloc_percpu() 76 gfp_t gf in kmemleak_vmalloc() [all...] |
H A D | slab.h | 184 void * __must_check krealloc(const void *, size_t, gfp_t); 318 static __always_inline enum kmalloc_cache_type kmalloc_type(gfp_t flags) in kmalloc_type() 389 void *__kmalloc(size_t size, gfp_t flags) __assume_kmalloc_alignment __malloc; 390 void *kmem_cache_alloc(struct kmem_cache *, gfp_t flags) __assume_slab_alignment __malloc; 401 int kmem_cache_alloc_bulk(struct kmem_cache *, gfp_t, size_t, void **); 413 void *__kmalloc_node(size_t size, gfp_t flags, int node) __assume_kmalloc_alignment __malloc; 414 void *kmem_cache_alloc_node(struct kmem_cache *, gfp_t flags, int node) __assume_slab_alignment __malloc; 416 static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) in __kmalloc_node() 421 static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) in kmem_cache_alloc_node() 428 extern void *kmem_cache_alloc_trace(struct kmem_cache *, gfp_t, size_ [all...] |
H A D | xarray.h | 263 #define XA_FLAGS_LOCK_IRQ ((__force gfp_t)XA_LOCK_IRQ) 264 #define XA_FLAGS_LOCK_BH ((__force gfp_t)XA_LOCK_BH) 265 #define XA_FLAGS_TRACK_FREE ((__force gfp_t)4U) 266 #define XA_FLAGS_ZERO_BUSY ((__force gfp_t)8U) 267 #define XA_FLAGS_ALLOC_WRAPPED ((__force gfp_t)16U) 268 #define XA_FLAGS_ACCOUNT ((__force gfp_t)32U) 269 #define XA_FLAGS_MARK(mark) ((__force gfp_t)((1U << __GFP_BITS_SHIFT) << \ 295 gfp_t xa_flags; 348 void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t); 351 void *entry, gfp_t); [all...] |
H A D | devcoredump.h | 56 gfp_t gfp); 59 void *data, size_t datalen, gfp_t gfp, 65 size_t datalen, gfp_t gfp); 68 size_t datalen, gfp_t gfp) in dev_coredumpv() 75 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() 84 size_t datalen, gfp_t gfp) in dev_coredumpsg()
|
H A D | mempool.h | 13 typedef void * (mempool_alloc_t)(gfp_t gfp_mask, void *pool_data); 36 gfp_t gfp_mask, int node_id); 44 gfp_t gfp_mask, int nid); 48 extern void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) __malloc; 56 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data); 77 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data); 96 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data);
|
H A D | cpuset.h | 69 extern bool __cpuset_node_allowed(int node, gfp_t gfp_mask); 71 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 78 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 83 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() 211 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 216 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 221 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
|
/kernel/linux/linux-5.10/include/net/sctp/ |
H A D | ulpevent.h | 81 gfp_t gfp); 90 gfp_t gfp); 96 gfp_t gfp); 103 gfp_t gfp); 108 gfp_t gfp); 113 __u32 flags, gfp_t gfp); 116 const struct sctp_association *asoc, gfp_t gfp); 120 gfp_t gfp); 124 __u32 indication, gfp_t gfp); 127 const struct sctp_association *asoc, gfp_t gf [all...] |
H A D | stream_interleave.h | 25 int len, __u8 flags, gfp_t gfp); 29 struct sctp_chunk *chunk, gfp_t gfp); 33 struct sctp_chunk *chunk, gfp_t gfp); 34 void (*start_pd)(struct sctp_ulpq *ulpq, gfp_t gfp); 35 void (*abort_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
|
/kernel/linux/linux-6.6/include/net/sctp/ |
H A D | ulpevent.h | 81 gfp_t gfp); 90 gfp_t gfp); 96 gfp_t gfp); 103 gfp_t gfp); 108 gfp_t gfp); 113 __u32 flags, gfp_t gfp); 116 const struct sctp_association *asoc, gfp_t gfp); 120 gfp_t gfp); 124 __u32 indication, gfp_t gfp); 127 const struct sctp_association *asoc, gfp_t gf [all...] |
H A D | stream_interleave.h | 25 int len, __u8 flags, gfp_t gfp); 29 struct sctp_chunk *chunk, gfp_t gfp); 33 struct sctp_chunk *chunk, gfp_t gfp); 34 void (*start_pd)(struct sctp_ulpq *ulpq, gfp_t gfp); 35 void (*abort_pd)(struct sctp_ulpq *ulpq, gfp_t gfp);
|
/kernel/linux/linux-5.10/net/wireless/ |
H A D | nl80211.h | 64 const u8 *buf, size_t len, gfp_t gfp); 67 const u8 *buf, size_t len, gfp_t gfp, 72 const u8 *buf, size_t len, gfp_t gfp); 75 const u8 *buf, size_t len, gfp_t gfp); 78 const u8 *addr, gfp_t gfp); 81 const u8 *addr, gfp_t gfp); 85 gfp_t gfp); 88 struct cfg80211_roam_info *info, gfp_t gfp); 99 int key_id, const u8 *tsc, gfp_t gfp); 108 gfp_t gf [all...] |
/kernel/linux/linux-6.6/net/wireless/ |
H A D | nl80211.h | 60 const u8 *buf, size_t len, gfp_t gfp); 67 bool reconnect, gfp_t gfp); 71 bool reconnect, gfp_t gfp); 74 const u8 *addr, gfp_t gfp); 77 const u8 *addr, gfp_t gfp); 81 gfp_t gfp); 84 struct cfg80211_roam_info *info, gfp_t gfp); 96 int key_id, const u8 *tsc, gfp_t gfp); 105 gfp_t gfp); 109 struct cfg80211_rx_info *info, gfp_t gf [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | swap.h | 35 gfp_t gfp, void **shadowp); 47 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, 51 struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, 55 struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t flag, 57 struct page *swapin_readahead(swp_entry_t entry, gfp_t flag, 84 gfp_t gfp_mask, struct vm_fault *vmf) in swap_cluster_readahead() 89 static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, in swapin_readahead() 128 gfp_t gfp_mask, void **shadowp) in add_to_swap_cache()
|
/kernel/linux/linux-6.6/include/net/ |
H A D | handshake.h | 36 int tls_client_hello_anon(const struct tls_handshake_args *args, gfp_t flags); 37 int tls_client_hello_x509(const struct tls_handshake_args *args, gfp_t flags); 38 int tls_client_hello_psk(const struct tls_handshake_args *args, gfp_t flags); 39 int tls_server_hello_x509(const struct tls_handshake_args *args, gfp_t flags); 40 int tls_server_hello_psk(const struct tls_handshake_args *args, gfp_t flags);
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | cpuset.h | 68 extern bool _cpuset_node_allowed(int node, gfp_t gfp_mask); 70 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 78 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 83 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() 228 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 233 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 238 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
|