/kernel/linux/linux-6.6/include/net/ |
H A D | xdp.h | 228 bool pfmemalloc) in xdp_update_skb_shared_info() 235 skb->pfmemalloc |= pfmemalloc; in xdp_update_skb_shared_info() 226 xdp_update_skb_shared_info(struct sk_buff *skb, u8 nr_frags, unsigned int size, unsigned int truesize, bool pfmemalloc) xdp_update_skb_shared_info() argument
|
/kernel/linux/linux-6.6/net/core/ |
H A D | skbuff.c | 212 #define NAPI_SMALL_PAGE_PFMEMALLOC(nc) ((nc).pfmemalloc) 223 bool pfmemalloc; member 240 nc->pfmemalloc = page_is_pfmemalloc(page); in page_frag_alloc_1k() 459 * takes care of skb->head and skb->pfmemalloc 525 * and skb->pfmemalloc when the data is a page or page fragment. 544 * the caller if emergency pfmemalloc reserves are being used. If it is and 550 bool *pfmemalloc) in kmalloc_reserve() 565 /* Try again but now we are using pfmemalloc reserves */ in kmalloc_reserve() 587 /* Try again but now we are using pfmemalloc reserves */ in kmalloc_reserve() 592 if (pfmemalloc) in kmalloc_reserve() 549 kmalloc_reserve(unsigned int *size, gfp_t flags, int node, bool *pfmemalloc) kmalloc_reserve() argument 626 bool pfmemalloc; __alloc_skb() local 703 bool pfmemalloc; __netdev_alloc_skb() local 777 bool pfmemalloc; __napi_alloc_skb() local [all...] |
H A D | dev.c | 5343 static int __netif_receive_skb_core(struct sk_buff **pskb, bool pfmemalloc, in __netif_receive_skb_core() argument 5394 if (pfmemalloc) in __netif_receive_skb_core() 5429 if (pfmemalloc && !skb_pfmemalloc_protocol(skb)) in __netif_receive_skb_core() 5547 static int __netif_receive_skb_one_core(struct sk_buff *skb, bool pfmemalloc) in __netif_receive_skb_one_core() argument 5553 ret = __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev); in __netif_receive_skb_one_core() 5566 * Caller must also take care of handling if ``(page_is_)pfmemalloc``. 5607 static void __netif_receive_skb_list_core(struct list_head *head, bool pfmemalloc) in __netif_receive_skb_list_core() argument 5631 __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev); in __netif_receive_skb_list_core() 5678 bool pfmemalloc = false; /* Is current sublist PF_MEMALLOC? */ in __netif_receive_skb_list() local 5681 if ((sk_memalloc_socks() && skb_pfmemalloc(skb)) != pfmemalloc) { in __netif_receive_skb_list() [all...] |
H A D | gro.c | 618 if (unlikely(skb->pfmemalloc)) { in napi_reuse_skb()
|
/kernel/linux/linux-5.10/net/core/ |
H A D | skbuff.c | 124 * the caller if emergency pfmemalloc reserves are being used. If it is and 129 #define kmalloc_reserve(size, gfp, node, pfmemalloc) \ 130 __kmalloc_reserve(size, gfp, node, _RET_IP_, pfmemalloc) 133 unsigned long ip, bool *pfmemalloc) in __kmalloc_reserve() 148 /* Try again but now we are using pfmemalloc reserves */ in __kmalloc_reserve() 153 if (pfmemalloc) in __kmalloc_reserve() 154 *pfmemalloc = ret_pfmemalloc; in __kmalloc_reserve() 189 bool pfmemalloc; in __alloc_skb() local 210 data = kmalloc_reserve(size, gfp_mask, node, &pfmemalloc); in __alloc_skb() 228 skb->pfmemalloc in __alloc_skb() 132 __kmalloc_reserve(size_t size, gfp_t flags, int node, unsigned long ip, bool *pfmemalloc) __kmalloc_reserve() argument 435 bool pfmemalloc; __netdev_alloc_skb() local [all...] |
H A D | dev.c | 5163 static int __netif_receive_skb_core(struct sk_buff **pskb, bool pfmemalloc, in __netif_receive_skb_core() argument 5216 if (pfmemalloc) in __netif_receive_skb_core() 5249 if (pfmemalloc && !skb_pfmemalloc_protocol(skb)) in __netif_receive_skb_core() 5367 static int __netif_receive_skb_one_core(struct sk_buff *skb, bool pfmemalloc) in __netif_receive_skb_one_core() argument 5373 ret = __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev); in __netif_receive_skb_one_core() 5386 * Caller must also take care of handling if ``(page_is_)pfmemalloc``. 5427 static void __netif_receive_skb_list_core(struct list_head *head, bool pfmemalloc) in __netif_receive_skb_list_core() argument 5451 __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev); in __netif_receive_skb_list_core() 5498 bool pfmemalloc = false; /* Is current sublist PF_MEMALLOC? */ in __netif_receive_skb_list() local 5501 if ((sk_memalloc_socks() && skb_pfmemalloc(skb)) != pfmemalloc) { in __netif_receive_skb_list() [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | mm_types.h | 269 bool pfmemalloc; member
|
H A D | skbuff.h | 667 * @pfmemalloc: skbuff was allocated from PFMEMALLOC reserves 790 pfmemalloc:1; member 948 return unlikely(skb->pfmemalloc); in skb_pfmemalloc() 2156 * Propagate page pfmemalloc to the skb if we can. The problem is in __skb_fill_page_desc() 2166 skb->pfmemalloc = true; in __skb_fill_page_desc() 2942 * skb_propagate_pfmemalloc - Propagate pfmemalloc if skb is allocated after RX page 2944 * @skb: The skb that may need pfmemalloc set 2950 skb->pfmemalloc = true; in skb_propagate_pfmemalloc()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | mm_types.h | 510 bool pfmemalloc; member
|
H A D | skbuff.h | 786 * @pfmemalloc: skbuff was allocated from PFMEMALLOC reserves 920 pfmemalloc:1, member 1100 return unlikely(skb->pfmemalloc); in skb_pfmemalloc() 2470 /* Propagate page pfmemalloc to the skb if we can. The problem is in __skb_fill_page_desc() 2476 skb->pfmemalloc = true; in __skb_fill_page_desc() 2509 * pfmemalloc, if page is not owned by us. 3354 * skb_propagate_pfmemalloc - Propagate pfmemalloc if skb is allocated after RX page 3356 * @skb: The skb that may need pfmemalloc set 3362 skb->pfmemalloc = true; in skb_propagate_pfmemalloc()
|
/kernel/linux/linux-5.10/mm/ |
H A D | slab.c | 2781 /* Try to find non-pfmemalloc slab if needed */ 2783 struct page *page, bool pfmemalloc) in get_valid_first_slab() 2788 if (pfmemalloc) in get_valid_first_slab() 2794 /* No need to keep pfmemalloc slab if we have enough free objects */ in get_valid_first_slab() 2800 /* Move pfmemalloc slab to the end of list to speed up next search */ in get_valid_first_slab() 2824 static struct page *get_first_slab(struct kmem_cache_node *n, bool pfmemalloc) in get_first_slab() argument 2840 page = get_valid_first_slab(n, page, pfmemalloc); in get_first_slab() 2955 /* Check if we can use obj in pfmemalloc slab */ in cache_alloc_refill() 2782 get_valid_first_slab(struct kmem_cache_node *n, struct page *page, bool pfmemalloc) get_valid_first_slab() argument
|
H A D | page_alloc.c | 2301 * page is set pfmemalloc when ALLOC_NO_WATERMARKS was necessary to in prep_new_page() 5219 nc->pfmemalloc = page_is_pfmemalloc(page); in page_frag_alloc() 5231 if (unlikely(nc->pfmemalloc)) { in page_frag_alloc()
|
/kernel/linux/linux-6.6/mm/ |
H A D | slab.c | 2741 /* Try to find non-pfmemalloc slab if needed */ 2743 struct slab *slab, bool pfmemalloc) in get_valid_first_slab() 2748 if (pfmemalloc) in get_valid_first_slab() 2754 /* No need to keep pfmemalloc slab if we have enough free objects */ in get_valid_first_slab() 2760 /* Move pfmemalloc slab to the end of list to speed up next search */ in get_valid_first_slab() 2784 static struct slab *get_first_slab(struct kmem_cache_node *n, bool pfmemalloc) in get_first_slab() argument 2800 slab = get_valid_first_slab(n, slab, pfmemalloc); in get_first_slab() 2915 /* Check if we can use obj in pfmemalloc slab */ in cache_alloc_refill() 2742 get_valid_first_slab(struct kmem_cache_node *n, struct slab *slab, bool pfmemalloc) get_valid_first_slab() argument
|
H A D | page_alloc.c | 1556 * page is set pfmemalloc when ALLOC_NO_WATERMARKS was necessary to in prep_new_page() 4637 nc->pfmemalloc = page_is_pfmemalloc(page); in page_frag_alloc_align() 4649 if (unlikely(nc->pfmemalloc)) { in page_frag_alloc_align()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_rx.c | 420 skb->pfmemalloc = 1; in fun_handle_cqe_pkt()
|
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_output.c | 1312 * Packets not looped back do not care about pfmemalloc. in __tcp_transmit_skb() 1314 skb->pfmemalloc = 0; in __tcp_transmit_skb()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_output.c | 1322 * Packets not looped back do not care about pfmemalloc. in __tcp_transmit_skb() 1324 skb->pfmemalloc = 0; in __tcp_transmit_skb()
|