/kernel/linux/linux-6.6/drivers/net/ethernet/actions/ |
H A D | owl-emac.c | 206 ring->skbs[i] = skb; in owl_emac_ring_prepare_rx() 252 owl_emac_dma_unmap_rx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_rx() 255 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_rx() 256 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_rx() 271 owl_emac_dma_unmap_tx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_tx() 274 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_tx() 275 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_tx() 288 ring->skbs = devm_kcalloc(dev, size, sizeof(struct sk_buff *), in owl_emac_ring_alloc() 290 if (!ring->skbs) in owl_emac_ring_alloc() 530 ring->skbs[tx_hea in owl_emac_setup_frame_xmit() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/sfc/ |
H A D | selftest.c | 74 * @skbs: An array of skbs transmitted 83 struct sk_buff **skbs; member 424 state->skbs[i] = skb; in efx_begin_loopback() 448 /* Defer cleaning up the other skbs for the caller */ in efx_begin_loopback() 476 * skbs not already completed will be free'd when the queue is flushed */ in efx_end_loopback() 478 skb = state->skbs[i]; in efx_end_loopback() 490 /* Don't free the skbs; they will be picked up on TX in efx_end_loopback() 534 state->skbs = kcalloc(state->packet_count, in efx_test_loopback() 535 sizeof(state->skbs[ in efx_test_loopback() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/sfc/falcon/ |
H A D | selftest.c | 71 * @skbs: An array of skbs transmitted 80 struct sk_buff **skbs; member 426 state->skbs[i] = skb; in ef4_begin_loopback() 450 /* Defer cleaning up the other skbs for the caller */ in ef4_begin_loopback() 478 * skbs not already completed will be free'd when the queue is flushed */ in ef4_end_loopback() 480 skb = state->skbs[i]; in ef4_end_loopback() 492 /* Don't free the skbs; they will be picked up on TX in ef4_end_loopback() 536 state->skbs = kcalloc(state->packet_count, in ef4_test_loopback() 537 sizeof(state->skbs[ in ef4_test_loopback() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/sfc/falcon/ |
H A D | selftest.c | 76 * @skbs: An array of skbs transmitted 85 struct sk_buff **skbs; member 434 state->skbs[i] = skb; in ef4_begin_loopback() 462 /* Defer cleaning up the other skbs for the caller */ in ef4_begin_loopback() 490 * skbs not already completed will be free'd when the queue is flushed */ in ef4_end_loopback() 492 skb = state->skbs[i]; in ef4_end_loopback() 504 /* Don't free the skbs; they will be picked up on TX in ef4_end_loopback() 548 state->skbs = kcalloc(state->packet_count, in ef4_test_loopback() 549 sizeof(state->skbs[ in ef4_test_loopback() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/sfc/ |
H A D | selftest.c | 78 * @skbs: An array of skbs transmitted 87 struct sk_buff **skbs; member 431 state->skbs[i] = skb; in efx_begin_loopback() 459 /* Defer cleaning up the other skbs for the caller */ in efx_begin_loopback() 487 * skbs not already completed will be free'd when the queue is flushed */ in efx_end_loopback() 489 skb = state->skbs[i]; in efx_end_loopback() 501 /* Don't free the skbs; they will be picked up on TX in efx_end_loopback() 545 state->skbs = kcalloc(state->packet_count, in efx_test_loopback() 546 sizeof(state->skbs[ in efx_test_loopback() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/sfc/siena/ |
H A D | selftest.c | 79 * @skbs: An array of skbs transmitted 88 struct sk_buff **skbs; member 432 state->skbs[i] = skb; in efx_begin_loopback() 460 /* Defer cleaning up the other skbs for the caller */ in efx_begin_loopback() 488 * skbs not already completed will be free'd when the queue is flushed */ in efx_end_loopback() 490 skb = state->skbs[i]; in efx_end_loopback() 502 /* Don't free the skbs; they will be picked up on TX in efx_end_loopback() 546 state->skbs = kcalloc(state->packet_count, in efx_test_loopback() 547 sizeof(state->skbs[ in efx_test_loopback() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/renesas/ |
H A D | rswitch.c | 239 if (gq->skbs[index]) in rswitch_gwca_queue_alloc_skb() 241 gq->skbs[index] = netdev_alloc_skb_ip_align(gq->ndev, in rswitch_gwca_queue_alloc_skb() 243 if (!gq->skbs[index]) in rswitch_gwca_queue_alloc_skb() 252 dev_kfree_skb(gq->skbs[index]); in rswitch_gwca_queue_alloc_skb() 253 gq->skbs[index] = NULL; in rswitch_gwca_queue_alloc_skb() 271 dev_kfree_skb(gq->skbs[i]); in rswitch_gwca_queue_free() 279 kfree(gq->skbs); in rswitch_gwca_queue_free() 280 gq->skbs = NULL; in rswitch_gwca_queue_free() 304 gq->skbs = kcalloc(gq->ring_size, sizeof(*gq->skbs), GFP_KERNE in rswitch_gwca_queue_alloc() [all...] |
/kernel/linux/linux-5.10/net/mac80211/ |
H A D | tx.c | 900 __skb_queue_tail(&tx->skbs, tmp); in ieee80211_fragment() 941 /* no matter what happens, tx->skb moves to tx->skbs */ in ieee80211_tx_h_fragment() 942 __skb_queue_tail(&tx->skbs, skb); in ieee80211_tx_h_fragment() 979 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_fragment() 985 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_fragment() 1015 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_stats() 1068 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_calculate_duration() 1072 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_calculate_duration() 1073 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); in ieee80211_tx_h_calculate_duration() 1180 __skb_queue_head_init(&tx->skbs); in ieee80211_tx_prepare() 1632 ieee80211_tx_frags(struct ieee80211_local *local, struct ieee80211_vif *vif, struct sta_info *sta, struct sk_buff_head *skbs, bool txpending) ieee80211_tx_frags() argument 1708 __ieee80211_tx(struct ieee80211_local *local, struct sk_buff_head *skbs, int led_len, struct sta_info *sta, bool txpending) __ieee80211_tx() argument 4443 struct sk_buff_head skbs; ieee80211_tx_pending_skb() local [all...] |
H A D | wpa.c | 255 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_tkip_encrypt() 485 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_ccmp_encrypt() 720 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_gcmp_encrypt() 953 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_encrypt() 956 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_encrypt() 1001 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_256_encrypt() 1004 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_256_encrypt() 1146 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_gmac_encrypt() 1149 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_gmac_encrypt() 1250 skb_queue_walk(&tx->skbs, sk in ieee80211_crypto_hw_encrypt() [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt7601u/ |
H A D | dma.c | 272 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local 275 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet() 284 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet() 288 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet() 289 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
|
/kernel/linux/linux-6.6/drivers/net/wireless/mediatek/mt7601u/ |
H A D | dma.c | 283 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local 286 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet() 295 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet() 299 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet() 300 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/socionext/ |
H A D | sni_ave.c | 226 struct sk_buff *skbs; member 585 skb = priv->rx.desc[entry].skbs; in ave_rxdesc_prepare() 618 priv->rx.desc[entry].skbs = skb; in ave_rxdesc_prepare() 715 if (priv->tx.desc[done_idx].skbs) { in ave_tx_complete() 718 dev_consume_skb_any(priv->tx.desc[done_idx].skbs); in ave_tx_complete() 719 priv->tx.desc[done_idx].skbs = NULL; in ave_tx_complete() 776 skb = priv->rx.desc[proc_idx].skbs; in ave_rx_receive() 777 priv->rx.desc[proc_idx].skbs = NULL; in ave_rx_receive() 1369 if (!priv->tx.desc[entry].skbs) in ave_stop() 1373 dev_kfree_skb_any(priv->tx.desc[entry].skbs); in ave_stop() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/socionext/ |
H A D | sni_ave.c | 227 struct sk_buff *skbs; member 586 skb = priv->rx.desc[entry].skbs; in ave_rxdesc_prepare() 619 priv->rx.desc[entry].skbs = skb; in ave_rxdesc_prepare() 716 if (priv->tx.desc[done_idx].skbs) { in ave_tx_complete() 719 dev_consume_skb_any(priv->tx.desc[done_idx].skbs); in ave_tx_complete() 720 priv->tx.desc[done_idx].skbs = NULL; in ave_tx_complete() 777 skb = priv->rx.desc[proc_idx].skbs; in ave_rx_receive() 778 priv->rx.desc[proc_idx].skbs = NULL; in ave_rx_receive() 1372 if (!priv->tx.desc[entry].skbs) in ave_stop() 1376 dev_kfree_skb_any(priv->tx.desc[entry].skbs); in ave_stop() [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | skb_array.h | 10 * Limited-size FIFO of skbs. Can be used more or less whenever 191 struct sk_buff **skbs, int n) in skb_array_unconsume() 193 ptr_ring_unconsume(&a->ring, (void **)skbs, n, __skb_array_destroy_skb); in skb_array_unconsume() 190 skb_array_unconsume(struct skb_array *a, struct sk_buff **skbs, int n) skb_array_unconsume() argument
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | skb_array.h | 10 * Limited-size FIFO of skbs. Can be used more or less whenever 191 struct sk_buff **skbs, int n) in skb_array_unconsume() 193 ptr_ring_unconsume(&a->ring, (void **)skbs, n, __skb_array_destroy_skb); in skb_array_unconsume() 190 skb_array_unconsume(struct skb_array *a, struct sk_buff **skbs, int n) skb_array_unconsume() argument
|
/kernel/linux/linux-6.6/net/mac80211/ |
H A D | wpa.c | 255 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_tkip_encrypt() 496 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_ccmp_encrypt() 695 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_gcmp_encrypt() 830 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_encrypt() 833 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_encrypt() 878 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_256_encrypt() 881 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_256_encrypt() 1023 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_gmac_encrypt() 1026 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_gmac_encrypt()
|
H A D | tx.c | 921 __skb_queue_tail(&tx->skbs, tmp); in ieee80211_fragment() 962 /* no matter what happens, tx->skb moves to tx->skbs */ in ieee80211_tx_h_fragment() 963 __skb_queue_tail(&tx->skbs, skb); in ieee80211_tx_h_fragment() 1000 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_fragment() 1006 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_fragment() 1036 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_stats() 1087 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_calculate_duration() 1091 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_calculate_duration() 1092 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); in ieee80211_tx_h_calculate_duration() 1219 __skb_queue_head_init(&tx->skbs); in ieee80211_tx_prepare() 1671 ieee80211_tx_frags(struct ieee80211_local *local, struct ieee80211_vif *vif, struct sta_info *sta, struct sk_buff_head *skbs, bool txpending) ieee80211_tx_frags() argument 1747 __ieee80211_tx(struct ieee80211_local *local, struct sk_buff_head *skbs, struct sta_info *sta, bool txpending) __ieee80211_tx() argument 4585 unsigned int skbs = 0, len = 0; ieee80211_8023_xmit() local 4801 struct sk_buff_head skbs; ieee80211_tx_pending_skb() local [all...] |
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | cpumap.c | 315 void *skbs[CPUMAP_BATCH]; in cpu_map_kthread_run() local 353 m = kmem_cache_alloc_bulk(skbuff_head_cache, gfp, nframes, skbs); in cpu_map_kthread_run() 356 skbs[i] = NULL; /* effect: xdp_return_frame */ in cpu_map_kthread_run() 364 struct sk_buff *skb = skbs[i]; in cpu_map_kthread_run()
|
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | cpumap.c | 281 void *skbs[CPUMAP_BATCH]; in cpu_map_kthread_run() local 332 m = kmem_cache_alloc_bulk(skbuff_cache, gfp, nframes, skbs); in cpu_map_kthread_run() 335 skbs[i] = NULL; /* effect: xdp_return_frame */ in cpu_map_kthread_run() 343 struct sk_buff *skb = skbs[i]; in cpu_map_kthread_run() 445 * stop the kthread prematurely and all pending frames or skbs in __cpu_map_entry_alloc()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/netronome/nfp/nfd3/ |
H A D | xsk.c | 391 unsigned int pkts_polled, skbs = 0; in nfp_nfd3_xsk_poll() local 393 pkts_polled = nfp_nfd3_xsk_rx(r_vec->rx_ring, budget, &skbs); in nfp_nfd3_xsk_poll() 404 if (pkts_polled < budget && napi_complete_done(napi, skbs)) in nfp_nfd3_xsk_poll()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mediatek/ |
H A D | mtk_star_emac.c | 229 struct sk_buff *skbs[MTK_STAR_RING_NUM_DESCS]; member 302 desc_data->skb = ring->skbs[ring->tail]; in mtk_star_ring_pop_tail() 305 ring->skbs[ring->tail] = NULL; in mtk_star_ring_pop_tail() 326 ring->skbs[ring->head] = desc_data->skb; in mtk_star_ring_push_head() 677 ring->skbs[i] = skb; in mtk_star_prepare_rx_skbs() 697 desc_data.skb = ring->skbs[i]; in mtk_star_ring_free_skbs()
|
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | tx.c | 1127 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local 1148 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx() 1177 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs); in iwlagn_rx_reply_tx() 1182 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx() 1253 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx() 1254 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
|
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | tx.c | 1123 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local 1144 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx() 1173 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs, false); in iwlagn_rx_reply_tx() 1178 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx() 1249 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx() 1250 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 78 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 83 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 123 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt() 124 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_tx_pkt() 135 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_failed_tx_pkt() 171 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_failed_tx_pkt() 172 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_failed_tx_pkt() 1537 txq->sw_tx_ring.skbs[idx].skb = skb; in qede_start_xmit() 1659 txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; in qede_start_xmit()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 79 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 84 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 124 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt() 125 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_tx_pkt() 136 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_failed_tx_pkt() 172 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_failed_tx_pkt() 173 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_failed_tx_pkt() 1519 txq->sw_tx_ring.skbs[idx].skb = skb; in qede_start_xmit() 1641 txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; in qede_start_xmit()
|