Searched refs:nb_pkts (Results 1 - 5 of 5) sorted by relevance
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_xsk.c | 578 static void i40e_fill_tx_hw_ring(struct i40e_ring *xdp_ring, struct xdp_desc *descs, u32 nb_pkts, in i40e_fill_tx_hw_ring() argument 583 batched = nb_pkts & ~(PKTS_PER_BATCH - 1); in i40e_fill_tx_hw_ring() 584 leftover = nb_pkts & (PKTS_PER_BATCH - 1); in i40e_fill_tx_hw_ring() 610 u32 nb_pkts, nb_processed = 0; in i40e_xmit_zc() local 613 nb_pkts = xsk_tx_peek_release_desc_batch(xdp_ring->xsk_pool, budget); in i40e_xmit_zc() 614 if (!nb_pkts) in i40e_xmit_zc() 617 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in i40e_xmit_zc() 623 i40e_fill_tx_hw_ring(xdp_ring, &descs[nb_processed], nb_pkts - nb_processed, in i40e_xmit_zc() 630 i40e_update_tx_stats(xdp_ring, nb_pkts, total_bytes); in i40e_xmit_zc() 632 return nb_pkts < budge in i40e_xmit_zc() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/ |
H A D | xskxceiver.c | 504 if (pkt_stream->current_pkt_nb >= pkt_stream->nb_pkts) in pkt_stream_get_next_tx_pkt() 512 while (pkt_stream->current_pkt_nb < pkt_stream->nb_pkts) { in pkt_stream_get_next_rx_pkt() 543 static struct pkt_stream *__pkt_stream_alloc(u32 nb_pkts) in __pkt_stream_alloc() argument 551 pkt_stream->pkts = calloc(nb_pkts, sizeof(*pkt_stream->pkts)); in __pkt_stream_alloc() 557 pkt_stream->nb_pkts = nb_pkts; in __pkt_stream_alloc() 590 while (next_frag++ < pkt_stream->nb_pkts) { in pkt_nb_frags() 614 static struct pkt_stream *pkt_stream_generate(struct xsk_umem_info *umem, u32 nb_pkts, u32 pkt_len) in pkt_stream_generate() argument 619 pkt_stream = __pkt_stream_alloc(nb_pkts); in pkt_stream_generate() 623 pkt_stream->nb_pkts in pkt_stream_generate() 641 pkt_stream_replace(struct test_spec *test, u32 nb_pkts, u32 pkt_len) pkt_stream_replace() argument 760 pkt_stream_generate_custom(struct test_spec *test, struct pkt *pkts, u32 nb_pkts) pkt_stream_generate_custom() argument 1389 u32 idx = 0, filled = 0, buffers_to_fill, nb_pkts; xsk_populate_fill_ring() local [all...] |
H A D | xskxceiver.h | 134 u32 nb_pkts; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_xsk.c | 1041 * @nb_pkts: count of packets to be send 1045 u32 nb_pkts, unsigned int *total_bytes) in ice_fill_tx_hw_ring() 1049 batched = ALIGN_DOWN(nb_pkts, PKTS_PER_BATCH); in ice_fill_tx_hw_ring() 1050 leftover = nb_pkts & (PKTS_PER_BATCH - 1); in ice_fill_tx_hw_ring() 1066 u32 nb_pkts, nb_processed = 0; in ice_xmit_zc() local 1075 nb_pkts = xsk_tx_peek_release_desc_batch(xdp_ring->xsk_pool, budget); in ice_xmit_zc() 1076 if (!nb_pkts) in ice_xmit_zc() 1079 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in ice_xmit_zc() 1085 ice_fill_tx_hw_ring(xdp_ring, &descs[nb_processed], nb_pkts - nb_processed, in ice_xmit_zc() 1090 ice_update_tx_ring_stats(xdp_ring, nb_pkts, total_byte in ice_xmit_zc() 1044 ice_fill_tx_hw_ring(struct ice_tx_ring *xdp_ring, struct xdp_desc *descs, u32 nb_pkts, unsigned int *total_bytes) ice_fill_tx_hw_ring() argument [all...] |
/kernel/linux/linux-6.6/net/xdp/ |
H A D | xsk.c | 452 u32 nb_pkts = 0; in xsk_tx_peek_release_fallback() local 454 while (nb_pkts < max_entries && xsk_tx_peek_desc(pool, &descs[nb_pkts])) in xsk_tx_peek_release_fallback() 455 nb_pkts++; in xsk_tx_peek_release_fallback() 458 return nb_pkts; in xsk_tx_peek_release_fallback() 461 u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 nb_pkts) in xsk_tx_peek_release_desc_batch() argument 469 return xsk_tx_peek_release_fallback(pool, nb_pkts); in xsk_tx_peek_release_desc_batch() 474 nb_pkts = 0; in xsk_tx_peek_release_desc_batch() 478 nb_pkts = xskq_cons_nb_entries(xs->tx, nb_pkts); in xsk_tx_peek_release_desc_batch() [all...] |
Completed in 8 milliseconds