Home
last modified time | relevance | path

Searched refs:pool (Results 1 - 25 of 818) sorted by relevance

12345678910>>...33

/kernel/linux/linux-6.6/net/xdp/
H A Dxsk_buff_pool.c11 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_add_xsk() argument
18 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_add_xsk()
19 list_add_rcu(&xs->tx_list, &pool->xsk_tx_list); in xp_add_xsk()
20 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_add_xsk()
23 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_del_xsk() argument
30 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_del_xsk()
32 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_del_xsk()
35 void xp_destroy(struct xsk_buff_pool *pool) in xp_destroy() argument
37 if (!pool) in xp_destroy()
40 kvfree(pool in xp_destroy()
45 xp_alloc_tx_descs(struct xsk_buff_pool *pool, struct xdp_sock *xs) xp_alloc_tx_descs() argument
59 struct xsk_buff_pool *pool; xp_create_and_assign_umem() local
117 xp_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xp_set_rxq_info() argument
126 xp_disable_drv_zc(struct xsk_buff_pool *pool) xp_disable_drv_zc() argument
149 xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *netdev, u16 queue_id, u16 flags) xp_assign_dev() argument
228 xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_sock *umem_xs, struct net_device *dev, u16 queue_id) xp_assign_dev_shared() argument
245 xp_clear_dev(struct xsk_buff_pool *pool) xp_clear_dev() argument
258 struct xsk_buff_pool *pool = container_of(work, struct xsk_buff_pool, xp_release_deferred() local
279 xp_get_pool(struct xsk_buff_pool *pool) xp_get_pool() argument
284 xp_put_pool(struct xsk_buff_pool *pool) xp_put_pool() argument
298 xp_find_dma_map(struct xsk_buff_pool *pool) xp_find_dma_map() argument
359 xp_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xp_dma_unmap() argument
395 xp_init_dma_info(struct xsk_buff_pool *pool, struct xsk_dma_map *dma_map) xp_init_dma_info() argument
420 xp_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs, struct page **pages, u32 nr_pages) xp_dma_map() argument
467 xp_addr_crosses_non_contig_pg(struct xsk_buff_pool *pool, u64 addr) xp_addr_crosses_non_contig_pg() argument
473 xp_check_unaligned(struct xsk_buff_pool *pool, u64 *addr) xp_check_unaligned() argument
483 xp_check_aligned(struct xsk_buff_pool *pool, u64 *addr) xp_check_aligned() argument
489 __xp_alloc(struct xsk_buff_pool *pool) __xp_alloc() argument
527 xp_alloc(struct xsk_buff_pool *pool) xp_alloc() argument
555 xp_alloc_new_from_fq(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) xp_alloc_new_from_fq() argument
598 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) xp_alloc_reused() argument
618 xp_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) xp_alloc_batch() argument
649 xp_can_alloc(struct xsk_buff_pool *pool, u32 count) xp_can_alloc() argument
667 xp_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xp_raw_get_data() argument
674 xp_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xp_raw_get_dma() argument
690 xp_dma_sync_for_device_slow(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xp_dma_sync_for_device_slow() argument
[all...]
/kernel/linux/linux-5.10/net/xdp/
H A Dxsk_buff_pool.c11 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_add_xsk() argument
18 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_add_xsk()
19 list_add_rcu(&xs->tx_list, &pool->xsk_tx_list); in xp_add_xsk()
20 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_add_xsk()
23 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_del_xsk() argument
30 spin_lock_irqsave(&pool->xsk_tx_list_lock, flags); in xp_del_xsk()
32 spin_unlock_irqrestore(&pool->xsk_tx_list_lock, flags); in xp_del_xsk()
35 void xp_destroy(struct xsk_buff_pool *pool) in xp_destroy() argument
37 if (!pool) in xp_destroy()
40 kvfree(pool in xp_destroy()
47 struct xsk_buff_pool *pool; xp_create_and_assign_umem() local
94 xp_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xp_set_rxq_info() argument
103 xp_disable_drv_zc(struct xsk_buff_pool *pool) xp_disable_drv_zc() argument
122 __xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *netdev, u16 queue_id, u16 flags) __xp_assign_dev() argument
195 xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *dev, u16 queue_id, u16 flags) xp_assign_dev() argument
201 xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_sock *umem_xs, struct net_device *dev, u16 queue_id) xp_assign_dev_shared() argument
218 xp_clear_dev(struct xsk_buff_pool *pool) xp_clear_dev() argument
231 struct xsk_buff_pool *pool = container_of(work, struct xsk_buff_pool, xp_release_deferred() local
252 xp_get_pool(struct xsk_buff_pool *pool) xp_get_pool() argument
257 xp_put_pool(struct xsk_buff_pool *pool) xp_put_pool() argument
271 xp_find_dma_map(struct xsk_buff_pool *pool) xp_find_dma_map() argument
332 xp_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xp_dma_unmap() argument
367 xp_init_dma_info(struct xsk_buff_pool *pool, struct xsk_dma_map *dma_map) xp_init_dma_info() argument
382 xp_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs, struct page **pages, u32 nr_pages) xp_dma_map() argument
429 xp_addr_crosses_non_contig_pg(struct xsk_buff_pool *pool, u64 addr) xp_addr_crosses_non_contig_pg() argument
435 xp_check_unaligned(struct xsk_buff_pool *pool, u64 *addr) xp_check_unaligned() argument
445 xp_check_aligned(struct xsk_buff_pool *pool, u64 *addr) xp_check_aligned() argument
451 __xp_alloc(struct xsk_buff_pool *pool) __xp_alloc() argument
492 xp_alloc(struct xsk_buff_pool *pool) xp_alloc() argument
519 xp_can_alloc(struct xsk_buff_pool *pool, u32 count) xp_can_alloc() argument
534 xp_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xp_raw_get_data() argument
541 xp_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xp_raw_get_dma() argument
557 xp_dma_sync_for_device_slow(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xp_dma_sync_for_device_slow() argument
[all...]
/kernel/linux/linux-6.6/net/core/
H A Dpage_pool.c33 #define alloc_stat_inc(pool, __stat) (pool->alloc_stats.__stat++)
35 #define recycle_stat_inc(pool, __stat) \
37 struct page_pool_recycle_stats __percpu *s = pool->recycle_stats; \
41 #define recycle_stat_add(pool, __stat, val) \
43 struct page_pool_recycle_stats __percpu *s = pool->recycle_stats; \
62 * page_pool_get_stats() - fetch page pool stats
63 * @pool: pool from which page was allocated
72 bool page_pool_get_stats(struct page_pool *pool, in page_pool_get_stats() argument
171 page_pool_init(struct page_pool *pool, const struct page_pool_params *params) page_pool_init() argument
248 struct page_pool *pool; page_pool_create() local
269 page_pool_refill_alloc_cache(struct page_pool *pool) page_pool_refill_alloc_cache() argument
322 __page_pool_get_cached(struct page_pool *pool) __page_pool_get_cached() argument
338 page_pool_dma_sync_for_device(struct page_pool *pool, struct page *page, unsigned int dma_sync_size) page_pool_dma_sync_for_device() argument
350 page_pool_dma_map(struct page_pool *pool, struct page *page) page_pool_dma_map() argument
374 page_pool_set_pp_info(struct page_pool *pool, struct page *page) page_pool_set_pp_info() argument
389 __page_pool_alloc_page_order(struct page_pool *pool, gfp_t gfp) __page_pool_alloc_page_order() argument
416 __page_pool_alloc_pages_slow(struct page_pool *pool, gfp_t gfp) __page_pool_alloc_pages_slow() argument
475 page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp) page_pool_alloc_pages() argument
495 page_pool_inflight(struct page_pool *pool) page_pool_inflight() argument
514 page_pool_return_page(struct page_pool *pool, struct page *page) page_pool_return_page() argument
548 page_pool_recycle_in_ring(struct page_pool *pool, struct page *page) page_pool_recycle_in_ring() argument
570 page_pool_recycle_in_cache(struct page *page, struct page_pool *pool) page_pool_recycle_in_cache() argument
591 __page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) __page_pool_put_page() argument
638 page_pool_put_defragged_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) page_pool_put_defragged_page() argument
665 page_pool_put_page_bulk(struct page_pool *pool, void **data, int count) page_pool_put_page_bulk() argument
711 page_pool_drain_frag(struct page_pool *pool, struct page *page) page_pool_drain_frag() argument
731 page_pool_free_frag(struct page_pool *pool) page_pool_free_frag() argument
744 page_pool_alloc_frag(struct page_pool *pool, unsigned int *offset, unsigned int size, gfp_t gfp) page_pool_alloc_frag() argument
790 page_pool_empty_ring(struct page_pool *pool) page_pool_empty_ring() argument
805 page_pool_free(struct page_pool *pool) page_pool_free() argument
821 page_pool_empty_alloc_cache_once(struct page_pool *pool) page_pool_empty_alloc_cache_once() argument
838 page_pool_scrub(struct page_pool *pool) page_pool_scrub() argument
849 page_pool_release(struct page_pool *pool) page_pool_release() argument
864 struct page_pool *pool = container_of(dwq, typeof(*pool), release_dw); page_pool_release_retry() local
884 page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *), struct xdp_mem_info *mem) page_pool_use_xdp_mem() argument
892 page_pool_unlink_napi(struct page_pool *pool) page_pool_unlink_napi() argument
907 page_pool_destroy(struct page_pool *pool) page_pool_destroy() argument
930 page_pool_update_nid(struct page_pool *pool, int new_nid) page_pool_update_nid() argument
[all...]
/kernel/linux/linux-5.10/drivers/staging/android/ion/
H A Dion_page_pool.c3 * ION Memory Allocator page pool helpers
15 static inline struct page *ion_page_pool_alloc_pages(struct ion_page_pool *pool) in ion_page_pool_alloc_pages() argument
19 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages()
22 static void ion_page_pool_free_pages(struct ion_page_pool *pool, in ion_page_pool_free_pages() argument
25 __free_pages(page, pool->order); in ion_page_pool_free_pages()
28 static void ion_page_pool_add(struct ion_page_pool *pool, struct page *page) in ion_page_pool_add() argument
30 mutex_lock(&pool->mutex); in ion_page_pool_add()
32 list_add_tail(&page->lru, &pool->high_items); in ion_page_pool_add()
33 pool in ion_page_pool_add()
44 ion_page_pool_remove(struct ion_page_pool *pool, bool high) ion_page_pool_remove() argument
64 ion_page_pool_alloc(struct ion_page_pool *pool) ion_page_pool_alloc() argument
83 ion_page_pool_free(struct ion_page_pool *pool, struct page *page) ion_page_pool_free() argument
90 ion_page_pool_total(struct ion_page_pool *pool, bool high) ion_page_pool_total() argument
100 ion_page_pool_shrink(struct ion_page_pool *pool, gfp_t gfp_mask, int nr_to_scan) ion_page_pool_shrink() argument
136 struct ion_page_pool *pool = kmalloc(sizeof(*pool), GFP_KERNEL); ion_page_pool_create() local
152 ion_page_pool_destroy(struct ion_page_pool *pool) ion_page_pool_destroy() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/ti/
H A Dk3-cppi-desc-pool.c2 /* TI K3 CPPI5 descriptors pool API
15 #include "k3-cppi-desc-pool.h"
27 void k3_cppi_desc_pool_destroy(struct k3_cppi_desc_pool *pool) in k3_cppi_desc_pool_destroy() argument
29 if (!pool) in k3_cppi_desc_pool_destroy()
32 WARN(gen_pool_size(pool->gen_pool) != gen_pool_avail(pool->gen_pool), in k3_cppi_desc_pool_destroy()
34 gen_pool_size(pool->gen_pool), in k3_cppi_desc_pool_destroy()
35 gen_pool_avail(pool->gen_pool)); in k3_cppi_desc_pool_destroy()
36 if (pool->cpumem) in k3_cppi_desc_pool_destroy()
37 dma_free_coherent(pool in k3_cppi_desc_pool_destroy()
48 struct k3_cppi_desc_pool *pool; k3_cppi_desc_pool_create_name() local
102 k3_cppi_desc_pool_virt2dma(struct k3_cppi_desc_pool *pool, void *addr) k3_cppi_desc_pool_virt2dma() argument
108 k3_cppi_desc_pool_dma2virt(struct k3_cppi_desc_pool *pool, dma_addr_t dma) k3_cppi_desc_pool_dma2virt() argument
113 k3_cppi_desc_pool_alloc(struct k3_cppi_desc_pool *pool) k3_cppi_desc_pool_alloc() argument
118 k3_cppi_desc_pool_free(struct k3_cppi_desc_pool *pool, void *addr) k3_cppi_desc_pool_free() argument
123 k3_cppi_desc_pool_avail(struct k3_cppi_desc_pool *pool) k3_cppi_desc_pool_avail() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/ti/
H A Dk3-cppi-desc-pool.c2 /* TI K3 CPPI5 descriptors pool API
15 #include "k3-cppi-desc-pool.h"
27 void k3_cppi_desc_pool_destroy(struct k3_cppi_desc_pool *pool) in k3_cppi_desc_pool_destroy() argument
29 if (!pool) in k3_cppi_desc_pool_destroy()
32 WARN(gen_pool_size(pool->gen_pool) != gen_pool_avail(pool->gen_pool), in k3_cppi_desc_pool_destroy()
34 gen_pool_size(pool->gen_pool), in k3_cppi_desc_pool_destroy()
35 gen_pool_avail(pool->gen_pool)); in k3_cppi_desc_pool_destroy()
36 if (pool->cpumem) in k3_cppi_desc_pool_destroy()
37 dma_free_coherent(pool in k3_cppi_desc_pool_destroy()
49 struct k3_cppi_desc_pool *pool; k3_cppi_desc_pool_create_name() local
104 k3_cppi_desc_pool_virt2dma(struct k3_cppi_desc_pool *pool, void *addr) k3_cppi_desc_pool_virt2dma() argument
111 k3_cppi_desc_pool_dma2virt(struct k3_cppi_desc_pool *pool, dma_addr_t dma) k3_cppi_desc_pool_dma2virt() argument
117 k3_cppi_desc_pool_alloc(struct k3_cppi_desc_pool *pool) k3_cppi_desc_pool_alloc() argument
123 k3_cppi_desc_pool_free(struct k3_cppi_desc_pool *pool, void *addr) k3_cppi_desc_pool_free() argument
129 k3_cppi_desc_pool_avail(struct k3_cppi_desc_pool *pool) k3_cppi_desc_pool_avail() argument
[all...]
/kernel/linux/linux-5.10/net/core/
H A Dpage_pool.c24 static int page_pool_init(struct page_pool *pool, in page_pool_init() argument
29 memcpy(&pool->p, params, sizeof(pool->p)); in page_pool_init()
32 if (pool->p.flags & ~(PP_FLAG_ALL)) in page_pool_init()
35 if (pool->p.pool_size) in page_pool_init()
36 ring_qsize = pool->p.pool_size; in page_pool_init()
46 if (pool->p.flags & PP_FLAG_DMA_MAP) { in page_pool_init()
47 if ((pool->p.dma_dir != DMA_FROM_DEVICE) && in page_pool_init()
48 (pool->p.dma_dir != DMA_BIDIRECTIONAL)) in page_pool_init()
52 if (pool in page_pool_init()
83 struct page_pool *pool; page_pool_create() local
104 page_pool_refill_alloc_cache(struct page_pool *pool) page_pool_refill_alloc_cache() argument
156 __page_pool_get_cached(struct page_pool *pool) __page_pool_get_cached() argument
171 page_pool_dma_sync_for_device(struct page_pool *pool, struct page *page, unsigned int dma_sync_size) page_pool_dma_sync_for_device() argument
185 __page_pool_alloc_pages_slow(struct page_pool *pool, gfp_t _gfp) __page_pool_alloc_pages_slow() argument
247 page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp) page_pool_alloc_pages() argument
267 page_pool_inflight(struct page_pool *pool) page_pool_inflight() argument
286 page_pool_release_page(struct page_pool *pool, struct page *page) page_pool_release_page() argument
314 page_pool_return_page(struct page_pool *pool, struct page *page) page_pool_return_page() argument
325 page_pool_recycle_in_ring(struct page_pool *pool, struct page *page) page_pool_recycle_in_ring() argument
342 page_pool_recycle_in_cache(struct page *page, struct page_pool *pool) page_pool_recycle_in_cache() argument
356 pool_page_reusable(struct page_pool *pool, struct page *page) pool_page_reusable() argument
367 page_pool_put_page(struct page_pool *pool, struct page *page, unsigned int dma_sync_size, bool allow_direct) page_pool_put_page() argument
413 page_pool_empty_ring(struct page_pool *pool) page_pool_empty_ring() argument
428 page_pool_free(struct page_pool *pool) page_pool_free() argument
441 page_pool_empty_alloc_cache_once(struct page_pool *pool) page_pool_empty_alloc_cache_once() argument
458 page_pool_scrub(struct page_pool *pool) page_pool_scrub() argument
469 page_pool_release(struct page_pool *pool) page_pool_release() argument
484 struct page_pool *pool = container_of(dwq, typeof(*pool), release_dw); page_pool_release_retry() local
504 page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *)) page_pool_use_xdp_mem() argument
510 page_pool_destroy(struct page_pool *pool) page_pool_destroy() argument
530 page_pool_update_nid(struct page_pool *pool, int new_nid) page_pool_update_nid() argument
[all...]
/kernel/linux/linux-5.10/drivers/md/
H A Ddm-thin.c41 * The block size of the device holding pool data must be
191 * A pool device ties together a metadata device and a data device. It
198 * The pool runs in various modes. Ordered in degraded order for comparisons.
229 struct pool { struct
231 struct dm_target *ti; /* Only set if a pool target is bound */
289 static void metadata_operation_failed(struct pool *pool, const char *op, int r);
291 static enum pool_mode get_pool_mode(struct pool *pool) in get_pool_mode() argument
293 return pool in get_pool_mode()
296 notify_of_pool_mode_change(struct pool *pool) notify_of_pool_mode_change() argument
326 struct pool *pool; global() member
345 struct pool *pool; global() member
366 block_size_is_power_of_two(struct pool *pool) block_size_is_power_of_two() argument
371 block_to_sectors(struct pool *pool, dm_block_t b) block_to_sectors() argument
436 wake_worker(struct pool *pool) wake_worker() argument
443 bio_detain(struct pool *pool, struct dm_cell_key *key, struct bio *bio, struct dm_bio_prison_cell **cell_result) bio_detain() argument
466 cell_release(struct pool *pool, struct dm_bio_prison_cell *cell, struct bio_list *bios) cell_release() argument
474 cell_visit_release(struct pool *pool, void (*fn)(void *, struct dm_bio_prison_cell *), void *context, struct dm_bio_prison_cell *cell) cell_visit_release() argument
483 cell_release_no_holder(struct pool *pool, struct dm_bio_prison_cell *cell, struct bio_list *bios) cell_release_no_holder() argument
491 cell_error_with_code(struct pool *pool, struct dm_bio_prison_cell *cell, blk_status_t error_code) cell_error_with_code() argument
498 get_pool_io_error_code(struct pool *pool) get_pool_io_error_code() argument
503 cell_error(struct pool *pool, struct dm_bio_prison_cell *cell) cell_error() argument
508 cell_success(struct pool *pool, struct dm_bio_prison_cell *cell) cell_success() argument
513 cell_requeue(struct pool *pool, struct dm_bio_prison_cell *cell) cell_requeue() argument
539 __pool_table_insert(struct pool *pool) __pool_table_insert() argument
545 __pool_table_remove(struct pool *pool) __pool_table_remove() argument
553 struct pool *pool = NULL, *tmp; __pool_table_lookup() local
569 struct pool *pool = NULL, *tmp; __pool_table_lookup_metadata_dev() local
626 struct pool *pool = tc->pool; requeue_deferred_cells() local
655 error_retry_list_with_code(struct pool *pool, blk_status_t error) error_retry_list_with_code() argument
665 error_retry_list(struct pool *pool) error_retry_list() argument
679 struct pool *pool = tc->pool; get_bio_block() local
696 struct pool *pool = tc->pool; get_bio_block_range() local
720 struct pool *pool = tc->pool; remap() local
744 inc_all_io_entry(struct pool *pool, struct bio *bio) inc_all_io_entry() argument
757 struct pool *pool = tc->pool; issue() local
832 struct pool *pool = m->tc->pool; __complete_mapping_preparation() local
843 struct pool *pool = m->tc->pool; complete_mapping_preparation() local
885 struct pool *pool = tc->pool; cell_defer_no_holder() local
963 struct pool *pool = tc->pool; complete_overwrite_bio() local
996 struct pool *pool = tc->pool; process_prepared_mapping() local
1087 struct pool *pool = tc->pool; passdown_double_checking_shared_status() local
1129 struct pool *pool = m->tc->pool; queue_passdown_pt2() local
1151 struct pool *pool = tc->pool; process_prepared_discard_passdown_pt1() local
1208 struct pool *pool = tc->pool; process_prepared_discard_passdown_pt2() local
1226 process_prepared(struct pool *pool, struct list_head *head, process_mapping_fn *fn) process_prepared() argument
1244 io_overlaps_block(struct pool *pool, struct bio *bio) io_overlaps_block() argument
1250 io_overwrites_block(struct pool *pool, struct bio *bio) io_overwrites_block() argument
1263 ensure_next_mapping(struct pool *pool) ensure_next_mapping() argument
1273 get_next_mapping(struct pool *pool) get_next_mapping() argument
1304 struct pool *pool = tc->pool; remap_and_issue_overwrite() local
1323 struct pool *pool = tc->pool; schedule_copy() local
1391 struct pool *pool = tc->pool; schedule_zero() local
1420 struct pool *pool = tc->pool; schedule_external_copy() local
1447 is_read_only(struct pool *pool) is_read_only() argument
1452 check_for_metadata_space(struct pool *pool) check_for_metadata_space() argument
1470 check_for_data_space(struct pool *pool) check_for_data_space() argument
1492 commit(struct pool *pool) commit() argument
1510 check_low_water_mark(struct pool *pool, dm_block_t free_blocks) check_low_water_mark() argument
1526 struct pool *pool = tc->pool; alloc_data_block() local
1599 should_error_unserviceable_bio(struct pool *pool) should_error_unserviceable_bio() argument
1623 handle_unserviceable_bio(struct pool *pool, struct bio *bio) handle_unserviceable_bio() argument
1634 retry_bios_on_resume(struct pool *pool, struct dm_bio_prison_cell *cell) retry_bios_on_resume() argument
1656 struct pool *pool = tc->pool; process_discard_cell_no_passdown() local
1676 struct pool *pool = tc->pool; break_up_discard_bio() local
1793 struct pool *pool = tc->pool; break_sharing() local
1861 struct pool *pool = tc->pool; process_shared_bio() local
1894 struct pool *pool = tc->pool; provision_block() local
1941 struct pool *pool = tc->pool; process_cell() local
1995 struct pool *pool = tc->pool; process_bio() local
2096 need_commit_due_to_time(struct pool *pool) need_commit_due_to_time() argument
2168 struct pool *pool = tc->pool; process_thin_deferred_bios() local
2242 sort_cells(struct pool *pool, struct list_head *cells) sort_cells() argument
2262 struct pool *pool = tc->pool; process_thin_deferred_cells() local
2315 get_first_thin(struct pool *pool) get_first_thin() argument
2329 get_next_thin(struct pool *pool, struct thin_c *tc) get_next_thin() argument
2346 process_deferred_bios(struct pool *pool) process_deferred_bios() argument
2404 struct pool *pool = container_of(ws, struct pool, worker); do_worker() local
2425 struct pool *pool = container_of(to_delayed_work(ws), struct pool, waker); do_waker() local
2437 struct pool *pool = container_of(to_delayed_work(ws), struct pool, do_no_space_timeout() local
2464 pool_work_wait(struct pool_work *pw, struct pool *pool, void (*fn)(struct work_struct *)) pool_work_wait() argument
2515 set_discard_callbacks(struct pool *pool) set_discard_callbacks() argument
2529 set_pool_mode(struct pool *pool, enum pool_mode new_mode) set_pool_mode() argument
2627 abort_transaction(struct pool *pool) abort_transaction() argument
2643 metadata_operation_failed(struct pool *pool, const char *op, int r) metadata_operation_failed() argument
2663 struct pool *pool = tc->pool; thin_defer_bio() local
2674 struct pool *pool = tc->pool; thin_defer_bio_with_throttle() local
2683 struct pool *pool = tc->pool; thin_defer_cell() local
2800 requeue_bios(struct pool *pool) requeue_bios() argument
2835 struct pool *pool = pt->pool; disable_passdown_if_not_supported() local
2856 bind_control_target(struct pool *pool, struct dm_target *ti) bind_control_target() argument
2882 unbind_control_target(struct pool *pool, struct dm_target *ti) unbind_control_target() argument
2901 __pool_destroy(struct pool *pool) __pool_destroy() argument
2936 struct pool *pool; pool_create() local
3065 __pool_inc(struct pool *pool) __pool_inc() argument
3071 __pool_dec(struct pool *pool) __pool_dec() argument
3085 struct pool *pool = __pool_table_lookup_metadata_dev(metadata_dev); __pool_find() local
3186 struct pool *pool = context; metadata_low_callback() local
3207 struct pool *pool = context; metadata_pre_commit_callback() local
3285 struct pool *pool; pool_ctr() local
3439 struct pool *pool = pt->pool; pool_map() local
3456 struct pool *pool = pt->pool; maybe_resize_data_dev() local
3504 struct pool *pool = pt->pool; maybe_resize_metadata_dev() local
3567 struct pool *pool = pt->pool; pool_preresume() local
3598 pool_suspend_active_thins(struct pool *pool) pool_suspend_active_thins() argument
3610 pool_resume_active_thins(struct pool *pool) pool_resume_active_thins() argument
3625 struct pool *pool = pt->pool; pool_resume() local
3645 struct pool *pool = pt->pool; pool_presuspend() local
3657 struct pool *pool = pt->pool; pool_presuspend_undo() local
3669 struct pool *pool = pt->pool; pool_postsuspend() local
3700 process_create_thin_mesg(unsigned argc, char **argv, struct pool *pool) process_create_thin_mesg() argument
3723 process_create_snap_mesg(unsigned argc, char **argv, struct pool *pool) process_create_snap_mesg() argument
3751 process_delete_mesg(unsigned argc, char **argv, struct pool *pool) process_delete_mesg() argument
3771 process_set_transaction_id_mesg(unsigned argc, char **argv, struct pool *pool) process_set_transaction_id_mesg() argument
3800 process_reserve_metadata_snap_mesg(unsigned argc, char **argv, struct pool *pool) process_reserve_metadata_snap_mesg() argument
3817 process_release_metadata_snap_mesg(unsigned argc, char **argv, struct pool *pool) process_release_metadata_snap_mesg() argument
3846 struct pool *pool = pt->pool; pool_message() local
3926 struct pool *pool = pt->pool; pool_status() local
4048 struct pool *pool = pt->pool; pool_io_hints() local
4332 struct pool *pool = h->tc->pool; thin_endio() local
4456 struct pool *pool = tc->pool; thin_iterate_devices() local
4476 struct pool *pool = tc->pool; thin_io_hints() local
[all...]
/kernel/linux/linux-6.6/drivers/md/
H A Ddm-thin.c42 * The block size of the device holding pool data must be
194 * A pool device ties together a metadata device and a data device. It
201 * The pool runs in various modes. Ordered in degraded order for comparisons.
232 struct pool { struct
234 struct dm_target *ti; /* Only set if a pool target is bound */
290 static void metadata_operation_failed(struct pool *pool, const char *op, int r);
292 static enum pool_mode get_pool_mode(struct pool *pool) in get_pool_mode() argument
294 return pool in get_pool_mode()
297 notify_of_pool_mode_change(struct pool *pool) notify_of_pool_mode_change() argument
327 struct pool *pool; global() member
346 struct pool *pool; global() member
367 block_size_is_power_of_two(struct pool *pool) block_size_is_power_of_two() argument
372 block_to_sectors(struct pool *pool, dm_block_t b) block_to_sectors() argument
436 wake_worker(struct pool *pool) wake_worker() argument
443 bio_detain(struct pool *pool, struct dm_cell_key *key, struct bio *bio, struct dm_bio_prison_cell **cell_result) bio_detain() argument
466 cell_release(struct pool *pool, struct dm_bio_prison_cell *cell, struct bio_list *bios) cell_release() argument
474 cell_visit_release(struct pool *pool, void (*fn)(void *, struct dm_bio_prison_cell *), void *context, struct dm_bio_prison_cell *cell) cell_visit_release() argument
483 cell_release_no_holder(struct pool *pool, struct dm_bio_prison_cell *cell, struct bio_list *bios) cell_release_no_holder() argument
491 cell_error_with_code(struct pool *pool, struct dm_bio_prison_cell *cell, blk_status_t error_code) cell_error_with_code() argument
498 get_pool_io_error_code(struct pool *pool) get_pool_io_error_code() argument
503 cell_error(struct pool *pool, struct dm_bio_prison_cell *cell) cell_error() argument
508 cell_success(struct pool *pool, struct dm_bio_prison_cell *cell) cell_success() argument
513 cell_requeue(struct pool *pool, struct dm_bio_prison_cell *cell) cell_requeue() argument
539 __pool_table_insert(struct pool *pool) __pool_table_insert() argument
545 __pool_table_remove(struct pool *pool) __pool_table_remove() argument
553 struct pool *pool = NULL, *tmp; __pool_table_lookup() local
569 struct pool *pool = NULL, *tmp; __pool_table_lookup_metadata_dev() local
626 struct pool *pool = tc->pool; requeue_deferred_cells() local
655 error_retry_list_with_code(struct pool *pool, blk_status_t error) error_retry_list_with_code() argument
665 error_retry_list(struct pool *pool) error_retry_list() argument
679 struct pool *pool = tc->pool; get_bio_block() local
696 struct pool *pool = tc->pool; get_bio_block_range() local
720 struct pool *pool = tc->pool; remap() local
744 inc_all_io_entry(struct pool *pool, struct bio *bio) inc_all_io_entry() argument
757 struct pool *pool = tc->pool; issue() local
832 struct pool *pool = m->tc->pool; __complete_mapping_preparation() local
843 struct pool *pool = m->tc->pool; complete_mapping_preparation() local
885 struct pool *pool = tc->pool; cell_defer_no_holder() local
965 struct pool *pool = tc->pool; complete_overwrite_bio() local
998 struct pool *pool = tc->pool; process_prepared_mapping() local
1090 struct pool *pool = tc->pool; passdown_double_checking_shared_status() local
1132 struct pool *pool = m->tc->pool; queue_passdown_pt2() local
1154 struct pool *pool = tc->pool; process_prepared_discard_passdown_pt1() local
1203 struct pool *pool = tc->pool; process_prepared_discard_passdown_pt2() local
1221 process_prepared(struct pool *pool, struct list_head *head, process_mapping_fn *fn) process_prepared() argument
1239 io_overlaps_block(struct pool *pool, struct bio *bio) io_overlaps_block() argument
1245 io_overwrites_block(struct pool *pool, struct bio *bio) io_overwrites_block() argument
1258 ensure_next_mapping(struct pool *pool) ensure_next_mapping() argument
1268 get_next_mapping(struct pool *pool) get_next_mapping() argument
1299 struct pool *pool = tc->pool; remap_and_issue_overwrite() local
1318 struct pool *pool = tc->pool; schedule_copy() local
1386 struct pool *pool = tc->pool; schedule_zero() local
1415 struct pool *pool = tc->pool; schedule_external_copy() local
1442 is_read_only(struct pool *pool) is_read_only() argument
1447 check_for_metadata_space(struct pool *pool) check_for_metadata_space() argument
1465 check_for_data_space(struct pool *pool) check_for_data_space() argument
1487 commit(struct pool *pool) commit() argument
1505 check_low_water_mark(struct pool *pool, dm_block_t free_blocks) check_low_water_mark() argument
1521 struct pool *pool = tc->pool; alloc_data_block() local
1594 should_error_unserviceable_bio(struct pool *pool) should_error_unserviceable_bio() argument
1618 handle_unserviceable_bio(struct pool *pool, struct bio *bio) handle_unserviceable_bio() argument
1629 retry_bios_on_resume(struct pool *pool, struct dm_bio_prison_cell *cell) retry_bios_on_resume() argument
1651 struct pool *pool = tc->pool; process_discard_cell_no_passdown() local
1671 struct pool *pool = tc->pool; break_up_discard_bio() local
1810 struct pool *pool = tc->pool; break_sharing() local
1878 struct pool *pool = tc->pool; process_shared_bio() local
1911 struct pool *pool = tc->pool; provision_block() local
1958 struct pool *pool = tc->pool; process_cell() local
2012 struct pool *pool = tc->pool; process_bio() local
2113 need_commit_due_to_time(struct pool *pool) need_commit_due_to_time() argument
2185 struct pool *pool = tc->pool; process_thin_deferred_bios() local
2259 sort_cells(struct pool *pool, struct list_head *cells) sort_cells() argument
2279 struct pool *pool = tc->pool; process_thin_deferred_cells() local
2332 get_first_thin(struct pool *pool) get_first_thin() argument
2346 get_next_thin(struct pool *pool, struct thin_c *tc) get_next_thin() argument
2363 process_deferred_bios(struct pool *pool) process_deferred_bios() argument
2421 struct pool *pool = container_of(ws, struct pool, worker); do_worker() local
2442 struct pool *pool = container_of(to_delayed_work(ws), struct pool, waker); do_waker() local
2455 struct pool *pool = container_of(to_delayed_work(ws), struct pool, do_no_space_timeout() local
2482 pool_work_wait(struct pool_work *pw, struct pool *pool, void (*fn)(struct work_struct *)) pool_work_wait() argument
2530 set_discard_callbacks(struct pool *pool) set_discard_callbacks() argument
2544 set_pool_mode(struct pool *pool, enum pool_mode new_mode) set_pool_mode() argument
2642 abort_transaction(struct pool *pool) abort_transaction() argument
2658 metadata_operation_failed(struct pool *pool, const char *op, int r) metadata_operation_failed() argument
2678 struct pool *pool = tc->pool; thin_defer_bio() local
2689 struct pool *pool = tc->pool; thin_defer_bio_with_throttle() local
2698 struct pool *pool = tc->pool; thin_defer_cell() local
2815 requeue_bios(struct pool *pool) requeue_bios() argument
2845 struct pool *pool = pt->pool; disable_discard_passdown_if_not_supported() local
2865 bind_control_target(struct pool *pool, struct dm_target *ti) bind_control_target() argument
2891 unbind_control_target(struct pool *pool, struct dm_target *ti) unbind_control_target() argument
2912 __pool_destroy(struct pool *pool) __pool_destroy() argument
2946 struct pool *pool; pool_create() local
3074 __pool_inc(struct pool *pool) __pool_inc() argument
3080 __pool_dec(struct pool *pool) __pool_dec() argument
3094 struct pool *pool = __pool_table_lookup_metadata_dev(metadata_dev); __pool_find() local
3197 struct pool *pool = context; metadata_low_callback() local
3218 struct pool *pool = context; metadata_pre_commit_callback() local
3291 struct pool *pool; pool_ctr() local
3445 struct pool *pool = pt->pool; pool_map() local
3461 struct pool *pool = pt->pool; maybe_resize_data_dev() local
3509 struct pool *pool = pt->pool; maybe_resize_metadata_dev() local
3572 struct pool *pool = pt->pool; pool_preresume() local
3603 pool_suspend_active_thins(struct pool *pool) pool_suspend_active_thins() argument
3615 pool_resume_active_thins(struct pool *pool) pool_resume_active_thins() argument
3630 struct pool *pool = pt->pool; pool_resume() local
3650 struct pool *pool = pt->pool; pool_presuspend() local
3662 struct pool *pool = pt->pool; pool_presuspend_undo() local
3674 struct pool *pool = pt->pool; pool_postsuspend() local
3705 process_create_thin_mesg(unsigned int argc, char **argv, struct pool *pool) process_create_thin_mesg() argument
3728 process_create_snap_mesg(unsigned int argc, char **argv, struct pool *pool) process_create_snap_mesg() argument
3756 process_delete_mesg(unsigned int argc, char **argv, struct pool *pool) process_delete_mesg() argument
3776 process_set_transaction_id_mesg(unsigned int argc, char **argv, struct pool *pool) process_set_transaction_id_mesg() argument
3805 process_reserve_metadata_snap_mesg(unsigned int argc, char **argv, struct pool *pool) process_reserve_metadata_snap_mesg() argument
3822 process_release_metadata_snap_mesg(unsigned int argc, char **argv, struct pool *pool) process_release_metadata_snap_mesg() argument
3851 struct pool *pool = pt->pool; pool_message() local
3931 struct pool *pool = pt->pool; pool_status() local
4057 struct pool *pool = pt->pool; pool_io_hints() local
4344 struct pool *pool = h->tc->pool; thin_endio() local
4472 struct pool *pool = tc->pool; thin_iterate_devices() local
4492 struct pool *pool = tc->pool; thin_io_hints() local
[all...]
/kernel/linux/linux-6.6/mm/
H A Dmempool.c5 * memory buffer pool support. Such pools are mostly used
24 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument
27 const int nr = pool->curr_nr; in poison_error()
33 pr_err("Mempool %p size %zu\n", pool, size); in poison_error()
41 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument
50 poison_error(pool, element, size, i); in __check_element()
57 static void check_element(mempool_t *pool, void *element) in check_element() argument
60 if (pool->free == mempool_kfree) { in check_element()
61 __check_element(pool, element, (size_t)pool in check_element()
82 poison_element(mempool_t *pool, void *element) poison_element() argument
99 check_element(mempool_t *pool, void *element) check_element() argument
102 poison_element(mempool_t *pool, void *element) poison_element() argument
107 kasan_poison_element(mempool_t *pool, void *element) kasan_poison_element() argument
116 kasan_unpoison_element(mempool_t *pool, void *element) kasan_unpoison_element() argument
127 add_element(mempool_t *pool, void *element) add_element() argument
135 remove_element(mempool_t *pool) remove_element() argument
156 mempool_exit(mempool_t *pool) mempool_exit() argument
175 mempool_destroy(mempool_t *pool) mempool_destroy() argument
185 mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, mempool_free_t *free_fn, void *pool_data, gfp_t gfp_mask, int node_id) mempool_init_node() argument
233 mempool_init(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, mempool_free_t *free_fn, void *pool_data) mempool_init() argument
270 mempool_t *pool; mempool_create_node() local
304 mempool_resize(mempool_t *pool, int new_min_nr) mempool_resize() argument
380 mempool_alloc(mempool_t *pool, gfp_t gfp_mask) mempool_alloc() argument
457 mempool_free(void *element, mempool_t *pool) mempool_free() argument
[all...]
H A Ddmapool.c14 * The current design of this allocator is fairly simple. The pool is
48 struct dma_pool { /* the pool */
74 struct dma_pool *pool; in pools_show() local
80 list_for_each_entry(pool, &dev->dma_pools, pools) { in pools_show()
81 /* per-pool info, no real statistics yet */ in pools_show()
83 pool->name, pool->nr_active, in pools_show()
84 pool->nr_blocks, pool->size, in pools_show()
85 pool in pools_show()
95 pool_check_block(struct dma_pool *pool, struct dma_block *block, gfp_t mem_flags) pool_check_block() argument
120 pool_find_page(struct dma_pool *pool, dma_addr_t dma) pool_find_page() argument
133 pool_block_err(struct dma_pool *pool, void *vaddr, dma_addr_t dma) pool_block_err() argument
159 pool_init_page(struct dma_pool *pool, struct dma_page *page) pool_init_page() argument
164 pool_check_block(struct dma_pool *pool, struct dma_block *block, gfp_t mem_flags) pool_check_block() argument
169 pool_block_err(struct dma_pool *pool, void *vaddr, dma_addr_t dma) pool_block_err() argument
176 pool_init_page(struct dma_pool *pool, struct dma_page *page) pool_init_page() argument
181 pool_block_pop(struct dma_pool *pool) pool_block_pop() argument
192 pool_block_push(struct dma_pool *pool, struct dma_block *block, dma_addr_t dma) pool_block_push() argument
300 pool_initialise_page(struct dma_pool *pool, struct dma_page *page) pool_initialise_page() argument
334 pool_alloc_page(struct dma_pool *pool, gfp_t mem_flags) pool_alloc_page() argument
360 dma_pool_destroy(struct dma_pool *pool) dma_pool_destroy() argument
404 dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, dma_addr_t *handle) dma_pool_alloc() argument
450 dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma) dma_pool_free() argument
469 struct dma_pool *pool = *(struct dma_pool **)res; dmam_pool_release() local
496 struct dma_pool **ptr, *pool; dmam_pool_create() local
518 dmam_pool_destroy(struct dma_pool *pool) dmam_pool_destroy() argument
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/
H A Drxe_pool.c81 static inline const char *pool_name(struct rxe_pool *pool) in pool_name() argument
83 return rxe_type_info[pool->type].name; in pool_name()
86 static int rxe_pool_init_index(struct rxe_pool *pool, u32 max, u32 min) in rxe_pool_init_index() argument
91 if ((max - min + 1) < pool->max_elem) { in rxe_pool_init_index()
97 pool->max_index = max; in rxe_pool_init_index()
98 pool->min_index = min; in rxe_pool_init_index()
101 pool->table = kmalloc(size, GFP_KERNEL); in rxe_pool_init_index()
102 if (!pool->table) { in rxe_pool_init_index()
107 pool->table_size = size; in rxe_pool_init_index()
108 bitmap_zero(pool in rxe_pool_init_index()
114 rxe_pool_init( struct rxe_dev *rxe, struct rxe_pool *pool, enum rxe_elem_type type, unsigned int max_elem) rxe_pool_init() argument
160 struct rxe_pool *pool = container_of(kref, struct rxe_pool, ref_cnt); rxe_pool_release() local
166 rxe_pool_put(struct rxe_pool *pool) rxe_pool_put() argument
171 rxe_pool_cleanup(struct rxe_pool *pool) rxe_pool_cleanup() argument
185 alloc_index(struct rxe_pool *pool) alloc_index() argument
200 insert_index(struct rxe_pool *pool, struct rxe_pool_entry *new) insert_index() argument
227 insert_key(struct rxe_pool *pool, struct rxe_pool_entry *new) insert_key() argument
261 struct rxe_pool *pool = elem->pool; rxe_add_key() local
273 struct rxe_pool *pool = elem->pool; rxe_drop_key() local
284 struct rxe_pool *pool = elem->pool; rxe_add_index() local
296 struct rxe_pool *pool = elem->pool; rxe_drop_index() local
305 rxe_alloc(struct rxe_pool *pool) rxe_alloc() argument
345 rxe_add_to_pool(struct rxe_pool *pool, struct rxe_pool_entry *elem) rxe_add_to_pool() argument
382 struct rxe_pool *pool = elem->pool; rxe_elem_release() local
394 rxe_pool_get_index(struct rxe_pool *pool, u32 index) rxe_pool_get_index() argument
425 rxe_pool_get_key(struct rxe_pool *pool, void *key) rxe_pool_get_key() argument
[all...]
/kernel/linux/linux-5.10/mm/
H A Dmempool.c5 * memory buffer pool support. Such pools are mostly used
25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument
28 const int nr = pool->curr_nr; in poison_error()
34 pr_err("Mempool %p size %zu\n", pool, size); in poison_error()
42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument
51 poison_error(pool, element, size, i); in __check_element()
58 static void check_element(mempool_t *pool, void *element) in check_element() argument
61 if (pool->free == mempool_free_slab || pool->free == mempool_kfree) { in check_element()
62 __check_element(pool, elemen in check_element()
81 poison_element(mempool_t *pool, void *element) poison_element() argument
96 check_element(mempool_t *pool, void *element) check_element() argument
99 poison_element(mempool_t *pool, void *element) poison_element() argument
104 kasan_poison_element(mempool_t *pool, void *element) kasan_poison_element() argument
112 kasan_unpoison_element(mempool_t *pool, void *element) kasan_unpoison_element() argument
120 add_element(mempool_t *pool, void *element) add_element() argument
128 remove_element(mempool_t *pool) remove_element() argument
149 mempool_exit(mempool_t *pool) mempool_exit() argument
168 mempool_destroy(mempool_t *pool) mempool_destroy() argument
178 mempool_init_node(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, mempool_free_t *free_fn, void *pool_data, gfp_t gfp_mask, int node_id) mempool_init_node() argument
226 mempool_init(mempool_t *pool, int min_nr, mempool_alloc_t *alloc_fn, mempool_free_t *free_fn, void *pool_data) mempool_init() argument
263 mempool_t *pool; mempool_create_node() local
297 mempool_resize(mempool_t *pool, int new_min_nr) mempool_resize() argument
373 mempool_alloc(mempool_t *pool, gfp_t gfp_mask) mempool_alloc() argument
450 mempool_free(void *element, mempool_t *pool) mempool_free() argument
[all...]
H A Dzbud.c63 * freelists maintained in each pool. NCHUNKS_ORDER of 6 means that the
67 * 63 freelists per pool.
77 * struct zbud_pool - stores metadata for each zbud pool
78 * @lock: protects all pool fields and first|last_chunk fields of any
79 * zbud page in the pool
87 * @pages_nr: number of zbud pages in the pool.
89 * pool creation time.
91 * This structure is allocated at pool creation time and maintains metadata
92 * pertaining to a particular zbud pool.
110 * @buddy: links the zbud page into the unbuddied/buddied lists in the pool
129 zbud_zpool_evict(struct zbud_pool *pool, unsigned long handle) zbud_zpool_evict() argument
145 struct zbud_pool *pool; zbud_zpool_create() local
155 zbud_zpool_destroy(void *pool) zbud_zpool_destroy() argument
160 zbud_zpool_malloc(void *pool, size_t size, gfp_t gfp, unsigned long *handle) zbud_zpool_malloc() argument
165 zbud_zpool_free(void *pool, unsigned long handle) zbud_zpool_free() argument
170 zbud_zpool_shrink(void *pool, unsigned int pages, unsigned int *reclaimed) zbud_zpool_shrink() argument
189 zbud_zpool_map(void *pool, unsigned long handle, enum zpool_mapmode mm) zbud_zpool_map() argument
194 zbud_zpool_unmap(void *pool, unsigned long handle) zbud_zpool_unmap() argument
199 zbud_zpool_total_size(void *pool) zbud_zpool_total_size() argument
309 struct zbud_pool *pool; zbud_create_pool() local
331 zbud_destroy_pool(struct zbud_pool *pool) zbud_destroy_pool() argument
355 zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp, unsigned long *handle) zbud_alloc() argument
430 zbud_free(struct zbud_pool *pool, unsigned long handle) zbud_free() argument
502 zbud_reclaim_page(struct zbud_pool *pool, unsigned int retries) zbud_reclaim_page() argument
584 zbud_map(struct zbud_pool *pool, unsigned long handle) zbud_map() argument
594 zbud_unmap(struct zbud_pool *pool, unsigned long handle) zbud_unmap() argument
605 zbud_get_pool_size(struct zbud_pool *pool) zbud_get_pool_size() argument
[all...]
/kernel/linux/linux-5.10/include/net/
H A Dxdp_sock_drv.h14 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries);
15 bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc);
16 void xsk_tx_release(struct xsk_buff_pool *pool);
19 void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool);
20 void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool);
21 void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool);
22 void xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool);
23 bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool);
25 static inline u32 xsk_pool_get_headroom(struct xsk_buff_pool *pool) in xsk_pool_get_headroom() argument
27 return XDP_PACKET_HEADROOM + pool in xsk_pool_get_headroom()
30 xsk_pool_get_chunk_size(struct xsk_buff_pool *pool) xsk_pool_get_chunk_size() argument
35 xsk_pool_get_rx_frame_size(struct xsk_buff_pool *pool) xsk_pool_get_rx_frame_size() argument
40 xsk_pool_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xsk_pool_set_rxq_info() argument
46 xsk_pool_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xsk_pool_dma_unmap() argument
52 xsk_pool_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs) xsk_pool_dma_map() argument
74 xsk_buff_alloc(struct xsk_buff_pool *pool) xsk_buff_alloc() argument
79 xsk_buff_can_alloc(struct xsk_buff_pool *pool, u32 count) xsk_buff_can_alloc() argument
91 xsk_buff_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_dma() argument
97 xsk_buff_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_data() argument
102 xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) xsk_buff_dma_sync_for_cpu() argument
112 xsk_buff_raw_dma_sync_for_device(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xsk_buff_raw_dma_sync_for_device() argument
121 xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) xsk_tx_completed() argument
125 xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc) xsk_tx_peek_desc() argument
131 xsk_tx_release(struct xsk_buff_pool *pool) xsk_tx_release() argument
141 xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool) xsk_set_rx_need_wakeup() argument
145 xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool) xsk_set_tx_need_wakeup() argument
149 xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool) xsk_clear_rx_need_wakeup() argument
153 xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool) xsk_clear_tx_need_wakeup() argument
157 xsk_uses_need_wakeup(struct xsk_buff_pool *pool) xsk_uses_need_wakeup() argument
162 xsk_pool_get_headroom(struct xsk_buff_pool *pool) xsk_pool_get_headroom() argument
167 xsk_pool_get_chunk_size(struct xsk_buff_pool *pool) xsk_pool_get_chunk_size() argument
172 xsk_pool_get_rx_frame_size(struct xsk_buff_pool *pool) xsk_pool_get_rx_frame_size() argument
177 xsk_pool_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xsk_pool_set_rxq_info() argument
182 xsk_pool_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xsk_pool_dma_unmap() argument
187 xsk_pool_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs) xsk_pool_dma_map() argument
203 xsk_buff_alloc(struct xsk_buff_pool *pool) xsk_buff_alloc() argument
208 xsk_buff_can_alloc(struct xsk_buff_pool *pool, u32 count) xsk_buff_can_alloc() argument
217 xsk_buff_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_dma() argument
223 xsk_buff_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_data() argument
228 xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) xsk_buff_dma_sync_for_cpu() argument
232 xsk_buff_raw_dma_sync_for_device(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xsk_buff_raw_dma_sync_for_device() argument
[all...]
/kernel/linux/linux-5.10/sound/core/seq/
H A Dseq_memory.c22 static inline int snd_seq_pool_available(struct snd_seq_pool *pool) in snd_seq_pool_available() argument
24 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available()
27 static inline int snd_seq_output_ok(struct snd_seq_pool *pool) in snd_seq_output_ok() argument
29 return snd_seq_pool_available(pool) >= pool->room; in snd_seq_output_ok()
166 static inline void free_cell(struct snd_seq_pool *pool, in free_cell() argument
169 cell->next = pool->free; in free_cell()
170 pool->free = cell; in free_cell()
171 atomic_dec(&pool in free_cell()
177 struct snd_seq_pool *pool; snd_seq_cell_free() local
210 snd_seq_cell_alloc(struct snd_seq_pool *pool, struct snd_seq_event_cell **cellp, int nonblock, struct file *file, struct mutex *mutexp) snd_seq_cell_alloc() argument
282 snd_seq_event_dup(struct snd_seq_pool *pool, struct snd_seq_event *event, struct snd_seq_event_cell **cellp, int nonblock, struct file *file, struct mutex *mutexp) snd_seq_event_dup() argument
363 snd_seq_pool_poll_wait(struct snd_seq_pool *pool, struct file *file, poll_table *wait) snd_seq_pool_poll_wait() argument
372 snd_seq_pool_init(struct snd_seq_pool *pool) snd_seq_pool_init() argument
412 snd_seq_pool_mark_closing(struct snd_seq_pool *pool) snd_seq_pool_mark_closing() argument
424 snd_seq_pool_done(struct snd_seq_pool *pool) snd_seq_pool_done() argument
459 struct snd_seq_pool *pool; snd_seq_pool_new() local
483 struct snd_seq_pool *pool = *ppool; snd_seq_pool_delete() local
495 snd_seq_info_pool(struct snd_info_buffer *buffer, struct snd_seq_pool *pool, char *space) snd_seq_info_pool() argument
[all...]
/kernel/linux/linux-6.6/sound/core/seq/
H A Dseq_memory.c22 static inline int snd_seq_pool_available(struct snd_seq_pool *pool) in snd_seq_pool_available() argument
24 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available()
27 static inline int snd_seq_output_ok(struct snd_seq_pool *pool) in snd_seq_output_ok() argument
29 return snd_seq_pool_available(pool) >= pool->room; in snd_seq_output_ok()
225 static inline void free_cell(struct snd_seq_pool *pool, in free_cell() argument
228 cell->next = pool->free; in free_cell()
229 pool->free = cell; in free_cell()
230 atomic_dec(&pool in free_cell()
236 struct snd_seq_pool *pool; snd_seq_cell_free() local
269 snd_seq_cell_alloc(struct snd_seq_pool *pool, struct snd_seq_event_cell **cellp, int nonblock, struct file *file, struct mutex *mutexp) snd_seq_cell_alloc() argument
341 snd_seq_event_dup(struct snd_seq_pool *pool, struct snd_seq_event *event, struct snd_seq_event_cell **cellp, int nonblock, struct file *file, struct mutex *mutexp) snd_seq_event_dup() argument
428 snd_seq_pool_poll_wait(struct snd_seq_pool *pool, struct file *file, poll_table *wait) snd_seq_pool_poll_wait() argument
437 snd_seq_pool_init(struct snd_seq_pool *pool) snd_seq_pool_init() argument
477 snd_seq_pool_mark_closing(struct snd_seq_pool *pool) snd_seq_pool_mark_closing() argument
489 snd_seq_pool_done(struct snd_seq_pool *pool) snd_seq_pool_done() argument
524 struct snd_seq_pool *pool; snd_seq_pool_new() local
548 struct snd_seq_pool *pool = *ppool; snd_seq_pool_delete() local
560 snd_seq_info_pool(struct snd_info_buffer *buffer, struct snd_seq_pool *pool, char *space) snd_seq_info_pool() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/ttm/
H A Dttm_page_alloc_dma.c27 * A simple DMA pool losely based on dmapool.c. It has certain advantages
70 * The pool structure. There are up to nine pools:
80 * @type: Type of the pool
82 * used with irqsave/irqrestore variants because pool allocator maybe called
89 * @nfrees: Stats when pool is shrinking.
90 * @nrefills: Stats when the pool is grown.
92 * @name: Name of the pool.
117 * huge pool
129 * Limits for the pool. They are handled without locks because only place where
144 * @dev: The 'struct device' associated with the 'pool'
150 struct dma_pool *pool; global() member
265 ttm_set_pages_caching(struct dma_pool *pool, struct page **pages, unsigned cpages) ttm_set_pages_caching() argument
285 __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page) __ttm_dma_free_page() argument
298 __ttm_dma_alloc_page(struct dma_pool *pool) __ttm_dma_alloc_page() argument
343 ttm_pool_update_free_locked(struct dma_pool *pool, unsigned freed_pages) ttm_pool_update_free_locked() argument
352 ttm_dma_page_put(struct dma_pool *pool, struct dma_page *d_page) ttm_dma_page_put() argument
369 ttm_dma_pages_put(struct dma_pool *pool, struct list_head *d_pages, struct page *pages[], unsigned npages) ttm_dma_pages_put() argument
403 ttm_dma_page_pool_free(struct dma_pool *pool, unsigned nr_free, bool use_static) ttm_dma_page_pool_free() argument
500 struct dma_pool *pool; ttm_dma_free_pool() local
542 struct dma_pool *pool = *(struct dma_pool **)res; ttm_dma_pool_release() local
559 struct dma_pool *pool = NULL, **ptr; ttm_dma_pool_init() local
638 struct dma_pool *pool, *tmp; ttm_dma_find_pool() local
665 ttm_dma_handle_caching_state_failure(struct dma_pool *pool, struct list_head *d_pages, struct page **failed_pages, unsigned cpages) ttm_dma_handle_caching_state_failure() argument
698 ttm_dma_pool_alloc_new_pages(struct dma_pool *pool, struct list_head *d_pages, unsigned count) ttm_dma_pool_alloc_new_pages() argument
786 ttm_dma_page_pool_fill_locked(struct dma_pool *pool, unsigned long *irq_flags) ttm_dma_page_pool_fill_locked() argument
833 ttm_dma_pool_get_pages(struct dma_pool *pool, struct ttm_dma_tt *ttm_dma, unsigned index) ttm_dma_pool_get_pages() argument
891 struct dma_pool *pool; ttm_dma_populate() local
995 struct dma_pool *pool; ttm_dma_unpopulate() local
1217 struct dma_pool *pool = NULL; ttm_dma_page_alloc_debugfs() local
[all...]
/kernel/linux/linux-5.10/net/ceph/
H A Dmsgpool.c14 struct ceph_msgpool *pool = arg; in msgpool_alloc() local
17 msg = ceph_msg_new2(pool->type, pool->front_len, pool->max_data_items, in msgpool_alloc()
20 dout("msgpool_alloc %s failed\n", pool->name); in msgpool_alloc()
22 dout("msgpool_alloc %s %p\n", pool->name, msg); in msgpool_alloc()
23 msg->pool = pool; in msgpool_alloc()
30 struct ceph_msgpool *pool = arg; in msgpool_free() local
33 dout("msgpool_release %s %p\n", pool in msgpool_free()
38 ceph_msgpool_init(struct ceph_msgpool *pool, int type, int front_len, int max_data_items, int size, const char *name) ceph_msgpool_init() argument
53 ceph_msgpool_destroy(struct ceph_msgpool *pool) ceph_msgpool_destroy() argument
59 ceph_msgpool_get(struct ceph_msgpool *pool, int front_len, int max_data_items) ceph_msgpool_get() argument
81 ceph_msgpool_put(struct ceph_msgpool *pool, struct ceph_msg *msg) ceph_msgpool_put() argument
[all...]
/kernel/linux/linux-6.6/net/ceph/
H A Dmsgpool.c14 struct ceph_msgpool *pool = arg; in msgpool_alloc() local
17 msg = ceph_msg_new2(pool->type, pool->front_len, pool->max_data_items, in msgpool_alloc()
20 dout("msgpool_alloc %s failed\n", pool->name); in msgpool_alloc()
22 dout("msgpool_alloc %s %p\n", pool->name, msg); in msgpool_alloc()
23 msg->pool = pool; in msgpool_alloc()
30 struct ceph_msgpool *pool = arg; in msgpool_free() local
33 dout("msgpool_release %s %p\n", pool in msgpool_free()
38 ceph_msgpool_init(struct ceph_msgpool *pool, int type, int front_len, int max_data_items, int size, const char *name) ceph_msgpool_init() argument
53 ceph_msgpool_destroy(struct ceph_msgpool *pool) ceph_msgpool_destroy() argument
59 ceph_msgpool_get(struct ceph_msgpool *pool, int front_len, int max_data_items) ceph_msgpool_get() argument
81 ceph_msgpool_put(struct ceph_msgpool *pool, struct ceph_msg *msg) ceph_msgpool_put() argument
[all...]
/kernel/linux/linux-5.10/drivers/staging/media/atomisp/pci/runtime/rmgr/src/
H A Drmgr_vbuf.c31 * @brief VBUF resource pool - refpool
42 * @brief VBUF resource pool - writepool
53 * @brief VBUF resource pool - hmmbufferpool
137 * @brief Initialize the resource pool (host, vbuf)
139 * @param pool The pointer to the pool
141 int ia_css_rmgr_init_vbuf(struct ia_css_rmgr_vbuf_pool *pool) in ia_css_rmgr_init_vbuf() argument
147 assert(pool); in ia_css_rmgr_init_vbuf()
148 if (!pool) in ia_css_rmgr_init_vbuf()
150 /* initialize the recycle pool i in ia_css_rmgr_init_vbuf()
174 ia_css_rmgr_uninit_vbuf(struct ia_css_rmgr_vbuf_pool *pool) ia_css_rmgr_uninit_vbuf() argument
212 rmgr_push_handle(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) rmgr_push_handle() argument
240 rmgr_pop_handle(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) rmgr_pop_handle() argument
268 ia_css_rmgr_acq_vbuf(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) ia_css_rmgr_acq_vbuf() argument
316 ia_css_rmgr_rel_vbuf(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) ia_css_rmgr_rel_vbuf() argument
[all...]
/kernel/linux/linux-6.6/drivers/staging/media/atomisp/pci/runtime/rmgr/src/
H A Drmgr_vbuf.c31 * @brief VBUF resource pool - refpool
36 * @brief VBUF resource pool - writepool
43 * @brief VBUF resource pool - hmmbufferpool
124 * @brief Initialize the resource pool (host, vbuf)
126 * @param pool The pointer to the pool
128 int ia_css_rmgr_init_vbuf(struct ia_css_rmgr_vbuf_pool *pool) in ia_css_rmgr_init_vbuf() argument
134 assert(pool); in ia_css_rmgr_init_vbuf()
135 if (!pool) in ia_css_rmgr_init_vbuf()
137 /* initialize the recycle pool i in ia_css_rmgr_init_vbuf()
161 ia_css_rmgr_uninit_vbuf(struct ia_css_rmgr_vbuf_pool *pool) ia_css_rmgr_uninit_vbuf() argument
197 rmgr_push_handle(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) rmgr_push_handle() argument
225 rmgr_pop_handle(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) rmgr_pop_handle() argument
254 ia_css_rmgr_acq_vbuf(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) ia_css_rmgr_acq_vbuf() argument
310 ia_css_rmgr_rel_vbuf(struct ia_css_rmgr_vbuf_pool *pool, struct ia_css_rmgr_vbuf_handle **handle) ia_css_rmgr_rel_vbuf() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_icm_pool.c24 struct mutex mutex; /* protect the ICM pool and ICM buddy */
82 u32 offset = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_mr_addr()
94 u32 size = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_icm_addr()
102 chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_byte_size()
111 dr_icm_pool_mr_create(struct mlx5dr_icm_pool *pool) in dr_icm_pool_mr_create() argument
113 struct mlx5_core_dev *mdev = pool->dmn->mdev; in dr_icm_pool_mr_create()
123 icm_mr->dmn = pool->dmn; in dr_icm_pool_mr_create()
125 icm_mr->dm.length = mlx5dr_icm_pool_chunk_size_to_byte(pool->max_log_chunk_sz, in dr_icm_pool_mr_create()
126 pool->icm_type); in dr_icm_pool_mr_create()
128 switch (pool in dr_icm_pool_mr_create()
263 dr_icm_buddy_create(struct mlx5dr_icm_pool *pool) dr_icm_buddy_create() argument
321 dr_icm_chunk_init(struct mlx5dr_icm_chunk *chunk, struct mlx5dr_icm_pool *pool, enum mlx5dr_icm_chunk_size chunk_size, struct mlx5dr_icm_buddy_mem *buddy_mem_pool, unsigned int seg) dr_icm_chunk_init() argument
341 dr_icm_pool_is_sync_required(struct mlx5dr_icm_pool *pool) dr_icm_pool_is_sync_required() argument
346 dr_icm_pool_clear_hot_chunks_arr(struct mlx5dr_icm_pool *pool) dr_icm_pool_clear_hot_chunks_arr() argument
365 dr_icm_pool_sync_all_buddy_pools(struct mlx5dr_icm_pool *pool) dr_icm_pool_sync_all_buddy_pools() argument
386 dr_icm_handle_buddies_get_mem(struct mlx5dr_icm_pool *pool, enum mlx5dr_icm_chunk_size chunk_size, struct mlx5dr_icm_buddy_mem **buddy, unsigned int *seg) dr_icm_handle_buddies_get_mem() argument
435 mlx5dr_icm_alloc_chunk(struct mlx5dr_icm_pool *pool, enum mlx5dr_icm_chunk_size chunk_size) mlx5dr_icm_alloc_chunk() argument
470 struct mlx5dr_icm_pool *pool = buddy->pool; mlx5dr_icm_free_chunk() local
495 mlx5dr_icm_pool_alloc_htbl(struct mlx5dr_icm_pool *pool) mlx5dr_icm_pool_alloc_htbl() argument
500 mlx5dr_icm_pool_free_htbl(struct mlx5dr_icm_pool *pool, struct mlx5dr_ste_htbl *htbl) mlx5dr_icm_pool_free_htbl() argument
509 struct mlx5dr_icm_pool *pool; mlx5dr_icm_pool_create() local
564 mlx5dr_icm_pool_destroy(struct mlx5dr_icm_pool *pool) mlx5dr_icm_pool_destroy() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/
H A Dirq_affinity.c8 static void cpu_put(struct mlx5_irq_pool *pool, int cpu) in cpu_put() argument
10 pool->irqs_per_cpu[cpu]--; in cpu_put()
13 static void cpu_get(struct mlx5_irq_pool *pool, int cpu) in cpu_get() argument
15 pool->irqs_per_cpu[cpu]++; in cpu_get()
19 static int cpu_get_least_loaded(struct mlx5_irq_pool *pool, in cpu_get_least_loaded() argument
27 if (!pool->irqs_per_cpu[cpu]) { in cpu_get_least_loaded()
33 if (pool->irqs_per_cpu[cpu] < pool->irqs_per_cpu[best_cpu]) in cpu_get_least_loaded()
38 mlx5_core_err(pool->dev, "NO online CPUs in req_mask (%*pbl)\n", in cpu_get_least_loaded()
42 pool in cpu_get_least_loaded()
48 irq_pool_request_irq(struct mlx5_irq_pool *pool, struct irq_affinity_desc *af_desc) irq_pool_request_irq() argument
82 irq_pool_find_least_loaded(struct mlx5_irq_pool *pool, const struct cpumask *req_mask) irq_pool_find_least_loaded() argument
121 mlx5_irq_affinity_request(struct mlx5_irq_pool *pool, struct irq_affinity_desc *af_desc) mlx5_irq_affinity_request() argument
161 struct mlx5_irq_pool *pool = mlx5_irq_pool_get(dev); mlx5_irq_affinity_irq_release() local
[all...]
/kernel/linux/linux-6.6/include/net/
H A Dxdp_sock_drv.h17 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries);
18 bool xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc);
19 u32 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 max);
20 void xsk_tx_release(struct xsk_buff_pool *pool);
23 void xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool);
24 void xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool);
25 void xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool);
26 void xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool);
27 bool xsk_uses_need_wakeup(struct xsk_buff_pool *pool);
29 static inline u32 xsk_pool_get_headroom(struct xsk_buff_pool *pool) in xsk_pool_get_headroom() argument
34 xsk_pool_get_chunk_size(struct xsk_buff_pool *pool) xsk_pool_get_chunk_size() argument
39 xsk_pool_get_rx_frame_size(struct xsk_buff_pool *pool) xsk_pool_get_rx_frame_size() argument
44 xsk_pool_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xsk_pool_set_rxq_info() argument
50 xsk_pool_get_napi_id(struct xsk_buff_pool *pool) xsk_pool_get_napi_id() argument
59 xsk_pool_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xsk_pool_dma_unmap() argument
65 xsk_pool_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs) xsk_pool_dma_map() argument
87 xsk_buff_alloc(struct xsk_buff_pool *pool) xsk_buff_alloc() argument
98 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) xsk_buff_alloc_batch() argument
103 xsk_buff_can_alloc(struct xsk_buff_pool *pool, u32 count) xsk_buff_can_alloc() argument
175 xsk_buff_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_dma() argument
181 xsk_buff_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_data() argument
186 xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) xsk_buff_dma_sync_for_cpu() argument
196 xsk_buff_raw_dma_sync_for_device(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xsk_buff_raw_dma_sync_for_device() argument
205 xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) xsk_tx_completed() argument
209 xsk_tx_peek_desc(struct xsk_buff_pool *pool, struct xdp_desc *desc) xsk_tx_peek_desc() argument
215 xsk_tx_peek_release_desc_batch(struct xsk_buff_pool *pool, u32 max) xsk_tx_peek_release_desc_batch() argument
220 xsk_tx_release(struct xsk_buff_pool *pool) xsk_tx_release() argument
230 xsk_set_rx_need_wakeup(struct xsk_buff_pool *pool) xsk_set_rx_need_wakeup() argument
234 xsk_set_tx_need_wakeup(struct xsk_buff_pool *pool) xsk_set_tx_need_wakeup() argument
238 xsk_clear_rx_need_wakeup(struct xsk_buff_pool *pool) xsk_clear_rx_need_wakeup() argument
242 xsk_clear_tx_need_wakeup(struct xsk_buff_pool *pool) xsk_clear_tx_need_wakeup() argument
246 xsk_uses_need_wakeup(struct xsk_buff_pool *pool) xsk_uses_need_wakeup() argument
251 xsk_pool_get_headroom(struct xsk_buff_pool *pool) xsk_pool_get_headroom() argument
256 xsk_pool_get_chunk_size(struct xsk_buff_pool *pool) xsk_pool_get_chunk_size() argument
261 xsk_pool_get_rx_frame_size(struct xsk_buff_pool *pool) xsk_pool_get_rx_frame_size() argument
266 xsk_pool_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) xsk_pool_set_rxq_info() argument
271 xsk_pool_get_napi_id(struct xsk_buff_pool *pool) xsk_pool_get_napi_id() argument
276 xsk_pool_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) xsk_pool_dma_unmap() argument
281 xsk_pool_dma_map(struct xsk_buff_pool *pool, struct device *dev, unsigned long attrs) xsk_pool_dma_map() argument
297 xsk_buff_alloc(struct xsk_buff_pool *pool) xsk_buff_alloc() argument
307 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) xsk_buff_alloc_batch() argument
312 xsk_buff_can_alloc(struct xsk_buff_pool *pool, u32 count) xsk_buff_can_alloc() argument
343 xsk_buff_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_dma() argument
349 xsk_buff_raw_get_data(struct xsk_buff_pool *pool, u64 addr) xsk_buff_raw_get_data() argument
354 xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) xsk_buff_dma_sync_for_cpu() argument
358 xsk_buff_raw_dma_sync_for_device(struct xsk_buff_pool *pool, dma_addr_t dma, size_t size) xsk_buff_raw_dma_sync_for_device() argument
[all...]

Completed in 239 milliseconds

12345678910>>...33