/kernel/linux/linux-6.6/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_xsk.c | 27 * @pool_present: is pool for XSK present 77 * i40e_xsk_pool_enable - Enable/associate an AF_XDP buffer pool to a 80 * @pool: buffer pool 81 * @qid: Rx ring to associate buffer pool with 86 struct xsk_buff_pool *pool, in i40e_xsk_pool_enable() 103 err = xsk_pool_dma_map(pool, &vsi->back->pdev->dev, I40E_RX_DMA_ATTR); in i40e_xsk_pool_enable() 134 * i40e_xsk_pool_disable - Disassociate an AF_XDP buffer pool from a 137 * @qid: Rx ring to associate buffer pool with 144 struct xsk_buff_pool *pool; in i40e_xsk_pool_disable() local 85 i40e_xsk_pool_enable(struct i40e_vsi *vsi, struct xsk_buff_pool *pool, u16 qid) i40e_xsk_pool_enable() argument 186 i40e_xsk_pool_setup(struct i40e_vsi *vsi, struct xsk_buff_pool *pool, u16 qid) i40e_xsk_pool_setup() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/ti/ |
H A D | cpsw_priv.c | 1103 struct page_pool *pool; in cpsw_fill_rx_channels() local 1110 pool = cpsw->page_pool[ch]; in cpsw_fill_rx_channels() 1113 page = page_pool_dev_alloc_pages(pool); in cpsw_fill_rx_channels() 1132 page_pool_recycle_direct(pool, page); in cpsw_fill_rx_channels() 1148 struct page_pool *pool; in cpsw_create_page_pool() local 1157 pool = page_pool_create(&pp_params); in cpsw_create_page_pool() 1158 if (IS_ERR(pool)) in cpsw_create_page_pool() 1159 dev_err(cpsw->dev, "cannot create rx page pool\n"); in cpsw_create_page_pool() 1161 return pool; in cpsw_create_page_pool() 1166 struct page_pool *pool; in cpsw_create_rx_pool() local 1183 struct page_pool *pool; cpsw_ndev_create_xdp_rxq() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm_sdma.c | 64 enum amdgpu_ib_pool_type pool = p->immediate ? AMDGPU_IB_POOL_IMMEDIATE in amdgpu_vm_sdma_prepare() local 69 r = amdgpu_job_alloc_with_ib(p->adev, ndw * 4, pool, &p->job); in amdgpu_vm_sdma_prepare() 204 enum amdgpu_ib_pool_type pool = p->immediate ? AMDGPU_IB_POOL_IMMEDIATE in amdgpu_vm_sdma_update() local 231 r = amdgpu_job_alloc_with_ib(p->adev, ndw * 4, pool, in amdgpu_vm_sdma_update()
|
/kernel/linux/linux-5.10/drivers/media/platform/vsp1/ |
H A D | vsp1_clu.c | 48 dlb = vsp1_dl_body_get(clu->pool); in clu_set_table() 225 vsp1_dl_body_pool_destroy(clu->pool); in clu_destroy() 258 * Pre-allocate a body pool, with 3 bodies allowing a userspace update in vsp1_clu_create() 263 clu->pool = vsp1_dl_body_pool_create(clu->entity.vsp1, 3, CLU_SIZE + 1, in vsp1_clu_create() 265 if (!clu->pool) in vsp1_clu_create()
|
/kernel/linux/linux-6.6/drivers/media/platform/renesas/vsp1/ |
H A D | vsp1_clu.c | 48 dlb = vsp1_dl_body_get(clu->pool); in clu_set_table() 226 vsp1_dl_body_pool_destroy(clu->pool); in clu_destroy() 259 * Pre-allocate a body pool, with 3 bodies allowing a userspace update in vsp1_clu_create() 264 clu->pool = vsp1_dl_body_pool_create(clu->entity.vsp1, 3, CLU_SIZE + 1, in vsp1_clu_create() 266 if (!clu->pool) in vsp1_clu_create()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 280 struct intel_gt_buffer_pool_node *pool; member 283 struct intel_gt_buffer_pool_node *reloc_pool; /** relocation pool for -EDEADLK handling */ 292 struct intel_gt_buffer_pool_node *batch_pool; /** pool node for batch buffer */ 987 cache->pool = NULL; in reloc_cache_clear() 1027 if (!cache->pool) in reloc_cache_put_pool() 1033 * dropping the reference. Fortunately we can only hold 1 pool node at in reloc_cache_put_pool() 1036 i915_gem_ww_unlock_single(cache->pool->obj); in reloc_cache_put_pool() 1037 intel_gt_buffer_pool_put(cache->pool); in reloc_cache_put_pool() 1038 cache->pool = NULL; in reloc_cache_put_pool() 1272 struct intel_gt_buffer_pool_node *pool in __reloc_gpu_alloc() local 2288 struct intel_gt_buffer_pool_node *pool = eb->batch_pool; eb_parse() local [all...] |
/kernel/linux/linux-5.10/fs/ceph/ |
H A D | addr.c | 1898 s64 pool, struct ceph_string *pool_ns) in __ceph_pool_perm_get() 1913 if (pool < perm->pool) in __ceph_pool_perm_get() 1915 else if (pool > perm->pool) in __ceph_pool_perm_get() 1936 dout("__ceph_pool_perm_get pool %lld ns %.*s no perm cached\n", in __ceph_pool_perm_get() 1937 pool, (int)pool_ns->len, pool_ns->str); in __ceph_pool_perm_get() 1939 dout("__ceph_pool_perm_get pool %lld no perm cached\n", pool); in __ceph_pool_perm_get() 1947 if (pool < per in __ceph_pool_perm_get() 1897 __ceph_pool_perm_get(struct ceph_inode_info *ci, s64 pool, struct ceph_string *pool_ns) __ceph_pool_perm_get() argument 2077 s64 pool; ceph_pool_perm_check() local [all...] |
/kernel/linux/linux-6.6/fs/ceph/ |
H A D | addr.c | 1978 s64 pool, struct ceph_string *pool_ns) in __ceph_pool_perm_get() 1993 if (pool < perm->pool) in __ceph_pool_perm_get() 1995 else if (pool > perm->pool) in __ceph_pool_perm_get() 2016 dout("__ceph_pool_perm_get pool %lld ns %.*s no perm cached\n", in __ceph_pool_perm_get() 2017 pool, (int)pool_ns->len, pool_ns->str); in __ceph_pool_perm_get() 2019 dout("__ceph_pool_perm_get pool %lld no perm cached\n", pool); in __ceph_pool_perm_get() 2027 if (pool < per in __ceph_pool_perm_get() 1977 __ceph_pool_perm_get(struct ceph_inode_info *ci, s64 pool, struct ceph_string *pool_ns) __ceph_pool_perm_get() argument 2155 s64 pool; ceph_pool_perm_check() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/display/dc/dce/ |
H A D | dce_i2c_hw.c | 370 struct resource_pool *pool, in acquire_i2c_hw_engine() 383 if (line < pool->res_cap->num_ddc) in acquire_i2c_hw_engine() 384 dce_i2c_hw = pool->hw_i2cs[line]; in acquire_i2c_hw_engine() 390 if (pool->i2c_hw_buffer_in_use || !is_engine_available(dce_i2c_hw)) in acquire_i2c_hw_engine() 417 pool->i2c_hw_buffer_in_use = true; in acquire_i2c_hw_engine() 558 struct resource_pool *pool, in dce_i2c_submit_command_hw() 584 pool->i2c_hw_buffer_in_use = false; in dce_i2c_submit_command_hw() 369 acquire_i2c_hw_engine( struct resource_pool *pool, struct ddc *ddc) acquire_i2c_hw_engine() argument 557 dce_i2c_submit_command_hw( struct resource_pool *pool, struct ddc *ddc, struct i2c_command *cmd, struct dce_i2c_hw *dce_i2c_hw) dce_i2c_submit_command_hw() argument
|
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm-io.c | 25 mempool_t pool; member 58 ret = mempool_init_slab_pool(&client->pool, min_ios, _dm_io_cache); in dm_io_client_create() 69 mempool_exit(&client->pool); in dm_io_client_create() 77 mempool_exit(&client->pool); in dm_io_client_destroy() 124 mempool_free(io, &io->client->pool); in complete_io() 446 io = mempool_alloc(&client->pool, GFP_NOIO); in sync_io() 478 io = mempool_alloc(&client->pool, GFP_NOIO); in async_io()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/display/dc/dce/ |
H A D | dce_i2c_hw.c | 394 struct resource_pool *pool, in acquire_i2c_hw_engine() 407 if (line < pool->res_cap->num_ddc) in acquire_i2c_hw_engine() 408 dce_i2c_hw = pool->hw_i2cs[line]; in acquire_i2c_hw_engine() 414 if (pool->i2c_hw_buffer_in_use || !is_engine_available(dce_i2c_hw)) in acquire_i2c_hw_engine() 441 pool->i2c_hw_buffer_in_use = true; in acquire_i2c_hw_engine() 580 struct resource_pool *pool, in dce_i2c_submit_command_hw() 606 pool->i2c_hw_buffer_in_use = false; in dce_i2c_submit_command_hw() 393 acquire_i2c_hw_engine( struct resource_pool *pool, struct ddc *ddc) acquire_i2c_hw_engine() argument 579 dce_i2c_submit_command_hw( struct resource_pool *pool, struct ddc *ddc, struct i2c_command *cmd, struct dce_i2c_hw *dce_i2c_hw) dce_i2c_submit_command_hw() argument
|
/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-io.c | 26 mempool_t pool; member 59 ret = mempool_init_slab_pool(&client->pool, min_ios, _dm_io_cache); in dm_io_client_create() 70 mempool_exit(&client->pool); in dm_io_client_create() 78 mempool_exit(&client->pool); in dm_io_client_destroy() 129 mempool_free(io, &io->client->pool); in complete_io() 441 io = mempool_alloc(&client->pool, GFP_NOIO); in sync_io() 474 io = mempool_alloc(&client->pool, GFP_NOIO); in async_io()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_cmd.c | 215 hr_dev->cmd.pool = dma_pool_create("hns_roce_cmd", dev, in hns_roce_cmd_init() 218 if (!hr_dev->cmd.pool) in hns_roce_cmd_init() 226 dma_pool_destroy(hr_dev->cmd.pool); in hns_roce_cmd_cleanup() 274 dma_pool_alloc(hr_dev->cmd.pool, GFP_KERNEL, &mailbox->dma); in hns_roce_alloc_cmd_mailbox() 289 dma_pool_free(hr_dev->cmd.pool, mailbox->buf, mailbox->dma); in hns_roce_free_cmd_mailbox()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/display/dc/inc/ |
H A D | core_types.h | 94 void (*destroy)(struct resource_pool **pool); 120 const struct resource_pool *pool, 139 const struct resource_pool *pool, 157 const struct resource_pool *pool, 164 const struct resource_pool *pool,
|
/kernel/linux/linux-5.10/drivers/gpu/drm/sis/ |
H A D | sis_mm.c | 86 void *data, int pool) in sis_drm_alloc() 97 if (0 == ((pool == 0) ? dev_priv->vram_initialized : in sis_drm_alloc() 112 if (pool == AGP_TYPE) { in sis_drm_alloc() 142 mem->offset = ((pool == 0) ? in sis_drm_alloc() 160 DRM_DEBUG("alloc %d, size = %ld, offset = %ld\n", pool, mem->size, in sis_drm_alloc() 85 sis_drm_alloc(struct drm_device *dev, struct drm_file *file, void *data, int pool) sis_drm_alloc() argument
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/display/dc/inc/ |
H A D | core_types.h | 69 void (*destroy)(struct resource_pool **pool); 132 const struct resource_pool *pool, 153 const struct resource_pool *pool, 170 const struct resource_pool *pool, 177 const struct resource_pool *pool,
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/ |
H A D | ttm_tt.c | 309 if (bdev->pool.use_dma32) in ttm_tt_populate() 328 ret = ttm_pool_alloc(&bdev->pool, ttm, ctx); in ttm_tt_populate() 346 if (bdev->pool.use_dma32) in ttm_tt_populate() 362 ttm_pool_free(&bdev->pool, ttm); in ttm_tt_unpopulate() 366 if (bdev->pool.use_dma32) in ttm_tt_unpopulate()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_send.c | 61 struct mlx5dr_send_info_pool *pool; member 69 static int dr_send_info_pool_fill(struct mlx5dr_send_info_pool *pool) in dr_send_info_pool_fill() argument 79 pool_obj->pool = pool; in dr_send_info_pool_fill() 80 list_add_tail(&pool_obj->list_node, &pool->free_list); in dr_send_info_pool_fill() 86 list_for_each_entry_safe(pool_obj, tmp_pool_obj, &pool->free_list, list_node) { in dr_send_info_pool_fill() 94 static void dr_send_info_pool_destroy(struct mlx5dr_send_info_pool *pool) in dr_send_info_pool_destroy() argument 98 list_for_each_entry_safe(pool_obj, tmp_pool_obj, &pool->free_list, list_node) { in dr_send_info_pool_destroy() 103 kfree(pool); in dr_send_info_pool_destroy() 114 struct mlx5dr_send_info_pool *pool; in dr_send_info_pool_create() local 152 struct mlx5dr_send_info_pool *pool; mlx5dr_send_info_alloc() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_xsk.h | 12 int ice_xsk_pool_setup(struct ice_vsi *vsi, struct xsk_buff_pool *pool, 24 struct xsk_buff_pool __always_unused *pool, in ice_xsk_pool_setup() 23 ice_xsk_pool_setup(struct ice_vsi __always_unused *vsi, struct xsk_buff_pool __always_unused *pool, u16 __always_unused qid) ice_xsk_pool_setup() argument
|
/kernel/linux/linux-5.10/scripts/clang-tools/ |
H A D | run-clang-tools.py | 67 pool = multiprocessing.Pool(initializer=init, initargs=(lock, args)) 71 pool.map(run_analysis, datastore)
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_xsk.h | 21 int ice_xsk_pool_setup(struct ice_vsi *vsi, struct xsk_buff_pool *pool, 39 struct xsk_buff_pool __always_unused *pool, in ice_xsk_pool_setup() 38 ice_xsk_pool_setup(struct ice_vsi __always_unused *vsi, struct xsk_buff_pool __always_unused *pool, u16 __always_unused qid) ice_xsk_pool_setup() argument
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/display/dc/dcn31/ |
H A D | dcn31_resource.h | 31 #define TO_DCN31_RES_POOL(pool)\ 32 container_of(pool, struct dcn31_resource_pool, base)
|
/kernel/linux/linux-6.6/scripts/gdb/linux/ |
H A D | stackdepot.py | 31 gdb.write("pool index %d out of bounds (%d) for stack id 0x%08x\n" % (parts['pool_index'], pool_index_cached, handle)) 37 pool = stack_pools[parts['pool_index']] 38 stack = (pool + gdb.Value(offset).cast(utils.get_size_t_type())).cast(stack_record_type.get_type().pointer())
|
/kernel/linux/linux-5.10/net/ceph/ |
H A D | debugfs.c | 73 seq_printf(s, "pool %lld '%s' type %d size %d min_size %d pg_num %u pg_num_mask %d flags 0x%llx lfor %u read_tier %lld write_tier %lld\n", in osdmap_show() 96 seq_printf(s, "pg_temp %llu.%x [", pg->pgid.pool, in osdmap_show() 107 seq_printf(s, "primary_temp %llu.%x %d\n", pg->pgid.pool, in osdmap_show() 114 seq_printf(s, "pg_upmap %llu.%x [", pg->pgid.pool, in osdmap_show() 125 seq_printf(s, "pg_upmap_items %llu.%x [", pg->pgid.pool, in osdmap_show() 178 seq_printf(s, "%llu.%x", spgid->pgid.pool, spgid->pgid.seed); in dump_spgid() 187 seq_printf(s, "osd%d\t%llu.%x\t", t->osd, t->pgid.pool, t->pgid.seed); in dump_target() 303 hoid->pool == S64_MIN) { in dump_hoid() 311 seq_printf(s, "%lld:%08x:", hoid->pool, hoid->hash_reverse_bits); in dump_hoid()
|
/kernel/linux/linux-6.6/net/ceph/ |
H A D | debugfs.c | 73 seq_printf(s, "pool %lld '%s' type %d size %d min_size %d pg_num %u pg_num_mask %d flags 0x%llx lfor %u read_tier %lld write_tier %lld\n", in osdmap_show() 96 seq_printf(s, "pg_temp %llu.%x [", pg->pgid.pool, in osdmap_show() 107 seq_printf(s, "primary_temp %llu.%x %d\n", pg->pgid.pool, in osdmap_show() 114 seq_printf(s, "pg_upmap %llu.%x [", pg->pgid.pool, in osdmap_show() 125 seq_printf(s, "pg_upmap_items %llu.%x [", pg->pgid.pool, in osdmap_show() 178 seq_printf(s, "%llu.%x", spgid->pgid.pool, spgid->pgid.seed); in dump_spgid() 187 seq_printf(s, "osd%d\t%llu.%x\t", t->osd, t->pgid.pool, t->pgid.seed); in dump_target() 303 hoid->pool == S64_MIN) { in dump_hoid() 311 seq_printf(s, "%lld:%08x:", hoid->pool, hoid->hash_reverse_bits); in dump_hoid()
|