Home
last modified time | relevance | path

Searched refs:pool (Results 1 - 25 of 588) sorted by relevance

12345678910>>...24

/third_party/node/deps/openssl/openssl/crypto/rand/
H A Drand_pool.c20 * Allocate memory and initialize a new random pool
25 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); in ossl_rand_pool_new() local
28 if (pool == NULL) { in ossl_rand_pool_new()
33 pool->min_len = min_len; in ossl_rand_pool_new()
34 pool->max_len = (max_len > RAND_POOL_MAX_LENGTH) ? in ossl_rand_pool_new()
36 pool->alloc_len = min_len < min_alloc_size ? min_alloc_size : min_len; in ossl_rand_pool_new()
37 if (pool->alloc_len > pool->max_len) in ossl_rand_pool_new()
38 pool in ossl_rand_pool_new()
68 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); ossl_rand_pool_attach() local
94 ossl_rand_pool_free(RAND_POOL *pool) ossl_rand_pool_free() argument
118 ossl_rand_pool_buffer(RAND_POOL *pool) ossl_rand_pool_buffer() argument
126 ossl_rand_pool_entropy(RAND_POOL *pool) ossl_rand_pool_entropy() argument
134 ossl_rand_pool_length(RAND_POOL *pool) ossl_rand_pool_length() argument
145 ossl_rand_pool_detach(RAND_POOL *pool) ossl_rand_pool_detach() argument
157 ossl_rand_pool_reattach(RAND_POOL *pool, unsigned char *buffer) ossl_rand_pool_reattach() argument
180 ossl_rand_pool_entropy_available(RAND_POOL *pool) ossl_rand_pool_entropy_available() argument
196 ossl_rand_pool_entropy_needed(RAND_POOL *pool) ossl_rand_pool_entropy_needed() argument
205 rand_pool_grow(RAND_POOL *pool, size_t len) rand_pool_grow() argument
246 ossl_rand_pool_bytes_needed(RAND_POOL *pool, unsigned int entropy_factor) ossl_rand_pool_bytes_needed() argument
291 ossl_rand_pool_bytes_remaining(RAND_POOL *pool) ossl_rand_pool_bytes_remaining() argument
305 ossl_rand_pool_add(RAND_POOL *pool, const unsigned char *buffer, size_t len, size_t entropy) ossl_rand_pool_add() argument
360 ossl_rand_pool_add_begin(RAND_POOL *pool, size_t len) ossl_rand_pool_add_begin() argument
399 ossl_rand_pool_add_end(RAND_POOL *pool, size_t len, size_t entropy) ossl_rand_pool_add_end() argument
[all...]
/third_party/openssl/crypto/rand/
H A Drand_pool.c20 * Allocate memory and initialize a new random pool
25 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); in ossl_rand_pool_new() local
28 if (pool == NULL) { in ossl_rand_pool_new()
33 pool->min_len = min_len; in ossl_rand_pool_new()
34 pool->max_len = (max_len > RAND_POOL_MAX_LENGTH) ? in ossl_rand_pool_new()
36 pool->alloc_len = min_len < min_alloc_size ? min_alloc_size : min_len; in ossl_rand_pool_new()
37 if (pool->alloc_len > pool->max_len) in ossl_rand_pool_new()
38 pool in ossl_rand_pool_new()
68 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); ossl_rand_pool_attach() local
94 ossl_rand_pool_free(RAND_POOL *pool) ossl_rand_pool_free() argument
118 ossl_rand_pool_buffer(RAND_POOL *pool) ossl_rand_pool_buffer() argument
126 ossl_rand_pool_entropy(RAND_POOL *pool) ossl_rand_pool_entropy() argument
134 ossl_rand_pool_length(RAND_POOL *pool) ossl_rand_pool_length() argument
145 ossl_rand_pool_detach(RAND_POOL *pool) ossl_rand_pool_detach() argument
157 ossl_rand_pool_reattach(RAND_POOL *pool, unsigned char *buffer) ossl_rand_pool_reattach() argument
180 ossl_rand_pool_entropy_available(RAND_POOL *pool) ossl_rand_pool_entropy_available() argument
196 ossl_rand_pool_entropy_needed(RAND_POOL *pool) ossl_rand_pool_entropy_needed() argument
205 rand_pool_grow(RAND_POOL *pool, size_t len) rand_pool_grow() argument
246 ossl_rand_pool_bytes_needed(RAND_POOL *pool, unsigned int entropy_factor) ossl_rand_pool_bytes_needed() argument
291 ossl_rand_pool_bytes_remaining(RAND_POOL *pool) ossl_rand_pool_bytes_remaining() argument
305 ossl_rand_pool_add(RAND_POOL *pool, const unsigned char *buffer, size_t len, size_t entropy) ossl_rand_pool_add() argument
360 ossl_rand_pool_add_begin(RAND_POOL *pool, size_t len) ossl_rand_pool_add_begin() argument
399 ossl_rand_pool_add_end(RAND_POOL *pool, size_t len, size_t entropy) ossl_rand_pool_add_end() argument
[all...]
/third_party/skia/tests/
H A DSkBlockAllocatorTest.cpp19 static size_t ScratchBlockSize(SkSBlockAllocator<N>& pool) { in ScratchBlockSize() argument
20 return (size_t) pool->scratchBlockSize(); in ScratchBlockSize()
26 static int block_count(const SkSBlockAllocator<N>& pool) { in block_count() argument
28 for (const Block* b : pool->blocks()) { in block_count()
36 static Block* get_block(SkSBlockAllocator<N>& pool, int blockIndex) { in get_block() argument
39 for (Block* b: pool->blocks()) { in get_block()
55 static size_t total_size(SkSBlockAllocator<N>& pool) { in total_size() argument
56 return pool->totalSize() - BlockAllocatorTestAccess::ScratchBlockSize(pool); in total_size()
60 static size_t add_block(SkSBlockAllocator<N>& pool) { in add_block() argument
70 alloc_byte(SkSBlockAllocator<N>& pool) alloc_byte() argument
414 SkSBlockAllocator<256> pool; DEF_TEST() local
451 SkSBlockAllocator<256> pool; DEF_TEST() local
621 run_owning_block_test(skiatest::Reporter* r, SkBlockAllocator* pool) run_owning_block_test() argument
638 run_owning_block_tests(skiatest::Reporter* r, SkBlockAllocator* pool) run_owning_block_tests() argument
[all...]
/third_party/skia/third_party/externals/microhttpd/src/microhttpd/
H A Dmemorypool.c22 * @brief memory pool
47 * Handle for a memory pool. Pools are not reentrant and must not be
54 * Pointer to the pool's memory
59 * Size of the pool.
74 * #MHD_NO if pool was malloc'ed, #MHD_YES if mmapped (VirtualAlloc'ed for W32).
81 * Create a memory pool.
83 * @param max maximum size of the pool
89 struct MemoryPool *pool; in MHD_pool_create() local
91 pool = malloc (sizeof (struct MemoryPool)); in MHD_pool_create()
92 if (NULL == pool) in MHD_pool_create()
135 MHD_pool_destroy(struct MemoryPool *pool) MHD_pool_destroy() argument
165 MHD_pool_allocate(struct MemoryPool *pool, size_t size, int from_end) MHD_pool_allocate() argument
208 MHD_pool_reallocate(struct MemoryPool *pool, void *old, size_t old_size, size_t new_size) MHD_pool_reallocate() argument
262 MHD_pool_reset(struct MemoryPool *pool, void *keep, size_t size) MHD_pool_reset() argument
[all...]
/third_party/mesa3d/src/gallium/frontends/nine/
H A Dthreadpool.c44 struct threadpool *pool = data; in threadpool_worker() local
46 pthread_mutex_lock(&pool->m); in threadpool_worker()
48 while (!pool->shutdown) { in threadpool_worker()
52 while (!pool->workqueue && !pool->shutdown) in threadpool_worker()
53 pthread_cond_wait(&pool->new_work, &pool->m); in threadpool_worker()
55 if (pool->shutdown) in threadpool_worker()
62 task = pool->workqueue; in threadpool_worker()
63 pool in threadpool_worker()
90 struct threadpool *pool = calloc(1, sizeof(*pool)); _mesa_threadpool_create() local
109 _mesa_threadpool_destroy(struct NineSwapChain9 *swapchain, struct threadpool *pool) _mesa_threadpool_destroy() argument
142 _mesa_threadpool_queue_task(struct threadpool *pool, threadpool_task_func work, void *data) _mesa_threadpool_queue_task() argument
184 _mesa_threadpool_wait_for_task(struct threadpool *pool, struct threadpool_task **task_handle) _mesa_threadpool_wait_for_task() argument
[all...]
/third_party/ffmpeg/libavfilter/
H A Dframepool.c58 FFFramePool *pool; variable
62 pool = av_mallocz(sizeof(FFFramePool));
63 if (!pool)
66 pool->type = AVMEDIA_TYPE_VIDEO;
67 pool->width = width;
68 pool->height = height;
69 pool->format = format;
70 pool->align = align;
76 if (!pool->linesize[0]) {
77 ret = av_image_fill_linesizes(pool
110 ff_frame_pool_uninit(&pool); global() variable
121 FFFramePool *pool; global() variable
148 ff_frame_pool_uninit(&pool); global() variable
152 ff_frame_pool_get_video_config(FFFramePool *pool, int *width, int *height, enum AVPixelFormat *format, int *align) ff_frame_pool_get_video_config() argument
171 ff_frame_pool_get_audio_config(FFFramePool *pool, int *channels, int *nb_samples, enum AVSampleFormat *format, int *align) ff_frame_pool_get_audio_config() argument
190 ff_frame_pool_get(FFFramePool *pool) ff_frame_pool_get() argument
283 ff_frame_pool_uninit(FFFramePool **pool) ff_frame_pool_uninit() argument
[all...]
/third_party/mesa3d/src/panfrost/vulkan/
H A Dpanvk_mempool.c32 * In "owned" mode, a single parent owns the entire pool, and the pool owns all
34 * panvk_pool_get_bo_handles. Freeing occurs at the level of an entire pool.
35 * This is useful for streaming uploads, where the batch owns the pool.
37 * In "unowned" mode, the pool is freestanding. It does not track created BOs
44 panvk_pool_alloc_backing(struct panvk_pool *pool, size_t bo_sz) in panvk_pool_alloc_backing() argument
48 /* If there's a free BO in our BO pool, let's pick it. */ in panvk_pool_alloc_backing()
49 if (pool->bo_pool && bo_sz == pool->base.slab_size && in panvk_pool_alloc_backing()
50 util_dynarray_num_elements(&pool in panvk_pool_alloc_backing()
75 panvk_pool_alloc_aligned(struct panvk_pool *pool, size_t sz, unsigned alignment) panvk_pool_alloc_aligned() argument
103 panvk_pool_init(struct panvk_pool *pool, struct panfrost_device *dev, struct panvk_bo_pool *bo_pool, unsigned create_flags, size_t slab_size, const char *label, bool prealloc) panvk_pool_init() argument
120 panvk_pool_reset(struct panvk_pool *pool) panvk_pool_reset() argument
142 panvk_pool_cleanup(struct panvk_pool *pool) panvk_pool_cleanup() argument
150 panvk_pool_get_bo_handles(struct panvk_pool *pool, uint32_t *handles) panvk_pool_get_bo_handles() argument
[all...]
/third_party/mesa3d/src/util/
H A Dslab.c48 * - a pointer to the child pool to which this element belongs, or
62 /* Next page in the same child pool. */
83 * pool has been destroyed). Mark the element as freed and free the whole page
99 * Create a parent pool for the allocation of same-sized objects.
123 * Create a child pool linked to the given parent.
125 void slab_create_child(struct slab_child_pool *pool, in slab_create_child() argument
128 pool->parent = parent; in slab_create_child()
129 pool->pages = NULL; in slab_create_child()
130 pool->free = NULL; in slab_create_child()
131 pool in slab_create_child()
140 slab_destroy_child(struct slab_child_pool *pool) slab_destroy_child() argument
177 slab_add_new_page(struct slab_child_pool *pool) slab_add_new_page() argument
207 slab_alloc(struct slab_child_pool *pool) slab_alloc() argument
238 slab_zalloc(struct slab_child_pool *pool) slab_zalloc() argument
255 slab_free(struct slab_child_pool *pool, void *ptr) slab_free() argument
[all...]
/third_party/mesa3d/src/gallium/drivers/r600/
H A Dcompute_memory_pool.c48 static void compute_memory_shadow(struct compute_memory_pool* pool,
51 static void compute_memory_defrag(struct compute_memory_pool *pool,
55 static int compute_memory_promote_item(struct compute_memory_pool *pool,
59 static void compute_memory_move_item(struct compute_memory_pool *pool,
64 static void compute_memory_transfer(struct compute_memory_pool* pool,
70 * Creates a new pool.
75 struct compute_memory_pool* pool = (struct compute_memory_pool*) in compute_memory_pool_new() local
77 if (!pool) in compute_memory_pool_new()
82 pool->screen = rscreen; in compute_memory_pool_new()
83 pool in compute_memory_pool_new()
98 compute_memory_pool_init(struct compute_memory_pool * pool, unsigned initial_size_in_dw) compute_memory_pool_init() argument
113 compute_memory_pool_delete(struct compute_memory_pool* pool) compute_memory_pool_delete() argument
132 compute_memory_grow_defrag_pool(struct compute_memory_pool *pool, struct pipe_context *pipe, int new_size_in_dw) compute_memory_grow_defrag_pool() argument
194 compute_memory_shadow(struct compute_memory_pool* pool, struct pipe_context * pipe, int device_to_host) compute_memory_shadow() argument
215 compute_memory_finalize_pending(struct compute_memory_pool* pool, struct pipe_context * pipe) compute_memory_finalize_pending() argument
325 compute_memory_defrag(struct compute_memory_pool *pool, struct pipe_resource *src, struct pipe_resource *dst, struct pipe_context *pipe) compute_memory_defrag() argument
355 compute_memory_promote_item(struct compute_memory_pool *pool, struct compute_memory_item *item, struct pipe_context *pipe, int64_t start_in_dw) compute_memory_promote_item() argument
404 compute_memory_demote_item(struct compute_memory_pool *pool, struct compute_memory_item *item, struct pipe_context *pipe) compute_memory_demote_item() argument
466 compute_memory_move_item(struct compute_memory_pool *pool, struct pipe_resource *src, struct pipe_resource *dst, struct compute_memory_item *item, uint64_t new_start_in_dw, struct pipe_context *pipe) compute_memory_move_item() argument
560 compute_memory_free(struct compute_memory_pool* pool, int64_t id) compute_memory_free() argument
598 compute_memory_alloc( struct compute_memory_pool* pool, int64_t size_in_dw) compute_memory_alloc() argument
631 compute_memory_transfer( struct compute_memory_pool* pool, struct pipe_context * pipe, int device_to_host, struct compute_memory_item* chunk, void* data, int offset_in_chunk, int size) compute_memory_transfer() argument
[all...]
/third_party/vk-gl-cts/framework/delibs/depool/
H A DdeMemPool.c21 * \brief Memory pool management.
48 * Represent a page of memory allocate by a memory pool.
69 * \brief Memory pool.
71 * A pool of memory from which individual memory allocations can be made.
73 * but rather all of the memory allocated from a pool is freed when the pool
76 * The pools can be arranged into a hierarchy. If a pool with children is
78 * the pool itself.
82 * creating the root pool with the deMemPool_createFailingRoot() function.
92 deMemPool* firstChild; /*!< Pointer to first child pool i
171 deMemPool* pool; createPoolInternal() local
243 deMemPool* pool = createPoolInternal(DE_NULL); deMemPool_createRoot() local
283 deMemPool* pool; deMemPool_create() local
298 deMemPool_destroy(deMemPool* pool) deMemPool_destroy() argument
378 deMemPool_getNumChildren(const deMemPool* pool) deMemPool_getNumChildren() argument
390 deMemPool_getNumAllocatedBytes(const deMemPool* pool, deBool recurse) deMemPool_getNumAllocatedBytes() argument
408 deMemPool_getCapacity(const deMemPool* pool, deBool recurse) deMemPool_getCapacity() argument
426 deMemPool_allocInternal(deMemPool* pool, size_t numBytes, deUint32 alignBytes) deMemPool_allocInternal() argument
499 deMemPool_alloc(deMemPool* pool, size_t numBytes) deMemPool_alloc() argument
517 deMemPool_alignedAlloc(deMemPool* pool, size_t numBytes, deUint32 alignBytes) deMemPool_alignedAlloc() argument
536 deMemPool_memDup(deMemPool* pool, const void* ptr, size_t numBytes) deMemPool_memDup() argument
550 deMemPool_strDup(deMemPool* pool, const char* str) deMemPool_strDup() argument
566 deMemPool_strnDup(deMemPool* pool, const char* str, int maxLength) deMemPool_strnDup() argument
583 deMemPool_getMaxNumAllocatedBytes(const deMemPool* pool) deMemPool_getMaxNumAllocatedBytes() argument
589 deMemPool_getMaxCapacity(const deMemPool* pool) deMemPool_getMaxCapacity() argument
[all...]
/third_party/mesa3d/src/gallium/drivers/llvmpipe/
H A Dlp_cs_tpool.c27 * compute shader thread pool.
38 struct lp_cs_tpool *pool = data; in lp_cs_tpool_worker() local
42 mtx_lock(&pool->m); in lp_cs_tpool_worker()
44 while (!pool->shutdown) { in lp_cs_tpool_worker()
48 while (list_is_empty(&pool->workqueue) && !pool->shutdown) in lp_cs_tpool_worker()
49 cnd_wait(&pool->new_work, &pool->m); in lp_cs_tpool_worker()
51 if (pool->shutdown) in lp_cs_tpool_worker()
54 task = list_first_entry(&pool in lp_cs_tpool_worker()
89 struct lp_cs_tpool *pool = CALLOC_STRUCT(lp_cs_tpool); lp_cs_tpool_create() local
110 lp_cs_tpool_destroy(struct lp_cs_tpool *pool) lp_cs_tpool_destroy() argument
130 lp_cs_tpool_queue_task(struct lp_cs_tpool *pool, lp_cs_tpool_task_func work, void *data, int num_iters) lp_cs_tpool_queue_task() argument
169 lp_cs_tpool_wait_for_task(struct lp_cs_tpool *pool, struct lp_cs_tpool_task **task_handle) lp_cs_tpool_wait_for_task() argument
[all...]
/third_party/mesa3d/src/broadcom/vulkan/
H A Dv3dv_query.c120 struct v3dv_query_pool *pool, in kperfmon_create()
123 for (uint32_t i = 0; i < pool->perfmon.nperfmons; i++) { in kperfmon_create()
124 assert(i * DRM_V3D_MAX_PERF_COUNTERS < pool->perfmon.ncounters); in kperfmon_create()
127 .ncounters = MIN2(pool->perfmon.ncounters - in kperfmon_create()
132 &pool->perfmon.counters[i * DRM_V3D_MAX_PERF_COUNTERS], in kperfmon_create()
141 pool->queries[query].perf.kperfmon_ids[i] = req.id; in kperfmon_create()
147 struct v3dv_query_pool *pool, in kperfmon_destroy()
151 if (!pool->queries[query].perf.kperfmon_ids[0]) in kperfmon_destroy()
154 for (uint32_t i = 0; i < pool->perfmon.nperfmons; i++) { in kperfmon_destroy()
156 .id = pool in kperfmon_destroy()
119 kperfmon_create(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query) kperfmon_create() argument
146 kperfmon_destroy(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query) kperfmon_destroy() argument
183 struct v3dv_query_pool *pool = v3dv_CreateQueryPool() local
380 write_occlusion_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_occlusion_query_result() argument
401 write_timestamp_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_timestamp_query_result() argument
417 write_performance_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_performance_query_result() argument
473 write_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_query_result() argument
496 query_is_available(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_wait, bool *available) query_is_available() argument
525 get_query_result_count(struct v3dv_query_pool *pool) get_query_result_count() argument
539 v3dv_get_query_pool_results(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t first, uint32_t count, void *data, VkDeviceSize stride, VkQueryResultFlags flags) v3dv_get_query_pool_results() argument
667 v3dv_reset_query_pools(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t first, uint32_t count) v3dv_reset_query_pools() argument
[all...]
/third_party/mesa3d/src/gallium/drivers/panfrost/
H A Dpan_mempool.c35 * In "owned" mode, a single parent owns the entire pool, and the pool owns all
37 * panfrost_pool_get_bo_handles. Freeing occurs at the level of an entire pool.
38 * This is useful for streaming uploads, where the batch owns the pool.
40 * In "unowned" mode, the pool is freestanding. It does not track created BOs
47 panfrost_pool_alloc_backing(struct panfrost_pool *pool, size_t bo_sz) in panfrost_pool_alloc_backing() argument
55 struct panfrost_bo *bo = panfrost_bo_create(pool->base.dev, bo_sz, in panfrost_pool_alloc_backing()
56 pool->base.create_flags, pool->base.label); in panfrost_pool_alloc_backing()
58 if (pool in panfrost_pool_alloc_backing()
70 panfrost_pool_init(struct panfrost_pool *pool, void *memctx, struct panfrost_device *dev, unsigned create_flags, size_t slab_size, const char *label, bool prealloc, bool owned) panfrost_pool_init() argument
87 panfrost_pool_cleanup(struct panfrost_pool *pool) panfrost_pool_cleanup() argument
101 panfrost_pool_get_bo_handles(struct panfrost_pool *pool, uint32_t *handles) panfrost_pool_get_bo_handles() argument
124 panfrost_pool_alloc_aligned(struct panfrost_pool *pool, size_t sz, unsigned alignment) panfrost_pool_alloc_aligned() argument
[all...]
/third_party/libunwind/libunwind/src/mi/
H A Dmempool.c69 free_object (struct mempool *pool, void *object) in free_object() argument
73 obj->next = pool->free_list; in free_object()
74 pool->free_list = obj; in free_object()
75 ++pool->num_free; in free_object()
79 add_memory (struct mempool *pool, char *mem, size_t size, size_t obj_size) in add_memory() argument
84 free_object (pool, obj); in add_memory()
88 expand (struct mempool *pool) in expand() argument
93 size = pool->chunk_size; in expand()
97 size = UNW_ALIGN(pool->obj_size, pg_size); in expand()
102 size = pool in expand()
110 mempool_init(struct mempool *pool, size_t obj_size, size_t reserve) mempool_init() argument
137 mempool_alloc(struct mempool *pool) mempool_alloc() argument
158 mempool_free(struct mempool *pool, void *object) mempool_free() argument
[all...]
/third_party/mesa3d/src/gallium/drivers/r300/compiler/
H A Dmemory_pool.c38 void memory_pool_init(struct memory_pool * pool) in memory_pool_init() argument
40 memset(pool, 0, sizeof(struct memory_pool)); in memory_pool_init()
44 void memory_pool_destroy(struct memory_pool * pool) in memory_pool_destroy() argument
46 while(pool->blocks) { in memory_pool_destroy()
47 struct memory_block * block = pool->blocks; in memory_pool_destroy()
48 pool->blocks = block->next; in memory_pool_destroy()
53 static void refill_pool(struct memory_pool * pool) in refill_pool() argument
55 unsigned int blocksize = pool->total_allocated; in refill_pool()
62 newblock->next = pool->blocks; in refill_pool()
63 pool in refill_pool()
71 memory_pool_malloc(struct memory_pool * pool, unsigned int bytes) memory_pool_malloc() argument
[all...]
/third_party/ffmpeg/libavutil/
H A Dbuffer.c263 AVBufferPool *pool = av_mallocz(sizeof(*pool)); in av_buffer_pool_init2() local
264 if (!pool) in av_buffer_pool_init2()
267 ff_mutex_init(&pool->mutex, NULL); in av_buffer_pool_init2()
269 pool->size = size; in av_buffer_pool_init2()
270 pool->opaque = opaque; in av_buffer_pool_init2()
271 pool->alloc2 = alloc; in av_buffer_pool_init2()
272 pool->alloc = av_buffer_alloc; // fallback in av_buffer_pool_init2()
273 pool->pool_free = pool_free; in av_buffer_pool_init2()
275 atomic_init(&pool in av_buffer_pool_init2()
282 AVBufferPool *pool = av_mallocz(sizeof(*pool)); av_buffer_pool_init() local
296 buffer_pool_flush(AVBufferPool *pool) buffer_pool_flush() argument
311 buffer_pool_free(AVBufferPool *pool) buffer_pool_free() argument
324 AVBufferPool *pool; av_buffer_pool_uninit() local
342 AVBufferPool *pool = buf->pool; pool_release_buffer() local
358 pool_alloc_buffer(AVBufferPool *pool) pool_alloc_buffer() argument
387 av_buffer_pool_get(AVBufferPool *pool) av_buffer_pool_get() argument
[all...]
/third_party/mesa3d/src/intel/vulkan/
H A Danv_allocator.c85 * At the next level we can use various sub-allocators. The state pool is a
86 * pool of smaller, fixed size objects, which operates much like the block
87 * pool. It uses a free list for freeing objects, but when it runs out of
88 * space it just allocates a new block from the block pool. This allocator is
193 /* Assert that we only ever grow the pool */ in anv_state_table_expand_range()
206 /* Just leak the old map until we destroy the pool. We can't munmap it in anv_state_table_expand_range()
237 /* The block pool is always initialized to a nonzero size and this function in anv_state_table_grow()
298 /* We allocated the first block outside the pool so we have to grow in anv_state_table_add()
299 * the pool. pool_state->next acts a mutex: threads who try to in anv_state_table_add()
365 anv_block_pool_expand_range(struct anv_block_pool *pool,
369 anv_block_pool_init(struct anv_block_pool *pool, struct anv_device *device, const char *name, uint64_t start_address, uint32_t initial_size) anv_block_pool_init() argument
444 anv_block_pool_finish(struct anv_block_pool *pool) anv_block_pool_finish() argument
461 anv_block_pool_expand_range(struct anv_block_pool *pool, uint32_t center_bo_offset, uint32_t size) anv_block_pool_expand_range() argument
581 anv_block_pool_map(struct anv_block_pool *pool, int32_t offset, uint32_t size) anv_block_pool_map() argument
628 anv_block_pool_grow(struct anv_block_pool *pool, struct anv_block_state *state, uint32_t contiguous_size) anv_block_pool_grow() argument
748 anv_block_pool_alloc_new(struct anv_block_pool *pool, struct anv_block_state *pool_state, uint32_t block_size, uint32_t *padding) anv_block_pool_alloc_new() argument
801 anv_block_pool_alloc(struct anv_block_pool *pool, uint32_t block_size, uint32_t *padding) anv_block_pool_alloc() argument
821 anv_block_pool_alloc_back(struct anv_block_pool *pool, uint32_t block_size) anv_block_pool_alloc_back() argument
837 anv_state_pool_init(struct anv_state_pool *pool, struct anv_device *device, const char *name, uint64_t base_address, int32_t start_offset, uint32_t block_size) anv_state_pool_init() argument
879 anv_state_pool_finish(struct anv_state_pool *pool) anv_state_pool_finish() argument
887 anv_fixed_size_state_pool_alloc_new(struct anv_fixed_size_state_pool *pool, struct anv_block_pool *block_pool, uint32_t state_size, uint32_t block_size, uint32_t *padding) anv_fixed_size_state_pool_alloc_new() argument
951 anv_state_pool_return_blocks(struct anv_state_pool *pool, uint32_t chunk_offset, uint32_t count, uint32_t block_size) anv_state_pool_return_blocks() argument
991 anv_state_pool_return_chunk(struct anv_state_pool *pool, uint32_t chunk_offset, uint32_t chunk_size, uint32_t small_size) anv_state_pool_return_chunk() argument
1033 anv_state_pool_alloc_no_vg(struct anv_state_pool *pool, uint32_t size, uint32_t align) anv_state_pool_alloc_no_vg() argument
1125 anv_state_pool_alloc(struct anv_state_pool *pool, uint32_t size, uint32_t align) anv_state_pool_alloc() argument
1136 anv_state_pool_alloc_back(struct anv_state_pool *pool) anv_state_pool_alloc_back() argument
1168 anv_state_pool_free_no_vg(struct anv_state_pool *pool, struct anv_state state) anv_state_pool_free_no_vg() argument
1184 anv_state_pool_free(struct anv_state_pool *pool, struct anv_state state) anv_state_pool_free() argument
1293 anv_state_reserved_pool_init(struct anv_state_reserved_pool *pool, struct anv_state_pool *parent, uint32_t count, uint32_t size, uint32_t alignment) anv_state_reserved_pool_init() argument
1308 anv_state_reserved_pool_finish(struct anv_state_reserved_pool *pool) anv_state_reserved_pool_finish() argument
1320 anv_state_reserved_pool_alloc(struct anv_state_reserved_pool *pool) anv_state_reserved_pool_alloc() argument
1326 anv_state_reserved_pool_free(struct anv_state_reserved_pool *pool, struct anv_state state) anv_state_reserved_pool_free() argument
1333 anv_bo_pool_init(struct anv_bo_pool *pool, struct anv_device *device, const char *name) anv_bo_pool_init() argument
1348 anv_bo_pool_finish(struct anv_bo_pool *pool) anv_bo_pool_finish() argument
1367 anv_bo_pool_alloc(struct anv_bo_pool *pool, uint32_t size, struct anv_bo **bo_out) anv_bo_pool_alloc() argument
1405 anv_bo_pool_free(struct anv_bo_pool *pool, struct anv_bo *bo) anv_bo_pool_free() argument
1423 anv_scratch_pool_init(struct anv_device *device, struct anv_scratch_pool *pool) anv_scratch_pool_init() argument
1429 anv_scratch_pool_finish(struct anv_device *device, struct anv_scratch_pool *pool) anv_scratch_pool_finish() argument
1447 anv_scratch_pool_alloc(struct anv_device *device, struct anv_scratch_pool *pool, gl_shader_stage stage, unsigned per_thread_scratch) anv_scratch_pool_alloc() argument
1512 anv_scratch_pool_get_surf(struct anv_device *device, struct anv_scratch_pool *pool, unsigned per_thread_scratch) anv_scratch_pool_get_surf() argument
[all...]
H A DgenX_query.c54 anv_query_address(struct anv_query_pool *pool, uint32_t query) in anv_query_address() argument
57 .bo = pool->bo, in anv_query_address()
58 .offset = query * pool->stride, in anv_query_address()
82 /* Query pool slots are made up of some number of 64-bit values packed in CreateQueryPool()
94 VK_MULTIALLOC_DECL(&ma, struct anv_query_pool, pool, 1); in CreateQueryPool()
174 pool->type = pCreateInfo->queryType; in CreateQueryPool()
175 pool->pipeline_statistics = pipeline_statistics; in CreateQueryPool()
176 pool->stride = uint64s_per_slot * sizeof(uint64_t); in CreateQueryPool()
177 pool->slots = pCreateInfo->queryCount; in CreateQueryPool()
179 if (pool in CreateQueryPool()
301 khr_perf_query_availability_offset(struct anv_query_pool *pool, uint32_t query, uint32_t pass) khr_perf_query_availability_offset() argument
307 khr_perf_query_data_offset(struct anv_query_pool *pool, uint32_t query, uint32_t pass, bool end) khr_perf_query_data_offset() argument
314 khr_perf_query_availability_address(struct anv_query_pool *pool, uint32_t query, uint32_t pass) khr_perf_query_availability_address() argument
322 khr_perf_query_data_address(struct anv_query_pool *pool, uint32_t query, uint32_t pass, bool end) khr_perf_query_data_address() argument
378 intel_perf_query_data_offset(struct anv_query_pool *pool, bool end) intel_perf_query_data_offset() argument
397 query_slot(struct anv_query_pool *pool, uint32_t query) query_slot() argument
403 query_is_available(struct anv_query_pool *pool, uint32_t query) query_is_available() argument
421 wait_for_available(struct anv_device *device, struct anv_query_pool *pool, uint32_t query) wait_for_available() argument
659 emit_zero_queries(struct anv_cmd_buffer *cmd_buffer, struct mi_builder *b, struct anv_query_pool *pool, uint32_t first_index, uint32_t num_queries) emit_zero_queries() argument
863 emit_perf_intel_query(struct anv_cmd_buffer *cmd_buffer, struct anv_query_pool *pool, struct mi_builder *b, struct anv_address query_addr, bool end) emit_perf_intel_query() argument
[all...]
/third_party/ffmpeg/libavcodec/
H A Dget_buffer.c57 FramePool *pool = (FramePool*)data; in frame_pool_free() local
60 for (i = 0; i < FF_ARRAY_ELEMS(pool->pools); i++) in frame_pool_free()
61 av_buffer_pool_uninit(&pool->pools[i]); in frame_pool_free()
68 FramePool *pool = av_mallocz(sizeof(*pool)); in frame_pool_alloc() local
71 if (!pool) in frame_pool_alloc()
74 buf = av_buffer_create((uint8_t*)pool, sizeof(*pool), in frame_pool_alloc()
77 av_freep(&pool); in frame_pool_alloc()
86 FramePool *pool in update_frame_pool() local
203 FramePool *pool = (FramePool*)avctx->internal->pool->data; audio_get_buffer() local
248 FramePool *pool = (FramePool*)s->internal->pool->data; video_get_buffer() local
[all...]
/third_party/node/deps/undici/src/lib/
H A Dbalanced-pool.js14 } = require('./pool-base')
15 const Pool = require('./pool')
70 if (this[kClients].find((pool) => (
71 pool[kUrl].origin === upstreamOrigin &&
72 pool.closed !== true &&
73 pool.destroyed !== true
77 const pool = this[kFactory](upstreamOrigin, Object.assign({}, this[kOptions]))
79 this[kAddClient](pool)
80 pool.on('connect', () => {
81 pool[kWeigh
[all...]
/third_party/mesa3d/src/asahi/lib/
H A Dpool.c28 #include "pool.h"
32 * into the pool and copy there */
37 agx_pool_alloc_backing(struct agx_pool *pool, size_t bo_sz) in agx_pool_alloc_backing() argument
39 struct agx_bo *bo = agx_bo_create(pool->dev, bo_sz, in agx_pool_alloc_backing()
40 pool->create_flags); in agx_pool_alloc_backing()
42 util_dynarray_append(&pool->bos, struct agx_bo *, bo); in agx_pool_alloc_backing()
43 pool->transient_bo = bo; in agx_pool_alloc_backing()
44 pool->transient_offset = 0; in agx_pool_alloc_backing()
50 agx_pool_init(struct agx_pool *pool, struct agx_device *dev, in agx_pool_init() argument
53 memset(pool, in agx_pool_init()
63 agx_pool_cleanup(struct agx_pool *pool) agx_pool_cleanup() argument
73 agx_pool_get_bo_handles(struct agx_pool *pool, uint32_t *handles) agx_pool_get_bo_handles() argument
82 agx_pool_alloc_aligned(struct agx_pool *pool, size_t sz, unsigned alignment) agx_pool_alloc_aligned() argument
109 agx_pool_upload(struct agx_pool *pool, const void *data, size_t sz) agx_pool_upload() argument
115 agx_pool_upload_aligned(struct agx_pool *pool, const void *data, size_t sz, unsigned alignment) agx_pool_upload_aligned() argument
[all...]
/third_party/mesa3d/src/freedreno/vulkan/
H A Dtu_query.c97 * pool. */
98 #define query_iova(type, pool, query, field) \
99 pool->bo->iova + pool->stride * (query) + offsetof(type, field)
101 #define occlusion_query_iova(pool, query, field) \
102 query_iova(struct occlusion_query_slot, pool, query, field)
104 #define pipeline_stat_query_iova(pool, query, field) \
105 pool->bo->iova + pool->stride * (query) + \
108 #define primitive_query_iova(pool, quer
174 slot_address(struct tu_query_pool *pool, uint32_t query) slot_address() argument
255 struct tu_query_pool *pool = tu_CreateQueryPool() local
352 get_result_count(struct tu_query_pool *pool) get_result_count() argument
437 wait_for_available(struct tu_device *device, struct tu_query_pool *pool, uint32_t query) wait_for_available() argument
468 get_query_pool_results(struct tu_device *device, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount, size_t dataSize, void *pData, VkDeviceSize stride, VkQueryResultFlags flags) get_query_pool_results() argument
604 emit_copy_query_pool_results(struct tu_cmd_buffer *cmdbuf, struct tu_cs *cs, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount, struct tu_buffer *buffer, VkDeviceSize dstOffset, VkDeviceSize stride, VkQueryResultFlags flags) emit_copy_query_pool_results() argument
726 emit_reset_query_pool(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount) emit_reset_query_pool() argument
813 emit_begin_occlusion_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_occlusion_query() argument
845 emit_begin_stat_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_stat_query() argument
905 emit_begin_perf_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_perf_query() argument
983 emit_begin_xfb_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query, uint32_t stream_id) emit_begin_xfb_query() argument
996 emit_begin_prim_generated_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_prim_generated_query() argument
1096 emit_end_occlusion_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_occlusion_query() argument
1216 emit_end_stat_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_stat_query() argument
1279 emit_end_perf_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_perf_query() argument
1358 emit_end_xfb_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query, uint32_t stream_id) emit_end_xfb_query() argument
1407 emit_end_prim_generated_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_prim_generated_query() argument
1485 handle_multiview_queries(struct tu_cmd_buffer *cmd, struct tu_query_pool *pool, uint32_t query) handle_multiview_queries() argument
[all...]
/third_party/mesa3d/src/gallium/drivers/iris/
H A Diris_border_color.c36 * To work around this, we maintain a single "border color pool" BO
40 * wasting a lot of space in the pool.
71 struct iris_border_color_pool *pool) in iris_init_border_color_pool()
73 simple_mtx_init(&pool->lock, mtx_plain); in iris_init_border_color_pool()
75 pool->ht = _mesa_hash_table_create(NULL, color_hash, color_equals); in iris_init_border_color_pool()
77 pool->bo = iris_bo_alloc(bufmgr, "border colors", in iris_init_border_color_pool()
80 pool->map = iris_bo_map(NULL, pool->bo, MAP_WRITE); in iris_init_border_color_pool()
83 pool->insert_point = BC_ALIGNMENT; in iris_init_border_color_pool()
86 ASSERTED uint32_t black_offset = iris_upload_border_color(pool, in iris_init_border_color_pool()
70 iris_init_border_color_pool(struct iris_bufmgr *bufmgr, struct iris_border_color_pool *pool) iris_init_border_color_pool() argument
91 iris_destroy_border_color_pool(struct iris_border_color_pool *pool) iris_destroy_border_color_pool() argument
105 iris_upload_border_color(struct iris_border_color_pool *pool, union pipe_color_union *color) iris_upload_border_color() argument
[all...]
/third_party/mesa3d/src/vulkan/runtime/
H A Dvk_command_pool.c34 vk_command_pool_init(struct vk_command_pool *pool, in vk_command_pool_init() argument
39 memset(pool, 0, sizeof(*pool)); in vk_command_pool_init()
40 vk_object_base_init(device, &pool->base, in vk_command_pool_init()
43 pool->flags = pCreateInfo->flags; in vk_command_pool_init()
44 pool->queue_family_index = pCreateInfo->queueFamilyIndex; in vk_command_pool_init()
45 pool->alloc = pAllocator ? *pAllocator : device->alloc; in vk_command_pool_init()
46 list_inithead(&pool->command_buffers); in vk_command_pool_init()
52 vk_command_pool_finish(struct vk_command_pool *pool) in vk_command_pool_finish() argument
55 &pool in vk_command_pool_finish()
70 struct vk_command_pool *pool; vk_common_CreateCommandPool() local
[all...]
/third_party/mesa3d/src/virtio/vulkan/
H A Dvn_feedback.c139 vn_feedback_pool_grow_locked(struct vn_feedback_pool *pool) in vn_feedback_pool_grow_locked() argument
145 result = vn_feedback_buffer_create(pool->device, pool->size, pool->alloc, in vn_feedback_pool_grow_locked()
150 pool->used = 0; in vn_feedback_pool_grow_locked()
152 list_add(&feedback_buf->head, &pool->feedback_buffers); in vn_feedback_pool_grow_locked()
159 struct vn_feedback_pool *pool, in vn_feedback_pool_init()
163 simple_mtx_init(&pool->mutex, mtx_plain); in vn_feedback_pool_init()
165 pool->device = dev; in vn_feedback_pool_init()
166 pool in vn_feedback_pool_init()
158 vn_feedback_pool_init(struct vn_device *dev, struct vn_feedback_pool *pool, uint32_t size, const VkAllocationCallbacks *alloc) vn_feedback_pool_init() argument
177 vn_feedback_pool_fini(struct vn_feedback_pool *pool) vn_feedback_pool_fini() argument
191 vn_feedback_pool_alloc_locked(struct vn_feedback_pool *pool, uint32_t size, uint32_t *out_offset) vn_feedback_pool_alloc_locked() argument
214 vn_feedback_pool_alloc(struct vn_feedback_pool *pool, enum vn_feedback_type type) vn_feedback_pool_alloc() argument
260 vn_feedback_pool_free(struct vn_feedback_pool *pool, struct vn_feedback_slot *slot) vn_feedback_pool_free() argument
388 vn_feedback_fence_cmd_alloc(VkDevice dev_handle, struct vn_feedback_cmd_pool *pool, struct vn_feedback_slot *slot, VkCommandBuffer *out_cmd_handle) vn_feedback_fence_cmd_alloc() argument
423 vn_feedback_fence_cmd_free(VkDevice dev_handle, struct vn_feedback_cmd_pool *pool, VkCommandBuffer cmd_handle) vn_feedback_fence_cmd_free() argument
[all...]

Completed in 13 milliseconds

12345678910>>...24