/third_party/node/deps/openssl/openssl/crypto/rand/ |
H A D | rand_pool.c | 20 * Allocate memory and initialize a new random pool 25 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); in ossl_rand_pool_new() local 28 if (pool == NULL) { in ossl_rand_pool_new() 33 pool->min_len = min_len; in ossl_rand_pool_new() 34 pool->max_len = (max_len > RAND_POOL_MAX_LENGTH) ? in ossl_rand_pool_new() 36 pool->alloc_len = min_len < min_alloc_size ? min_alloc_size : min_len; in ossl_rand_pool_new() 37 if (pool->alloc_len > pool->max_len) in ossl_rand_pool_new() 38 pool in ossl_rand_pool_new() 68 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); ossl_rand_pool_attach() local 94 ossl_rand_pool_free(RAND_POOL *pool) ossl_rand_pool_free() argument 118 ossl_rand_pool_buffer(RAND_POOL *pool) ossl_rand_pool_buffer() argument 126 ossl_rand_pool_entropy(RAND_POOL *pool) ossl_rand_pool_entropy() argument 134 ossl_rand_pool_length(RAND_POOL *pool) ossl_rand_pool_length() argument 145 ossl_rand_pool_detach(RAND_POOL *pool) ossl_rand_pool_detach() argument 157 ossl_rand_pool_reattach(RAND_POOL *pool, unsigned char *buffer) ossl_rand_pool_reattach() argument 180 ossl_rand_pool_entropy_available(RAND_POOL *pool) ossl_rand_pool_entropy_available() argument 196 ossl_rand_pool_entropy_needed(RAND_POOL *pool) ossl_rand_pool_entropy_needed() argument 205 rand_pool_grow(RAND_POOL *pool, size_t len) rand_pool_grow() argument 246 ossl_rand_pool_bytes_needed(RAND_POOL *pool, unsigned int entropy_factor) ossl_rand_pool_bytes_needed() argument 291 ossl_rand_pool_bytes_remaining(RAND_POOL *pool) ossl_rand_pool_bytes_remaining() argument 305 ossl_rand_pool_add(RAND_POOL *pool, const unsigned char *buffer, size_t len, size_t entropy) ossl_rand_pool_add() argument 360 ossl_rand_pool_add_begin(RAND_POOL *pool, size_t len) ossl_rand_pool_add_begin() argument 399 ossl_rand_pool_add_end(RAND_POOL *pool, size_t len, size_t entropy) ossl_rand_pool_add_end() argument [all...] |
/third_party/openssl/crypto/rand/ |
H A D | rand_pool.c | 20 * Allocate memory and initialize a new random pool 25 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); in ossl_rand_pool_new() local 28 if (pool == NULL) { in ossl_rand_pool_new() 33 pool->min_len = min_len; in ossl_rand_pool_new() 34 pool->max_len = (max_len > RAND_POOL_MAX_LENGTH) ? in ossl_rand_pool_new() 36 pool->alloc_len = min_len < min_alloc_size ? min_alloc_size : min_len; in ossl_rand_pool_new() 37 if (pool->alloc_len > pool->max_len) in ossl_rand_pool_new() 38 pool in ossl_rand_pool_new() 68 RAND_POOL *pool = OPENSSL_zalloc(sizeof(*pool)); ossl_rand_pool_attach() local 94 ossl_rand_pool_free(RAND_POOL *pool) ossl_rand_pool_free() argument 118 ossl_rand_pool_buffer(RAND_POOL *pool) ossl_rand_pool_buffer() argument 126 ossl_rand_pool_entropy(RAND_POOL *pool) ossl_rand_pool_entropy() argument 134 ossl_rand_pool_length(RAND_POOL *pool) ossl_rand_pool_length() argument 145 ossl_rand_pool_detach(RAND_POOL *pool) ossl_rand_pool_detach() argument 157 ossl_rand_pool_reattach(RAND_POOL *pool, unsigned char *buffer) ossl_rand_pool_reattach() argument 180 ossl_rand_pool_entropy_available(RAND_POOL *pool) ossl_rand_pool_entropy_available() argument 196 ossl_rand_pool_entropy_needed(RAND_POOL *pool) ossl_rand_pool_entropy_needed() argument 205 rand_pool_grow(RAND_POOL *pool, size_t len) rand_pool_grow() argument 246 ossl_rand_pool_bytes_needed(RAND_POOL *pool, unsigned int entropy_factor) ossl_rand_pool_bytes_needed() argument 291 ossl_rand_pool_bytes_remaining(RAND_POOL *pool) ossl_rand_pool_bytes_remaining() argument 305 ossl_rand_pool_add(RAND_POOL *pool, const unsigned char *buffer, size_t len, size_t entropy) ossl_rand_pool_add() argument 360 ossl_rand_pool_add_begin(RAND_POOL *pool, size_t len) ossl_rand_pool_add_begin() argument 399 ossl_rand_pool_add_end(RAND_POOL *pool, size_t len, size_t entropy) ossl_rand_pool_add_end() argument [all...] |
/third_party/skia/tests/ |
H A D | SkBlockAllocatorTest.cpp | 19 static size_t ScratchBlockSize(SkSBlockAllocator<N>& pool) { in ScratchBlockSize() argument 20 return (size_t) pool->scratchBlockSize(); in ScratchBlockSize() 26 static int block_count(const SkSBlockAllocator<N>& pool) { in block_count() argument 28 for (const Block* b : pool->blocks()) { in block_count() 36 static Block* get_block(SkSBlockAllocator<N>& pool, int blockIndex) { in get_block() argument 39 for (Block* b: pool->blocks()) { in get_block() 55 static size_t total_size(SkSBlockAllocator<N>& pool) { in total_size() argument 56 return pool->totalSize() - BlockAllocatorTestAccess::ScratchBlockSize(pool); in total_size() 60 static size_t add_block(SkSBlockAllocator<N>& pool) { in add_block() argument 70 alloc_byte(SkSBlockAllocator<N>& pool) alloc_byte() argument 414 SkSBlockAllocator<256> pool; DEF_TEST() local 451 SkSBlockAllocator<256> pool; DEF_TEST() local 621 run_owning_block_test(skiatest::Reporter* r, SkBlockAllocator* pool) run_owning_block_test() argument 638 run_owning_block_tests(skiatest::Reporter* r, SkBlockAllocator* pool) run_owning_block_tests() argument [all...] |
/third_party/skia/third_party/externals/microhttpd/src/microhttpd/ |
H A D | memorypool.c | 22 * @brief memory pool 47 * Handle for a memory pool. Pools are not reentrant and must not be 54 * Pointer to the pool's memory 59 * Size of the pool. 74 * #MHD_NO if pool was malloc'ed, #MHD_YES if mmapped (VirtualAlloc'ed for W32). 81 * Create a memory pool. 83 * @param max maximum size of the pool 89 struct MemoryPool *pool; in MHD_pool_create() local 91 pool = malloc (sizeof (struct MemoryPool)); in MHD_pool_create() 92 if (NULL == pool) in MHD_pool_create() 135 MHD_pool_destroy(struct MemoryPool *pool) MHD_pool_destroy() argument 165 MHD_pool_allocate(struct MemoryPool *pool, size_t size, int from_end) MHD_pool_allocate() argument 208 MHD_pool_reallocate(struct MemoryPool *pool, void *old, size_t old_size, size_t new_size) MHD_pool_reallocate() argument 262 MHD_pool_reset(struct MemoryPool *pool, void *keep, size_t size) MHD_pool_reset() argument [all...] |
/third_party/mesa3d/src/gallium/frontends/nine/ |
H A D | threadpool.c | 44 struct threadpool *pool = data; in threadpool_worker() local 46 pthread_mutex_lock(&pool->m); in threadpool_worker() 48 while (!pool->shutdown) { in threadpool_worker() 52 while (!pool->workqueue && !pool->shutdown) in threadpool_worker() 53 pthread_cond_wait(&pool->new_work, &pool->m); in threadpool_worker() 55 if (pool->shutdown) in threadpool_worker() 62 task = pool->workqueue; in threadpool_worker() 63 pool in threadpool_worker() 90 struct threadpool *pool = calloc(1, sizeof(*pool)); _mesa_threadpool_create() local 109 _mesa_threadpool_destroy(struct NineSwapChain9 *swapchain, struct threadpool *pool) _mesa_threadpool_destroy() argument 142 _mesa_threadpool_queue_task(struct threadpool *pool, threadpool_task_func work, void *data) _mesa_threadpool_queue_task() argument 184 _mesa_threadpool_wait_for_task(struct threadpool *pool, struct threadpool_task **task_handle) _mesa_threadpool_wait_for_task() argument [all...] |
/third_party/ffmpeg/libavfilter/ |
H A D | framepool.c | 58 FFFramePool *pool; variable 62 pool = av_mallocz(sizeof(FFFramePool)); 63 if (!pool) 66 pool->type = AVMEDIA_TYPE_VIDEO; 67 pool->width = width; 68 pool->height = height; 69 pool->format = format; 70 pool->align = align; 76 if (!pool->linesize[0]) { 77 ret = av_image_fill_linesizes(pool 110 ff_frame_pool_uninit(&pool); global() variable 121 FFFramePool *pool; global() variable 148 ff_frame_pool_uninit(&pool); global() variable 152 ff_frame_pool_get_video_config(FFFramePool *pool, int *width, int *height, enum AVPixelFormat *format, int *align) ff_frame_pool_get_video_config() argument 171 ff_frame_pool_get_audio_config(FFFramePool *pool, int *channels, int *nb_samples, enum AVSampleFormat *format, int *align) ff_frame_pool_get_audio_config() argument 190 ff_frame_pool_get(FFFramePool *pool) ff_frame_pool_get() argument 283 ff_frame_pool_uninit(FFFramePool **pool) ff_frame_pool_uninit() argument [all...] |
/third_party/mesa3d/src/panfrost/vulkan/ |
H A D | panvk_mempool.c | 32 * In "owned" mode, a single parent owns the entire pool, and the pool owns all 34 * panvk_pool_get_bo_handles. Freeing occurs at the level of an entire pool. 35 * This is useful for streaming uploads, where the batch owns the pool. 37 * In "unowned" mode, the pool is freestanding. It does not track created BOs 44 panvk_pool_alloc_backing(struct panvk_pool *pool, size_t bo_sz) in panvk_pool_alloc_backing() argument 48 /* If there's a free BO in our BO pool, let's pick it. */ in panvk_pool_alloc_backing() 49 if (pool->bo_pool && bo_sz == pool->base.slab_size && in panvk_pool_alloc_backing() 50 util_dynarray_num_elements(&pool in panvk_pool_alloc_backing() 75 panvk_pool_alloc_aligned(struct panvk_pool *pool, size_t sz, unsigned alignment) panvk_pool_alloc_aligned() argument 103 panvk_pool_init(struct panvk_pool *pool, struct panfrost_device *dev, struct panvk_bo_pool *bo_pool, unsigned create_flags, size_t slab_size, const char *label, bool prealloc) panvk_pool_init() argument 120 panvk_pool_reset(struct panvk_pool *pool) panvk_pool_reset() argument 142 panvk_pool_cleanup(struct panvk_pool *pool) panvk_pool_cleanup() argument 150 panvk_pool_get_bo_handles(struct panvk_pool *pool, uint32_t *handles) panvk_pool_get_bo_handles() argument [all...] |
/third_party/mesa3d/src/util/ |
H A D | slab.c | 48 * - a pointer to the child pool to which this element belongs, or 62 /* Next page in the same child pool. */ 83 * pool has been destroyed). Mark the element as freed and free the whole page 99 * Create a parent pool for the allocation of same-sized objects. 123 * Create a child pool linked to the given parent. 125 void slab_create_child(struct slab_child_pool *pool, in slab_create_child() argument 128 pool->parent = parent; in slab_create_child() 129 pool->pages = NULL; in slab_create_child() 130 pool->free = NULL; in slab_create_child() 131 pool in slab_create_child() 140 slab_destroy_child(struct slab_child_pool *pool) slab_destroy_child() argument 177 slab_add_new_page(struct slab_child_pool *pool) slab_add_new_page() argument 207 slab_alloc(struct slab_child_pool *pool) slab_alloc() argument 238 slab_zalloc(struct slab_child_pool *pool) slab_zalloc() argument 255 slab_free(struct slab_child_pool *pool, void *ptr) slab_free() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/r600/ |
H A D | compute_memory_pool.c | 48 static void compute_memory_shadow(struct compute_memory_pool* pool, 51 static void compute_memory_defrag(struct compute_memory_pool *pool, 55 static int compute_memory_promote_item(struct compute_memory_pool *pool, 59 static void compute_memory_move_item(struct compute_memory_pool *pool, 64 static void compute_memory_transfer(struct compute_memory_pool* pool, 70 * Creates a new pool. 75 struct compute_memory_pool* pool = (struct compute_memory_pool*) in compute_memory_pool_new() local 77 if (!pool) in compute_memory_pool_new() 82 pool->screen = rscreen; in compute_memory_pool_new() 83 pool in compute_memory_pool_new() 98 compute_memory_pool_init(struct compute_memory_pool * pool, unsigned initial_size_in_dw) compute_memory_pool_init() argument 113 compute_memory_pool_delete(struct compute_memory_pool* pool) compute_memory_pool_delete() argument 132 compute_memory_grow_defrag_pool(struct compute_memory_pool *pool, struct pipe_context *pipe, int new_size_in_dw) compute_memory_grow_defrag_pool() argument 194 compute_memory_shadow(struct compute_memory_pool* pool, struct pipe_context * pipe, int device_to_host) compute_memory_shadow() argument 215 compute_memory_finalize_pending(struct compute_memory_pool* pool, struct pipe_context * pipe) compute_memory_finalize_pending() argument 325 compute_memory_defrag(struct compute_memory_pool *pool, struct pipe_resource *src, struct pipe_resource *dst, struct pipe_context *pipe) compute_memory_defrag() argument 355 compute_memory_promote_item(struct compute_memory_pool *pool, struct compute_memory_item *item, struct pipe_context *pipe, int64_t start_in_dw) compute_memory_promote_item() argument 404 compute_memory_demote_item(struct compute_memory_pool *pool, struct compute_memory_item *item, struct pipe_context *pipe) compute_memory_demote_item() argument 466 compute_memory_move_item(struct compute_memory_pool *pool, struct pipe_resource *src, struct pipe_resource *dst, struct compute_memory_item *item, uint64_t new_start_in_dw, struct pipe_context *pipe) compute_memory_move_item() argument 560 compute_memory_free(struct compute_memory_pool* pool, int64_t id) compute_memory_free() argument 598 compute_memory_alloc( struct compute_memory_pool* pool, int64_t size_in_dw) compute_memory_alloc() argument 631 compute_memory_transfer( struct compute_memory_pool* pool, struct pipe_context * pipe, int device_to_host, struct compute_memory_item* chunk, void* data, int offset_in_chunk, int size) compute_memory_transfer() argument [all...] |
/third_party/vk-gl-cts/framework/delibs/depool/ |
H A D | deMemPool.c | 21 * \brief Memory pool management. 48 * Represent a page of memory allocate by a memory pool. 69 * \brief Memory pool. 71 * A pool of memory from which individual memory allocations can be made. 73 * but rather all of the memory allocated from a pool is freed when the pool 76 * The pools can be arranged into a hierarchy. If a pool with children is 78 * the pool itself. 82 * creating the root pool with the deMemPool_createFailingRoot() function. 92 deMemPool* firstChild; /*!< Pointer to first child pool i 171 deMemPool* pool; createPoolInternal() local 243 deMemPool* pool = createPoolInternal(DE_NULL); deMemPool_createRoot() local 283 deMemPool* pool; deMemPool_create() local 298 deMemPool_destroy(deMemPool* pool) deMemPool_destroy() argument 378 deMemPool_getNumChildren(const deMemPool* pool) deMemPool_getNumChildren() argument 390 deMemPool_getNumAllocatedBytes(const deMemPool* pool, deBool recurse) deMemPool_getNumAllocatedBytes() argument 408 deMemPool_getCapacity(const deMemPool* pool, deBool recurse) deMemPool_getCapacity() argument 426 deMemPool_allocInternal(deMemPool* pool, size_t numBytes, deUint32 alignBytes) deMemPool_allocInternal() argument 499 deMemPool_alloc(deMemPool* pool, size_t numBytes) deMemPool_alloc() argument 517 deMemPool_alignedAlloc(deMemPool* pool, size_t numBytes, deUint32 alignBytes) deMemPool_alignedAlloc() argument 536 deMemPool_memDup(deMemPool* pool, const void* ptr, size_t numBytes) deMemPool_memDup() argument 550 deMemPool_strDup(deMemPool* pool, const char* str) deMemPool_strDup() argument 566 deMemPool_strnDup(deMemPool* pool, const char* str, int maxLength) deMemPool_strnDup() argument 583 deMemPool_getMaxNumAllocatedBytes(const deMemPool* pool) deMemPool_getMaxNumAllocatedBytes() argument 589 deMemPool_getMaxCapacity(const deMemPool* pool) deMemPool_getMaxCapacity() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/llvmpipe/ |
H A D | lp_cs_tpool.c | 27 * compute shader thread pool. 38 struct lp_cs_tpool *pool = data; in lp_cs_tpool_worker() local 42 mtx_lock(&pool->m); in lp_cs_tpool_worker() 44 while (!pool->shutdown) { in lp_cs_tpool_worker() 48 while (list_is_empty(&pool->workqueue) && !pool->shutdown) in lp_cs_tpool_worker() 49 cnd_wait(&pool->new_work, &pool->m); in lp_cs_tpool_worker() 51 if (pool->shutdown) in lp_cs_tpool_worker() 54 task = list_first_entry(&pool in lp_cs_tpool_worker() 89 struct lp_cs_tpool *pool = CALLOC_STRUCT(lp_cs_tpool); lp_cs_tpool_create() local 110 lp_cs_tpool_destroy(struct lp_cs_tpool *pool) lp_cs_tpool_destroy() argument 130 lp_cs_tpool_queue_task(struct lp_cs_tpool *pool, lp_cs_tpool_task_func work, void *data, int num_iters) lp_cs_tpool_queue_task() argument 169 lp_cs_tpool_wait_for_task(struct lp_cs_tpool *pool, struct lp_cs_tpool_task **task_handle) lp_cs_tpool_wait_for_task() argument [all...] |
/third_party/mesa3d/src/broadcom/vulkan/ |
H A D | v3dv_query.c | 120 struct v3dv_query_pool *pool, in kperfmon_create() 123 for (uint32_t i = 0; i < pool->perfmon.nperfmons; i++) { in kperfmon_create() 124 assert(i * DRM_V3D_MAX_PERF_COUNTERS < pool->perfmon.ncounters); in kperfmon_create() 127 .ncounters = MIN2(pool->perfmon.ncounters - in kperfmon_create() 132 &pool->perfmon.counters[i * DRM_V3D_MAX_PERF_COUNTERS], in kperfmon_create() 141 pool->queries[query].perf.kperfmon_ids[i] = req.id; in kperfmon_create() 147 struct v3dv_query_pool *pool, in kperfmon_destroy() 151 if (!pool->queries[query].perf.kperfmon_ids[0]) in kperfmon_destroy() 154 for (uint32_t i = 0; i < pool->perfmon.nperfmons; i++) { in kperfmon_destroy() 156 .id = pool in kperfmon_destroy() 119 kperfmon_create(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query) kperfmon_create() argument 146 kperfmon_destroy(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query) kperfmon_destroy() argument 183 struct v3dv_query_pool *pool = v3dv_CreateQueryPool() local 380 write_occlusion_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_occlusion_query_result() argument 401 write_timestamp_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_timestamp_query_result() argument 417 write_performance_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_performance_query_result() argument 473 write_query_result(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_64bit, void *data, uint32_t slot) write_query_result() argument 496 query_is_available(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t query, bool do_wait, bool *available) query_is_available() argument 525 get_query_result_count(struct v3dv_query_pool *pool) get_query_result_count() argument 539 v3dv_get_query_pool_results(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t first, uint32_t count, void *data, VkDeviceSize stride, VkQueryResultFlags flags) v3dv_get_query_pool_results() argument 667 v3dv_reset_query_pools(struct v3dv_device *device, struct v3dv_query_pool *pool, uint32_t first, uint32_t count) v3dv_reset_query_pools() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/panfrost/ |
H A D | pan_mempool.c | 35 * In "owned" mode, a single parent owns the entire pool, and the pool owns all 37 * panfrost_pool_get_bo_handles. Freeing occurs at the level of an entire pool. 38 * This is useful for streaming uploads, where the batch owns the pool. 40 * In "unowned" mode, the pool is freestanding. It does not track created BOs 47 panfrost_pool_alloc_backing(struct panfrost_pool *pool, size_t bo_sz) in panfrost_pool_alloc_backing() argument 55 struct panfrost_bo *bo = panfrost_bo_create(pool->base.dev, bo_sz, in panfrost_pool_alloc_backing() 56 pool->base.create_flags, pool->base.label); in panfrost_pool_alloc_backing() 58 if (pool in panfrost_pool_alloc_backing() 70 panfrost_pool_init(struct panfrost_pool *pool, void *memctx, struct panfrost_device *dev, unsigned create_flags, size_t slab_size, const char *label, bool prealloc, bool owned) panfrost_pool_init() argument 87 panfrost_pool_cleanup(struct panfrost_pool *pool) panfrost_pool_cleanup() argument 101 panfrost_pool_get_bo_handles(struct panfrost_pool *pool, uint32_t *handles) panfrost_pool_get_bo_handles() argument 124 panfrost_pool_alloc_aligned(struct panfrost_pool *pool, size_t sz, unsigned alignment) panfrost_pool_alloc_aligned() argument [all...] |
/third_party/libunwind/libunwind/src/mi/ |
H A D | mempool.c | 69 free_object (struct mempool *pool, void *object) in free_object() argument 73 obj->next = pool->free_list; in free_object() 74 pool->free_list = obj; in free_object() 75 ++pool->num_free; in free_object() 79 add_memory (struct mempool *pool, char *mem, size_t size, size_t obj_size) in add_memory() argument 84 free_object (pool, obj); in add_memory() 88 expand (struct mempool *pool) in expand() argument 93 size = pool->chunk_size; in expand() 97 size = UNW_ALIGN(pool->obj_size, pg_size); in expand() 102 size = pool in expand() 110 mempool_init(struct mempool *pool, size_t obj_size, size_t reserve) mempool_init() argument 137 mempool_alloc(struct mempool *pool) mempool_alloc() argument 158 mempool_free(struct mempool *pool, void *object) mempool_free() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/r300/compiler/ |
H A D | memory_pool.c | 38 void memory_pool_init(struct memory_pool * pool) in memory_pool_init() argument 40 memset(pool, 0, sizeof(struct memory_pool)); in memory_pool_init() 44 void memory_pool_destroy(struct memory_pool * pool) in memory_pool_destroy() argument 46 while(pool->blocks) { in memory_pool_destroy() 47 struct memory_block * block = pool->blocks; in memory_pool_destroy() 48 pool->blocks = block->next; in memory_pool_destroy() 53 static void refill_pool(struct memory_pool * pool) in refill_pool() argument 55 unsigned int blocksize = pool->total_allocated; in refill_pool() 62 newblock->next = pool->blocks; in refill_pool() 63 pool in refill_pool() 71 memory_pool_malloc(struct memory_pool * pool, unsigned int bytes) memory_pool_malloc() argument [all...] |
/third_party/ffmpeg/libavutil/ |
H A D | buffer.c | 263 AVBufferPool *pool = av_mallocz(sizeof(*pool)); in av_buffer_pool_init2() local 264 if (!pool) in av_buffer_pool_init2() 267 ff_mutex_init(&pool->mutex, NULL); in av_buffer_pool_init2() 269 pool->size = size; in av_buffer_pool_init2() 270 pool->opaque = opaque; in av_buffer_pool_init2() 271 pool->alloc2 = alloc; in av_buffer_pool_init2() 272 pool->alloc = av_buffer_alloc; // fallback in av_buffer_pool_init2() 273 pool->pool_free = pool_free; in av_buffer_pool_init2() 275 atomic_init(&pool in av_buffer_pool_init2() 282 AVBufferPool *pool = av_mallocz(sizeof(*pool)); av_buffer_pool_init() local 296 buffer_pool_flush(AVBufferPool *pool) buffer_pool_flush() argument 311 buffer_pool_free(AVBufferPool *pool) buffer_pool_free() argument 324 AVBufferPool *pool; av_buffer_pool_uninit() local 342 AVBufferPool *pool = buf->pool; pool_release_buffer() local 358 pool_alloc_buffer(AVBufferPool *pool) pool_alloc_buffer() argument 387 av_buffer_pool_get(AVBufferPool *pool) av_buffer_pool_get() argument [all...] |
/third_party/mesa3d/src/intel/vulkan/ |
H A D | anv_allocator.c | 85 * At the next level we can use various sub-allocators. The state pool is a 86 * pool of smaller, fixed size objects, which operates much like the block 87 * pool. It uses a free list for freeing objects, but when it runs out of 88 * space it just allocates a new block from the block pool. This allocator is 193 /* Assert that we only ever grow the pool */ in anv_state_table_expand_range() 206 /* Just leak the old map until we destroy the pool. We can't munmap it in anv_state_table_expand_range() 237 /* The block pool is always initialized to a nonzero size and this function in anv_state_table_grow() 298 /* We allocated the first block outside the pool so we have to grow in anv_state_table_add() 299 * the pool. pool_state->next acts a mutex: threads who try to in anv_state_table_add() 365 anv_block_pool_expand_range(struct anv_block_pool *pool, 369 anv_block_pool_init(struct anv_block_pool *pool, struct anv_device *device, const char *name, uint64_t start_address, uint32_t initial_size) anv_block_pool_init() argument 444 anv_block_pool_finish(struct anv_block_pool *pool) anv_block_pool_finish() argument 461 anv_block_pool_expand_range(struct anv_block_pool *pool, uint32_t center_bo_offset, uint32_t size) anv_block_pool_expand_range() argument 581 anv_block_pool_map(struct anv_block_pool *pool, int32_t offset, uint32_t size) anv_block_pool_map() argument 628 anv_block_pool_grow(struct anv_block_pool *pool, struct anv_block_state *state, uint32_t contiguous_size) anv_block_pool_grow() argument 748 anv_block_pool_alloc_new(struct anv_block_pool *pool, struct anv_block_state *pool_state, uint32_t block_size, uint32_t *padding) anv_block_pool_alloc_new() argument 801 anv_block_pool_alloc(struct anv_block_pool *pool, uint32_t block_size, uint32_t *padding) anv_block_pool_alloc() argument 821 anv_block_pool_alloc_back(struct anv_block_pool *pool, uint32_t block_size) anv_block_pool_alloc_back() argument 837 anv_state_pool_init(struct anv_state_pool *pool, struct anv_device *device, const char *name, uint64_t base_address, int32_t start_offset, uint32_t block_size) anv_state_pool_init() argument 879 anv_state_pool_finish(struct anv_state_pool *pool) anv_state_pool_finish() argument 887 anv_fixed_size_state_pool_alloc_new(struct anv_fixed_size_state_pool *pool, struct anv_block_pool *block_pool, uint32_t state_size, uint32_t block_size, uint32_t *padding) anv_fixed_size_state_pool_alloc_new() argument 951 anv_state_pool_return_blocks(struct anv_state_pool *pool, uint32_t chunk_offset, uint32_t count, uint32_t block_size) anv_state_pool_return_blocks() argument 991 anv_state_pool_return_chunk(struct anv_state_pool *pool, uint32_t chunk_offset, uint32_t chunk_size, uint32_t small_size) anv_state_pool_return_chunk() argument 1033 anv_state_pool_alloc_no_vg(struct anv_state_pool *pool, uint32_t size, uint32_t align) anv_state_pool_alloc_no_vg() argument 1125 anv_state_pool_alloc(struct anv_state_pool *pool, uint32_t size, uint32_t align) anv_state_pool_alloc() argument 1136 anv_state_pool_alloc_back(struct anv_state_pool *pool) anv_state_pool_alloc_back() argument 1168 anv_state_pool_free_no_vg(struct anv_state_pool *pool, struct anv_state state) anv_state_pool_free_no_vg() argument 1184 anv_state_pool_free(struct anv_state_pool *pool, struct anv_state state) anv_state_pool_free() argument 1293 anv_state_reserved_pool_init(struct anv_state_reserved_pool *pool, struct anv_state_pool *parent, uint32_t count, uint32_t size, uint32_t alignment) anv_state_reserved_pool_init() argument 1308 anv_state_reserved_pool_finish(struct anv_state_reserved_pool *pool) anv_state_reserved_pool_finish() argument 1320 anv_state_reserved_pool_alloc(struct anv_state_reserved_pool *pool) anv_state_reserved_pool_alloc() argument 1326 anv_state_reserved_pool_free(struct anv_state_reserved_pool *pool, struct anv_state state) anv_state_reserved_pool_free() argument 1333 anv_bo_pool_init(struct anv_bo_pool *pool, struct anv_device *device, const char *name) anv_bo_pool_init() argument 1348 anv_bo_pool_finish(struct anv_bo_pool *pool) anv_bo_pool_finish() argument 1367 anv_bo_pool_alloc(struct anv_bo_pool *pool, uint32_t size, struct anv_bo **bo_out) anv_bo_pool_alloc() argument 1405 anv_bo_pool_free(struct anv_bo_pool *pool, struct anv_bo *bo) anv_bo_pool_free() argument 1423 anv_scratch_pool_init(struct anv_device *device, struct anv_scratch_pool *pool) anv_scratch_pool_init() argument 1429 anv_scratch_pool_finish(struct anv_device *device, struct anv_scratch_pool *pool) anv_scratch_pool_finish() argument 1447 anv_scratch_pool_alloc(struct anv_device *device, struct anv_scratch_pool *pool, gl_shader_stage stage, unsigned per_thread_scratch) anv_scratch_pool_alloc() argument 1512 anv_scratch_pool_get_surf(struct anv_device *device, struct anv_scratch_pool *pool, unsigned per_thread_scratch) anv_scratch_pool_get_surf() argument [all...] |
H A D | genX_query.c | 54 anv_query_address(struct anv_query_pool *pool, uint32_t query) in anv_query_address() argument 57 .bo = pool->bo, in anv_query_address() 58 .offset = query * pool->stride, in anv_query_address() 82 /* Query pool slots are made up of some number of 64-bit values packed in CreateQueryPool() 94 VK_MULTIALLOC_DECL(&ma, struct anv_query_pool, pool, 1); in CreateQueryPool() 174 pool->type = pCreateInfo->queryType; in CreateQueryPool() 175 pool->pipeline_statistics = pipeline_statistics; in CreateQueryPool() 176 pool->stride = uint64s_per_slot * sizeof(uint64_t); in CreateQueryPool() 177 pool->slots = pCreateInfo->queryCount; in CreateQueryPool() 179 if (pool in CreateQueryPool() 301 khr_perf_query_availability_offset(struct anv_query_pool *pool, uint32_t query, uint32_t pass) khr_perf_query_availability_offset() argument 307 khr_perf_query_data_offset(struct anv_query_pool *pool, uint32_t query, uint32_t pass, bool end) khr_perf_query_data_offset() argument 314 khr_perf_query_availability_address(struct anv_query_pool *pool, uint32_t query, uint32_t pass) khr_perf_query_availability_address() argument 322 khr_perf_query_data_address(struct anv_query_pool *pool, uint32_t query, uint32_t pass, bool end) khr_perf_query_data_address() argument 378 intel_perf_query_data_offset(struct anv_query_pool *pool, bool end) intel_perf_query_data_offset() argument 397 query_slot(struct anv_query_pool *pool, uint32_t query) query_slot() argument 403 query_is_available(struct anv_query_pool *pool, uint32_t query) query_is_available() argument 421 wait_for_available(struct anv_device *device, struct anv_query_pool *pool, uint32_t query) wait_for_available() argument 659 emit_zero_queries(struct anv_cmd_buffer *cmd_buffer, struct mi_builder *b, struct anv_query_pool *pool, uint32_t first_index, uint32_t num_queries) emit_zero_queries() argument 863 emit_perf_intel_query(struct anv_cmd_buffer *cmd_buffer, struct anv_query_pool *pool, struct mi_builder *b, struct anv_address query_addr, bool end) emit_perf_intel_query() argument [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | get_buffer.c | 57 FramePool *pool = (FramePool*)data; in frame_pool_free() local 60 for (i = 0; i < FF_ARRAY_ELEMS(pool->pools); i++) in frame_pool_free() 61 av_buffer_pool_uninit(&pool->pools[i]); in frame_pool_free() 68 FramePool *pool = av_mallocz(sizeof(*pool)); in frame_pool_alloc() local 71 if (!pool) in frame_pool_alloc() 74 buf = av_buffer_create((uint8_t*)pool, sizeof(*pool), in frame_pool_alloc() 77 av_freep(&pool); in frame_pool_alloc() 86 FramePool *pool in update_frame_pool() local 203 FramePool *pool = (FramePool*)avctx->internal->pool->data; audio_get_buffer() local 248 FramePool *pool = (FramePool*)s->internal->pool->data; video_get_buffer() local [all...] |
/third_party/node/deps/undici/src/lib/ |
H A D | balanced-pool.js | 14 } = require('./pool-base') 15 const Pool = require('./pool') 70 if (this[kClients].find((pool) => ( 71 pool[kUrl].origin === upstreamOrigin && 72 pool.closed !== true && 73 pool.destroyed !== true 77 const pool = this[kFactory](upstreamOrigin, Object.assign({}, this[kOptions])) 79 this[kAddClient](pool) 80 pool.on('connect', () => { 81 pool[kWeigh [all...] |
/third_party/mesa3d/src/asahi/lib/ |
H A D | pool.c | 28 #include "pool.h" 32 * into the pool and copy there */ 37 agx_pool_alloc_backing(struct agx_pool *pool, size_t bo_sz) in agx_pool_alloc_backing() argument 39 struct agx_bo *bo = agx_bo_create(pool->dev, bo_sz, in agx_pool_alloc_backing() 40 pool->create_flags); in agx_pool_alloc_backing() 42 util_dynarray_append(&pool->bos, struct agx_bo *, bo); in agx_pool_alloc_backing() 43 pool->transient_bo = bo; in agx_pool_alloc_backing() 44 pool->transient_offset = 0; in agx_pool_alloc_backing() 50 agx_pool_init(struct agx_pool *pool, struct agx_device *dev, in agx_pool_init() argument 53 memset(pool, in agx_pool_init() 63 agx_pool_cleanup(struct agx_pool *pool) agx_pool_cleanup() argument 73 agx_pool_get_bo_handles(struct agx_pool *pool, uint32_t *handles) agx_pool_get_bo_handles() argument 82 agx_pool_alloc_aligned(struct agx_pool *pool, size_t sz, unsigned alignment) agx_pool_alloc_aligned() argument 109 agx_pool_upload(struct agx_pool *pool, const void *data, size_t sz) agx_pool_upload() argument 115 agx_pool_upload_aligned(struct agx_pool *pool, const void *data, size_t sz, unsigned alignment) agx_pool_upload_aligned() argument [all...] |
/third_party/mesa3d/src/freedreno/vulkan/ |
H A D | tu_query.c | 97 * pool. */ 98 #define query_iova(type, pool, query, field) \ 99 pool->bo->iova + pool->stride * (query) + offsetof(type, field) 101 #define occlusion_query_iova(pool, query, field) \ 102 query_iova(struct occlusion_query_slot, pool, query, field) 104 #define pipeline_stat_query_iova(pool, query, field) \ 105 pool->bo->iova + pool->stride * (query) + \ 108 #define primitive_query_iova(pool, quer 174 slot_address(struct tu_query_pool *pool, uint32_t query) slot_address() argument 255 struct tu_query_pool *pool = tu_CreateQueryPool() local 352 get_result_count(struct tu_query_pool *pool) get_result_count() argument 437 wait_for_available(struct tu_device *device, struct tu_query_pool *pool, uint32_t query) wait_for_available() argument 468 get_query_pool_results(struct tu_device *device, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount, size_t dataSize, void *pData, VkDeviceSize stride, VkQueryResultFlags flags) get_query_pool_results() argument 604 emit_copy_query_pool_results(struct tu_cmd_buffer *cmdbuf, struct tu_cs *cs, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount, struct tu_buffer *buffer, VkDeviceSize dstOffset, VkDeviceSize stride, VkQueryResultFlags flags) emit_copy_query_pool_results() argument 726 emit_reset_query_pool(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t firstQuery, uint32_t queryCount) emit_reset_query_pool() argument 813 emit_begin_occlusion_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_occlusion_query() argument 845 emit_begin_stat_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_stat_query() argument 905 emit_begin_perf_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_perf_query() argument 983 emit_begin_xfb_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query, uint32_t stream_id) emit_begin_xfb_query() argument 996 emit_begin_prim_generated_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_begin_prim_generated_query() argument 1096 emit_end_occlusion_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_occlusion_query() argument 1216 emit_end_stat_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_stat_query() argument 1279 emit_end_perf_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_perf_query() argument 1358 emit_end_xfb_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query, uint32_t stream_id) emit_end_xfb_query() argument 1407 emit_end_prim_generated_query(struct tu_cmd_buffer *cmdbuf, struct tu_query_pool *pool, uint32_t query) emit_end_prim_generated_query() argument 1485 handle_multiview_queries(struct tu_cmd_buffer *cmd, struct tu_query_pool *pool, uint32_t query) handle_multiview_queries() argument [all...] |
/third_party/mesa3d/src/gallium/drivers/iris/ |
H A D | iris_border_color.c | 36 * To work around this, we maintain a single "border color pool" BO 40 * wasting a lot of space in the pool. 71 struct iris_border_color_pool *pool) in iris_init_border_color_pool() 73 simple_mtx_init(&pool->lock, mtx_plain); in iris_init_border_color_pool() 75 pool->ht = _mesa_hash_table_create(NULL, color_hash, color_equals); in iris_init_border_color_pool() 77 pool->bo = iris_bo_alloc(bufmgr, "border colors", in iris_init_border_color_pool() 80 pool->map = iris_bo_map(NULL, pool->bo, MAP_WRITE); in iris_init_border_color_pool() 83 pool->insert_point = BC_ALIGNMENT; in iris_init_border_color_pool() 86 ASSERTED uint32_t black_offset = iris_upload_border_color(pool, in iris_init_border_color_pool() 70 iris_init_border_color_pool(struct iris_bufmgr *bufmgr, struct iris_border_color_pool *pool) iris_init_border_color_pool() argument 91 iris_destroy_border_color_pool(struct iris_border_color_pool *pool) iris_destroy_border_color_pool() argument 105 iris_upload_border_color(struct iris_border_color_pool *pool, union pipe_color_union *color) iris_upload_border_color() argument [all...] |
/third_party/mesa3d/src/vulkan/runtime/ |
H A D | vk_command_pool.c | 34 vk_command_pool_init(struct vk_command_pool *pool, in vk_command_pool_init() argument 39 memset(pool, 0, sizeof(*pool)); in vk_command_pool_init() 40 vk_object_base_init(device, &pool->base, in vk_command_pool_init() 43 pool->flags = pCreateInfo->flags; in vk_command_pool_init() 44 pool->queue_family_index = pCreateInfo->queueFamilyIndex; in vk_command_pool_init() 45 pool->alloc = pAllocator ? *pAllocator : device->alloc; in vk_command_pool_init() 46 list_inithead(&pool->command_buffers); in vk_command_pool_init() 52 vk_command_pool_finish(struct vk_command_pool *pool) in vk_command_pool_finish() argument 55 &pool in vk_command_pool_finish() 70 struct vk_command_pool *pool; vk_common_CreateCommandPool() local [all...] |
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_feedback.c | 139 vn_feedback_pool_grow_locked(struct vn_feedback_pool *pool) in vn_feedback_pool_grow_locked() argument 145 result = vn_feedback_buffer_create(pool->device, pool->size, pool->alloc, in vn_feedback_pool_grow_locked() 150 pool->used = 0; in vn_feedback_pool_grow_locked() 152 list_add(&feedback_buf->head, &pool->feedback_buffers); in vn_feedback_pool_grow_locked() 159 struct vn_feedback_pool *pool, in vn_feedback_pool_init() 163 simple_mtx_init(&pool->mutex, mtx_plain); in vn_feedback_pool_init() 165 pool->device = dev; in vn_feedback_pool_init() 166 pool in vn_feedback_pool_init() 158 vn_feedback_pool_init(struct vn_device *dev, struct vn_feedback_pool *pool, uint32_t size, const VkAllocationCallbacks *alloc) vn_feedback_pool_init() argument 177 vn_feedback_pool_fini(struct vn_feedback_pool *pool) vn_feedback_pool_fini() argument 191 vn_feedback_pool_alloc_locked(struct vn_feedback_pool *pool, uint32_t size, uint32_t *out_offset) vn_feedback_pool_alloc_locked() argument 214 vn_feedback_pool_alloc(struct vn_feedback_pool *pool, enum vn_feedback_type type) vn_feedback_pool_alloc() argument 260 vn_feedback_pool_free(struct vn_feedback_pool *pool, struct vn_feedback_slot *slot) vn_feedback_pool_free() argument 388 vn_feedback_fence_cmd_alloc(VkDevice dev_handle, struct vn_feedback_cmd_pool *pool, struct vn_feedback_slot *slot, VkCommandBuffer *out_cmd_handle) vn_feedback_fence_cmd_alloc() argument 423 vn_feedback_fence_cmd_free(VkDevice dev_handle, struct vn_feedback_cmd_pool *pool, VkCommandBuffer cmd_handle) vn_feedback_fence_cmd_free() argument [all...] |