/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
H A D | id_table.c | 44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument 49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc() 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc() 52 if (obj >= alloc->max) in c4iw_id_alloc() 53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc() 55 if (obj < alloc->max) { in c4iw_id_alloc() 56 if (alloc in c4iw_id_alloc() 71 c4iw_id_free(struct c4iw_id_table *alloc, u32 obj) c4iw_id_free() argument 82 c4iw_id_table_alloc(struct c4iw_id_table *alloc, u32 start, u32 num, u32 reserved, u32 flags) c4iw_id_table_alloc() argument 108 c4iw_id_table_free(struct c4iw_id_table *alloc) c4iw_id_table_free() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/ |
H A D | id_table.c | 44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument 49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc() 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc() 52 if (obj >= alloc->max) in c4iw_id_alloc() 53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc() 55 if (obj < alloc->max) { in c4iw_id_alloc() 56 if (alloc in c4iw_id_alloc() 71 c4iw_id_free(struct c4iw_id_table *alloc, u32 obj) c4iw_id_free() argument 82 c4iw_id_table_alloc(struct c4iw_id_table *alloc, u32 start, u32 num, u32 reserved, u32 flags) c4iw_id_table_alloc() argument 103 c4iw_id_table_free(struct c4iw_id_table *alloc) c4iw_id_table_free() argument [all...] |
/kernel/linux/linux-6.6/drivers/android/ |
H A D | binder_alloc.c | 60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument 63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size() 64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument 71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer() 79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer() 83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer() 90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer() 98 rb_insert_color(&new_buffer->rb_node, &alloc in binder_insert_free_buffer() 101 binder_insert_allocated_buffer_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer) binder_insert_allocated_buffer_locked() argument 126 binder_alloc_prepare_to_free_locked( struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free_locked() argument 170 binder_alloc_prepare_to_free(struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free() argument 181 binder_update_page_range(struct binder_alloc *alloc, int allocate, void __user *start, void __user *end) binder_update_page_range() argument 312 binder_alloc_set_vma(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_set_vma() argument 319 binder_alloc_get_vma( struct binder_alloc *alloc) binder_alloc_get_vma() argument 326 debug_low_async_space_locked(struct binder_alloc *alloc, int pid) debug_low_async_space_locked() argument 368 binder_alloc_new_buf_locked( struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf_locked() argument 561 binder_alloc_new_buf(struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf() argument 588 binder_delete_free_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_delete_free_buffer() argument 637 binder_free_buf_locked(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_free_buf_locked() argument 701 binder_alloc_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_free_buf() argument 734 binder_alloc_mmap_handler(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_mmap_handler() argument 804 binder_alloc_deferred_release(struct binder_alloc *alloc) binder_alloc_deferred_release() argument 888 binder_alloc_print_allocated(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_allocated() argument 905 binder_alloc_print_pages(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_pages() argument 941 binder_alloc_get_allocated_count(struct binder_alloc *alloc) binder_alloc_get_allocated_count() argument 962 binder_alloc_vma_close(struct binder_alloc *alloc) binder_alloc_vma_close() argument 986 struct binder_alloc *alloc; __must_hold() local 1071 binder_alloc_init(struct binder_alloc *alloc) binder_alloc_init() argument 1117 check_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t offset, size_t bytes) check_buffer() argument 1149 binder_alloc_get_page(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, pgoff_t *pgoffp) binder_alloc_get_page() argument 1172 binder_alloc_clear_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_clear_buf() argument 1205 binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, const void __user *from, size_t bytes) binder_alloc_copy_user_to_buffer() argument 1236 binder_alloc_do_buffer_copy(struct binder_alloc *alloc, bool to_buffer, struct binder_buffer *buffer, binder_size_t buffer_offset, void *ptr, size_t bytes) binder_alloc_do_buffer_copy() argument 1267 binder_alloc_copy_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, void *src, size_t bytes) binder_alloc_copy_to_buffer() argument 1277 binder_alloc_copy_from_buffer(struct binder_alloc *alloc, void *dest, struct binder_buffer *buffer, binder_size_t buffer_offset, size_t bytes) binder_alloc_copy_from_buffer() argument [all...] |
H A D | binder_alloc_selftest.c | 82 pr_err("alloc sizes: "); in pr_err_size_seq() 92 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument 103 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated() 104 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated() 105 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated() 106 pr_err("expect alloc but is %s at page index %d\n", in check_buffer_pages_allocated() 107 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated() 115 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument 122 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0, 0); in binder_selftest_alloc_buf() 124 !check_buffer_pages_allocated(alloc, buffer in binder_selftest_alloc_buf() 132 binder_selftest_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffers[], size_t *sizes, int *seq, size_t end) binder_selftest_free_buf() argument 156 binder_selftest_free_page(struct binder_alloc *alloc) binder_selftest_free_page() argument 176 binder_selftest_alloc_free(struct binder_alloc *alloc, size_t *sizes, int *seq, size_t end) binder_selftest_alloc_free() argument 205 binder_selftest_free_seq(struct binder_alloc *alloc, size_t *sizes, int *seq, int index, size_t end) binder_selftest_free_seq() argument 223 binder_selftest_alloc_size(struct binder_alloc *alloc, size_t *end_offset) binder_selftest_alloc_size() argument 249 binder_selftest_alloc_offset(struct binder_alloc *alloc, size_t *end_offset, int index) binder_selftest_alloc_offset() argument 283 binder_selftest_alloc(struct binder_alloc *alloc) binder_selftest_alloc() argument [all...] |
H A D | binder_alloc.h | 23 * @entry: entry alloc->buffers 67 * @alloc: binder_alloc for a proc 72 struct binder_alloc *alloc; member 117 void binder_selftest_alloc(struct binder_alloc *alloc); 119 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument 124 extern struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc, 130 extern void binder_alloc_init(struct binder_alloc *alloc); 133 extern void binder_alloc_vma_close(struct binder_alloc *alloc); 135 binder_alloc_prepare_to_free(struct binder_alloc *alloc, 137 extern void binder_alloc_free_buf(struct binder_alloc *alloc, 155 binder_alloc_get_free_async_space(struct binder_alloc *alloc) binder_alloc_get_free_async_space() argument [all...] |
/kernel/linux/linux-5.10/drivers/android/ |
H A D | binder_alloc.c | 60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument 63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size() 64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument 71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer() 79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer() 83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer() 90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer() 98 rb_insert_color(&new_buffer->rb_node, &alloc in binder_insert_free_buffer() 101 binder_insert_allocated_buffer_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer) binder_insert_allocated_buffer_locked() argument 126 binder_alloc_prepare_to_free_locked( struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free_locked() argument 170 binder_alloc_prepare_to_free(struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free() argument 181 binder_update_page_range(struct binder_alloc *alloc, int allocate, void __user *start, void __user *end) binder_update_page_range() argument 313 binder_alloc_set_vma(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_set_vma() argument 328 binder_alloc_get_vma( struct binder_alloc *alloc) binder_alloc_get_vma() argument 341 debug_low_async_space_locked(struct binder_alloc *alloc, int pid) debug_low_async_space_locked() argument 377 binder_alloc_new_buf_locked( struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf_locked() argument 566 binder_alloc_new_buf(struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf() argument 593 binder_delete_free_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_delete_free_buffer() argument 642 binder_free_buf_locked(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_free_buf_locked() argument 706 binder_alloc_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_free_buf() argument 739 binder_alloc_mmap_handler(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_mmap_handler() argument 801 binder_alloc_deferred_release(struct binder_alloc *alloc) binder_alloc_deferred_release() argument 885 binder_alloc_print_allocated(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_allocated() argument 902 binder_alloc_print_pages(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_pages() argument 938 binder_alloc_get_allocated_count(struct binder_alloc *alloc) binder_alloc_get_allocated_count() argument 959 binder_alloc_vma_close(struct binder_alloc *alloc) binder_alloc_vma_close() argument 983 struct binder_alloc *alloc; __must_hold() local 1072 binder_alloc_init(struct binder_alloc *alloc) binder_alloc_init() argument 1116 check_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t offset, size_t bytes) check_buffer() argument 1148 binder_alloc_get_page(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, pgoff_t *pgoffp) binder_alloc_get_page() argument 1171 binder_alloc_clear_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_clear_buf() argument 1207 binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, const void __user *from, size_t bytes) binder_alloc_copy_user_to_buffer() argument 1238 binder_alloc_do_buffer_copy(struct binder_alloc *alloc, bool to_buffer, struct binder_buffer *buffer, binder_size_t buffer_offset, void *ptr, size_t bytes) binder_alloc_do_buffer_copy() argument 1278 binder_alloc_copy_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, void *src, size_t bytes) binder_alloc_copy_to_buffer() argument 1288 binder_alloc_copy_from_buffer(struct binder_alloc *alloc, void *dest, struct binder_buffer *buffer, binder_size_t buffer_offset, size_t bytes) binder_alloc_copy_from_buffer() argument [all...] |
H A D | binder_alloc_selftest.c | 82 pr_err("alloc sizes: "); in pr_err_size_seq() 92 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument 103 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated() 104 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated() 105 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated() 106 pr_err("expect alloc but is %s at page index %d\n", in check_buffer_pages_allocated() 107 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated() 115 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument 122 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0, 0); in binder_selftest_alloc_buf() 124 !check_buffer_pages_allocated(alloc, buffer in binder_selftest_alloc_buf() 132 binder_selftest_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffers[], size_t *sizes, int *seq, size_t end) binder_selftest_free_buf() argument 156 binder_selftest_free_page(struct binder_alloc *alloc) binder_selftest_free_page() argument 176 binder_selftest_alloc_free(struct binder_alloc *alloc, size_t *sizes, int *seq, size_t end) binder_selftest_alloc_free() argument 205 binder_selftest_free_seq(struct binder_alloc *alloc, size_t *sizes, int *seq, int index, size_t end) binder_selftest_free_seq() argument 223 binder_selftest_alloc_size(struct binder_alloc *alloc, size_t *end_offset) binder_selftest_alloc_size() argument 249 binder_selftest_alloc_offset(struct binder_alloc *alloc, size_t *end_offset, int index) binder_selftest_alloc_offset() argument 283 binder_selftest_alloc(struct binder_alloc *alloc) binder_selftest_alloc() argument [all...] |
H A D | binder_alloc.h | 23 * @entry: entry alloc->buffers 64 * @alloc: binder_alloc for a proc 69 struct binder_alloc *alloc; member 113 void binder_selftest_alloc(struct binder_alloc *alloc); 115 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument 120 extern struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc, 126 extern void binder_alloc_init(struct binder_alloc *alloc); 129 extern void binder_alloc_vma_close(struct binder_alloc *alloc); 131 binder_alloc_prepare_to_free(struct binder_alloc *alloc, 133 extern void binder_alloc_free_buf(struct binder_alloc *alloc, 151 binder_alloc_get_free_async_space(struct binder_alloc *alloc) binder_alloc_get_free_async_space() argument [all...] |
/third_party/skia/src/gpu/vk/ |
H A D | GrVkMemory.cpp | 48 GrVkAlloc* alloc, 51 GrVkAlloc* alloc) { 71 allocator->getAllocInfo(memory, alloc); 74 alloc->fBytes = size; 80 GR_VK_CALL_RESULT(gpu, err, BindBufferMemory(gpu->device(), buffer, alloc->fMemory, 81 alloc->fOffset)); 83 FreeBufferMemory(gpu, *alloc); 93 GrVkAlloc* alloc) { 137 alloc->fMemory = memory; 138 alloc 44 AllocAndBindBufferMemory(GrVkGpu* gpu, VkBuffer buffer, BufferUsage usage, GrVkAlloc* alloc, size_t size) global() argument 90 ImportAndBindBufferMemory(GrVkGpu* gpu, OH_NativeBuffer *nativeBuffer, VkBuffer buffer, GrVkAlloc* alloc) global() argument 146 FreeBufferMemory(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument 162 AllocAndBindImageMemory(GrVkGpu* gpu, VkImage image, GrMemoryless memoryless, GrVkAlloc* alloc, int memorySize) global() argument 231 FreeImageMemory(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument 243 MapAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument 258 UnmapAlloc(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument 267 GetNonCoherentMappedMemoryRange(const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size, VkDeviceSize alignment, VkMappedMemoryRange* range) global() argument 290 FlushMappedAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size) global() argument 305 InvalidateMappedAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size) global() argument [all...] |
/kernel/linux/linux-5.10/sound/isa/gus/ |
H A D | gus_mem.c | 18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument 21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock() 23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock() 27 static struct snd_gf1_mem_block *snd_gf1_mem_xalloc(struct snd_gf1_mem * alloc, in snd_gf1_mem_xalloc() argument 36 pblock = alloc->first; in snd_gf1_mem_xalloc() 42 if (pblock == alloc->first) in snd_gf1_mem_xalloc() 43 alloc->first = nblock; in snd_gf1_mem_xalloc() 46 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc() 52 if (alloc->last == NULL) { in snd_gf1_mem_xalloc() 54 alloc in snd_gf1_mem_xalloc() 63 snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) snd_gf1_mem_xfree() argument 93 snd_gf1_mem_look(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_look() argument 106 snd_gf1_mem_share(struct snd_gf1_mem * alloc, unsigned int *share_id) snd_gf1_mem_share() argument 121 snd_gf1_mem_find(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block, unsigned int size, int w_16, int align) snd_gf1_mem_find() argument 173 snd_gf1_mem_alloc(struct snd_gf1_mem * alloc, int owner, char *name, int size, int w_16, int align, unsigned int *share_id) snd_gf1_mem_alloc() argument 207 snd_gf1_mem_free(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_free() argument 224 struct snd_gf1_mem *alloc; snd_gf1_mem_init() local 255 struct snd_gf1_mem *alloc; snd_gf1_mem_done() local 273 struct snd_gf1_mem *alloc; snd_gf1_mem_info_read() local [all...] |
/kernel/linux/linux-6.6/sound/isa/gus/ |
H A D | gus_mem.c | 18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument 21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock() 23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock() 28 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument 43 pblock = alloc->first; in snd_gf1_mem_xalloc() 49 if (pblock == alloc->first) in snd_gf1_mem_xalloc() 50 alloc->first = nblock; in snd_gf1_mem_xalloc() 53 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc() 59 if (alloc->last == NULL) { in snd_gf1_mem_xalloc() 61 alloc in snd_gf1_mem_xalloc() 70 snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) snd_gf1_mem_xfree() argument 100 snd_gf1_mem_look(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_look() argument 113 snd_gf1_mem_share(struct snd_gf1_mem * alloc, unsigned int *share_id) snd_gf1_mem_share() argument 128 snd_gf1_mem_find(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block, unsigned int size, int w_16, int align) snd_gf1_mem_find() argument 180 snd_gf1_mem_alloc(struct snd_gf1_mem * alloc, int owner, char *name, int size, int w_16, int align, unsigned int *share_id) snd_gf1_mem_alloc() argument 213 snd_gf1_mem_free(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_free() argument 231 struct snd_gf1_mem *alloc; snd_gf1_mem_init() local 260 struct snd_gf1_mem *alloc; snd_gf1_mem_done() local 278 struct snd_gf1_mem *alloc; snd_gf1_mem_info_read() local [all...] |
/third_party/mesa3d/src/vulkan/util/ |
H A D | vk_alloc.h | 44 vk_alloc(const VkAllocationCallbacks *alloc, in vk_alloc() argument 48 return alloc->pfnAllocation(alloc->pUserData, size, align, scope); in vk_alloc() 52 vk_zalloc(const VkAllocationCallbacks *alloc, in vk_zalloc() argument 56 void *mem = vk_alloc(alloc, size, align, scope); in vk_zalloc() 66 vk_realloc(const VkAllocationCallbacks *alloc, in vk_realloc() argument 70 return alloc->pfnReallocation(alloc->pUserData, ptr, size, align, scope); in vk_realloc() 74 vk_free(const VkAllocationCallbacks *alloc, void *data) in vk_free() argument 79 alloc in vk_free() 83 vk_strdup(const VkAllocationCallbacks *alloc, const char *s, VkSystemAllocationScope scope) vk_strdup() argument 100 vk_vasprintf(const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope, const char *fmt, va_list args) vk_vasprintf() argument 113 vk_asprintf(const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope, const char *fmt, ...) vk_asprintf() argument 126 vk_alloc2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, size_t size, size_t align, VkSystemAllocationScope scope) vk_alloc2() argument 138 vk_zalloc2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, size_t size, size_t align, VkSystemAllocationScope scope) vk_zalloc2() argument 153 vk_free2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, void *data) vk_free2() argument 230 vk_multialloc_alloc(struct vk_multialloc *ma, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_alloc() argument 269 vk_multialloc_alloc2(struct vk_multialloc *ma, const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_alloc2() argument 278 vk_multialloc_zalloc(struct vk_multialloc *ma, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_zalloc() argument 293 vk_multialloc_zalloc2(struct vk_multialloc *ma, const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_zalloc2() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory_ump.c | 22 mali_mem_allocation *alloc; in mali_mem_ump_map() local 34 alloc = mem_backend->mali_allocation; in mali_mem_ump_map() 35 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map() 37 session = alloc->session; in mali_mem_ump_map() 63 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map() 73 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 83 if (alloc->flags & _MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map() 84 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 98 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument 101 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap() 110 mali_mem_bind_ump_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 secure_id, u32 flags) mali_mem_bind_ump_buf() argument 143 mali_mem_allocation *alloc; mali_mem_unbind_ump_buf() local [all...] |
H A D | mali_memory_external.c | 20 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local 23 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf() 24 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf() 27 session = alloc->session; in mali_mem_unbind_ext_buf() 30 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_unbind_ext_buf() 31 alloc->flags); in mali_mem_unbind_ext_buf() 35 _mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, in mali_mem_bind_ext_buf() argument 44 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf() 45 size = alloc in mali_mem_bind_ext_buf() [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory_ump.c | 23 mali_mem_allocation *alloc; in mali_mem_ump_map() local 35 alloc = mem_backend->mali_allocation; in mali_mem_ump_map() 36 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map() 38 session = alloc->session; in mali_mem_ump_map() 64 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map() 74 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 84 if (alloc->flags & MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map() 85 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 99 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument 102 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap() 110 mali_mem_bind_ump_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 secure_id, u32 flags) mali_mem_bind_ump_buf() argument 145 mali_mem_allocation *alloc; mali_mem_unbind_ump_buf() local [all...] |
H A D | mali_memory_external.c | 21 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local 24 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf() 25 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf() 28 session = alloc->session; in mali_mem_unbind_ext_buf() 31 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_unbind_ext_buf() 35 mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 phys_addr, in mali_mem_bind_ext_buf() argument 42 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf() 43 size = alloc in mali_mem_bind_ext_buf() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument 45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc() 47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc() 48 if (obj >= alloc->max) { in mthca_alloc() 49 alloc->top = (alloc->top + alloc->max) & alloc in mthca_alloc() 64 mthca_free(struct mthca_alloc *alloc, u32 obj) mthca_free() argument 79 mthca_alloc_init(struct mthca_alloc *alloc, u32 num, u32 mask, u32 reserved) mthca_alloc_init() argument 105 mthca_alloc_cleanup(struct mthca_alloc *alloc) mthca_alloc_cleanup() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_allocator.c | 40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument 45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc() 47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc() 48 if (obj >= alloc->max) { in mthca_alloc() 49 alloc->top = (alloc->top + alloc->max) & alloc in mthca_alloc() 64 mthca_free(struct mthca_alloc *alloc, u32 obj) mthca_free() argument 79 mthca_alloc_init(struct mthca_alloc *alloc, u32 num, u32 mask, u32 reserved) mthca_alloc_init() argument 100 mthca_alloc_cleanup(struct mthca_alloc *alloc) mthca_alloc_cleanup() argument [all...] |
/kernel/linux/linux-5.10/fs/ocfs2/ |
H A D | localalloc.c | 22 #include "alloc.h" 37 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc); 40 struct ocfs2_dinode *alloc, 44 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc); 48 struct ocfs2_dinode *alloc, 66 * the local alloc. 68 * Generally, we'd like to pick as large a local alloc as 74 * Some things work against us when trying to choose a large local alloc: 89 * alloc maximums at various cluster sizes (4k blocksize) 144 * local alloc siz in ocfs2_la_default_mb() 273 struct ocfs2_dinode *alloc = NULL; ocfs2_load_local_alloc() local 377 struct ocfs2_dinode *alloc = NULL; ocfs2_shutdown_local_alloc() local 485 struct ocfs2_dinode *alloc; ocfs2_begin_local_alloc_recovery() local 548 ocfs2_complete_local_alloc_recovery(struct ocfs2_super *osb, struct ocfs2_dinode *alloc) ocfs2_complete_local_alloc_recovery() argument 621 struct ocfs2_dinode *alloc; ocfs2_reserve_local_alloc_bits() local 725 struct ocfs2_dinode *alloc; ocfs2_claim_local_alloc_bits() local 781 struct ocfs2_dinode *alloc; ocfs2_free_local_alloc_bits() local 813 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc) ocfs2_local_alloc_count_bits() argument 824 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument 912 ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc) ocfs2_clear_local_alloc() argument 949 ocfs2_sync_local_to_main(struct ocfs2_super *osb, handle_t *handle, struct ocfs2_dinode *alloc, struct inode *main_bm_inode, struct buffer_head *main_bm_bh) ocfs2_sync_local_to_main() argument 1154 struct ocfs2_dinode *alloc = NULL; ocfs2_local_alloc_new_window() local 1243 struct ocfs2_dinode *alloc; ocfs2_local_alloc_slide_window() local [all...] |
/kernel/linux/linux-6.6/fs/ocfs2/ |
H A D | localalloc.c | 20 #include "alloc.h" 35 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc); 38 struct ocfs2_dinode *alloc, 42 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc); 46 struct ocfs2_dinode *alloc, 64 * the local alloc. 66 * Generally, we'd like to pick as large a local alloc as 72 * Some things work against us when trying to choose a large local alloc: 87 * alloc maximums at various cluster sizes (4k blocksize) 142 * local alloc siz in ocfs2_la_default_mb() 271 struct ocfs2_dinode *alloc = NULL; ocfs2_load_local_alloc() local 375 struct ocfs2_dinode *alloc = NULL; ocfs2_shutdown_local_alloc() local 483 struct ocfs2_dinode *alloc; ocfs2_begin_local_alloc_recovery() local 546 ocfs2_complete_local_alloc_recovery(struct ocfs2_super *osb, struct ocfs2_dinode *alloc) ocfs2_complete_local_alloc_recovery() argument 619 struct ocfs2_dinode *alloc; ocfs2_reserve_local_alloc_bits() local 723 struct ocfs2_dinode *alloc; ocfs2_claim_local_alloc_bits() local 779 struct ocfs2_dinode *alloc; ocfs2_free_local_alloc_bits() local 811 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc) ocfs2_local_alloc_count_bits() argument 822 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument 910 ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc) ocfs2_clear_local_alloc() argument 947 ocfs2_sync_local_to_main(struct ocfs2_super *osb, handle_t *handle, struct ocfs2_dinode *alloc, struct inode *main_bm_inode, struct buffer_head *main_bm_bh) ocfs2_sync_local_to_main() argument 1152 struct ocfs2_dinode *alloc = NULL; ocfs2_local_alloc_new_window() local 1241 struct ocfs2_dinode *alloc; ocfs2_local_alloc_slide_window() local [all...] |
/kernel/linux/linux-5.10/fs/xfs/libxfs/ |
H A D | xfs_alloc_btree.c | 130 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 133 if (be32_to_cpu(rec->alloc.ar_blockcount) <= in xfs_allocbt_update_lastrec() 136 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 186 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec() 187 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec() 197 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec() 198 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec() 199 key->alloc in xfs_bnobt_init_high_key_from_rec() [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem.h | 67 struct kbase_mem_phy_alloc *alloc; member 85 * but with alloc instead of a gpu va (handle) */ 87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 107 struct kref kref; /* number of users of this alloc */ 175 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 177 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 179 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped() 180 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 184 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 186 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 202 kbase_mem_phy_alloc_get(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_get() argument 208 kbase_mem_phy_alloc_put(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_put() argument 351 struct kbase_mem_phy_alloc *alloc; kbase_alloc_create() local [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem.h | 69 struct kbase_mem_phy_alloc *alloc; member 87 * but with alloc instead of a gpu va (handle) */ 89 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 109 struct kref kref; /* number of users of this alloc */ 177 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 179 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 181 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped() 182 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 185 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 187 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 202 kbase_mem_phy_alloc_get(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_get() argument 208 kbase_mem_phy_alloc_put(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_put() argument 349 struct kbase_mem_phy_alloc *alloc; kbase_alloc_create() local [all...] |
/kernel/linux/linux-6.6/fs/xfs/libxfs/ |
H A D | xfs_alloc_btree.c | 129 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 132 if (be32_to_cpu(rec->alloc.ar_blockcount) <= in xfs_allocbt_update_lastrec() 135 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 185 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec() 186 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec() 196 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec() 197 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec() 198 key->alloc in xfs_bnobt_init_high_key_from_rec() [all...] |
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_image.c | 113 const VkAllocationCallbacks *alloc) in vn_image_deferred_info_init() 118 info = vk_zalloc(alloc, sizeof(*info), VN_DEFAULT_ALIGN, in vn_image_deferred_info_init() 145 alloc, size, VN_DEFAULT_ALIGN, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); in vn_image_deferred_info_init() 147 vk_free(alloc, info); in vn_image_deferred_info_init() 184 const VkAllocationCallbacks *alloc) in vn_image_deferred_info_fini() 190 vk_free(alloc, (void *)img->deferred_info->list.pViewFormats); in vn_image_deferred_info_fini() 192 vk_free(alloc, img->deferred_info); in vn_image_deferred_info_fini() 220 const VkAllocationCallbacks *alloc, in vn_image_create() 226 img = vk_zalloc(alloc, sizeof(*img), VN_DEFAULT_ALIGN, in vn_image_create() 236 vk_free(alloc, im in vn_image_create() 111 vn_image_deferred_info_init(struct vn_image *img, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc) vn_image_deferred_info_init() argument 183 vn_image_deferred_info_fini(struct vn_image *img, const VkAllocationCallbacks *alloc) vn_image_deferred_info_fini() argument 218 vn_image_create(struct vn_device *dev, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc, struct vn_image **out_img) vn_image_create() argument 256 vn_image_create_deferred(struct vn_device *dev, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc, struct vn_image **out_img) vn_image_create_deferred() argument 293 const VkAllocationCallbacks *alloc = vn_CreateImage() local 349 const VkAllocationCallbacks *alloc = vn_DestroyImage() local 447 const VkAllocationCallbacks *alloc = &dev->base.base.alloc; vn_BindImageMemory2() local 582 const VkAllocationCallbacks *alloc = vn_CreateImageView() local 621 const VkAllocationCallbacks *alloc = vn_DestroyImageView() local 642 const VkAllocationCallbacks *alloc = vn_CreateSampler() local 669 const VkAllocationCallbacks *alloc = vn_DestroySampler() local 691 const VkAllocationCallbacks *alloc = vn_CreateSamplerYcbcrConversion() local 739 const VkAllocationCallbacks *alloc = vn_DestroySamplerYcbcrConversion() local [all...] |