Home
last modified time | relevance | path

Searched refs:alloc (Results 1 - 25 of 2222) sorted by relevance

12345678910>>...89

/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/
H A Did_table.c44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument
49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc()
51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc()
52 if (obj >= alloc->max) in c4iw_id_alloc()
53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc()
55 if (obj < alloc->max) { in c4iw_id_alloc()
56 if (alloc in c4iw_id_alloc()
71 c4iw_id_free(struct c4iw_id_table *alloc, u32 obj) c4iw_id_free() argument
82 c4iw_id_table_alloc(struct c4iw_id_table *alloc, u32 start, u32 num, u32 reserved, u32 flags) c4iw_id_table_alloc() argument
108 c4iw_id_table_free(struct c4iw_id_table *alloc) c4iw_id_table_free() argument
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/
H A Did_table.c44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument
49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc()
51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc()
52 if (obj >= alloc->max) in c4iw_id_alloc()
53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc()
55 if (obj < alloc->max) { in c4iw_id_alloc()
56 if (alloc in c4iw_id_alloc()
71 c4iw_id_free(struct c4iw_id_table *alloc, u32 obj) c4iw_id_free() argument
82 c4iw_id_table_alloc(struct c4iw_id_table *alloc, u32 start, u32 num, u32 reserved, u32 flags) c4iw_id_table_alloc() argument
103 c4iw_id_table_free(struct c4iw_id_table *alloc) c4iw_id_table_free() argument
[all...]
/kernel/linux/linux-6.6/drivers/android/
H A Dbinder_alloc.c60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument
63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size()
64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size()
68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument
71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer()
79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer()
83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer()
90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer()
98 rb_insert_color(&new_buffer->rb_node, &alloc in binder_insert_free_buffer()
101 binder_insert_allocated_buffer_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer) binder_insert_allocated_buffer_locked() argument
126 binder_alloc_prepare_to_free_locked( struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free_locked() argument
170 binder_alloc_prepare_to_free(struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free() argument
181 binder_update_page_range(struct binder_alloc *alloc, int allocate, void __user *start, void __user *end) binder_update_page_range() argument
312 binder_alloc_set_vma(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_set_vma() argument
319 binder_alloc_get_vma( struct binder_alloc *alloc) binder_alloc_get_vma() argument
326 debug_low_async_space_locked(struct binder_alloc *alloc, int pid) debug_low_async_space_locked() argument
368 binder_alloc_new_buf_locked( struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf_locked() argument
561 binder_alloc_new_buf(struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf() argument
588 binder_delete_free_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_delete_free_buffer() argument
637 binder_free_buf_locked(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_free_buf_locked() argument
701 binder_alloc_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_free_buf() argument
734 binder_alloc_mmap_handler(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_mmap_handler() argument
804 binder_alloc_deferred_release(struct binder_alloc *alloc) binder_alloc_deferred_release() argument
888 binder_alloc_print_allocated(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_allocated() argument
905 binder_alloc_print_pages(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_pages() argument
941 binder_alloc_get_allocated_count(struct binder_alloc *alloc) binder_alloc_get_allocated_count() argument
962 binder_alloc_vma_close(struct binder_alloc *alloc) binder_alloc_vma_close() argument
986 struct binder_alloc *alloc; __must_hold() local
1071 binder_alloc_init(struct binder_alloc *alloc) binder_alloc_init() argument
1117 check_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t offset, size_t bytes) check_buffer() argument
1149 binder_alloc_get_page(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, pgoff_t *pgoffp) binder_alloc_get_page() argument
1172 binder_alloc_clear_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_clear_buf() argument
1205 binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, const void __user *from, size_t bytes) binder_alloc_copy_user_to_buffer() argument
1236 binder_alloc_do_buffer_copy(struct binder_alloc *alloc, bool to_buffer, struct binder_buffer *buffer, binder_size_t buffer_offset, void *ptr, size_t bytes) binder_alloc_do_buffer_copy() argument
1267 binder_alloc_copy_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, void *src, size_t bytes) binder_alloc_copy_to_buffer() argument
1277 binder_alloc_copy_from_buffer(struct binder_alloc *alloc, void *dest, struct binder_buffer *buffer, binder_size_t buffer_offset, size_t bytes) binder_alloc_copy_from_buffer() argument
[all...]
H A Dbinder_alloc_selftest.c82 pr_err("alloc sizes: "); in pr_err_size_seq()
92 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument
103 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated()
104 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated()
105 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated()
106 pr_err("expect alloc but is %s at page index %d\n", in check_buffer_pages_allocated()
107 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated()
115 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument
122 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0, 0); in binder_selftest_alloc_buf()
124 !check_buffer_pages_allocated(alloc, buffer in binder_selftest_alloc_buf()
132 binder_selftest_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffers[], size_t *sizes, int *seq, size_t end) binder_selftest_free_buf() argument
156 binder_selftest_free_page(struct binder_alloc *alloc) binder_selftest_free_page() argument
176 binder_selftest_alloc_free(struct binder_alloc *alloc, size_t *sizes, int *seq, size_t end) binder_selftest_alloc_free() argument
205 binder_selftest_free_seq(struct binder_alloc *alloc, size_t *sizes, int *seq, int index, size_t end) binder_selftest_free_seq() argument
223 binder_selftest_alloc_size(struct binder_alloc *alloc, size_t *end_offset) binder_selftest_alloc_size() argument
249 binder_selftest_alloc_offset(struct binder_alloc *alloc, size_t *end_offset, int index) binder_selftest_alloc_offset() argument
283 binder_selftest_alloc(struct binder_alloc *alloc) binder_selftest_alloc() argument
[all...]
H A Dbinder_alloc.h23 * @entry: entry alloc->buffers
67 * @alloc: binder_alloc for a proc
72 struct binder_alloc *alloc; member
117 void binder_selftest_alloc(struct binder_alloc *alloc);
119 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument
124 extern struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc,
130 extern void binder_alloc_init(struct binder_alloc *alloc);
133 extern void binder_alloc_vma_close(struct binder_alloc *alloc);
135 binder_alloc_prepare_to_free(struct binder_alloc *alloc,
137 extern void binder_alloc_free_buf(struct binder_alloc *alloc,
155 binder_alloc_get_free_async_space(struct binder_alloc *alloc) binder_alloc_get_free_async_space() argument
[all...]
/kernel/linux/linux-5.10/drivers/android/
H A Dbinder_alloc.c60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument
63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size()
64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size()
68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument
71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer()
79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer()
83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer()
90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer()
98 rb_insert_color(&new_buffer->rb_node, &alloc in binder_insert_free_buffer()
101 binder_insert_allocated_buffer_locked( struct binder_alloc *alloc, struct binder_buffer *new_buffer) binder_insert_allocated_buffer_locked() argument
126 binder_alloc_prepare_to_free_locked( struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free_locked() argument
170 binder_alloc_prepare_to_free(struct binder_alloc *alloc, uintptr_t user_ptr) binder_alloc_prepare_to_free() argument
181 binder_update_page_range(struct binder_alloc *alloc, int allocate, void __user *start, void __user *end) binder_update_page_range() argument
313 binder_alloc_set_vma(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_set_vma() argument
328 binder_alloc_get_vma( struct binder_alloc *alloc) binder_alloc_get_vma() argument
341 debug_low_async_space_locked(struct binder_alloc *alloc, int pid) debug_low_async_space_locked() argument
377 binder_alloc_new_buf_locked( struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf_locked() argument
566 binder_alloc_new_buf(struct binder_alloc *alloc, size_t data_size, size_t offsets_size, size_t extra_buffers_size, int is_async, int pid) binder_alloc_new_buf() argument
593 binder_delete_free_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_delete_free_buffer() argument
642 binder_free_buf_locked(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_free_buf_locked() argument
706 binder_alloc_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_free_buf() argument
739 binder_alloc_mmap_handler(struct binder_alloc *alloc, struct vm_area_struct *vma) binder_alloc_mmap_handler() argument
801 binder_alloc_deferred_release(struct binder_alloc *alloc) binder_alloc_deferred_release() argument
885 binder_alloc_print_allocated(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_allocated() argument
902 binder_alloc_print_pages(struct seq_file *m, struct binder_alloc *alloc) binder_alloc_print_pages() argument
938 binder_alloc_get_allocated_count(struct binder_alloc *alloc) binder_alloc_get_allocated_count() argument
959 binder_alloc_vma_close(struct binder_alloc *alloc) binder_alloc_vma_close() argument
983 struct binder_alloc *alloc; __must_hold() local
1072 binder_alloc_init(struct binder_alloc *alloc) binder_alloc_init() argument
1116 check_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t offset, size_t bytes) check_buffer() argument
1148 binder_alloc_get_page(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, pgoff_t *pgoffp) binder_alloc_get_page() argument
1171 binder_alloc_clear_buf(struct binder_alloc *alloc, struct binder_buffer *buffer) binder_alloc_clear_buf() argument
1207 binder_alloc_copy_user_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, const void __user *from, size_t bytes) binder_alloc_copy_user_to_buffer() argument
1238 binder_alloc_do_buffer_copy(struct binder_alloc *alloc, bool to_buffer, struct binder_buffer *buffer, binder_size_t buffer_offset, void *ptr, size_t bytes) binder_alloc_do_buffer_copy() argument
1278 binder_alloc_copy_to_buffer(struct binder_alloc *alloc, struct binder_buffer *buffer, binder_size_t buffer_offset, void *src, size_t bytes) binder_alloc_copy_to_buffer() argument
1288 binder_alloc_copy_from_buffer(struct binder_alloc *alloc, void *dest, struct binder_buffer *buffer, binder_size_t buffer_offset, size_t bytes) binder_alloc_copy_from_buffer() argument
[all...]
H A Dbinder_alloc_selftest.c82 pr_err("alloc sizes: "); in pr_err_size_seq()
92 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument
103 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated()
104 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated()
105 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated()
106 pr_err("expect alloc but is %s at page index %d\n", in check_buffer_pages_allocated()
107 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated()
115 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument
122 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0, 0); in binder_selftest_alloc_buf()
124 !check_buffer_pages_allocated(alloc, buffer in binder_selftest_alloc_buf()
132 binder_selftest_free_buf(struct binder_alloc *alloc, struct binder_buffer *buffers[], size_t *sizes, int *seq, size_t end) binder_selftest_free_buf() argument
156 binder_selftest_free_page(struct binder_alloc *alloc) binder_selftest_free_page() argument
176 binder_selftest_alloc_free(struct binder_alloc *alloc, size_t *sizes, int *seq, size_t end) binder_selftest_alloc_free() argument
205 binder_selftest_free_seq(struct binder_alloc *alloc, size_t *sizes, int *seq, int index, size_t end) binder_selftest_free_seq() argument
223 binder_selftest_alloc_size(struct binder_alloc *alloc, size_t *end_offset) binder_selftest_alloc_size() argument
249 binder_selftest_alloc_offset(struct binder_alloc *alloc, size_t *end_offset, int index) binder_selftest_alloc_offset() argument
283 binder_selftest_alloc(struct binder_alloc *alloc) binder_selftest_alloc() argument
[all...]
H A Dbinder_alloc.h23 * @entry: entry alloc->buffers
64 * @alloc: binder_alloc for a proc
69 struct binder_alloc *alloc; member
113 void binder_selftest_alloc(struct binder_alloc *alloc);
115 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument
120 extern struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc,
126 extern void binder_alloc_init(struct binder_alloc *alloc);
129 extern void binder_alloc_vma_close(struct binder_alloc *alloc);
131 binder_alloc_prepare_to_free(struct binder_alloc *alloc,
133 extern void binder_alloc_free_buf(struct binder_alloc *alloc,
151 binder_alloc_get_free_async_space(struct binder_alloc *alloc) binder_alloc_get_free_async_space() argument
[all...]
/third_party/skia/src/gpu/vk/
H A DGrVkMemory.cpp48 GrVkAlloc* alloc,
51 GrVkAlloc* alloc) {
71 allocator->getAllocInfo(memory, alloc);
74 alloc->fBytes = size;
80 GR_VK_CALL_RESULT(gpu, err, BindBufferMemory(gpu->device(), buffer, alloc->fMemory,
81 alloc->fOffset));
83 FreeBufferMemory(gpu, *alloc);
93 GrVkAlloc* alloc) {
137 alloc->fMemory = memory;
138 alloc
44 AllocAndBindBufferMemory(GrVkGpu* gpu, VkBuffer buffer, BufferUsage usage, GrVkAlloc* alloc, size_t size) global() argument
90 ImportAndBindBufferMemory(GrVkGpu* gpu, OH_NativeBuffer *nativeBuffer, VkBuffer buffer, GrVkAlloc* alloc) global() argument
146 FreeBufferMemory(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument
162 AllocAndBindImageMemory(GrVkGpu* gpu, VkImage image, GrMemoryless memoryless, GrVkAlloc* alloc, int memorySize) global() argument
231 FreeImageMemory(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument
243 MapAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument
258 UnmapAlloc(const GrVkGpu* gpu, const GrVkAlloc& alloc) global() argument
267 GetNonCoherentMappedMemoryRange(const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size, VkDeviceSize alignment, VkMappedMemoryRange* range) global() argument
290 FlushMappedAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size) global() argument
305 InvalidateMappedAlloc(GrVkGpu* gpu, const GrVkAlloc& alloc, VkDeviceSize offset, VkDeviceSize size) global() argument
[all...]
/kernel/linux/linux-5.10/sound/isa/gus/
H A Dgus_mem.c18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument
21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock()
23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock()
27 static struct snd_gf1_mem_block *snd_gf1_mem_xalloc(struct snd_gf1_mem * alloc, in snd_gf1_mem_xalloc() argument
36 pblock = alloc->first; in snd_gf1_mem_xalloc()
42 if (pblock == alloc->first) in snd_gf1_mem_xalloc()
43 alloc->first = nblock; in snd_gf1_mem_xalloc()
46 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc()
52 if (alloc->last == NULL) { in snd_gf1_mem_xalloc()
54 alloc in snd_gf1_mem_xalloc()
63 snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) snd_gf1_mem_xfree() argument
93 snd_gf1_mem_look(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_look() argument
106 snd_gf1_mem_share(struct snd_gf1_mem * alloc, unsigned int *share_id) snd_gf1_mem_share() argument
121 snd_gf1_mem_find(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block, unsigned int size, int w_16, int align) snd_gf1_mem_find() argument
173 snd_gf1_mem_alloc(struct snd_gf1_mem * alloc, int owner, char *name, int size, int w_16, int align, unsigned int *share_id) snd_gf1_mem_alloc() argument
207 snd_gf1_mem_free(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_free() argument
224 struct snd_gf1_mem *alloc; snd_gf1_mem_init() local
255 struct snd_gf1_mem *alloc; snd_gf1_mem_done() local
273 struct snd_gf1_mem *alloc; snd_gf1_mem_info_read() local
[all...]
/kernel/linux/linux-6.6/sound/isa/gus/
H A Dgus_mem.c18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument
21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock()
23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock()
28 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument
43 pblock = alloc->first; in snd_gf1_mem_xalloc()
49 if (pblock == alloc->first) in snd_gf1_mem_xalloc()
50 alloc->first = nblock; in snd_gf1_mem_xalloc()
53 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc()
59 if (alloc->last == NULL) { in snd_gf1_mem_xalloc()
61 alloc in snd_gf1_mem_xalloc()
70 snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) snd_gf1_mem_xfree() argument
100 snd_gf1_mem_look(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_look() argument
113 snd_gf1_mem_share(struct snd_gf1_mem * alloc, unsigned int *share_id) snd_gf1_mem_share() argument
128 snd_gf1_mem_find(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block, unsigned int size, int w_16, int align) snd_gf1_mem_find() argument
180 snd_gf1_mem_alloc(struct snd_gf1_mem * alloc, int owner, char *name, int size, int w_16, int align, unsigned int *share_id) snd_gf1_mem_alloc() argument
213 snd_gf1_mem_free(struct snd_gf1_mem * alloc, unsigned int address) snd_gf1_mem_free() argument
231 struct snd_gf1_mem *alloc; snd_gf1_mem_init() local
260 struct snd_gf1_mem *alloc; snd_gf1_mem_done() local
278 struct snd_gf1_mem *alloc; snd_gf1_mem_info_read() local
[all...]
/third_party/mesa3d/src/vulkan/util/
H A Dvk_alloc.h44 vk_alloc(const VkAllocationCallbacks *alloc, in vk_alloc() argument
48 return alloc->pfnAllocation(alloc->pUserData, size, align, scope); in vk_alloc()
52 vk_zalloc(const VkAllocationCallbacks *alloc, in vk_zalloc() argument
56 void *mem = vk_alloc(alloc, size, align, scope); in vk_zalloc()
66 vk_realloc(const VkAllocationCallbacks *alloc, in vk_realloc() argument
70 return alloc->pfnReallocation(alloc->pUserData, ptr, size, align, scope); in vk_realloc()
74 vk_free(const VkAllocationCallbacks *alloc, void *data) in vk_free() argument
79 alloc in vk_free()
83 vk_strdup(const VkAllocationCallbacks *alloc, const char *s, VkSystemAllocationScope scope) vk_strdup() argument
100 vk_vasprintf(const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope, const char *fmt, va_list args) vk_vasprintf() argument
113 vk_asprintf(const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope, const char *fmt, ...) vk_asprintf() argument
126 vk_alloc2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, size_t size, size_t align, VkSystemAllocationScope scope) vk_alloc2() argument
138 vk_zalloc2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, size_t size, size_t align, VkSystemAllocationScope scope) vk_zalloc2() argument
153 vk_free2(const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, void *data) vk_free2() argument
230 vk_multialloc_alloc(struct vk_multialloc *ma, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_alloc() argument
269 vk_multialloc_alloc2(struct vk_multialloc *ma, const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_alloc2() argument
278 vk_multialloc_zalloc(struct vk_multialloc *ma, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_zalloc() argument
293 vk_multialloc_zalloc2(struct vk_multialloc *ma, const VkAllocationCallbacks *parent_alloc, const VkAllocationCallbacks *alloc, VkSystemAllocationScope scope) vk_multialloc_zalloc2() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory_ump.c22 mali_mem_allocation *alloc; in mali_mem_ump_map() local
34 alloc = mem_backend->mali_allocation; in mali_mem_ump_map()
35 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map()
37 session = alloc->session; in mali_mem_ump_map()
63 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map()
73 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
83 if (alloc->flags & _MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map()
84 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
98 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument
101 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap()
110 mali_mem_bind_ump_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 secure_id, u32 flags) mali_mem_bind_ump_buf() argument
143 mali_mem_allocation *alloc; mali_mem_unbind_ump_buf() local
[all...]
H A Dmali_memory_external.c20 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local
23 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf()
24 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf()
27 session = alloc->session; in mali_mem_unbind_ext_buf()
30 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_unbind_ext_buf()
31 alloc->flags); in mali_mem_unbind_ext_buf()
35 _mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, in mali_mem_bind_ext_buf() argument
44 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf()
45 size = alloc in mali_mem_bind_ext_buf()
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory_ump.c23 mali_mem_allocation *alloc; in mali_mem_ump_map() local
35 alloc = mem_backend->mali_allocation; in mali_mem_ump_map()
36 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map()
38 session = alloc->session; in mali_mem_ump_map()
64 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map()
74 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
84 if (alloc->flags & MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map()
85 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
99 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument
102 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap()
110 mali_mem_bind_ump_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 secure_id, u32 flags) mali_mem_bind_ump_buf() argument
145 mali_mem_allocation *alloc; mali_mem_unbind_ump_buf() local
[all...]
H A Dmali_memory_external.c21 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local
24 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf()
25 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf()
28 session = alloc->session; in mali_mem_unbind_ext_buf()
31 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_unbind_ext_buf()
35 mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend, u32 phys_addr, in mali_mem_bind_ext_buf() argument
42 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf()
43 size = alloc in mali_mem_bind_ext_buf()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument
45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc()
47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc()
48 if (obj >= alloc->max) { in mthca_alloc()
49 alloc->top = (alloc->top + alloc->max) & alloc in mthca_alloc()
64 mthca_free(struct mthca_alloc *alloc, u32 obj) mthca_free() argument
79 mthca_alloc_init(struct mthca_alloc *alloc, u32 num, u32 mask, u32 reserved) mthca_alloc_init() argument
105 mthca_alloc_cleanup(struct mthca_alloc *alloc) mthca_alloc_cleanup() argument
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument
45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc()
47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc()
48 if (obj >= alloc->max) { in mthca_alloc()
49 alloc->top = (alloc->top + alloc->max) & alloc in mthca_alloc()
64 mthca_free(struct mthca_alloc *alloc, u32 obj) mthca_free() argument
79 mthca_alloc_init(struct mthca_alloc *alloc, u32 num, u32 mask, u32 reserved) mthca_alloc_init() argument
100 mthca_alloc_cleanup(struct mthca_alloc *alloc) mthca_alloc_cleanup() argument
[all...]
/kernel/linux/linux-5.10/fs/ocfs2/
H A Dlocalalloc.c22 #include "alloc.h"
37 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc);
40 struct ocfs2_dinode *alloc,
44 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc);
48 struct ocfs2_dinode *alloc,
66 * the local alloc.
68 * Generally, we'd like to pick as large a local alloc as
74 * Some things work against us when trying to choose a large local alloc:
89 * alloc maximums at various cluster sizes (4k blocksize)
144 * local alloc siz in ocfs2_la_default_mb()
273 struct ocfs2_dinode *alloc = NULL; ocfs2_load_local_alloc() local
377 struct ocfs2_dinode *alloc = NULL; ocfs2_shutdown_local_alloc() local
485 struct ocfs2_dinode *alloc; ocfs2_begin_local_alloc_recovery() local
548 ocfs2_complete_local_alloc_recovery(struct ocfs2_super *osb, struct ocfs2_dinode *alloc) ocfs2_complete_local_alloc_recovery() argument
621 struct ocfs2_dinode *alloc; ocfs2_reserve_local_alloc_bits() local
725 struct ocfs2_dinode *alloc; ocfs2_claim_local_alloc_bits() local
781 struct ocfs2_dinode *alloc; ocfs2_free_local_alloc_bits() local
813 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc) ocfs2_local_alloc_count_bits() argument
824 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument
912 ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc) ocfs2_clear_local_alloc() argument
949 ocfs2_sync_local_to_main(struct ocfs2_super *osb, handle_t *handle, struct ocfs2_dinode *alloc, struct inode *main_bm_inode, struct buffer_head *main_bm_bh) ocfs2_sync_local_to_main() argument
1154 struct ocfs2_dinode *alloc = NULL; ocfs2_local_alloc_new_window() local
1243 struct ocfs2_dinode *alloc; ocfs2_local_alloc_slide_window() local
[all...]
/kernel/linux/linux-6.6/fs/ocfs2/
H A Dlocalalloc.c20 #include "alloc.h"
35 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc);
38 struct ocfs2_dinode *alloc,
42 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc);
46 struct ocfs2_dinode *alloc,
64 * the local alloc.
66 * Generally, we'd like to pick as large a local alloc as
72 * Some things work against us when trying to choose a large local alloc:
87 * alloc maximums at various cluster sizes (4k blocksize)
142 * local alloc siz in ocfs2_la_default_mb()
271 struct ocfs2_dinode *alloc = NULL; ocfs2_load_local_alloc() local
375 struct ocfs2_dinode *alloc = NULL; ocfs2_shutdown_local_alloc() local
483 struct ocfs2_dinode *alloc; ocfs2_begin_local_alloc_recovery() local
546 ocfs2_complete_local_alloc_recovery(struct ocfs2_super *osb, struct ocfs2_dinode *alloc) ocfs2_complete_local_alloc_recovery() argument
619 struct ocfs2_dinode *alloc; ocfs2_reserve_local_alloc_bits() local
723 struct ocfs2_dinode *alloc; ocfs2_claim_local_alloc_bits() local
779 struct ocfs2_dinode *alloc; ocfs2_free_local_alloc_bits() local
811 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc) ocfs2_local_alloc_count_bits() argument
822 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument
910 ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc) ocfs2_clear_local_alloc() argument
947 ocfs2_sync_local_to_main(struct ocfs2_super *osb, handle_t *handle, struct ocfs2_dinode *alloc, struct inode *main_bm_inode, struct buffer_head *main_bm_bh) ocfs2_sync_local_to_main() argument
1152 struct ocfs2_dinode *alloc = NULL; ocfs2_local_alloc_new_window() local
1241 struct ocfs2_dinode *alloc; ocfs2_local_alloc_slide_window() local
[all...]
/kernel/linux/linux-5.10/fs/xfs/libxfs/
H A Dxfs_alloc_btree.c130 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec()
133 if (be32_to_cpu(rec->alloc.ar_blockcount) <= in xfs_allocbt_update_lastrec()
136 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec()
186 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec()
187 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec()
197 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec()
198 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec()
199 key->alloc in xfs_bnobt_init_high_key_from_rec()
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.h67 struct kbase_mem_phy_alloc *alloc; member
85 * but with alloc instead of a gpu va (handle) */
87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
107 struct kref kref; /* number of users of this alloc */
175 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
177 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
179 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped()
180 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
184 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
186 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
202 kbase_mem_phy_alloc_get(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_get() argument
208 kbase_mem_phy_alloc_put(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_put() argument
351 struct kbase_mem_phy_alloc *alloc; kbase_alloc_create() local
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.h69 struct kbase_mem_phy_alloc *alloc; member
87 * but with alloc instead of a gpu va (handle) */
89 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
109 struct kref kref; /* number of users of this alloc */
177 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
179 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
181 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped()
182 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
185 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
187 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
202 kbase_mem_phy_alloc_get(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_get() argument
208 kbase_mem_phy_alloc_put(struct kbase_mem_phy_alloc *alloc) kbase_mem_phy_alloc_put() argument
349 struct kbase_mem_phy_alloc *alloc; kbase_alloc_create() local
[all...]
/kernel/linux/linux-6.6/fs/xfs/libxfs/
H A Dxfs_alloc_btree.c129 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec()
132 if (be32_to_cpu(rec->alloc.ar_blockcount) <= in xfs_allocbt_update_lastrec()
135 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec()
185 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec()
186 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec()
196 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec()
197 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec()
198 key->alloc in xfs_bnobt_init_high_key_from_rec()
[all...]
/third_party/mesa3d/src/virtio/vulkan/
H A Dvn_image.c113 const VkAllocationCallbacks *alloc) in vn_image_deferred_info_init()
118 info = vk_zalloc(alloc, sizeof(*info), VN_DEFAULT_ALIGN, in vn_image_deferred_info_init()
145 alloc, size, VN_DEFAULT_ALIGN, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); in vn_image_deferred_info_init()
147 vk_free(alloc, info); in vn_image_deferred_info_init()
184 const VkAllocationCallbacks *alloc) in vn_image_deferred_info_fini()
190 vk_free(alloc, (void *)img->deferred_info->list.pViewFormats); in vn_image_deferred_info_fini()
192 vk_free(alloc, img->deferred_info); in vn_image_deferred_info_fini()
220 const VkAllocationCallbacks *alloc, in vn_image_create()
226 img = vk_zalloc(alloc, sizeof(*img), VN_DEFAULT_ALIGN, in vn_image_create()
236 vk_free(alloc, im in vn_image_create()
111 vn_image_deferred_info_init(struct vn_image *img, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc) vn_image_deferred_info_init() argument
183 vn_image_deferred_info_fini(struct vn_image *img, const VkAllocationCallbacks *alloc) vn_image_deferred_info_fini() argument
218 vn_image_create(struct vn_device *dev, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc, struct vn_image **out_img) vn_image_create() argument
256 vn_image_create_deferred(struct vn_device *dev, const VkImageCreateInfo *create_info, const VkAllocationCallbacks *alloc, struct vn_image **out_img) vn_image_create_deferred() argument
293 const VkAllocationCallbacks *alloc = vn_CreateImage() local
349 const VkAllocationCallbacks *alloc = vn_DestroyImage() local
447 const VkAllocationCallbacks *alloc = &dev->base.base.alloc; vn_BindImageMemory2() local
582 const VkAllocationCallbacks *alloc = vn_CreateImageView() local
621 const VkAllocationCallbacks *alloc = vn_DestroyImageView() local
642 const VkAllocationCallbacks *alloc = vn_CreateSampler() local
669 const VkAllocationCallbacks *alloc = vn_DestroySampler() local
691 const VkAllocationCallbacks *alloc = vn_CreateSamplerYcbcrConversion() local
739 const VkAllocationCallbacks *alloc = vn_DestroySamplerYcbcrConversion() local
[all...]

Completed in 13 milliseconds

12345678910>>...89