Home
last modified time | relevance | path

Searched refs:order (Results 1 - 25 of 78) sorted by relevance

1234

/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_osk_locks.c26 static const char *const lock_order_to_string(_mali_osk_lock_order_t order);
30 _mali_osk_lock_order_t order) in _mali_osk_locks_debug_init()
36 checker->order = order; in _mali_osk_locks_debug_init()
48 printk(KERN_ERR "%d: ERROR lock %p taken while holding a lock of a higher order.\n", mali_osk_get_tid(), in _mali_osk_locks_debug_add()
67 /* Lock order checking
72 * make sure that a lock that is taken has a higher order than the current highest-order lock a
80 * lock taken by this thread. If the last lock's order was lower than the
97 printk(" [lock: %p, tid_owner: %d, order in dump_lock_tracking_list()
29 _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_locks_debug_init() argument
225 lock_order_to_string(_mali_osk_lock_order_t order) lock_order_to_string() argument
[all...]
H A Dmali_osk_locks.h31 /* When DEBUG is enabled, this struct will be used to track owner, mode and order checking */
36 _mali_osk_lock_order_t order; member
80 * init/lock/unlock a lock/mutex, we could track lock order of a given tid. */
82 _mali_osk_lock_order_t order);
104 static inline _mali_osk_spinlock_t *_mali_osk_spinlock_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_init() argument
113 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_init()
147 _mali_osk_lock_order_t order) in mali_osk_spinlock_irq_init()
157 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in mali_osk_spinlock_irq_init()
193 static inline _mali_osk_mutex_rw_t *_mali_osk_mutex_rw_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_mutex_rw_init() argument
203 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_mutex_rw_init()
146 mali_osk_spinlock_irq_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) mali_osk_spinlock_irq_init() argument
264 _mali_osk_mutex_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_mutex_init() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_osk_locks.c26 static const char *const lock_order_to_string(_mali_osk_lock_order_t order);
29 void _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_locks_debug_init() argument
35 checker->order = order; in _mali_osk_locks_debug_init()
47 printk(KERN_ERR "%d: ERROR lock %p taken while holding a lock of a higher order.\n", in _mali_osk_locks_debug_add()
68 /* Lock order checking
73 * make sure that a lock that is taken has a higher order than the current highest-order lock a
81 * lock taken by this thread. If the last lock's order was lower than the
98 printk(" [lock: %p, tid_owner: %d, order in dump_lock_tracking_list()
228 lock_order_to_string(_mali_osk_lock_order_t order) lock_order_to_string() argument
[all...]
H A Dmali_osk_locks.h30 /* When DEBUG is enabled, this struct will be used to track owner, mode and order checking */
35 _mali_osk_lock_order_t order; member
79 * init/lock/unlock a lock/mutex, we could track lock order of a given tid. */
80 void _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order);
96 static inline _mali_osk_spinlock_t *_mali_osk_spinlock_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_init() argument
105 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_init()
138 static inline _mali_osk_spinlock_irq_t *_mali_osk_spinlock_irq_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_irq_init() argument
149 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_irq_init()
185 static inline _mali_osk_mutex_rw_t *_mali_osk_mutex_rw_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_mutex_rw_init() argument
196 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_mutex_rw_init()
257 _mali_osk_mutex_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_mutex_init() argument
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_pool.c130 (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_sync_page()
138 for (i = 0; i < (1U << pool->order); i++) in kbase_mem_pool_zero_page()
162 /* don't warn on higher order failures */ in kbase_mem_alloc_page()
163 if (pool->order) in kbase_mem_alloc_page()
167 pool->group_id, gfp, pool->order); in kbase_mem_alloc_page()
171 dma_addr = dma_map_page(dev, p, 0, (PAGE_SIZE << pool->order), in kbase_mem_alloc_page()
176 pool->group_id, p, pool->order); in kbase_mem_alloc_page()
181 for (i = 0; i < (1u << pool->order); i++) in kbase_mem_alloc_page()
195 dma_unmap_page(dev, dma_addr, (PAGE_SIZE << pool->order), in kbase_mem_pool_free_page()
197 for (i = 0; i < (1u << pool->order); in kbase_mem_pool_free_page()
358 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument
[all...]
H A Dmali_kbase_native_mgm.c36 * @order: Page order for physical page size (order=0 means 4 KiB,
37 * order=9 means 2 MiB).
46 gfp_t gfp_mask, unsigned int order) in kbase_native_mgm_alloc()
64 return alloc_pages(gfp_mask, order); in kbase_native_mgm_alloc()
76 * @order: Page order for physical page size (order=0 means 4 KiB,
77 * order
44 kbase_native_mgm_alloc( struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, unsigned int order) kbase_native_mgm_alloc() argument
81 kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, unsigned int order) kbase_native_mgm_free() argument
[all...]
H A Dmali_kbase_sync_android.c48 int order; member
74 new_mpt->order = mpt->order; in timeline_dup()
87 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled()
100 int diff = ma->order - mb->order; in timeline_compare()
120 snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str()
194 * Sync points must be triggered in *exactly* the same order as they are
208 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc()
301 * Sync points must be triggered in *exactly* the same order a
[all...]
/device/soc/rockchip/common/vendor/drivers/dma-buf/heaps/
H A Dpage_pool.c26 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages()
31 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages()
48 mod_node_page_state(page_pgdat(page), NR_KERNEL_MISC_RECLAIMABLE, 1 << pool->order); in dmabuf_page_pool_add()
60 mod_node_page_state(page_pgdat(page), NR_KERNEL_MISC_RECLAIMABLE, -(1 << pool->order)); in dmabuf_page_pool_remove()
97 if (WARN_ON(pool->order != compound_order(page))) { in dmabuf_page_pool_free()
113 return count << pool->order; in dmabuf_page_pool_total()
116 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order) in dmabuf_page_pool_create() argument
130 pool->order = order; in dmabuf_page_pool_create()
191 freed += (1 << pool->order); in dmabuf_page_pool_do_shrink()
[all...]
H A Dpage_pool.h35 * @order: order of pages in the pool
45 unsigned int order; member
49 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order);
/device/soc/rockchip/rk3588/kernel/drivers/dma-buf/heaps/
H A Dpage_pool.c26 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages()
32 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages()
48 1 << pool->order); in dmabuf_page_pool_add()
62 -(1 << pool->order)); in dmabuf_page_pool_remove()
97 if (WARN_ON(pool->order != compound_order(page))) in dmabuf_page_pool_free()
111 return count << pool->order; in dmabuf_page_pool_total()
114 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order) in dmabuf_page_pool_create() argument
127 pool->order = order; in dmabuf_page_pool_create()
184 freed += (1 << pool->order); in dmabuf_page_pool_do_shrink()
[all...]
H A Dpage_pool.h35 * @order: order of pages in the pool
45 unsigned int order; member
50 unsigned int order);
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_native_mgm.c37 * @order: Page order for physical page size (order=0 means 4 KiB,
38 * order=9 means 2 MiB).
46 unsigned int order) in kbase_native_mgm_alloc()
62 return alloc_pages(gfp_mask, order); in kbase_native_mgm_alloc()
74 * @order: Page order for physical page size (order=0 means 4 KiB,
75 * order
45 kbase_native_mgm_alloc(struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, unsigned int order) kbase_native_mgm_alloc() argument
79 kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, unsigned int order) kbase_native_mgm_free() argument
[all...]
H A Dmali_kbase_mem_pool.c124 dma_sync_single_for_device(dev, kbase_dma_addr(p), (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_sync_page()
131 for (i = 0; i < (1U << pool->order); i++) { in kbase_mem_pool_zero_page()
162 /* don't warn on higher order failures */ in kbase_mem_alloc_page()
163 if (pool->order) { in kbase_mem_alloc_page()
167 p = kbdev->mgm_dev->ops.mgm_alloc_page(kbdev->mgm_dev, pool->group_id, gfp, pool->order); in kbase_mem_alloc_page()
171 dma_addr = dma_map_page(dev, p, 0, (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_alloc_page()
173 kbdev->mgm_dev->ops.mgm_free_page(kbdev->mgm_dev, pool->group_id, p, pool->order); in kbase_mem_alloc_page()
178 for (i = 0; i < (1u << pool->order); i++) { in kbase_mem_alloc_page()
192 dma_unmap_page(dev, dma_addr, (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_free_page()
193 for (i = 0; i < (1u << pool->order); in kbase_mem_pool_free_page()
361 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument
[all...]
H A Dmali_kbase_sync_android.c49 int order; member
83 new_mpt->order = mpt->order; in timeline_dup()
95 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled()
108 int diff = ma->order - mb->order; in timeline_compare()
128 (void)snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str()
203 * Sync points must be triggered in *exactly* the same order as they are
217 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc()
331 * Sync points must be triggered in *exactly* the same order a
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_sync_android.c42 int order; member
76 new_mpt->order = mpt->order; in timeline_dup()
88 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled()
101 int diff = ma->order - mb->order; in timeline_compare()
121 snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str()
196 * Sync points must be triggered in *exactly* the same order as they are
210 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc()
324 * Sync points must be triggered in *exactly* the same order a
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_sync_android.c44 int order; member
79 new_mpt->order = mpt->order; in timeline_dup()
92 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled()
105 int diff = ma->order - mb->order; in timeline_compare()
125 snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str()
198 * Sync points must be triggered in *exactly* the same order as they are
212 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc()
324 * Sync points must be triggered in *exactly* the same order a
[all...]
/device/soc/rockchip/rk3588/kernel/include/linux/
H A Dprotected_memory_allocator.h31 * @order: Size of memory allocation in pages, as a base-2 logarithm.
35 unsigned int order; member
54 * @order: How many pages to allocate, as a base-2 logarithm.
60 unsigned int order);
96 * In order for a system integrator to provide custom behaviors for protected
H A Dpage_pool.h35 * @order: order of pages in the pool
45 unsigned int order; member
50 unsigned int order);
H A Dmemory_group_manager.h59 * @order: Page order for physical page size (order=0 means 4 KiB,
60 * order=9 means 2 MiB).
66 gfp_t gfp_mask, unsigned int order);
79 * @group_id and @order.
80 * @order: Page order for physical page size (order=0 means 4 KiB,
81 * order
[all...]
/device/soc/rockchip/common/vendor/include/
H A Dpage_pool.h35 * @order: order of pages in the pool
45 unsigned int order; member
49 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order);
/device/soc/rockchip/common/sdk_linux/include/linux/
H A Dmemory_group_manager.h59 * @order: Page order for physical page size (order=0 means 4 KiB,
60 * order=9 means 2 MiB).
65 unsigned int order);
78 * @group_id and @order.
79 * @order: Page order for physical page size (order=0 means 4 KiB,
80 * order
[all...]
H A Ddma-map-ops.h121 unsigned int order, bool no_warn);
145 unsigned int order, in dma_alloc_from_contiguous()
181 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr);
187 int dma_release_from_global_coherent(int order, void *vaddr);
200 #define dma_release_from_dev_coherent(dev, order, vaddr) (0)
201 #define dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0)
209 static inline int dma_release_from_global_coherent(int order, void *vaddr) in dma_release_from_global_coherent() argument
143 dma_alloc_from_contiguous(struct device *dev, size_t count, unsigned int order, bool no_warn) dma_alloc_from_contiguous() argument
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/osal/include/
H A Dosal_mmz.h67 unsigned int order; member
159 unsigned int __phys_addr_align__ order; member
248 unsigned long gfp, const char *mmz_name, unsigned int order);
266 hil_mmz_t *_user_mmz, unsigned int order);
/device/soc/hisilicon/hi3516dv300/sdk_linux/include/
H A Dosal_mmz.h64 unsigned int order; member
156 unsigned int __phys_addr_align__ order; member
245 unsigned long gfp, const char *mmz_name, unsigned int order);
263 hil_mmz_t *_user_mmz, unsigned int order);
/device/soc/hisilicon/hi3516dv300/sdk_liteos/include/
H A Dosal_mmz.h65 unsigned int order; member
157 unsigned int __phys_addr_align__ order; member
231 unsigned long gfp, const char *mmz_name, unsigned int order);
249 hil_mmz_t *_user_mmz, unsigned int order);

Completed in 12 milliseconds

1234