/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_osk_locks.c | 26 static const char *const lock_order_to_string(_mali_osk_lock_order_t order); 30 _mali_osk_lock_order_t order) in _mali_osk_locks_debug_init() 36 checker->order = order; in _mali_osk_locks_debug_init() 48 printk(KERN_ERR "%d: ERROR lock %p taken while holding a lock of a higher order.\n", mali_osk_get_tid(), in _mali_osk_locks_debug_add() 67 /* Lock order checking 72 * make sure that a lock that is taken has a higher order than the current highest-order lock a 80 * lock taken by this thread. If the last lock's order was lower than the 97 printk(" [lock: %p, tid_owner: %d, order in dump_lock_tracking_list() 29 _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_locks_debug_init() argument 225 lock_order_to_string(_mali_osk_lock_order_t order) lock_order_to_string() argument [all...] |
H A D | mali_osk_locks.h | 31 /* When DEBUG is enabled, this struct will be used to track owner, mode and order checking */ 36 _mali_osk_lock_order_t order; member 80 * init/lock/unlock a lock/mutex, we could track lock order of a given tid. */ 82 _mali_osk_lock_order_t order); 104 static inline _mali_osk_spinlock_t *_mali_osk_spinlock_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_init() argument 113 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_init() 147 _mali_osk_lock_order_t order) in mali_osk_spinlock_irq_init() 157 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in mali_osk_spinlock_irq_init() 193 static inline _mali_osk_mutex_rw_t *_mali_osk_mutex_rw_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_mutex_rw_init() argument 203 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_mutex_rw_init() 146 mali_osk_spinlock_irq_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) mali_osk_spinlock_irq_init() argument 264 _mali_osk_mutex_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_mutex_init() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_osk_locks.c | 26 static const char *const lock_order_to_string(_mali_osk_lock_order_t order); 29 void _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_locks_debug_init() argument 35 checker->order = order; in _mali_osk_locks_debug_init() 47 printk(KERN_ERR "%d: ERROR lock %p taken while holding a lock of a higher order.\n", in _mali_osk_locks_debug_add() 68 /* Lock order checking 73 * make sure that a lock that is taken has a higher order than the current highest-order lock a 81 * lock taken by this thread. If the last lock's order was lower than the 98 printk(" [lock: %p, tid_owner: %d, order in dump_lock_tracking_list() 228 lock_order_to_string(_mali_osk_lock_order_t order) lock_order_to_string() argument [all...] |
H A D | mali_osk_locks.h | 30 /* When DEBUG is enabled, this struct will be used to track owner, mode and order checking */ 35 _mali_osk_lock_order_t order; member 79 * init/lock/unlock a lock/mutex, we could track lock order of a given tid. */ 80 void _mali_osk_locks_debug_init(struct _mali_osk_lock_debug_s *checker, _mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order); 96 static inline _mali_osk_spinlock_t *_mali_osk_spinlock_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_init() argument 105 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_init() 138 static inline _mali_osk_spinlock_irq_t *_mali_osk_spinlock_irq_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_spinlock_irq_init() argument 149 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_spinlock_irq_init() 185 static inline _mali_osk_mutex_rw_t *_mali_osk_mutex_rw_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) in _mali_osk_mutex_rw_init() argument 196 _mali_osk_locks_debug_init((struct _mali_osk_lock_debug_s *)lock, flags, order); in _mali_osk_mutex_rw_init() 257 _mali_osk_mutex_init(_mali_osk_lock_flags_t flags, _mali_osk_lock_order_t order) _mali_osk_mutex_init() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_pool.c | 130 (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_sync_page() 138 for (i = 0; i < (1U << pool->order); i++) in kbase_mem_pool_zero_page() 162 /* don't warn on higher order failures */ in kbase_mem_alloc_page() 163 if (pool->order) in kbase_mem_alloc_page() 167 pool->group_id, gfp, pool->order); in kbase_mem_alloc_page() 171 dma_addr = dma_map_page(dev, p, 0, (PAGE_SIZE << pool->order), in kbase_mem_alloc_page() 176 pool->group_id, p, pool->order); in kbase_mem_alloc_page() 181 for (i = 0; i < (1u << pool->order); i++) in kbase_mem_alloc_page() 195 dma_unmap_page(dev, dma_addr, (PAGE_SIZE << pool->order), in kbase_mem_pool_free_page() 197 for (i = 0; i < (1u << pool->order); in kbase_mem_pool_free_page() 358 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument [all...] |
H A D | mali_kbase_native_mgm.c | 36 * @order: Page order for physical page size (order=0 means 4 KiB, 37 * order=9 means 2 MiB). 46 gfp_t gfp_mask, unsigned int order) in kbase_native_mgm_alloc() 64 return alloc_pages(gfp_mask, order); in kbase_native_mgm_alloc() 76 * @order: Page order for physical page size (order=0 means 4 KiB, 77 * order 44 kbase_native_mgm_alloc( struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, unsigned int order) kbase_native_mgm_alloc() argument 81 kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, unsigned int order) kbase_native_mgm_free() argument [all...] |
H A D | mali_kbase_sync_android.c | 48 int order; member 74 new_mpt->order = mpt->order; in timeline_dup() 87 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled() 100 int diff = ma->order - mb->order; in timeline_compare() 120 snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str() 194 * Sync points must be triggered in *exactly* the same order as they are 208 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc() 301 * Sync points must be triggered in *exactly* the same order a [all...] |
/device/soc/rockchip/common/vendor/drivers/dma-buf/heaps/ |
H A D | page_pool.c | 26 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages() 31 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages() 48 mod_node_page_state(page_pgdat(page), NR_KERNEL_MISC_RECLAIMABLE, 1 << pool->order); in dmabuf_page_pool_add() 60 mod_node_page_state(page_pgdat(page), NR_KERNEL_MISC_RECLAIMABLE, -(1 << pool->order)); in dmabuf_page_pool_remove() 97 if (WARN_ON(pool->order != compound_order(page))) { in dmabuf_page_pool_free() 113 return count << pool->order; in dmabuf_page_pool_total() 116 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order) in dmabuf_page_pool_create() argument 130 pool->order = order; in dmabuf_page_pool_create() 191 freed += (1 << pool->order); in dmabuf_page_pool_do_shrink() [all...] |
H A D | page_pool.h | 35 * @order: order of pages in the pool 45 unsigned int order; member 49 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order);
|
/device/soc/rockchip/rk3588/kernel/drivers/dma-buf/heaps/ |
H A D | page_pool.c | 26 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages() 32 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages() 48 1 << pool->order); in dmabuf_page_pool_add() 62 -(1 << pool->order)); in dmabuf_page_pool_remove() 97 if (WARN_ON(pool->order != compound_order(page))) in dmabuf_page_pool_free() 111 return count << pool->order; in dmabuf_page_pool_total() 114 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order) in dmabuf_page_pool_create() argument 127 pool->order = order; in dmabuf_page_pool_create() 184 freed += (1 << pool->order); in dmabuf_page_pool_do_shrink() [all...] |
H A D | page_pool.h | 35 * @order: order of pages in the pool 45 unsigned int order; member 50 unsigned int order);
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_native_mgm.c | 37 * @order: Page order for physical page size (order=0 means 4 KiB, 38 * order=9 means 2 MiB). 46 unsigned int order) in kbase_native_mgm_alloc() 62 return alloc_pages(gfp_mask, order); in kbase_native_mgm_alloc() 74 * @order: Page order for physical page size (order=0 means 4 KiB, 75 * order 45 kbase_native_mgm_alloc(struct memory_group_manager_device *mgm_dev, int group_id, gfp_t gfp_mask, unsigned int order) kbase_native_mgm_alloc() argument 79 kbase_native_mgm_free(struct memory_group_manager_device *mgm_dev, int group_id, struct page *page, unsigned int order) kbase_native_mgm_free() argument [all...] |
H A D | mali_kbase_mem_pool.c | 124 dma_sync_single_for_device(dev, kbase_dma_addr(p), (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_sync_page() 131 for (i = 0; i < (1U << pool->order); i++) { in kbase_mem_pool_zero_page() 162 /* don't warn on higher order failures */ in kbase_mem_alloc_page() 163 if (pool->order) { in kbase_mem_alloc_page() 167 p = kbdev->mgm_dev->ops.mgm_alloc_page(kbdev->mgm_dev, pool->group_id, gfp, pool->order); in kbase_mem_alloc_page() 171 dma_addr = dma_map_page(dev, p, 0, (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_alloc_page() 173 kbdev->mgm_dev->ops.mgm_free_page(kbdev->mgm_dev, pool->group_id, p, pool->order); in kbase_mem_alloc_page() 178 for (i = 0; i < (1u << pool->order); i++) { in kbase_mem_alloc_page() 192 dma_unmap_page(dev, dma_addr, (PAGE_SIZE << pool->order), DMA_BIDIRECTIONAL); in kbase_mem_pool_free_page() 193 for (i = 0; i < (1u << pool->order); in kbase_mem_pool_free_page() 361 kbase_mem_pool_init(struct kbase_mem_pool *pool, const struct kbase_mem_pool_config *config, unsigned int order, int group_id, struct kbase_device *kbdev, struct kbase_mem_pool *next_pool) kbase_mem_pool_init() argument [all...] |
H A D | mali_kbase_sync_android.c | 49 int order; member 83 new_mpt->order = mpt->order; in timeline_dup() 95 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled() 108 int diff = ma->order - mb->order; in timeline_compare() 128 (void)snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str() 203 * Sync points must be triggered in *exactly* the same order as they are 217 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc() 331 * Sync points must be triggered in *exactly* the same order a [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_sync_android.c | 42 int order;
member 76 new_mpt->order = mpt->order;
in timeline_dup() 88 int diff = atomic_read(&mtl->signaled) - mpt->order;
in timeline_has_signaled() 101 int diff = ma->order - mb->order;
in timeline_compare() 121 snprintf(str, size, "%d(%d)", mpt->order, mpt->result);
in pt_value_str() 196 * Sync points must be triggered in *exactly* the same order as they are
210 mpt->order = atomic_inc_return(&mtl->counter);
in kbase_sync_pt_alloc() 324 * Sync points must be triggered in *exactly* the same order a [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_sync_android.c | 44 int order; member 79 new_mpt->order = mpt->order; in timeline_dup() 92 int diff = atomic_read(&mtl->signaled) - mpt->order; in timeline_has_signaled() 105 int diff = ma->order - mb->order; in timeline_compare() 125 snprintf(str, size, "%d(%d)", mpt->order, mpt->result); in pt_value_str() 198 * Sync points must be triggered in *exactly* the same order as they are 212 mpt->order = atomic_inc_return(&mtl->counter); in kbase_sync_pt_alloc() 324 * Sync points must be triggered in *exactly* the same order a [all...] |
/device/soc/rockchip/rk3588/kernel/include/linux/ |
H A D | protected_memory_allocator.h | 31 * @order: Size of memory allocation in pages, as a base-2 logarithm. 35 unsigned int order; member 54 * @order: How many pages to allocate, as a base-2 logarithm. 60 unsigned int order); 96 * In order for a system integrator to provide custom behaviors for protected
|
H A D | page_pool.h | 35 * @order: order of pages in the pool 45 unsigned int order; member 50 unsigned int order);
|
H A D | memory_group_manager.h | 59 * @order: Page order for physical page size (order=0 means 4 KiB, 60 * order=9 means 2 MiB). 66 gfp_t gfp_mask, unsigned int order); 79 * @group_id and @order. 80 * @order: Page order for physical page size (order=0 means 4 KiB, 81 * order [all...] |
/device/soc/rockchip/common/vendor/include/ |
H A D | page_pool.h | 35 * @order: order of pages in the pool 45 unsigned int order; member 49 struct dmabuf_page_pool *dmabuf_page_pool_create(gfp_t gfp_mask, unsigned int order);
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | memory_group_manager.h | 59 * @order: Page order for physical page size (order=0 means 4 KiB, 60 * order=9 means 2 MiB). 65 unsigned int order); 78 * @group_id and @order. 79 * @order: Page order for physical page size (order=0 means 4 KiB, 80 * order [all...] |
H A D | dma-map-ops.h | 121 unsigned int order, bool no_warn); 145 unsigned int order, in dma_alloc_from_contiguous() 181 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr); 187 int dma_release_from_global_coherent(int order, void *vaddr); 200 #define dma_release_from_dev_coherent(dev, order, vaddr) (0) 201 #define dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0) 209 static inline int dma_release_from_global_coherent(int order, void *vaddr) in dma_release_from_global_coherent() argument 143 dma_alloc_from_contiguous(struct device *dev, size_t count, unsigned int order, bool no_warn) dma_alloc_from_contiguous() argument
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/osal/include/ |
H A D | osal_mmz.h | 67 unsigned int order; member 159 unsigned int __phys_addr_align__ order; member 248 unsigned long gfp, const char *mmz_name, unsigned int order); 266 hil_mmz_t *_user_mmz, unsigned int order);
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/include/ |
H A D | osal_mmz.h | 64 unsigned int order; member 156 unsigned int __phys_addr_align__ order; member 245 unsigned long gfp, const char *mmz_name, unsigned int order); 263 hil_mmz_t *_user_mmz, unsigned int order);
|
/device/soc/hisilicon/hi3516dv300/sdk_liteos/include/ |
H A D | osal_mmz.h | 65 unsigned int order; member 157 unsigned int __phys_addr_align__ order; member 231 unsigned long gfp, const char *mmz_name, unsigned int order); 249 hil_mmz_t *_user_mmz, unsigned int order);
|