/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem.c | 881 stride = alloc->imported.alias.stride;
in kbase_gpu_mmap() 882 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased);
in kbase_gpu_mmap() 883 for (i = 0; i < alloc->imported.alias.nents; i++) {
in kbase_gpu_mmap() 884 if (alloc->imported.alias.aliased[i].alloc) {
in kbase_gpu_mmap() 886 alloc->imported.alias.aliased[i].alloc->pages +
in kbase_gpu_mmap() 887 alloc->imported.alias.aliased[i].offset,
in kbase_gpu_mmap() 888 alloc->imported.alias.aliased[i].length, reg->flags);
in kbase_gpu_mmap() 893 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc);
in kbase_gpu_mmap() 897 alloc->imported.alias.aliased[i].length, (reg->flags & mask) | attr);
in kbase_gpu_mmap() 918 stride = reg->gpu_alloc->imported in kbase_gpu_mmap() [all...] |
H A D | mali_kbase_mem.h | 163 } imported; member 167 * used to signify that a buffer was pinned when it was imported. Since the 357 alloc_size += nr_pages * sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 358 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 393 alloc->imported.user_buf.dma_addrs = (void *)(alloc->pages + nr_pages); in kbase_alloc_create() 412 reg->cpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native() 416 reg->gpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
|
H A D | mali_kbase_mem_linux.c | 280 aliased = reg->cpu_alloc->imported.alias.aliased;
in kbase_mem_query() 281 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) {
in kbase_mem_query() 469 struct kbase_context *kctx = alloc->imported.kctx;
in kbase_mem_evictable_mark_reclaim() 485 struct kbase_context *kctx = alloc->imported.kctx;
in kbase_mem_evictable_unmark_reclaim() 501 struct kbase_context *kctx = gpu_alloc->imported.kctx;
in kbase_mem_evictable_make() 525 struct kbase_context *kctx = gpu_alloc->imported.kctx;
in kbase_mem_evictable_unmake() 639 /* limit to imported memory */
in kbase_mem_flags_change() 656 /* Currently supporting only imported memory */
in kbase_mem_flags_change() 670 WARN_ON(reg->gpu_alloc->imported.umm.current_mapping_usage_count);
in kbase_mem_flags_change() 750 reg->gpu_alloc->imported in kbase_mem_from_ump() [all...] |
H A D | mali_kbase_softjobs.c | 173 * - security, this is currently how imported user bufs are checked.
in kbase_dump_cpu_gpu_time() 648 unsigned long nr_pages2 = alloc->imported.user_buf.nr_pages;
in kbase_debug_copy_prepare() 650 if (alloc->imported.user_buf.mm != current->mm) {
in kbase_debug_copy_prepare() 660 ret = get_user_pages_fast(alloc->imported.user_buf.address, nr_pages2, 0, buffers[i].extres_pages);
in kbase_debug_copy_prepare() 741 struct sg_table *sgt = gpu_alloc->imported.umm.sgt;
in dma_buf_kmap_page() 810 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf;
in kbase_mem_copy_from_extres()
|
H A D | mali_kbase_jd.c | 402 resv = reg->gpu_alloc->imported.umm.dma_buf->resv;
in kbase_jd_pre_external_resources()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_mem.c | 909 stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 910 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 911 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 912 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 915 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap() 916 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 921 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap() 926 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 949 stride = reg->gpu_alloc->imported in kbase_gpu_mmap() [all...] |
H A D | mali_kbase_mem_linux.c | 287 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query() 288 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) in kbase_mem_query() 472 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_mark_reclaim() 492 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_unmark_reclaim() 511 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_make() 536 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_unmake() 644 /* limit to imported memory */ in kbase_mem_flags_change() 660 /* Currently supporting only imported memory */ in kbase_mem_flags_change() 673 WARN_ON(reg->gpu_alloc->imported.umm.current_mapping_usage_count); in kbase_mem_flags_change() 745 reg->gpu_alloc->imported in kbase_mem_from_ump() [all...] |
H A D | mali_kbase_mem.h | 165 } imported; member 169 * used to signify that a buffer was pinned when it was imported. Since the 356 sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 357 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 389 alloc->imported.user_buf.dma_addrs = in kbase_alloc_create() 409 reg->cpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native() 415 reg->gpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
|
H A D | mali_kbase_softjobs.c | 180 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time() 668 alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare() 670 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare() 682 alloc->imported.user_buf.address, in kbase_debug_copy_prepare() 766 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page() 835 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
|
H A D | mali_kbase_jd.c | 421 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_linux.c | 74 * gets imported twice on Mali DDK side and so the problem of sync happening 296 WARN_ON(kctx != kern_mapping->cpu_alloc->imported.native.kctx); in kbase_phy_alloc_mapping_put() 583 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query() 584 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) { in kbase_mem_query() 818 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_mark_reclaim() 837 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_unmark_reclaim() 855 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_make() 879 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_unmake() 997 /* limit to imported memory */ in kbase_mem_flags_change() 1018 /* Currently supporting only imported memor in kbase_mem_flags_change() [all...] |
H A D | mali_kbase_mem.c | 1261 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 1263 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 1264 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 1265 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 1268 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap() 1269 alloc->imported.alias.aliased[i].length, reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap() 1274 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap() 1277 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 1295 reg->gpu_alloc->imported in kbase_gpu_mmap() [all...] |
H A D | mali_kbase_trace_gpu_mem.c | 29 * struct kbase_dma_buf - Object instantiated when a dma-buf imported allocation 34 * @dma_buf: Reference to dma_buf been imported. 36 * @import_count: The number of times the dma_buf was imported. 167 dev_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage() 169 prcs_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage() 198 unique_dev_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage() 200 unique_prcs_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage()
|
H A D | mali_kbase_mem.h | 125 * @imported: member in union valid based on @a type 175 } imported; member 179 * used to signify that a buffer was pinned when it was imported. Since the 220 * kbase_mem_is_imported - Indicate whether a memory type is imported 224 * Return: true if the memory type is imported, false otherwise 567 alloc_size += nr_pages * sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 568 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 591 alloc->imported.native.nr_struct_pages = (alloc_size + (PAGE_SIZE - 1)) >> PAGE_SHIFT; in kbase_alloc_create() 592 kbase_process_page_usage_inc(kctx, alloc->imported.native.nr_struct_pages); in kbase_alloc_create() 609 alloc->imported in kbase_alloc_create() [all...] |
H A D | mali_kbase_softjobs.c | 180 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time() 650 unsigned long nr_pages2 = alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare() 652 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare() 662 ret = get_user_pages_fast(alloc->imported.user_buf.address, nr_pages2, 0, buffers[i].extres_pages); in kbase_debug_copy_prepare() 705 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page() 773 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
|
H A D | mali_kbase_jd.c | 308 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem_linux.c | 72 * gets imported twice on Mali DDK side and so the problem of sync happening 283 WARN_ON(kctx != kern_mapping->cpu_alloc->imported.native.kctx); in kbase_phy_alloc_mapping_put() 549 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query() 550 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) in kbase_mem_query() 758 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_mark_reclaim() 781 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_unmark_reclaim() 803 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_make() 829 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_unmake() 953 /* limit to imported memory */ in kbase_mem_flags_change() 972 /* Currently supporting only imported memor in kbase_mem_flags_change() [all...] |
H A D | mali_kbase_mem.c | 1536 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 1538 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 1539 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 1540 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 1544 alloc->imported.alias.aliased[i] in kbase_gpu_mmap() 1546 alloc->imported.alias.aliased[i] in kbase_gpu_mmap() 1548 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 1561 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 1584 reg->gpu_alloc->imported.umm.current_mapping_usage_count) { in kbase_gpu_mmap() 1585 /* For padded imported dm in kbase_gpu_mmap() [all...] |
H A D | mali_kbase_trace_gpu_mem.c | 28 * struct kbase_dma_buf - Object instantiated when a dma-buf imported allocation 33 * @dma_buf: Reference to dma_buf been imported. 35 * @import_count: The number of times the dma_buf was imported. 171 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage() 174 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage() 202 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage() 205 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage()
|
H A D | mali_kbase_mem.h | 131 * @imported: member in union valid based on @a type 182 } imported; member 186 * used to signify that a buffer was pinned when it was imported. Since the 251 * kbase_mem_is_imported - Indicate whether a memory type is imported 255 * Return: true if the memory type is imported, false otherwise 648 sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 649 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create() 670 alloc->imported.native.nr_struct_pages = in kbase_alloc_create() 673 alloc->imported.native.nr_struct_pages); in kbase_alloc_create() 690 alloc->imported in kbase_alloc_create() [all...] |
H A D | mali_kbase_softjobs.c | 190 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time() 677 alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare() 679 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare() 691 alloc->imported.user_buf.address, in kbase_debug_copy_prepare() 737 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page() 805 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
|
H A D | mali_kbase_jd.c | 330 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf_kcpu.c | 120 reg->gpu_alloc->imported.user_buf.nr_pages) { in kbase_kcpu_unmap_import_prepare_internal()
|