Home
last modified time | relevance | path

Searched refs:imported (Results 1 - 23 of 23) sorted by relevance

/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.c881 stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
882 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
883 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
884 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
886 alloc->imported.alias.aliased[i].alloc->pages + in kbase_gpu_mmap()
887 alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
888 alloc->imported.alias.aliased[i].length, reg->flags); in kbase_gpu_mmap()
893 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
897 alloc->imported.alias.aliased[i].length, (reg->flags & mask) | attr); in kbase_gpu_mmap()
918 stride = reg->gpu_alloc->imported in kbase_gpu_mmap()
[all...]
H A Dmali_kbase_mem.h163 } imported; member
167 * used to signify that a buffer was pinned when it was imported. Since the
357 alloc_size += nr_pages * sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
358 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
393 alloc->imported.user_buf.dma_addrs = (void *)(alloc->pages + nr_pages); in kbase_alloc_create()
412 reg->cpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
416 reg->gpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
H A Dmali_kbase_mem_linux.c280 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query()
281 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) { in kbase_mem_query()
469 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_mark_reclaim()
485 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_unmark_reclaim()
501 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_make()
525 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_unmake()
639 /* limit to imported memory */ in kbase_mem_flags_change()
656 /* Currently supporting only imported memory */ in kbase_mem_flags_change()
670 WARN_ON(reg->gpu_alloc->imported.umm.current_mapping_usage_count); in kbase_mem_flags_change()
750 reg->gpu_alloc->imported in kbase_mem_from_ump()
[all...]
H A Dmali_kbase_softjobs.c173 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time()
648 unsigned long nr_pages2 = alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare()
650 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare()
660 ret = get_user_pages_fast(alloc->imported.user_buf.address, nr_pages2, 0, buffers[i].extres_pages); in kbase_debug_copy_prepare()
741 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page()
810 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
H A Dmali_kbase_jd.c402 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.c909 stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
910 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
911 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
912 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
915 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
916 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
921 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
926 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
949 stride = reg->gpu_alloc->imported in kbase_gpu_mmap()
[all...]
H A Dmali_kbase_mem_linux.c287 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query()
288 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) in kbase_mem_query()
472 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_mark_reclaim()
492 struct kbase_context *kctx = alloc->imported.kctx; in kbase_mem_evictable_unmark_reclaim()
511 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_make()
536 struct kbase_context *kctx = gpu_alloc->imported.kctx; in kbase_mem_evictable_unmake()
644 /* limit to imported memory */ in kbase_mem_flags_change()
660 /* Currently supporting only imported memory */ in kbase_mem_flags_change()
673 WARN_ON(reg->gpu_alloc->imported.umm.current_mapping_usage_count); in kbase_mem_flags_change()
745 reg->gpu_alloc->imported in kbase_mem_from_ump()
[all...]
H A Dmali_kbase_mem.h165 } imported; member
169 * used to signify that a buffer was pinned when it was imported. Since the
356 sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
357 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
389 alloc->imported.user_buf.dma_addrs = in kbase_alloc_create()
409 reg->cpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
415 reg->gpu_alloc->imported.kctx = kctx; in kbase_reg_prepare_native()
H A Dmali_kbase_softjobs.c180 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time()
668 alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare()
670 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare()
682 alloc->imported.user_buf.address, in kbase_debug_copy_prepare()
766 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page()
835 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
H A Dmali_kbase_jd.c421 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c74 * gets imported twice on Mali DDK side and so the problem of sync happening
296 WARN_ON(kctx != kern_mapping->cpu_alloc->imported.native.kctx); in kbase_phy_alloc_mapping_put()
583 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query()
584 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) { in kbase_mem_query()
818 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_mark_reclaim()
837 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_unmark_reclaim()
855 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_make()
879 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_unmake()
997 /* limit to imported memory */ in kbase_mem_flags_change()
1018 /* Currently supporting only imported memor in kbase_mem_flags_change()
[all...]
H A Dmali_kbase_mem.c1261 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
1263 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
1264 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
1265 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
1268 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
1269 alloc->imported.alias.aliased[i].length, reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap()
1274 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
1277 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
1295 reg->gpu_alloc->imported in kbase_gpu_mmap()
[all...]
H A Dmali_kbase_trace_gpu_mem.c29 * struct kbase_dma_buf - Object instantiated when a dma-buf imported allocation
34 * @dma_buf: Reference to dma_buf been imported.
36 * @import_count: The number of times the dma_buf was imported.
167 dev_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage()
169 prcs_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage()
198 unique_dev_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage()
200 unique_prcs_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage()
H A Dmali_kbase_mem.h125 * @imported: member in union valid based on @a type
175 } imported; member
179 * used to signify that a buffer was pinned when it was imported. Since the
220 * kbase_mem_is_imported - Indicate whether a memory type is imported
224 * Return: true if the memory type is imported, false otherwise
567 alloc_size += nr_pages * sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
568 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
591 alloc->imported.native.nr_struct_pages = (alloc_size + (PAGE_SIZE - 1)) >> PAGE_SHIFT; in kbase_alloc_create()
592 kbase_process_page_usage_inc(kctx, alloc->imported.native.nr_struct_pages); in kbase_alloc_create()
609 alloc->imported in kbase_alloc_create()
[all...]
H A Dmali_kbase_softjobs.c180 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time()
650 unsigned long nr_pages2 = alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare()
652 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare()
662 ret = get_user_pages_fast(alloc->imported.user_buf.address, nr_pages2, 0, buffers[i].extres_pages); in kbase_debug_copy_prepare()
705 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page()
773 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
H A Dmali_kbase_jd.c308 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c72 * gets imported twice on Mali DDK side and so the problem of sync happening
283 WARN_ON(kctx != kern_mapping->cpu_alloc->imported.native.kctx); in kbase_phy_alloc_mapping_put()
549 aliased = reg->cpu_alloc->imported.alias.aliased; in kbase_mem_query()
550 for (i = 0; i < reg->cpu_alloc->imported.alias.nents; i++) in kbase_mem_query()
758 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_mark_reclaim()
781 struct kbase_context *kctx = alloc->imported.native.kctx; in kbase_mem_evictable_unmark_reclaim()
803 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_make()
829 struct kbase_context *kctx = gpu_alloc->imported.native.kctx; in kbase_mem_evictable_unmake()
953 /* limit to imported memory */ in kbase_mem_flags_change()
972 /* Currently supporting only imported memor in kbase_mem_flags_change()
[all...]
H A Dmali_kbase_mem.c1536 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
1538 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
1539 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
1540 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
1544 alloc->imported.alias.aliased[i] in kbase_gpu_mmap()
1546 alloc->imported.alias.aliased[i] in kbase_gpu_mmap()
1548 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
1561 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
1584 reg->gpu_alloc->imported.umm.current_mapping_usage_count) { in kbase_gpu_mmap()
1585 /* For padded imported dm in kbase_gpu_mmap()
[all...]
H A Dmali_kbase_trace_gpu_mem.c28 * struct kbase_dma_buf - Object instantiated when a dma-buf imported allocation
33 * @dma_buf: Reference to dma_buf been imported.
35 * @import_count: The number of times the dma_buf was imported.
171 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage()
174 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage()
202 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage()
205 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage()
H A Dmali_kbase_mem.h131 * @imported: member in union valid based on @a type
182 } imported; member
186 * used to signify that a buffer was pinned when it was imported. Since the
251 * kbase_mem_is_imported - Indicate whether a memory type is imported
255 * Return: true if the memory type is imported, false otherwise
648 sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
649 per_page_size += sizeof(*alloc->imported.user_buf.dma_addrs); in kbase_alloc_create()
670 alloc->imported.native.nr_struct_pages = in kbase_alloc_create()
673 alloc->imported.native.nr_struct_pages); in kbase_alloc_create()
690 alloc->imported in kbase_alloc_create()
[all...]
H A Dmali_kbase_softjobs.c190 * - security, this is currently how imported user bufs are checked. in kbase_dump_cpu_gpu_time()
677 alloc->imported.user_buf.nr_pages; in kbase_debug_copy_prepare()
679 if (alloc->imported.user_buf.mm != current->mm) { in kbase_debug_copy_prepare()
691 alloc->imported.user_buf.address, in kbase_debug_copy_prepare()
737 struct sg_table *sgt = gpu_alloc->imported.umm.sgt; in dma_buf_kmap_page()
805 struct dma_buf *dma_buf = gpu_alloc->imported.umm.dma_buf; in kbase_mem_copy_from_extres()
H A Dmali_kbase_jd.c330 resv = reg->gpu_alloc->imported.umm.dma_buf->resv; in kbase_jd_pre_external_resources()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_kcpu.c120 reg->gpu_alloc->imported.user_buf.nr_pages) { in kbase_kcpu_unmap_import_prepare_internal()

Completed in 43 milliseconds