/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf_tiler_heap.c | 29 * encode_chunk_ptr - Encode the address and size of a chunk as an integer. 31 * @chunk_size: Size of a tiler heap chunk, in bytes. 32 * @chunk_addr: GPU virtual address of the same tiler heap chunk. 34 * The size and address of the next chunk in a list are packed into a single 35 * 64-bit value for storage in a chunk's header. This function returns that 38 * Return: Next chunk pointer suitable for writing into a chunk header. 60 * get_last_chunk - Get the last chunk of a tiler heap 64 * Return: The address of the most recently-linked chunk, or NULL if none. 77 * link_chunk - Link a chunk int 89 link_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk) link_chunk() argument 132 init_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk, bool link_with_prev) init_chunk() argument 185 struct kbase_csf_tiler_heap_chunk *chunk = NULL; create_chunk() local 250 delete_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk) delete_chunk() argument 277 struct kbase_csf_tiler_heap_chunk *chunk = list_entry( delete_all_chunks() local [all...] |
H A D | mali_kbase_csf_tiler_heap_debugfs.c | 42 struct kbase_csf_tiler_heap_chunk *chunk; in kbasep_csf_tiler_heap_debugfs_show() local 58 list_for_each_entry(chunk, &heap->chunks_list, link) in kbasep_csf_tiler_heap_debugfs_show() 60 chunk->gpu_va); in kbasep_csf_tiler_heap_debugfs_show()
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/extend/ |
H A D | ext_hash.c | 63 static hi_s32 ext_hash_update_from_user(mbedtls_md_context_t *md, hi_u8 *chunk, hi_u32 chunk_len) in ext_hash_update_from_user() argument 81 ret = crypto_copy_from_user(ptr, HASH_MAX_BUFFER_SIZE, chunk + offset, len); in ext_hash_update_from_user() 107 hi_s32 mbedtls_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in mbedtls_hash_update() argument 122 ret = ext_hash_update_from_user(md, chunk, chunk_len); in mbedtls_hash_update() 128 ret = mbedtls_md_update(md, chunk, chunk_len); in mbedtls_hash_update()
|
H A D | ext_sm3.c | 323 hi_s32 ext_sm3_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in ext_sm3_update() argument 351 crypto_chk_err_goto(crypto_copy_from_user(ptr, HASH_MAX_BUFFER_SIZE, chunk + offset, length)); in ext_sm3_update() 357 if (chunk == HI_NULL) { in ext_sm3_update() 358 hi_log_error("Invalid point, chunk is null!\n"); in ext_sm3_update() 362 crypto_chk_err_goto(sm3_update(ctx, chunk, chunk_len)); in ext_sm3_update()
|
/device/soc/rockchip/rk3568/hardware/omx_il/osal/ |
H A D | Rockchip_OSAL_ETC.c | 45 size_t const chunk = 512; in getline() local 47 size_t defaultBufferSize = chunk + 1; in getline() 75 i = fread(pCurrentPointer, 1, chunk, pStream); in getline() 76 if (i < chunk && ferror(pStream)) { in getline() 110 i = ((readByte + (chunk * 2)) / chunk) * chunk; // 2:byte alignment in getline()
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/crypto/ |
H A D | cryp_hash.c | 318 static hi_s32 cryp_hash_chunk_copy(const hi_void *chunk, hi_u32 chunk_len, in cryp_hash_chunk_copy() argument 331 hi_log_chk_param_return(chunk == HI_NULL); in cryp_hash_chunk_copy() 335 if (memcpy_s(dma, dma_len, chunk, chunk_len) != EOK) { in cryp_hash_chunk_copy() 340 ret = crypto_copy_from_user(dma, dma_len, chunk, chunk_len); in cryp_hash_chunk_copy() 522 static hi_s32 cryp_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in cryp_hash_update() argument 546 ret = cryp_hash_chunk_copy(chunk, inverse_len, hisi_ctx->tail + hisi_ctx->tail_len, inverse_len, src); in cryp_hash_update() 552 chunk += inverse_len; in cryp_hash_update() 568 /* new chunk len align at block size */ in cryp_hash_update() 572 ret = cryp_hash_chunk_copy(chunk + chunk_len, hisi_ctx->tail_len, hisi_ctx->tail, hisi_ctx->tail_len, src); in cryp_hash_update() 582 ret = cryp_hash_process(hisi_ctx, chunk, chunk_le in cryp_hash_update() [all...] |
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/extend/include/ |
H A D | ext_alg.h | 272 * brief Hash message chunk calculation 275 * which get from cryp_hash_get_cpu_addr, and the length of message chunk 279 * param chunk hash message to update 283 hi_s32 mbedtls_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src); 304 hi_s32 ext_sm3_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src);
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/crypto/include/ |
H A D | cryp_hash.h | 47 * \brief Hash message chunk calculation 50 * which get from cryp_hash_get_cpu_addr, and the length of message chunk 54 * \param chunk hash message to update 58 typedef hi_s32 (*func_hash_update)(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src);
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | dmaengine.h | 90 * A chunk is collection of contiguous bytes to be transferred. 91 * The gap(in bytes) between two chunks is called inter-chunk-gap(ICG). 93 * A FRAME is the smallest series of contiguous {chunk,icg} pairs, 115 * chunk and before first src/dst address for next chunk. 119 * chunk and before the first dst address for next chunk. 122 * chunk and before the first src address for next chunk. 135 * @src_start: Bus address of source for the first chunk 1285 dmaengine_get_dst_icg(struct dma_interleaved_template *xt, struct data_chunk *chunk) dmaengine_get_dst_icg() argument 1290 dmaengine_get_src_icg(struct dma_interleaved_template *xt, struct data_chunk *chunk) dmaengine_get_src_icg() argument [all...] |
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_softjobs.c | 700 size_t chunk = PAGE_SIZE - offset;
in kbase_mem_copy_from_extres_page() local 710 chunk = min(chunk, *to_copy);
in kbase_mem_copy_from_extres_page() 712 memcpy(target_page + offset, extres_page, chunk);
in kbase_mem_copy_from_extres_page() 713 *to_copy -= chunk;
in kbase_mem_copy_from_extres_page() 731 chunk = min(offset, *to_copy);
in kbase_mem_copy_from_extres_page() 732 memcpy(target_page, extres_page + PAGE_SIZE - offset, chunk);
in kbase_mem_copy_from_extres_page() 733 *to_copy -= chunk;
in kbase_mem_copy_from_extres_page()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_kbase_softjobs.c | 725 size_t chunk = PAGE_SIZE-offset; in kbase_mem_copy_from_extres_page() local 735 chunk = min(chunk, *to_copy); in kbase_mem_copy_from_extres_page() 737 memcpy(target_page + offset, extres_page, chunk); in kbase_mem_copy_from_extres_page() 738 *to_copy -= chunk; in kbase_mem_copy_from_extres_page() 755 chunk = min(offset, *to_copy); in kbase_mem_copy_from_extres_page() 756 memcpy(target_page, extres_page + PAGE_SIZE-offset, chunk); in kbase_mem_copy_from_extres_page() 757 *to_copy -= chunk; in kbase_mem_copy_from_extres_page()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem.c | 1956 /* only if we actually have a chunk left <512. If more it indicates in kbase_alloc_phy_pages_helper() 2132 /* only if we actually have a chunk left <512. If more it in kbase_alloc_phy_pages_helper_locked() 2607 /* To have an allocation lie within a 4GB chunk is required only for in kbase_check_alloc_flags() 3209 * 'extent' sized chunk, but didn't actually write to it, so we in kbase_mem_jit_trim_pages_from_region() 3223 * chunk, hence automatically bringing the backed size up to in kbase_mem_jit_trim_pages_from_region() 4253 size_t chunk = PAGE_SIZE - offset; in kbase_mem_copy_to_pinned_user_pages() local 4260 chunk = min(chunk, *to_copy); in kbase_mem_copy_to_pinned_user_pages() 4262 memcpy(target_page + offset, src_page, chunk); in kbase_mem_copy_to_pinned_user_pages() 4263 *to_copy -= chunk; in kbase_mem_copy_to_pinned_user_pages() [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_mem.c | 2288 /* only if we actually have a chunk left <512. If more it indicates in kbase_alloc_phy_pages_helper() 2484 /* only if we actually have a chunk left <512. If more it in kbase_alloc_phy_pages_helper_locked() 3013 /* To have an allocation lie within a 4GB chunk is required only for in kbase_check_alloc_flags() 3636 * 'extension' sized chunk, but didn't actually write to it, so we in kbase_mem_jit_trim_pages_from_region() 3650 * chunk, hence automatically bringing the backed size up to in kbase_mem_jit_trim_pages_from_region() 4746 size_t chunk = PAGE_SIZE-offset; in kbase_mem_copy_to_pinned_user_pages() local 4753 chunk = min(chunk, *to_copy); in kbase_mem_copy_to_pinned_user_pages() 4755 memcpy(target_page + offset, src_page, chunk); in kbase_mem_copy_to_pinned_user_pages() 4756 *to_copy -= chunk; in kbase_mem_copy_to_pinned_user_pages() [all...] |