Home
last modified time | relevance | path

Searched refs:chunk (Results 1 - 13 of 13) sorted by relevance

/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_tiler_heap.c29 * encode_chunk_ptr - Encode the address and size of a chunk as an integer.
31 * @chunk_size: Size of a tiler heap chunk, in bytes.
32 * @chunk_addr: GPU virtual address of the same tiler heap chunk.
34 * The size and address of the next chunk in a list are packed into a single
35 * 64-bit value for storage in a chunk's header. This function returns that
38 * Return: Next chunk pointer suitable for writing into a chunk header.
60 * get_last_chunk - Get the last chunk of a tiler heap
64 * Return: The address of the most recently-linked chunk, or NULL if none.
77 * link_chunk - Link a chunk int
89 link_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk) link_chunk() argument
132 init_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk, bool link_with_prev) init_chunk() argument
185 struct kbase_csf_tiler_heap_chunk *chunk = NULL; create_chunk() local
250 delete_chunk(struct kbase_csf_tiler_heap *const heap, struct kbase_csf_tiler_heap_chunk *const chunk) delete_chunk() argument
277 struct kbase_csf_tiler_heap_chunk *chunk = list_entry( delete_all_chunks() local
[all...]
H A Dmali_kbase_csf_tiler_heap_debugfs.c42 struct kbase_csf_tiler_heap_chunk *chunk; in kbasep_csf_tiler_heap_debugfs_show() local
58 list_for_each_entry(chunk, &heap->chunks_list, link) in kbasep_csf_tiler_heap_debugfs_show()
60 chunk->gpu_va); in kbasep_csf_tiler_heap_debugfs_show()
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/extend/
H A Dext_hash.c63 static hi_s32 ext_hash_update_from_user(mbedtls_md_context_t *md, hi_u8 *chunk, hi_u32 chunk_len) in ext_hash_update_from_user() argument
81 ret = crypto_copy_from_user(ptr, HASH_MAX_BUFFER_SIZE, chunk + offset, len); in ext_hash_update_from_user()
107 hi_s32 mbedtls_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in mbedtls_hash_update() argument
122 ret = ext_hash_update_from_user(md, chunk, chunk_len); in mbedtls_hash_update()
128 ret = mbedtls_md_update(md, chunk, chunk_len); in mbedtls_hash_update()
H A Dext_sm3.c323 hi_s32 ext_sm3_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in ext_sm3_update() argument
351 crypto_chk_err_goto(crypto_copy_from_user(ptr, HASH_MAX_BUFFER_SIZE, chunk + offset, length)); in ext_sm3_update()
357 if (chunk == HI_NULL) { in ext_sm3_update()
358 hi_log_error("Invalid point, chunk is null!\n"); in ext_sm3_update()
362 crypto_chk_err_goto(sm3_update(ctx, chunk, chunk_len)); in ext_sm3_update()
/device/soc/rockchip/rk3568/hardware/omx_il/osal/
H A DRockchip_OSAL_ETC.c45 size_t const chunk = 512; in getline() local
47 size_t defaultBufferSize = chunk + 1; in getline()
75 i = fread(pCurrentPointer, 1, chunk, pStream); in getline()
76 if (i < chunk && ferror(pStream)) { in getline()
110 i = ((readByte + (chunk * 2)) / chunk) * chunk; // 2:byte alignment in getline()
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/crypto/
H A Dcryp_hash.c318 static hi_s32 cryp_hash_chunk_copy(const hi_void *chunk, hi_u32 chunk_len, in cryp_hash_chunk_copy() argument
331 hi_log_chk_param_return(chunk == HI_NULL); in cryp_hash_chunk_copy()
335 if (memcpy_s(dma, dma_len, chunk, chunk_len) != EOK) { in cryp_hash_chunk_copy()
340 ret = crypto_copy_from_user(dma, dma_len, chunk, chunk_len); in cryp_hash_chunk_copy()
522 static hi_s32 cryp_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src) in cryp_hash_update() argument
546 ret = cryp_hash_chunk_copy(chunk, inverse_len, hisi_ctx->tail + hisi_ctx->tail_len, inverse_len, src); in cryp_hash_update()
552 chunk += inverse_len; in cryp_hash_update()
568 /* new chunk len align at block size */ in cryp_hash_update()
572 ret = cryp_hash_chunk_copy(chunk + chunk_len, hisi_ctx->tail_len, hisi_ctx->tail, hisi_ctx->tail_len, src); in cryp_hash_update()
582 ret = cryp_hash_process(hisi_ctx, chunk, chunk_le in cryp_hash_update()
[all...]
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/extend/include/
H A Dext_alg.h272 * brief Hash message chunk calculation
275 * which get from cryp_hash_get_cpu_addr, and the length of message chunk
279 * param chunk hash message to update
283 hi_s32 mbedtls_hash_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src);
304 hi_s32 ext_sm3_update(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src);
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/interdrv/common/cipher/src/drv/cipher_v1.0/drivers/crypto/include/
H A Dcryp_hash.h47 * \brief Hash message chunk calculation
50 * which get from cryp_hash_get_cpu_addr, and the length of message chunk
54 * \param chunk hash message to update
58 typedef hi_s32 (*func_hash_update)(hi_void *ctx, const hi_u8 *chunk, hi_u32 chunk_len, hash_chunk_src src);
/device/soc/rockchip/common/sdk_linux/include/linux/
H A Ddmaengine.h90 * A chunk is collection of contiguous bytes to be transferred.
91 * The gap(in bytes) between two chunks is called inter-chunk-gap(ICG).
93 * A FRAME is the smallest series of contiguous {chunk,icg} pairs,
115 * chunk and before first src/dst address for next chunk.
119 * chunk and before the first dst address for next chunk.
122 * chunk and before the first src address for next chunk.
135 * @src_start: Bus address of source for the first chunk
1285 dmaengine_get_dst_icg(struct dma_interleaved_template *xt, struct data_chunk *chunk) dmaengine_get_dst_icg() argument
1290 dmaengine_get_src_icg(struct dma_interleaved_template *xt, struct data_chunk *chunk) dmaengine_get_src_icg() argument
[all...]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_softjobs.c700 size_t chunk = PAGE_SIZE - offset; in kbase_mem_copy_from_extres_page() local
710 chunk = min(chunk, *to_copy); in kbase_mem_copy_from_extres_page()
712 memcpy(target_page + offset, extres_page, chunk); in kbase_mem_copy_from_extres_page()
713 *to_copy -= chunk; in kbase_mem_copy_from_extres_page()
731 chunk = min(offset, *to_copy); in kbase_mem_copy_from_extres_page()
732 memcpy(target_page, extres_page + PAGE_SIZE - offset, chunk); in kbase_mem_copy_from_extres_page()
733 *to_copy -= chunk; in kbase_mem_copy_from_extres_page()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_softjobs.c725 size_t chunk = PAGE_SIZE-offset; in kbase_mem_copy_from_extres_page() local
735 chunk = min(chunk, *to_copy); in kbase_mem_copy_from_extres_page()
737 memcpy(target_page + offset, extres_page, chunk); in kbase_mem_copy_from_extres_page()
738 *to_copy -= chunk; in kbase_mem_copy_from_extres_page()
755 chunk = min(offset, *to_copy); in kbase_mem_copy_from_extres_page()
756 memcpy(target_page, extres_page + PAGE_SIZE-offset, chunk); in kbase_mem_copy_from_extres_page()
757 *to_copy -= chunk; in kbase_mem_copy_from_extres_page()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem.c1956 /* only if we actually have a chunk left <512. If more it indicates in kbase_alloc_phy_pages_helper()
2132 /* only if we actually have a chunk left <512. If more it in kbase_alloc_phy_pages_helper_locked()
2607 /* To have an allocation lie within a 4GB chunk is required only for in kbase_check_alloc_flags()
3209 * 'extent' sized chunk, but didn't actually write to it, so we in kbase_mem_jit_trim_pages_from_region()
3223 * chunk, hence automatically bringing the backed size up to in kbase_mem_jit_trim_pages_from_region()
4253 size_t chunk = PAGE_SIZE - offset; in kbase_mem_copy_to_pinned_user_pages() local
4260 chunk = min(chunk, *to_copy); in kbase_mem_copy_to_pinned_user_pages()
4262 memcpy(target_page + offset, src_page, chunk); in kbase_mem_copy_to_pinned_user_pages()
4263 *to_copy -= chunk; in kbase_mem_copy_to_pinned_user_pages()
[all...]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem.c2288 /* only if we actually have a chunk left <512. If more it indicates in kbase_alloc_phy_pages_helper()
2484 /* only if we actually have a chunk left <512. If more it in kbase_alloc_phy_pages_helper_locked()
3013 /* To have an allocation lie within a 4GB chunk is required only for in kbase_check_alloc_flags()
3636 * 'extension' sized chunk, but didn't actually write to it, so we in kbase_mem_jit_trim_pages_from_region()
3650 * chunk, hence automatically bringing the backed size up to in kbase_mem_jit_trim_pages_from_region()
4746 size_t chunk = PAGE_SIZE-offset; in kbase_mem_copy_to_pinned_user_pages() local
4753 chunk = min(chunk, *to_copy); in kbase_mem_copy_to_pinned_user_pages()
4755 memcpy(target_page + offset, src_page, chunk); in kbase_mem_copy_to_pinned_user_pages()
4756 *to_copy -= chunk; in kbase_mem_copy_to_pinned_user_pages()
[all...]

Completed in 23 milliseconds