Home
last modified time | relevance | path

Searched refs:SHA512_BLOCK_SIZE (Results 1 - 25 of 71) sorted by relevance

123

/kernel/linux/linux-5.10/include/crypto/
H A Dsha512_base.h61 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update()
67 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update()
71 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update()
80 blocks = len / SHA512_BLOCK_SIZE; in sha512_base_do_update()
81 len %= SHA512_BLOCK_SIZE; in sha512_base_do_update()
85 data += blocks * SHA512_BLOCK_SIZE; in sha512_base_do_update()
98 const int bit_offset = SHA512_BLOCK_SIZE - sizeof(__be64[2]); in sha512_base_do_finalize()
101 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize()
105 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize()
H A Dsha.h24 #define SHA512_BLOCK_SIZE 128 macro
93 u8 buf[SHA512_BLOCK_SIZE];
/kernel/linux/linux-6.6/include/crypto/
H A Dsha512_base.h62 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update()
68 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update()
72 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update()
81 blocks = len / SHA512_BLOCK_SIZE; in sha512_base_do_update()
82 len %= SHA512_BLOCK_SIZE; in sha512_base_do_update()
86 data += blocks * SHA512_BLOCK_SIZE; in sha512_base_do_update()
99 const int bit_offset = SHA512_BLOCK_SIZE - sizeof(__be64[2]); in sha512_base_do_finalize()
102 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize()
106 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize()
H A Dsha2.h21 #define SHA512_BLOCK_SIZE 128 macro
76 u8 buf[SHA512_BLOCK_SIZE];
/kernel/linux/linux-6.6/arch/sparc/crypto/
H A Dsha512_glue.c37 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update()
41 if (len - done >= SHA512_BLOCK_SIZE) { in __sha512_sparc64_update()
42 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update()
45 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update()
55 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update()
58 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update()
74 static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, }; in sha512_sparc64_final()
81 index = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_final()
82 padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index); in sha512_sparc64_final()
126 .cra_blocksize = SHA512_BLOCK_SIZE,
[all...]
/kernel/linux/linux-5.10/arch/sparc/crypto/
H A Dsha512_glue.c68 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update()
72 if (len - done >= SHA512_BLOCK_SIZE) { in __sha512_sparc64_update()
73 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update()
76 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update()
86 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update()
89 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update()
105 static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, }; in sha512_sparc64_final()
112 index = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_final()
113 padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index); in sha512_sparc64_final()
157 .cra_blocksize = SHA512_BLOCK_SIZE,
[all...]
/kernel/linux/linux-6.6/drivers/crypto/nx/
H A Dnx-sha512.c21 u8 buf[SHA512_BLOCK_SIZE];
73 u64 buf_len = (sctx->count[0] % SHA512_BLOCK_SIZE); in nx_sha512_update()
78 * 1: < SHA512_BLOCK_SIZE: copy into state, return 0 in nx_sha512_update()
79 * 2: >= SHA512_BLOCK_SIZE: process X blocks, copy in leftover in nx_sha512_update()
81 total = (sctx->count[0] % SHA512_BLOCK_SIZE) + len; in nx_sha512_update()
82 if (total < SHA512_BLOCK_SIZE) { in nx_sha512_update()
124 /* to_process: SHA512_BLOCK_SIZE aligned chunk to be in nx_sha512_update()
133 to_process = to_process & ~(SHA512_BLOCK_SIZE - 1); in nx_sha512_update()
172 } while (leftover >= SHA512_BLOCK_SIZE); in nx_sha512_update()
205 if (sctx->count[0] >= SHA512_BLOCK_SIZE) { in nx_sha512_final()
[all...]
/kernel/linux/linux-5.10/drivers/crypto/nx/
H A Dnx-sha512.c68 u64 buf_len = (sctx->count[0] % SHA512_BLOCK_SIZE); in nx_sha512_update()
73 * 1: < SHA512_BLOCK_SIZE: copy into state, return 0 in nx_sha512_update()
74 * 2: >= SHA512_BLOCK_SIZE: process X blocks, copy in leftover in nx_sha512_update()
76 total = (sctx->count[0] % SHA512_BLOCK_SIZE) + len; in nx_sha512_update()
77 if (total < SHA512_BLOCK_SIZE) { in nx_sha512_update()
119 /* to_process: SHA512_BLOCK_SIZE aligned chunk to be in nx_sha512_update()
128 to_process = to_process & ~(SHA512_BLOCK_SIZE - 1); in nx_sha512_update()
167 } while (leftover >= SHA512_BLOCK_SIZE); in nx_sha512_update()
200 if (sctx->count[0] >= SHA512_BLOCK_SIZE) { in nx_sha512_final()
218 len = sctx->count[0] & (SHA512_BLOCK_SIZE in nx_sha512_final()
[all...]
/kernel/linux/linux-5.10/arch/mips/cavium-octeon/crypto/
H A Docteon-sha512.c119 index = sctx->count[0] % SHA512_BLOCK_SIZE; in __octeon_sha512_update()
125 part_len = SHA512_BLOCK_SIZE - index; in __octeon_sha512_update()
132 for (i = part_len; i + SHA512_BLOCK_SIZE <= len; in __octeon_sha512_update()
133 i += SHA512_BLOCK_SIZE) in __octeon_sha512_update()
157 if ((sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in octeon_sha512_update()
234 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-6.6/arch/mips/cavium-octeon/crypto/
H A Docteon-sha512.c86 index = sctx->count[0] % SHA512_BLOCK_SIZE; in __octeon_sha512_update()
92 part_len = SHA512_BLOCK_SIZE - index; in __octeon_sha512_update()
99 for (i = part_len; i + SHA512_BLOCK_SIZE <= len; in __octeon_sha512_update()
100 i += SHA512_BLOCK_SIZE) in __octeon_sha512_update()
124 if ((sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in octeon_sha512_update()
201 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-5.10/arch/arm/crypto/
H A Dsha512-neon-glue.c32 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_neon_update()
91 .cra_blocksize = SHA512_BLOCK_SIZE,
H A Dsha512-glue.c63 .cra_blocksize = SHA512_BLOCK_SIZE,
77 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-6.6/arch/arm/crypto/
H A Dsha512-neon-glue.c32 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_neon_update()
89 .cra_blocksize = SHA512_BLOCK_SIZE,
H A Dsha512-glue.c61 .cra_blocksize = SHA512_BLOCK_SIZE,
75 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-5.10/arch/x86/crypto/
H A Dsha512_ssse3_glue.c50 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_update()
110 .cra_blocksize = SHA512_BLOCK_SIZE,
186 .cra_blocksize = SHA512_BLOCK_SIZE,
252 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-6.6/arch/x86/crypto/
H A Dsha512_ssse3_glue.c51 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_update()
111 .cra_blocksize = SHA512_BLOCK_SIZE,
187 .cra_blocksize = SHA512_BLOCK_SIZE,
253 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-5.10/arch/arm64/crypto/
H A Dsha512-ce-glue.c43 src += (blocks - rem) * SHA512_BLOCK_SIZE; in __sha512_ce_transform()
94 .base.cra_blocksize = SHA512_BLOCK_SIZE,
106 .base.cra_blocksize = SHA512_BLOCK_SIZE,
H A Dsha512-glue.c65 .base.cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-6.6/arch/arm64/crypto/
H A Dsha512-ce-glue.c43 src += (blocks - rem) * SHA512_BLOCK_SIZE; in __sha512_ce_transform()
94 .base.cra_blocksize = SHA512_BLOCK_SIZE,
106 .base.cra_blocksize = SHA512_BLOCK_SIZE,
H A Dsha512-glue.c65 .base.cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-6.6/drivers/crypto/aspeed/
H A Daspeed-hace.h166 u8 ipad[SHA512_BLOCK_SIZE];
167 u8 opad[SHA512_BLOCK_SIZE];
194 u8 buffer[SHA512_BLOCK_SIZE * 2];
/kernel/linux/linux-6.6/drivers/crypto/intel/keembay/
H A Dkeembay-ocs-hcu-core.c47 u8 key[SHA512_BLOCK_SIZE];
86 u8 buffer[2 * SHA512_BLOCK_SIZE];
598 rctx->blk_sz = SHA512_BLOCK_SIZE; in kmb_ocs_hcu_init()
1108 .cra_blocksize = SHA512_BLOCK_SIZE,
1134 .cra_blocksize = SHA512_BLOCK_SIZE,
/kernel/linux/linux-5.10/drivers/crypto/ccree/
H A Dcc_hash.h19 #define CC_MAX_HASH_BLCK_SIZE SHA512_BLOCK_SIZE
/kernel/linux/linux-6.6/drivers/crypto/ccree/
H A Dcc_hash.h19 #define CC_MAX_HASH_BLCK_SIZE SHA512_BLOCK_SIZE
/kernel/linux/linux-6.6/drivers/crypto/inside-secure/
H A Dsafexcel_hash.c1507 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_sha512_init()
1543 .cra_blocksize = SHA512_BLOCK_SIZE,
1564 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_sha384_init()
1627 req->len = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init()
1628 req->processed = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init()
1634 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init()
1672 .cra_blocksize = SHA512_BLOCK_SIZE,
1699 req->len = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init()
1700 req->processed = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init()
1706 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init()
[all...]

Completed in 13 milliseconds

123