/kernel/linux/linux-5.10/include/crypto/ |
H A D | sha512_base.h | 61 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update() 67 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update() 71 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update() 80 blocks = len / SHA512_BLOCK_SIZE; in sha512_base_do_update() 81 len %= SHA512_BLOCK_SIZE; in sha512_base_do_update() 85 data += blocks * SHA512_BLOCK_SIZE; in sha512_base_do_update() 98 const int bit_offset = SHA512_BLOCK_SIZE - sizeof(__be64[2]); in sha512_base_do_finalize() 101 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize() 105 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize()
|
H A D | sha.h | 24 #define SHA512_BLOCK_SIZE 128 macro 93 u8 buf[SHA512_BLOCK_SIZE];
|
/kernel/linux/linux-6.6/include/crypto/ |
H A D | sha512_base.h | 62 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update() 68 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update() 72 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update() 81 blocks = len / SHA512_BLOCK_SIZE; in sha512_base_do_update() 82 len %= SHA512_BLOCK_SIZE; in sha512_base_do_update() 86 data += blocks * SHA512_BLOCK_SIZE; in sha512_base_do_update() 99 const int bit_offset = SHA512_BLOCK_SIZE - sizeof(__be64[2]); in sha512_base_do_finalize() 102 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize() 106 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize()
|
H A D | sha2.h | 21 #define SHA512_BLOCK_SIZE 128 macro 76 u8 buf[SHA512_BLOCK_SIZE];
|
/kernel/linux/linux-6.6/arch/sparc/crypto/ |
H A D | sha512_glue.c | 37 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update() 41 if (len - done >= SHA512_BLOCK_SIZE) { in __sha512_sparc64_update() 42 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 45 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 55 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() 58 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update() 74 static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, }; in sha512_sparc64_final() 81 index = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_final() 82 padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index); in sha512_sparc64_final() 126 .cra_blocksize = SHA512_BLOCK_SIZE, [all...] |
/kernel/linux/linux-5.10/arch/sparc/crypto/ |
H A D | sha512_glue.c | 68 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update() 72 if (len - done >= SHA512_BLOCK_SIZE) { in __sha512_sparc64_update() 73 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 76 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 86 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() 89 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update() 105 static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, }; in sha512_sparc64_final() 112 index = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_final() 113 padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index); in sha512_sparc64_final() 157 .cra_blocksize = SHA512_BLOCK_SIZE, [all...] |
/kernel/linux/linux-6.6/drivers/crypto/nx/ |
H A D | nx-sha512.c | 21 u8 buf[SHA512_BLOCK_SIZE]; 73 u64 buf_len = (sctx->count[0] % SHA512_BLOCK_SIZE); in nx_sha512_update() 78 * 1: < SHA512_BLOCK_SIZE: copy into state, return 0 in nx_sha512_update() 79 * 2: >= SHA512_BLOCK_SIZE: process X blocks, copy in leftover in nx_sha512_update() 81 total = (sctx->count[0] % SHA512_BLOCK_SIZE) + len; in nx_sha512_update() 82 if (total < SHA512_BLOCK_SIZE) { in nx_sha512_update() 124 /* to_process: SHA512_BLOCK_SIZE aligned chunk to be in nx_sha512_update() 133 to_process = to_process & ~(SHA512_BLOCK_SIZE - 1); in nx_sha512_update() 172 } while (leftover >= SHA512_BLOCK_SIZE); in nx_sha512_update() 205 if (sctx->count[0] >= SHA512_BLOCK_SIZE) { in nx_sha512_final() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/nx/ |
H A D | nx-sha512.c | 68 u64 buf_len = (sctx->count[0] % SHA512_BLOCK_SIZE); in nx_sha512_update() 73 * 1: < SHA512_BLOCK_SIZE: copy into state, return 0 in nx_sha512_update() 74 * 2: >= SHA512_BLOCK_SIZE: process X blocks, copy in leftover in nx_sha512_update() 76 total = (sctx->count[0] % SHA512_BLOCK_SIZE) + len; in nx_sha512_update() 77 if (total < SHA512_BLOCK_SIZE) { in nx_sha512_update() 119 /* to_process: SHA512_BLOCK_SIZE aligned chunk to be in nx_sha512_update() 128 to_process = to_process & ~(SHA512_BLOCK_SIZE - 1); in nx_sha512_update() 167 } while (leftover >= SHA512_BLOCK_SIZE); in nx_sha512_update() 200 if (sctx->count[0] >= SHA512_BLOCK_SIZE) { in nx_sha512_final() 218 len = sctx->count[0] & (SHA512_BLOCK_SIZE in nx_sha512_final() [all...] |
/kernel/linux/linux-5.10/arch/mips/cavium-octeon/crypto/ |
H A D | octeon-sha512.c | 119 index = sctx->count[0] % SHA512_BLOCK_SIZE; in __octeon_sha512_update() 125 part_len = SHA512_BLOCK_SIZE - index; in __octeon_sha512_update() 132 for (i = part_len; i + SHA512_BLOCK_SIZE <= len; in __octeon_sha512_update() 133 i += SHA512_BLOCK_SIZE) in __octeon_sha512_update() 157 if ((sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in octeon_sha512_update() 234 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-6.6/arch/mips/cavium-octeon/crypto/ |
H A D | octeon-sha512.c | 86 index = sctx->count[0] % SHA512_BLOCK_SIZE; in __octeon_sha512_update() 92 part_len = SHA512_BLOCK_SIZE - index; in __octeon_sha512_update() 99 for (i = part_len; i + SHA512_BLOCK_SIZE <= len; in __octeon_sha512_update() 100 i += SHA512_BLOCK_SIZE) in __octeon_sha512_update() 124 if ((sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in octeon_sha512_update() 201 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-5.10/arch/arm/crypto/ |
H A D | sha512-neon-glue.c | 32 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_neon_update() 91 .cra_blocksize = SHA512_BLOCK_SIZE,
|
H A D | sha512-glue.c | 63 .cra_blocksize = SHA512_BLOCK_SIZE, 77 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-6.6/arch/arm/crypto/ |
H A D | sha512-neon-glue.c | 32 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_neon_update() 89 .cra_blocksize = SHA512_BLOCK_SIZE,
|
H A D | sha512-glue.c | 61 .cra_blocksize = SHA512_BLOCK_SIZE, 75 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | sha512_ssse3_glue.c | 50 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_update() 110 .cra_blocksize = SHA512_BLOCK_SIZE, 186 .cra_blocksize = SHA512_BLOCK_SIZE, 252 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | sha512_ssse3_glue.c | 51 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_update() 111 .cra_blocksize = SHA512_BLOCK_SIZE, 187 .cra_blocksize = SHA512_BLOCK_SIZE, 253 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-5.10/arch/arm64/crypto/ |
H A D | sha512-ce-glue.c | 43 src += (blocks - rem) * SHA512_BLOCK_SIZE; in __sha512_ce_transform() 94 .base.cra_blocksize = SHA512_BLOCK_SIZE, 106 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
H A D | sha512-glue.c | 65 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-6.6/arch/arm64/crypto/ |
H A D | sha512-ce-glue.c | 43 src += (blocks - rem) * SHA512_BLOCK_SIZE; in __sha512_ce_transform() 94 .base.cra_blocksize = SHA512_BLOCK_SIZE, 106 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
H A D | sha512-glue.c | 65 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-6.6/drivers/crypto/aspeed/ |
H A D | aspeed-hace.h | 166 u8 ipad[SHA512_BLOCK_SIZE]; 167 u8 opad[SHA512_BLOCK_SIZE]; 194 u8 buffer[SHA512_BLOCK_SIZE * 2];
|
/kernel/linux/linux-6.6/drivers/crypto/intel/keembay/ |
H A D | keembay-ocs-hcu-core.c | 47 u8 key[SHA512_BLOCK_SIZE]; 86 u8 buffer[2 * SHA512_BLOCK_SIZE]; 598 rctx->blk_sz = SHA512_BLOCK_SIZE; in kmb_ocs_hcu_init() 1108 .cra_blocksize = SHA512_BLOCK_SIZE, 1134 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/kernel/linux/linux-5.10/drivers/crypto/ccree/ |
H A D | cc_hash.h | 19 #define CC_MAX_HASH_BLCK_SIZE SHA512_BLOCK_SIZE
|
/kernel/linux/linux-6.6/drivers/crypto/ccree/ |
H A D | cc_hash.h | 19 #define CC_MAX_HASH_BLCK_SIZE SHA512_BLOCK_SIZE
|
/kernel/linux/linux-6.6/drivers/crypto/inside-secure/ |
H A D | safexcel_hash.c | 1507 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_sha512_init() 1543 .cra_blocksize = SHA512_BLOCK_SIZE, 1564 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_sha384_init() 1627 req->len = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init() 1628 req->processed = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init() 1634 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_hmac_sha512_init() 1672 .cra_blocksize = SHA512_BLOCK_SIZE, 1699 req->len = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init() 1700 req->processed = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init() 1706 req->block_sz = SHA512_BLOCK_SIZE; in safexcel_hmac_sha384_init() [all...] |