Searched refs:PADLOCK_ALIGNMENT (Results 1 - 8 of 8) sorted by relevance
/kernel/linux/linux-5.10/drivers/crypto/ |
H A D | padlock-aes.c | 51 } __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 62 __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 64 __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 90 unsigned long align = PADLOCK_ALIGNMENT; in aes_ctx_common() 218 u8 buf[AES_BLOCK_SIZE * (MAX_ECB_FETCH_BLOCKS - 1) + PADLOCK_ALIGNMENT - 1]; in ecb_crypt_copy() 219 u8 *tmp = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in ecb_crypt_copy() 232 u8 buf[AES_BLOCK_SIZE * (MAX_CBC_FETCH_BLOCKS - 1) + PADLOCK_ALIGNMENT - 1]; in cbc_crypt_copy() 233 u8 *tmp = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in cbc_crypt_copy() 331 .cra_alignmask = PADLOCK_ALIGNMENT - 1, 400 .base.cra_alignmask = PADLOCK_ALIGNMENT [all...] |
H A D | padlock-sha.c | 77 char buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha1_finup() 79 char *result = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha1_finup() 138 char buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha256_finup() 140 char *result = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha256_finup() 286 u8 buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha1_update_nano() 288 u8 *dst = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha1_update_nano() 370 u8 buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha256_update_nano() 372 u8 *dst = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha256_update_nano()
|
/kernel/linux/linux-6.6/drivers/crypto/ |
H A D | padlock-aes.c | 51 } __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 62 __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 64 __attribute__ ((__aligned__(PADLOCK_ALIGNMENT))); 90 unsigned long align = PADLOCK_ALIGNMENT; in aes_ctx_common() 218 u8 buf[AES_BLOCK_SIZE * (MAX_ECB_FETCH_BLOCKS - 1) + PADLOCK_ALIGNMENT - 1]; in ecb_crypt_copy() 219 u8 *tmp = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in ecb_crypt_copy() 232 u8 buf[AES_BLOCK_SIZE * (MAX_CBC_FETCH_BLOCKS - 1) + PADLOCK_ALIGNMENT - 1]; in cbc_crypt_copy() 233 u8 *tmp = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in cbc_crypt_copy() 331 .cra_alignmask = PADLOCK_ALIGNMENT - 1, 400 .base.cra_alignmask = PADLOCK_ALIGNMENT [all...] |
H A D | padlock-sha.c | 78 char buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha1_finup() 80 char *result = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha1_finup() 139 char buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha256_finup() 141 char *result = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha256_finup() 287 u8 buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha1_update_nano() 289 u8 *dst = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha1_update_nano() 371 u8 buf[128 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in padlock_sha256_update_nano() 373 u8 *dst = PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in padlock_sha256_update_nano()
|
/kernel/linux/linux-5.10/drivers/char/hw_random/ |
H A D | via-rng.c | 89 char buf[16 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in via_rng_data_present() 91 u32 *via_rng_datum = (u32 *)PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in via_rng_data_present()
|
/kernel/linux/linux-6.6/drivers/char/hw_random/ |
H A D | via-rng.c | 89 char buf[16 + PADLOCK_ALIGNMENT - STACK_ALIGN] __attribute__ in via_rng_data_present() 91 u32 *via_rng_datum = (u32 *)PTR_ALIGN(&buf[0], PADLOCK_ALIGNMENT); in via_rng_data_present()
|
/kernel/linux/linux-5.10/include/crypto/ |
H A D | padlock.h | 11 #define PADLOCK_ALIGNMENT 16 macro
|
/kernel/linux/linux-6.6/include/crypto/ |
H A D | padlock.h | 11 #define PADLOCK_ALIGNMENT 16 macro
|
Completed in 4 milliseconds