/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | cmpxchg_64.h | 24 u128 full; 44 static __always_inline u128 arch_cmpxchg128(volatile u128 *ptr, u128 old, u128 new) in arch_cmpxchg128() 50 static __always_inline u128 arch_cmpxchg128_local(volatile u128 *ptr, u128 old, u128 new) in arch_cmpxchg128_local() 76 static __always_inline bool arch_try_cmpxchg128(volatile u128 *pt [all...] |
/kernel/linux/linux-5.10/include/crypto/ |
H A D | b128ops.h | 54 } u128; typedef 64 static inline void u128_xor(u128 *r, const u128 *p, const u128 *q) in u128_xor() 72 u128_xor((u128 *)r, (u128 *)p, (u128 *)q); in be128_xor() 77 u128_xor((u128 *)r, (u128 *)p, (u128 *) in le128_xor() [all...] |
/kernel/linux/linux-5.10/lib/crypto/ |
H A D | curve25519-hacl64.c | 17 typedef __uint128_t u128; typedef 53 static __always_inline void fproduct_copy_from_wide_(u64 *output, u128 *input) in fproduct_copy_from_wide_() 56 u128 xi = input[0]; in fproduct_copy_from_wide_() 60 u128 xi = input[1]; in fproduct_copy_from_wide_() 64 u128 xi = input[2]; in fproduct_copy_from_wide_() 68 u128 xi = input[3]; in fproduct_copy_from_wide_() 72 u128 xi = input[4]; in fproduct_copy_from_wide_() 78 fproduct_sum_scalar_multiplication_(u128 *output, u64 *input, u64 s) in fproduct_sum_scalar_multiplication_() 80 output[0] += (u128)input[0] * s; in fproduct_sum_scalar_multiplication_() 81 output[1] += (u128)inpu in fproduct_sum_scalar_multiplication_() [all...] |
H A D | poly1305-donna64.c | 13 typedef __uint128_t u128; typedef 44 u128 d0, d1, d2, d; in poly1305_core_blocks() 74 d0 = (u128)h0 * r0; in poly1305_core_blocks() 75 d = (u128)h1 * s2; in poly1305_core_blocks() 77 d = (u128)h2 * s1; in poly1305_core_blocks() 79 d1 = (u128)h0 * r1; in poly1305_core_blocks() 80 d = (u128)h1 * r0; in poly1305_core_blocks() 82 d = (u128)h2 * s2; in poly1305_core_blocks() 84 d2 = (u128)h0 * r2; in poly1305_core_blocks() 85 d = (u128)h in poly1305_core_blocks() [all...] |
/kernel/linux/linux-6.6/lib/crypto/ |
H A D | curve25519-hacl64.c | 51 static __always_inline void fproduct_copy_from_wide_(u64 *output, u128 *input) in fproduct_copy_from_wide_() 54 u128 xi = input[0]; in fproduct_copy_from_wide_() 58 u128 xi = input[1]; in fproduct_copy_from_wide_() 62 u128 xi = input[2]; in fproduct_copy_from_wide_() 66 u128 xi = input[3]; in fproduct_copy_from_wide_() 70 u128 xi = input[4]; in fproduct_copy_from_wide_() 76 fproduct_sum_scalar_multiplication_(u128 *output, u64 *input, u64 s) in fproduct_sum_scalar_multiplication_() 78 output[0] += (u128)input[0] * s; in fproduct_sum_scalar_multiplication_() 79 output[1] += (u128)input[1] * s; in fproduct_sum_scalar_multiplication_() 80 output[2] += (u128)inpu in fproduct_sum_scalar_multiplication_() [all...] |
H A D | poly1305-donna64.c | 42 u128 d0, d1, d2, d; in poly1305_core_blocks() 72 d0 = (u128)h0 * r0; in poly1305_core_blocks() 73 d = (u128)h1 * s2; in poly1305_core_blocks() 75 d = (u128)h2 * s1; in poly1305_core_blocks() 77 d1 = (u128)h0 * r1; in poly1305_core_blocks() 78 d = (u128)h1 * r0; in poly1305_core_blocks() 80 d = (u128)h2 * s2; in poly1305_core_blocks() 82 d2 = (u128)h0 * r2; in poly1305_core_blocks() 83 d = (u128)h1 * r1; in poly1305_core_blocks() 85 d = (u128)h in poly1305_core_blocks() [all...] |
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | twofish_glue_3way.c | 41 u128 ivs[2]; in twofish_dec_blk_cbc_3way() 42 u128 *dst = (u128 *)d; in twofish_dec_blk_cbc_3way() 43 const u128 *src = (const u128 *)s; in twofish_dec_blk_cbc_3way() 58 u128 *dst = (u128 *)d; in twofish_enc_blk_ctr() 59 const u128 *src = (const u128 *)s; in twofish_enc_blk_ctr() 68 u128_xor(dst, dst, (u128 *) in twofish_enc_blk_ctr() [all...] |
H A D | serpent_sse2_glue.c | 36 u128 ivs[SERPENT_PARALLEL_BLOCKS - 1]; in serpent_decrypt_cbc_xway() 37 u128 *dst = (u128 *)d; in serpent_decrypt_cbc_xway() 38 const u128 *src = (const u128 *)s; in serpent_decrypt_cbc_xway() 53 u128 *dst = (u128 *)d; in serpent_crypt_ctr() 54 const u128 *src = (const u128 *)s; in serpent_crypt_ctr() 60 u128_xor(dst, src, (u128 *) in serpent_crypt_ctr() [all...] |
H A D | glue_helper.c | 78 const u128 *src = (u128 *)walk.src.virt.addr; in glue_cbc_encrypt_req_128bit() 79 u128 *dst = (u128 *)walk.dst.virt.addr; in glue_cbc_encrypt_req_128bit() 80 u128 *iv = (u128 *)walk.iv; in glue_cbc_encrypt_req_128bit() 91 *(u128 *)walk.iv = *iv; in glue_cbc_encrypt_req_128bit() 111 const u128 *src = walk.src.virt.addr; in glue_cbc_decrypt_req_128bit() 112 u128 *dst = walk.dst.virt.addr; in glue_cbc_decrypt_req_128bit() 115 u128 last_i in glue_cbc_decrypt_req_128bit() [all...] |
H A D | serpent_avx_glue.c | 50 u128 *dst = (u128 *)d; in __serpent_crypt_ctr() 51 const u128 *src = (const u128 *)s; in __serpent_crypt_ctr() 57 u128_xor(dst, src, (u128 *)&ctrblk); in __serpent_crypt_ctr()
|
H A D | cast6_avx_glue.c | 54 u128 *dst = (u128 *)d; in cast6_crypt_ctr() 55 const u128 *src = (const u128 *)s; in cast6_crypt_ctr() 61 u128_xor(dst, src, (u128 *)&ctrblk); in cast6_crypt_ctr()
|
H A D | ghash-clmulni-intel_glue.c | 27 void clmul_ghash_mul(char *dst, const u128 *shash); 30 const u128 *shash); 37 u128 shash;
|
H A D | camellia_glue.c | 1267 u128 *dst = (u128 *)d; in camellia_decrypt_cbc_2way() 1268 const u128 *src = (const u128 *)s; in camellia_decrypt_cbc_2way() 1269 u128 iv = *src; in camellia_decrypt_cbc_2way() 1280 u128 *dst = (u128 *)d; in camellia_crypt_ctr() 1281 const u128 *src = (const u128 *)s; in camellia_crypt_ctr() 1296 u128 *ds in camellia_crypt_ctr_2way() [all...] |
/kernel/linux/linux-6.6/arch/s390/include/asm/ |
H A D | cmpxchg.h | 195 static __always_inline u128 arch_cmpxchg128(volatile u128 *ptr, u128 old, u128 new) in arch_cmpxchg128()
|
H A D | cpu_mf.h | 143 u128 val;
|
/kernel/linux/linux-5.10/drivers/crypto/vmx/ |
H A D | ghash.c | 27 void gcm_init_p8(u128 htable[16], const u64 Xi[2]); 28 void gcm_gmult_p8(u64 Xi[2], const u128 htable[16]); 29 void gcm_ghash_p8(u64 Xi[2], const u128 htable[16], 34 u128 htable[16];
|
/kernel/linux/linux-6.6/drivers/crypto/vmx/ |
H A D | ghash.c | 28 void gcm_init_p8(u128 htable[16], const u64 Xi[2]); 29 void gcm_gmult_p8(u64 Xi[2], const u128 htable[16]); 30 void gcm_ghash_p8(u64 Xi[2], const u128 htable[16], 35 u128 htable[16];
|
/kernel/linux/linux-6.6/arch/arm64/include/asm/ |
H A D | atomic_ll_sc.h | 298 u128 full; 305 static __always_inline u128 \ 306 __ll_sc__cmpxchg128##name(volatile u128 *ptr, u128 old, u128 new) \ 322 : [v] "+Q" (*(u128 *)ptr), \
|
H A D | atomic_lse.h | 285 static __always_inline u128 \ 286 __lse__cmpxchg128##name(volatile u128 *ptr, u128 old, u128 new) \ 300 [v] "+Q" (*(u128 *)ptr) \
|
H A D | cmpxchg.h | 134 static inline u128 __cmpxchg128##name(volatile u128 *ptr, \ 135 u128 old, u128 new) \
|
/kernel/linux/linux-6.6/rust/alloc/vec/ |
H A D | is_zero.rs | 36 impl_is_zero!(u128, |x| x == 0); 168 impl_is_zero_option_of_num!(u8, u16, u32, u64, u128, i8, i16, i32, i64, i128, usize, isize,);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/microchip/vcap/ |
H A D | vcap_api_client.h | 73 struct vcap_u128_key u128; member 131 struct vcap_u128_action u128; member
|
H A D | vcap_api_debugfs.c | 89 value = data->u128.value; in vcap_debugfs_show_rule_keyfield() 90 mask = data->u128.mask; in vcap_debugfs_show_rule_keyfield() 94 vcap_netbytes_copy(nvalue, data->u128.value, in vcap_debugfs_show_rule_keyfield() 96 vcap_netbytes_copy(nmask, data->u128.mask, in vcap_debugfs_show_rule_keyfield()
|
H A D | vcap_api.c | 489 value = kf->data.u128.value; in vcap_encode_keyfield() 490 mask = kf->data.u128.mask; in vcap_encode_keyfield() 580 vcap_copy_to_w32be(ddata->u128.value, sdata->u128.value, size); in vcap_copy_from_client_keyfield() 581 vcap_copy_to_w32be(ddata->u128.mask, sdata->u128.mask, size); in vcap_copy_from_client_keyfield() 630 vcap_copy_to_w32be(ddata->u128.value, sdata->u128.value, size); in vcap_copy_from_client_actionfield() 774 value = af->data.u128.value; in vcap_encode_actionfield() 1135 vcap_copy_from_w32be(field->data.u128 in vcap_copy_to_client_actionfield() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | types.h | 15 typedef __u128 u128; typedef
|