Home
last modified time | relevance | path

Searched refs:K2 (Results 1 - 25 of 39) sorted by relevance

12

/kernel/linux/linux-5.10/arch/x86/crypto/
H A Dnh-sse2-x86_64.S18 #define K2 %xmm6 define
78 movdqu 0x20(KEY), K2
88 _nh_stride K0, K1, K2, K3, 0x00
89 _nh_stride K1, K2, K3, K0, 0x10
90 _nh_stride K2, K3, K0, K1, 0x20
91 _nh_stride K3, K0, K1, K2, 0x30
100 _nh_stride K0, K1, K2, K3, 0x00
104 _nh_stride K1, K2, K3, K0, 0x10
108 _nh_stride K2, K3, K0, K1, 0x20
H A Dnh-avx2-x86_64.S20 #define K2 %ymm6 define
86 vmovdqu 0x00(KEY), K2
88 _nh_2xstride K0, K1, K2, K3
93 _nh_2xstride K2, K3, K0, K1
109 vmovdqu 0x00(KEY), K2
111 _nh_2xstride K0, K1, K2, K3
116 vmovdqa K2, K0
126 _nh_2xstride K0, K1, K2, K3
H A Dsha1_avx2_x86_64_asm.S685 #define K2 0x6ed9eba1 define
693 .long K2, K2, K2, K2
694 .long K2, K2, K2, K2
H A Dsha1_ssse3_asm.S433 #define K2 0x6ed9eba1 define
442 .long K2, K2, K2, K2
H A Dcrc32c-pcl-intel-asm_64.S222 pmovzxdq (%bufp,%rax), %xmm0 # 2 consts: K1:K2
227 pclmulqdq $0x00, %xmm0, %xmm1 # Multiply by K2
H A Dserpent-avx-x86_64-asm_64.S374 #define K2(x0, x1, x2, x3, x4, i) \ define
571 K2(RA, RB, RC, RD, RE, 0);
603 S(S7, RD, RE, RB, RC, RA); K2(RA, RB, RC, RD, RE, 32);
625 K2(RA, RB, RC, RD, RE, 32);
657 S(SI0, RE, RB, RC, RA, RD); K2(RC, RD, RB, RE, RA, 0);
H A Dserpent-avx2-asm_64.S380 #define K2(x0, x1, x2, x3, x4, i) \ define
577 K2(RA, RB, RC, RD, RE, 0);
609 S(S7, RD, RE, RB, RC, RA); K2(RA, RB, RC, RD, RE, 32);
631 K2(RA, RB, RC, RD, RE, 32);
663 S(SI0, RE, RB, RC, RA, RD); K2(RC, RD, RB, RE, RA, 0);
/kernel/linux/linux-6.6/arch/x86/crypto/
H A Dnh-sse2-x86_64.S19 #define K2 %xmm6 define
79 movdqu 0x20(KEY), K2
89 _nh_stride K0, K1, K2, K3, 0x00
90 _nh_stride K1, K2, K3, K0, 0x10
91 _nh_stride K2, K3, K0, K1, 0x20
92 _nh_stride K3, K0, K1, K2, 0x30
101 _nh_stride K0, K1, K2, K3, 0x00
105 _nh_stride K1, K2, K3, K0, 0x10
109 _nh_stride K2, K3, K0, K1, 0x20
H A Dnh-avx2-x86_64.S21 #define K2 %ymm6 define
87 vmovdqu 0x00(KEY), K2
89 _nh_2xstride K0, K1, K2, K3
94 _nh_2xstride K2, K3, K0, K1
110 vmovdqu 0x00(KEY), K2
112 _nh_2xstride K0, K1, K2, K3
117 vmovdqa K2, K0
127 _nh_2xstride K0, K1, K2, K3
H A Dsha1_avx2_x86_64_asm.S674 #define K2 0x6ed9eba1 define
682 .long K2, K2, K2, K2
683 .long K2, K2, K2, K2
H A Dsha1_ssse3_asm.S434 #define K2 0x6ed9eba1 define
443 .long K2, K2, K2, K2
H A Dcrc32c-pcl-intel-asm_64.S221 pmovzxdq (%bufp,%rax), %xmm0 # 2 consts: K1:K2
226 pclmulqdq $0x00, %xmm0, %xmm1 # Multiply by K2
/kernel/linux/linux-5.10/fs/ext4/
H A Dhash.c47 #define K2 013240474631UL macro
68 ROUND(G, a, b, c, d, in[1] + K2, 3); in half_md4_transform()
69 ROUND(G, d, a, b, c, in[3] + K2, 5); in half_md4_transform()
70 ROUND(G, c, d, a, b, in[5] + K2, 9); in half_md4_transform()
71 ROUND(G, b, c, d, a, in[7] + K2, 13); in half_md4_transform()
72 ROUND(G, a, b, c, d, in[0] + K2, 3); in half_md4_transform()
73 ROUND(G, d, a, b, c, in[2] + K2, 5); in half_md4_transform()
74 ROUND(G, c, d, a, b, in[4] + K2, 9); in half_md4_transform()
75 ROUND(G, b, c, d, a, in[6] + K2, 13); in half_md4_transform()
96 #undef K2 macro
[all...]
/kernel/linux/linux-6.6/arch/s390/crypto/
H A Dchacha-s390.S54 #define K2 %v18 define
91 VL K2,16,,KEY
110 VREPF XC0,K2,0
111 VREPF XC1,K2,1
112 VREPF XC2,K2,2
113 VREPF XC3,K2,3
285 VAF XC0,XC0,K2
308 VAF XC0,XC1,K2
335 VAF XC0,XC2,K2
362 VAF XC0,XC3,K2
437 #define K2 %v25 global() define
[all...]
/kernel/linux/linux-5.10/crypto/
H A Drmd160.c27 #define K2 RMD_K2 macro
86 ROUND(ee, aa, bb, cc, dd, F2, K2, in[7], 7); in rmd160_transform()
87 ROUND(dd, ee, aa, bb, cc, F2, K2, in[4], 6); in rmd160_transform()
88 ROUND(cc, dd, ee, aa, bb, F2, K2, in[13], 8); in rmd160_transform()
89 ROUND(bb, cc, dd, ee, aa, F2, K2, in[1], 13); in rmd160_transform()
90 ROUND(aa, bb, cc, dd, ee, F2, K2, in[10], 11); in rmd160_transform()
91 ROUND(ee, aa, bb, cc, dd, F2, K2, in[6], 9); in rmd160_transform()
92 ROUND(dd, ee, aa, bb, cc, F2, K2, in[15], 7); in rmd160_transform()
93 ROUND(cc, dd, ee, aa, bb, F2, K2, in[3], 15); in rmd160_transform()
94 ROUND(bb, cc, dd, ee, aa, F2, K2, i in rmd160_transform()
[all...]
H A Drmd256.c27 #define K2 RMD_K2 macro
101 ROUND(aa, bb, cc, dd, F2, K2, in[7], 7); in rmd256_transform()
102 ROUND(dd, aa, bb, cc, F2, K2, in[4], 6); in rmd256_transform()
103 ROUND(cc, dd, aa, bb, F2, K2, in[13], 8); in rmd256_transform()
104 ROUND(bb, cc, dd, aa, F2, K2, in[1], 13); in rmd256_transform()
105 ROUND(aa, bb, cc, dd, F2, K2, in[10], 11); in rmd256_transform()
106 ROUND(dd, aa, bb, cc, F2, K2, in[6], 9); in rmd256_transform()
107 ROUND(cc, dd, aa, bb, F2, K2, in[15], 7); in rmd256_transform()
108 ROUND(bb, cc, dd, aa, F2, K2, in[3], 15); in rmd256_transform()
109 ROUND(aa, bb, cc, dd, F2, K2, i in rmd256_transform()
[all...]
H A Drmd128.c27 #define K2 RMD_K2 macro
80 ROUND(aa, bb, cc, dd, F2, K2, in[7], 7); in rmd128_transform()
81 ROUND(dd, aa, bb, cc, F2, K2, in[4], 6); in rmd128_transform()
82 ROUND(cc, dd, aa, bb, F2, K2, in[13], 8); in rmd128_transform()
83 ROUND(bb, cc, dd, aa, F2, K2, in[1], 13); in rmd128_transform()
84 ROUND(aa, bb, cc, dd, F2, K2, in[10], 11); in rmd128_transform()
85 ROUND(dd, aa, bb, cc, F2, K2, in[6], 9); in rmd128_transform()
86 ROUND(cc, dd, aa, bb, F2, K2, in[15], 7); in rmd128_transform()
87 ROUND(bb, cc, dd, aa, F2, K2, in[3], 15); in rmd128_transform()
88 ROUND(aa, bb, cc, dd, F2, K2, i in rmd128_transform()
[all...]
H A Drmd320.c27 #define K2 RMD_K2 macro
107 ROUND(ee, aa, bb, cc, dd, F2, K2, in[7], 7); in rmd320_transform()
108 ROUND(dd, ee, aa, bb, cc, F2, K2, in[4], 6); in rmd320_transform()
109 ROUND(cc, dd, ee, aa, bb, F2, K2, in[13], 8); in rmd320_transform()
110 ROUND(bb, cc, dd, ee, aa, F2, K2, in[1], 13); in rmd320_transform()
111 ROUND(aa, bb, cc, dd, ee, F2, K2, in[10], 11); in rmd320_transform()
112 ROUND(ee, aa, bb, cc, dd, F2, K2, in[6], 9); in rmd320_transform()
113 ROUND(dd, ee, aa, bb, cc, F2, K2, in[15], 7); in rmd320_transform()
114 ROUND(cc, dd, ee, aa, bb, F2, K2, in[3], 15); in rmd320_transform()
115 ROUND(bb, cc, dd, ee, aa, F2, K2, i in rmd320_transform()
[all...]
H A Danubis.c491 u32 K0, K1, K2, K3; in anubis_setkey() local
497 K2 = T4[(kappa[N - 1] >> 8) & 0xff]; in anubis_setkey()
510 K2 = T4[(kappa[i] >> 8) & 0xff] ^ in anubis_setkey()
511 (T5[(K2 >> 24) ] & 0xff000000U) ^ in anubis_setkey()
512 (T5[(K2 >> 16) & 0xff] & 0x00ff0000U) ^ in anubis_setkey()
513 (T5[(K2 >> 8) & 0xff] & 0x0000ff00U) ^ in anubis_setkey()
514 (T5[(K2 ) & 0xff] & 0x000000ffU); in anubis_setkey()
524 ctx->E[r][2] = K2; in anubis_setkey()
H A Dkhazad.c763 u64 K2, K1; in khazad_setkey() local
766 K2 = ((u64)be32_to_cpu(key[0]) << 32) | be32_to_cpu(key[1]); in khazad_setkey()
779 c[r] ^ K2; in khazad_setkey()
780 K2 = K1; in khazad_setkey()
/kernel/linux/linux-6.6/crypto/
H A Drmd160.c27 #define K2 RMD_K2 macro
86 ROUND(ee, aa, bb, cc, dd, F2, K2, in[7], 7); in rmd160_transform()
87 ROUND(dd, ee, aa, bb, cc, F2, K2, in[4], 6); in rmd160_transform()
88 ROUND(cc, dd, ee, aa, bb, F2, K2, in[13], 8); in rmd160_transform()
89 ROUND(bb, cc, dd, ee, aa, F2, K2, in[1], 13); in rmd160_transform()
90 ROUND(aa, bb, cc, dd, ee, F2, K2, in[10], 11); in rmd160_transform()
91 ROUND(ee, aa, bb, cc, dd, F2, K2, in[6], 9); in rmd160_transform()
92 ROUND(dd, ee, aa, bb, cc, F2, K2, in[15], 7); in rmd160_transform()
93 ROUND(cc, dd, ee, aa, bb, F2, K2, in[3], 15); in rmd160_transform()
94 ROUND(bb, cc, dd, ee, aa, F2, K2, i in rmd160_transform()
[all...]
H A Danubis.c491 u32 K0, K1, K2, K3; in anubis_setkey() local
497 K2 = T4[(kappa[N - 1] >> 8) & 0xff]; in anubis_setkey()
510 K2 = T4[(kappa[i] >> 8) & 0xff] ^ in anubis_setkey()
511 (T5[(K2 >> 24) ] & 0xff000000U) ^ in anubis_setkey()
512 (T5[(K2 >> 16) & 0xff] & 0x00ff0000U) ^ in anubis_setkey()
513 (T5[(K2 >> 8) & 0xff] & 0x0000ff00U) ^ in anubis_setkey()
514 (T5[(K2 ) & 0xff] & 0x000000ffU); in anubis_setkey()
524 ctx->E[r][2] = K2; in anubis_setkey()
H A Dkhazad.c763 u64 K2, K1; in khazad_setkey() local
766 K2 = ((u64)be32_to_cpu(key[0]) << 32) | be32_to_cpu(key[1]); in khazad_setkey()
779 c[r] ^ K2; in khazad_setkey()
780 K2 = K1; in khazad_setkey()
/kernel/linux/linux-6.6/fs/ext4/
H A Dhash.c47 #define K2 013240474631UL macro
68 ROUND(G, a, b, c, d, in[1] + K2, 3); in half_md4_transform()
69 ROUND(G, d, a, b, c, in[3] + K2, 5); in half_md4_transform()
70 ROUND(G, c, d, a, b, in[5] + K2, 9); in half_md4_transform()
71 ROUND(G, b, c, d, a, in[7] + K2, 13); in half_md4_transform()
72 ROUND(G, a, b, c, d, in[0] + K2, 3); in half_md4_transform()
73 ROUND(G, d, a, b, c, in[2] + K2, 5); in half_md4_transform()
74 ROUND(G, c, d, a, b, in[4] + K2, 9); in half_md4_transform()
75 ROUND(G, b, c, d, a, in[6] + K2, 13); in half_md4_transform()
96 #undef K2 macro
[all...]
/kernel/linux/linux-5.10/arch/arm/crypto/
H A Dsha1-armv7-neon.S28 #define K2 0x6ED9EBA1 define
34 .LK2: .long K2, K2, K2, K2
314 vld1.32 {qK1-qK2}, [RT3]!; /* Load K1,K2 */

Completed in 16 milliseconds

12