/kernel/linux/linux-6.6/arch/arm64/crypto/ |
H A D | sm3-neon-core.S | 76 #define XTMP0 v6 define 197 rev32 XTMP0.16b, W0.16b; 205 eor XTMP4.16b, XTMP1.16b, XTMP0.16b; 209 st1 {XTMP0.16b}, [addr0], #16; 216 ext W0.16b, XTMP0.16b, XTMP0.16b, #8; /* W0: xx, w0, xx, xx */ 223 ext W1.16b, XTMP0.16b, XTMP0.16b, #4; /* W1: xx, w3, w2, w1 */ 245 /* Load (w[i - 16]) => XTMP0 */ \ 247 ext XTMP0 [all...] |
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | sm3-avx-asm_64.S | 124 #define XTMP0 %xmm6 define 241 vmovdqu 0*16(RDATA), XTMP0; /* XTMP0: w3, w2, w1, w0 */ \ 245 vpshufb BSWAP_REG, XTMP0, XTMP0; \ 249 vpxor XTMP0, XTMP1, XTMP4; \ 253 vmovdqa XTMP0, IW_W1_ADDR(0, 0); \ 263 vpshufd $0b00000000, XTMP0, W0; /* W0: xx, w0, xx, xx */ \ 264 vpshufd $0b11111001, XTMP0, W1; /* W1: xx, w3, w2, w1 */ \ 272 /* Load (w[i - 16]) => XTMP0 */ \ [all...] |
H A D | sha256-avx-asm.S | 86 XTMP0 = %xmm0 define 161 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 168 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 228 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 262 vpaddd XTMP4, XTMP0, XTMP [all...] |
H A D | sha256-avx2-asm.S | 79 XTMP0 = %ymm0 define 163 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 169 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16]# y1 = (e >> 6)# S1 245 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 286 vpaddd XTMP4, XTMP0, XTMP [all...] |
H A D | sha256-ssse3-asm.S | 80 XTMP0 = %xmm0 define 151 movdqa X3, XTMP0 155 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 163 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 228 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 265 paddd XTMP4, XTMP0 # XTMP0 [all...] |
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | sha256-avx-asm.S | 85 XTMP0 = %xmm0 define 160 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 167 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 227 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 261 vpaddd XTMP4, XTMP0, XTMP [all...] |
H A D | sha256-avx2-asm.S | 78 XTMP0 = %ymm0 define 164 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 170 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16]# y1 = (e >> 6)# S1 246 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 287 vpaddd XTMP4, XTMP0, XTMP [all...] |
H A D | sha256-ssse3-asm.S | 79 XTMP0 = %xmm0 define 150 movdqa X3, XTMP0 154 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 162 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 227 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 264 paddd XTMP4, XTMP0 # XTMP0 [all...] |