/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | sm4-aesni-avx-asm_64.S | 32 #define RA1 %xmm9 define 158 vmovdqa RA0, RA1; 163 vmovdqu 1*16(%rdx), RA1; 174 vpshufb RTMP2, RA1, RA1; 186 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 217 ROUND(0, RA0, RA1, RA2, RA3); 218 ROUND(1, RA1, RA2, RA3, RA0); 219 ROUND(2, RA2, RA3, RA0, RA1); 220 ROUND(3, RA3, RA0, RA1, RA [all...] |
H A D | sm4-aesni-avx2-asm_64.S | 33 #define RA1 %ymm9 define 160 * RA0, RA1, RA2, RA3, RB0, RB1, RB2, RB3: sixteen parallel 163 * RA0, RA1, RA2, RA3, RB0, RB1, RB2, RB3: sixteen parallel 170 vpshufb RTMP2, RA1, RA1; 179 transpose_4x4(RA0, RA1, RA2, RA3, RTMP0, RTMP1); 245 ROUND(0, RA0, RA1, RA2, RA3, RB0, RB1, RB2, RB3); 246 ROUND(1, RA1, RA2, RA3, RA0, RB1, RB2, RB3, RB0); 247 ROUND(2, RA2, RA3, RA0, RA1, RB2, RB3, RB0, RB1); 248 ROUND(3, RA3, RA0, RA1, RA [all...] |
H A D | twofish-avx-x86_64-asm_64.S | 37 #define RA1 %xmm0 define 234 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 236 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 245 inpack_blocks(RA1, RB1, RC1, RD1, RK1, RX0, RY0, RK2); 246 preload_rgi(RA1); 266 outunpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2); 275 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 277 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 285 inpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2); 287 rotate_1l(RA1); [all...] |
H A D | cast6-avx-x86_64-asm_64.S | 37 #define RA1 %xmm0 define 255 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 257 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 269 inpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 293 outunpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 303 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 305 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 317 inpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 340 outunpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 358 load_8way(%rdx, RA1, RB [all...] |
H A D | serpent-avx2-asm_64.S | 30 #define RA1 %ymm2 define 556 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: plaintext 558 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 563 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 600 write_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 609 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 616 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 669 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 673 store_16way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 691 load_16way(%rdx, RA1, RB [all...] |
H A D | serpent-avx-x86_64-asm_64.S | 29 #define RA1 %xmm0 define 556 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 558 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 563 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 600 write_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 609 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 616 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 667 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 671 store_8way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 685 load_8way(%rdx, RA1, RB [all...] |
H A D | serpent-sse2-x86_64-asm_64.S | 22 #define RA1 %xmm0 define 633 read_blocks(%rdx, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 675 write_blocks(%rsi, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 681 xor_blocks(%rsi, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 697 read_blocks(%rdx, RA1, RB1, RC1, RD1, RK0, RK1, RK2);
|
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | cast6-avx-x86_64-asm_64.S | 37 #define RA1 %xmm0 define 253 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 255 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 267 inpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 291 outunpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 301 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 303 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 315 inpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 338 outunpack_blocks(RA1, RB1, RC1, RD1, RTMP, RX, RKRF, RKM); 356 load_8way(%rdx, RA1, RB [all...] |
H A D | twofish-avx-x86_64-asm_64.S | 42 #define RA1 %xmm0 define 240 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 242 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 251 inpack_blocks(RA1, RB1, RC1, RD1, RK1, RX0, RY0, RK2); 252 preload_rgi(RA1); 272 outunpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2); 282 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 284 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 292 inpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2); 294 rotate_1l(RA1); [all...] |
H A D | serpent-avx-x86_64-asm_64.S | 33 #define RA1 %xmm0 define 561 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 563 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 568 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 605 write_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 615 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 622 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 673 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 677 store_8way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 691 load_8way(%rdx, RA1, RB [all...] |
H A D | serpent-avx2-asm_64.S | 40 #define RA1 %ymm2 define 567 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: plaintext 569 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 574 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 611 write_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 621 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 628 read_blocks(RA1, RB1, RC1, RD1, RK0, RK1, RK2); 681 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 685 store_16way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 703 load_16way(%rdx, RA1, RB [all...] |
H A D | serpent-sse2-x86_64-asm_64.S | 22 #define RA1 %xmm0 define 633 read_blocks(%rdx, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 675 write_blocks(%rsi, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 681 xor_blocks(%rsi, RA1, RB1, RC1, RD1, RK0, RK1, RK2); 697 read_blocks(%rdx, RA1, RB1, RC1, RD1, RK0, RK1, RK2);
|