Lines Matching refs:TMP1

199 .macro PRECOMPUTE SUBKEY TMP1 TMP2 TMP3 TMP4 TMP5 TMP6 TMP7
210 movdqa \TMP2, \TMP1
212 psrldq $8, \TMP1
217 pshufd $0x24, \TMP1, \TMP2
224 pshufd $78, \TMP3, \TMP1
225 pxor \TMP3, \TMP1
226 movdqu \TMP1, HashKey_k(%arg2)
228 GHASH_MUL \TMP5, \TMP3, \TMP1, \TMP2, \TMP4, \TMP6, \TMP7
232 pshufd $78, \TMP5, \TMP1
233 pxor \TMP5, \TMP1
234 movdqu \TMP1, HashKey_2_k(%arg2)
236 GHASH_MUL \TMP5, \TMP3, \TMP1, \TMP2, \TMP4, \TMP6, \TMP7
239 pshufd $78, \TMP5, \TMP1
240 pxor \TMP5, \TMP1
241 movdqu \TMP1, HashKey_3_k(%arg2)
243 GHASH_MUL \TMP5, \TMP3, \TMP1, \TMP2, \TMP4, \TMP6, \TMP7
246 pshufd $78, \TMP5, \TMP1
247 pxor \TMP5, \TMP1
248 movdqu \TMP1, HashKey_4_k(%arg2)
514 .macro GHASH_MUL GH HK TMP1 TMP2 TMP3 TMP4 TMP5
515 movdqa \GH, \TMP1
520 pclmulqdq $0x11, \HK, \TMP1 # TMP1 = a1*b1
524 pxor \TMP1, \TMP2 # TMP2 = (a0*b0)+(a1*b0)
529 pxor \TMP2, \TMP1 # TMP2:GH holds the result of GH*HK
562 pxor \TMP1, \GH # result is in TMP1
598 .macro CALC_AAD_HASH HASHKEY AAD AADLEN TMP1 TMP2 TMP3 TMP4 TMP5 \
612 GHASH_MUL \TMP6, \HASHKEY, \TMP1, \TMP2, \TMP3, \TMP4, \TMP5
625 READ_PARTIAL_BLOCK %r10, %r11, \TMP1, \TMP7
628 GHASH_MUL \TMP7, \HASHKEY, \TMP1, \TMP2, \TMP3, \TMP4, \TMP5
791 .macro INITIAL_BLOCKS_ENC_DEC TMP1 TMP2 TMP3 TMP4 TMP5 XMM0 XMM1 \
803 MOVADQ ONE(%RIP),\TMP1
806 paddd \TMP1, \XMM0 # INCR Y0
821 MOVADQ (%r10),\TMP1
823 aesenc \TMP1, %xmm\index
829 MOVADQ (%r10), \TMP1
831 aesenclast \TMP1, %xmm\index # Last Round
834 movdqu (%arg4 , %r11, 1), \TMP1
835 pxor \TMP1, %xmm\index
841 movdqa \TMP1, %xmm\index
853 GHASH_MUL %xmm6, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
855 GHASH_MUL %xmm7, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
857 GHASH_MUL %xmm8, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
860 GHASH_MUL %xmm7, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
862 GHASH_MUL %xmm8, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
865 GHASH_MUL %xmm8, \TMP3, \TMP1, \TMP2, \TMP4, \TMP5, \XMM1
875 MOVADQ ONE(%RIP),\TMP1
876 paddd \TMP1, \XMM0 # INCR Y0
880 paddd \TMP1, \XMM0 # INCR Y0
884 paddd \TMP1, \XMM0 # INCR Y0
888 paddd \TMP1, \XMM0 # INCR Y0
892 MOVADQ 0(%arg1),\TMP1
893 pxor \TMP1, \XMM1
894 pxor \TMP1, \XMM2
895 pxor \TMP1, \XMM3
896 pxor \TMP1, \XMM4
898 movaps 0x10*\index(%arg1), \TMP1
899 aesenc \TMP1, \XMM1
900 aesenc \TMP1, \XMM2
901 aesenc \TMP1, \XMM3
902 aesenc \TMP1, \XMM4
905 movaps 0x10*\index(%arg1), \TMP1
906 aesenc \TMP1, \XMM1
907 aesenc \TMP1, \XMM2
908 aesenc \TMP1, \XMM3
909 aesenc \TMP1, \XMM4
932 movdqu 16*0(%arg4 , %r11 , 1), \TMP1
933 pxor \TMP1, \XMM1
936 movdqa \TMP1, \XMM1
938 movdqu 16*1(%arg4 , %r11 , 1), \TMP1
939 pxor \TMP1, \XMM2
942 movdqa \TMP1, \XMM2
944 movdqu 16*2(%arg4 , %r11 , 1), \TMP1
945 pxor \TMP1, \XMM3
948 movdqa \TMP1, \XMM3
950 movdqu 16*3(%arg4 , %r11 , 1), \TMP1
951 pxor \TMP1, \XMM4
954 movdqa \TMP1, \XMM4
980 .macro GHASH_4_ENCRYPT_4_PARALLEL_enc TMP1 TMP2 TMP3 TMP4 TMP5 \
1016 movaps 0x10(%arg1), \TMP1
1017 aesenc \TMP1, \XMM1 # Round 1
1018 aesenc \TMP1, \XMM2
1019 aesenc \TMP1, \XMM3
1020 aesenc \TMP1, \XMM4
1021 movaps 0x20(%arg1), \TMP1
1022 aesenc \TMP1, \XMM1 # Round 2
1023 aesenc \TMP1, \XMM2
1024 aesenc \TMP1, \XMM3
1025 aesenc \TMP1, \XMM4
1026 movdqa \XMM6, \TMP1
1030 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1 * b1
1049 pxor \TMP1, \TMP4
1053 movdqa \XMM7, \TMP1
1060 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1079 pxor \TMP1, \TMP4
1087 movdqa \XMM8, \TMP1
1091 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1138 pxor \TMP4, \TMP1
1141 pxor \TMP1, \TMP2
1147 pxor \TMP2, \TMP1 # accumulate the results in TMP1:XMM5
1177 pxor \TMP1, \XMM5 # result is in TMP1
1188 .macro GHASH_4_ENCRYPT_4_PARALLEL_dec TMP1 TMP2 TMP3 TMP4 TMP5 \
1224 movaps 0x10(%arg1), \TMP1
1225 aesenc \TMP1, \XMM1 # Round 1
1226 aesenc \TMP1, \XMM2
1227 aesenc \TMP1, \XMM3
1228 aesenc \TMP1, \XMM4
1229 movaps 0x20(%arg1), \TMP1
1230 aesenc \TMP1, \XMM1 # Round 2
1231 aesenc \TMP1, \XMM2
1232 aesenc \TMP1, \XMM3
1233 aesenc \TMP1, \XMM4
1234 movdqa \XMM6, \TMP1
1238 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1 * b1
1257 pxor \TMP1, \TMP4
1261 movdqa \XMM7, \TMP1
1268 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1287 pxor \TMP1, \TMP4
1295 movdqa \XMM8, \TMP1
1299 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1350 pxor \TMP4, \TMP1
1353 pxor \TMP1, \TMP2
1359 pxor \TMP2, \TMP1 # accumulate the results in TMP1:XMM5
1389 pxor \TMP1, \XMM5 # result is in TMP1
1395 .macro GHASH_LAST_4 TMP1 TMP2 TMP3 TMP4 TMP5 TMP6 \
1411 # Multiply TMP1 * HashKey (using Karatsuba)
1413 movdqa \XMM2, \TMP1
1417 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1421 pxor \TMP1, \TMP6
1426 # Multiply TMP1 * HashKey (using Karatsuba)
1428 movdqa \XMM3, \TMP1
1432 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1436 pxor \TMP1, \TMP6
1440 # Multiply TMP1 * HashKey (using Karatsuba)
1441 movdqa \XMM4, \TMP1
1445 pclmulqdq $0x11, \TMP5, \TMP1 # TMP1 = a1*b1
1449 pxor \TMP1, \TMP6
1496 .macro ENCRYPT_SINGLE_BLOCK XMM0 TMP1
1505 MOVADQ (%r10),\TMP1
1506 aesenc \TMP1,\XMM0
1511 MOVADQ (%r10),\TMP1
1512 aesenclast \TMP1,\XMM0