Lines Matching refs:Xi

138 $Xi="%rdi";
262 movzb 15($Xi),$Zlo
265 &loop ($Xi);
267 mov $Zlo,8($Xi)
268 mov $Zhi,($Xi)
345 &mov ($Zlo,"8($Xi)");
346 &mov ($Zhi,"0($Xi)");
356 &mov ("($Xi)",$Zhi);
357 &mov ("8($Xi)","%rdx");
392 &mov ($dat,"$j($Xi)") if (--$j%4==0);
433 mov $Zlo,8($Xi)
434 mov $Zhi,($Xi)
465 ($Xi,$Xhi)=("%xmm0","%xmm1"); $Hkey="%xmm2";
469 my ($Xhi,$Xi,$Hkey,$HK)=@_;
473 movdqa $Xi,$Xhi #
474 pshufd \$0b01001110,$Xi,$T1
476 pxor $Xi,$T1 #
481 movdqa $Xi,$Xhi #
482 pshufd \$0b01001110,$Xi,$T1
483 pxor $Xi,$T1 #
487 pclmulqdq \$0x00,$Hkey,$Xi #######
490 pxor $Xi,$T1 #
497 pxor $T2,$Xi #
502 my ($Xhi,$Xi) = @_;
506 movdqa $Xi,$T2 #
507 movdqa $Xi,$T1
508 psllq \$5,$Xi
509 pxor $Xi,$T1 #
510 psllq \$1,$Xi
511 pxor $T1,$Xi #
512 psllq \$57,$Xi #
513 movdqa $Xi,$T1 #
514 pslldq \$8,$Xi
516 pxor $T2,$Xi
520 movdqa $Xi,$T2
521 psrlq \$1,$Xi
523 pxor $Xi,$T2
524 psrlq \$5,$Xi
525 pxor $T2,$Xi #
526 psrlq \$1,$Xi #
527 pxor $Xhi,$Xi #
569 movdqa $Hkey,$Xi
572 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK);
573 &reduction_alg9 ($Xhi,$Xi);
576 pshufd \$0b01001110,$Xi,$T2
579 pxor $Xi,$T2 # Karatsuba pre-processing
580 movdqu $Xi,0x10($Htbl) # save H^2
585 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H^3
586 &reduction_alg9 ($Xhi,$Xi);
588 movdqa $Xi,$T3
590 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H^4
591 &reduction_alg9 ($Xhi,$Xi);
594 pshufd \$0b01001110,$Xi,$T2
597 pxor $Xi,$T2 # Karatsuba pre-processing
598 movdqu $Xi,0x40($Htbl) # save H^4
625 movdqu ($Xip),$Xi
629 pshufb $T3,$Xi
631 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$T2);
632 $code.=<<___ if (0 || (&reduction_alg9($Xhi,$Xi)&&0));
641 pand $Xi,$T3
642 pshufb $T3,$T2 # ($Xi&7)·0xE0
644 pclmulqdq \$0x00,$Xi,$T1 # ·(0xE1<<1)
645 pxor $Xi,$T2
648 pxor $T2,$Xi
649 pclmulqdq \$0x01,$T3,$Xi
653 pslldq \$7,$Xi
654 pxor $Xhi,$Xi
657 pshufb $T3,$Xi
658 movdqu $Xi,($Xip)
698 movdqu ($Xip),$Xi
701 pshufb $T3,$Xi
726 # Xi+4 =[(H*Ii+3) + (H^2*Ii+2) + (H^3*Ii+1) + H^4*(Ii+Xi)] mod P
756 pxor $T1,$Xi
759 movdqa $Xi,$Xhi
760 pshufd \$0b01001110,$Xi,$T1
761 pxor $Xi,$T1
774 pclmulqdq \$0x00,$Hkey4,$Xi
779 xorps $Xln,$Xi
792 pxor $Xi,$T1 # aggregated Karatsuba post-processing
800 pxor $T1,$Xi
805 pand $Xi,$T1 # 1st phase
807 pxor $Xi,$T2 #
814 pxor $T2,$Xi
818 movdqa $Xi,$T2 # 2nd phase
819 psrlq \$1,$Xi
829 pxor $Xi,$T2
830 psrlq \$5,$Xi
835 pxor $T2,$Xi #
839 psrlq \$1,$Xi #
840 pxor $Xhi,$Xi #
841 movdqa $Xi,$Xhi
844 pshufd \$0b01001110,$Xi,$T1
845 pxor $Xi,$T1
855 pclmulqdq \$0x00,$Hkey4,$Xi
859 xorps $Xln,$Xi
861 pxor $Xi,$Xhi # aggregated Karatsuba post-processing
865 pxor $Xi,$Xhi
871 pxor $T2,$Xi #
873 &reduction_alg9($Xhi,$Xi);
885 # Xi+2 =[H*(Ii+1 + Xi+1)] mod P =
886 # [(H*Ii+1) + (H*Xi+1)] mod P =
887 # [(H*Ii+1) + H^2*(Ii+Xi)] mod P
893 pxor $T1,$Xi # Ii+Xi
911 movdqa $Xi,$Xhi
913 pshufd \$0b01001110,$Xi,$Xmn #
914 pxor $Xi,$Xmn #
916 pclmulqdq \$0x00,$Hkey2,$Xi
920 pxor $Xln,$Xi # (H*Ii+1) + H^2*(Ii+Xi)
923 pxor $Xi,$T1 # aggregated Karatsuba post-processing
928 pxor $T2,$Xhi # "Ii+Xi", consume early
935 pxor $Xmn,$Xi #
939 movdqa $Xi,$T2 # 1st phase
940 movdqa $Xi,$T1
941 psllq \$5,$Xi
942 pxor $Xi,$T1 #
944 psllq \$1,$Xi
945 pxor $T1,$Xi #
946 psllq \$57,$Xi #
947 movdqa $Xi,$T1 #
948 pslldq \$8,$Xi
950 pxor $T2,$Xi
955 movdqa $Xi,$T2 # 2nd phase
956 psrlq \$1,$Xi
959 pxor $Xi,$T2
960 psrlq \$5,$Xi
961 pxor $T2,$Xi #
963 psrlq \$1,$Xi #
965 pxor $Xhi,$Xi #
971 movdqa $Xi,$Xhi
973 pshufd \$0b01001110,$Xi,$Xmn #
974 pxor $Xi,$Xmn #
976 pclmulqdq \$0x00,$Hkey2,$Xi
980 pxor $Xln,$Xi # (H*Ii+1) + H^2*(Ii+Xi)
982 pxor $Xi,$T1
989 pxor $Xmn,$Xi #
991 &reduction_alg9 ($Xhi,$Xi);
999 pxor $T1,$Xi # Ii+Xi
1001 &clmul64x64_T2 ($Xhi,$Xi,$Hkey,$HK); # H*(Ii+Xi)
1002 &reduction_alg9 ($Xhi,$Xi);
1005 pshufb $T3,$Xi
1006 movdqu $Xi,($Xip)
1067 vmovdqa $Hkey,$Xi
1074 my ($Xhi,$Xi,$Hkey,$HK)=@_;
1078 vpunpckhqdq $Xi,$Xi,$T1
1080 vpxor $Xi,$T1,$T1 #
1085 vpunpckhqdq $Xi,$Xi,$T1
1086 vpxor $Xi,$T1,$T1 #
1090 vpclmulqdq \$0x11,$Hkey,$Xi,$Xhi #######
1091 vpclmulqdq \$0x00,$Hkey,$Xi,$Xi #######
1093 vpxor $Xi,$Xhi,$T2 #
1098 vpxor $T2,$Xi,$Xi #
1104 my ($Xhi,$Xi) = @_;
1107 vpsllq \$57,$Xi,$T1 # 1st phase
1108 vpsllq \$62,$Xi,$T2
1110 vpsllq \$63,$Xi,$T1
1114 vpxor $T1,$Xi,$Xi #
1117 vpsrlq \$1,$Xi,$T2 # 2nd phase
1118 vpxor $Xi,$Xhi,$Xhi
1119 vpxor $T2,$Xi,$Xi #
1121 vpxor $T2,$Xi,$Xi #
1122 vpsrlq \$1,$Xi,$Xi #
1123 vpxor $Xhi,$Xi,$Xi #
1133 &clmul64x64_avx ($Xhi,$Xi,$Hkey,$HK); # calculate H^3,5,7
1134 &reduction_avx ($Xhi,$Xi);
1137 vmovdqa $Xi,$T3
1139 &clmul64x64_avx ($Xhi,$Xi,$Hkey,$HK); # calculate H^2,4,6,8
1140 &reduction_avx ($Xhi,$Xi);
1143 vpshufd \$0b01001110,$Xi,$T2
1146 vpxor $Xi,$T2,$T2 # Karatsuba pre-processing
1147 vmovdqu $Xi,0x10($Htbl) # save H^2,4,6,8
1201 $Xi,$Xo,$Tred,$bswap,$Ii,$Ij) = map("%xmm$_",(0..15));
1222 vmovdqu ($Xip),$Xi # load $Xi
1226 vpshufb $bswap,$Xi,$Xi
1321 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1331 vpclmulqdq \$0x00,$Hkey,$Ij,$Xi
1344 vpxor $Zlo,$Xi,$Xi # collect result
1355 vpxor $Xi,$Tred,$Tred # aggregated Karatsuba post-processing
1362 vpxor $T2, $Xi, $Xi
1374 vpalignr \$8,$Xi,$Xi,$Tred # 1st phase
1387 vpclmulqdq \$0x10,(%r10),$Xi,$Xi
1411 vxorps $Tred,$Xi,$Xi
1414 vpalignr \$8,$Xi,$Xi,$Tred # 2nd phase
1420 vpclmulqdq \$0x10,(%r10),$Xi,$Xi
1436 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1551 vpxor $Xi,$Ij,$Ij # accumulate $Xi
1564 vpxor $Xlo,$Zlo,$Xi
1568 vpxor $Xi, $Zmi,$Zmi # aggregated Karatsuba post-processing
1572 vpxor $T2, $Xi, $Xi
1575 vpclmulqdq \$0x10,$Tred,$Xi,$T2 # 1st phase
1576 vpalignr \$8,$Xi,$Xi,$Xi
1577 vpxor $T2,$Xi,$Xi
1579 vpclmulqdq \$0x10,$Tred,$Xi,$T2 # 2nd phase
1580 vpalignr \$8,$Xi,$Xi,$Xi
1581 vpxor $Xo,$Xi,$Xi
1582 vpxor $T2,$Xi,$Xi
1587 vpshufb $bswap,$Xi,$Xi
1588 vmovdqu $Xi,($Xip)