Lines Matching refs:D4

392 my ($D0,$D1,$D2,$D3,$D4,$T0,$T1,$T2)=map("xmm$_",(0..7));
397 &movdqu ($D4,&QWP(4*6,"edi")); # key base 2^32
403 #&pand ($D4,&QWP(96,"ebx")); # magic mask
406 &movdqa ($D0,$D4);
407 &movdqa ($D1,$D4);
408 &movdqa ($D2,$D4);
419 &psrldq ($D4,13);
428 &movdqa (&QWP(16*4,"esp"),$D4);
439 &movdqa ($T0,$D4);
443 &paddd ($T0,$D4); # *5
452 &pshufd ($D4,$D4,0b01000100);
457 &movdqa (&QWP(16*4,"edx"),$D4);
466 &pmuludq ($D4,$D0); # h4*r0
486 &paddq ($D4,$T0);
502 &paddq ($D4,$T2);
518 &paddq ($D4,$T1);
534 &paddq ($D4,$T0);
563 &paddq ($T0,$D4); # h3 -> h4
567 &movdqa ($D4,$T0);
570 &pand ($D4,$MASK);
592 &paddd ($D4,$T1); # h3 -> h4
603 &punpcklqdq ($D4,&QWP(16*4,"esp"));
611 &psllq ($D4,32);
616 &por ($D4,&QWP(16*4,"esp"));
622 &pshufd ($D4,$D4,0b10001101);
628 &movdqu (&QWP(16*4,"edi"),$D4);
639 &movdqa ($T0,$D4);
643 &paddd ($T0,$D4); # *5
696 &movd ($D4,"ebp");
707 &movd ($D4,&DWP(4*4,"edi"));
752 &paddd ($D4,$T1);
754 &paddd ($D4,$T0);
760 &movdqa (&QWP(16*4,"esp"),$D4);
774 &pmuludq ($D4,$T1); # h0*r0
850 &movdqa (&QWP($offbase+16*4,"esp"),$D4);
856 &movdqa ($D4,$T0);
858 &punpckhqdq ($D4,$T1); # 4
865 &psrlq ($D4,40); # 4
871 &por ($D4,&QWP(0,"ebx")); # padbit, yes, always
895 &movdqa (&QWP(16*4,"eax"),$D4);
910 &pmuludq ($D4,$T2); # h4*r0
931 &paddq ($D4,$T1);
940 &paddq ($D4,$T1);
953 &paddq ($D4,$T2);
966 &paddq ($D4,$T0);
988 &paddd ($D4,&QWP(16*(5+4),"esp"));
998 &movdqa (&QWP(16*4,"eax"),$D4);
1014 &pmuludq ($D4,$T2); # h4*r0
1019 &paddq ($D4,&QWP(16*4,"esp"));
1041 &paddd ($D4,&QWP(16*9,"esp"));
1049 &movdqa (&QWP(16*4,"eax"),$D4);
1065 &pmuludq ($D4,$T2); # h4*r0
1080 &paddd ($D4,&QWP(16*9,"esp"));
1100 &movdqa ($T1,$D4);
1101 &pmuludq ($D4,$T2); # h4*r0
1102 &paddq ($D4,&QWP(16*4,"esp"));
1114 &pshufd ($T1,$D4,0b01001110);
1116 &paddq ($D4,$T1);
1132 &movd (&DWP(-16*3+4*4,"edi"),$D4);
1238 &vmovdqu ($D4,&QWP(4*6,"edi")); # key base 2^32
1244 #&vpand ($D4,$D4,&QWP(96,"ebx")); # magic mask
1247 &vpand ($D0,$D4,$MASK); # -> base 2^26
1248 &vpsrlq ($D1,$D4,26);
1249 &vpsrldq ($D3,$D4,6);
1255 &vpsrldq ($D4,$D4,13);
1264 &vmovdqa (&QWP(16*4,"esp"),$D4);
1273 &vpslld ($T0,$D4,2);
1275 &vpaddd ($T0,$T0,$D4); # *5
1284 &vpshufd ($D4,$D4,0b01000100);
1289 &vmovdqa (&QWP(16*4,"edx"),$D4);
1298 &vpmuludq ($D4,$D4,$D0); # h4*r0
1305 &vpaddq ($D4,$D4,$T0);
1318 &vpaddq ($D4,$D4,$T1);
1331 &vpaddq ($D4,$D4,$T2);
1345 &vpaddq ($D4,$D4,$T2);
1362 &vpaddq ($D4,$D4,$T0); # h3 -> h4
1364 &vpsrlq ($T0,$D4,26);
1365 &vpand ($D4,$D4,$MASK);
1380 &vpaddd ($D4,$D4,$T1); # h3 -> h4
1389 &vpunpcklqdq ($D4,$D4,&QWP(16*4,"esp"));
1397 &vpsllq ($D4,$D4,32);
1402 &vpor ($D4,$D4,&QWP(16*4,"esp"));
1408 &vpshufd ($D4,$D4,0b10001101);
1414 &vmovdqu (&QWP(16*4,"edi"),$D4);
1423 &vpslld ($T0,$D4,2);
1425 &vpaddd ($T0,$T0,$D4); # *5
1437 my ($D0,$D1,$D2,$D3,$D4,$T0,$T1,$T2)=map("ymm$_",(0..7));
1513 &vmovdqu (&X($D4),&QWP(16*(3+4),"edi"));
1519 &vpermq ($D4,$D4,0b01000000);
1524 &vpshufd ($D4,$D4,0b11001000);
1533 &vmovdqa (&QWP(32*4-128,"edx"),$D4);
1550 &vmovd (&X($D4),&DWP(-16*3+4*4,"edi"));
1640 &vpaddq ($D1,$D1,$D4);
1651 &vpmuludq ($D4,$D2,&$addr(2)); # d4 = h2*r2
1661 &vpaddq ($D4,$D4,$T1); # d4 + h0*r4
1673 &vpaddq ($D4,$D4,$T0); # d4 += h1*r3
1685 &vpaddq ($D4,$D4,$T2); # d4 += h3*r1
1699 &vpaddq ($D4,$D4,$T2); # d4 += h4*r0
1716 &vpaddq ($D4,$D4,$T0); # h3 -> h4
1718 &vpsrlq ($T0,$D4,26);
1719 &vpand ($D4,$D4,$MASK);
1734 &vpaddq ($D4,$D4,$T1); # h3 -> h4
1755 &vpsrldq ($T0,$D4,8);
1757 &vpaddq ($D4,$D4,$T0);
1764 &vpermq ($T1,$D4,2); # keep folding
1767 &vpaddq ($D4,$D4,$T1);
1789 &vpshufd (&X($D4),&X($D4),0b11111100);
1797 &vmovd (&DWP(-16*3+4*4,"edi"),&X($D4));