/third_party/node/deps/openssl/openssl/crypto/sha/asm/ |
H A D | sha1-mb-x86_64.pl | 101 @Xi=map("%xmm$_",(10..14)); 107 @Xi=map("%xmm$_",(0..4)); 140 movd (@ptr[0]),@Xi[0] 142 movd (@ptr[1]),@Xi[2] # borrow @Xi[2] 144 movd (@ptr[2]),@Xi[3] # borrow @Xi[3] 146 movd (@ptr[3]),@Xi[4] # borrow @Xi[4] 148 punpckldq @Xi[ [all...] |
H A D | sha1-586.pl | 564 my $Xi=4; # 4xSIMD Xupdate round, start pre-seeded 565 my @X=map("xmm$_",(4..7,0..3)); # pre-seeded for $Xi=4 678 sub Xupdate_ssse3_16_31() # recall that $Xi starts with 4 693 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]);# save X[] to backtrace buffer 711 &movdqa (&QWP(0+16*(($Xi-1)&3),"esp"),@X[3]); # X[]+K xfer to IALU 740 &movdqa (@X[2],&QWP(64+16*(($Xi-6)%3),"esp")) if ($Xi>5); # restore X[] from backtrace buffer 748 &movdqa (@X[4],&QWP(112-16+16*(($Xi)/5),"esp")); # K_XX_XX 753 &pshufd (@X[1],@X[-3&7],0xee) if ($Xi<7); # was &movdqa (@X[1],@X[-2&7]) 754 &pshufd (@X[3],@X[-1&7],0xee) if ($Xi [all...] |
H A D | sha256-mb-x86_64.pl | 104 ($t1,$t2,$t3,$axb,$bxc,$Xi,$Xn,$sigma)=map("%xmm$_",(0..7)); 119 movd `4*$i`(@ptr[0]),$Xi 123 punpckldq $t2,$Xi 125 punpckldq $t1,$Xi 128 movd `4*$i`(@ptr[0]),$Xi 136 punpckldq $t2,$Xi 138 punpckldq $t1,$Xi 142 `"pshufb $Xn,$Xi" if ($i<=15 && ($i&1)==0)` 144 `"pshufb $Xn,$Xi" if ($i<=15 && ($i&1)==1)` 148 movdqa $Xi,` [all...] |
H A D | sha1-armv4-large.pl | 108 $Xi="r14"; 114 ldr $t0,[$Xi,#15*4] 115 ldr $t1,[$Xi,#13*4] 116 ldr $t2,[$Xi,#7*4] 118 ldr $t3,[$Xi,#2*4] 125 str $t0,[$Xi,#-4]! 158 str $t0,[$Xi,#-4]! 226 mov $Xi,sp 239 teq $Xi,$t3 241 teq $Xi,s [all...] |
H A D | sha1-sparcv9.pl | 40 $Xi="%g4"; 62 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 80 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 99 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 104 or $tmp1,$Xi,$Xi 106 xor $Xi,@X[$j%8],@X[$j%8] 107 srlx @X[$j%8],31,$Xi 109 and $Xi,$rot1m,$Xi [all...] |
H A D | sha1-x86_64.pl | 479 my $Xi=4; 583 sub Xupdate_ssse3_16_31() # recall that $Xi starts with 4 616 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 651 &movdqa (@Tx[2],eval(2*16*(($Xi)/5)-64)."($K_XX_XX)"); # K_XX_XX 657 &pshufd (@Tx[1],@X[-1&7],0xee) if ($Xi==7); # was &movdqa (@Tx[0],@X[-1&7]) in Xupdate_ssse3_32_79 661 $Xi++; push(@X,shift(@X)); # "rotate" X[] 671 eval(shift(@insns)) if ($Xi==8); 673 eval(shift(@insns)) if ($Xi==8); 685 if ($Xi%5) { 688 &movdqa (@Tx[2],eval(2*16*($Xi/ [all...] |
/third_party/openssl/crypto/sha/asm/ |
H A D | sha1-mb-x86_64.pl | 101 @Xi=map("%xmm$_",(10..14)); 107 @Xi=map("%xmm$_",(0..4)); 140 movd (@ptr[0]),@Xi[0] 142 movd (@ptr[1]),@Xi[2] # borrow @Xi[2] 144 movd (@ptr[2]),@Xi[3] # borrow @Xi[3] 146 movd (@ptr[3]),@Xi[4] # borrow @Xi[4] 148 punpckldq @Xi[ [all...] |
H A D | sha1-586.pl | 564 my $Xi=4; # 4xSIMD Xupdate round, start pre-seeded 565 my @X=map("xmm$_",(4..7,0..3)); # pre-seeded for $Xi=4 678 sub Xupdate_ssse3_16_31() # recall that $Xi starts with 4 693 &movdqa (&QWP(64+16*(($Xi-4)%3),"esp"),@X[-4&7]);# save X[] to backtrace buffer 711 &movdqa (&QWP(0+16*(($Xi-1)&3),"esp"),@X[3]); # X[]+K xfer to IALU 740 &movdqa (@X[2],&QWP(64+16*(($Xi-6)%3),"esp")) if ($Xi>5); # restore X[] from backtrace buffer 748 &movdqa (@X[4],&QWP(112-16+16*(($Xi)/5),"esp")); # K_XX_XX 753 &pshufd (@X[1],@X[-3&7],0xee) if ($Xi<7); # was &movdqa (@X[1],@X[-2&7]) 754 &pshufd (@X[3],@X[-1&7],0xee) if ($Xi [all...] |
H A D | sha256-mb-x86_64.pl | 104 ($t1,$t2,$t3,$axb,$bxc,$Xi,$Xn,$sigma)=map("%xmm$_",(0..7)); 119 movd `4*$i`(@ptr[0]),$Xi 123 punpckldq $t2,$Xi 125 punpckldq $t1,$Xi 128 movd `4*$i`(@ptr[0]),$Xi 136 punpckldq $t2,$Xi 138 punpckldq $t1,$Xi 142 `"pshufb $Xn,$Xi" if ($i<=15 && ($i&1)==0)` 144 `"pshufb $Xn,$Xi" if ($i<=15 && ($i&1)==1)` 148 movdqa $Xi,` [all...] |
H A D | sha1-armv4-large.pl | 108 $Xi="r14"; 114 ldr $t0,[$Xi,#15*4] 115 ldr $t1,[$Xi,#13*4] 116 ldr $t2,[$Xi,#7*4] 118 ldr $t3,[$Xi,#2*4] 125 str $t0,[$Xi,#-4]! 158 str $t0,[$Xi,#-4]! 226 mov $Xi,sp 239 teq $Xi,$t3 241 teq $Xi,s [all...] |
H A D | sha1-sparcv9.pl | 40 $Xi="%g4"; 62 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 80 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 99 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 104 or $tmp1,$Xi,$Xi 106 xor $Xi,@X[$j%8],@X[$j%8] 107 srlx @X[$j%8],31,$Xi 109 and $Xi,$rot1m,$Xi [all...] |
/third_party/node/deps/openssl/openssl/crypto/modes/asm/ |
H A D | ghash-x86_64.pl | 138 $Xi="%rdi"; 262 movzb 15($Xi),$Zlo 265 &loop ($Xi); 267 mov $Zlo,8($Xi) 268 mov $Zhi,($Xi) 345 &mov ($Zlo,"8($Xi)"); 346 &mov ($Zhi,"0($Xi)"); 356 &mov ("($Xi)",$Zhi); 357 &mov ("8($Xi)","%rdx"); 392 &mov ($dat,"$j($Xi)") i [all...] |
H A D | ghash-x86.pl | 273 &mov ($inp,&wparam(0)); # load Xi 276 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 283 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack 306 &mov ($Zll,&wparam(0)); # load Xi 313 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16] 344 &mov ($inp,&wparam(0)); # load Xi 434 &mov ($inp,&wparam(0)); # load Xi 446 &mov ($inp,&wparam(0)); # load Xi 458 &mov ($Zhh,&wparam(0)); # load Xi 472 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[1 [all...] |
H A D | aesni-gcm-x86_64.pl | 83 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8)); 141 vpxor $Z0,$Xi,$Xi # modulo-scheduled 152 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi] 216 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0] 230 vpclmulqdq \$0x10,$Hkey,$Xi, [all...] |
H A D | ghash-s390x.pl | 67 $Xi="%r2"; # argument block 103 la %r1,0($Xi) # H lies right after Xi in gcm128_context 115 aghi $Xi,-1 120 lg $Zlo,8+1($Xi) # Xi 136 la %r1,0($Xi) # H lies right after Xi in gcm128_context 149 aghi $Xi,-1 154 lg $Zlo,8+1($Xi) # X [all...] |
H A D | ghash-c64xplus.pl | 66 || MV $Xip,${xip} ; reassign Xi 70 || LDBU *++${xip}[15],$x1 ; Xi[15] 72 || LDBU *--${xip},$x0 ; Xi[14] 94 || MV $Xip,${xip} ; reassign Xi 113 || [B0] XOR $H0x,$Z0,$Z0 ; Xi^=inp 118 || [B0] SHRU $Z1,24,$xia ; Xi[15], avoid cross-path stall 120 || [B0] SHRU $Z1,16,$x0 ; Xi[14] 125 || [B0] MV $Z0,$xia ; Xi[15], avoid cross-path stall 127 || [B0] SHRU $Z0,8,$x0 ; Xi[14] 163 XORMPY $H0,$xia,$H0x ; 0 ; H·(Xi[ [all...] |
/third_party/openssl/crypto/modes/asm/ |
H A D | ghash-x86_64.pl | 138 $Xi="%rdi"; 262 movzb 15($Xi),$Zlo 265 &loop ($Xi); 267 mov $Zlo,8($Xi) 268 mov $Zhi,($Xi) 345 &mov ($Zlo,"8($Xi)"); 346 &mov ($Zhi,"0($Xi)"); 356 &mov ("($Xi)",$Zhi); 357 &mov ("8($Xi)","%rdx"); 392 &mov ($dat,"$j($Xi)") i [all...] |
H A D | ghash-x86.pl | 273 &mov ($inp,&wparam(0)); # load Xi 276 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 283 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack 306 &mov ($Zll,&wparam(0)); # load Xi 313 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16] 344 &mov ($inp,&wparam(0)); # load Xi 434 &mov ($inp,&wparam(0)); # load Xi 446 &mov ($inp,&wparam(0)); # load Xi 458 &mov ($Zhh,&wparam(0)); # load Xi 472 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[1 [all...] |
H A D | aesni-gcm-x86_64.pl | 83 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8)); 141 vpxor $Z0,$Xi,$Xi # modulo-scheduled 152 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi] 216 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0] 230 vpclmulqdq \$0x10,$Hkey,$Xi, [all...] |
H A D | ghash-s390x.pl | 67 $Xi="%r2"; # argument block 103 la %r1,0($Xi) # H lies right after Xi in gcm128_context 115 aghi $Xi,-1 120 lg $Zlo,8+1($Xi) # Xi 136 la %r1,0($Xi) # H lies right after Xi in gcm128_context 149 aghi $Xi,-1 154 lg $Zlo,8+1($Xi) # X [all...] |
H A D | ghash-c64xplus.pl | 66 || MV $Xip,${xip} ; reassign Xi 70 || LDBU *++${xip}[15],$x1 ; Xi[15] 72 || LDBU *--${xip},$x0 ; Xi[14] 94 || MV $Xip,${xip} ; reassign Xi 113 || [B0] XOR $H0x,$Z0,$Z0 ; Xi^=inp 118 || [B0] SHRU $Z1,24,$xia ; Xi[15], avoid cross-path stall 120 || [B0] SHRU $Z1,16,$x0 ; Xi[14] 125 || [B0] MV $Z0,$xia ; Xi[15], avoid cross-path stall 127 || [B0] SHRU $Z0,8,$x0 ; Xi[14] 163 XORMPY $H0,$xia,$H0x ; 0 ; H·(Xi[ [all...] |
/third_party/node/deps/openssl/openssl/crypto/modes/ |
H A D | gcm128.c | 104 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) in gcm_gmult_8bit() argument 107 const u8 *xi = (const u8 *)Xi + 15; in gcm_gmult_8bit() 181 if ((u8 *)Xi == xi) in gcm_gmult_8bit() 197 Xi[0] = BSWAP8(Z.hi); in gcm_gmult_8bit() 198 Xi[1] = BSWAP8(Z.lo); in gcm_gmult_8bit() 200 u8 *p = (u8 *)Xi; in gcm_gmult_8bit() 212 Xi[0] = Z.hi; in gcm_gmult_8bit() 213 Xi[1] = Z.lo; in gcm_gmult_8bit() 217 # define GCM_MUL(ctx) gcm_gmult_8bit(ctx->Xi.u,ctx->Htable) 300 static void gcm_gmult_4bit(u64 Xi[ argument 375 gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16], const u8 *inp, size_t len) global() argument 562 gcm_gmult_1bit(u64 Xi[2], const u64 H[2]) global() argument [all...] |
/third_party/openssl/crypto/modes/ |
H A D | gcm128.c | 104 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) in gcm_gmult_8bit() argument 107 const u8 *xi = (const u8 *)Xi + 15; in gcm_gmult_8bit() 181 if ((u8 *)Xi == xi) in gcm_gmult_8bit() 197 Xi[0] = BSWAP8(Z.hi); in gcm_gmult_8bit() 198 Xi[1] = BSWAP8(Z.lo); in gcm_gmult_8bit() 200 u8 *p = (u8 *)Xi; in gcm_gmult_8bit() 212 Xi[0] = Z.hi; in gcm_gmult_8bit() 213 Xi[1] = Z.lo; in gcm_gmult_8bit() 217 # define GCM_MUL(ctx) gcm_gmult_8bit(ctx->Xi.u,ctx->Htable) 300 static void gcm_gmult_4bit(u64 Xi[ argument 375 gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16], const u8 *inp, size_t len) global() argument 562 gcm_gmult_1bit(u64 Xi[2], const u64 H[2]) global() argument [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm/crypto/modes/ |
H A D | ghashv8-armx.S | 68 vld1.64 {q9},[r0] @ load Xi 77 INST(0x86,0x0e,0xa8,0xf2) @ pmull q0,q12,q3 @ H.lo·Xi.lo 79 INST(0x87,0x4e,0xa9,0xf2) @ pmull2 q2,q12,q3 @ H.hi·Xi.hi 80 INST(0xa2,0x2e,0xaa,0xf2) @ pmull q1,q13,q9 @ (H.lo+H.hi)·(Xi.lo+Xi.hi) 101 vst1.64 {q0},[r0] @ write out Xi 110 vld1.64 {q0},[r0] @ load [rotated] Xi 131 vext.8 q0,q0,q0,#8 @ rotate Xi 145 veor q3,q3,q0 @ I[i]^=Xi 155 INST(0x86,0x0e,0xac,0xf2) @ pmull q0,q14,q3 @ H^2.lo·Xi [all...] |
/third_party/node/deps/openssl/config/archs/linux-armv4/asm_avx2/crypto/modes/ |
H A D | ghashv8-armx.S | 68 vld1.64 {q9},[r0] @ load Xi 77 INST(0x86,0x0e,0xa8,0xf2) @ pmull q0,q12,q3 @ H.lo·Xi.lo 79 INST(0x87,0x4e,0xa9,0xf2) @ pmull2 q2,q12,q3 @ H.hi·Xi.hi 80 INST(0xa2,0x2e,0xaa,0xf2) @ pmull q1,q13,q9 @ (H.lo+H.hi)·(Xi.lo+Xi.hi) 101 vst1.64 {q0},[r0] @ write out Xi 110 vld1.64 {q0},[r0] @ load [rotated] Xi 131 vext.8 q0,q0,q0,#8 @ rotate Xi 145 veor q3,q3,q0 @ I[i]^=Xi 155 INST(0x86,0x0e,0xac,0xf2) @ pmull q0,q14,q3 @ H^2.lo·Xi [all...] |