/third_party/node/deps/openssl/openssl/crypto/chacha/asm/ |
H A D | chacha-x86_64.pl | 780 my ($xd0,$xd1,$xd2,$xd3, $xt0,$xt1,$xt2,$xt3, 790 my ($xc,$xc_,$t0,$t1)=map("\"$_\"",$xt0,$xt1,$xt2,$xt3); 986 pshufd \$0xaa,$xt3,$xt2 # "$xc2" 989 movdqa $xt2,0xe0-0x100(%rcx) 1015 movdqa 0xe0-0x100(%rcx),$xt2 # "$xc2" 1024 movdqa $xt2,0x20(%rsp) # SIMD equivalent of "@x[10]" 1045 movdqa $xa0,$xt2 # "de-interlace" data 1049 punpckhdq $xa1,$xt2 1053 movdqa $xt2,$xa3 1054 punpcklqdq $xt3,$xt2 # "a [all...] |
H A D | chacha-s390x.pl | 445 my ($xt0,$xt1,$xt2,$xt3)=map("%v$_",(27..30)); 514 vmrlf ($xt2,$xa0,$xa1); 518 vpdi ($xa2,$xt2,$xt3,0b0000); 519 vpdi ($xa3,$xt2,$xt3,0b0101); 523 vmrlf ($xt2,$xb0,$xb1); 527 vpdi ($xb2,$xt2,$xt3,0b0000); 528 vpdi ($xb3,$xt2,$xt3,0b0101); 532 vmrlf ($xt2,$xc0,$xc1); 536 vpdi ($xc2,$xt2,$xt3,0b0000); 537 vpdi ($xc3,$xt2, [all...] |
H A D | chacha-ppc.pl | 919 my ($xt0,$xt1,$xt2,$xt3) = map("v$_",(27..30)); 920 my ($sixteen,$twelve,$eight,$seven) = ($xt0,$xt1,$xt2,$xt3); 1081 vmrgew $xt2,$xb0,$xb1 1094 vpermdi $xb0,$xt2,$xt3,0b00 1095 vpermdi $xb2,$xt2,$xt3,0b11 1099 vmrgew $xt2,$xd0,$xd1 1112 vpermdi $xd0,$xt2,$xt3,0b00 1113 vpermdi $xd2,$xt2,$xt3,0b11 1130 lvx_4w $xt2,$x20,$inp 1135 vxor $xt2, [all...] |
H A D | chacha-x86.pl | 592 my ($xa0,$xa1,$xa2,$xa3,$xt0,$xt1,$xt2,$xt3)=map("xmm$_",(0..7)); 605 &movdqa ($xt2,$xa0); # "de-interlace" data 609 &punpckhdq ($xt2,$xa1); 613 &movdqa ($xa3,$xt2); 614 &punpcklqdq ($xt2,$xt3); # "a2" 618 #($xa2,$xt2)=($xt2,$xa2); 629 &pxor ($xt2,$xa2); 635 &movdqu (&QWP(64*2-128,$out),$xt2); 988 my ($xa0,$xa1,$xa2,$xa3,$xt0,$xt1,$xt2, [all...] |
H A D | chacha-armv8.pl | 352 my ($xt0,$xt1,$xt2,$xt3, $CTR,$ROT24) = map("v$_.4s",(4..9)); 384 "&eor ('$xt2',@x[$b2],@x[$c2])", 388 "&ushr (@x[$b2],'$xt2',20)", 392 "&sli (@x[$b2],'$xt2',12)", 401 "&eor ('$xt2',@x[$d2],@x[$a2])", 405 "&tbl (@x[$d2],'{$xt2}','$ROT24')", 414 "&eor ('$xt2',@x[$b2],@x[$c2])", 418 "&ushr (@x[$b2],'$xt2',25)", 422 "&sli (@x[$b2],'$xt2',7)", 522 zip2 $xt2, [all...] |
/third_party/openssl/crypto/chacha/asm/ |
H A D | chacha-x86_64.pl | 780 my ($xd0,$xd1,$xd2,$xd3, $xt0,$xt1,$xt2,$xt3, 790 my ($xc,$xc_,$t0,$t1)=map("\"$_\"",$xt0,$xt1,$xt2,$xt3); 986 pshufd \$0xaa,$xt3,$xt2 # "$xc2" 989 movdqa $xt2,0xe0-0x100(%rcx) 1015 movdqa 0xe0-0x100(%rcx),$xt2 # "$xc2" 1024 movdqa $xt2,0x20(%rsp) # SIMD equivalent of "@x[10]" 1045 movdqa $xa0,$xt2 # "de-interlace" data 1049 punpckhdq $xa1,$xt2 1053 movdqa $xt2,$xa3 1054 punpcklqdq $xt3,$xt2 # "a [all...] |
H A D | chacha-s390x.pl | 445 my ($xt0,$xt1,$xt2,$xt3)=map("%v$_",(27..30)); 514 vmrlf ($xt2,$xa0,$xa1); 518 vpdi ($xa2,$xt2,$xt3,0b0000); 519 vpdi ($xa3,$xt2,$xt3,0b0101); 523 vmrlf ($xt2,$xb0,$xb1); 527 vpdi ($xb2,$xt2,$xt3,0b0000); 528 vpdi ($xb3,$xt2,$xt3,0b0101); 532 vmrlf ($xt2,$xc0,$xc1); 536 vpdi ($xc2,$xt2,$xt3,0b0000); 537 vpdi ($xc3,$xt2, [all...] |
H A D | chacha-ppc.pl | 919 my ($xt0,$xt1,$xt2,$xt3) = map("v$_",(27..30)); 920 my ($sixteen,$twelve,$eight,$seven) = ($xt0,$xt1,$xt2,$xt3); 1081 vmrgew $xt2,$xb0,$xb1 1094 vpermdi $xb0,$xt2,$xt3,0b00 1095 vpermdi $xb2,$xt2,$xt3,0b11 1099 vmrgew $xt2,$xd0,$xd1 1112 vpermdi $xd0,$xt2,$xt3,0b00 1113 vpermdi $xd2,$xt2,$xt3,0b11 1130 lvx_4w $xt2,$x20,$inp 1135 vxor $xt2, [all...] |
H A D | chacha-x86.pl | 592 my ($xa0,$xa1,$xa2,$xa3,$xt0,$xt1,$xt2,$xt3)=map("xmm$_",(0..7)); 605 &movdqa ($xt2,$xa0); # "de-interlace" data 609 &punpckhdq ($xt2,$xa1); 613 &movdqa ($xa3,$xt2); 614 &punpcklqdq ($xt2,$xt3); # "a2" 618 #($xa2,$xt2)=($xt2,$xa2); 629 &pxor ($xt2,$xa2); 635 &movdqu (&QWP(64*2-128,$out),$xt2); 988 my ($xa0,$xa1,$xa2,$xa3,$xt0,$xt1,$xt2, [all...] |
H A D | chacha-armv8.pl | 355 my ($xt0,$xt1,$xt2,$xt3, $CTR,$ROT24) = map("v$_.4s",(4..9)); 387 "&eor ('$xt2',@x[$b2],@x[$c2])", 391 "&ushr (@x[$b2],'$xt2',20)", 395 "&sli (@x[$b2],'$xt2',12)", 404 "&eor ('$xt2',@x[$d2],@x[$a2])", 408 "&tbl (@x[$d2],'{$xt2}','$ROT24')", 417 "&eor ('$xt2',@x[$b2],@x[$c2])", 421 "&ushr (@x[$b2],'$xt2',25)", 425 "&sli (@x[$b2],'$xt2',7)", 526 zip2 $xt2, [all...] |
/third_party/mesa3d/src/intel/isl/ |
H A D | isl_tiled_memcpy_normal.c | 36 _isl_memcpy_linear_to_tiled(uint32_t xt1, uint32_t xt2, in _isl_memcpy_linear_to_tiled() argument 44 linear_to_tiled(xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, in _isl_memcpy_linear_to_tiled() 49 _isl_memcpy_tiled_to_linear(uint32_t xt1, uint32_t xt2, in _isl_memcpy_tiled_to_linear() argument 57 tiled_to_linear(xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, in _isl_memcpy_tiled_to_linear()
|
H A D | isl_tiled_memcpy_sse41.c | 37 _isl_memcpy_linear_to_tiled_sse41(uint32_t xt1, uint32_t xt2, in _isl_memcpy_linear_to_tiled_sse41() argument 45 linear_to_tiled(xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, in _isl_memcpy_linear_to_tiled_sse41() 50 _isl_memcpy_tiled_to_linear_sse41(uint32_t xt1, uint32_t xt2, in _isl_memcpy_tiled_to_linear_sse41() argument 58 tiled_to_linear(xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, in _isl_memcpy_tiled_to_linear_sse41()
|
H A D | isl_priv.h | 165 _isl_memcpy_linear_to_tiled(uint32_t xt1, uint32_t xt2, 174 _isl_memcpy_tiled_to_linear(uint32_t xt1, uint32_t xt2, 183 _isl_memcpy_linear_to_tiled_sse41(uint32_t xt1, uint32_t xt2, 192 _isl_memcpy_tiled_to_linear_sse41(uint32_t xt1, uint32_t xt2,
|
H A D | isl_tiled_memcpy.c | 822 * Divide the region given by X range [xt1, xt2) and Y range [yt1, yt2) into 831 linear_to_tiled(uint32_t xt1, uint32_t xt2, in linear_to_tiled() argument 862 xt3 = ALIGN_UP (xt2, tw); in linear_to_tiled() 879 uint32_t x3 = MIN2(xt2, xt + tw); in linear_to_tiled() 913 * Divide the region given by X range [xt1, xt2) and Y range [yt1, yt2) into 922 tiled_to_linear(uint32_t xt1, uint32_t xt2, in tiled_to_linear() argument 962 xt3 = ALIGN_UP (xt2, tw); in tiled_to_linear() 979 uint32_t x3 = MIN2(xt2, xt + tw); in tiled_to_linear()
|
H A D | isl.c | 40 isl_memcpy_linear_to_tiled(uint32_t xt1, uint32_t xt2, in isl_memcpy_linear_to_tiled() argument 51 xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, has_swizzling, in isl_memcpy_linear_to_tiled() 58 xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, has_swizzling, in isl_memcpy_linear_to_tiled() 63 isl_memcpy_tiled_to_linear(uint32_t xt1, uint32_t xt2, in isl_memcpy_tiled_to_linear() argument 74 xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, has_swizzling, in isl_memcpy_tiled_to_linear() 81 xt1, xt2, yt1, yt2, dst, src, dst_pitch, src_pitch, has_swizzling, in isl_memcpy_tiled_to_linear()
|
H A D | isl.h | 2899 isl_memcpy_linear_to_tiled(uint32_t xt1, uint32_t xt2, 2912 isl_memcpy_tiled_to_linear(uint32_t xt1, uint32_t xt2,
|
/third_party/vixl/src/aarch64/ |
H A D | assembler-aarch64.cc | 1112 const Register& xt2, in ldpsw() 1114 VIXL_ASSERT(xt.Is64Bits() && xt2.Is64Bits()); in ldpsw() 1115 LoadStorePair(xt, xt2, src, LDPSW_x); in ldpsw() 2014 const Register& xt2, in stgp() 2032 Rt2(xt2) | Rt(xt1)); in stgp() 1111 ldpsw(const Register& xt, const Register& xt2, const MemOperand& src) ldpsw() argument 2013 stgp(const Register& xt1, const Register& xt2, const MemOperand& addr) stgp() argument
|
H A D | assembler-aarch64.h | 1298 void ldpsw(const Register& xt, const Register& xt2, const MemOperand& src); 6951 void stgp(const Register& xt1, const Register& xt2, const MemOperand& addr);
|