Home
last modified time | relevance | path

Searched refs:tp1 (Results 1 - 25 of 28) sorted by relevance

12

/third_party/ffmpeg/libavcodec/mips/
H A Dh264chroma_msa.c889 uint16_t tp0, tp1, tp2, tp3; in avc_chroma_hz_and_aver_dst_2x4_msa() local
902 tp1 = LH(dst + stride); in avc_chroma_hz_and_aver_dst_2x4_msa()
906 dst_data = (v16u8) __msa_insert_h((v8i16) dst_data, 1, tp1); in avc_chroma_hz_and_aver_dst_2x4_msa()
973 uint32_t tp0, tp1, tp2, tp3; in avc_chroma_hz_and_aver_dst_4x4_msa() local
985 LW4(dst, stride, tp0, tp1, tp2, tp3); in avc_chroma_hz_and_aver_dst_4x4_msa()
986 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst_data); in avc_chroma_hz_and_aver_dst_4x4_msa()
1002 uint32_t tp0, tp1, tp2, tp3; in avc_chroma_hz_and_aver_dst_4x8_msa() local
1014 LW4(dst, stride, tp0, tp1, tp2, tp3); in avc_chroma_hz_and_aver_dst_4x8_msa()
1015 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in avc_chroma_hz_and_aver_dst_4x8_msa()
1016 LW4(dst + 4 * stride, stride, tp0, tp1, tp in avc_chroma_hz_and_aver_dst_4x8_msa()
1047 uint64_t tp0, tp1, tp2, tp3; avc_chroma_hz_and_aver_dst_8x4_msa() local
1077 uint64_t tp0, tp1, tp2, tp3; avc_chroma_hz_and_aver_dst_8x8_msa() local
1169 uint16_t tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_2x4_msa() local
1254 uint32_t tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_4x4_msa() local
1284 uint32_t tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_4x8_msa() local
1333 uint64_t tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_8x4_msa() local
1362 uint64_t tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_8x8_msa() local
1461 uint16_t tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_2x4_msa() local
1523 uint32_t tp0, tp1; avc_chroma_hv_and_aver_dst_4x2_msa() local
1559 uint32_t tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_4x4_msa() local
1598 uint32_t tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_4x8_msa() local
1667 uint64_t tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_8x4_msa() local
1715 uint64_t tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_8x8_msa() local
1797 uint32_t tp0, tp1, tp2, tp3, tp4, tp5, tp6, tp7; copy_width4_msa() local
1838 uint32_t tp0, tp1, tp2, tp3; avg_width4_msa() local
1873 uint64_t tp0, tp1, tp2, tp3, tp4, tp5, tp6, tp7; avg_width8_msa() local
[all...]
H A Dh264dsp_msa.c28 uint32_t tp0, tp1, offset_val; in avc_wgt_4x2_msa() local
39 LW2(data, stride, tp0, tp1); in avc_wgt_4x2_msa()
40 INSERT_W2_UB(tp0, tp1, src0); in avc_wgt_4x2_msa()
55 uint32_t tp0, tp1, tp2, tp3, offset_val; in avc_wgt_4x4_msa() local
65 LW4(data, stride, tp0, tp1, tp2, tp3); in avc_wgt_4x4_msa()
66 INSERT_W4_UB(tp0, tp1, tp2, tp3, src0); in avc_wgt_4x4_msa()
82 uint32_t tp0, tp1, tp2, tp3, offset_val; in avc_wgt_4x8_msa() local
93 LW4(data, stride, tp0, tp1, tp2, tp3); in avc_wgt_4x8_msa()
94 INSERT_W4_UB(tp0, tp1, tp2, tp3, src0); in avc_wgt_4x8_msa()
95 LW4(data + 4 * stride, stride, tp0, tp1, tp in avc_wgt_4x8_msa()
115 uint64_t tp0, tp1, tp2, tp3; avc_wgt_8x4_msa() local
146 uint64_t tp0, tp1, tp2, tp3; avc_wgt_8x8_msa() local
189 uint64_t tp0, tp1, tp2, tp3; avc_wgt_8x16_msa() local
234 uint32_t tp0, tp1; avc_biwgt_4x2_msa() local
267 uint32_t tp0, tp1, tp2, tp3; avc_biwgt_4x4_msa() local
301 uint32_t tp0, tp1, tp2, tp3; avc_biwgt_4x8_msa() local
341 uint64_t tp0, tp1, tp2, tp3; avc_biwgt_8x4_msa() local
379 uint64_t tp0, tp1, tp2, tp3; avc_biwgt_8x8_msa() local
431 uint64_t tp0, tp1, tp2, tp3; avc_biwgt_8x16_msa() local
950 v8i16 tp0, tp1, tp2, tp3, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7; avc_loopfilter_luma_intra_edge_ver_msa() local
[all...]
H A Dvp9_mc_msa.c991 uint32_t tp0, tp1, tp2, tp3; in common_hz_8t_and_aver_dst_4x4_msa() local
1012 LW4(dst, dst_stride, tp0, tp1, tp2, tp3); in common_hz_8t_and_aver_dst_4x4_msa()
1013 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in common_hz_8t_and_aver_dst_4x4_msa()
1026 uint32_t tp0, tp1, tp2, tp3; in common_hz_8t_and_aver_dst_4x8_msa() local
1046 LW4(dst, dst_stride, tp0, tp1, tp2, tp3); in common_hz_8t_and_aver_dst_4x8_msa()
1047 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in common_hz_8t_and_aver_dst_4x8_msa()
1048 LW4(dst + 4 * dst_stride, dst_stride, tp0, tp1, tp2, tp3); in common_hz_8t_and_aver_dst_4x8_msa()
1049 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst1); in common_hz_8t_and_aver_dst_4x8_msa()
1088 int64_t tp0, tp1, tp2, tp3; in common_hz_8t_and_aver_dst_8w_msa() local
1111 LD4(dst, dst_stride, tp0, tp1, tp in common_hz_8t_and_aver_dst_8w_msa()
1308 uint32_t tp0, tp1, tp2, tp3; common_vt_8t_and_aver_dst_4w_msa() local
1367 uint64_t tp0, tp1, tp2, tp3; common_vt_8t_and_aver_dst_8w_msa() local
1557 uint32_t tp0, tp1, tp2, tp3; common_hv_8ht_8vt_and_aver_dst_4w_msa() local
1640 uint64_t tp0, tp1, tp2, tp3; common_hv_8ht_8vt_and_aver_dst_8w_msa() local
2823 uint32_t tp0, tp1, tp2, tp3; common_hz_2t_and_aver_dst_4x4_msa() local
2852 uint32_t tp0, tp1, tp2, tp3; common_hz_2t_and_aver_dst_4x8_msa() local
2901 int64_t tp0, tp1, tp2, tp3; common_hz_2t_and_aver_dst_8x4_msa() local
2931 int64_t tp0, tp1, tp2, tp3; common_hz_2t_and_aver_dst_8x8mult_msa() local
3176 uint32_t tp0, tp1, tp2, tp3; common_vt_2t_and_aver_dst_4x4_msa() local
3212 uint32_t tp0, tp1, tp2, tp3; common_vt_2t_and_aver_dst_4x8_msa() local
3267 int64_t tp0, tp1, tp2, tp3; common_vt_2t_and_aver_dst_8x4_msa() local
3298 int64_t tp0, tp1, tp2, tp3; common_vt_2t_and_aver_dst_8x8mult_msa() local
3588 uint32_t tp0, tp1, tp2, tp3; common_hv_2ht_2vt_and_aver_dst_4x4_msa() local
3632 uint32_t tp0, tp1, tp2, tp3; common_hv_2ht_2vt_and_aver_dst_4x8_msa() local
3700 uint64_t tp0, tp1, tp2, tp3; common_hv_2ht_2vt_and_aver_dst_8x4_msa() local
3752 uint64_t tp0, tp1, tp2, tp3; common_hv_2ht_2vt_and_aver_dst_8x8mult_msa() local
4059 uint32_t tp0, tp1, tp2, tp3; avg_width4_msa() local
4089 uint64_t tp0, tp1, tp2, tp3, tp4, tp5, tp6, tp7; avg_width8_msa() local
[all...]
H A Dh264qpel_msa.c360 uint32_t tp0, tp1, tp2, tp3; in avc_luma_hv_qrt_and_aver_dst_4x4_msa() local
411 LW4(dst, stride, tp0, tp1, tp2, tp3); in avc_luma_hv_qrt_and_aver_dst_4x4_msa()
412 INSERT_W4_UB(tp0, tp1, tp2, tp3, dst0); in avc_luma_hv_qrt_and_aver_dst_4x4_msa()
432 uint64_t tp0, tp1, tp2, tp3; in avc_luma_hv_qrt_and_aver_dst_8x8_msa() local
492 LD4(dst, stride, tp0, tp1, tp2, tp3); in avc_luma_hv_qrt_and_aver_dst_8x8_msa()
493 INSERT_D2_UB(tp0, tp1, dst0); in avc_luma_hv_qrt_and_aver_dst_8x8_msa()
533 LD4(dst, stride, tp0, tp1, tp2, tp3); in avc_luma_hv_qrt_and_aver_dst_8x8_msa()
534 INSERT_D2_UB(tp0, tp1, dst0); in avc_luma_hv_qrt_and_aver_dst_8x8_msa()
556 uint64_t tp0, tp1, tp2, tp3; in avc_luma_hv_qrt_and_aver_dst_16x16_msa() local
620 LD4(dst, stride, tp0, tp1, tp in avc_luma_hv_qrt_and_aver_dst_16x16_msa()
702 uint64_t tp0, tp1, tp2, tp3, tp4, tp5, tp6, tp7; ff_avg_h264_qpel8_mc00_msa() local
730 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc00_msa() local
3292 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc10_msa() local
3357 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc30_msa() local
3422 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc10_msa() local
3458 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc30_msa() local
3557 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc20_msa() local
3615 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc20_msa() local
3797 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc01_msa() local
3863 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc03_msa() local
3929 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc01_msa() local
3974 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc03_msa() local
4119 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel16_mc21_msa() local
4251 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel16_mc23_msa() local
4383 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc21_msa() local
4529 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc23_msa() local
4672 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc21_msa() local
4742 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc23_msa() local
4877 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc02_msa() local
4943 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc02_msa() local
5154 uint64_t tp0, tp1; ff_avg_h264_qpel8_mc12_msa() local
5233 uint64_t tp0, tp1; ff_avg_h264_qpel8_mc32_msa() local
5312 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc12_msa() local
5408 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc32_msa() local
5512 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel16_mc22_msa() local
5614 uint64_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel8_mc22_msa() local
5736 uint32_t tp0, tp1, tp2, tp3; ff_avg_h264_qpel4_mc22_msa() local
[all...]
H A Dhevc_mc_bi_msa.c69 uint32_t loop_cnt, tp0, tp1, tp2, tp3; in hevc_bi_copy_4w_msa() local
77 LW2(src0_ptr, src_stride, tp0, tp1); in hevc_bi_copy_4w_msa()
78 INSERT_W2_SB(tp0, tp1, src0); in hevc_bi_copy_4w_msa()
91 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_bi_copy_4w_msa()
92 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_bi_copy_4w_msa()
103 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_bi_copy_4w_msa()
105 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_bi_copy_4w_msa()
106 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_bi_copy_4w_msa()
108 INSERT_W4_SB(tp0, tp1, tp2, tp3, src1); in hevc_bi_copy_4w_msa()
138 uint64_t tp0, tp1, tp in hevc_bi_copy_6w_msa() local
189 uint64_t tp0, tp1, tp2, tp3; hevc_bi_copy_8w_msa() local
1630 uint64_t tp0, tp1; hevc_hv_bi_8t_4w_msa() local
1905 uint64_t tp0, tp1; hevc_hv_bi_8t_12w_msa() local
3793 uint64_t tp0, tp1; hevc_hv_bi_4t_4x2_msa() local
3860 uint64_t tp0, tp1; hevc_hv_bi_4t_4x4_msa() local
3938 uint64_t tp0, tp1; hevc_hv_bi_4t_4multx8mult_msa() local
4081 uint64_t tp0, tp1; hevc_hv_bi_4t_6w_msa() local
4669 uint64_t tp0, tp1; hevc_hv_bi_4t_12w_msa() local
[all...]
H A Dhevc_mc_biw_msa.c95 uint32_t loop_cnt, tp0, tp1, tp2, tp3; in hevc_biwgt_copy_4w_msa() local
114 LW2(src0_ptr, src_stride, tp0, tp1); in hevc_biwgt_copy_4w_msa()
115 INSERT_W2_SB(tp0, tp1, src0); in hevc_biwgt_copy_4w_msa()
131 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_biwgt_copy_4w_msa()
132 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_biwgt_copy_4w_msa()
144 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_biwgt_copy_4w_msa()
146 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_biwgt_copy_4w_msa()
147 LW4(src0_ptr, src_stride, tp0, tp1, tp2, tp3); in hevc_biwgt_copy_4w_msa()
149 INSERT_W4_SB(tp0, tp1, tp2, tp3, src1); in hevc_biwgt_copy_4w_msa()
186 uint64_t tp0, tp1, tp in hevc_biwgt_copy_6w_msa() local
238 uint64_t tp0, tp1, tp2, tp3; hevc_biwgt_copy_8w_msa() local
1938 uint64_t tp0, tp1; hevc_hv_biwgt_8t_4w_msa() local
2291 uint64_t tp0, tp1; hevc_hv_biwgt_8t_12w_msa() local
4513 uint64_t tp0, tp1; hevc_hv_biwgt_4t_4x2_msa() local
4597 uint64_t tp0, tp1; hevc_hv_biwgt_4t_4x4_msa() local
4696 uint64_t tp0, tp1; hevc_hv_biwgt_4t_4multx8mult_msa() local
4872 uint64_t tp0, tp1; hevc_hv_biwgt_4t_6w_msa() local
5611 uint64_t tp0, tp1; hevc_hv_biwgt_4t_12w_msa() local
[all...]
H A Dhevc_mc_uniw_msa.c66 uint32_t loop_cnt, tp0, tp1, tp2, tp3; in hevc_uniwgt_copy_4w_msa() local
81 LW2(src, src_stride, tp0, tp1); in hevc_uniwgt_copy_4w_msa()
82 INSERT_W2_SB(tp0, tp1, src0); in hevc_uniwgt_copy_4w_msa()
95 LW4(src, src_stride, tp0, tp1, tp2, tp3); in hevc_uniwgt_copy_4w_msa()
96 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_uniwgt_copy_4w_msa()
105 LW4(src, src_stride, tp0, tp1, tp2, tp3); in hevc_uniwgt_copy_4w_msa()
107 INSERT_W4_SB(tp0, tp1, tp2, tp3, src0); in hevc_uniwgt_copy_4w_msa()
108 LW4(src, src_stride, tp0, tp1, tp2, tp3); in hevc_uniwgt_copy_4w_msa()
110 INSERT_W4_SB(tp0, tp1, tp2, tp3, src1); in hevc_uniwgt_copy_4w_msa()
134 uint64_t tp0, tp1, tp in hevc_uniwgt_copy_6w_msa() local
196 uint64_t tp0, tp1, tp2, tp3; hevc_uniwgt_copy_8w_msa() local
[all...]
/third_party/node/deps/openssl/openssl/crypto/bn/asm/
H A Drsaz-avx2.pl123 my $tp1=$r3;
228 lea 448(%rsp), $tp1 # 64+128+256=448
256 vmovdqu $ACC9, 32*10-448($tp1)
258 vmovdqu $ACC9, 32*11-448($tp1)
260 vmovdqu $ACC9, 32*12-448($tp1)
262 vmovdqu $ACC9, 32*13-448($tp1)
264 vmovdqu $ACC9, 32*14-448($tp1)
266 vmovdqu $ACC9, 32*15-448($tp1)
268 vmovdqu $ACC9, 32*16-448($tp1)
271 vmovdqu $ACC9, 32*17-448($tp1)
[all...]
/third_party/openssl/crypto/bn/asm/
H A Drsaz-avx2.pl123 my $tp1=$r3;
228 lea 448(%rsp), $tp1 # 64+128+256=448
256 vmovdqu $ACC9, 32*10-448($tp1)
258 vmovdqu $ACC9, 32*11-448($tp1)
260 vmovdqu $ACC9, 32*12-448($tp1)
262 vmovdqu $ACC9, 32*13-448($tp1)
264 vmovdqu $ACC9, 32*14-448($tp1)
266 vmovdqu $ACC9, 32*15-448($tp1)
268 vmovdqu $ACC9, 32*16-448($tp1)
271 vmovdqu $ACC9, 32*17-448($tp1)
[all...]
/third_party/ltp/testcases/open_posix_testsuite/conformance/interfaces/clock_getcpuclockid/
H A D1-2.c29 struct timespec tp1; in main()
48 if (clock_gettime(clockid, &tp1) != 0) { in main()
H A D1-1.c38 struct timespec tp1; in main()
57 if (clock_gettime(clockid, &tp1) != 0) { in main()
/third_party/ffmpeg/libavcodec/loongarch/
H A Dh264chroma_lasx.c735 uint32_t tp0, tp1, tp2, tp3, tp4, tp5, tp6, tp7; in copy_width4_lasx() local
745 "ldx.wu %[tp1], %[src], %[stride] \n\t" in copy_width4_lasx()
754 "stx.w %[tp1], %[dst], %[stride] \n\t" in copy_width4_lasx()
763 [src]"+&r"(src), [dst]"+&r"(dst), [tp0]"+&r"(tp0), [tp1]"+&r"(tp1), in copy_width4_lasx()
776 "ldx.wu %[tp1], %[src], %[stride] \n\t" in copy_width4_lasx()
780 "stx.w %[tp1], %[dst], %[stride] \n\t" in copy_width4_lasx()
784 [src]"+&r"(src), [dst]"+&r"(dst), [tp0]"+&r"(tp0), [tp1]"+&r"(tp1), in copy_width4_lasx()
792 "ldx.wu %[tp1], in copy_width4_lasx()
852 __m256i tp0, tp1, tp2, tp3; avc_chroma_hv_and_aver_dst_8x4_lasx() local
896 __m256i tp0, tp1, tp2, tp3, dst0, dst1; avc_chroma_hv_and_aver_dst_8x8_lasx() local
966 __m256i tp0, tp1, tp2, tp3; avc_chroma_hz_and_aver_dst_8x4_lasx() local
1000 __m256i tp0, tp1, tp2, tp3, dst0, dst1; avc_chroma_hz_and_aver_dst_8x8_lasx() local
1053 __m256i tp0, tp1, tp2, tp3; avc_chroma_vt_and_aver_dst_8x4_lasx() local
1087 __m256i tp0, tp1, tp2, tp3, dst0, dst1; avc_chroma_vt_and_aver_dst_8x8_lasx() local
[all...]
/third_party/node/deps/openssl/openssl/crypto/aes/
H A Daes_x86core.c603 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; in AES_set_decrypt_key() local
605 tp1 = rk[j]; in AES_set_decrypt_key()
606 m = tp1 & 0x80808080; in AES_set_decrypt_key()
607 tp2 = ((tp1 & 0x7f7f7f7f) << 1) ^ in AES_set_decrypt_key()
615 tp9 = tp8 ^ tp1; in AES_set_decrypt_key()
915 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; in AES_decrypt() local
918 tp1 = t[i]; in AES_decrypt()
919 m = tp1 & 0x80808080; in AES_decrypt()
920 tp2 = ((tp1 & 0x7f7f7f7f) << 1) ^ in AES_decrypt()
928 tp9 = tp8 ^ tp1; in AES_decrypt()
992 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; AES_decrypt() local
[all...]
H A Daes_core.c1984 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; in AES_set_decrypt_key() local
1986 tp1 = rk[j]; in AES_set_decrypt_key()
1987 m = tp1 & 0x80808080; in AES_set_decrypt_key()
1988 tp2 = ((tp1 & 0x7f7f7f7f) << 1) ^ in AES_set_decrypt_key()
1996 tp9 = tp8 ^ tp1; in AES_set_decrypt_key()
/third_party/openssl/crypto/aes/
H A Daes_x86core.c603 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; in AES_set_decrypt_key() local
605 tp1 = rk[j]; in AES_set_decrypt_key()
606 m = tp1 & 0x80808080; in AES_set_decrypt_key()
607 tp2 = ((tp1 & 0x7f7f7f7f) << 1) ^ in AES_set_decrypt_key()
615 tp9 = tp8 ^ tp1; in AES_set_decrypt_key()
915 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; in AES_decrypt() local
918 tp1 = t[i]; in AES_decrypt()
919 m = tp1 & 0x80808080; in AES_decrypt()
920 tp2 = ((tp1 & 0x7f7f7f7f) << 1) ^ in AES_decrypt()
928 tp9 = tp8 ^ tp1; in AES_decrypt()
992 u32 tp1, tp2, tp4, tp8, tp9, tpb, tpd, tpe, m; AES_decrypt() local
[all...]
/third_party/node/deps/openssl/openssl/crypto/aes/asm/
H A Daes-mips.pl1670 my ($tp1,$tp2,$tp4,$tp8,$tp9,$tpb,$tpd,$tpe)=($a4,$a5,$a6,$a7,$s0,$s1,$s2,$s3);
1733 lw $tp1,16($key) # modulo-scheduled
1744 and $m,$tp1,$x80808080
1745 and $tp2,$tp1,$x7f7f7f7f
1768 xor $tp9,$tp8,$tp1
1774 rotr $tp1,$tpd,16
1777 xor $tpe,$tp1
1780 lw $tp1,4($key) # modulo-scheduled
1783 _ror $tp1,$tpd,16
1786 xor $tpe,$tp1
[all...]
H A Daes-586.pl1305 &xor ($tp2,$s[$i]); # tp2^tp1
1316 &xor ($tp4,$s[$i]); # tp4^tp1
1317 &rotl ($s[$i],8); # = ROTATE(tp1,8)
1325 &xor ($s[$i],$tp8); # ^= tp8^(tp4^tp1)^(tp2^tp1)
1327 &xor ($s[$i],$tp2); # ^= ROTATE(tp8^tp2^tp1,24)
1329 &xor ($s[$i],$tp4); # ^= ROTATE(tp8^tp4^tp1,16)
2881 { my ($i,$key,$tp1,$tp2,$tp4,$tp8) = @_;
2885 &and ($tmp,$tp1);
2886 &lea ($tp2,&DWP(0,$tp1,
[all...]
H A Daes-x86_64.pl1094 xor $tp10,$tp20 # tp2^=tp1
1096 xor $tp18,$tp28 # tp2^=tp1
1101 xor $tp10,$tp40 # tp4^=tp1
1102 xor $tp18,$tp48 # tp4^=tp1
1110 xor $tp80,$tp10 # tp1^=tp8
1111 xor $tp88,$tp18 # tp1^=tp8
1112 xor $tp80,$tp20 # tp2^tp1^=tp8
1113 xor $tp88,$tp28 # tp2^tp1^=tp8
1116 xor $tp80,$tp40 # tp4^tp1^=tp8
1118 xor $tp88,$tp48 # tp4^tp1
[all...]
/third_party/openssl/crypto/aes/asm/
H A Daes-mips.pl1670 my ($tp1,$tp2,$tp4,$tp8,$tp9,$tpb,$tpd,$tpe)=($a4,$a5,$a6,$a7,$s0,$s1,$s2,$s3);
1733 lw $tp1,16($key) # modulo-scheduled
1744 and $m,$tp1,$x80808080
1745 and $tp2,$tp1,$x7f7f7f7f
1768 xor $tp9,$tp8,$tp1
1774 rotr $tp1,$tpd,16
1777 xor $tpe,$tp1
1780 lw $tp1,4($key) # modulo-scheduled
1783 _ror $tp1,$tpd,16
1786 xor $tpe,$tp1
[all...]
H A Daes-586.pl1305 &xor ($tp2,$s[$i]); # tp2^tp1
1316 &xor ($tp4,$s[$i]); # tp4^tp1
1317 &rotl ($s[$i],8); # = ROTATE(tp1,8)
1325 &xor ($s[$i],$tp8); # ^= tp8^(tp4^tp1)^(tp2^tp1)
1327 &xor ($s[$i],$tp2); # ^= ROTATE(tp8^tp2^tp1,24)
1329 &xor ($s[$i],$tp4); # ^= ROTATE(tp8^tp4^tp1,16)
2881 { my ($i,$key,$tp1,$tp2,$tp4,$tp8) = @_;
2885 &and ($tmp,$tp1);
2886 &lea ($tp2,&DWP(0,$tp1,
[all...]
H A Daes-x86_64.pl1094 xor $tp10,$tp20 # tp2^=tp1
1096 xor $tp18,$tp28 # tp2^=tp1
1101 xor $tp10,$tp40 # tp4^=tp1
1102 xor $tp18,$tp48 # tp4^=tp1
1110 xor $tp80,$tp10 # tp1^=tp8
1111 xor $tp88,$tp18 # tp1^=tp8
1112 xor $tp80,$tp20 # tp2^tp1^=tp8
1113 xor $tp88,$tp28 # tp2^tp1^=tp8
1116 xor $tp80,$tp40 # tp4^tp1^=tp8
1118 xor $tp88,$tp48 # tp4^tp1
[all...]
/third_party/node/deps/openssl/config/archs/linux32-s390x/asm/crypto/aes/
H A Daes-s390x.S1346 .Lmix: l %r8,16(%r4) # tp1
1374 xr %r9,%r8 # tp2^tp1
1375 xr %r10,%r8 # tp4^tp1
1376 rll %r8,%r8,24 # = ROTATE(tp1,8)
1378 xr %r8,%r9 # ^=tp2^tp1
1379 xr %r9,%r11 # tp2^tp1^tp8
1380 xr %r8,%r10 # ^=tp4^tp1^tp8
1383 xr %r8,%r9 # ^= ROTATE(tp8^tp2^tp1,24)
1385 xr %r8,%r10 # ^= ROTATE(tp8^tp4^tp1,16)
/third_party/node/deps/openssl/config/archs/linux32-s390x/asm_avx2/crypto/aes/
H A Daes-s390x.S1346 .Lmix: l %r8,16(%r4) # tp1
1374 xr %r9,%r8 # tp2^tp1
1375 xr %r10,%r8 # tp4^tp1
1376 rll %r8,%r8,24 # = ROTATE(tp1,8)
1378 xr %r8,%r9 # ^=tp2^tp1
1379 xr %r9,%r11 # tp2^tp1^tp8
1380 xr %r8,%r10 # ^=tp4^tp1^tp8
1383 xr %r8,%r9 # ^= ROTATE(tp8^tp2^tp1,24)
1385 xr %r8,%r10 # ^= ROTATE(tp8^tp4^tp1,16)
/third_party/node/deps/openssl/config/archs/linux64-s390x/asm/crypto/aes/
H A Daes-s390x.S1346 .Lmix: l %r8,16(%r4) # tp1
1374 xr %r9,%r8 # tp2^tp1
1375 xr %r10,%r8 # tp4^tp1
1376 rll %r8,%r8,24 # = ROTATE(tp1,8)
1378 xr %r8,%r9 # ^=tp2^tp1
1379 xr %r9,%r11 # tp2^tp1^tp8
1380 xr %r8,%r10 # ^=tp4^tp1^tp8
1383 xr %r8,%r9 # ^= ROTATE(tp8^tp2^tp1,24)
1385 xr %r8,%r10 # ^= ROTATE(tp8^tp4^tp1,16)
/third_party/node/deps/openssl/config/archs/linux64-s390x/asm_avx2/crypto/aes/
H A Daes-s390x.S1346 .Lmix: l %r8,16(%r4) # tp1
1374 xr %r9,%r8 # tp2^tp1
1375 xr %r10,%r8 # tp4^tp1
1376 rll %r8,%r8,24 # = ROTATE(tp1,8)
1378 xr %r8,%r9 # ^=tp2^tp1
1379 xr %r9,%r11 # tp2^tp1^tp8
1380 xr %r8,%r10 # ^=tp4^tp1^tp8
1383 xr %r8,%r9 # ^= ROTATE(tp8^tp2^tp1,24)
1385 xr %r8,%r10 # ^= ROTATE(tp8^tp4^tp1,16)

Completed in 67 milliseconds

12