/third_party/node/deps/zlib/ |
H A D | crc_folding.c | 33 __m128i xmm_crc3 = _mm_loadu_si128((__m128i *)s->crc0 + 3);\ 40 _mm_storeu_si128((__m128i *)s->crc0 + 3, xmm_crc3);\ 51 xmm_crc3 = _mm_setzero_si128(); in crc_fold_init() 60 __m128i *xmm_crc2, __m128i *xmm_crc3) in fold_1() 69 x_tmp3 = *xmm_crc3; in fold_1() 71 *xmm_crc3 = *xmm_crc0; in fold_1() 73 *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10); in fold_1() 75 ps_crc3 = _mm_castsi128_ps(*xmm_crc3); in fold_1() 81 *xmm_crc3 in fold_1() 58 fold_1(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_1() argument 84 fold_2(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_2() argument 118 fold_3(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_3() argument 158 fold_4(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_4() argument 224 partial_fold(deflate_state *const s, const size_t len, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3, __m128i *xmm_crc_part) partial_fold() argument 441 __m128i xmm_crc3 = _mm_loadu_si128((__m128i *)s->crc0 + 3); crc_fold_512to32() local [all...] |
/third_party/node/deps/v8/third_party/zlib/ |
H A D | crc_folding.c | 33 __m128i xmm_crc3 = _mm_loadu_si128((__m128i *)s->crc0 + 3);\ 40 _mm_storeu_si128((__m128i *)s->crc0 + 3, xmm_crc3);\ 51 xmm_crc3 = _mm_setzero_si128(); in crc_fold_init() 60 __m128i *xmm_crc2, __m128i *xmm_crc3) in fold_1() 69 x_tmp3 = *xmm_crc3; in fold_1() 71 *xmm_crc3 = *xmm_crc0; in fold_1() 73 *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10); in fold_1() 75 ps_crc3 = _mm_castsi128_ps(*xmm_crc3); in fold_1() 81 *xmm_crc3 in fold_1() 58 fold_1(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_1() argument 84 fold_2(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_2() argument 118 fold_3(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_3() argument 158 fold_4(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_4() argument 224 partial_fold(deflate_state *const s, const size_t len, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3, __m128i *xmm_crc_part) partial_fold() argument [all...] |
/third_party/skia/third_party/externals/zlib/ |
H A D | crc_folding.c | 33 __m128i xmm_crc3 = _mm_loadu_si128((__m128i *)s->crc0 + 3);\ 40 _mm_storeu_si128((__m128i *)s->crc0 + 3, xmm_crc3);\ 51 xmm_crc3 = _mm_setzero_si128(); in crc_fold_init() 60 __m128i *xmm_crc2, __m128i *xmm_crc3) in fold_1() 69 x_tmp3 = *xmm_crc3; in fold_1() 71 *xmm_crc3 = *xmm_crc0; in fold_1() 73 *xmm_crc3 = _mm_clmulepi64_si128(*xmm_crc3, xmm_fold4, 0x10); in fold_1() 75 ps_crc3 = _mm_castsi128_ps(*xmm_crc3); in fold_1() 81 *xmm_crc3 in fold_1() 58 fold_1(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_1() argument 84 fold_2(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_2() argument 118 fold_3(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_3() argument 158 fold_4(deflate_state *const s, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3) fold_4() argument 224 partial_fold(deflate_state *const s, const size_t len, __m128i *xmm_crc0, __m128i *xmm_crc1, __m128i *xmm_crc2, __m128i *xmm_crc3, __m128i *xmm_crc_part) partial_fold() argument [all...] |