Lines Matching refs:x2
51 __m512i x0, x1, x2, x3, x4, x5, x6, x7, x8, y5, y6, y7, y8;
58 x2 = _mm512_loadu_si512((__m512i *)(buf + 0x40));
75 x6 = _mm512_clmulepi64_epi128(x2, x0, 0x00);
81 x2 = _mm512_clmulepi64_epi128(x2, x0, 0x11);
91 x2 = _mm512_xor_si512(x2, x6);
96 x2 = _mm512_xor_si512(x2, y6);
111 x1 = _mm512_xor_si512(x1, x2);
129 x2 = _mm512_loadu_si512((__m512i *)buf);
133 x1 = _mm512_xor_si512(x1, x2);
228 __m128i x0, x1, x2, x3, x4, x5, x6, x7, x8, y5, y6, y7, y8;
234 x2 = _mm_loadu_si128((__m128i *)(buf + 0x10));
251 x6 = _mm_clmulepi64_si128(x2, x0, 0x00);
256 x2 = _mm_clmulepi64_si128(x2, x0, 0x11);
266 x2 = _mm_xor_si128(x2, x6);
271 x2 = _mm_xor_si128(x2, y6);
286 x1 = _mm_xor_si128(x1, x2);
304 x2 = _mm_loadu_si128((__m128i *)buf);
308 x1 = _mm_xor_si128(x1, x2);
318 x2 = _mm_clmulepi64_si128(x1, x0, 0x10);
321 x1 = _mm_xor_si128(x1, x2);
325 x2 = _mm_srli_si128(x1, 4);
328 x1 = _mm_xor_si128(x1, x2);
335 x2 = _mm_and_si128(x1, x3);
336 x2 = _mm_clmulepi64_si128(x2, x0, 0x10);
337 x2 = _mm_and_si128(x2, x3);
338 x2 = _mm_clmulepi64_si128(x2, x0, 0x00);
339 x1 = _mm_xor_si128(x1, x2);
498 uint64x2_t x0, x1, x2, x3, x4, x5, x6, x7, x8, y5, y6, y7, y8;
504 x2 = vld1q_u64((const uint64_t *)(buf + 0x10));
521 x6 = (uint64x2_t) pmull_lo(x2, x0);
531 x2 = (uint64x2_t) pmull_hi(x2, x0);
536 x2 = veorq_u64(x2, x6);
541 x2 = veorq_u64(x2, y6);
556 x1 = veorq_u64(x1, x2);
574 x2 = vld1q_u64((const uint64_t *)buf);
578 x1 = veorq_u64(x1, x2);
590 x2 = (uint64x2_t) pmull_01(x1, x0);
593 x1 = veorq_u64(x1, x2);
597 x2 = (uint64x2_t) pmull_01(x2, x0);
598 x2 = (uint64x2_t) vextq_u8(vreinterpretq_u8_u64(x1), vdupq_n_u8(0), 4);
601 x1 = veorq_u64(x1, x2);
608 x2 = vandq_u64(x1, x3);
609 x2 = (uint64x2_t) pmull_01(x2, x0);
610 x2 = vandq_u64(x2, x3);
611 x2 = (uint64x2_t) pmull_lo(x2, x0);
612 x1 = veorq_u64(x1, x2);