Lines Matching refs:z16

868   __ Clasta(z16.VnS(), p4, z16.VnS(), z0.VnS());
909 ASSERT_EQUAL_SVE(z16_expected, z16.VnD());
1172 InsrHelper(&masm, z16.VnS(), z16_inputs);
1176 __ Cmpgt(p10.VnS(), p3.Zeroing(), z16.VnS(), z17.VnS());
1179 __ Cmpne(p11.VnS(), p3.Zeroing(), z16.VnS(), z17.VnS());
1186 __ Cmplt(p15.VnS(), p3.Zeroing(), z17.VnS(), z16.VnS()); // GT
1389 __ dupm(z16.VnB(), 0xc3);
1813 __ Incp(z16.VnS(), p0, z2.VnS());
1875 ASSERT_EQUAL_SVE(z6_expected, z16.VnS());
1961 __ Sqincp(z16.VnS(), p0, z2.VnS());
2023 ASSERT_EQUAL_SVE(z6_expected, z16.VnS());
2109 __ Uqincp(z16.VnS(), p0, z2.VnS());
2179 ASSERT_EQUAL_SVE(z6_expected, z16.VnS());
2543 InsrHelper(&masm, z16.VnD(), z16_inputs);
2546 __ Cmple(p10.VnD(), p0.Zeroing(), z16.VnD(), -1);
2548 __ Cmple(p11.VnD(), p0.Zeroing(), z16.VnD(), LLONG_MIN);
2550 __ Cmpne(p12.VnD(), p0.Zeroing(), z16.VnD(), -1);
2552 __ Cmpne(p13.VnD(), p0.Zeroing(), z16.VnD(), LLONG_MAX);
4357 (masm.*fn)(z16.WithLaneSize(lane_size_in_bits), SVE_POW2, multiplier);
4403 ASSERT_EQUAL_SVE(z0, z16);
5633 __ Dup(z16.VnB(), z9.VnB(), index[6]);
5731 ASSERT_EQUAL_SVE_LANE(expected_z16, z16.VnB(), i);
5754 __ Uunpkhi(z16.VnH(), z9.VnB());
5826 uint16_t expected = core.zreg_lane<uint16_t>(z16.GetCode(), i);
6454 __ Cpy(z16.VnB(), pg, sp);
6494 ASSERT_EQUAL_SVE(expected_b_sp, z16.VnD());
6661 __ Fmov(z16.VnD(), pg.Merging(), RawbitsToDouble(0x7ff0000012340000)); // NaN
6749 ASSERT_EQUAL_SVE(z12.VnD(), z16.VnD());
6886 __ Ldr(z16, SVEMemOperand(x0, 6 * vl));
6919 ASSERT_EQUAL_SVE(z6, z16);
7095 __ Ld1sb(z16.VnD(), p4.Zeroing(), SVEMemOperand(x1, x2));
7136 // Ld1sb(z16.VnD(), ...)
7225 ASSERT_EQUAL_SVE(z26, z16);
7254 __ Index(z16.VnH(), -2, 5);
7257 __ St2h(z16.VnH(), z17.VnH(), p1, SVEMemOperand(x0, 8, SVE_MUL_VL));
7293 // Ld2h(z16.VnH(), z17.VnH(), ...)
7296 __ Mov(z6.VnH(), p1.Merging(), z16.VnH());
7350 // st2h { z16.h, z17.h }, SVE_MUL3
7602 // cleared). Registers z16-z27 will hold the values that were loaded.
7671 __ Mov(z16, z31);
7761 ASSERT_EQUAL_SVE(z4, z16);
7808 // cleared). Registers z16-z27 will hold the values that were loaded.
7881 __ Mov(z16, z31);
7974 ASSERT_EQUAL_SVE(z4, z16);
8105 __ Dup(z16.VnD(), 0);
8109 __ Mov(z16.VnD(), p3.Merging(), z21.VnD());
8240 ASSERT_EQUAL_SVE(z16, z0);
8363 __ Dup(z16.VnD(), 0);
8367 __ Mov(z16.VnD(), p4.Merging(), z0.VnD());
8501 ASSERT_EQUAL_SVE(z16, z0);
8558 __ Dup(z16.VnH(), 0);
8561 __ Mov(z16.VnH(), p1.Merging(), z4.VnH());
8602 ASSERT_EQUAL_SVE(z4, z16);
8654 __ Dup(z16.VnH(), 0);
8657 __ Mov(z16.VnH(), p1.Merging(), z4.VnH());
8943 ldff1)(z16.WithLaneSize(esize_in_bits), all.Zeroing(), SVEMemOperand(x20));
8951 (masm.*ld1)(z16.WithLaneSize(esize_in_bits),
8985 z16.WithLaneSize(esize_in_bits));
9509 __ Ldff1w(z16.VnD(), all.Zeroing(), SVEMemOperand(x0, z31.VnD(), UXTW, 2));
9604 ASSERT_EQUAL_SVE(expected_z16, z16.VnD());
9703 __ Ld1w(z16.VnD(), all.Zeroing(), SVEMemOperand(x0, z31.VnD(), UXTW, 2));
9784 ASSERT_EQUAL_SVE(expected_z16, z16.VnD());
14465 __ Uzp1(z16.VnD(), z8.VnD(), z9.VnD());
14496 ASSERT_EQUAL_SVE(z0.VnD(), z16.VnD());
14584 __ Mov(z16, z15);
14585 __ Ext(z16.VnB(), z16.VnB(), z16.VnB(), 8);
14586 __ Sel(z16.VnD(), p2, z16.VnD(), z30.VnD());
14594 __ Fcadd(z15.VnD(), p0.Merging(), z15.VnD(), z16.VnD(), 270);
14851 __ Dup(z16.VnS(), z25.VnS(), 0);
14852 FPSegmentPatternHelper(&masm, z16.VnH(), p0.Merging(), z16.VnH());
14853 __ Fmul(z16.VnS(), z1.VnS(), z16.VnS());
14879 ASSERT_EQUAL_SVE(z16.VnS(), z6.VnS());
15185 __ Mov(z16, z11);
15186 __ Fminnm(z16.VnS(), p0m, z16.VnS(), 0.0);
15250 ASSERT_EQUAL_SVE(expected_z16, z16.VnD());
17001 (masm.*macro)(z16.VnH(), p0.Merging(), z2.VnH(), z1.VnH(), z0.VnH(), option);
17039 ASSERT_EQUAL_SVE(z16.VnH(), z4.VnH());
18448 __ Adr(z16.VnS(), SVEMemOperand(z0.VnS(), z2.VnS(), LSL, 1));
18486 ASSERT_EQUAL_SVE(expected_z16, z16.VnD());
18745 __ Index(z16.VnD(), 0, 42);
18747 __ Saba(z17.VnD(), z17.VnD(), z16.VnD(), z17.VnD());
18762 ASSERT_EQUAL_SVE(z16, z17);
19228 __ Sqrdcmlah(z16.VnS(), z31.VnS(), z0.VnS(), z1.VnS(), 180);
19249 ASSERT_EQUAL_SVE(z16, z12);
19712 __ Usdot(z16.VnS(), z0.VnS(), z3.VnB(), z3.VnB(), 1);
19727 ASSERT_EQUAL_SVE(z17, z16);