Lines Matching defs:load

247    /* Volatile prevents us from eliminating a load by combining it with
336 /* We only load g[2] once. */
339 nir_intrinsic_instr *load = get_intrinsic(nir_intrinsic_load_deref, i);
340 if (nir_intrinsic_get_var(load, 0) == g[2])
348 /* Invalidating a load in the end of loop body will apply to the whole loop
634 /* Only the second load will remain after the optimization. */
636 nir_intrinsic_instr *load = get_intrinsic(nir_intrinsic_load_deref, 0);
637 ASSERT_EQ(nir_intrinsic_get_var(load, 0), v[1]);
699 nir_intrinsic_instr *load;
701 load = get_intrinsic(nir_intrinsic_load_deref, 0);
702 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[0]);
703 load = get_intrinsic(nir_intrinsic_load_deref, 1);
704 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[1]);
706 load = get_intrinsic(nir_intrinsic_load_deref, 2);
707 ASSERT_EQ(nir_intrinsic_get_var(load, 0), y[0]);
708 load = get_intrinsic(nir_intrinsic_load_deref, 3);
709 ASSERT_EQ(nir_intrinsic_get_var(load, 0), y[1]);
711 load = get_intrinsic(nir_intrinsic_load_deref, 4);
712 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[0]);
713 load = get_intrinsic(nir_intrinsic_load_deref, 5);
714 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[1]);
776 nir_intrinsic_instr *load;
778 load = get_intrinsic(nir_intrinsic_load_deref, 0);
779 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[0]);
780 load = get_intrinsic(nir_intrinsic_load_deref, 1);
781 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[1]);
783 load = get_intrinsic(nir_intrinsic_load_deref, 2);
784 ASSERT_EQ(nir_intrinsic_get_var(load, 0), y[0]);
785 load = get_intrinsic(nir_intrinsic_load_deref, 3);
786 ASSERT_EQ(nir_intrinsic_get_var(load, 0), y[1]);
863 nir_intrinsic_instr *load;
865 load = get_intrinsic(nir_intrinsic_load_deref, 0);
866 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[0]);
867 load = get_intrinsic(nir_intrinsic_load_deref, 1);
868 ASSERT_EQ(nir_intrinsic_get_var(load, 0), x[1]);
1087 /* This load will be dropped, as vec.y (or vec[1]) is already known. */
1120 /* This load will be replaced with one from in0. */
1133 nir_intrinsic_instr *load = get_intrinsic(nir_intrinsic_load_deref, 0);
1134 ASSERT_EQ(nir_intrinsic_get_var(load, 0), in0);
1151 /* This load will be dropped, since both vec.x and vec.y are known. */
1229 /* Load from vec[idx] to store in out. This load should be dropped. */
1264 /* Load from vec[idx]. This load is currently not dropped. */
1269 /* Load from vec[idx] again. This load should be dropped. */
1306 /* Load from arr[idx] to store in out. This load should be dropped. */
1353 /* Load ssbo0.x and store it in out.x. This load should be dropped */
1399 /* Load ssbo0.x and store it in out.x. This load should not be dropped */
1443 /* Load ssbo_arr[0].x and store it in out.x. This load should not be dropped */
1486 /* Load ssbo_arr[0].x and store it in out.x. This load should be dropped */
1535 /* Load ssbo_arr[0].x and store it in out.x. This load should not be dropped */
1859 nir_intrinsic_instr *load = nir_src_as_intrinsic(vec->src[i].src);
1860 ASSERT_EQ(load->intrinsic, nir_intrinsic_load_deref);
1861 ASSERT_EQ(nir_intrinsic_get_var(load, 0), v[i])