Lines Matching refs:vec_ld
77 vzo2 = vec_ld(i2, &(out[0])); // zo2.r zo2.i z(o2+1).r z(o2+1).i
78 vzo2plus1 = vec_ld(i2+16, &(out[0]));
79 vzo3 = vec_ld(i3, &(out[0])); // zo3.r zo3.i z(o3+1).r z(o3+1).i
80 vzo3plus1 = vec_ld(i3+16, &(out[0]));
81 vz0 = vec_ld(0, &(out[0])); // z0.r z0.i z1.r z1.i
82 vz0plus1 = vec_ld(16, &(out[0]));
83 vzo1 = vec_ld(i1, &(out[0])); // zo1.r zo1.i z(o1+1).r z(o1+1).i
84 vzo1plus1 = vec_ld(i1+16, &(out[0]));
166 vzo2 = vec_ld(i2, &(out[0])); // zo2.r zo2.i z(o2+1).r z(o2+1).i
167 vzo2plus1 = vec_ld(i2+16, &(out[0]));
168 vzo3 = vec_ld(i3, &(out[0])); // zo3.r zo3.i z(o3+1).r z(o3+1).i
169 vzo3plus1 = vec_ld(i3+16, &(out[0]));
170 vz0 = vec_ld(0, &(out[0])); // z0.r z0.i z1.r z1.i
171 vz0plus1 = vec_ld(16, &(out[0]));
172 vzo1 = vec_ld(i1, &(out[0])); // zo1.r zo1.i z(o1+1).r z(o1+1).i
173 vzo1plus1 = vec_ld(i1+16, &(out[0]));
263 a = vec_ld(0, &(out[0]));
264 b = vec_ld(byte_2complex, &(out[0]));
296 vz0 = vec_ld(0, &(out[0]));
297 vz1 = vec_ld(byte_2complex, &(out[0]));
298 vz2 = vec_ld(byte_4complex, &(out[0]));
299 vz3 = vec_ld(byte_6complex, &(out[0]));
380 vz0 = vec_ld(0, &(out[0]));
381 vz1 = vec_ld(byte_2complex, &(out[0]));
382 vz2 = vec_ld(byte_4complex, &(out[0]));
383 vz3 = vec_ld(byte_6complex, &(out[0]));
384 vz4 = vec_ld(byte_8complex, &(out[0]));
385 vz5 = vec_ld(byte_10complex, &(out[0]));
386 vz6 = vec_ld(byte_12complex, &(out[0]));
387 vz7 = vec_ld(byte_14complex, &(out[0]));
505 a = vec_ld(0, &(out[0]));
506 b = vec_ld(byte_2complex, &(out[0]));
537 vz0 = vec_ld(0, &(out[0]));
538 vz1 = vec_ld(byte_2complex, &(out[0]));
539 vz2 = vec_ld(byte_4complex, &(out[0]));
540 vz3 = vec_ld(byte_6complex, &(out[0]));
601 vz0 = vec_ld(byte_8complex, &(out[0]));
602 vz1 = vec_ld(byte_10complex, &(out[0]));
603 vz2 = vec_ld(byte_12complex, &(out[0]));
604 vz3 = vec_ld(byte_14complex, &(out[0]));
632 vz0 = vec_ld(0, &(out[0]));
633 vz1 = vec_ld(byte_2complex, &(out[0]));
634 vz2 = vec_ld(byte_4complex, &(out[0]));
635 vz3 = vec_ld(byte_6complex, &(out[0]));
726 v8 = vec_ld(0, &(wre[0]));
727 v10 = vec_ld(0, &(wim[0]));
728 v9 = vec_ld(0, &(wim[-4]));
731 v4 = vec_ld(i2, &(out[0]));
732 v5 = vec_ld(i2+16, &(out[0]));
733 v6 = vec_ld(i3, &(out[0]));
734 v7 = vec_ld(i3+16, &(out[0]));
740 v0 = vec_ld(0, &(out[0])); // r0
741 v3 = vec_ld(i1+16, &(out[0])); // i1
747 v1 = vec_ld(16, &(out[0])); // i0
748 v2 = vec_ld(i1, &(out[0])); // r1
778 v8 = vec_ld(0, &(wre[0]));
779 v10 = vec_ld(0, &(wim[0]));
780 v9 = vec_ld(0, &(wim[-4]));
783 v4 = vec_ld(i2, &(out[0])); // r2
784 v5 = vec_ld(i2+16, &(out[0])); // i2
785 v6 = vec_ld(i3, &(out[0])); // r3
786 v7 = vec_ld(i3+16, &(out[0]));// i3
792 v0 = vec_ld(0, &(out[0])); // r0
793 v3 = vec_ld(i1+16, &(out[0])); // i1
799 v1 = vec_ld(16, &(out[0])); // i0
800 v2 = vec_ld(i1, &(out[0])); // r1