Lines Matching defs:dst
139 dst = arg0
162 add dst,len,end
168 xor src,dst,t0
179 extru dst,31,3,t1
182 21: stb,ma t1,1(dstspc,dst)
197 12: std,ma t1,8(dstspc,dst)
198 13: std,ma t2,8(dstspc,dst)
202 16: std,ma t1,8(dstspc,dst)
203 17: std,ma t2,8(dstspc,dst)
220 21: stw,ma t1,4(dstspc,dst)
231 extru dst,31,2,t1
234 21: stb,ma t1,1(dstspc,dst)
247 12: stw,ma t1,4(dstspc,dst)
248 13: stw,ma t2,4(dstspc,dst)
252 16: stw,ma t1,4(dstspc,dst)
253 17: stw,ma t2,4(dstspc,dst)
272 21: stb,ma t1,1(dstspc,dst)
281 sub end,dst,ret0
284 /* src and dst are not aligned the same way. */
287 /* align until dst is 32bit-word-aligned */
288 extru dst,31,2,t1
292 21: stb,ma t1,1(dstspc,dst)
301 /* store src, dst and len in safe place */
303 copy dst,save_dst
310 * Copy from a not-aligned src to an aligned dst using shifts.
346 1: stw,ma t0, 4(dstspc,dst)
352 1: stw,ma t0, 4(dstspc,dst)
358 1: stw,ma t0, 4(dstspc,dst)
364 1: stw,ma t0, 4(dstspc,dst)
371 1: stw,ma t0, 4(dstspc,dst)
376 /* calculate new src, dst and len and jump to byte-copy loop */
377 sub dst,save_dst,t0
402 10: std,ma t1,8(dstspc,dst)
408 10: stw,ma t1,4(dstspc,dst)