Lines Matching defs:len
311 * csum_partial_copy_nocheck(src, dst, len)
312 * __csum_partial_copy_kernel(src, dst, len)
320 #define len a2
443 * Note: dst & src may be unaligned, len may be 0
451 * If len < NBYTES use byte operations.
453 sltu t2, len, NBYTES
466 SRL t0, len, LOG_NBYTES+3 # +3 for 8 units/iter
467 beqz t0, .Lcleanup_both_aligned\@ # len < 8*NBYTES
469 SUB len, 8*NBYTES # subtract here for bgez loop
480 SUB len, len, 8*NBYTES
500 bgez len, 1b
502 ADD len, 8*NBYTES # revert len (see above)
505 * len == the number of bytes left to copy < 8*NBYTES
509 beqz len, .Ldone\@
510 sltu t0, len, 4*NBYTES
512 and rem, len, (NBYTES-1) # rem = len % NBYTES
514 * len >= 4*NBYTES
520 SUB len, len, 4*NBYTES
532 beqz len, .Ldone\@
536 * rem = len % NBYTES
538 beq rem, len, .Lcopy_bytes\@
543 SUB len, len, NBYTES
548 bne rem, len, 1b
563 beqz len, .Ldone\@
564 ADD t1, dst, len # t1 is just past last byte of dst
566 SLL rem, len, 3 # rem = number of bits to keep
581 * len >= NBYTES
597 beq len, t2, .Ldone\@
598 SUB len, len, t2
604 SRL t0, len, LOG_NBYTES+2 # +2 for 4 units/iter
606 and rem, len, (4*NBYTES-1) # rem = len % 4*NBYTES
616 SUB len, len, 4*NBYTES
637 bne len, rem, 1b
641 beqz len, .Ldone\@
642 and rem, len, NBYTES-1 # rem = len % NBYTES
643 beq rem, len, .Lcopy_bytes\@
649 SUB len, len, NBYTES
654 bne len, rem, 1b
658 beqz len, .Ldone\@
661 /* 0 < len < NBYTES */
673 SUB len, len, 1; \
677 beqz len, .Lcopy_bytes_done\@; \
689 SUB len, len, 1