Lines Matching refs:SZREG

58 	 * Byte copy if copying less than (2 * SZREG) bytes. This can
62 andi t0, a2, -(2 * SZREG)
68 andi t5, t3, -SZREG
69 andi t6, t4, -SZREG
72 * aligned SZREG address, does not equal dest, then add SZREG
73 * to find the low-bound of SZREG alignment in the dest memory
75 * region if n is less than SZREG. This is one reason why
76 * we always byte copy if n is less than SZREG.
77 * Otherwise, dest is already naturally aligned to SZREG.
80 addi t5, t5, SZREG
84 * If the dest and src are co-aligned to SZREG, then there is
89 andi t1, t0, (SZREG - 1)
99 andi a5, a1, (SZREG - 1) /* Find the alignment offset of src (a1) */
102 andi a1, a1, -SZREG /* Align the src pointer */
103 addi a2, t6, SZREG /* The other breakpoint for the unrolled loop*/
109 * Add that to XLEN. XLEN = SZREG * 8.
112 addi a7, a7, (SZREG * 8 + 1)
133 REG_L t0, (0 * SZREG)(a1)
135 REG_L t1, (1 * SZREG)(a1)
136 addi t3, t3, (2 * SZREG)
140 REG_S t2, ((0 * SZREG) - (2 * SZREG))(t3)
144 REG_L t0, (2 * SZREG)(a1)
145 addi a1, a1, (2 * SZREG)
149 REG_S t2, ((1 * SZREG) - (2 * SZREG))(t3)
161 andi a5, a4, (SZREG - 1) /* Find the alignment offset of src (a4) */
164 andi a4, a4, -SZREG /* Align the src pointer */
165 addi a2, t5, -SZREG /* The other breakpoint for the unrolled loop*/
171 * Add that to XLEN. XLEN = SZREG * 8.
174 addi a7, a7, (SZREG * 8 + 1)
195 REG_L t1, ( 0 * SZREG)(a4)
197 REG_L t0, (-1 * SZREG)(a4)
198 addi t4, t4, (-2 * SZREG)
202 REG_S t2, ( 1 * SZREG)(t4)
206 REG_L t1, (-2 * SZREG)(a4)
207 addi a4, a4, (-2 * SZREG)
211 REG_S t2, ( 0 * SZREG)(t4)
221 * Simple copy loops for SZREG co-aligned memory locations.
232 REG_L t1, ( 0 * SZREG)(a1)
233 addi a1, a1, SZREG
234 addi t3, t3, SZREG
235 REG_S t1, (-1 * SZREG)(t3)
244 REG_L t1, (-1 * SZREG)(a4)
245 addi a4, a4, -SZREG
246 addi t4, t4, -SZREG
247 REG_S t1, ( 0 * SZREG)(t4)