/kernel/linux/linux-5.10/arch/arm/crypto/ |
H A D | sha1-armv4-large.S | 77 mov r5,r5,ror#30 78 mov r6,r6,ror#30 79 mov r7,r7,ror#30 @ [6] 85 add r7,r8,r7,ror#2 @ E+=K_00_19 90 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 94 add r7,r8,r7,ror#2 @ E+=K_00_19 96 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 101 and r10,r4,r10,ror#2 103 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 110 add r6,r8,r6,ror# [all...] |
H A D | sha256-armv4.pl | 80 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 82 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 97 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 99 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 107 add $h,$h,$t0,ror#$Sigma1[0] @ h+=Sigma1(e) 111 eor $t0,$a,$a,ror#`$Sigma0[1]-$Sigma0[0]` 129 eor $t0,$t0,$a,ror#`$Sigma0[2]-$Sigma0[0]` @ Sigma0(a) 133 add $h,$h,$t0,ror#$Sigma0[0] @ h+=Sigma0(a) 145 mov $t0,$t1,ror#$sigma0[0] 147 mov $t2,$t4,ror# [all...] |
/kernel/linux/linux-6.6/arch/arm/crypto/ |
H A D | sha1-armv4-large.S | 77 mov r5,r5,ror#30 78 mov r6,r6,ror#30 79 mov r7,r7,ror#30 @ [6] 85 add r7,r8,r7,ror#2 @ E+=K_00_19 90 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 94 add r7,r8,r7,ror#2 @ E+=K_00_19 96 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 101 and r10,r4,r10,ror#2 103 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 110 add r6,r8,r6,ror# [all...] |
H A D | sha256-armv4.pl | 80 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 82 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 97 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 99 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 107 add $h,$h,$t0,ror#$Sigma1[0] @ h+=Sigma1(e) 111 eor $t0,$a,$a,ror#`$Sigma0[1]-$Sigma0[0]` 129 eor $t0,$t0,$a,ror#`$Sigma0[2]-$Sigma0[0]` @ Sigma0(a) 133 add $h,$h,$t0,ror#$Sigma0[0] @ h+=Sigma0(a) 145 mov $t0,$t1,ror#$sigma0[0] 147 mov $t2,$t4,ror# [all...] |
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | sha512-ssse3-asm.S | 124 ror $23, tmp0 # 41 # tmp = e ror 23 126 xor e_64, tmp0 # tmp = (e ror 23) ^ e 130 ror $4, tmp0 # 18 # tmp = ((e ror 23) ^ e) ror 4 131 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 134 ror $14, tmp0 # 14 # tmp = ((((e ror23)^e)ror4)^e)ror14 = S1(e) 142 ror [all...] |
H A D | sha256-ssse3-asm.S | 152 ror $(25-11), y0 # y0 = e >> (25-11) 155 ror $(22-13), y1 # y1 = a >> (22-13) 158 ror $(11-6), y0 # y0 = (e >> (11-6)) ^ (e >> (25-6)) 165 ror $(13-2), y1 # y1 = (a >> (13-2)) ^ (a >> (22-2)) 169 ror $6, y0 # y0 = S1 = (e>>6) & (e>>11) ^ (e>>25) 172 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 186 por XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 195 ror $(25-11), y0 # y0 = e >> (25-11) 198 ror $(22-13), y1 # y1 = a >> (22-13) 201 ror [all...] |
H A D | twofish-i586-asm_32.S | 74 ror $16, b ## D;\ 77 ror $16, a ## D;\ 84 ror $15, b ## D;\ 112 ror $16, b ## D;\ 115 ror $16, a ## D;\ 122 ror $16, b ## D;\ 131 ror $1, c ## D;\ 149 ror $16, a ## D;\ 152 ror $16, b ## D;\ 159 ror [all...] |
H A D | twofish-x86_64-asm_64.S | 72 ror $16, b ## D;\ 75 ror $16, a ## D;\ 82 ror $15, b ## D;\ 110 ror $16, b ## D;\ 113 ror $16, a ## D;\ 128 ror $1, c ## D;\ 145 ror $16, a ## D;\ 148 ror $16, b ## D;\ 155 ror $15, a ## D;\ 181 ror [all...] |
H A D | sha512-avx-asm.S | 121 # shld is faster than ror on Sandybridge 130 RORQ tmp0, 23 # 41 # tmp = e ror 23 132 xor e_64, tmp0 # tmp = (e ror 23) ^ e 136 RORQ tmp0, 4 # 18 # tmp = ((e ror 23) ^ e) ror 4 137 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 148 RORQ tmp0, 5 # 39 # tmp = a ror 5 149 xor a_64, tmp0 # tmp = (a ror 5) ^ a 151 RORQ tmp0, 6 # 34 # tmp = ((a ror [all...] |
H A D | sha512-avx2-asm.S | 178 # Calculate w[t-15] ror 1 181 vpor YTMP2, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 224 # Calculate w[t-15] ror 8 227 vpor YTMP2, YTMP1, YTMP1 # YTMP1 = W[-15] ror 8 229 vpxor YTMP4, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 ^ W[-15] >> 7 291 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 19 {BABA} 292 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {BABA} 295 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 61 {BABA} 296 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = s1 = (W[-2] ror 19) ^ 297 # (W[-2] ror 6 [all...] |
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | sha512-ssse3-asm.S | 121 ror $23, tmp0 # 41 # tmp = e ror 23 123 xor e_64, tmp0 # tmp = (e ror 23) ^ e 127 ror $4, tmp0 # 18 # tmp = ((e ror 23) ^ e) ror 4 128 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 131 ror $14, tmp0 # 14 # tmp = ((((e ror23)^e)ror4)^e)ror14 = S1(e) 139 ror [all...] |
H A D | sha256-ssse3-asm.S | 153 ror $(25-11), y0 # y0 = e >> (25-11) 156 ror $(22-13), y1 # y1 = a >> (22-13) 159 ror $(11-6), y0 # y0 = (e >> (11-6)) ^ (e >> (25-6)) 166 ror $(13-2), y1 # y1 = (a >> (13-2)) ^ (a >> (22-2)) 170 ror $6, y0 # y0 = S1 = (e>>6) & (e>>11) ^ (e>>25) 173 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 187 por XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 196 ror $(25-11), y0 # y0 = e >> (25-11) 199 ror $(22-13), y1 # y1 = a >> (22-13) 202 ror [all...] |
H A D | twofish-i586-asm_32.S | 74 ror $16, b ## D;\ 77 ror $16, a ## D;\ 84 ror $15, b ## D;\ 112 ror $16, b ## D;\ 115 ror $16, a ## D;\ 122 ror $16, b ## D;\ 131 ror $1, c ## D;\ 149 ror $16, a ## D;\ 152 ror $16, b ## D;\ 159 ror [all...] |
H A D | twofish-x86_64-asm_64.S | 72 ror $16, b ## D;\ 75 ror $16, a ## D;\ 82 ror $15, b ## D;\ 110 ror $16, b ## D;\ 113 ror $16, a ## D;\ 128 ror $1, c ## D;\ 145 ror $16, a ## D;\ 148 ror $16, b ## D;\ 155 ror $15, a ## D;\ 181 ror [all...] |
H A D | sha512-avx-asm.S | 118 # shld is faster than ror on Sandybridge 127 RORQ tmp0, 23 # 41 # tmp = e ror 23 129 xor e_64, tmp0 # tmp = (e ror 23) ^ e 133 RORQ tmp0, 4 # 18 # tmp = ((e ror 23) ^ e) ror 4 134 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 145 RORQ tmp0, 5 # 39 # tmp = a ror 5 146 xor a_64, tmp0 # tmp = (a ror 5) ^ a 148 RORQ tmp0, 6 # 34 # tmp = ((a ror [all...] |
H A D | sha512-avx2-asm.S | 175 # Calculate w[t-15] ror 1 178 vpor YTMP2, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 221 # Calculate w[t-15] ror 8 224 vpor YTMP2, YTMP1, YTMP1 # YTMP1 = W[-15] ror 8 226 vpxor YTMP4, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 ^ W[-15] >> 7 288 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 19 {BABA} 289 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {BABA} 292 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 61 {BABA} 293 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = s1 = (W[-2] ror 19) ^ 294 # (W[-2] ror 6 [all...] |
/kernel/linux/linux-5.10/arch/arm64/crypto/ |
H A D | sha512-armv8.pl | 133 # 'eor x,y,z,ror#n', it was found to negatively affect performance 140 ror $t0,$e,#$Sigma1[0] 142 eor $T0,$e,$e,ror#`$Sigma1[2]-$Sigma1[1]` 148 eor $t0,$t0,$T0,ror#$Sigma1[1] // Sigma1(e) 149 ror $T0,$a,#$Sigma0[0] 151 eor $t1,$a,$a,ror#`$Sigma0[2]-$Sigma0[1]` 156 eor $t1,$T0,$t1,ror#$Sigma0[1] // Sigma0(a) 162 ror $t0,$e,#$Sigma1[0] 164 ror $T1,@X[($j+1)&15],#$sigma0[0] 166 ror [all...] |
H A D | chacha-neon-core.S | 265 ror a12, a12, #16 267 ror a13, a13, #16 269 ror a14, a14, #16 271 ror a15, a15, #16 301 ror a4, a4, #20 303 ror a5, a5, #20 305 ror a6, a6, #20 307 ror a7, a7, #20 332 ror a12, a12, #24 334 ror a1 [all...] |
/kernel/linux/linux-6.6/arch/arm64/crypto/ |
H A D | sha512-armv8.pl | 133 # 'eor x,y,z,ror#n', it was found to negatively affect performance 140 ror $t0,$e,#$Sigma1[0] 142 eor $T0,$e,$e,ror#`$Sigma1[2]-$Sigma1[1]` 148 eor $t0,$t0,$T0,ror#$Sigma1[1] // Sigma1(e) 149 ror $T0,$a,#$Sigma0[0] 151 eor $t1,$a,$a,ror#`$Sigma0[2]-$Sigma0[1]` 156 eor $t1,$T0,$t1,ror#$Sigma0[1] // Sigma0(a) 162 ror $t0,$e,#$Sigma1[0] 164 ror $T1,@X[($j+1)&15],#$sigma0[0] 166 ror [all...] |
H A D | chacha-neon-core.S | 264 ror a12, a12, #16 266 ror a13, a13, #16 268 ror a14, a14, #16 270 ror a15, a15, #16 300 ror a4, a4, #20 302 ror a5, a5, #20 304 ror a6, a6, #20 306 ror a7, a7, #20 331 ror a12, a12, #24 333 ror a1 [all...] |
/kernel/linux/linux-5.10/arch/arm/lib/ |
H A D | csumpartial.S | 39 movne sum, sum, ror #8 74 movne r0, r0, ror #8 @ rotate checksum by 8 bits 105 movne sum, sum, ror #8
|
H A D | io-readsw-armv4.S | 99 _BE_ONLY_( mov ip, ip, ror #8 ) 122 _BE_ONLY_( movne ip, ip, ror #8 )
|
/kernel/linux/linux-6.6/arch/arm/lib/ |
H A D | csumpartial.S | 39 movne sum, sum, ror #8 74 movne r0, r0, ror #8 @ rotate checksum by 8 bits 105 movne sum, sum, ror #8
|
/kernel/linux/linux-5.10/arch/arm64/include/asm/ |
H A D | kvm_mmu.h | 68 ror \reg, \reg, #1 /* rotate to the first tag bit */ 71 ror \reg, \reg, #63 /* rotate back */ 90 "ror %0, %0, #1\n" 93 "ror %0, %0, #63\n",
|
/kernel/linux/linux-5.10/arch/arm64/kvm/hyp/nvhe/ |
H A D | hyp-init.S | 66 ror x0, x0, #24 68 ror x0, x0, #4
|