/kernel/linux/linux-6.6/net/sunrpc/auth_gss/ |
H A D | gss_krb5_wrap.c | 74 int shifted = 0; in _rotate_left() local 78 while (shifted < shift) { in _rotate_left() 79 this_shift = min(shift - shifted, LOCAL_BUF_LEN); in _rotate_left() 81 shifted += this_shift; in _rotate_left()
|
/kernel/linux/linux-5.10/arch/h8300/lib/ |
H A D | udivsi3.S | 55 ;; er0 contains shifted dividend 57 ;; er2 contains shifted divisor
|
/kernel/linux/linux-5.10/net/sunrpc/auth_gss/ |
H A D | gss_krb5_wrap.c | 389 int shifted = 0; in _rotate_left() local 393 while (shifted < shift) { in _rotate_left() 394 this_shift = min(shift - shifted, LOCAL_BUF_LEN); in _rotate_left() 396 shifted += this_shift; in _rotate_left()
|
/kernel/linux/linux-6.6/arch/arm/mm/ |
H A D | cache-v7.S | 51 mov r3, r3, lsl r1 @ NumWays-1 shifted into bits [31:...] 52 movs r1, r2, lsl r1 @ #1 shifted left by same amount 157 movne r4, r4, lsl r5 @ # of ways shifted into bits [31:...] 158 movne r6, r6, lsl r5 @ 1 shifted left by same amount
|
/kernel/linux/linux-5.10/drivers/media/rc/ |
H A D | serial_ir.c | 181 unsigned char chunk, shifted; in send_pulse_irdeo() local 190 shifted = chunk << (i * 3); in send_pulse_irdeo() 191 shifted >>= 1; in send_pulse_irdeo() 192 output &= (~shifted); in send_pulse_irdeo()
|
/kernel/linux/linux-6.6/drivers/media/rc/ |
H A D | serial_ir.c | 181 unsigned char chunk, shifted; in send_pulse_irdeo() local 190 shifted = chunk << (i * 3); in send_pulse_irdeo() 191 shifted >>= 1; in send_pulse_irdeo() 192 output &= (~shifted); in send_pulse_irdeo()
|
/kernel/linux/linux-5.10/arch/m68k/fpsp040/ |
H A D | sgetem.S | 109 | shifted bits in d0 and d1 134 | ;be shifted into ms mant
|
H A D | round.S | 320 | ;it is just shifted 375 | ;be shifted into ms mant 445 | ; shifted off in denorm routine) 458 | ; shifted off in denorm routine)
|
H A D | binstr.S | 17 | shift and a mul by 8 shift. The bits shifted out of the
|
/kernel/linux/linux-6.6/arch/m68k/fpsp040/ |
H A D | sgetem.S | 109 | shifted bits in d0 and d1 134 | ;be shifted into ms mant
|
H A D | round.S | 320 | ;it is just shifted 375 | ;be shifted into ms mant 445 | ; shifted off in denorm routine) 458 | ; shifted off in denorm routine)
|
H A D | binstr.S | 17 | shift and a mul by 8 shift. The bits shifted out of the
|
/kernel/linux/linux-5.10/arch/sh/kernel/ |
H A D | entry-common.S | 262 * Note: When we're first called, the TRA value must be shifted
|
/kernel/linux/linux-6.6/arch/sh/kernel/ |
H A D | entry-common.S | 262 * Note: When we're first called, the TRA value must be shifted
|
/kernel/linux/linux-5.10/arch/powerpc/boot/ |
H A D | div64.S | 39 divwu r11,r11,r9 # then we divide the shifted quantities
|
/kernel/linux/linux-6.6/arch/powerpc/boot/ |
H A D | div64.S | 39 divwu r11,r11,r9 # then we divide the shifted quantities
|
/kernel/linux/linux-5.10/arch/parisc/kernel/ |
H A D | perf_asm.S | 86 ;* is shifted shifted backup immediately. This is to compensate 984 ;* is shifted shifted backup immediately. This is to compensate
|
/kernel/linux/linux-6.6/arch/parisc/kernel/ |
H A D | perf_asm.S | 86 ;* is shifted shifted backup immediately. This is to compensate 984 ;* is shifted shifted backup immediately. This is to compensate
|
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | aesni-intel_asm.S | 541 pxor \TMP3, \TMP2 # xor the shifted versions 558 pxor \TMP3,\TMP2 # xor the shifted versions 1158 pxor \TMP3, \TMP2 # xor the shifted versions 1173 pxor \TMP3,\TMP2 # xor the shifted versions 1370 pxor \TMP3, \TMP2 # xor the shifted versions 1385 pxor \TMP3,\TMP2 # xor the shifted versions 1469 pxor \TMP3, \TMP2 # xor the shifted versions 1484 pxor \TMP3, \TMP2 # xor the shifted versions
|
H A D | aesni-intel_avx-x86_64.S | 918 vpxor \T3, \T2, \T2 # xor the shifted versions 931 vpxor \T3, \T2, \T2 # xor the shifted versions 1553 vpxor \T3, \T2, \T2 # xor the shifted versions 1577 vpxor \T3, \T2, \T2 # xor the shifted versions 1741 vpxor \T3, \T2, \T2 # xor the shifted versions 1755 vpxor \T3, \T2, \T2 # xor the shifted versions
|
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | aesni-intel_asm.S | 537 pxor \TMP3, \TMP2 # xor the shifted versions 554 pxor \TMP3,\TMP2 # xor the shifted versions 1154 pxor \TMP3, \TMP2 # xor the shifted versions 1169 pxor \TMP3,\TMP2 # xor the shifted versions 1366 pxor \TMP3, \TMP2 # xor the shifted versions 1381 pxor \TMP3,\TMP2 # xor the shifted versions 1465 pxor \TMP3, \TMP2 # xor the shifted versions 1480 pxor \TMP3, \TMP2 # xor the shifted versions
|
H A D | aesni-intel_avx-x86_64.S | 886 vpxor \T3, \T2, \T2 # xor the shifted versions 899 vpxor \T3, \T2, \T2 # xor the shifted versions 1521 vpxor \T3, \T2, \T2 # xor the shifted versions 1545 vpxor \T3, \T2, \T2 # xor the shifted versions 1709 vpxor \T3, \T2, \T2 # xor the shifted versions 1723 vpxor \T3, \T2, \T2 # xor the shifted versions
|
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_input.c | 1431 unsigned int pcount, int shifted, int mss, in tcp_shifted_skb() 1436 u32 end_seq = start_seq + shifted; /* end of newly-SACKed */ in tcp_shifted_skb() 1454 TCP_SKB_CB(prev)->end_seq += shifted; in tcp_shifted_skb() 1455 TCP_SKB_CB(skb)->seq += shifted; in tcp_shifted_skb() 1428 tcp_shifted_skb(struct sock *sk, struct sk_buff *prev, struct sk_buff *skb, struct tcp_sacktag_state *state, unsigned int pcount, int shifted, int mss, bool dup_sack) tcp_shifted_skb() argument
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_input.c | 1450 unsigned int pcount, int shifted, int mss, in tcp_shifted_skb() 1455 u32 end_seq = start_seq + shifted; /* end of newly-SACKed */ in tcp_shifted_skb() 1473 TCP_SKB_CB(prev)->end_seq += shifted; in tcp_shifted_skb() 1474 TCP_SKB_CB(skb)->seq += shifted; in tcp_shifted_skb() 1447 tcp_shifted_skb(struct sock *sk, struct sk_buff *prev, struct sk_buff *skb, struct tcp_sacktag_state *state, unsigned int pcount, int shifted, int mss, bool dup_sack) tcp_shifted_skb() argument
|
/kernel/linux/linux-5.10/arch/m68k/ifpsp060/src/ |
H A D | ilsp.S | 322 # digit (word). After subtraction, the dividend is shifted and the
|