/kernel/linux/linux-6.6/drivers/char/tpm/ |
H A D | xen-tpmfront.c | 24 struct vtpm_shared_page *shr; member 103 switch (priv->shr->state) { in vtpm_status() 124 priv->shr->state = VTPM_STATE_CANCEL; in vtpm_cancel() 129 static size_t shr_data_offset(struct vtpm_shared_page *shr) in shr_data_offset() argument 131 return struct_size(shr, extra_pages, shr->nr_extra_pages); in shr_data_offset() 137 struct vtpm_shared_page *shr = priv->shr; in vtpm_send() local 138 size_t offset = shr_data_offset(shr); in vtpm_send() 156 memcpy(offset + (u8 *)shr, bu in vtpm_send() 179 struct vtpm_shared_page *shr = priv->shr; vtpm_recv() local [all...] |
/kernel/linux/linux-5.10/drivers/char/tpm/ |
H A D | xen-tpmfront.c | 24 struct vtpm_shared_page *shr; member 103 switch (priv->shr->state) { in vtpm_status() 124 priv->shr->state = VTPM_STATE_CANCEL; in vtpm_cancel() 129 static unsigned int shr_data_offset(struct vtpm_shared_page *shr) in shr_data_offset() argument 131 return sizeof(*shr) + sizeof(u32) * shr->nr_extra_pages; in shr_data_offset() 137 struct vtpm_shared_page *shr = priv->shr; in vtpm_send() local 138 unsigned int offset = shr_data_offset(shr); in vtpm_send() 156 memcpy(offset + (u8 *)shr, bu in vtpm_send() 179 struct vtpm_shared_page *shr = priv->shr; vtpm_recv() local [all...] |
/kernel/linux/linux-5.10/arch/ia64/lib/ |
H A D | ip_fast_csum.S | 59 shr.u ret0=r20,16 // now need to add the carry 64 shr.u ret0=r20,16 // add carry again 69 shr.u ret0=r20,16 117 shr.u r15=r15,16 129 shr.u r10=r8,32 // now fold sum into short 134 shr.u r10=r8,16 // yeah, keep it rolling 139 shr.u r10=r8,16 // three times lucky
|
H A D | do_csum.S | 170 shr.u tmask=tmask,tmp1 // build tail mask, mask off ]8,lastoff] 177 shr.u count=count,3 // how many 8-byte? 212 shr.u count=count,1 // we do 16 bytes per loop 264 shr.u tmp2=result1[0],32 269 shr.u tmp2=result1[0],16 274 shr.u tmp2=result1[0],16 279 shr.u tmp2=result1[0],16 289 (p15) shr.u ret0=ret0,64-16 // + shift back to position = swap bytes 296 // shr.u word1[0]=result1[0],32 303 // shr [all...] |
H A D | memcpy.S | 69 shr.u cnt=in2,3 // number of 8-byte words to copy 198 shr.u t0=t0,t2 209 (p3) shr.u t0=t0,8 213 (p4) shr.u t0=t0,16 232 shr.u t2=cnt,3 // t2 = number of 8-byte words left to copy 263 (p5) shr.u t0=t0,32 267 (p4) shr.u t0=t0,16
|
/kernel/linux/linux-6.6/arch/ia64/lib/ |
H A D | ip_fast_csum.S | 59 shr.u ret0=r20,16 // now need to add the carry 64 shr.u ret0=r20,16 // add carry again 69 shr.u ret0=r20,16 117 shr.u r15=r15,16 129 shr.u r10=r8,32 // now fold sum into short 134 shr.u r10=r8,16 // yeah, keep it rolling 139 shr.u r10=r8,16 // three times lucky
|
H A D | do_csum.S | 170 shr.u tmask=tmask,tmp1 // build tail mask, mask off ]8,lastoff] 177 shr.u count=count,3 // how many 8-byte? 212 shr.u count=count,1 // we do 16 bytes per loop 264 shr.u tmp2=result1[0],32 269 shr.u tmp2=result1[0],16 274 shr.u tmp2=result1[0],16 279 shr.u tmp2=result1[0],16 289 (p15) shr.u ret0=ret0,64-16 // + shift back to position = swap bytes 296 // shr.u word1[0]=result1[0],32 303 // shr [all...] |
H A D | memcpy.S | 69 shr.u cnt=in2,3 // number of 8-byte words to copy 198 shr.u t0=t0,t2 209 (p3) shr.u t0=t0,8 213 (p4) shr.u t0=t0,16 232 shr.u t2=cnt,3 // t2 = number of 8-byte words left to copy 263 (p5) shr.u t0=t0,32 267 (p4) shr.u t0=t0,16
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | vmscan.h | 186 TP_PROTO(struct shrinker *shr, struct shrink_control *sc, 191 TP_ARGS(shr, sc, nr_objects_to_shrink, cache_items, delta, total_scan, 195 __field(struct shrinker *, shr) 207 __entry->shr = shr; 208 __entry->shrink = shr->scan_objects; 220 __entry->shr, 231 TP_PROTO(struct shrinker *shr, int nid, int shrinker_retval, 234 TP_ARGS(shr, nid, shrinker_retval, unused_scan_cnt, new_scan_cnt, 238 __field(struct shrinker *, shr) [all...] |
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | vmscan.h | 200 TP_PROTO(struct shrinker *shr, struct shrink_control *sc, 205 TP_ARGS(shr, sc, nr_objects_to_shrink, cache_items, delta, total_scan, 209 __field(struct shrinker *, shr) 221 __entry->shr = shr; 222 __entry->shrink = shr->scan_objects; 234 __entry->shr, 245 TP_PROTO(struct shrinker *shr, int nid, int shrinker_retval, 248 TP_ARGS(shr, nid, shrinker_retval, unused_scan_cnt, new_scan_cnt, 252 __field(struct shrinker *, shr) [all...] |
/kernel/linux/linux-5.10/arch/x86/math-emu/ |
H A D | wm_shrx.S | 52 shr %cl,%edx 68 shr %cl,%edx 81 shr %cl,%eax 137 shr %cl,%edx 158 shr %cl,%edx 179 shr %cl,%eax
|
H A D | shr_Xsig.S | 41 shr %cl,%edx 58 shr %cl,%edx 72 shr %cl,%eax
|
H A D | reg_u_add.S | 85 shr %cl,%eax 96 shr %cl,%eax
|
/kernel/linux/linux-6.6/arch/x86/math-emu/ |
H A D | wm_shrx.S | 52 shr %cl,%edx 68 shr %cl,%edx 81 shr %cl,%eax 137 shr %cl,%edx 158 shr %cl,%edx 179 shr %cl,%eax
|
H A D | shr_Xsig.S | 41 shr %cl,%edx 58 shr %cl,%edx 72 shr %cl,%eax
|
H A D | reg_u_add.S | 85 shr %cl,%eax 96 shr %cl,%eax
|
/kernel/linux/linux-5.10/arch/mips/crypto/ |
H A D | poly1305-mips.pl | 228 my ($shr,$shl) = ($s6,$s7); # used on R6 271 andi $shr,$inp,7 272 dsubu $inp,$inp,$shr # align $inp 273 sll $shr,$shr,3 # byte to bit offset 274 subu $shl,$zero,$shr 294 beqz $shr,.Laligned_inp 298 dsllv $in0,$in0,$shr 300 dsllv $in1,$in1,$shr 303 dsrlv $in0,$in0,$shr [all...] |
/kernel/linux/linux-6.6/arch/mips/crypto/ |
H A D | poly1305-mips.pl | 228 my ($shr,$shl) = ($s6,$s7); # used on R6 271 andi $shr,$inp,7 272 dsubu $inp,$inp,$shr # align $inp 273 sll $shr,$shr,3 # byte to bit offset 274 subu $shl,$zero,$shr 294 beqz $shr,.Laligned_inp 298 dsllv $in0,$in0,$shr 300 dsllv $in1,$in1,$shr 303 dsrlv $in0,$in0,$shr [all...] |
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | poly1305-x86_64-cryptogams.pl | 219 shr \$2,$d3 276 shr \$32,%r9 308 shr \$4,$len 335 shr \$2,$s1 393 shr \$2,%r10 # did 130-bit value overflow? 456 shr \$26,$d1 458 shr \$26,$d2 469 shr \$26,$d1 471 shr \$26,$d2 492 shr \ [all...] |
H A D | twofish-x86_64-asm_64.S | 221 shr $32, R1 223 shr $32, R3 273 shr $32, R1 276 shr $32, R3
|
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | poly1305-x86_64-cryptogams.pl | 218 shr \$2,$d3 275 shr \$32,%r9 307 shr \$4,$len 334 shr \$2,$s1 392 shr \$2,%r10 # did 130-bit value overflow? 455 shr \$26,$d1 457 shr \$26,$d2 468 shr \$26,$d1 470 shr \$26,$d2 491 shr \ [all...] |
/kernel/linux/linux-5.10/arch/ia64/kernel/ |
H A D | ivt.S | 124 shr.u r17=r16,61 // get the region number into r17 126 shr.u r22=r21,3 134 (p8) shr r22=r22,r27 138 shr.u r18=r22,PGDIR_SHIFT // get bottom portion of pgd index bit 146 (p6) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT 147 (p7) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT-3 153 shr.u r28=r22,PUD_SHIFT // shift pud index into position 155 shr.u r18=r22,PMD_SHIFT // shift pmd index into position 164 shr.u r18=r22,PMD_SHIFT // shift pmd index into position 174 shr [all...] |
H A D | gate.S | 254 shr.u r17=r17,16 257 shr.u r17=r17,3 // r17 <- (sc_loadrs >> 19) 264 shr.u r14=r14,6 // r14 <- (rse_slot_num(bspstore1) + (sc_loadrs >> 19))/0x40 282 shr r18=r18,63 284 shr r17=r17,5
|
/kernel/linux/linux-6.6/arch/ia64/kernel/ |
H A D | ivt.S | 123 shr.u r17=r16,61 // get the region number into r17 125 shr.u r22=r21,3 133 (p8) shr r22=r22,r27 137 shr.u r18=r22,PGDIR_SHIFT // get bottom portion of pgd index bit 145 (p6) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT 146 (p7) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT-3 152 shr.u r28=r22,PUD_SHIFT // shift pud index into position 154 shr.u r18=r22,PMD_SHIFT // shift pmd index into position 163 shr.u r18=r22,PMD_SHIFT // shift pmd index into position 173 shr [all...] |
H A D | gate.S | 254 shr.u r17=r17,16 257 shr.u r17=r17,3 // r17 <- (sc_loadrs >> 19) 264 shr.u r14=r14,6 // r14 <- (rse_slot_num(bspstore1) + (sc_loadrs >> 19))/0x40 282 shr r18=r18,63 284 shr r17=r17,5
|