/kernel/linux/linux-5.10/arch/ia64/kernel/ |
H A D | ivt.S | 137 cmp.eq p6,p7=5,r17 // is IFA pointing into to region 5? 140 (p7) dep r17=r17,r19,(PAGE_SHIFT-3),3 // put region number bits in place 145 .pred.rel "mutex", p6, p7 147 (p7) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT-3 150 (p7) dep r17=r18,r17,3,(PAGE_SHIFT-6) // r17=pgd_offset for region[0-4] 151 cmp.eq p7,p6=0,r21 // unused address bits all zeroes? 160 (p7) cmp.eq p6,p7=r17,r0 // was pgd_present(*pgd) == NULL? 165 (p7) ld8 r29=[r28] // get *pud (may be 0) 167 (p7) cm [all...] |
H A D | fsys.S | 102 tnat.z p6,p7=r32 // check argument register for being NaT 122 (p7) st8 [r18]=r17 180 // p6,p7 short term use 244 (p13) cmp.gt.unc p6,p7 = r3,r0 // check if it is less than last. p6,p7 cleared 248 (p7) mov ar.ccv = r25 // more than last_cycle. Prep for cmpxchg 250 (p7) cmpxchg8.rel r3 = [r19],r2,ar.ccv 252 (p7) cmp.ne p7,p0 = r25,r3 // if cmpxchg not successful 254 (p7) su [all...] |
H A D | entry.S | 74 cmp4.ge p6,p7=r8,r0 81 (p7) br.ret.sptk.many rp 180 cmp.eq p7,p6=r25,in0 185 (p6) cmp.eq p7,p6=r26,r27 504 cmp.leu p6,p7=r15,r3 507 (p7) movl r20=sys_ni_syscall 608 cmp.ge p6,p7=r8,r0 // syscall executed successfully? 611 (p7) br.cond.spnt handle_syscall_error // handle potential syscall failure 1012 # define pReturn p7 1165 cmp.eq p6,p7 [all...] |
H A D | gate.S | 271 cmp.lt p7,p0=r14,r0 // p7 <- (r14 < 0)? 273 (p7) adds r18=-62,r18 // delta -= 62 339 cmp.geu p6,p7=r19,r17 // A (sysnr > 0 && sysnr < 1024+NR_syscalls)?
|
/kernel/linux/linux-6.6/arch/ia64/kernel/ |
H A D | ivt.S | 136 cmp.eq p6,p7=5,r17 // is IFA pointing into to region 5? 139 (p7) dep r17=r17,r19,(PAGE_SHIFT-3),3 // put region number bits in place 144 .pred.rel "mutex", p6, p7 146 (p7) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT-3 149 (p7) dep r17=r18,r17,3,(PAGE_SHIFT-6) // r17=pgd_offset for region[0-4] 150 cmp.eq p7,p6=0,r21 // unused address bits all zeroes? 159 (p7) cmp.eq p6,p7=r17,r0 // was pgd_present(*pgd) == NULL? 164 (p7) ld8 r29=[r28] // get *pud (may be 0) 166 (p7) cm [all...] |
H A D | fsys.S | 102 tnat.z p6,p7=r32 // check argument register for being NaT 122 (p7) st8 [r18]=r17 180 // p6,p7 short term use 244 (p13) cmp.gt.unc p6,p7 = r3,r0 // check if it is less than last. p6,p7 cleared 248 (p7) mov ar.ccv = r25 // more than last_cycle. Prep for cmpxchg 250 (p7) cmpxchg8.rel r3 = [r19],r2,ar.ccv 252 (p7) cmp.ne p7,p0 = r25,r3 // if cmpxchg not successful 254 (p7) su [all...] |
H A D | entry.S | 73 cmp4.ge p6,p7=r8,r0 80 (p7) br.ret.sptk.many rp 179 cmp.eq p7,p6=r25,in0 184 (p6) cmp.eq p7,p6=r26,r27 503 cmp.leu p6,p7=r15,r3 506 (p7) movl r20=sys_ni_syscall 607 cmp.ge p6,p7=r8,r0 // syscall executed successfully? 610 (p7) br.cond.spnt handle_syscall_error // handle potential syscall failure 1011 # define pReturn p7 1164 cmp.eq p6,p7 [all...] |
H A D | gate.S | 271 cmp.lt p7,p0=r14,r0 // p7 <- (r14 < 0)? 273 (p7) adds r18=-62,r18 // delta -= 62 339 cmp.geu p6,p7=r19,r17 // A (sysnr > 0 && sysnr < 1024+NR_syscalls)?
|
/kernel/linux/linux-5.10/arch/ia64/lib/ |
H A D | clear_user.S | 160 // error handling. We also reuse p6/p7 to return correct value. 171 tbit.nz p7,p6=len,2 173 EX( .Lexit2, (p7) st4 [buf]=r0,4 ) // at least 4 bytes 174 (p7) adds len2=-4,len3 175 tbit.nz p6,p7=len,1 179 tbit.nz p7,p6=len,0 181 EX( .Lexit2, (p7) st1 [buf]=r0 ) // only 1 byte left 196 // if p7 -> coming from st4 or st1 : len3 contains what's left 199 .pred.rel "mutex", p6, p7 201 (p7) mo [all...] |
H A D | copy_user.S | 105 cmp.lt p10,p7=COPY_BREAK,len // if len > COPY_BREAK then long copy 113 // p7 is necessarily false by now 264 SWITCH(p7, 16) 272 CASE(p7, 16) 321 tbit.nz p6,p7=src1,0 // odd alignment 336 tbit.nz p7,p0=src1,1 338 EX(.failure_in1,(p7) ld2 val1[1]=[src1],2) // 2-byte aligned 339 (p7) adds len1=-2,len1;; 356 EX(.failure_out,(p7) st2 [dst1]=val1[1],2) 365 cmp.eq p7,p [all...] |
H A D | memcpy_mck.S | 120 cmp.lt p7,p8=1,cnt 128 (p7) mov ar.lc=cnt // prefetch count 149 cmp.le p7,p0=16,tmp 173 EX(.ex_hndlr_s, (p7) ld8 t2=[src1],8) 180 EX(.ex_hndlr_d, (p7) st8 [dst1]=t2,8) // store byte 2 292 cmp.lt p6,p7=blocksize,in2 302 cmp.lt p7,p8=1,r21 309 (p7) mov ar.lc = cnt 321 cmp.le p6,p7=8,curlen; // have at least 8 byte remaining? 430 cmp.le p6,p7 [all...] |
H A D | strlen.S | 131 tnat.nz p6,p7=val1 // test NaT on val1 135 // if we come here p7 is true, i.e., initialized for // cmp 137 cmp.eq.and p7,p0=8,val1// val1==8? 138 tnat.nz.and p7,p0=val2 // test NaT if val2 139 (p7) br.cond.spnt .recover // jump to recovery if val2 is NaT
|
H A D | ip_fast_csum.S | 38 cmp.ne p6,p7=5,in1 // size other than 20 byte? 42 cmp.ne.or.andcm p6,p7=r14,r0 44 (p7) ld4 r20=[in0],8 45 (p7) ld4 r21=[r15],8
|
H A D | do_csum.S | 246 cmp.ltu p7,p0=result2[LOAD_LATENCY+1],carry2 249 (p7) adds result2[LOAD_LATENCY+1]=1,result2[LOAD_LATENCY+1]
|
H A D | memcpy.S | 74 cmp.gtu p7,p0=16,in2 // copying less than 16 bytes? 84 (p7) br.cond.spnt.few .memcpy_short
|
/kernel/linux/linux-6.6/arch/ia64/lib/ |
H A D | clear_user.S | 160 // error handling. We also reuse p6/p7 to return correct value. 171 tbit.nz p7,p6=len,2 173 EX( .Lexit2, (p7) st4 [buf]=r0,4 ) // at least 4 bytes 174 (p7) adds len2=-4,len3 175 tbit.nz p6,p7=len,1 179 tbit.nz p7,p6=len,0 181 EX( .Lexit2, (p7) st1 [buf]=r0 ) // only 1 byte left 196 // if p7 -> coming from st4 or st1 : len3 contains what's left 199 .pred.rel "mutex", p6, p7 201 (p7) mo [all...] |
H A D | copy_user.S | 105 cmp.lt p10,p7=COPY_BREAK,len // if len > COPY_BREAK then long copy 113 // p7 is necessarily false by now 264 SWITCH(p7, 16) 272 CASE(p7, 16) 321 tbit.nz p6,p7=src1,0 // odd alignment 336 tbit.nz p7,p0=src1,1 338 EX(.failure_in1,(p7) ld2 val1[1]=[src1],2) // 2-byte aligned 339 (p7) adds len1=-2,len1;; 356 EX(.failure_out,(p7) st2 [dst1]=val1[1],2) 365 cmp.eq p7,p [all...] |
H A D | memcpy_mck.S | 120 cmp.lt p7,p8=1,cnt 128 (p7) mov ar.lc=cnt // prefetch count 149 cmp.le p7,p0=16,tmp 173 EX(.ex_hndlr_s, (p7) ld8 t2=[src1],8) 180 EX(.ex_hndlr_d, (p7) st8 [dst1]=t2,8) // store byte 2 292 cmp.lt p6,p7=blocksize,in2 302 cmp.lt p7,p8=1,r21 309 (p7) mov ar.lc = cnt 321 cmp.le p6,p7=8,curlen; // have at least 8 byte remaining? 430 cmp.le p6,p7 [all...] |
H A D | strlen.S | 131 tnat.nz p6,p7=val1 // test NaT on val1 135 // if we come here p7 is true, i.e., initialized for // cmp 137 cmp.eq.and p7,p0=8,val1// val1==8? 138 tnat.nz.and p7,p0=val2 // test NaT if val2 139 (p7) br.cond.spnt .recover // jump to recovery if val2 is NaT
|
H A D | ip_fast_csum.S | 38 cmp.ne p6,p7=5,in1 // size other than 20 byte? 42 cmp.ne.or.andcm p6,p7=r14,r0 44 (p7) ld4 r20=[in0],8 45 (p7) ld4 r21=[r15],8
|
/kernel/linux/linux-5.10/tools/testing/selftests/net/ |
H A D | fib-onlink-tests.sh | 20 NETIFS[p7]=veth7 31 V4ADDRS[p7]=169.254.7.1 42 V6ADDRS[p7]=2001:db8:701::1 192 ip li add ${NETIFS[p7]} type veth peer name ${NETIFS[p8]} 214 ip -6 ro add table ${VRF_TABLE} default via ${V6ADDRS[p7]/::[0-9]/::64} 325 run_ip ${VRF_TABLE} ${TEST_NET4[2]}.103 ${V4ADDRS[p7]} ${NETIFS[p5]} 2 \ 449 run_ip6 ${VRF_TABLE} ${TEST_NET6[2]}::103 ${V6ADDRS[p7]/::[0-9]/::64} ${NETIFS[p5]} 2 \
|
/kernel/linux/linux-6.6/tools/testing/selftests/net/ |
H A D | fib-onlink-tests.sh | 20 NETIFS[p7]=veth7 31 V4ADDRS[p7]=169.254.7.1 42 V6ADDRS[p7]=2001:db8:701::1 192 ip li add ${NETIFS[p7]} type veth peer name ${NETIFS[p8]} 214 ip -6 ro add table ${VRF_TABLE} default via ${V6ADDRS[p7]/::[0-9]/::64} 325 run_ip ${VRF_TABLE} ${TEST_NET4[2]}.103 ${V4ADDRS[p7]} ${NETIFS[p5]} 2 \ 449 run_ip6 ${VRF_TABLE} ${TEST_NET6[2]}::103 ${V6ADDRS[p7]/::[0-9]/::64} ${NETIFS[p5]} 2 \
|
/kernel/linux/linux-5.10/include/media/i2c/ |
H A D | mt9t112.h | 13 u8 p1, p2, p3, p4, p5, p6, p7; member
|
/kernel/linux/linux-6.6/include/media/i2c/ |
H A D | mt9t112.h | 13 u8 p1, p2, p3, p4, p5, p6, p7; member
|
/kernel/linux/linux-5.10/samples/bpf/ |
H A D | spintest_kern.c | 55 SEC("kprobe/_raw_spin_unlock_bh")PROG(p7)
|