Lines Matching refs:t0

117 my ($t0,$t1,$t2,$t3,$t4)=("%rax","%rdx","%rcx","%r12","%r13");
141 mov $a0, $t0
155 cmovc $t0, $a0
191 mov $a0, $t0
205 test \$1, $t0
207 cmovz $t0, $a0
213 mov $a1, $t0 # a0:a3>>1
215 shl \$63, $t0
218 or $t0, $a0
265 mov $a0, $t0
279 cmovc $t0, $a0
287 mov $a0, $t0
301 cmovc $t0, $a0
343 mov $a0, $t0
357 cmovc $t0, $a0
399 mov $a0, $t0
413 cmovz $t0, $a0
455 mov $a0, $t0
469 cmovz $t0, $a0
493 my ($t0,$t1,$t2,$t3,$t4)=("%rcx","%rbp","%rbx","%rdx","%rax");
536 mov %rax, $t0
539 mov $t0, %rax
544 mov $t0, %rax
550 mov $t0, %rax
569 mov %rdx, $t0
575 add $t0, $acc1
594 mov %rax, $t0
597 mov $t0, %rax
605 mov $t0, %rax
613 mov $t0, %rax
616 mov $acc1, $t0
632 add %rax, $t0 # guaranteed to be zero
634 adc %rdx, $t0
640 add $t0, $acc2
659 mov %rax, $t0
662 mov $t0, %rax
670 mov $t0, %rax
678 mov $t0, %rax
681 mov $acc2, $t0
697 add %rax, $t0 # guaranteed to be zero
699 adc %rdx, $t0
705 add $t0, $acc3
724 mov %rax, $t0
727 mov $t0, %rax
735 mov $t0, %rax
743 mov $t0, %rax
746 mov $acc3, $t0
762 add %rax, $t0 # guaranteed to be zero
764 adc %rdx, $t0
770 add $t0, $acc4
792 mov $acc0, $t0
800 cmovc $t0, $acc0
944 mov $acc0, $t0
956 add %rax, $t0 # guaranteed to be zero
958 adc %rdx, $t0
964 add $t0, $acc1
972 mov $acc1, $t0
987 add %rax, $t0 # guaranteed to be zero
989 adc %rdx, $t0
995 add $t0, $acc2
1003 mov $acc2, $t0
1018 add %rax, $t0 # guaranteed to be zero
1020 adc %rdx, $t0
1026 add $t0, $acc3
1034 mov $acc3, $t0
1049 add %rax, $t0 # guaranteed to be zero
1051 adc %rdx, $t0
1057 add $t0, $acc0
1161 mulx $acc2, $t0, $acc2
1163 add $t0, $acc1
1164 mulx $acc4, $t0, $acc4
1168 adc $t0, $acc3
1173 mulx 8*0+128(%r14), $t0, $t1
1174 adcx $t0, $acc0 # guaranteed to be zero
1177 mulx 8*1+128(%r14), $t0, $t1
1178 adcx $t0, $acc1
1181 mulx 8*2+128(%r14), $t0, $t1
1182 adcx $t0, $acc2
1185 mulx 8*3+128(%r14), $t0, $t1
1187 adcx $t0, $acc3
1194 mulx 8*0+128($a_ptr), $t0, $t1
1195 adcx $t0, $acc1
1198 mulx 8*1+128($a_ptr), $t0, $t1
1199 adcx $t0, $acc2
1202 mulx 8*2+128($a_ptr), $t0, $t1
1203 adcx $t0, $acc3
1206 mulx 8*3+128($a_ptr), $t0, $t1
1209 adcx $t0, $acc4
1217 mulx 8*0+128(%r14), $t0, $t1
1218 adcx $t0, $acc1 # guaranteed to be zero
1221 mulx 8*1+128(%r14), $t0, $t1
1222 adcx $t0, $acc2
1225 mulx 8*2+128(%r14), $t0, $t1
1226 adcx $t0, $acc3
1229 mulx 8*3+128(%r14), $t0, $t1
1231 adcx $t0, $acc4
1238 mulx 8*0+128($a_ptr), $t0, $t1
1239 adcx $t0, $acc2
1242 mulx 8*1+128($a_ptr), $t0, $t1
1243 adcx $t0, $acc3
1246 mulx 8*2+128($a_ptr), $t0, $t1
1247 adcx $t0, $acc4
1250 mulx 8*3+128($a_ptr), $t0, $t1
1253 adcx $t0, $acc5
1261 mulx 8*0+128(%r14), $t0, $t1
1262 adcx $t0, $acc2 # guaranteed to be zero
1265 mulx 8*1+128(%r14), $t0, $t1
1266 adcx $t0, $acc3
1269 mulx 8*2+128(%r14), $t0, $t1
1270 adcx $t0, $acc4
1273 mulx 8*3+128(%r14), $t0, $t1
1275 adcx $t0, $acc5
1282 mulx 8*0+128($a_ptr), $t0, $t1
1283 adcx $t0, $acc3
1286 mulx 8*1+128($a_ptr), $t0, $t1
1287 adcx $t0, $acc4
1290 mulx 8*2+128($a_ptr), $t0, $t1
1291 adcx $t0, $acc5
1294 mulx 8*3+128($a_ptr), $t0, $t1
1297 adcx $t0, $acc0
1305 mulx 8*0+128(%r14), $t0, $t1
1306 adcx $t0, $acc3 # guaranteed to be zero
1309 mulx 8*1+128(%r14), $t0, $t1
1310 adcx $t0, $acc4
1313 mulx 8*2+128(%r14), $t0, $t1
1314 adcx $t0, $acc5
1317 mulx 8*3+128(%r14), $t0, $t1
1320 adcx $t0, $acc0
1329 mov $acc0, $t0
1339 cmovc $t0, $acc0
1396 mulx $acc7, $t0, $acc3 # a[0]*a[2]
1401 add $t0, $acc2
1407 mulx $acc7, $t0, $t1 # a[1]*a[2]
1408 adcx $t0, $acc3
1411 mulx $acc0, $t0, $t1 # a[1]*a[3]
1413 adcx $t0, $acc4
1417 mulx $acc0, $t0, $acc6 # a[2]*a[3]
1422 adox $t0, $acc5
1432 mulx %rdx, $t0, $t4
1435 adox $t0, $acc2
1437 mulx %rdx, $t0, $t1
1442 adox $t0, $acc4
1444 mulx %rdx, $t0, $t4
1445 adox $t0, $acc6
1450 mulx 8*4($a_ptr), %rdx, $t0
1453 mulx 8*0($a_ptr), $t0, $t1
1454 adcx $t0, $acc0 # guaranteed to be zero
1456 mulx 8*1($a_ptr), $t0, $t1
1457 adcx $t0, $acc1
1459 mulx 8*2($a_ptr), $t0, $t1
1460 adcx $t0, $acc2
1462 mulx 8*3($a_ptr), $t0, $t1
1463 adcx $t0, $acc3
1469 mulx 8*4($a_ptr), %rdx, $t0
1471 mulx 8*0($a_ptr), $t0, $t1
1472 adox $t0, $acc1 # guaranteed to be zero
1474 mulx 8*1($a_ptr), $t0, $t1
1475 adox $t0, $acc2
1477 mulx 8*2($a_ptr), $t0, $t1
1478 adox $t0, $acc3
1480 mulx 8*3($a_ptr), $t0, $t1
1481 adox $t0, $acc0
1487 mulx 8*4($a_ptr), %rdx, $t0
1489 mulx 8*0($a_ptr), $t0, $t1
1490 adcx $t0, $acc2 # guaranteed to be zero
1492 mulx 8*1($a_ptr), $t0, $t1
1493 adcx $t0, $acc3
1495 mulx 8*2($a_ptr), $t0, $t1
1496 adcx $t0, $acc0
1498 mulx 8*3($a_ptr), $t0, $t1
1499 adcx $t0, $acc1
1505 mulx 8*4($a_ptr), %rdx, $t0
1507 mulx 8*0($a_ptr), $t0, $t1
1508 adox $t0, $acc3 # guaranteed to be zero
1510 mulx 8*1($a_ptr), $t0, $t1
1511 adox $t0, $acc0
1513 mulx 8*2($a_ptr), $t0, $t1
1514 adox $t0, $acc1
1516 mulx 8*3($a_ptr), $t0, $t1
1517 adox $t0, $acc2
1739 mov %rdx, $t0
1742 add $t0, $acc2
1747 mov %rdx, $t0
1750 add $t0, $acc3
1755 mov %rdx, $t0
1758 add $t0, $acc4
1786 mov %rdx, $t0
1789 add $t0, $acc3
1794 mov %rdx, $t0
1797 add $t0, $acc4
1802 mov %rdx, $t0
1805 add $t0, $acc5
1833 mov %rdx, $t0
1836 add $t0, $acc4
1841 mov %rdx, $t0
1844 add $t0, $acc5
1849 mov %rdx, $t0
1852 add $t0, $acc0
1867 mov $acc4, $t0
1883 cmovc $t0, $acc4
2030 mov %rdx, $t0
2033 add $t0, $acc1
2037 mov %rdx, $t0
2040 add $t0, $acc3
2044 mov %rdx, $t0
2047 add $t0, $acc5
2058 mov $acc0, $t0
2061 shr \$32, $t0
2063 adc $t0, $acc2
2070 mov $acc1, $t0
2074 shr \$32, $t0
2076 adc $t0, $acc3
2083 mov $acc2, $t0
2087 shr \$32, $t0
2089 adc $t0, $acc0
2096 mov $acc3, $t0
2100 shr \$32, $t0
2102 adc $t0, $acc1
2121 mov $acc7, $t0
2130 cmovc $t0, $acc7
2148 mulx $acc2, $t0, $acc2
2153 adc $t0, $acc1
2154 mulx $acc4, $t0, $acc4
2158 adc $t0, $acc3
2159 shrx $poly1,$acc0,$t0
2165 adc $t0, $acc2
2167 mulx $poly3, $t0, $t1
2169 adc $t0, $acc3
2176 mulx 8*0+128($a_ptr), $t0, $t1
2177 adcx $t0, $acc1
2180 mulx 8*1+128($a_ptr), $t0, $t1
2181 adcx $t0, $acc2
2184 mulx 8*2+128($a_ptr), $t0, $t1
2185 adcx $t0, $acc3
2188 mulx 8*3+128($a_ptr), $t0, $t1
2190 adcx $t0, $acc4
2191 shlx $poly1, $acc1, $t0
2201 add $t0, $acc2
2204 mulx $poly3, $t0, $t1
2206 adc $t0, $acc4
2213 mulx 8*0+128($a_ptr), $t0, $t1
2214 adcx $t0, $acc2
2217 mulx 8*1+128($a_ptr), $t0, $t1
2218 adcx $t0, $acc3
2221 mulx 8*2+128($a_ptr), $t0, $t1
2222 adcx $t0, $acc4
2225 mulx 8*3+128($a_ptr), $t0, $t1
2227 adcx $t0, $acc5
2228 shlx $poly1, $acc2, $t0
2238 add $t0, $acc3
2241 mulx $poly3, $t0, $t1
2243 adc $t0, $acc5
2250 mulx 8*0+128($a_ptr), $t0, $t1
2251 adcx $t0, $acc3
2254 mulx 8*1+128($a_ptr), $t0, $t1
2255 adcx $t0, $acc4
2258 mulx 8*2+128($a_ptr), $t0, $t1
2259 adcx $t0, $acc5
2262 mulx 8*3+128($a_ptr), $t0, $t1
2264 adcx $t0, $acc0
2265 shlx $poly1, $acc3, $t0
2275 add $t0, $acc4
2278 mulx $poly3, $t0, $t1
2281 adc $t0, $acc0
2289 mov $acc0, $t0
2300 cmovc $t0, $acc0
2315 mulx $acc7, $t0, $acc3 # a[0]*a[2]
2317 adc $t0, $acc2
2325 mulx $acc7, $t0, $t1 # a[1]*a[2]
2326 adcx $t0, $acc3
2329 mulx $acc0, $t0, $t1 # a[1]*a[3]
2331 adcx $t0, $acc4
2336 mulx $acc0, $t0, $acc6 # a[2]*a[3]
2340 adox $t0, $acc5
2349 mulx %rdx, $t0, $t4
2352 adox $t0, $acc2
2355 mulx %rdx, $t0, $t1
2359 adox $t0, $acc4
2363 mulx %rdx, $t0, $t4
2365 adox $t0, $acc6
2366 shlx $a_ptr, $acc0, $t0
2372 add $t0, $acc1
2375 mulx $acc0, $t0, $acc0
2376 adc $t0, $acc3
2377 shlx $a_ptr, $acc1, $t0
2382 add $t0, $acc2
2385 mulx $acc1, $t0, $acc1
2386 adc $t0, $acc0
2387 shlx $a_ptr, $acc2, $t0
2392 add $t0, $acc3
2395 mulx $acc2, $t0, $acc2
2396 adc $t0, $acc1
2397 shlx $a_ptr, $acc3, $t0
2402 add $t0, $acc0
2405 mulx $acc3, $t0, $acc3
2406 adc $t0, $acc2
2445 my ($t0,$t1,$t2)=("%rcx","%r12","%r13");
2475 mov %rax, $t0
2478 shr \$32, $t0
2480 adc $t0, $acc2
2487 mov $acc1, $t0
2491 shr \$32, $t0
2493 adc $t0, $acc3
2500 mov $acc2, $t0
2504 shr \$32, $t0
2506 adc $t0, $acc0
2513 mov $acc3, $t0
2517 shr \$32, $t0
2519 adc $t0, $acc1
2520 mov $acc0, $t0
2535 cmovnz $t0, $acc0
3056 my ($t0,$t1,$t2,$t3,$t4)=("%rax","%rbp","%rcx",$acc4,$acc4);
3097 mov $a0, $t0
3111 cmovc $t0, $a0
3130 mov $a0, $t0
3144 cmovz $t0, $a0
3161 sub $a0, $t0
3163 mov $t0, $a0
3169 add \$-1, $t0
3177 cmovnz $t0, $a0
3193 mov $a0, $t0
3207 cmovc $t0, $a0
3347 mov $a0, $t0
3357 test \$1, $t0
3359 cmovz $t0, $a0
3365 mov $a1, $t0 # a0:a3>>1
3367 shl \$63, $t0
3370 or $t0, $a0
3416 mov $S+8*0(%rsp), $t0
3693 mov $acc0, $t0
3707 cmovc $t0, $acc0
3708 mov 8*0($a_ptr), $t0
3722 mov $U2+8*0(%rsp), $t0
4018 mov $acc0, $t0
4032 cmovc $t0, $acc0
4033 mov 8*0($a_ptr), $t0
4047 mov $U2+8*0(%rsp), $t0
4189 mov $a0, $t0
4204 cmovc $t0, $a0
4224 mov $a0, $t0
4239 cmovnc $t0, $a0
4257 sbb $a0, $t0
4259 mov $t0, $a0
4266 adc \$-1, $t0
4274 cmovc $t0, $a0
4290 mov $a0, $t0
4305 cmovc $t0, $a0