/third_party/littlefs/scripts/ |
H A D | code.py | 109 # code size results 159 results = [] 308 results.append(r._replace(file=file)) 310 return results 313 def fold(Result, results, *, 328 for r in results: 331 results = results_ 333 # organize results into conflicts 335 for r in results: 348 def table(Result, results, diff_result [all...] |
H A D | data.py | 109 # data size results 159 results = [] 308 results.append(r._replace(file=file)) 310 return results 313 def fold(Result, results, *, 328 for r in results: 331 results = results_ 333 # organize results into conflicts 335 for r in results: 348 def table(Result, results, diff_result [all...] |
/third_party/vk-gl-cts/scripts/ |
H A D | run_nightly.py | 117 self.results = [] 121 results = [] 125 results.append(TestCaseResult(items[0], items[1])) 127 return results 134 batchResult.results = readTestCaseResultsFromCSV(filename) 161 for result in batchResult.results: 172 for result in batchResult.results: 178 results = [] 188 results.append(MultiResult(caseName, statusCodes)) 190 return results [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/ |
H A D | kselftest_harness.h | 126 snprintf(_metadata->results->reason, \ 127 sizeof(_metadata->results->reason), fmt, ##__VA_ARGS__); \ 130 _metadata->results->reason); \ 827 struct __test_results *results; member 970 memset(t->results->reason, 0, sizeof(t->results->reason)); in __run_test() 1000 ksft_test_result_skip("%s\n", t->results->reason[0] ? in __run_test() 1001 t->results->reason : "unknown"); in __run_test() 1013 struct __test_results *results; in test_harness_run() local 1028 results in test_harness_run() [all...] |
/kernel/linux/linux-6.6/tools/perf/tests/shell/lib/ |
H A D | perf_metric_validation.py | 21 self.allresults = dict() # metric results of all workload 22 self.allignoremetrics = dict() # metrics with no results or negative results 28 self.results = dict() # metric results of current workload 30 self.ignoremetrics= set() # metrics with no results or negative results, neg result counts as a failed test 73 return self.results[idx] 118 Get value of the metric from self.results. 123 @returns: list with value found in self.results; lis [all...] |
/third_party/python/Lib/test/ |
H A D | test_trace.py | 142 self.assertEqual(self.tracer.results().counts, expected) 154 self.assertEqual(self.tracer.results().counts, expected) 166 self.assertEqual(self.tracer.results().counts, expected) 181 self.assertEqual(self.tracer.results().counts, expected) 196 self.assertEqual(self.tracer.results().counts, expected) 215 self.assertEqual(self.tracer.results().counts, expected) 231 self.assertEqual(tracer.results().counts, expected) 259 self.assertEqual(self.tracer.results().counts[k], expected[k]) 281 self.assertEqual(self.tracer.results().calledfuncs, expected) 301 self.assertEqual(self.tracer.results() [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/ |
H A D | test_tcp_check_syncookie_user.c | 131 log_err("Can't clear results"); in run_test() 136 log_err("Can't clear results"); in run_test() 141 log_err("Can't clear results"); in run_test() 208 int results = -1; in main() local 217 results = get_map_fd_by_prog_id(atoi(argv[1]), &xdp); in main() 218 if (results < 0) { in main() 242 if (run_test(server, results, xdp)) in main() 245 if (run_test(server_v6, results, xdp)) in main() 255 close(results); in main()
|
/third_party/icu/icu4c/source/i18n/ |
H A D | csr2022.cpp | 149 UBool CharsetRecog_2022JP::match(InputText *textIn, CharsetMatch *results) const { in match() 154 results->set(textIn, this, confidence); in match() 165 UBool CharsetRecog_2022KR::match(InputText *textIn, CharsetMatch *results) const { in match() 170 results->set(textIn, this, confidence); in match() 180 UBool CharsetRecog_2022CN::match(InputText *textIn, CharsetMatch *results) const { in match() 185 results->set(textIn, this, confidence); in match()
|
/third_party/node/tools/ |
H A D | test-v8.bat | 23 echo running 'python tools\run-tests.py %common_v8_test_options% %v8_test_options% --slow-tests-cutoff 1000000 --json-test-results v8-tap.xml' 24 call python tools\run-tests.py %common_v8_test_options% %v8_test_options% --slow-tests-cutoff 1000000 --json-test-results v8-tap.xml 29 echo running 'python tools\run-tests.py %common_v8_test_options% intl --slow-tests-cutoff 1000000 --json-test-results v8-intl-tap.xml' 30 call python tools\run-tests.py %common_v8_test_options% intl --slow-tests-cutoff 1000000 --json-test-results ./v8-intl-tap.xml 35 echo running 'python tools\run-tests.py %common_v8_test_options% benchmarks --slow-tests-cutoff 1000000 --json-test-results v8-benchmarks-tap.xml' 36 call python tools\run-tests.py %common_v8_test_options% benchmarks --slow-tests-cutoff 1000000 --json-test-results ./v8-benchmarks-tap.xml
|
/third_party/node/deps/icu-small/source/i18n/ |
H A D | csr2022.cpp | 149 UBool CharsetRecog_2022JP::match(InputText *textIn, CharsetMatch *results) const { in match() 154 results->set(textIn, this, confidence); in match() 165 UBool CharsetRecog_2022KR::match(InputText *textIn, CharsetMatch *results) const { in match() 170 results->set(textIn, this, confidence); in match() 180 UBool CharsetRecog_2022CN::match(InputText *textIn, CharsetMatch *results) const { in match() 185 results->set(textIn, this, confidence); in match()
|
/third_party/skia/third_party/externals/icu/source/i18n/ |
H A D | csr2022.cpp | 149 UBool CharsetRecog_2022JP::match(InputText *textIn, CharsetMatch *results) const { in match() 154 results->set(textIn, this, confidence); in match() 165 UBool CharsetRecog_2022KR::match(InputText *textIn, CharsetMatch *results) const { in match() 170 results->set(textIn, this, confidence); in match() 180 UBool CharsetRecog_2022CN::match(InputText *textIn, CharsetMatch *results) const { in match() 185 results->set(textIn, this, confidence); in match()
|
/third_party/node/deps/v8/src/torque/ |
H A D | csa-generator.cc | 128 std::vector<std::string> results; in EmitInstruction() local 132 results.push_back(DefinitionToVariable(instruction.GetValueDefinition(i))); in EmitInstruction() 133 stack->Push(results.back()); in EmitInstruction() 141 PrintCommaSeparatedList(out(), results); in EmitInstruction() local 143 } else if (results.size() == 1) { in EmitInstruction() 144 out() << results[0] << " = "; in EmitInstruction() 185 std::vector<std::string> results; in EmitInstruction() local 189 results.push_back(DefinitionToVariable(instruction.GetValueDefinition(i))); in EmitInstruction() 190 stack->Push(results.back()); in EmitInstruction() 198 PrintCommaSeparatedList(out(), results); in EmitInstruction() local 321 std::vector<std::string> results; EmitInstruction() local 337 PrintCommaSeparatedList(out(), results); EmitInstruction() local 372 std::vector<std::string> results; EmitInstruction() local 412 PrintCommaSeparatedList(out(), results); EmitInstruction() local [all...] |
/third_party/skia/tools/ |
H A D | test_all.py | 22 results = unittest.TextTestRunner(verbosity=2).run(suite) 23 print(repr(results)) 24 if not results.wasSuccessful():
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/ |
H A D | test_tcp_check_syncookie_user.c | 133 log_err("Can't clear results"); in run_test() 138 log_err("Can't clear results"); in run_test() 143 log_err("Can't clear results"); in run_test() 228 int results = -1; in main() local 240 results = get_map_fd_by_prog_id(atoi(argv[1]), &xdp); in main() 241 if (results < 0) { in main() 277 if (run_test(server, results, xdp, in main() 281 if (run_test(server_v6, results, xdp, in main() 285 if (run_test(server_dual, results, xdp, in main() 297 close(results); in main() [all...] |
/third_party/ltp/ |
H A D | ltpmenu | 46 # April 23 2003 - Added PID to results filename. 51 # April 30, 2003 - Recoded results display to allow selection 52 # of results file. 53 # - Created variable to hold results filename 54 # - Added time to results filename. 152 # Description: The results generated after the ltp execution located under 153 # ltp-mmddyy/results/ directory in a text (ASCII) file called 154 # results.todaysdate. This function displays this file in a 155 # window. If the results file does not exit it displays an 157 # need to be executed inorder to view results [all...] |
/third_party/notofonts/scripts/ |
H A D | gather.py | 64 results = {} variable 79 results[repo_name] = { 87 results[repo_name]["issues"] = [] 89 results[repo_name]["issues"].append( 159 results[repo_name]["families"] = state[repo_name].get("families", {}) 168 for result in results.values(): 180 json.dump(results, open("docs/noto.json", "w"), indent=True, sort_keys=True)
|
/third_party/skia/infra/lottiecap/gold/ |
H A D | lottie-web-aggregator.go | 64 // contains all the results reported in through report_gold_data 65 var results []jsonio.Result var 80 results = []jsonio.Result{} 125 results = append(results, jsonio.Result{ 149 // dumpJSON writes out a JSON file with all the results, typically at the end of 166 results := jsonio.GoldResults{ 172 Results: results, 179 if err := enc.Encode(&results); err != nil {
|
/third_party/vk-gl-cts/scripts/verify/ |
H A D | verify.py | 54 results = readTestLog(filename) 61 # Verify number of results 62 if len(results) != len(mustpass): 63 messages.append(error(filename, "Wrong number of test results, expected %d, found %d" % (len(mustpass), len(results)))) 66 for ndx in range(len(results)): 67 result = results[ndx] 71 messages.append(error(filename, "Multiple results for " + result.name)) 73 # Verify that all results are present and valid 79 result = results[resultNd [all...] |
/third_party/mesa3d/src/amd/compiler/tests/ |
H A D | check_output.py | 461 def print_results(results, output, expected): 462 results = {name: result for name, result in results.items() if result.result == output} 463 results = {name: result for name, result in results.items() if (result.result == result.expected) == expected} 465 if not results: 469 for test, result in results.items(): 477 return len(results) 489 results = {} variable 511 results[full_nam [all...] |
/third_party/python/Lib/xmlrpc/ |
H A D | server.py | 351 results = [] 359 results.append([self._dispatch(method_name, params)]) 361 results.append( 366 results.append( 370 return results 712 results = [] 727 results.append(escape(text[here:start])) 732 results.append('<a href="%s">%s</a>' % (url, url)) 735 results.append('<a href="%s">%s</a>' % (url, escape(all))) 738 results [all...] |
/third_party/libwebsockets/minimal-examples/api-tests/api-test-smtp_client/ |
H A D | main.c | 16 static int interrupted, results[10], count_tests, count_passes; variable 227 args.results = results; in main() 228 args.results_max = LWS_ARRAY_SIZE(results); in main() 244 /* describe the overall test results */ in main() 250 lws_unit_test_result_name(results[n])); in main()
|
/third_party/node/deps/npm/node_modules/fastest-levenshtein/ |
H A D | bench.js | 81 var results = new Map(); 85 if (results.has(event.target.name[0])) { 86 results.get(event.target.name[0]).push(event.target.hz); 89 results.set(event.target.name[0], [event.target.hz]); 93 console.log(results);
|
/third_party/vk-gl-cts/external/vulkan-docs/src/scripts/Retired/ |
H A D | findBalance.py | 149 results = parser.parse_args() variable 151 setLogFile(True, True, results.logFile) 152 setLogFile(True, False, results.diagFile) 153 setLogFile(False, True, results.warnFile) 156 for file in results.files:
|
/third_party/python/Lib/lib2to3/fixes/ |
H A D | fix_xreadlines.py | 19 def transform(self, node, results): 20 no_call = results.get("no_call") 25 node.replace([x.clone() for x in results["call"]])
|
/third_party/skia/buildtools/checkdeps/ |
H A D | PRESUBMIT.py | 12 results = [] 13 results.extend(input_api.canned_checks.RunUnitTests( 17 return results
|