/third_party/mesa3d/src/intel/vulkan/tests/ |
H A D | state_pool_no_free.c | 38 } jobs[NUM_THREADS]; variable 71 jobs[i].pool = &state_pool; in run_test() 72 jobs[i].id = i; in run_test() 73 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_test() 77 pthread_join(jobs[i].thread, NULL); in run_test() 92 if (thread_max < jobs[i].offsets[next[i]]) { in run_test() 93 thread_max = jobs[i].offsets[next[i]]; in run_test() 105 ASSERT(jobs[max_thread_idx].offsets[next[max_thread_idx]] > highest); in run_test() 107 highest = jobs[max_thread_id in run_test() [all...] |
H A D | block_pool_no_free.c | 39 } jobs[NUM_THREADS]; variable 45 uint32_t job_id = job - jobs; in alloc_blocks() 126 jobs[i].pool = &pool; in run_test() 127 jobs[i].id = i; in run_test() 128 pthread_create(&jobs[i].thread, NULL, alloc_blocks, &jobs[i]); in run_test() 132 pthread_join(jobs[i].thread, NULL); in run_test() 137 block_ptrs[i] = jobs[i].blocks; in run_test() 142 block_ptrs[i] = jobs[i].back_blocks; in run_test()
|
H A D | state_pool_test_helper.h | 30 } jobs[NUM_THREADS]; variable 64 jobs[i].pool = state_pool; in run_state_pool_test() 65 jobs[i].id = i; in run_state_pool_test() 66 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_state_pool_test() 70 pthread_join(jobs[i].thread, NULL); in run_state_pool_test()
|
/third_party/mesa3d/src/util/ |
H A D | u_queue.c | 295 job = queue->jobs[queue->read_idx]; in util_queue_thread_func() 296 memset(&queue->jobs[queue->read_idx], 0, sizeof(struct util_queue_job)); in util_queue_thread_func() 314 /* signal remaining jobs if all threads are being terminated */ in util_queue_thread_func() 319 if (queue->jobs[i].job) { in util_queue_thread_func() 320 if (queue->jobs[i].fence) in util_queue_thread_func() 321 util_queue_fence_signal(queue->jobs[i].fence); in util_queue_thread_func() 322 queue->jobs[i].job = NULL; in util_queue_thread_func() 448 queue->jobs = (struct util_queue_job*) in util_queue_init() 450 if (!queue->jobs) in util_queue_init() 477 if (queue->jobs) { in util_queue_init() 583 struct util_queue_job *jobs = util_queue_add_job() local [all...] |
/third_party/libdrm/tests/exynos/ |
H A D | exynos_fimg2d_event.c | 103 * We need to wait until all G2D jobs are finished, otherwise we 110 static void wait_all_jobs(struct g2d_job* jobs, unsigned num_jobs) in wait_all_jobs() argument 115 while (jobs[i].busy) in wait_all_jobs() 121 static struct g2d_job* free_job(struct g2d_job* jobs, unsigned num_jobs) in free_job() argument 126 if (jobs[i].busy == 0) in free_job() 127 return &jobs[i]; in free_job() 136 struct g2d_job *jobs = calloc(num_jobs, sizeof(struct g2d_job)); in g2d_work() local 142 jobs[i].id = i; in g2d_work() 150 j = free_job(jobs, num_jobs); in g2d_work() 189 wait_all_jobs(jobs, num_job in g2d_work() [all...] |
/third_party/vixl/tools/ |
H A D | test.py | 149 '--jobs', '-j', metavar='N', type=int, nargs='?', 152 help='''Runs the tests using N jobs. If the option is set but no value is 153 provided, the script will use as many jobs as it thinks useful.''') 259 def RunLinter(jobs): 261 jobs = args.jobs, progress_prefix = 'cpp lint: ') 264 def RunClangFormat(clang_path, jobs): 267 jobs = jobs, 270 def RunClangTidy(clang_path, jobs) [all...] |
H A D | lint.py | 61 parser.add_argument('--jobs', '-j', metavar='N', type=int, nargs='?', 64 help='''Runs the tests using N jobs. If the option is set 65 but no value is provided, the script will use as many jobs 131 jobs = 1, 158 pool = multiprocessing.Pool(jobs) 228 def RunLinter(files, jobs=1, progress_prefix='', cached=True): 232 jobs=jobs, 247 retcode = RunLinter(files, jobs=args.jobs, cache [all...] |
H A D | clang_tidy.py | 49 parser.add_argument('--jobs', '-j', metavar='N', type=int, nargs='?', 52 help='''Runs the tests using N jobs. If the option is set 53 but no value is provided, the script will use as many jobs 117 def ClangTidyFiles(files, clang_tidy, jobs = 1, progress_prefix = ''): 146 return queue.Run(jobs, True, RunTest) 153 rc = ClangTidyFiles(files, args.clang_tidy, args.jobs)
|
H A D | clang_format.py | 64 parser.add_argument('--jobs', '-j', metavar = 'N', type = int, nargs = '?', 67 help = '''Runs the tests using N jobs. If the option is set 68 but no value is provided, the script will use as many jobs 170 def ClangFormatFiles(files, clang_format, in_place = False, jobs = 1, 185 rc = queue.Run(jobs, True, RunTest) 200 in_place = args.in_place, jobs = args.jobs)
|
/third_party/skia/tools/skqp/ |
H A D | find_commit_with_best_gold_results.py | 23 path = skia_directory + '/infra/bots/jobs.json' 36 jobs = json.load(f) 37 for job in jobs: 62 def get_results_for_commit(commit, jobs): 81 for job in jobs for config in CONFIGS] 94 jobs = [j for j in get_jobs()] 98 results.append((commit_hash, get_results_for_commit(commit_hash, jobs)))
|
H A D | cut_release.py | 38 jobs = [] 44 jobs.append((rec['testName'], urls, exe)) 46 pool.map(make_skqp_model, jobs) 48 return set((n for n, _, _ in jobs))
|
/third_party/node/deps/openssl/openssl/crypto/async/ |
H A D | async.c | 112 * no max size and no pre-created jobs in async_get_pool_job() 119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job() 147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job() 324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool() 328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool() 374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread() 375 if (pool->jobs == NULL) { in ASYNC_init_thread() 383 /* Pre-create jobs as required */ in ASYNC_init_thread() 390 * skip creation of any more jobs in ASYNC_init_thread() 396 sk_ASYNC_JOB_push(pool->jobs, jo in ASYNC_init_thread() [all...] |
/third_party/openssl/crypto/async/ |
H A D | async.c | 112 * no max size and no pre-created jobs in async_get_pool_job() 119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job() 147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job() 324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool() 328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool() 374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread() 375 if (pool->jobs == NULL) { in ASYNC_init_thread() 383 /* Pre-create jobs as required */ in ASYNC_init_thread() 390 * skip creation of any more jobs in ASYNC_init_thread() 396 sk_ASYNC_JOB_push(pool->jobs, jo in ASYNC_init_thread() [all...] |
/third_party/mbedtls/tests/scripts/ |
H A D | depends.py | 14 Then, test domains (groups of jobs, tests) are built based on predefined data 42 Lastly, the collected jobs are executed and (optionally) tested, with 313 """Initialize the jobs container""" 314 self.jobs = [] 322 The domain contains a set of jobs that enable one of the elements 325 If exclude is a regular expression, skip generated jobs whose description 340 self.jobs.append(job) 346 If exclude is a regular expression, skip generated jobs whose description 360 self.jobs.append(job) 365 each call adds respective jobs [all...] |
/third_party/rust/crates/regex/src/ |
H A D | backtrack.rs | 56 jobs: Vec<Job>, 63 Cache { jobs: vec![], visited: vec![] } in new() 104 self.m.jobs.clear(); in clear() 168 self.m.jobs.push(Job::Inst { ip: 0, at: start }); in backtrack() 169 while let Some(job) = self.m.jobs.pop() { in backtrack() 215 self.m.jobs.push(Job::SaveRestore { in step() 224 self.m.jobs.push(Job::Inst { ip: inst.goto2, at }); in step()
|
/third_party/node/benchmark/crypto/ |
H A D | webcrypto-digest.js | 41 const jobs = new Array(n); 44 jobs[i] = subtle.digest(method, data); 45 Promise.all(jobs).then(() => bench.end(n)).catch((err) => {
|
/third_party/openssl/test/ |
H A D | run_tests.pl | 34 my $jobs = $ENV{HARNESS_JOBS} // 1; 50 if ($jobs > 1) { 54 $tapargs{jobs} = $jobs; 55 print "Using HARNESS_JOBS=$jobs\n"; 85 if ($jobs > 1 && $key =~ m/test_ssl_new|test_fuzz/) {
|
/third_party/protobuf/kokoro/linux/ |
H A D | make_test_output.py | 33 with open(basedir + "/joblog") as jobs: 34 firstline = next(jobs) 35 for line in jobs:
|
/third_party/mesa3d/src/panfrost/lib/ |
H A D | pan_indirect_draw.c | 103 struct jobs_data jobs; member 126 * the DDK seems to put all min/max search jobs at the beginning of the job chain 128 * the draw patching jobs which have the suppress_prefetch flag set. Merging the 312 builder->jobs.first_vertex_sysval = get_input_field(b, first_vertex_sysval); in extract_inputs() 313 builder->jobs.base_vertex_sysval = get_input_field(b, base_vertex_sysval); in extract_inputs() 314 builder->jobs.base_instance_sysval = get_input_field(b, base_instance_sysval); in extract_inputs() 315 builder->jobs.vertex_job = get_input_field(b, vertex_job); in extract_inputs() 316 builder->jobs.tiler_job = get_input_field(b, tiler_job); in extract_inputs() 385 builder->jobs.offset_start); in update_dcd() 396 builder->jobs in update_job() [all...] |
/third_party/mesa3d/.gitlab-ci/bin/ |
H A D | ci_run_n_monitor.py | 11 Helper script to restrict running only required CI jobs 46 # dependency graph to see which jobs the target jobs need 125 """Monitors pipeline and delegate canceling jobs""" 138 for job in pipeline.jobs.list(all=True, sort="desc"): 139 # target jobs 154 # all jobs 193 pjob = project.jobs.get(job.id, lazy=True) 204 pjob = project.jobs.get(job.id, lazy=True) 210 """Cancel unwanted GitLab jobs""" [all...] |
/third_party/node/tools/gyp/pylib/gyp/ |
H A D | xcode_ninja.py | 62 jobs = 0 68 jobs = params.get("generator_flags", {}).get("xcode_ninja_jobs", 0) 132 if jobs > 0: 133 ninja_target["actions"][0]["action"].extend(("-j", jobs))
|
/third_party/node/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/ |
H A D | xcode_ninja.py | 62 jobs = 0 68 jobs = params.get("generator_flags", {}).get("xcode_ninja_jobs", 0) 132 if jobs > 0: 133 ninja_target["actions"][0]["action"].extend(("-j", jobs))
|
/third_party/mesa3d/.gitlab-ci/tests/ |
H A D | conftest.py | 39 proxy_submit_mock = proxy_mock.scheduler.jobs.submit 48 proxy_logs_mock = proxy_mock.scheduler.jobs.logs
|
/third_party/jerryscript/tools/ |
H A D | run-tests.py | 316 def iterate_test_runner_jobs(jobs, options): 320 for job in jobs: 455 jobs = [] 457 jobs.extend(TEST262_TEST_SUITE_OPTIONS) 459 jobs.extend(TEST262_ES2015_TEST_SUITE_OPTIONS) 461 for job in jobs:
|
/third_party/mesa3d/src/gallium/drivers/v3d/ |
H A D | v3d_job.c | 26 * Functions for submitting V3D render jobs to the kernel. 48 _mesa_hash_table_remove_key(v3d->jobs, &job->key); in v3d_job_free() 154 hash_table_foreach(v3d->jobs, entry) { in v3d_flush_jobs_using_bo() 198 * graphics-compute dependency but nowadays all compute jobs in v3d_flush_jobs_writing_resource() 252 hash_table_foreach(v3d->jobs, entry) { in v3d_flush_jobs_reading_resource() 272 /* Reminder: v3d->jobs is safe to keep iterating even in v3d_flush_jobs_reading_resource() 305 struct hash_entry *entry = _mesa_hash_table_search(v3d->jobs, in v3d_get_job() 310 /* Creating a new job. Make sure that any previous jobs reading or in v3d_get_job() 365 _mesa_hash_table_insert(v3d->jobs, &job->key, job); in v3d_get_job() 392 * been bound, so set them all to ~0 when switching between jobs in v3d_get_job_for_fbo() [all...] |