/third_party/icu/icu4j/main/tests/core/src/com/ibm/icu/dev/test/format/ |
H A D | DateIntervalFormatTest.java | 2049 // These final variables are accessed directly by the concurrent threads. in TestTicket11669() 2095 List<TestThread> threads = new ArrayList<>(); in TestTicket11669() 2097 threads.add(new TestThread()); in TestTicket11669() 2099 for (Thread t: threads) { in TestTicket11669() 2102 for (TestThread t: threads) { in TestTicket11669()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/api/ |
H A D | vktApiObjectManagementTests.cpp | 95 * Thread group manages collection of threads that are expected to be 100 * an exception is safe, and other threads will continue execution. The 109 * threads. 2921 ThreadGroup threads; in multithreadedCreateSharedResourcesTest() local 2926 threads.add(MovePtr<ThreadGroupThread>(new CreateThread<Object>(env, res, params))); in multithreadedCreateSharedResourcesTest() 2928 return threads.run(); in multithreadedCreateSharedResourcesTest() 2942 ThreadGroup threads; in multithreadedCreatePerThreadResourcesTest() local 2949 threads.add(MovePtr<ThreadGroupThread>(new CreateThread<Object>(env, *resources[ndx], params))); in multithreadedCreatePerThreadResourcesTest() 2952 return threads.run(); in multithreadedCreatePerThreadResourcesTest() 3001 ThreadGroup threads; in multithreadedCreatePerThreadDeviceTest() local [all...] |
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/api/ |
H A D | vktApiObjectManagementTests.cpp | 95 * Thread group manages collection of threads that are expected to be 100 * an exception is safe, and other threads will continue execution. The 109 * threads. 2921 ThreadGroup threads; in multithreadedCreateSharedResourcesTest() local 2926 threads.add(MovePtr<ThreadGroupThread>(new CreateThread<Object>(env, res, params))); in multithreadedCreateSharedResourcesTest() 2928 return threads.run(); in multithreadedCreateSharedResourcesTest() 2942 ThreadGroup threads; in multithreadedCreatePerThreadResourcesTest() local 2949 threads.add(MovePtr<ThreadGroupThread>(new CreateThread<Object>(env, *resources[ndx], params))); in multithreadedCreatePerThreadResourcesTest() 2952 return threads.run(); in multithreadedCreatePerThreadResourcesTest() 3001 ThreadGroup threads; in multithreadedCreatePerThreadDeviceTest() local [all...] |
/third_party/skia/third_party/externals/swiftshader/tests/ReactorUnitTests/ |
H A D | ReactorUnitTests.cpp | 2625 // Test generating one coroutine, and executing it on multiple threads. This makes 2653 std::vector<std::thread> threads; 2658 threads.emplace_back([&] { 2670 for(auto &t : threads) 3387 auto threads = std::unique_ptr<std::thread[]>(new std::thread[numThreads]); 3406 threads[t] = std::thread(threadFunc, t); 3411 threads[t].join(); 3442 auto threads = std::unique_ptr<std::thread[]>(new std::thread[numThreads]); 3468 threads[t] = std::thread(threadFunc, t); 3473 threads[ [all...] |
/third_party/python/Lib/test/ |
H A D | _test_multiprocessing.py | 196 ALLOWED_TYPES = ('processes', 'manager', 'threads') 245 ALLOWED_TYPES = ('processes', 'threads') 248 if self.TYPE == 'threads': 262 if self.TYPE == 'threads': 301 if self.TYPE == "threads": 318 if cls.TYPE != 'threads': 323 if self.TYPE == "threads": 343 if self.TYPE == "threads": 392 if self.TYPE != 'threads': 409 if self.TYPE != 'threads' [all...] |
/third_party/ffmpeg/libavcodec/ |
H A D | libvpxdec.c | 92 .threads = FFMIN(avctx->thread_count ? avctx->thread_count : av_cpu_count(), 16) in vpx_init()
|
/third_party/ffmpeg/tests/fate/ |
H A D | hevc.mak | 248 fate-hevc-two-first-slice: CMD = threads=2 framemd5 -i $(TARGET_SAMPLES)/hevc/two_first_slice.mp4 -sws_flags bitexact -t 00:02.00 -an
|
/third_party/libunwind/libunwind/doc/ |
H A D | libunwind-ptrace.tex | 98 structure is never shared between threads. Because of this, no
|
/third_party/json/ |
H A D | Makefile | 221 gtar --sort=name --mtime="@$(shell git log -1 --pretty=%ct)" --owner=0 --group=0 --numeric-owner --pax-option=exthdr.name=%d/PaxHeaders/%f,delete=atime,delete=ctime --create --file - json | xz --compress -9e --threads=2 - > json.tar.xz
|
/third_party/mesa3d/src/drm-shim/ |
H A D | drm_shim.c | 48 #include <c11/threads.h>
|
/third_party/mesa3d/src/gallium/winsys/amdgpu/drm/ |
H A D | amdgpu_winsys.c | 343 util_set_thread_affinity(ws->cs_queue.threads[0], in amdgpu_pin_threads_to_L3_cache() 559 * other threads attempting to create the winsys from the same fd will in amdgpu_winsys_create()
|
/third_party/musl/libc-test/src/functionalext/supplement/thread/thread_gtest/ |
H A D | thread_rwlock_test.cpp | 10 #include <threads.h>
|
H A D | thread_mutex_test.cpp | 4 #include <threads.h>
|
/third_party/skia/modules/skottie/src/ |
H A D | SkottieTool.cpp | 46 static DEFINE_int(threads, 0, "Number of worker threads (0 -> cores count).");
|
/third_party/mesa3d/src/intel/vulkan/ |
H A D | genX_pipeline.c | 1559 * threads in play than any previous gfx9 hardware. in emit_3dstate_vs() 1626 * Hang occurs when the number of max threads is less than 2 times in emit_3dstate_hs_te_ds() 1627 * the number of instance count. The number of max threads must be in emit_3dstate_hs_te_ds() 1927 * we don't at least set the maximum number of threads. in emit_3dstate_ps() 2195 task.NumberofThreadsinGPGPUThreadGroup = task_dispatch.threads; in emit_task_state() 2266 mesh.NumberofThreadsinGPGPUThreadGroup = mesh_dispatch.threads; in emit_mesh_state() 2416 ALIGN(cs_prog_data->push.per_thread.regs * dispatch.threads + in compute_pipeline_emit() 2500 .NumberofThreadsinGPGPUThreadGroup = dispatch.threads, in compute_pipeline_emit()
|
H A D | anv_cmd_buffer.c | 749 brw_cs_push_const_total_size(cs_prog_data, dispatch.threads); in anv_cmd_buffer_cs_push_constants() 778 for (unsigned t = 0; t < dispatch.threads; t++) { in anv_cmd_buffer_cs_push_constants()
|
/third_party/mesa3d/src/broadcom/compiler/ |
H A D | v3d_compiler.h | 51 * which has 16 / threads per-thread entries, meaning that the maximum number 727 /* Minimum number of threads we are willing to use to register allocate 897 uint8_t threads; 927 uint8_t threads; 929 /* For threads > 1, whether the program should be dispatched in the
|
/third_party/icu/icu4c/source/test/intltest/ |
H A D | dtifmtts.cpp | 1785 static const DateIntervalFormat *gIntervalFormatter = NULL; // The Formatter to be used concurrently by test threads. 1833 ThreadPool<DateIntervalFormatTest> threads(this, 4, &DateIntervalFormatTest::threadFunc11669); in testTicket11669() 1834 threads.start(); in testTicket11669() 1835 threads.join(); in testTicket11669()
|
/third_party/mesa3d/src/intel/common/ |
H A D | intel_batch_decoder.c | 635 uint64_t threads = 0; in decode_mesh_task_ksp() local 645 threads = iter.raw_value; in decode_mesh_task_ksp() 654 if (threads && local_x_maximum) { in decode_mesh_task_ksp()
|
/third_party/node/tools/ |
H A D | test.py | 154 threads = [] 155 # Spawn N-1 threads and then use this thread as the last one. 160 threads.append(thread) 164 # Wait for the remaining threads 165 for thread in threads: 172 # remaining threads.
|
/third_party/skia/third_party/externals/d3d12allocator/src/ |
H A D | Tests.cpp | 1217 // Launch threads.
in TestMultithreading() 1218 std::thread threads[threadCount];
in TestMultithreading() local 1328 threads[threadIndex] = std::thread(threadFunc);
in TestMultithreading() 1331 // Wait for threads to finish.
in TestMultithreading() 1334 threads[threadIndex].join();
in TestMultithreading()
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/ray_query/ |
H A D | vktRayQueryAccelerationStructuresTests.cpp | 3840 const deUint32 threads[] = { 1, 2, 3, 4, 8, std::numeric_limits<deUint32>::max() }; in addHostThreadingOperationTests() local 3842 for (size_t threadsNdx = 0; threadsNdx < DE_LENGTH_OF_ARRAY(threads); ++threadsNdx) in addHostThreadingOperationTests() 3844 const std::string groupName = threads[threadsNdx] != std::numeric_limits<deUint32>::max() in addHostThreadingOperationTests() 3845 ? de::toString(threads[threadsNdx]) in addHostThreadingOperationTests() 3850 addOperationTestsImpl(threadGroup.get(), threads[threadsNdx]); in addHostThreadingOperationTests()
|
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/ray_query/ |
H A D | vktRayQueryAccelerationStructuresTests.cpp | 3782 const deUint32 threads[] = { 1, 2, 3, 4, 8, std::numeric_limits<deUint32>::max() }; in addHostThreadingOperationTests() local 3784 for (size_t threadsNdx = 0; threadsNdx < DE_LENGTH_OF_ARRAY(threads); ++threadsNdx) in addHostThreadingOperationTests() 3786 const std::string groupName = threads[threadsNdx] != std::numeric_limits<deUint32>::max() in addHostThreadingOperationTests() 3787 ? de::toString(threads[threadsNdx]) in addHostThreadingOperationTests() 3792 addOperationTestsImpl(threadGroup.get(), threads[threadsNdx]); in addHostThreadingOperationTests()
|
/third_party/benchmark/src/ |
H A D | benchmark.cc | 79 // on the total cpu time used by all threads that make up the test. For 81 // benchmark execution, regardless of number of threads. 180 << "thread_index must be less than threads"; in State() 390 // Count the number of benchmarks with threads to warn the user in case in RunBenchmarks() 399 benchmarks_with_threads += (benchmark.threads() > 1); in RunBenchmarks() 408 // The use of performance counters with threads would be unintuitive for in RunBenchmarks() 413 << " benchmarks with threads and " << perfcounters.num_counters() in RunBenchmarks() 416 "threads.\n"; in RunBenchmarks()
|
/third_party/alsa-utils/alsaloop/ |
H A D | alsaloop.c | 57 struct loopback_thread *threads; variable 818 thread = &threads[i]; in send_to_all() 839 thread = &threads[i]; in signal_handler_state() 914 threads = calloc(1, sizeof(struct loopback_thread) * j); in main() 915 if (threads == NULL) { in main() 919 /* sort all threads */ in main() 924 threads[k].loopbacks = malloc(l * sizeof(struct loopback *)); in main() 925 threads[k].loopbacks_count = l; in main() 926 threads[k].output = output; in main() 927 threads[ in main() [all...] |