/third_party/libdrm/intel/ |
H A D | test_decode.c | 74 size_t batch_size; in dump_batch() local 76 read_file(batch_filename, &batch_ptr, &batch_size); in dump_batch() 79 batch_size / 4); in dump_batch() 93 size_t ref_size, batch_size; in compare_batch() local 101 read_file(batch_filename, &batch_ptr, &batch_size); in compare_batch() 116 batch_size / 4); in compare_batch()
|
H A D | intel_bufmgr.h | 175 drm_intel_bufmgr *drm_intel_bufmgr_gem_init(int fd, int batch_size);
|
/third_party/benchmark/test/ |
H A D | basic_test.cc | 115 const benchmark::IterationCount batch_size = 1009; in BM_KeepRunningBatch() local 119 while (state.KeepRunningBatch(batch_size)) { in BM_KeepRunningBatch() 120 iter_count += batch_size; in BM_KeepRunningBatch() 126 assert(iter_count == batch_size // max_iterations == 1 in BM_KeepRunningBatch() 127 || iter_count > prior_iter_count); // max_iterations > batch_size in BM_KeepRunningBatch()
|
/third_party/mesa3d/src/intel/common/ |
H A D | intel_measure.c | 83 /* default batch_size allows for 64k renders in a single batch */ in intel_measure_init() 85 config.batch_size = DEFAULT_BATCH_SIZE; in intel_measure_init() 99 const char *batch_size_s = strstr(env_copy, "batch_size="); in intel_measure_init() 180 const int batch_size = atoi(batch_size_s); in intel_measure_init() local 181 if (batch_size < DEFAULT_BATCH_SIZE) { in intel_measure_init() 182 fprintf(stderr, "INTEL_MEASURE minimum batch_size is 4k: " in intel_measure_init() 183 "%d\n", batch_size); in intel_measure_init() 186 if (batch_size > DEFAULT_BATCH_SIZE * 1024) { in intel_measure_init() 187 fprintf(stderr, "INTEL_MEASURE batch_size limited to 4M: " in intel_measure_init() 188 "%d\n", batch_size); in intel_measure_init() [all...] |
H A D | intel_batch_decoder_stub.c | 51 const uint32_t *batch, uint32_t batch_size, in intel_print_batch() 50 intel_print_batch(struct intel_batch_decode_ctx *ctx, const uint32_t *batch, uint32_t batch_size, uint64_t batch_addr, bool from_ring) intel_print_batch() argument
|
H A D | intel_measure.h | 86 * INTEL_MEASURE=batch_size={num}. Additional snapshots will be dropped. 88 unsigned batch_size; member
|
H A D | intel_decoder.h | 275 const uint32_t *batch, uint32_t batch_size,
|
/third_party/skia/third_party/externals/abseil-cpp/absl/container/ |
H A D | btree_benchmark.cc | 75 const int batch_size = (kBenchmarkValues + 9) / 10; in BM_InsertImpl() local 76 while (state.KeepRunningBatch(batch_size)) { in BM_InsertImpl() 80 for (int j = i; j < i + batch_size; j++) { in BM_InsertImpl() 87 for (int j = i; j < i + batch_size; j++) { in BM_InsertImpl() 165 const int batch_size = (kBenchmarkValues + 9) / 10; in BM_Delete() local 166 while (state.KeepRunningBatch(batch_size)) { in BM_Delete() 169 for (int j = i; j < i + batch_size; j++) { in BM_Delete() 175 for (int j = i; j < i + batch_size; j++) { in BM_Delete() 192 const int batch_size = (kBenchmarkValues + 9) / 10; in BM_DeleteRange() local 193 while (state.KeepRunningBatch(batch_size)) { in BM_DeleteRange() [all...] |
/third_party/ffmpeg/libavfilter/dnn/ |
H A D | dnn_backend_openvino.c | 43 int batch_size; member 80 { "batch_size", "batch size per request", OFFSET(options.batch_size), AV_OPT_TYPE_INT, { .i64 = 1 }, 1, 1000, FLAGS}, 161 for (int i = 0; i < ctx->options.batch_size; ++i) { in fill_model_input_ov() 314 if (ctx->options.batch_size <= 0) { in init_model_ov() 315 ctx->options.batch_size = 1; in init_model_ov() 318 if (ctx->options.batch_size > 1) { in init_model_ov() 326 input_shapes.shapes[i].shape.dims[0] = ctx->options.batch_size; in init_model_ov() 416 item->lltasks = av_malloc_array(ctx->options.batch_size, sizeof(*item->lltasks)); in init_model_ov() 829 while (ff_queue_size(ov_model->lltask_queue) >= ctx->options.batch_size) { in ff_dnn_execute_model_ov() [all...] |
/third_party/node/deps/v8/src/baseline/ |
H A D | baseline-batch-compiler.cc | 96 int batch_size) { in BaselineBatchCompilerJob() 98 tasks_.reserve(batch_size); in BaselineBatchCompilerJob() 99 for (int i = 0; i < batch_size; i++) { in BaselineBatchCompilerJob() 202 void CompileBatch(Handle<WeakFixedArray> task_queue, int batch_size) { in CompileBatch() argument 206 isolate_, task_queue, batch_size)); in CompileBatch() 95 BaselineBatchCompilerJob(Isolate* isolate, Handle<WeakFixedArray> task_queue, int batch_size) BaselineBatchCompilerJob() argument
|
/third_party/mesa3d/src/gallium/drivers/iris/ |
H A D | iris_measure.c | 111 * snapshot objects based on the run-time configurable batch_size in iris_init_batch_measure() 114 config->batch_size * sizeof(struct intel_measure_snapshot); in iris_init_batch_measure() 121 config->batch_size * sizeof(uint64_t), 8, in iris_init_batch_measure() 160 if (measure_batch->index == config->batch_size) { in measure_start_snapshot() 169 "Increase setting with INTEL_MEASURE=batch_size={count}\n", in measure_start_snapshot() 170 config->batch_size); in measure_start_snapshot() 177 assert(index < config->batch_size); in measure_start_snapshot()
|
H A D | iris_batch.c | 637 unsigned batch_size = iris_batch_bytes_used(batch); in record_batch_sizes() local 639 VG(VALGRIND_CHECK_MEM_IS_DEFINED(batch->map, batch_size)); in record_batch_sizes() 642 batch->primary_batch_size = batch_size; in record_batch_sizes() 644 batch->total_chained_batch_size += batch_size; in record_batch_sizes()
|
/third_party/ltp/testcases/kernel/mem/tunable/ |
H A D | overcommit_memory.c | 237 long batch_size = MAX(ncpus * 2L, in calculate_total_batch_size() local 246 * batch_size. So the maximum error for __vm_enough_memory is in calculate_total_batch_size() 247 * batch_size * ncpus. */ in calculate_total_batch_size() 248 total_batch_size = (batch_size * ncpus * pagesize) / KB; in calculate_total_batch_size()
|
/third_party/node/deps/v8/tools/testrunner/testproc/ |
H A D | sequence_test.py | 64 def _test(self, tests, batch_size, max_heavy): 77 loader.load_initial_tests(batch_size)
|
/third_party/icu/tools/commit-checker/ |
H A D | check.py | 244 batch_size = 100 # https://jira.atlassian.com/browse/JRACLOUD-67570 246 issues = jira.search_issues(jira_query, startAt=start, maxResults=batch_size) 253 if len(issues) < batch_size: 255 start += batch_size
|
/third_party/mesa3d/src/intel/vulkan/ |
H A D | anv_measure.c | 106 * snapshot objects based on the run-time configurable batch_size in anv_measure_init() 109 config->batch_size * sizeof(struct intel_measure_snapshot); in anv_measure_init() 118 config->batch_size * sizeof(uint64_t), in anv_measure_init() 275 if (measure->base.index == config->batch_size) { in _anv_measure_snapshot() 284 "Increase setting with INTEL_MEASURE=batch_size={count}\n", in _anv_measure_snapshot() 285 config->batch_size); in _anv_measure_snapshot()
|
H A D | anv_batch_chain.c | 2430 uint32_t batch_size = align_u32(batch->next - batch->start, 8); in anv_queue_submit_simple_batch() local 2433 result = anv_bo_pool_alloc(&device->batch_bo_pool, batch_size, &batch_bo); in anv_queue_submit_simple_batch() 2437 memcpy(batch_bo->map, batch->start, batch_size); in anv_queue_submit_simple_batch() 2439 intel_flush_range(batch_bo->map, batch_size); in anv_queue_submit_simple_batch() 2461 .batch_len = batch_size, in anv_queue_submit_simple_batch()
|
/third_party/skia/third_party/externals/angle2/src/tests/restricted_traces/ |
H A D | restricted_trace_gold_tests.py | 324 def _get_batches(traces, batch_size): 325 for i in range(0, len(traces), batch_size): 326 yield traces[i:i + batch_size] 358 batches = _get_batches(traces, args.batch_size)
|
/third_party/mesa3d/src/intel/tools/ |
H A D | aubinator_viewer.h | 99 const void *batch, uint32_t batch_size,
|
H A D | aubinator_viewer_decoder.cpp | 898 const void *_batch, uint32_t batch_size, in aub_viewer_render_batch() 902 const uint32_t *p, *batch = (const uint32_t *) _batch, *end = batch + batch_size / sizeof(uint32_t); in aub_viewer_render_batch() 897 aub_viewer_render_batch(struct aub_viewer_decode_ctx *ctx, const void *_batch, uint32_t batch_size, uint64_t batch_addr, bool from_ring) aub_viewer_render_batch() argument
|
/third_party/libbpf/src/ |
H A D | bpf.h | 636 __u32 batch_size; member 638 #define bpf_test_run_opts__last_field batch_size
|
/third_party/node/deps/v8/src/wasm/ |
H A D | wasm-serialization.cc | 671 size_t batch_size = 0; in Read() local 676 batch_size += unit.code->instructions().size(); in Read() 678 if (batch_size >= batch_limit) { in Read() 681 batch_size = 0; in Read()
|
/third_party/benchmark/src/ |
H A D | sysinfo.cc | 800 static constexpr size_t batch_size = 10000; in GetCPUCyclesPerSecond() local 801 rng.discard(batch_size); in GetCPUCyclesPerSecond()
|
/third_party/rust/crates/nix/test/sys/ |
H A D | test_socket.rs | 520 let batch_size = 15; in udp_sendmmsg() 521 let mut iovs = Vec::with_capacity(1 + batch_size); in udp_sendmmsg() 522 let mut addrs = Vec::with_capacity(1 + batch_size); in udp_sendmmsg() 523 let mut data = MultiHeaders::preallocate(1 + batch_size, None); in udp_sendmmsg() 529 for _ in 0..batch_size { in udp_sendmmsg()
|
/third_party/mesa3d/src/broadcom/vulkan/ |
H A D | v3dv_meta_copy.c | 2735 uint32_t batch_size = 1; in v3dv_CmdCopyBufferToImage2KHR() local 2763 batch_size++; in v3dv_CmdCopyBufferToImage2KHR() 2767 batch_size, &info->pRegions[r], true)) { in v3dv_CmdCopyBufferToImage2KHR() 2779 batch_size = 1; in v3dv_CmdCopyBufferToImage2KHR() 2784 batch_size, &info->pRegions[r], false)) { in v3dv_CmdCopyBufferToImage2KHR() 2791 r += batch_size; in v3dv_CmdCopyBufferToImage2KHR()
|