/foundation/graphic/graphic_3d/lume/LumeRender/src/nodecontext/ |
H A D | render_node_graph_share_manager.cpp | 71 // NOTE: needs to have outputs at from which render node they are coming from 75 const array_view<const IRenderNodeGraphShareManager::NamedResource> outputs) in BeginFrame() 83 if ((outputs.size() > 0) && (rngOutputResources_.size() > 0) && (outputs.size() != rngOutputResources_.size())) { in BeginFrame() 86 "RENDER_VALIDATION: given output size missmatch with render node graph outputs."); in BeginFrame() 89 const uint32_t outputCount = static_cast<uint32_t>(Math::max(rngOutputResources_.size(), outputs.size())); in BeginFrame() 92 inOut_.outputView = { inOut_.outputs, outputCount }; in BeginFrame() 101 if (idx < outputs.size()) { in BeginFrame() 102 inOut_.outputView[idx] = outputs[idx].handle; in BeginFrame() 103 inOut_.namedOutputView[idx] = outputs[id in BeginFrame() 72 BeginFrame(RenderNodeGraphGlobalShareDataManager* rngGlobalShareDataMgr, const RenderNodeGraphShareDataManager* prevRngShareDataMgr, const uint32_t renderNodeCount, const array_view<const IRenderNodeGraphShareManager::NamedResource> inputs, const array_view<const IRenderNodeGraphShareManager::NamedResource> outputs) BeginFrame() argument [all...] |
H A D | render_node_graph_share_manager.h | 70 const BASE_NS::array_view<const IRenderNodeGraphShareManager::NamedResource> outputs); 99 // render node graph specified outputs 115 RenderHandle outputs[MAX_RENDER_NODE_GRAPH_RES_COUNT] { {}, {}, {}, {}, {}, {}, {}, {} }; 122 // render node graph inputs/outputs which can be set from the render node graph or render node graph manager 126 BASE_NS::vector<IRenderNodeGraphShareManager::NamedResource> outputs; member 131 // per render node registered inputs/outputs 160 void RegisterRenderNodeOutputs(const BASE_NS::array_view<const RenderHandle> outputs) override;
|
H A D | render_node_graph_manager.cpp | 230 // check for updated render node graph inputs / outputs 468 dstData.outputs[idx] = { "", srcData.outputs[idx].GetHandle() }; 577 const array_view<const RenderHandleReference> inputs, const array_view<const RenderHandleReference> outputs) 598 if (outputs.size() > RenderNodeGraphShareData::MAX_RENDER_NODE_GRAPH_RES_COUNT) { 600 static_cast<uint32_t>(outputs.size()), RenderNodeGraphShareData::MAX_RENDER_NODE_GRAPH_RES_COUNT); 608 RenderNodeGraphShareData::MAX_RENDER_NODE_GRAPH_RES_COUNT, static_cast<uint32_t>(outputs.size())); 618 clientData.outputs[idx] = outputs[idx]; 620 if (!outputs[id [all...] |
H A D | render_node_graph_manager.h | 77 const BASE_NS::array_view<const RenderHandleReference> outputs) override; 156 RenderHandleReference outputs[RenderNodeGraphShareData::MAX_RENDER_NODE_GRAPH_RES_COUNT] { {}, {}, {}, {} };
|
/foundation/ai/neural_network_runtime/test/unittest/common/v1_0/ |
H A D | executor_mock_device.cpp | 57 OH_NN_ReturnCode HDIPreparedModelV1_0::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 60 if (inputs.empty() || outputs.empty()) { in Run() 70 outputsDims.emplace_back(outputs[0].dimensions); in Run()
|
/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/hdi_prepared_model/ |
H A D | hdi_prepared_model_test.cpp | 92 std::vector<IOTensor> outputs; in Run() local 107 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in Run() 257 std::vector<IOTensor> outputs; in HWTEST_F() local 271 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 287 std::vector<IOTensor> outputs; in HWTEST_F() local 308 std::vector<IOTensor> outputs; in HWTEST_F() local 328 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 359 std::vector<IOTensor> outputs; in HWTEST_F() local 361 InitTensor(outputs, nullptr, 0); in HWTEST_F() 372 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDim in HWTEST_F() 391 std::vector<IOTensor> outputs; HWTEST_F() local 420 std::vector<IOTensor> outputs; RunFail() local 649 std::vector<NN_Tensor*> outputs; HWTEST_F() local 673 std::vector<NN_Tensor*> outputs; HWTEST_F() local 701 std::vector<NN_Tensor*> outputs; HWTEST_F() local 731 std::vector<NN_Tensor*> outputs; HWTEST_F() local [all...] |
/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/hdi_prepared_model/ |
H A D | hdi_prepared_model_test.cpp | 93 std::vector<IOTensor> outputs; in Run() local 109 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in Run() 260 std::vector<IOTensor> outputs; in HWTEST_F() local 274 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 290 std::vector<IOTensor> outputs; in HWTEST_F() local 311 std::vector<IOTensor> outputs; in HWTEST_F() local 332 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 363 std::vector<IOTensor> outputs; in HWTEST_F() local 365 InitTensor(outputs, nullptr, 0); in HWTEST_F() 376 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDim in HWTEST_F() 395 std::vector<IOTensor> outputs; HWTEST_F() local 425 std::vector<IOTensor> outputs; RunFail() local 654 std::vector<NN_Tensor*> outputs; HWTEST_F() local 678 std::vector<NN_Tensor*> outputs; HWTEST_F() local 706 std::vector<NN_Tensor*> outputs; HWTEST_F() local 736 std::vector<NN_Tensor*> outputs; HWTEST_F() local [all...] |
/foundation/ai/neural_network_runtime/test/unittest/components/v2_1/hdi_prepared_model/ |
H A D | hdi_prepared_model_test.cpp | 92 std::vector<IOTensor> outputs; in Run() local 107 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in Run() 258 std::vector<IOTensor> outputs; in HWTEST_F() local 272 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 288 std::vector<IOTensor> outputs; in HWTEST_F() local 309 std::vector<IOTensor> outputs; in HWTEST_F() local 329 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDims, isOutputBufferEnough); in HWTEST_F() 360 std::vector<IOTensor> outputs; in HWTEST_F() local 362 InitTensor(outputs, nullptr, 0); in HWTEST_F() 373 OH_NN_ReturnCode result = preparedModel->Run(inputs, outputs, outputsDim in HWTEST_F() 392 std::vector<IOTensor> outputs; HWTEST_F() local 419 std::vector<IOTensor> outputs; RunFail() local 648 std::vector<NN_Tensor*> outputs; HWTEST_F() local 672 std::vector<NN_Tensor*> outputs; HWTEST_F() local 700 std::vector<NN_Tensor*> outputs; HWTEST_F() local 730 std::vector<NN_Tensor*> outputs; HWTEST_F() local [all...] |
/foundation/ai/neural_network_runtime/test/unittest/common/v2_0/ |
H A D | executor_mock_device.cpp | 65 OH_NN_ReturnCode HDIPreparedModelV2_0::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 68 if (inputs.empty() || outputs.empty()) { in Run() 78 outputsDims.emplace_back(outputs[0].dimensions); in Run()
|
/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/ |
H A D | prepared_model.h | 34 const std::vector<IOTensor>& outputs, 39 const std::vector<NN_Tensor*>& outputs,
|
H A D | hdi_prepared_model_v1_0.h | 41 const std::vector<IOTensor>& outputs, 46 const std::vector<NN_Tensor*>& outputs,
|
H A D | hdi_prepared_model_v2_0.h | 42 const std::vector<IOTensor>& outputs, 47 const std::vector<NN_Tensor*>& outputs,
|
H A D | hdi_prepared_model_v2_1.h | 42 const std::vector<IOTensor>& outputs, 47 const std::vector<NN_Tensor*>& outputs,
|
H A D | hdi_prepared_model_v2_1.cpp | 209 OH_NN_ReturnCode HDIPreparedModelV2_1::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 224 for (const auto& output: outputs) { in Run() 227 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 246 const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, in Run() 265 for (const auto& output: outputs) { in Run() 272 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 245 Run(const std::vector<NN_Tensor*>& inputs, const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, std::vector<bool>& isOutputBufferEnough) Run() argument
|
H A D | hdi_prepared_model_v2_0.cpp | 209 OH_NN_ReturnCode HDIPreparedModelV2_0::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 224 for (const auto& output: outputs) { in Run() 227 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 246 const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, in Run() 265 for (const auto& output: outputs) { in Run() 272 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 245 Run(const std::vector<NN_Tensor*>& inputs, const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, std::vector<bool>& isOutputBufferEnough) Run() argument
|
H A D | hdi_prepared_model_v1_0.cpp | 209 OH_NN_ReturnCode HDIPreparedModelV1_0::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 224 for (const auto& output: outputs) { in Run() 227 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 243 const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, in Run() 262 for (const auto& output: outputs) { in Run() 269 LOGE("Transform outputs tensor failed, cannot find data file descriptor."); in Run() 242 Run(const std::vector<NN_Tensor*>& inputs, const std::vector<NN_Tensor*>& outputs, std::vector<std::vector<int32_t>>& outputsDims, std::vector<bool>& isOutputBufferEnough) Run() argument
|
/foundation/ai/neural_network_runtime/test/system_test/common/ |
H A D | nnrt_test.cpp | 104 const OH_NN_UInt32Array outputs = TransformUInt32Array(outputIndices); in AddOperation() local 106 OH_NN_ReturnCode status = OH_NNModel_AddOperation(m_model, opType, ¶ms, &inputs, &outputs); in AddOperation() 111 .outputs = outputIndices, in AddOperation() 124 const OH_NN_UInt32Array outputs = TransformUInt32Array(outputIndices); in SpecifyInputAndOutput() local 126 OH_NN_ReturnCode status = OH_NNModel_SpecifyInputsAndOutputs(m_model, &inputs, &outputs); in SpecifyInputAndOutput()
|
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/src/ |
H A D | prepared_model_service.cpp | 96 int32_t PreparedModelService::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 106 ret = SetOutputs(outputs); in Run() 122 ret = UpdateOutput(outputs, outputsDims, isOutputBufferEnough); in Run() 172 NNRT_ReturnCode PreparedModelService::UpdateOutput(const std::vector<IOTensor>& outputs, in UpdateOutput() argument 179 auto& output = outputs[i]; in UpdateOutput() 258 HDF_LOGE("Model without inputs or outputs is invalid."); in Compile() 285 HDF_LOGE("Model without inputs or outputs is invalid."); in Compile() 338 HDF_LOGE("Get ms inputs or outputs failed after resize."); in SetInputs() 359 NNRT_ReturnCode PreparedModelService::SetOutputs(const std::vector<IOTensor>& outputs) in SetOutputs() argument 361 HDF_LOGI("Start Set outputs, m_output in SetOutputs() [all...] |
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/ |
H A D | prepared_model_service.cpp | 93 int32_t PreparedModelService::Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, in Run() argument 103 ret = SetOutputs(outputs); in Run() 118 ret = UpdateOutput(outputs, outputsDims, isOutputBufferEnough); in Run() 130 int32_t PreparedModelService::UpdateOutput(const std::vector<IOTensor>& outputs, in UpdateOutput() argument 138 auto& output = outputs[i]; in UpdateOutput() 218 HDF_LOGE("Model without inputs or outputs is invalid."); in Compile() 240 HDF_LOGE("Model without inputs or outputs is invalid."); in Compile() 288 HDF_LOGE("Get ms inputs or outputs failed after resize."); in SetInputs() 309 int32_t PreparedModelService::SetOutputs(const std::vector<IOTensor>& outputs) in SetOutputs() argument 311 HDF_LOGI("Start Set outputs, m_output in SetOutputs() [all...] |
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/include/ |
H A D | prepared_model_service.h | 46 int32_t Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, 51 int32_t SetOutputs(const std::vector<IOTensor>& outputs); 55 int32_t UpdateOutput(const std::vector<IOTensor>& outputs,
|
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/include/ |
H A D | prepared_model_service.h | 46 int32_t Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs, 54 NNRT_ReturnCode SetOutputs(const std::vector<IOTensor>& outputs); 58 NNRT_ReturnCode UpdateOutput(const std::vector<IOTensor>& outputs,
|
/foundation/ai/neural_network_runtime/test/fuzztest/ |
H A D | nnrt_prepare_model_impl.h | 44 int32_t Run(const std::vector<IOTensor>& inputs, const std::vector<IOTensor>& outputs,
in Run() argument
|
/foundation/ai/neural_network_runtime/example/deep_learning_framework/tflite/label_classify/ |
H A D | label_classify.cpp | 119 const std::vector<int32_t> outputs = interpreter->outputs(); in PrepareModel() local 123 LOG(INFO) << "number of outputs: " << outputs.size(); in PrepareModel()
|
/foundation/ai/neural_network_runtime/example/deep_learning_framework/tflite/tools/ |
H A D | utils.cpp | 133 for (int32_t index = 0; index < interpreter->outputs().size(); ++index) { in PrintResult() 134 int32_t output_index = interpreter->outputs()[index]; in PrintResult() 173 int32_t output = interpreter->outputs()[0]; in AnalysisResults()
|
/foundation/graphic/graphic_3d/lume/Lume_3D/src/render/node/ |
H A D | render_node_morph.cpp | 97 const DescriptorCounts dc { { // weight/indexset for all prims + number of inputs and outputs in PreExecuteFrame() 115 allDescriptorSets_.outputs.resize(maxObjectCount_); in PreExecuteFrame() 119 allDescriptorSets_.outputs[idx] = descriptorSetMgr.CreateDescriptorSetBinder( in PreExecuteFrame() 231 // Bind outputs = pos/nor/tangent buffers (set 2) in ComputeMorphs() 232 auto& outputBinder = *allDescriptorSets_.outputs[outputIdx++]; in ComputeMorphs()
|