/foundation/ai/neural_network_runtime/frameworks/native/neural_network_runtime/ |
H A D | hdi_device_v1_0.h | 47 OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 50 OH_NN_ReturnCode PrepareModel(const void* metaGraph,
|
H A D | device.h | 51 virtual OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 54 virtual OH_NN_ReturnCode PrepareModel(const void* metaGraph,
|
H A D | hdi_device_v2_0.h | 47 OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 50 OH_NN_ReturnCode PrepareModel(const void* metaGraph,
|
H A D | hdi_device_v2_1.h | 47 OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 50 OH_NN_ReturnCode PrepareModel(const void* metaGraph,
|
H A D | hdi_device_v1_0.cpp | 240 OH_NN_ReturnCode HDIDeviceV1_0::PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV1_0 272 auto preparedRet = m_iDevice->PrepareModel(*iModel, iModelConfig, iPreparedModel); in PrepareModel() 294 OH_NN_ReturnCode HDIDeviceV1_0::PrepareModel(const void* metaGraph, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV1_0
|
H A D | hdi_device_v2_1.cpp | 279 OH_NN_ReturnCode HDIDeviceV2_1::PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV2_1 310 ret = m_iDevice->PrepareModel(*iModel, iModelConfig, iPreparedModel); in PrepareModel() 331 OH_NN_ReturnCode HDIDeviceV2_1::PrepareModel(const void* metaGraph, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV2_1
|
H A D | hdi_device_v2_0.cpp | 284 OH_NN_ReturnCode HDIDeviceV2_0::PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV2_0 315 ret = m_iDevice->PrepareModel(*iModel, iModelConfig, iPreparedModel); in PrepareModel() 336 OH_NN_ReturnCode HDIDeviceV2_0::PrepareModel(const void* metaGraph, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV2_0
|
H A D | nncompiler.cpp | 383 ret = m_device->PrepareModel(m_liteGraph, config, m_preparedModel); in NormalBuild() 386 ret = m_device->PrepareModel(m_metaGraph, config, m_preparedModel); in NormalBuild()
|
/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/device_registrar/ |
H A D | device_registrar_test.cpp | 62 virtual int32_t PrepareModel(const V2_0::Model& model, const V2_0::ModelConfig& config, 137 OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 174 MOCK_METHOD3(PrepareModel,
|
/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/device_registrar/ |
H A D | device_registrar_test.cpp | 62 virtual int32_t PrepareModel(const V1_0::Model& model, const V1_0::ModelConfig& config, 137 OH_NN_ReturnCode PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, 174 MOCK_METHOD3(PrepareModel,
|
/foundation/ai/neural_network_runtime/test/unittest/common/v2_0/ |
H A D | mock_idevice.h | 40 MOCK_METHOD3(PrepareModel, int32_t(const Model&, const ModelConfig&, OHOS::sptr<IPreparedModel>&));
|
H A D | compilation_mock_idevice.cpp | 141 OH_NN_ReturnCode HDIDeviceV2_0::PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV2_0 145 LOGE("HDIDeviceV2_0 mock PrepareModel failed, the model is nullptr"); in PrepareModel() 150 LOGE("HDIDeviceV2_0 mock PrepareModel failed, the enableFloat16 is false"); in PrepareModel() 157 LOGE("HDIDeviceV2_0 mock PrepareModel failed, error happened when new sptr"); in PrepareModel() 234 LOGE("HDIDeviceV2_0 mock PrepareModel failed, the modelCache size equals 0 or enableFloat16 is false"); in PrepareModelFromModelCache()
|
/foundation/ai/neural_network_runtime/test/unittest/common/v1_0/ |
H A D | mock_idevice.h | 40 MOCK_METHOD3(PrepareModel, int32_t(const Model&, const ModelConfig&, OHOS::sptr<IPreparedModel>&));
|
H A D | compilation_mock_idevice.cpp | 141 OH_NN_ReturnCode HDIDeviceV1_0::PrepareModel(std::shared_ptr<const mindspore::lite::LiteGraph> model, in PrepareModel() function in OHOS::NeuralNetworkRuntime::HDIDeviceV1_0 145 LOGE("HDIDeviceV1_0 mock PrepareModel failed, the model is nullptr"); in PrepareModel() 150 LOGE("HDIDeviceV1_0 mock PrepareModel failed, the enableFloat16 is false"); in PrepareModel() 157 LOGE("HDIDeviceV1_0 mock PrepareModel failed, error happened when new sptr"); in PrepareModel() 234 LOGE("HDIDeviceV1_0 mock PrepareModel failed, the modelCache size equals 0 or enableFloat16 is false"); in PrepareModelFromModelCache()
|
/foundation/ai/neural_network_runtime/test/fuzztest/ |
H A D | nnrt_device_impl.h | 84 int32_t PrepareModel(const Model& model, const ModelConfig& config, sptr<IPreparedModel>& preparedModel)
in PrepareModel() function in OHOS::HDI::Nnrt::V2_0::NnrtDeviceImpl
|
/foundation/ai/neural_network_runtime/test/unittest/common/v2_1/ |
H A D | mock_idevice.h | 40 MOCK_METHOD3(PrepareModel, int32_t(const Model&, const ModelConfig&, OHOS::sptr<IPreparedModel>&));
|
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/include/ |
H A D | nnrt_device_service.h | 54 int32_t PrepareModel(const Model& model, const ModelConfig& config, sptr<IPreparedModel>& preparedModel) override;
|
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v2_0/hdi_cpu_service/include/ |
H A D | nnrt_device_service.h | 54 int32_t PrepareModel(const Model& model, const ModelConfig& config, sptr<IPreparedModel>& preparedModel) override;
|
/foundation/ai/neural_network_runtime/example/deep_learning_framework/tflite/label_classify/ |
H A D | label_classify.cpp | 115 void PrepareModel(Settings& settings, std::unique_ptr<tflite::Interpreter>& interpreter, in PrepareModel() function 221 PrepareModel(settings, interpreter, delegateProviders); in InferenceModel()
|
/foundation/ai/neural_network_runtime/test/unittest/components/v1_0/hdi_device/ |
H A D | hdi_device_test.cpp | 65 OH_NN_ReturnCode PrepareModel(int32_t allocBufferType, int32_t prepareType); 94 OH_NN_ReturnCode HDIDeviceTest::PrepareModel(int32_t allocBufferType, int32_t prepareType) in PrepareModel() function in OHOS::NeuralNetworkRuntime::UnitTest::HDIDeviceTest 111 EXPECT_CALL(*sp, PrepareModel(::testing::_, ::testing::_, ::testing::_)) in PrepareModel() 116 OH_NN_ReturnCode result = hdiDevice->PrepareModel(model, config, preparedModel); in PrepareModel() 760 * @tc.desc: Verify the PrepareModel function return success. 767 OH_NN_ReturnCode result = PrepareModel(allocBufferType, prepareType); in HWTEST_F() 773 * @tc.desc: Verify the PrepareModel function return invalid parameter. 785 OH_NN_ReturnCode result = hdiDevice->PrepareModel(model, config, preparedModel); in HWTEST_F() 791 * @tc.desc: Verify the PrepareModel function return failed. 798 OH_NN_ReturnCode result = PrepareModel(allocBufferTyp in HWTEST_F() [all...] |
/foundation/ai/neural_network_runtime/test/unittest/components/nn_compiler/ |
H A D | nn_compiler_test.cpp | 54 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(std::shared_ptr<const mindspore::lite::LiteGraph>, 57 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(const void*, 995 PrepareModel(testing::A<std::shared_ptr<const mindspore::lite::LiteGraph>>(), ::testing::_, ::testing::_)) in HWTEST_F() 1171 PrepareModel(testing::A<std::shared_ptr<const mindspore::lite::LiteGraph>>(), ::testing::_, ::testing::_)) in HWTEST_F()
|
/foundation/ai/neural_network_runtime/test/unittest/components/nn_compiled_cache/ |
H A D | nn_compiled_cache_test.cpp | 53 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(std::shared_ptr<const mindspore::lite::LiteGraph>, 56 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(const void*,
|
/foundation/ai/neural_network_runtime/test/unittest/components/v2_0/hdi_device/ |
H A D | hdi_device_test.cpp | 65 MOCK_METHOD3(PrepareModel, int32_t(const Model&, const ModelConfig&, OHOS::sptr<IPreparedModel>&)); 153 OH_NN_ReturnCode PrepareModel(int32_t allocBufferType, int32_t prepareType); 183 OH_NN_ReturnCode HDIDeviceTest::PrepareModel(int32_t allocBufferType, int32_t prepareType) in PrepareModel() function in OHOS::NeuralNetworkRuntime::UnitTest::HDIDeviceTest 200 EXPECT_CALL(*sp, PrepareModel(::testing::_, ::testing::_, ::testing::_)) in PrepareModel() 205 OH_NN_ReturnCode result = hdiDevice->PrepareModel(model, config, preparedModel); in PrepareModel() 226 EXPECT_CALL(*sp, PrepareModel(::testing::_, ::testing::_, ::testing::_)) in PrepareModelv2() 231 OH_NN_ReturnCode result = hdiDevice->PrepareModel(model, config, preparedModel); in PrepareModelv2() 902 * @tc.desc: Verify the PrepareModel function return success. 909 OH_NN_ReturnCode result = PrepareModel(allocBufferType, prepareType); in HWTEST_F() 915 * @tc.desc: Verify the PrepareModel functio [all...] |
/foundation/ai/neural_network_runtime/test/unittest/components/nn_backend/ |
H A D | nn_backend_test.cpp | 51 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(std::shared_ptr<const mindspore::lite::LiteGraph>, 54 MOCK_METHOD3(PrepareModel, OH_NN_ReturnCode(const void*,
|
/foundation/ai/neural_network_runtime/example/drivers/nnrt/v1_0/hdi_cpu_service/src/ |
H A D | nnrt_device_service.cpp | 105 int32_t NnrtDeviceService::PrepareModel(const Model& model, const ModelConfig& config, in PrepareModel() function in OHOS::HDI::Nnrt::V1_0::NnrtDeviceService
|