| // |
| // Copyright © 2020-2023 Arm Ltd and Contributors. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| |
| #include <DelegateOptions.hpp> |
| #include <armnn/utility/NumericCast.hpp> |
| #include <armnn/utility/StringUtils.hpp> |
| |
| namespace armnnDelegate |
| { |
| |
| struct DelegateOptionsImpl |
| { |
| ~DelegateOptionsImpl() = default; |
| DelegateOptionsImpl() = default; |
| |
| explicit DelegateOptionsImpl(armnn::Compute computeDevice, |
| const std::vector<armnn::BackendOptions>& backendOptions, |
| const armnn::Optional<armnn::LogSeverity> logSeverityLevel) |
| : p_Backends({computeDevice}), p_RuntimeOptions(), m_LoggingSeverity(logSeverityLevel) |
| { |
| p_RuntimeOptions.m_BackendOptions = backendOptions; |
| } |
| |
| explicit DelegateOptionsImpl(const std::vector<armnn::BackendId>& backends, |
| const std::vector<armnn::BackendOptions>& backendOptions, |
| const armnn::Optional<armnn::LogSeverity> logSeverityLevel) |
| : p_Backends(backends), p_RuntimeOptions(), m_LoggingSeverity(logSeverityLevel) |
| { |
| p_RuntimeOptions.m_BackendOptions = backendOptions; |
| } |
| |
| explicit DelegateOptionsImpl(armnn::Compute computeDevice, |
| const armnn::OptimizerOptions& optimizerOptions, |
| const armnn::Optional<armnn::LogSeverity>& logSeverityLevel, |
| const armnn::Optional<armnn::DebugCallbackFunction>& func) |
| : p_Backends({computeDevice}), |
| p_RuntimeOptions(), |
| p_OptimizerOptions(optimizerOptions), |
| m_LoggingSeverity(logSeverityLevel), |
| p_DebugCallbackFunc(func) |
| { |
| } |
| |
| explicit DelegateOptionsImpl(const std::vector<armnn::BackendId>& backends, |
| const armnn::OptimizerOptions& optimizerOptions, |
| const armnn::Optional<armnn::LogSeverity>& logSeverityLevel, |
| const armnn::Optional<armnn::DebugCallbackFunction>& func) |
| : p_Backends(backends), |
| p_RuntimeOptions(), |
| p_OptimizerOptions(optimizerOptions), |
| m_LoggingSeverity(logSeverityLevel), |
| p_DebugCallbackFunc(func) |
| { |
| } |
| |
| /// Which backend to run Delegate on. |
| /// Examples of possible values are: CpuRef, CpuAcc, GpuAcc. |
| /// CpuRef as default. |
| std::vector<armnn::BackendId> p_Backends = {armnn::Compute::CpuRef }; |
| |
| /// Creation options for the ArmNN runtime |
| /// Contains options for global settings that are valid for the whole lifetime of ArmNN |
| /// i.e. BackendOptions, DynamicBackendPath, ExternalProfilingOptions and more |
| armnn::IRuntime::CreationOptions p_RuntimeOptions; |
| |
| /// Options for the optimization step for the network |
| armnn::OptimizerOptions p_OptimizerOptions; |
| |
| /// Internal profiling options. Written to INetworkProperties during model load. |
| /// Indicates whether internal profiling is enabled or not. |
| bool m_InternalProfilingEnabled = false; |
| |
| /// Sets the level of detail output by the profiling. Options are DetailsWithEvents = 1 and DetailsOnly = 2 |
| armnn::ProfilingDetailsMethod p_InternalProfilingDetail = armnn::ProfilingDetailsMethod::DetailsWithEvents; |
| |
| /// Severity level for logging within ArmNN that will be used on creation of the delegate |
| armnn::Optional<armnn::LogSeverity> m_LoggingSeverity; |
| |
| /// A callback function to debug layers performing custom computations on intermediate tensors. |
| /// If a function is not registered, and debug is enabled in OptimizerOptions, |
| /// debug will print information of the intermediate tensors. |
| armnn::Optional<armnn::DebugCallbackFunction> p_DebugCallbackFunc; |
| |
| /// If not empty then the optimized model will be serialized to a file with this file name in "dot" format. |
| std::string m_SerializeToDot = ""; |
| |
| /// Option to disable TfLite Runtime fallback for unsupported operators. |
| bool m_DisableTfLiteRuntimeFallback = false; |
| |
| }; |
| |
| DelegateOptions::~DelegateOptions() = default; |
| |
| DelegateOptions::DelegateOptions() |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>()) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(DelegateOptions const &other) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>(*other.p_DelegateOptionsImpl)) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(armnn::Compute computeDevice, |
| const std::vector<armnn::BackendOptions>& backendOptions, |
| const armnn::Optional<armnn::LogSeverity> logSeverityLevel) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>(computeDevice, backendOptions, logSeverityLevel)) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(const std::vector<armnn::BackendId>& backends, |
| const std::vector<armnn::BackendOptions>& backendOptions, |
| const armnn::Optional<armnn::LogSeverity> logSeverityLevel) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>(backends, backendOptions, logSeverityLevel)) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(armnn::Compute computeDevice, |
| const armnn::OptimizerOptions& optimizerOptions, |
| const armnn::Optional<armnn::LogSeverity>& logSeverityLevel, |
| const armnn::Optional<armnn::DebugCallbackFunction>& func) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>(computeDevice, optimizerOptions, |
| logSeverityLevel, func)) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(const std::vector<armnn::BackendId>& backends, |
| const armnn::OptimizerOptions& optimizerOptions, |
| const armnn::Optional<armnn::LogSeverity>& logSeverityLevel, |
| const armnn::Optional<armnn::DebugCallbackFunction>& func) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>(backends, optimizerOptions, |
| logSeverityLevel, func)) |
| { |
| } |
| |
| DelegateOptions::DelegateOptions(char const* const* options_keys, |
| char const* const* options_values, |
| size_t num_options, |
| void (*report_error)(const char*)) |
| : p_DelegateOptionsImpl(std::make_unique<DelegateOptionsImpl>()) |
| { |
| armnn::IRuntime::CreationOptions runtimeOptions; |
| armnn::OptimizerOptions optimizerOptions; |
| bool internalProfilingState = false; |
| armnn::ProfilingDetailsMethod internalProfilingDetail = armnn::ProfilingDetailsMethod::DetailsWithEvents; |
| for (size_t i = 0; i < num_options; ++i) |
| { |
| // Process backends |
| if (std::string(options_keys[i]) == std::string("backends")) |
| { |
| // The backend option is a comma separated string of backendIDs that needs to be split |
| std::vector<armnn::BackendId> backends; |
| char* dup = strdup(options_values[i]); |
| char* pch = std::strtok(dup, ","); |
| while (pch != NULL) |
| { |
| backends.push_back(pch); |
| pch = strtok (NULL, ","); |
| } |
| SetBackends(backends); |
| } |
| // Process dynamic-backends-path |
| else if (std::string(options_keys[i]) == std::string("dynamic-backends-path")) |
| { |
| runtimeOptions.m_DynamicBackendsPath = std::string(options_values[i]); |
| } |
| // Process logging level |
| else if (std::string(options_keys[i]) == std::string("logging-severity")) |
| { |
| SetLoggingSeverity(options_values[i]); |
| } |
| // Process GPU backend options |
| else if (std::string(options_keys[i]) == std::string("gpu-tuning-level")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"TuningLevel", |
| atoi(options_values[i])}}); |
| runtimeOptions.m_BackendOptions.push_back(option); |
| } |
| else if (std::string(options_keys[i]) == std::string("gpu-mlgo-tuning-file")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"MLGOTuningFilePath", |
| std::string(options_values[i])}}); |
| optimizerOptions.m_ModelOptions.push_back(option); |
| } |
| else if (std::string(options_keys[i]) == std::string("gpu-tuning-file")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"TuningFile", |
| std::string(options_values[i])}}); |
| runtimeOptions.m_BackendOptions.push_back(option); |
| } |
| else if (std::string(options_keys[i]) == std::string("gpu-enable-profiling")) |
| { |
| runtimeOptions.m_EnableGpuProfiling = (*options_values[i] != '0'); |
| } |
| else if (std::string(options_keys[i]) == std::string("gpu-kernel-profiling-enabled")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"KernelProfilingEnabled", |
| armnn::stringUtils::StringToBool(options_values[i])}}); |
| runtimeOptions.m_BackendOptions.push_back(option); |
| } |
| else if (std::string(options_keys[i]) == std::string("save-cached-network")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"SaveCachedNetwork", |
| armnn::stringUtils::StringToBool(options_values[i])}}); |
| optimizerOptions.m_ModelOptions.push_back(option); |
| } |
| else if (std::string(options_keys[i]) == std::string("cached-network-filepath")) |
| { |
| armnn::BackendOptions option("GpuAcc", {{"CachedNetworkFilePath", |
| std::string(options_values[i])}}); |
| optimizerOptions.m_ModelOptions.push_back(option); |
| } |
| // Process GPU & CPU backend options |
| else if (std::string(options_keys[i]) == std::string("enable-fast-math")) |
| { |
| armnn::BackendOptions modelOptionGpu("GpuAcc", {{"FastMathEnabled", |
| armnn::stringUtils::StringToBool(options_values[i])}}); |
| optimizerOptions.m_ModelOptions.push_back(modelOptionGpu); |
| |
| armnn::BackendOptions modelOptionCpu("CpuAcc", {{"FastMathEnabled", |
| armnn::stringUtils::StringToBool(options_values[i])}}); |
| optimizerOptions.m_ModelOptions.push_back(modelOptionCpu); |
| } |
| // Process CPU backend options |
| else if (std::string(options_keys[i]) == std::string("number-of-threads")) |
| { |
| unsigned int numberOfThreads = armnn::numeric_cast<unsigned int>(atoi(options_values[i])); |
| armnn::BackendOptions modelOption("CpuAcc", |
| {{"NumberOfThreads", numberOfThreads}}); |
| optimizerOptions.m_ModelOptions.push_back(modelOption); |
| } |
| // Process reduce-fp32-to-fp16 option |
| else if (std::string(options_keys[i]) == std::string("reduce-fp32-to-fp16")) |
| { |
| optimizerOptions.m_ReduceFp32ToFp16 = armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Process debug-data |
| else if (std::string(options_keys[i]) == std::string("debug-data")) |
| { |
| optimizerOptions.m_Debug = armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Infer output-shape |
| else if (std::string(options_keys[i]) == std::string("infer-output-shape")) |
| { |
| armnn::BackendOptions backendOption("ShapeInferenceMethod", |
| { |
| { "InferAndValidate", armnn::stringUtils::StringToBool(options_values[i]) } |
| }); |
| optimizerOptions.m_ModelOptions.push_back(backendOption); |
| } |
| // Allow expanded dims |
| else if (std::string(options_keys[i]) == std::string("allow-expanded-dims")) |
| { |
| armnn::BackendOptions backendOption("AllowExpandedDims", |
| { |
| { "AllowExpandedDims", armnn::stringUtils::StringToBool(options_values[i]) } |
| }); |
| optimizerOptions.m_ModelOptions.push_back(backendOption); |
| } |
| // Process memory-import |
| else if (std::string(options_keys[i]) == std::string("memory-import")) |
| { |
| optimizerOptions.m_ImportEnabled = armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Process enable-internal-profiling |
| else if (std::string(options_keys[i]) == std::string("enable-internal-profiling")) |
| { |
| internalProfilingState = *options_values[i] != '0'; |
| optimizerOptions.m_ProfilingEnabled = internalProfilingState; |
| } |
| // Process internal-profiling-detail |
| else if (std::string(options_keys[i]) == std::string("internal-profiling-detail")) |
| { |
| uint32_t detailLevel = static_cast<uint32_t>(std::stoul(options_values[i])); |
| switch (detailLevel) |
| { |
| case 1: |
| internalProfilingDetail = armnn::ProfilingDetailsMethod::DetailsWithEvents; |
| break; |
| case 2: |
| internalProfilingDetail = armnn::ProfilingDetailsMethod::DetailsOnly; |
| break; |
| default: |
| internalProfilingDetail = armnn::ProfilingDetailsMethod::Undefined; |
| break; |
| } |
| } |
| // Process enable-external-profiling |
| else if (std::string(options_keys[i]) == std::string("enable-external-profiling")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_EnableProfiling = |
| armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Process timeline-profiling |
| else if (std::string(options_keys[i]) == std::string("timeline-profiling")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_TimelineEnabled = |
| armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Process outgoing-capture-file |
| else if (std::string(options_keys[i]) == std::string("outgoing-capture-file")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_OutgoingCaptureFile = options_values[i]; |
| } |
| // Process incoming-capture-file |
| else if (std::string(options_keys[i]) == std::string("incoming-capture-file")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_IncomingCaptureFile = options_values[i]; |
| } |
| // Process file-only-external-profiling |
| else if (std::string(options_keys[i]) == std::string("file-only-external-profiling")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_FileOnly = armnn::stringUtils::StringToBool(options_values[i]); |
| } |
| // Process counter-capture-period |
| else if (std::string(options_keys[i]) == std::string("counter-capture-period")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_CapturePeriod = |
| static_cast<uint32_t>(std::stoul(options_values[i])); |
| } |
| // Process profiling-file-format |
| else if (std::string(options_keys[i]) == std::string("profiling-file-format")) |
| { |
| runtimeOptions.m_ProfilingOptions.m_FileFormat = options_values[i]; |
| } |
| // Process serialize-to-dot |
| else if (std::string(options_keys[i]) == std::string("serialize-to-dot")) |
| { |
| SetSerializeToDot(options_values[i]); |
| } |
| |
| // Process disable-tflite-runtime-fallback |
| else if (std::string(options_keys[i]) == std::string("disable-tflite-runtime-fallback")) |
| { |
| this->DisableTfLiteRuntimeFallback(armnn::stringUtils::StringToBool(options_values[i])); |
| } |
| else |
| { |
| throw armnn::Exception("Unknown option for the ArmNN Delegate given: " + |
| std::string(options_keys[i])); |
| } |
| } |
| |
| SetRuntimeOptions(runtimeOptions); |
| SetOptimizerOptions(optimizerOptions); |
| SetInternalProfilingParams(internalProfilingState, internalProfilingDetail); |
| } |
| |
| const std::vector<armnn::BackendId>& DelegateOptions::GetBackends() const |
| { |
| return p_DelegateOptionsImpl->p_Backends; |
| } |
| |
| void DelegateOptions::SetBackends(const std::vector<armnn::BackendId>& backends) |
| { |
| p_DelegateOptionsImpl->p_Backends = backends; |
| } |
| |
| void DelegateOptions::SetDynamicBackendsPath(const std::string& dynamicBackendsPath) |
| { |
| p_DelegateOptionsImpl->p_RuntimeOptions.m_DynamicBackendsPath = dynamicBackendsPath; |
| } |
| |
| const std::string& DelegateOptions::GetDynamicBackendsPath() const |
| { |
| return p_DelegateOptionsImpl->p_RuntimeOptions.m_DynamicBackendsPath; |
| } |
| |
| void DelegateOptions::SetGpuProfilingState(bool gpuProfilingState) |
| { |
| p_DelegateOptionsImpl->p_RuntimeOptions.m_EnableGpuProfiling = gpuProfilingState; |
| } |
| |
| bool DelegateOptions::GetGpuProfilingState() |
| { |
| return p_DelegateOptionsImpl->p_RuntimeOptions.m_EnableGpuProfiling; |
| } |
| |
| const std::vector<armnn::BackendOptions>& DelegateOptions::GetBackendOptions() const |
| { |
| return p_DelegateOptionsImpl->p_RuntimeOptions.m_BackendOptions; |
| } |
| |
| void DelegateOptions::AddBackendOption(const armnn::BackendOptions& option) |
| { |
| p_DelegateOptionsImpl->p_RuntimeOptions.m_BackendOptions.push_back(option); |
| } |
| |
| void DelegateOptions::SetLoggingSeverity(const armnn::LogSeverity& level) |
| { |
| p_DelegateOptionsImpl->m_LoggingSeverity = level; |
| } |
| |
| void DelegateOptions::SetLoggingSeverity(const std::string& level) |
| { |
| p_DelegateOptionsImpl->m_LoggingSeverity = armnn::StringToLogLevel(level); |
| } |
| |
| armnn::LogSeverity DelegateOptions::GetLoggingSeverity() |
| { |
| return p_DelegateOptionsImpl->m_LoggingSeverity.value(); |
| } |
| |
| bool DelegateOptions::IsLoggingEnabled() |
| { |
| return p_DelegateOptionsImpl->m_LoggingSeverity.has_value(); |
| } |
| |
| const armnn::OptimizerOptions& DelegateOptions::GetOptimizerOptions() const |
| { |
| return p_DelegateOptionsImpl->p_OptimizerOptions; |
| } |
| |
| void DelegateOptions::SetOptimizerOptions(const armnn::OptimizerOptions& optimizerOptions) |
| { |
| p_DelegateOptionsImpl->p_OptimizerOptions = optimizerOptions; |
| } |
| |
| const armnn::Optional<armnn::DebugCallbackFunction>& DelegateOptions::GetDebugCallbackFunction() const |
| { |
| return p_DelegateOptionsImpl->p_DebugCallbackFunc; |
| } |
| |
| void DelegateOptions::SetInternalProfilingParams(bool internalProfilingState, |
| const armnn::ProfilingDetailsMethod& internalProfilingDetail) |
| { |
| p_DelegateOptionsImpl->m_InternalProfilingEnabled = internalProfilingState; |
| p_DelegateOptionsImpl->p_InternalProfilingDetail = internalProfilingDetail; |
| } |
| |
| bool DelegateOptions::GetInternalProfilingState() const |
| { |
| return p_DelegateOptionsImpl->m_InternalProfilingEnabled; |
| } |
| |
| const armnn::ProfilingDetailsMethod& DelegateOptions::GetInternalProfilingDetail() const |
| { |
| return p_DelegateOptionsImpl->p_InternalProfilingDetail; |
| } |
| |
| void DelegateOptions::SetSerializeToDot(const std::string& serializeToDotFile) |
| { |
| p_DelegateOptionsImpl->m_SerializeToDot = serializeToDotFile; |
| } |
| |
| const std::string& DelegateOptions::GetSerializeToDot() const |
| { |
| return p_DelegateOptionsImpl->m_SerializeToDot; |
| } |
| |
| void DelegateOptions::SetRuntimeOptions(const armnn::IRuntime::CreationOptions& runtimeOptions) |
| { |
| p_DelegateOptionsImpl->p_RuntimeOptions = runtimeOptions; |
| } |
| |
| const armnn::IRuntime::CreationOptions& DelegateOptions::GetRuntimeOptions() |
| { |
| return p_DelegateOptionsImpl->p_RuntimeOptions; |
| } |
| |
| void DelegateOptions::DisableTfLiteRuntimeFallback(bool fallbackState) |
| { |
| p_DelegateOptionsImpl->m_DisableTfLiteRuntimeFallback = fallbackState; |
| } |
| |
| bool DelegateOptions::TfLiteRuntimeFallbackDisabled() |
| { |
| return p_DelegateOptionsImpl->m_DisableTfLiteRuntimeFallback; |
| } |
| |
| } // namespace armnnDelegate |