Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 1 | // |
| 2 | // Copyright © 2020 Arm Ltd and Contributors. All rights reserved. |
| 3 | // SPDX-License-Identifier: MIT |
| 4 | // |
| 5 | |
| 6 | #include "ExecuteNetworkProgramOptions.hpp" |
| 7 | #include "NetworkExecutionUtils/NetworkExecutionUtils.hpp" |
| 8 | #include "InferenceTest.hpp" |
| 9 | |
| 10 | #include <armnn/BackendRegistry.hpp> |
| 11 | #include <armnn/Exceptions.hpp> |
| 12 | #include <armnn/utility/Assert.hpp> |
| 13 | #include <armnn/utility/StringUtils.hpp> |
| 14 | #include <armnn/Logging.hpp> |
| 15 | |
| 16 | #include <fmt/format.h> |
| 17 | |
| 18 | bool CheckOption(const cxxopts::ParseResult& result, |
| 19 | const char* option) |
| 20 | { |
| 21 | // Check that the given option is valid. |
| 22 | if (option == nullptr) |
| 23 | { |
| 24 | return false; |
| 25 | } |
| 26 | |
| 27 | // Check whether 'option' is provided. |
| 28 | return ((result.count(option)) ? true : false); |
| 29 | } |
| 30 | |
| 31 | void CheckOptionDependency(const cxxopts::ParseResult& result, |
| 32 | const char* option, |
| 33 | const char* required) |
| 34 | { |
| 35 | // Check that the given options are valid. |
| 36 | if (option == nullptr || required == nullptr) |
| 37 | { |
| 38 | throw cxxopts::OptionParseException("Invalid option to check dependency for"); |
| 39 | } |
| 40 | |
| 41 | // Check that if 'option' is provided, 'required' is also provided. |
| 42 | if (CheckOption(result, option) && !result[option].has_default()) |
| 43 | { |
| 44 | if (CheckOption(result, required) == 0 || result[required].has_default()) |
| 45 | { |
| 46 | throw cxxopts::OptionParseException( |
| 47 | std::string("Option '") + option + "' requires option '" + required + "'."); |
| 48 | } |
| 49 | } |
| 50 | } |
| 51 | |
| 52 | void CheckOptionDependencies(const cxxopts::ParseResult& result) |
| 53 | { |
| 54 | CheckOptionDependency(result, "model-path", "model-format"); |
| 55 | CheckOptionDependency(result, "input-tensor-shape", "model-path"); |
| 56 | CheckOptionDependency(result, "tuning-level", "tuning-path"); |
| 57 | } |
| 58 | |
| 59 | void RemoveDuplicateDevices(std::vector<armnn::BackendId>& computeDevices) |
| 60 | { |
| 61 | // Mark the duplicate devices as 'Undefined'. |
| 62 | for (auto i = computeDevices.begin(); i != computeDevices.end(); ++i) |
| 63 | { |
| 64 | for (auto j = std::next(i); j != computeDevices.end(); ++j) |
| 65 | { |
| 66 | if (*j == *i) |
| 67 | { |
| 68 | *j = armnn::Compute::Undefined; |
| 69 | } |
| 70 | } |
| 71 | } |
| 72 | |
| 73 | // Remove 'Undefined' devices. |
| 74 | computeDevices.erase(std::remove(computeDevices.begin(), computeDevices.end(), armnn::Compute::Undefined), |
| 75 | computeDevices.end()); |
| 76 | } |
| 77 | |
Jan Eilers | c5b84b5 | 2021-02-16 12:40:43 +0000 | [diff] [blame] | 78 | /// Takes a vector of backend strings and returns a vector of backendIDs. |
| 79 | /// Removes duplicate entries. |
| 80 | /// Can handle backend strings that contain multiple backends separated by comma e.g "CpuRef,CpuAcc" |
| 81 | std::vector<armnn::BackendId> GetBackendIDs(const std::vector<std::string>& backendStringsVec) |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 82 | { |
| 83 | std::vector<armnn::BackendId> backendIDs; |
Jan Eilers | c5b84b5 | 2021-02-16 12:40:43 +0000 | [diff] [blame] | 84 | for (const auto& backendStrings : backendStringsVec) |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 85 | { |
Jan Eilers | c5b84b5 | 2021-02-16 12:40:43 +0000 | [diff] [blame] | 86 | // Each backendStrings might contain multiple backends separated by comma e.g "CpuRef,CpuAcc" |
| 87 | std::vector<std::string> backendStringVec = ParseStringList(backendStrings, ","); |
| 88 | for (const auto& b : backendStringVec) |
| 89 | { |
| 90 | backendIDs.push_back(armnn::BackendId(b)); |
| 91 | } |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 92 | } |
| 93 | |
| 94 | RemoveDuplicateDevices(backendIDs); |
| 95 | |
| 96 | return backendIDs; |
| 97 | } |
| 98 | |
| 99 | /// Provides a segfault safe way to get cxxopts option values by checking if the option was defined. |
| 100 | /// If the option wasn't defined it returns an empty object. |
| 101 | template<typename optionType> |
| 102 | optionType GetOptionValue(std::string&& optionName, const cxxopts::ParseResult& result) |
| 103 | { |
| 104 | optionType out; |
| 105 | if(result.count(optionName)) |
| 106 | { |
| 107 | out = result[optionName].as<optionType>(); |
| 108 | } |
| 109 | return out; |
| 110 | } |
| 111 | |
| 112 | void LogAndThrowFatal(std::string errorMessage) |
| 113 | { |
| 114 | throw armnn::InvalidArgumentException (errorMessage); |
| 115 | } |
| 116 | |
| 117 | void CheckRequiredOptions(const cxxopts::ParseResult& result) |
| 118 | { |
| 119 | |
| 120 | // For each option in option-group "a) Required |
| 121 | std::vector<std::string> requiredOptions{"compute", |
| 122 | "model-format", |
| 123 | "model-path", |
| 124 | "input-name", |
| 125 | "output-name"}; |
| 126 | |
| 127 | bool requiredMissing = false; |
| 128 | for(auto const& str : requiredOptions) |
| 129 | { |
| 130 | if(!(result.count(str) > 0)) |
| 131 | { |
| 132 | ARMNN_LOG(error) << fmt::format("The program option '{}' is mandatory but wasn't provided.", str); |
| 133 | requiredMissing = true; |
| 134 | } |
| 135 | } |
| 136 | if(requiredMissing) |
| 137 | { |
| 138 | throw armnn::InvalidArgumentException ("Some required arguments are missing"); |
| 139 | } |
| 140 | } |
| 141 | |
| 142 | void ProgramOptions::ValidateExecuteNetworkParams() |
| 143 | { |
| 144 | m_ExNetParams.ValidateParams(); |
| 145 | } |
| 146 | |
| 147 | void ProgramOptions::ValidateRuntimeOptions() |
| 148 | { |
| 149 | if (m_RuntimeOptions.m_ProfilingOptions.m_TimelineEnabled && |
| 150 | !m_RuntimeOptions.m_ProfilingOptions.m_EnableProfiling) |
| 151 | { |
| 152 | LogAndThrowFatal("Timeline profiling requires external profiling to be turned on"); |
| 153 | } |
| 154 | } |
| 155 | |
| 156 | |
| 157 | ProgramOptions::ProgramOptions() : m_CxxOptions{"ExecuteNetwork", |
| 158 | "Executes a neural network model using the provided input " |
| 159 | "tensor. Prints the resulting output tensor."} |
| 160 | { |
| 161 | try |
| 162 | { |
| 163 | // cxxopts doesn't provide a mechanism to ensure required options are given. There is a |
| 164 | // separate function CheckRequiredOptions() for that. |
| 165 | m_CxxOptions.add_options("a) Required") |
| 166 | ("c,compute", |
Jan Eilers | c5b84b5 | 2021-02-16 12:40:43 +0000 | [diff] [blame] | 167 | "Which device to run layers on by default. If a single device doesn't support all layers in the model " |
| 168 | "you can specify a second or third to fall back on. Possible choices: " |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 169 | + armnn::BackendRegistryInstance().GetBackendIdsAsString() |
Jan Eilers | c5b84b5 | 2021-02-16 12:40:43 +0000 | [diff] [blame] | 170 | + " NOTE: Multiple compute devices need to be passed as a comma separated list without whitespaces " |
| 171 | "e.g. GpuAcc,CpuAcc,CpuRef or by repeating the program option e.g. '-c Cpuacc -c CpuRef'. " |
| 172 | "Duplicates are ignored.", |
Jan Eilers | 3dda41d | 2020-11-11 11:44:14 +0000 | [diff] [blame] | 173 | cxxopts::value<std::vector<std::string>>()) |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 174 | |
| 175 | ("f,model-format", |
Nikhil Raj | 5d955cf | 2021-04-19 16:59:48 +0100 | [diff] [blame] | 176 | "armnn-binary, onnx-binary, onnx-text, tflite-binary", |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 177 | cxxopts::value<std::string>()) |
| 178 | |
| 179 | ("m,model-path", |
Nikhil Raj | 6dd178f | 2021-04-02 22:04:39 +0100 | [diff] [blame] | 180 | "Path to model file, e.g. .armnn, , .prototxt, .tflite, .onnx", |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 181 | cxxopts::value<std::string>(m_ExNetParams.m_ModelPath)) |
| 182 | |
| 183 | ("i,input-name", |
| 184 | "Identifier of the input tensors in the network separated by comma.", |
| 185 | cxxopts::value<std::string>()) |
| 186 | |
| 187 | ("o,output-name", |
| 188 | "Identifier of the output tensors in the network separated by comma.", |
| 189 | cxxopts::value<std::string>()); |
| 190 | |
| 191 | m_CxxOptions.add_options("b) General") |
| 192 | ("b,dynamic-backends-path", |
| 193 | "Path where to load any available dynamic backend from. " |
| 194 | "If left empty (the default), dynamic backends will not be used.", |
| 195 | cxxopts::value<std::string>(m_RuntimeOptions.m_DynamicBackendsPath)) |
| 196 | |
Sadik Armagan | a04a9d7 | 2021-04-27 10:02:10 +0100 | [diff] [blame] | 197 | ("n,concurrent", |
| 198 | "If this option is enabled inferences will be executed in parallel asynchronously.", |
| 199 | cxxopts::value<bool>(m_ExNetParams.m_Concurrent)->default_value("false")->implicit_value("true")) |
| 200 | |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 201 | ("d,input-tensor-data", |
| 202 | "Path to files containing the input data as a flat array separated by whitespace. " |
| 203 | "Several paths can be passed by separating them with a comma. If not specified, the network will be " |
| 204 | "run with dummy data (useful for profiling).", |
| 205 | cxxopts::value<std::string>()->default_value("")) |
| 206 | |
| 207 | ("h,help", "Display usage information") |
| 208 | |
| 209 | ("infer-output-shape", |
| 210 | "Infers output tensor shape from input tensor shape and validate where applicable (where supported by " |
| 211 | "parser)", |
| 212 | cxxopts::value<bool>(m_ExNetParams.m_InferOutputShape)->default_value("false")->implicit_value("true")) |
| 213 | |
| 214 | ("iterations", |
| 215 | "Number of iterations to run the network for, default is set to 1", |
| 216 | cxxopts::value<size_t>(m_ExNetParams.m_Iterations)->default_value("1")) |
| 217 | |
| 218 | ("l,dequantize-output", |
| 219 | "If this option is enabled, all quantized outputs will be dequantized to float. " |
| 220 | "If unset, default to not get dequantized. " |
| 221 | "Accepted values (true or false)", |
| 222 | cxxopts::value<bool>(m_ExNetParams.m_DequantizeOutput)->default_value("false")->implicit_value("true")) |
| 223 | |
| 224 | ("p,print-intermediate-layers", |
| 225 | "If this option is enabled, the output of every graph layer will be printed.", |
| 226 | cxxopts::value<bool>(m_ExNetParams.m_PrintIntermediate)->default_value("false") |
| 227 | ->implicit_value("true")) |
| 228 | |
| 229 | ("parse-unsupported", |
| 230 | "Add unsupported operators as stand-in layers (where supported by parser)", |
| 231 | cxxopts::value<bool>(m_ExNetParams.m_ParseUnsupported)->default_value("false")->implicit_value("true")) |
| 232 | |
| 233 | ("q,quantize-input", |
| 234 | "If this option is enabled, all float inputs will be quantized to qasymm8. " |
| 235 | "If unset, default to not quantized. Accepted values (true or false)", |
| 236 | cxxopts::value<bool>(m_ExNetParams.m_QuantizeInput)->default_value("false")->implicit_value("true")) |
| 237 | |
| 238 | ("r,threshold-time", |
| 239 | "Threshold time is the maximum allowed time for inference measured in milliseconds. If the actual " |
| 240 | "inference time is greater than the threshold time, the test will fail. By default, no threshold " |
| 241 | "time is used.", |
| 242 | cxxopts::value<double>(m_ExNetParams.m_ThresholdTime)->default_value("0.0")) |
| 243 | |
| 244 | ("s,input-tensor-shape", |
| 245 | "The shape of the input tensors in the network as a flat array of integers separated by comma." |
| 246 | "Several shapes can be passed by separating them with a colon (:).", |
| 247 | cxxopts::value<std::string>()) |
| 248 | |
| 249 | ("v,visualize-optimized-model", |
| 250 | "Enables built optimized model visualizer. If unset, defaults to off.", |
| 251 | cxxopts::value<bool>(m_ExNetParams.m_EnableLayerDetails)->default_value("false") |
| 252 | ->implicit_value("true")) |
| 253 | |
| 254 | ("w,write-outputs-to-file", |
| 255 | "Comma-separated list of output file paths keyed with the binding-id of the output slot. " |
| 256 | "If left empty (the default), the output tensors will not be written to a file.", |
| 257 | cxxopts::value<std::string>()) |
| 258 | |
| 259 | ("x,subgraph-number", |
| 260 | "Id of the subgraph to be executed. Defaults to 0.", |
| 261 | cxxopts::value<size_t>(m_ExNetParams.m_SubgraphId)->default_value("0")) |
| 262 | |
| 263 | ("y,input-type", |
| 264 | "The type of the input tensors in the network separated by comma. " |
| 265 | "If unset, defaults to \"float\" for all defined inputs. " |
| 266 | "Accepted values (float, int or qasymm8).", |
| 267 | cxxopts::value<std::string>()) |
| 268 | |
| 269 | ("z,output-type", |
| 270 | "The type of the output tensors in the network separated by comma. " |
| 271 | "If unset, defaults to \"float\" for all defined outputs. " |
| 272 | "Accepted values (float, int or qasymm8).", |
Finn Williams | f806c4d | 2021-02-22 15:13:12 +0000 | [diff] [blame] | 273 | cxxopts::value<std::string>()) |
| 274 | |
| 275 | ("T,tflite-executor", |
| 276 | "Set the executor for the tflite model: parser, delegate, tflite" |
| 277 | "parser is the ArmNNTfLiteParser, " |
| 278 | "delegate is the ArmNNTfLiteDelegate, " |
| 279 | "tflite is the TfliteInterpreter", |
| 280 | cxxopts::value<std::string>()->default_value("parser")) |
| 281 | |
| 282 | ("D,armnn-tflite-delegate", |
| 283 | "Enable Arm NN TfLite delegate. " |
| 284 | "This option is depreciated please use tflite-executor instead", |
Sadik Armagan | a04a9d7 | 2021-04-27 10:02:10 +0100 | [diff] [blame] | 285 | cxxopts::value<bool>(m_ExNetParams.m_EnableDelegate)->default_value("false")->implicit_value("true")) |
| 286 | |
| 287 | ("simultaneous-iterations", |
| 288 | "Number of simultaneous iterations to async-run the network for, default is set to 1", |
| 289 | cxxopts::value<size_t>(m_ExNetParams.m_SimultaneousIterations)->default_value("1")); |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 290 | |
| 291 | m_CxxOptions.add_options("c) Optimization") |
| 292 | ("bf16-turbo-mode", |
| 293 | "If this option is enabled, FP32 layers, " |
| 294 | "weights and biases will be converted to BFloat16 where the backend supports it", |
| 295 | cxxopts::value<bool>(m_ExNetParams.m_EnableBf16TurboMode) |
| 296 | ->default_value("false")->implicit_value("true")) |
| 297 | |
| 298 | ("enable-fast-math", |
| 299 | "Enables fast_math options in backends that support it. Using the fast_math flag can lead to " |
| 300 | "performance improvements but may result in reduced or different precision.", |
| 301 | cxxopts::value<bool>(m_ExNetParams.m_EnableFastMath)->default_value("false")->implicit_value("true")) |
| 302 | |
Matthew Sloyan | 0a7dc6b | 2021-02-10 16:50:53 +0000 | [diff] [blame] | 303 | ("number-of-threads", |
| 304 | "Assign the number of threads used by the CpuAcc backend. " |
| 305 | "Input value must be between 1 and 64. " |
| 306 | "Default is set to 0 (Backend will decide number of threads to use).", |
| 307 | cxxopts::value<unsigned int>(m_ExNetParams.m_NumberOfThreads)->default_value("0")) |
| 308 | |
Matthew Sloyan | 4243211 | 2021-01-08 10:30:51 +0000 | [diff] [blame] | 309 | ("save-cached-network", |
Matthew Sloyan | 9d7a332 | 2021-01-12 16:19:43 +0000 | [diff] [blame] | 310 | "Enables saving of the cached network to a file given with the cached-network-filepath option. " |
Matthew Sloyan | 4243211 | 2021-01-08 10:30:51 +0000 | [diff] [blame] | 311 | "See also --cached-network-filepath", |
| 312 | cxxopts::value<bool>(m_ExNetParams.m_SaveCachedNetwork) |
| 313 | ->default_value("false")->implicit_value("true")) |
| 314 | |
| 315 | ("cached-network-filepath", |
Matthew Sloyan | 9d7a332 | 2021-01-12 16:19:43 +0000 | [diff] [blame] | 316 | "If non-empty, the given file will be used to load/save the cached network. " |
| 317 | "If save-cached-network is given then the cached network will be saved to the given file. " |
| 318 | "To save the cached network a file must already exist. " |
| 319 | "If save-cached-network is not given then the cached network will be loaded from the given file. " |
| 320 | "This will remove initial compilation time of kernels and speed up the first execution.", |
Matthew Sloyan | 4243211 | 2021-01-08 10:30:51 +0000 | [diff] [blame] | 321 | cxxopts::value<std::string>(m_ExNetParams.m_CachedNetworkFilePath)->default_value("")) |
| 322 | |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 323 | ("fp16-turbo-mode", |
| 324 | "If this option is enabled, FP32 layers, " |
| 325 | "weights and biases will be converted to FP16 where the backend supports it", |
| 326 | cxxopts::value<bool>(m_ExNetParams.m_EnableFp16TurboMode) |
| 327 | ->default_value("false")->implicit_value("true")) |
| 328 | |
| 329 | ("tuning-level", |
| 330 | "Sets the tuning level which enables a tuning run which will update/create a tuning file. " |
| 331 | "Available options are: 1 (Rapid), 2 (Normal), 3 (Exhaustive). " |
| 332 | "Requires tuning-path to be set, default is set to 0 (No tuning run)", |
| 333 | cxxopts::value<int>(m_ExNetParams.m_TuningLevel)->default_value("0")) |
| 334 | |
| 335 | ("tuning-path", |
| 336 | "Path to tuning file. Enables use of CL tuning", |
Finn Williams | 4064632 | 2021-02-11 16:16:42 +0000 | [diff] [blame] | 337 | cxxopts::value<std::string>(m_ExNetParams.m_TuningPath)) |
| 338 | |
| 339 | ("MLGOTuningFilePath", |
| 340 | "Path to tuning file. Enables use of CL MLGO tuning", |
| 341 | cxxopts::value<std::string>(m_ExNetParams.m_MLGOTuningFilePath)); |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 342 | |
| 343 | m_CxxOptions.add_options("d) Profiling") |
| 344 | ("a,enable-external-profiling", |
| 345 | "If enabled external profiling will be switched on", |
| 346 | cxxopts::value<bool>(m_RuntimeOptions.m_ProfilingOptions.m_EnableProfiling) |
| 347 | ->default_value("false")->implicit_value("true")) |
| 348 | |
| 349 | ("e,event-based-profiling", |
| 350 | "Enables built in profiler. If unset, defaults to off.", |
| 351 | cxxopts::value<bool>(m_ExNetParams.m_EnableProfiling)->default_value("false")->implicit_value("true")) |
| 352 | |
| 353 | ("g,file-only-external-profiling", |
| 354 | "If enabled then the 'file-only' test mode of external profiling will be enabled", |
| 355 | cxxopts::value<bool>(m_RuntimeOptions.m_ProfilingOptions.m_FileOnly) |
| 356 | ->default_value("false")->implicit_value("true")) |
| 357 | |
| 358 | ("file-format", |
| 359 | "If profiling is enabled specifies the output file format", |
| 360 | cxxopts::value<std::string>(m_RuntimeOptions.m_ProfilingOptions.m_FileFormat)->default_value("binary")) |
| 361 | |
| 362 | ("j,outgoing-capture-file", |
| 363 | "If specified the outgoing external profiling packets will be captured in this binary file", |
| 364 | cxxopts::value<std::string>(m_RuntimeOptions.m_ProfilingOptions.m_OutgoingCaptureFile)) |
| 365 | |
| 366 | ("k,incoming-capture-file", |
| 367 | "If specified the incoming external profiling packets will be captured in this binary file", |
| 368 | cxxopts::value<std::string>(m_RuntimeOptions.m_ProfilingOptions.m_IncomingCaptureFile)) |
| 369 | |
| 370 | ("timeline-profiling", |
| 371 | "If enabled timeline profiling will be switched on, requires external profiling", |
| 372 | cxxopts::value<bool>(m_RuntimeOptions.m_ProfilingOptions.m_TimelineEnabled) |
| 373 | ->default_value("false")->implicit_value("true")) |
| 374 | |
| 375 | ("u,counter-capture-period", |
| 376 | "If profiling is enabled in 'file-only' mode this is the capture period that will be used in the test", |
| 377 | cxxopts::value<uint32_t>(m_RuntimeOptions.m_ProfilingOptions.m_CapturePeriod)->default_value("150")); |
| 378 | } |
| 379 | catch (const std::exception& e) |
| 380 | { |
| 381 | ARMNN_ASSERT_MSG(false, "Caught unexpected exception"); |
| 382 | ARMNN_LOG(fatal) << "Fatal internal error: " << e.what(); |
| 383 | exit(EXIT_FAILURE); |
| 384 | } |
| 385 | } |
| 386 | |
| 387 | ProgramOptions::ProgramOptions(int ac, const char* av[]): ProgramOptions() |
| 388 | { |
| 389 | ParseOptions(ac, av); |
| 390 | } |
| 391 | |
| 392 | void ProgramOptions::ParseOptions(int ac, const char* av[]) |
| 393 | { |
| 394 | // Parses the command-line. |
| 395 | m_CxxResult = m_CxxOptions.parse(ac, av); |
| 396 | |
| 397 | if (m_CxxResult.count("help") || ac <= 1) |
| 398 | { |
| 399 | std::cout << m_CxxOptions.help() << std::endl; |
| 400 | exit(EXIT_SUCCESS); |
| 401 | } |
| 402 | |
| 403 | CheckRequiredOptions(m_CxxResult); |
| 404 | CheckOptionDependencies(m_CxxResult); |
| 405 | |
| 406 | // Some options can't be assigned directly because they need some post-processing: |
Jan Eilers | 3dda41d | 2020-11-11 11:44:14 +0000 | [diff] [blame] | 407 | auto computeDevices = GetOptionValue<std::vector<std::string>>("compute", m_CxxResult); |
| 408 | m_ExNetParams.m_ComputeDevices = GetBackendIDs(computeDevices); |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 409 | m_ExNetParams.m_ModelFormat = |
| 410 | armnn::stringUtils::StringTrimCopy(GetOptionValue<std::string>("model-format", m_CxxResult)); |
| 411 | m_ExNetParams.m_InputNames = |
| 412 | ParseStringList(GetOptionValue<std::string>("input-name", m_CxxResult), ","); |
| 413 | m_ExNetParams.m_InputTensorDataFilePaths = |
| 414 | ParseStringList(GetOptionValue<std::string>("input-tensor-data", m_CxxResult), ","); |
| 415 | m_ExNetParams.m_OutputNames = |
| 416 | ParseStringList(GetOptionValue<std::string>("output-name", m_CxxResult), ","); |
| 417 | m_ExNetParams.m_InputTypes = |
| 418 | ParseStringList(GetOptionValue<std::string>("input-type", m_CxxResult), ","); |
| 419 | m_ExNetParams.m_OutputTypes = |
| 420 | ParseStringList(GetOptionValue<std::string>("output-type", m_CxxResult), ","); |
| 421 | m_ExNetParams.m_OutputTensorFiles = |
| 422 | ParseStringList(GetOptionValue<std::string>("write-outputs-to-file", m_CxxResult), ","); |
| 423 | m_ExNetParams.m_GenerateTensorData = |
| 424 | m_ExNetParams.m_InputTensorDataFilePaths.empty(); |
Francis Murtagh | bf18a26 | 2020-10-27 15:20:40 +0000 | [diff] [blame] | 425 | m_ExNetParams.m_DynamicBackendsPath = m_RuntimeOptions.m_DynamicBackendsPath; |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 426 | |
Sadik Armagan | 8c7a28b | 2021-04-01 17:27:21 +0100 | [diff] [blame] | 427 | m_RuntimeOptions.m_EnableGpuProfiling = m_ExNetParams.m_EnableProfiling; |
Finn Williams | f806c4d | 2021-02-22 15:13:12 +0000 | [diff] [blame] | 428 | |
| 429 | std::string tfliteExecutor = GetOptionValue<std::string>("tflite-executor", m_CxxResult); |
| 430 | |
| 431 | if (tfliteExecutor.size() == 0 || tfliteExecutor == "parser") |
| 432 | { |
| 433 | m_ExNetParams.m_TfLiteExecutor = ExecuteNetworkParams::TfLiteExecutor::ArmNNTfLiteParser; |
| 434 | } |
| 435 | else if (tfliteExecutor == "delegate") |
| 436 | { |
| 437 | m_ExNetParams.m_TfLiteExecutor = ExecuteNetworkParams::TfLiteExecutor::ArmNNTfLiteDelegate; |
| 438 | } |
| 439 | else if (tfliteExecutor == "tflite") |
| 440 | { |
| 441 | m_ExNetParams.m_TfLiteExecutor = ExecuteNetworkParams::TfLiteExecutor::TfliteInterpreter; |
| 442 | } |
| 443 | else |
| 444 | { |
| 445 | ARMNN_LOG(info) << fmt::format("Invalid tflite-executor option '{}'.", tfliteExecutor); |
| 446 | throw armnn::InvalidArgumentException ("Invalid tflite-executor option"); |
| 447 | } |
| 448 | |
| 449 | if (m_ExNetParams.m_EnableDelegate) |
| 450 | { |
| 451 | m_ExNetParams.m_TfLiteExecutor = ExecuteNetworkParams::TfLiteExecutor::ArmNNTfLiteDelegate; |
| 452 | ARMNN_LOG(info) << fmt::format("armnn-tflite-delegate option is being depreciated, " |
| 453 | "please use tflite-executor instead."); |
| 454 | } |
| 455 | |
| 456 | |
| 457 | |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 458 | // Parse input tensor shape from the string we got from the command-line. |
| 459 | std::vector<std::string> inputTensorShapesVector = |
| 460 | ParseStringList(GetOptionValue<std::string>("input-tensor-shape", m_CxxResult), ":"); |
| 461 | |
| 462 | if (!inputTensorShapesVector.empty()) |
| 463 | { |
| 464 | m_ExNetParams.m_InputTensorShapes.reserve(inputTensorShapesVector.size()); |
| 465 | |
| 466 | for(const std::string& shape : inputTensorShapesVector) |
| 467 | { |
| 468 | std::stringstream ss(shape); |
| 469 | std::vector<unsigned int> dims = ParseArray(ss); |
| 470 | |
| 471 | m_ExNetParams.m_InputTensorShapes.push_back( |
| 472 | std::make_unique<armnn::TensorShape>(static_cast<unsigned int>(dims.size()), dims.data())); |
| 473 | } |
| 474 | } |
| 475 | |
| 476 | // We have to validate ExecuteNetworkParams first so that the tuning path and level is validated |
| 477 | ValidateExecuteNetworkParams(); |
| 478 | |
| 479 | // Parse CL tuning parameters to runtime options |
| 480 | if (!m_ExNetParams.m_TuningPath.empty()) |
| 481 | { |
| 482 | m_RuntimeOptions.m_BackendOptions.emplace_back( |
| 483 | armnn::BackendOptions |
| 484 | { |
| 485 | "GpuAcc", |
| 486 | { |
| 487 | {"TuningLevel", m_ExNetParams.m_TuningLevel}, |
| 488 | {"TuningFile", m_ExNetParams.m_TuningPath.c_str()}, |
Finn Williams | 4064632 | 2021-02-11 16:16:42 +0000 | [diff] [blame] | 489 | {"KernelProfilingEnabled", m_ExNetParams.m_EnableProfiling}, |
| 490 | {"MLGOTuningFilePath", m_ExNetParams.m_MLGOTuningFilePath} |
Jan Eilers | 4527490 | 2020-10-15 18:34:43 +0100 | [diff] [blame] | 491 | } |
| 492 | } |
| 493 | ); |
| 494 | } |
| 495 | |
| 496 | ValidateRuntimeOptions(); |
| 497 | } |
| 498 | |