blob: 7c51011a228f9d920e1331d9e2f5f1b3adcf1e1f [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
Sadik Armagana9c2ce12020-07-14 10:02:22 +01002// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
Matteo Martincighc601aa62019-10-29 15:03:22 +00005
telsoa014fcda012018-03-09 14:13:49 +00006#pragma once
Matteo Martincighc601aa62019-10-29 15:03:22 +00007
David Beckf0b48452018-10-19 15:20:56 +01008#include <armnn/ArmNN.hpp>
alered01a7227ac2020-05-07 14:58:29 +01009#include <armnn/Logging.hpp>
10#include <armnn/utility/Timer.hpp>
Matteo Martincighc601aa62019-10-29 15:03:22 +000011#include <armnn/BackendRegistry.hpp>
Narumol Prangnawaratac2770a2020-04-01 16:51:23 +010012#include <armnn/utility/Assert.hpp>
Matthew Sloyan80c6b142020-09-08 12:00:32 +010013#include <armnn/utility/NumericCast.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010014
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000015#if defined(ARMNN_SERIALIZER)
Derek Lamberti0028d1b2019-02-20 13:57:42 +000016#include "armnnDeserializer/IDeserializer.hpp"
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000017#endif
telsoa01c577f2c2018-08-31 09:22:23 +010018#if defined(ARMNN_TF_LITE_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010019#include <armnnTfLiteParser/ITfLiteParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010020#endif
telsoa01c577f2c2018-08-31 09:22:23 +010021#if defined(ARMNN_ONNX_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010022#include <armnnOnnxParser/IOnnxParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010023#endif
telsoa014fcda012018-03-09 14:13:49 +000024
Francis Murtagh532a29d2020-06-29 11:50:01 +010025#include <Filesystem.hpp>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000026#include <HeapProfiling.hpp>
Jim Flynn2fd61002019-05-03 12:54:26 +010027#include <TensorIOUtils.hpp>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000028
David Monahana8837bf2020-04-16 10:01:56 +010029#include "armnn/utility/StringUtils.hpp"
James Wardc89829f2020-10-12 14:17:36 +010030#include <cxxopts/cxxopts.hpp>
31#include "CxxoptsUtils.hpp"
James Ward08f40162020-09-07 16:45:07 +010032#include <fmt/format.h>
James Ward6d9f5c52020-09-28 11:56:35 +010033#include <mapbox/variant.hpp>
telsoa014fcda012018-03-09 14:13:49 +000034
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000035#include <algorithm>
36#include <iterator>
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010037#include <fstream>
telsoa014fcda012018-03-09 14:13:49 +000038#include <map>
39#include <string>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000040#include <vector>
telsoa01c577f2c2018-08-31 09:22:23 +010041#include <type_traits>
42
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010043namespace
44{
45
46inline bool CheckRequestedBackendsAreValid(const std::vector<armnn::BackendId>& backendIds,
47 armnn::Optional<std::string&> invalidBackendIds = armnn::EmptyOptional())
48{
49 if (backendIds.empty())
50 {
51 return false;
52 }
53
54 armnn::BackendIdSet validBackendIds = armnn::BackendRegistryInstance().GetBackendIds();
55
56 bool allValid = true;
57 for (const auto& backendId : backendIds)
58 {
59 if (std::find(validBackendIds.begin(), validBackendIds.end(), backendId) == validBackendIds.end())
60 {
61 allValid = false;
62 if (invalidBackendIds)
63 {
64 if (!invalidBackendIds.value().empty())
65 {
66 invalidBackendIds.value() += ", ";
67 }
68 invalidBackendIds.value() += backendId;
69 }
70 }
71 }
72 return allValid;
73}
74
75} // anonymous namespace
76
telsoa01c577f2c2018-08-31 09:22:23 +010077namespace InferenceModelInternal
78{
Jim Flynnb4d7eae2019-05-01 14:44:27 +010079using BindingPointInfo = armnn::BindingPointInfo;
telsoa01c577f2c2018-08-31 09:22:23 +010080
81using QuantizationParams = std::pair<float,int32_t>;
82
83struct Params
84{
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000085 std::string m_ModelPath;
86 std::vector<std::string> m_InputBindings;
87 std::vector<armnn::TensorShape> m_InputShapes;
88 std::vector<std::string> m_OutputBindings;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000089 std::vector<armnn::BackendId> m_ComputeDevices;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010090 std::string m_DynamicBackendsPath;
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000091 size_t m_SubgraphId;
92 bool m_IsModelBinary;
93 bool m_VisualizePostOptimizationModel;
94 bool m_EnableFp16TurboMode;
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +000095 bool m_EnableBf16TurboMode;
Matthew Jackson54658b92019-08-27 15:35:59 +010096 bool m_PrintIntermediateLayers;
Derek Lamberti132563c2019-12-02 16:06:40 +000097 bool m_ParseUnsupported;
Sadik Armagana9c2ce12020-07-14 10:02:22 +010098 bool m_InferOutputShape;
Sadik Armagana25886e2020-09-15 17:17:08 +010099 bool m_EnableFastMath;
Matthew Sloyan42432112021-01-08 10:30:51 +0000100 bool m_SaveCachedNetwork;
101 std::string m_CachedNetworkFilePath;
Matthew Sloyan0a7dc6b2021-02-10 16:50:53 +0000102 unsigned int m_NumberOfThreads;
Finn Williams40646322021-02-11 16:16:42 +0000103 std::string m_MLGOTuningFilePath;
Sadik Armagana04a9d72021-04-27 10:02:10 +0100104 bool m_AsyncEnabled;
Kevin Mayb4b3ac92021-05-21 16:42:21 +0100105 size_t m_ThreadPoolSize;
Finn Williams40646322021-02-11 16:16:42 +0000106
telsoa01c577f2c2018-08-31 09:22:23 +0100107
108 Params()
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100109 : m_ComputeDevices{}
telsoa01c577f2c2018-08-31 09:22:23 +0100110 , m_SubgraphId(0)
111 , m_IsModelBinary(true)
112 , m_VisualizePostOptimizationModel(false)
113 , m_EnableFp16TurboMode(false)
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000114 , m_EnableBf16TurboMode(false)
Matthew Jackson54658b92019-08-27 15:35:59 +0100115 , m_PrintIntermediateLayers(false)
Derek Lamberti132563c2019-12-02 16:06:40 +0000116 , m_ParseUnsupported(false)
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100117 , m_InferOutputShape(false)
Sadik Armagana25886e2020-09-15 17:17:08 +0100118 , m_EnableFastMath(false)
Matthew Sloyan42432112021-01-08 10:30:51 +0000119 , m_SaveCachedNetwork(false)
120 , m_CachedNetworkFilePath("")
Matthew Sloyan0a7dc6b2021-02-10 16:50:53 +0000121 , m_NumberOfThreads(0)
Finn Williams40646322021-02-11 16:16:42 +0000122 , m_MLGOTuningFilePath("")
Sadik Armagana04a9d72021-04-27 10:02:10 +0100123 , m_AsyncEnabled(false)
Kevin Mayb4b3ac92021-05-21 16:42:21 +0100124 , m_ThreadPoolSize(1)
telsoa01c577f2c2018-08-31 09:22:23 +0100125 {}
126};
127
128} // namespace InferenceModelInternal
129
130template <typename IParser>
131struct CreateNetworkImpl
132{
133public:
134 using Params = InferenceModelInternal::Params;
telsoa01c577f2c2018-08-31 09:22:23 +0100135
136 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100137 std::vector<armnn::BindingPointInfo>& inputBindings,
138 std::vector<armnn::BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100139 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000140 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100141
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000142 // Create a network from a file on disk
143 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100144
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000145 std::map<std::string, armnn::TensorShape> inputShapes;
146 if (!params.m_InputShapes.empty())
147 {
148 const size_t numInputShapes = params.m_InputShapes.size();
149 const size_t numInputBindings = params.m_InputBindings.size();
150 if (numInputShapes < numInputBindings)
151 {
James Ward08f40162020-09-07 16:45:07 +0100152 throw armnn::Exception(fmt::format(
153 "Not every input has its tensor shape specified: expected={0}, got={1}",
154 numInputBindings, numInputShapes));
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000155 }
telsoa01c577f2c2018-08-31 09:22:23 +0100156
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000157 for (size_t i = 0; i < numInputShapes; i++)
158 {
159 inputShapes[params.m_InputBindings[i]] = params.m_InputShapes[i];
160 }
161 }
telsoa01c577f2c2018-08-31 09:22:23 +0100162
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000163 std::vector<std::string> requestedOutputs = params.m_OutputBindings;
164 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
165
166 {
167 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
168 // Handle text and binary input differently by calling the corresponding parser function
169 network = (params.m_IsModelBinary ?
170 parser->CreateNetworkFromBinaryFile(modelPath.c_str(), inputShapes, requestedOutputs) :
171 parser->CreateNetworkFromTextFile(modelPath.c_str(), inputShapes, requestedOutputs));
172 }
173
174 for (const std::string& inputLayerName : params.m_InputBindings)
175 {
176 inputBindings.push_back(parser->GetNetworkInputBindingInfo(inputLayerName));
177 }
178
179 for (const std::string& outputLayerName : params.m_OutputBindings)
180 {
181 outputBindings.push_back(parser->GetNetworkOutputBindingInfo(outputLayerName));
182 }
183
184 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100185 }
186};
187
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000188#if defined(ARMNN_SERIALIZER)
189template <>
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000190struct CreateNetworkImpl<armnnDeserializer::IDeserializer>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000191{
192public:
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000193 using IParser = armnnDeserializer::IDeserializer;
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000194 using Params = InferenceModelInternal::Params;
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000195
196 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100197 std::vector<armnn::BindingPointInfo>& inputBindings,
198 std::vector<armnn::BindingPointInfo>& outputBindings)
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000199 {
200 auto parser(IParser::Create());
Narumol Prangnawaratac2770a2020-04-01 16:51:23 +0100201 ARMNN_ASSERT(parser);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000202
203 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
204
205 {
206 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
Derek Lamberti2b183fb2019-02-18 16:36:57 +0000207
Francis Murtagh532a29d2020-06-29 11:50:01 +0100208 std::error_code errorCode;
209 fs::path pathToFile(params.m_ModelPath);
210 if (!fs::exists(pathToFile, errorCode))
Derek Lamberti2b183fb2019-02-18 16:36:57 +0000211 {
James Ward08f40162020-09-07 16:45:07 +0100212 throw armnn::FileNotFoundException(fmt::format("Cannot find the file ({0}) errorCode: {1} {2}",
213 params.m_ModelPath,
214 errorCode.message(),
Derek Lamberti2b183fb2019-02-18 16:36:57 +0000215 CHECK_LOCATION().AsString()));
216 }
217 std::ifstream file(params.m_ModelPath, std::ios::binary);
218
219 network = parser->CreateNetworkFromBinary(file);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000220 }
221
Matthew Sloyan80c6b142020-09-08 12:00:32 +0100222 unsigned int subgraphId = armnn::numeric_cast<unsigned int>(params.m_SubgraphId);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000223
224 for (const std::string& inputLayerName : params.m_InputBindings)
225 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000226 armnnDeserializer::BindingPointInfo inputBinding =
Derek Lambertiff05cc52019-04-26 13:05:17 +0100227 parser->GetNetworkInputBindingInfo(subgraphId, inputLayerName);
Derek Lamberti8ddae332019-02-21 16:29:43 +0000228 inputBindings.push_back(std::make_pair(inputBinding.m_BindingId, inputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000229 }
230
231 for (const std::string& outputLayerName : params.m_OutputBindings)
232 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000233 armnnDeserializer::BindingPointInfo outputBinding =
Derek Lambertiff05cc52019-04-26 13:05:17 +0100234 parser->GetNetworkOutputBindingInfo(subgraphId, outputLayerName);
Derek Lamberti8ddae332019-02-21 16:29:43 +0000235 outputBindings.push_back(std::make_pair(outputBinding.m_BindingId, outputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000236 }
237
238 return network;
239 }
240};
241#endif
242
telsoa01c577f2c2018-08-31 09:22:23 +0100243#if defined(ARMNN_TF_LITE_PARSER)
244template <>
245struct CreateNetworkImpl<armnnTfLiteParser::ITfLiteParser>
246{
247public:
248 using IParser = armnnTfLiteParser::ITfLiteParser;
249 using Params = InferenceModelInternal::Params;
telsoa01c577f2c2018-08-31 09:22:23 +0100250
251 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100252 std::vector<armnn::BindingPointInfo>& inputBindings,
253 std::vector<armnn::BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100254 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000255 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100256
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000257 // Create a network from a file on disk
Derek Lamberti132563c2019-12-02 16:06:40 +0000258 IParser::TfLiteParserOptions options;
259 options.m_StandInLayerForUnsupported = params.m_ParseUnsupported;
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100260 options.m_InferAndValidate = params.m_InferOutputShape;
Derek Lamberti132563c2019-12-02 16:06:40 +0000261 auto parser(IParser::Create(options));
telsoa01c577f2c2018-08-31 09:22:23 +0100262
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000263 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100264
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000265 {
266 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
267 network = parser->CreateNetworkFromBinaryFile(modelPath.c_str());
268 }
telsoa01c577f2c2018-08-31 09:22:23 +0100269
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000270 for (const std::string& inputLayerName : params.m_InputBindings)
271 {
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100272 armnn::BindingPointInfo inputBinding =
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000273 parser->GetNetworkInputBindingInfo(params.m_SubgraphId, inputLayerName);
274 inputBindings.push_back(inputBinding);
275 }
276
277 for (const std::string& outputLayerName : params.m_OutputBindings)
278 {
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100279 armnn::BindingPointInfo outputBinding =
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000280 parser->GetNetworkOutputBindingInfo(params.m_SubgraphId, outputLayerName);
281 outputBindings.push_back(outputBinding);
282 }
283
284 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100285 }
286};
287#endif
288
289#if defined(ARMNN_ONNX_PARSER)
290template <>
291struct CreateNetworkImpl<armnnOnnxParser::IOnnxParser>
292{
293public:
294 using IParser = armnnOnnxParser::IOnnxParser;
295 using Params = InferenceModelInternal::Params;
296 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
297
298 static armnn::INetworkPtr Create(const Params& params,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000299 std::vector<BindingPointInfo>& inputBindings,
300 std::vector<BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100301 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000302 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100303
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000304 // Create a network from a file on disk
305 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100306
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000307 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100308
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000309 {
310 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
311 network = (params.m_IsModelBinary ?
312 parser->CreateNetworkFromBinaryFile(modelPath.c_str()) :
313 parser->CreateNetworkFromTextFile(modelPath.c_str()));
314 }
telsoa01c577f2c2018-08-31 09:22:23 +0100315
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000316 for (const std::string& inputLayerName : params.m_InputBindings)
317 {
318 BindingPointInfo inputBinding = parser->GetNetworkInputBindingInfo(inputLayerName);
319 inputBindings.push_back(inputBinding);
320 }
321
322 for (const std::string& outputLayerName : params.m_OutputBindings)
323 {
324 BindingPointInfo outputBinding = parser->GetNetworkOutputBindingInfo(outputLayerName);
325 outputBindings.push_back(outputBinding);
326 }
327
328 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100329 }
330};
331#endif
telsoa014fcda012018-03-09 14:13:49 +0000332
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000333
telsoa014fcda012018-03-09 14:13:49 +0000334
335template <typename IParser, typename TDataType>
336class InferenceModel
337{
338public:
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000339 using DataType = TDataType;
340 using Params = InferenceModelInternal::Params;
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000341 using QuantizationParams = InferenceModelInternal::QuantizationParams;
Finn Williamsf806c4d2021-02-22 15:13:12 +0000342 using TContainer
343 = mapbox::util::variant<std::vector<float>, std::vector<int>, std::vector<unsigned char>, std::vector<int8_t>>;
telsoa014fcda012018-03-09 14:13:49 +0000344
345 struct CommandLineOptions
346 {
347 std::string m_ModelDir;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000348 std::vector<std::string> m_ComputeDevices;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100349 std::string m_DynamicBackendsPath;
surmeh013537c2c2018-05-18 16:31:43 +0100350 bool m_VisualizePostOptimizationModel;
telsoa01c577f2c2018-08-31 09:22:23 +0100351 bool m_EnableFp16TurboMode;
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000352 bool m_EnableBf16TurboMode;
Pablo Tello507f39d2019-04-15 15:44:39 +0100353 std::string m_Labels;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000354
355 std::vector<armnn::BackendId> GetComputeDevicesAsBackendIds()
356 {
357 std::vector<armnn::BackendId> backendIds;
358 std::copy(m_ComputeDevices.begin(), m_ComputeDevices.end(), std::back_inserter(backendIds));
359 return backendIds;
360 }
telsoa014fcda012018-03-09 14:13:49 +0000361 };
362
James Wardc89829f2020-10-12 14:17:36 +0100363 static void AddCommandLineOptions(cxxopts::Options& options,
364 CommandLineOptions& cLineOptions, std::vector<std::string>& required)
telsoa014fcda012018-03-09 14:13:49 +0000365 {
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000366 const std::vector<std::string> defaultComputes = { "CpuAcc", "CpuRef" };
David Beckf0b48452018-10-19 15:20:56 +0100367
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100368 const std::string backendsMessage = "Which device to run layers on by default. Possible choices: "
369 + armnn::BackendRegistryInstance().GetBackendIdsAsString();
370
James Wardc89829f2020-10-12 14:17:36 +0100371 options
372 .allow_unrecognised_options()
373 .add_options()
Nikhil Raj6dd178f2021-04-02 22:04:39 +0100374 ("m,model-dir", "Path to directory containing model files (.prototxt/.tflite)",
James Wardc89829f2020-10-12 14:17:36 +0100375 cxxopts::value<std::string>(cLineOptions.m_ModelDir))
376 ("c,compute", backendsMessage.c_str(),
377 cxxopts::value<std::vector<std::string>>(cLineOptions.m_ComputeDevices)->default_value("CpuRef"))
378 ("b,dynamic-backends-path",
379 "Path where to load any available dynamic backend from. "
380 "If left empty (the default), dynamic backends will not be used.",
381 cxxopts::value(cLineOptions.m_DynamicBackendsPath))
382 ("l,labels",
383 "Text file containing one image filename - correct label pair per line, "
384 "used to test the accuracy of the network.", cxxopts::value<std::string>(cLineOptions.m_Labels))
385 ("v,visualize-optimized-model",
386 "Produce a dot file useful for visualizing the graph post optimization."
387 "The file will have the same name as the model with the .dot extention.",
388 cxxopts::value<bool>(cLineOptions.m_VisualizePostOptimizationModel)->default_value("false"))
389 ("fp16-turbo-mode",
390 "If this option is enabled FP32 layers, weights and biases will be converted "
391 "to FP16 where the backend supports it.",
392 cxxopts::value<bool>(cLineOptions.m_EnableFp16TurboMode)->default_value("false"))
393 ("bf16-turbo-mode",
394 "If this option is enabled FP32 layers, weights and biases will be converted "
395 "to BF16 where the backend supports it.",
396 cxxopts::value<bool>(cLineOptions.m_EnableBf16TurboMode)->default_value("false"));
397
398 required.emplace_back("model-dir");
telsoa014fcda012018-03-09 14:13:49 +0000399 }
400
Matthew Bentham3e68b972019-04-09 13:10:46 +0100401 InferenceModel(const Params& params,
402 bool enableProfiling,
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100403 const std::string& dynamicBackendsPath,
Matthew Bentham3e68b972019-04-09 13:10:46 +0100404 const std::shared_ptr<armnn::IRuntime>& runtime = nullptr)
405 : m_EnableProfiling(enableProfiling)
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100406 , m_DynamicBackendsPath(dynamicBackendsPath)
telsoa014fcda012018-03-09 14:13:49 +0000407 {
telsoa01c577f2c2018-08-31 09:22:23 +0100408 if (runtime)
telsoa014fcda012018-03-09 14:13:49 +0000409 {
telsoa01c577f2c2018-08-31 09:22:23 +0100410 m_Runtime = runtime;
telsoa014fcda012018-03-09 14:13:49 +0000411 }
telsoa01c577f2c2018-08-31 09:22:23 +0100412 else
telsoa014fcda012018-03-09 14:13:49 +0000413 {
telsoa01c577f2c2018-08-31 09:22:23 +0100414 armnn::IRuntime::CreationOptions options;
Nina Drozd549ae372018-09-10 14:26:44 +0100415 options.m_EnableGpuProfiling = m_EnableProfiling;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100416 options.m_DynamicBackendsPath = m_DynamicBackendsPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100417 m_Runtime = std::move(armnn::IRuntime::Create(options));
surmeh013537c2c2018-05-18 16:31:43 +0100418 }
telsoa014fcda012018-03-09 14:13:49 +0000419
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100420 std::string invalidBackends;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000421 if (!CheckRequestedBackendsAreValid(params.m_ComputeDevices, armnn::Optional<std::string&>(invalidBackends)))
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100422 {
423 throw armnn::Exception("Some backend IDs are invalid: " + invalidBackends);
424 }
425
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100426 armnn::IOptimizedNetworkPtr optNet{nullptr, [](armnn::IOptimizedNetwork*){}};
surmeh013537c2c2018-05-18 16:31:43 +0100427 {
Finn Williams4422cec2021-03-22 17:51:06 +0000428 const auto parsing_start_time = armnn::GetTimeNow();
429 armnn::INetworkPtr network = CreateNetworkImpl<IParser>::Create(params, m_InputBindings, m_OutputBindings);
430
431 ARMNN_LOG(info) << "Network parsing time: " << std::setprecision(2)
432 << std::fixed << armnn::GetTimeDuration(parsing_start_time).count() << " ms\n";
433
surmeh013537c2c2018-05-18 16:31:43 +0100434 ARMNN_SCOPED_HEAP_PROFILING("Optimizing");
telsoa01c577f2c2018-08-31 09:22:23 +0100435
436 armnn::OptimizerOptions options;
437 options.m_ReduceFp32ToFp16 = params.m_EnableFp16TurboMode;
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000438 options.m_ReduceFp32ToBf16 = params.m_EnableBf16TurboMode;
Matthew Jackson54658b92019-08-27 15:35:59 +0100439 options.m_Debug = params.m_PrintIntermediateLayers;
telsoa01c577f2c2018-08-31 09:22:23 +0100440
Sadik Armagana25886e2020-09-15 17:17:08 +0100441 armnn::BackendOptions gpuAcc("GpuAcc",
442 {
Matthew Sloyan42432112021-01-08 10:30:51 +0000443 { "FastMathEnabled", params.m_EnableFastMath },
444 { "SaveCachedNetwork", params.m_SaveCachedNetwork },
Finn Williams40646322021-02-11 16:16:42 +0000445 { "CachedNetworkFilePath", params.m_CachedNetworkFilePath },
446 { "MLGOTuningFilePath", params.m_MLGOTuningFilePath }
Sadik Armagana25886e2020-09-15 17:17:08 +0100447 });
Finn Williams40646322021-02-11 16:16:42 +0000448
Sadik Armagana25886e2020-09-15 17:17:08 +0100449 armnn::BackendOptions cpuAcc("CpuAcc",
450 {
Matthew Sloyan0a7dc6b2021-02-10 16:50:53 +0000451 { "FastMathEnabled", params.m_EnableFastMath },
452 { "NumberOfThreads", params.m_NumberOfThreads }
Sadik Armagana25886e2020-09-15 17:17:08 +0100453 });
454 options.m_ModelOptions.push_back(gpuAcc);
455 options.m_ModelOptions.push_back(cpuAcc);
456
alered01a7227ac2020-05-07 14:58:29 +0100457 const auto optimization_start_time = armnn::GetTimeNow();
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000458 optNet = armnn::Optimize(*network, params.m_ComputeDevices, m_Runtime->GetDeviceSpec(), options);
alered01a7227ac2020-05-07 14:58:29 +0100459
460 ARMNN_LOG(info) << "Optimization time: " << std::setprecision(2)
461 << std::fixed << armnn::GetTimeDuration(optimization_start_time).count() << " ms\n";
462
telsoa01c577f2c2018-08-31 09:22:23 +0100463 if (!optNet)
464 {
465 throw armnn::Exception("Optimize returned nullptr");
466 }
Finn Williams4422cec2021-03-22 17:51:06 +0000467
468
surmeh013537c2c2018-05-18 16:31:43 +0100469 }
telsoa014fcda012018-03-09 14:13:49 +0000470
surmeh013537c2c2018-05-18 16:31:43 +0100471 if (params.m_VisualizePostOptimizationModel)
472 {
Francis Murtagh532a29d2020-06-29 11:50:01 +0100473 fs::path filename = params.m_ModelPath;
surmeh013537c2c2018-05-18 16:31:43 +0100474 filename.replace_extension("dot");
Rob Hughes9e10c2b2019-07-23 15:37:19 +0100475 std::fstream file(filename.c_str(), std::ios_base::out);
surmeh013537c2c2018-05-18 16:31:43 +0100476 optNet->SerializeToDot(file);
477 }
478
479 armnn::Status ret;
480 {
481 ARMNN_SCOPED_HEAP_PROFILING("LoadNetwork");
Matthew Sloyan4f29f152021-01-18 16:10:20 +0000482
483 const auto loading_start_time = armnn::GetTimeNow();
Francis Murtagh73d3e2e2021-04-29 14:23:04 +0100484 armnn::INetworkProperties networkProperties(params.m_AsyncEnabled,
485 armnn::MemorySource::Undefined,
Kevin Mayb4b3ac92021-05-21 16:42:21 +0100486 armnn::MemorySource::Undefined,
487 params.m_ThreadPoolSize);
Sadik Armagana04a9d72021-04-27 10:02:10 +0100488 std::string errorMessage;
489 ret = m_Runtime->LoadNetwork(m_NetworkIdentifier, std::move(optNet), errorMessage, networkProperties);
Matthew Sloyan4f29f152021-01-18 16:10:20 +0000490
491 ARMNN_LOG(info) << "Network loading time: " << std::setprecision(2)
492 << std::fixed << armnn::GetTimeDuration(loading_start_time).count() << " ms\n";
surmeh013537c2c2018-05-18 16:31:43 +0100493 }
494
telsoa014fcda012018-03-09 14:13:49 +0000495 if (ret == armnn::Status::Failure)
496 {
497 throw armnn::Exception("IRuntime::LoadNetwork failed");
498 }
499 }
500
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000501 void CheckInputIndexIsValid(unsigned int inputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000502 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000503 if (m_InputBindings.size() < inputIndex + 1)
504 {
James Ward08f40162020-09-07 16:45:07 +0100505 throw armnn::Exception(fmt::format("Input index out of range: {}", inputIndex));
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000506 }
telsoa014fcda012018-03-09 14:13:49 +0000507 }
508
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000509 void CheckOutputIndexIsValid(unsigned int outputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000510 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000511 if (m_OutputBindings.size() < outputIndex + 1)
512 {
James Ward08f40162020-09-07 16:45:07 +0100513 throw armnn::Exception(fmt::format("Output index out of range: {}", outputIndex));
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000514 }
515 }
516
Aron Virginas-Tarc82c8732019-10-24 17:07:43 +0100517 unsigned int GetInputSize(unsigned int inputIndex = 0u) const
518 {
519 CheckInputIndexIsValid(inputIndex);
520 return m_InputBindings[inputIndex].second.GetNumElements();
521 }
522
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000523 unsigned int GetOutputSize(unsigned int outputIndex = 0u) const
524 {
525 CheckOutputIndexIsValid(outputIndex);
526 return m_OutputBindings[outputIndex].second.GetNumElements();
527 }
528
James Conroy7b4886f2019-04-11 10:23:58 +0100529 std::chrono::duration<double, std::milli> Run(
530 const std::vector<TContainer>& inputContainers,
531 std::vector<TContainer>& outputContainers)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000532 {
Ferran Balaguerc602f292019-02-08 17:09:55 +0000533 for (unsigned int i = 0; i < outputContainers.size(); ++i)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000534 {
535 const unsigned int expectedOutputDataSize = GetOutputSize(i);
Ferran Balaguerc602f292019-02-08 17:09:55 +0000536
James Ward6d9f5c52020-09-28 11:56:35 +0100537 mapbox::util::apply_visitor([expectedOutputDataSize, i](auto&& value)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000538 {
Matthew Sloyan80c6b142020-09-08 12:00:32 +0100539 const unsigned int actualOutputDataSize = armnn::numeric_cast<unsigned int>(value.size());
Ferran Balaguerc602f292019-02-08 17:09:55 +0000540 if (actualOutputDataSize < expectedOutputDataSize)
541 {
Matthew Sloyan80c6b142020-09-08 12:00:32 +0100542 unsigned int outputIndex = i;
Ferran Balaguerc602f292019-02-08 17:09:55 +0000543 throw armnn::Exception(
James Ward08f40162020-09-07 16:45:07 +0100544 fmt::format("Not enough data for output #{0}: expected "
545 "{1} elements, got {2}", outputIndex, expectedOutputDataSize, actualOutputDataSize));
Ferran Balaguerc602f292019-02-08 17:09:55 +0000546 }
547 },
548 outputContainers[i]);
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000549 }
telsoa01c577f2c2018-08-31 09:22:23 +0100550
551 std::shared_ptr<armnn::IProfiler> profiler = m_Runtime->GetProfiler(m_NetworkIdentifier);
552 if (profiler)
553 {
554 profiler->EnableProfiling(m_EnableProfiling);
555 }
556
James Conroy7b4886f2019-04-11 10:23:58 +0100557 // Start timer to record inference time in EnqueueWorkload (in milliseconds)
alered01a7227ac2020-05-07 14:58:29 +0100558 const auto start_time = armnn::GetTimeNow();
James Conroy7b4886f2019-04-11 10:23:58 +0100559
telsoa014fcda012018-03-09 14:13:49 +0000560 armnn::Status ret = m_Runtime->EnqueueWorkload(m_NetworkIdentifier,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000561 MakeInputTensors(inputContainers),
562 MakeOutputTensors(outputContainers));
alered01a7227ac2020-05-07 14:58:29 +0100563 const auto duration = armnn::GetTimeDuration(start_time);
James Conroy7b4886f2019-04-11 10:23:58 +0100564
Sadik Armagan2b7a1582018-09-05 16:33:58 +0100565 // if profiling is enabled print out the results
566 if (profiler && profiler->IsProfilingEnabled())
567 {
568 profiler->Print(std::cout);
569 }
570
telsoa014fcda012018-03-09 14:13:49 +0000571 if (ret == armnn::Status::Failure)
572 {
573 throw armnn::Exception("IRuntime::EnqueueWorkload failed");
574 }
James Conroy7b4886f2019-04-11 10:23:58 +0100575 else
576 {
alered01a7227ac2020-05-07 14:58:29 +0100577 return duration;
James Conroy7b4886f2019-04-11 10:23:58 +0100578 }
telsoa014fcda012018-03-09 14:13:49 +0000579 }
580
Sadik Armagana04a9d72021-04-27 10:02:10 +0100581 std::tuple<armnn::profiling::ProfilingGuid, std::chrono::duration<double, std::milli>> RunAsync(
582 armnn::experimental::IWorkingMemHandle& workingMemHandleRef,
583 const std::vector<TContainer>& inputContainers,
584 std::vector<TContainer>& outputContainers)
585 {
586 for (unsigned int i = 0; i < outputContainers.size(); ++i)
587 {
588 const unsigned int expectedOutputDataSize = GetOutputSize(i);
589
590 mapbox::util::apply_visitor([expectedOutputDataSize, i](auto&& value)
591 {
592 const unsigned int actualOutputDataSize = armnn::numeric_cast<unsigned int>(value.size());
593 if (actualOutputDataSize < expectedOutputDataSize)
594 {
595 unsigned int outputIndex = i;
596 throw armnn::Exception(
597 fmt::format("Not enough data for output #{0}: expected "
598 "{1} elements, got {2}", outputIndex, expectedOutputDataSize, actualOutputDataSize));
599 }
600 },
601 outputContainers[i]);
602 }
603
604 std::shared_ptr<armnn::IProfiler> profiler = m_Runtime->GetProfiler(m_NetworkIdentifier);
605 if (profiler)
606 {
607 profiler->EnableProfiling(m_EnableProfiling);
608 }
609
610 // Start timer to record inference time in EnqueueWorkload (in milliseconds)
611 const auto start_time = armnn::GetTimeNow();
612
613 armnn::Status ret = m_Runtime->Execute(workingMemHandleRef,
614 MakeInputTensors(inputContainers),
615 MakeOutputTensors(outputContainers));
616 auto inferenceID = workingMemHandleRef.GetInferenceId();
617
618 const auto duration = armnn::GetTimeDuration(start_time);
619
620 // if profiling is enabled print out the results
621 if (profiler && profiler->IsProfilingEnabled())
622 {
623 profiler->Print(std::cout);
624 }
625
626 if (ret == armnn::Status::Failure)
627 {
628 throw armnn::Exception(
629 fmt::format("IRuntime::Execute asynchronously failed for network #{0} on inference #{1}",
630 m_NetworkIdentifier, inferenceID));
631 }
632 else
633 {
634 return std::make_tuple(inferenceID, duration);
635 }
636 }
637
Kevin Mayb4b3ac92021-05-21 16:42:21 +0100638 void RunAsync(const std::vector<TContainer>& inputContainers,
639 std::vector<TContainer>& outputContainers,
640 armnn::experimental::IAsyncExecutionCallbackPtr cb)
641 {
642 for (unsigned int i = 0; i < outputContainers.size(); ++i)
643 {
644 const unsigned int expectedOutputDataSize = GetOutputSize(i);
645
646 mapbox::util::apply_visitor([expectedOutputDataSize, i](auto&& value)
647 {
648 const unsigned int actualOutputDataSize = armnn::numeric_cast<unsigned int>(value.size());
649 if (actualOutputDataSize < expectedOutputDataSize)
650 {
651 unsigned int outputIndex = i;
652 throw armnn::Exception(
653 fmt::format("Not enough data for output #{0}: expected "
654 "{1} elements, got {2}", outputIndex, expectedOutputDataSize, actualOutputDataSize));
655 }
656 },
657 outputContainers[i]);
658 }
659
660 std::shared_ptr<armnn::IProfiler> profiler = m_Runtime->GetProfiler(m_NetworkIdentifier);
661 if (profiler)
662 {
663 profiler->EnableProfiling(m_EnableProfiling);
664 }
665
666 m_Runtime->Schedule(m_NetworkIdentifier,
667 MakeInputTensors(inputContainers),
668 MakeOutputTensors(outputContainers),
669 armnn::QosExecPriority::Medium,
670 cb);
671
672 // if profiling is enabled print out the results
673 if (profiler && profiler->IsProfilingEnabled())
674 {
675 profiler->Print(std::cout);
676 }
677 }
678
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100679 const armnn::BindingPointInfo& GetInputBindingInfo(unsigned int inputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100680 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000681 CheckInputIndexIsValid(inputIndex);
682 return m_InputBindings[inputIndex];
telsoa01c577f2c2018-08-31 09:22:23 +0100683 }
684
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100685 const std::vector<armnn::BindingPointInfo>& GetInputBindingInfos() const
telsoa01c577f2c2018-08-31 09:22:23 +0100686 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000687 return m_InputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100688 }
689
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100690 const armnn::BindingPointInfo& GetOutputBindingInfo(unsigned int outputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100691 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000692 CheckOutputIndexIsValid(outputIndex);
693 return m_OutputBindings[outputIndex];
694 }
695
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100696 const std::vector<armnn::BindingPointInfo>& GetOutputBindingInfos() const
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000697 {
698 return m_OutputBindings;
699 }
700
701 QuantizationParams GetQuantizationParams(unsigned int outputIndex = 0u) const
702 {
703 CheckOutputIndexIsValid(outputIndex);
704 return std::make_pair(m_OutputBindings[outputIndex].second.GetQuantizationScale(),
705 m_OutputBindings[outputIndex].second.GetQuantizationOffset());
706 }
707
Narumol Prangnawarat4628d052019-02-25 17:26:05 +0000708 QuantizationParams GetInputQuantizationParams(unsigned int inputIndex = 0u) const
709 {
710 CheckInputIndexIsValid(inputIndex);
711 return std::make_pair(m_InputBindings[inputIndex].second.GetQuantizationScale(),
712 m_InputBindings[inputIndex].second.GetQuantizationOffset());
713 }
714
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000715 std::vector<QuantizationParams> GetAllQuantizationParams() const
716 {
717 std::vector<QuantizationParams> quantizationParams;
718 for (unsigned int i = 0u; i < m_OutputBindings.size(); i++)
719 {
720 quantizationParams.push_back(GetQuantizationParams(i));
721 }
722 return quantizationParams;
telsoa01c577f2c2018-08-31 09:22:23 +0100723 }
724
Sadik Armagana04a9d72021-04-27 10:02:10 +0100725 std::unique_ptr<armnn::experimental::IWorkingMemHandle> CreateWorkingMemHandle()
726 {
727 return m_Runtime->CreateWorkingMemHandle(m_NetworkIdentifier);
728 }
729
telsoa014fcda012018-03-09 14:13:49 +0000730private:
telsoa01c577f2c2018-08-31 09:22:23 +0100731 armnn::NetworkId m_NetworkIdentifier;
732 std::shared_ptr<armnn::IRuntime> m_Runtime;
733
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100734 std::vector<armnn::BindingPointInfo> m_InputBindings;
735 std::vector<armnn::BindingPointInfo> m_OutputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100736 bool m_EnableProfiling;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100737 std::string m_DynamicBackendsPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100738
telsoa014fcda012018-03-09 14:13:49 +0000739 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000740 armnn::InputTensors MakeInputTensors(const std::vector<TContainer>& inputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000741 {
Jim Flynn2fd61002019-05-03 12:54:26 +0100742 return armnnUtils::MakeInputTensors(m_InputBindings, inputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000743 }
744
745 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000746 armnn::OutputTensors MakeOutputTensors(std::vector<TContainer>& outputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000747 {
Jim Flynn2fd61002019-05-03 12:54:26 +0100748 return armnnUtils::MakeOutputTensors(m_OutputBindings, outputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000749 }
Ferran Balaguerc602f292019-02-08 17:09:55 +0000750};