blob: 6e73f5275fc9fa5bfdfccea4bfb34975eac227c9 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#pragma once
David Beckf0b48452018-10-19 15:20:56 +01006#include <armnn/ArmNN.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +01007
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +00008#if defined(ARMNN_SERIALIZER)
Derek Lamberti0028d1b2019-02-20 13:57:42 +00009#include "armnnDeserializer/IDeserializer.hpp"
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000010#endif
telsoa01c577f2c2018-08-31 09:22:23 +010011#if defined(ARMNN_TF_LITE_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010012#include <armnnTfLiteParser/ITfLiteParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010013#endif
telsoa01c577f2c2018-08-31 09:22:23 +010014#if defined(ARMNN_ONNX_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010015#include <armnnOnnxParser/IOnnxParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010016#endif
telsoa014fcda012018-03-09 14:13:49 +000017
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000018#include <HeapProfiling.hpp>
19
David Beck1b61be52018-11-08 09:19:14 +000020#include <backendsCommon/BackendRegistry.hpp>
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010021
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000022#include <boost/algorithm/string/join.hpp>
surmeh013537c2c2018-05-18 16:31:43 +010023#include <boost/exception/exception.hpp>
24#include <boost/exception/diagnostic_information.hpp>
telsoa014fcda012018-03-09 14:13:49 +000025#include <boost/log/trivial.hpp>
26#include <boost/format.hpp>
27#include <boost/program_options.hpp>
surmeh013537c2c2018-05-18 16:31:43 +010028#include <boost/filesystem.hpp>
David Beckf0b48452018-10-19 15:20:56 +010029#include <boost/lexical_cast.hpp>
Ferran Balaguerc602f292019-02-08 17:09:55 +000030#include <boost/variant.hpp>
telsoa014fcda012018-03-09 14:13:49 +000031
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000032#include <algorithm>
33#include <iterator>
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010034#include <fstream>
telsoa014fcda012018-03-09 14:13:49 +000035#include <map>
36#include <string>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000037#include <vector>
telsoa01c577f2c2018-08-31 09:22:23 +010038#include <type_traits>
39
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010040namespace
41{
42
43inline bool CheckRequestedBackendsAreValid(const std::vector<armnn::BackendId>& backendIds,
44 armnn::Optional<std::string&> invalidBackendIds = armnn::EmptyOptional())
45{
46 if (backendIds.empty())
47 {
48 return false;
49 }
50
51 armnn::BackendIdSet validBackendIds = armnn::BackendRegistryInstance().GetBackendIds();
52
53 bool allValid = true;
54 for (const auto& backendId : backendIds)
55 {
56 if (std::find(validBackendIds.begin(), validBackendIds.end(), backendId) == validBackendIds.end())
57 {
58 allValid = false;
59 if (invalidBackendIds)
60 {
61 if (!invalidBackendIds.value().empty())
62 {
63 invalidBackendIds.value() += ", ";
64 }
65 invalidBackendIds.value() += backendId;
66 }
67 }
68 }
69 return allValid;
70}
71
72} // anonymous namespace
73
telsoa01c577f2c2018-08-31 09:22:23 +010074namespace InferenceModelInternal
75{
76// This needs to go when the armnnCaffeParser, armnnTfParser and armnnTfLiteParser
77// definitions of BindingPointInfo gets consolidated.
78using BindingPointInfo = std::pair<armnn::LayerBindingId, armnn::TensorInfo>;
79
80using QuantizationParams = std::pair<float,int32_t>;
81
82struct Params
83{
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000084 std::string m_ModelPath;
85 std::vector<std::string> m_InputBindings;
86 std::vector<armnn::TensorShape> m_InputShapes;
87 std::vector<std::string> m_OutputBindings;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000088 std::vector<armnn::BackendId> m_ComputeDevices;
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000089 bool m_EnableProfiling;
90 size_t m_SubgraphId;
91 bool m_IsModelBinary;
92 bool m_VisualizePostOptimizationModel;
93 bool m_EnableFp16TurboMode;
telsoa01c577f2c2018-08-31 09:22:23 +010094
95 Params()
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000096 : m_ComputeDevices{"CpuRef"}
telsoa01c577f2c2018-08-31 09:22:23 +010097 , m_EnableProfiling(false)
98 , m_SubgraphId(0)
99 , m_IsModelBinary(true)
100 , m_VisualizePostOptimizationModel(false)
101 , m_EnableFp16TurboMode(false)
102 {}
103};
104
105} // namespace InferenceModelInternal
106
107template <typename IParser>
108struct CreateNetworkImpl
109{
110public:
111 using Params = InferenceModelInternal::Params;
112 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
113
114 static armnn::INetworkPtr Create(const Params& params,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000115 std::vector<BindingPointInfo>& inputBindings,
116 std::vector<BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100117 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000118 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100119
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000120 // Create a network from a file on disk
121 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100122
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000123 std::map<std::string, armnn::TensorShape> inputShapes;
124 if (!params.m_InputShapes.empty())
125 {
126 const size_t numInputShapes = params.m_InputShapes.size();
127 const size_t numInputBindings = params.m_InputBindings.size();
128 if (numInputShapes < numInputBindings)
129 {
130 throw armnn::Exception(boost::str(boost::format(
131 "Not every input has its tensor shape specified: expected=%1%, got=%2%")
132 % numInputBindings % numInputShapes));
133 }
telsoa01c577f2c2018-08-31 09:22:23 +0100134
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000135 for (size_t i = 0; i < numInputShapes; i++)
136 {
137 inputShapes[params.m_InputBindings[i]] = params.m_InputShapes[i];
138 }
139 }
telsoa01c577f2c2018-08-31 09:22:23 +0100140
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000141 std::vector<std::string> requestedOutputs = params.m_OutputBindings;
142 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
143
144 {
145 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
146 // Handle text and binary input differently by calling the corresponding parser function
147 network = (params.m_IsModelBinary ?
148 parser->CreateNetworkFromBinaryFile(modelPath.c_str(), inputShapes, requestedOutputs) :
149 parser->CreateNetworkFromTextFile(modelPath.c_str(), inputShapes, requestedOutputs));
150 }
151
152 for (const std::string& inputLayerName : params.m_InputBindings)
153 {
154 inputBindings.push_back(parser->GetNetworkInputBindingInfo(inputLayerName));
155 }
156
157 for (const std::string& outputLayerName : params.m_OutputBindings)
158 {
159 outputBindings.push_back(parser->GetNetworkOutputBindingInfo(outputLayerName));
160 }
161
162 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100163 }
164};
165
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000166#if defined(ARMNN_SERIALIZER)
167template <>
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000168struct CreateNetworkImpl<armnnDeserializer::IDeserializer>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000169{
170public:
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000171 using IParser = armnnDeserializer::IDeserializer;
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000172 using Params = InferenceModelInternal::Params;
173 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
174
175 static armnn::INetworkPtr Create(const Params& params,
176 std::vector<BindingPointInfo>& inputBindings,
177 std::vector<BindingPointInfo>& outputBindings)
178 {
179 auto parser(IParser::Create());
180 BOOST_ASSERT(parser);
181
182 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
183
184 {
185 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
Derek Lamberti2b183fb2019-02-18 16:36:57 +0000186
187 boost::system::error_code errorCode;
188 boost::filesystem::path pathToFile(params.m_ModelPath);
189 if (!boost::filesystem::exists(pathToFile, errorCode))
190 {
191 throw armnn::FileNotFoundException(boost::str(
192 boost::format("Cannot find the file (%1%) errorCode: %2% %3%") %
193 params.m_ModelPath %
194 errorCode %
195 CHECK_LOCATION().AsString()));
196 }
197 std::ifstream file(params.m_ModelPath, std::ios::binary);
198
199 network = parser->CreateNetworkFromBinary(file);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000200 }
201
202 unsigned int subGraphId = boost::numeric_cast<unsigned int>(params.m_SubgraphId);
203
204 for (const std::string& inputLayerName : params.m_InputBindings)
205 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000206 armnnDeserializer::BindingPointInfo inputBinding =
207 parser->GetNetworkInputBindingInfo(subGraphId, inputLayerName);
208 inputBindings.push_back(std::make_pair(inputBinding.m_BindingId, inputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000209 }
210
211 for (const std::string& outputLayerName : params.m_OutputBindings)
212 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000213 armnnDeserializer::BindingPointInfo outputBinding =
214 parser->GetNetworkOutputBindingInfo(subGraphId, outputLayerName);
215 outputBindings.push_back(std::make_pair(outputBinding.m_BindingId, outputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000216 }
217
218 return network;
219 }
220};
221#endif
222
telsoa01c577f2c2018-08-31 09:22:23 +0100223#if defined(ARMNN_TF_LITE_PARSER)
224template <>
225struct CreateNetworkImpl<armnnTfLiteParser::ITfLiteParser>
226{
227public:
228 using IParser = armnnTfLiteParser::ITfLiteParser;
229 using Params = InferenceModelInternal::Params;
230 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
231
232 static armnn::INetworkPtr Create(const Params& params,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000233 std::vector<BindingPointInfo>& inputBindings,
234 std::vector<BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100235 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000236 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100237
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000238 // Create a network from a file on disk
239 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100240
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000241 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100242
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000243 {
244 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
245 network = parser->CreateNetworkFromBinaryFile(modelPath.c_str());
246 }
telsoa01c577f2c2018-08-31 09:22:23 +0100247
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000248 for (const std::string& inputLayerName : params.m_InputBindings)
249 {
250 BindingPointInfo inputBinding =
251 parser->GetNetworkInputBindingInfo(params.m_SubgraphId, inputLayerName);
252 inputBindings.push_back(inputBinding);
253 }
254
255 for (const std::string& outputLayerName : params.m_OutputBindings)
256 {
257 BindingPointInfo outputBinding =
258 parser->GetNetworkOutputBindingInfo(params.m_SubgraphId, outputLayerName);
259 outputBindings.push_back(outputBinding);
260 }
261
262 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100263 }
264};
265#endif
266
267#if defined(ARMNN_ONNX_PARSER)
268template <>
269struct CreateNetworkImpl<armnnOnnxParser::IOnnxParser>
270{
271public:
272 using IParser = armnnOnnxParser::IOnnxParser;
273 using Params = InferenceModelInternal::Params;
274 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
275
276 static armnn::INetworkPtr Create(const Params& params,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000277 std::vector<BindingPointInfo>& inputBindings,
278 std::vector<BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100279 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000280 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100281
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000282 // Create a network from a file on disk
283 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100284
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000285 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100286
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000287 {
288 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
289 network = (params.m_IsModelBinary ?
290 parser->CreateNetworkFromBinaryFile(modelPath.c_str()) :
291 parser->CreateNetworkFromTextFile(modelPath.c_str()));
292 }
telsoa01c577f2c2018-08-31 09:22:23 +0100293
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000294 for (const std::string& inputLayerName : params.m_InputBindings)
295 {
296 BindingPointInfo inputBinding = parser->GetNetworkInputBindingInfo(inputLayerName);
297 inputBindings.push_back(inputBinding);
298 }
299
300 for (const std::string& outputLayerName : params.m_OutputBindings)
301 {
302 BindingPointInfo outputBinding = parser->GetNetworkOutputBindingInfo(outputLayerName);
303 outputBindings.push_back(outputBinding);
304 }
305
306 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100307 }
308};
309#endif
telsoa014fcda012018-03-09 14:13:49 +0000310
311template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000312inline armnn::InputTensors MakeInputTensors(
313 const std::vector<InferenceModelInternal::BindingPointInfo>& inputBindings,
314 const std::vector<TContainer>& inputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000315{
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000316 armnn::InputTensors inputTensors;
317
318 const size_t numInputs = inputBindings.size();
319 if (numInputs != inputDataContainers.size())
telsoa014fcda012018-03-09 14:13:49 +0000320 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000321 throw armnn::Exception(boost::str(boost::format("Number of inputs does not match number of "
322 "tensor data containers: %1% != %2%") % numInputs % inputDataContainers.size()));
telsoa014fcda012018-03-09 14:13:49 +0000323 }
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000324
325 for (size_t i = 0; i < numInputs; i++)
326 {
327 const InferenceModelInternal::BindingPointInfo& inputBinding = inputBindings[i];
328 const TContainer& inputData = inputDataContainers[i];
329
Ferran Balaguerc602f292019-02-08 17:09:55 +0000330 boost::apply_visitor([&](auto&& value)
331 {
332 if (value.size() != inputBinding.second.GetNumElements())
333 {
334 throw armnn::Exception("Input tensor has incorrect size");
335 }
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000336
Ferran Balaguerc602f292019-02-08 17:09:55 +0000337 armnn::ConstTensor inputTensor(inputBinding.second, value.data());
338 inputTensors.push_back(std::make_pair(inputBinding.first, inputTensor));
339 },
340 inputData);
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000341 }
342
343 return inputTensors;
telsoa014fcda012018-03-09 14:13:49 +0000344}
345
346template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000347inline armnn::OutputTensors MakeOutputTensors(
348 const std::vector<InferenceModelInternal::BindingPointInfo>& outputBindings,
349 std::vector<TContainer>& outputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000350{
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000351 armnn::OutputTensors outputTensors;
352
353 const size_t numOutputs = outputBindings.size();
354 if (numOutputs != outputDataContainers.size())
telsoa014fcda012018-03-09 14:13:49 +0000355 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000356 throw armnn::Exception(boost::str(boost::format("Number of outputs does not match number of "
357 "tensor data containers: %1% != %2%") % numOutputs % outputDataContainers.size()));
telsoa014fcda012018-03-09 14:13:49 +0000358 }
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000359
360 for (size_t i = 0; i < numOutputs; i++)
361 {
362 const InferenceModelInternal::BindingPointInfo& outputBinding = outputBindings[i];
363 TContainer& outputData = outputDataContainers[i];
364
Ferran Balaguerc602f292019-02-08 17:09:55 +0000365 boost::apply_visitor([&](auto&& value)
366 {
367 if (value.size() != outputBinding.second.GetNumElements())
368 {
369 throw armnn::Exception("Output tensor has incorrect size");
370 }
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000371
Ferran Balaguerc602f292019-02-08 17:09:55 +0000372 armnn::Tensor outputTensor(outputBinding.second, value.data());
373 outputTensors.push_back(std::make_pair(outputBinding.first, outputTensor));
374 },
375 outputData);
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000376 }
377
378 return outputTensors;
telsoa014fcda012018-03-09 14:13:49 +0000379}
380
381template <typename IParser, typename TDataType>
382class InferenceModel
383{
384public:
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000385 using DataType = TDataType;
386 using Params = InferenceModelInternal::Params;
387 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
388 using QuantizationParams = InferenceModelInternal::QuantizationParams;
Ferran Balaguerc602f292019-02-08 17:09:55 +0000389 using TContainer = boost::variant<std::vector<float>, std::vector<int>, std::vector<unsigned char>>;
telsoa014fcda012018-03-09 14:13:49 +0000390
391 struct CommandLineOptions
392 {
393 std::string m_ModelDir;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000394 std::vector<std::string> m_ComputeDevices;
surmeh013537c2c2018-05-18 16:31:43 +0100395 bool m_VisualizePostOptimizationModel;
telsoa01c577f2c2018-08-31 09:22:23 +0100396 bool m_EnableFp16TurboMode;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000397
398 std::vector<armnn::BackendId> GetComputeDevicesAsBackendIds()
399 {
400 std::vector<armnn::BackendId> backendIds;
401 std::copy(m_ComputeDevices.begin(), m_ComputeDevices.end(), std::back_inserter(backendIds));
402 return backendIds;
403 }
telsoa014fcda012018-03-09 14:13:49 +0000404 };
405
406 static void AddCommandLineOptions(boost::program_options::options_description& desc, CommandLineOptions& options)
407 {
408 namespace po = boost::program_options;
409
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000410 const std::vector<std::string> defaultComputes = { "CpuAcc", "CpuRef" };
David Beckf0b48452018-10-19 15:20:56 +0100411
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100412 const std::string backendsMessage = "Which device to run layers on by default. Possible choices: "
413 + armnn::BackendRegistryInstance().GetBackendIdsAsString();
414
telsoa014fcda012018-03-09 14:13:49 +0000415 desc.add_options()
416 ("model-dir,m", po::value<std::string>(&options.m_ModelDir)->required(),
telsoa01c577f2c2018-08-31 09:22:23 +0100417 "Path to directory containing model files (.caffemodel/.prototxt/.tflite)")
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000418 ("compute,c", po::value<std::vector<std::string>>(&options.m_ComputeDevices)->
419 default_value(defaultComputes, boost::algorithm::join(defaultComputes, ", "))->
420 multitoken(), backendsMessage.c_str())
surmeh013537c2c2018-05-18 16:31:43 +0100421 ("visualize-optimized-model,v",
422 po::value<bool>(&options.m_VisualizePostOptimizationModel)->default_value(false),
423 "Produce a dot file useful for visualizing the graph post optimization."
telsoa01c577f2c2018-08-31 09:22:23 +0100424 "The file will have the same name as the model with the .dot extention.")
425 ("fp16-turbo-mode", po::value<bool>(&options.m_EnableFp16TurboMode)->default_value(false),
426 "If this option is enabled FP32 layers, weights and biases will be converted "
427 "to FP16 where the backend supports it.");
telsoa014fcda012018-03-09 14:13:49 +0000428 }
429
telsoa01c577f2c2018-08-31 09:22:23 +0100430 InferenceModel(const Params& params, const std::shared_ptr<armnn::IRuntime>& runtime = nullptr)
431 : m_EnableProfiling(params.m_EnableProfiling)
telsoa014fcda012018-03-09 14:13:49 +0000432 {
telsoa01c577f2c2018-08-31 09:22:23 +0100433 if (runtime)
telsoa014fcda012018-03-09 14:13:49 +0000434 {
telsoa01c577f2c2018-08-31 09:22:23 +0100435 m_Runtime = runtime;
telsoa014fcda012018-03-09 14:13:49 +0000436 }
telsoa01c577f2c2018-08-31 09:22:23 +0100437 else
telsoa014fcda012018-03-09 14:13:49 +0000438 {
telsoa01c577f2c2018-08-31 09:22:23 +0100439 armnn::IRuntime::CreationOptions options;
Nina Drozd549ae372018-09-10 14:26:44 +0100440 options.m_EnableGpuProfiling = m_EnableProfiling;
telsoa01c577f2c2018-08-31 09:22:23 +0100441 m_Runtime = std::move(armnn::IRuntime::Create(options));
surmeh013537c2c2018-05-18 16:31:43 +0100442 }
telsoa014fcda012018-03-09 14:13:49 +0000443
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100444 std::string invalidBackends;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000445 if (!CheckRequestedBackendsAreValid(params.m_ComputeDevices, armnn::Optional<std::string&>(invalidBackends)))
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100446 {
447 throw armnn::Exception("Some backend IDs are invalid: " + invalidBackends);
448 }
449
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000450 armnn::INetworkPtr network =
451 CreateNetworkImpl<IParser>::Create(params, m_InputBindings, m_OutputBindings);
telsoa014fcda012018-03-09 14:13:49 +0000452
surmeh013537c2c2018-05-18 16:31:43 +0100453 armnn::IOptimizedNetworkPtr optNet{nullptr, [](armnn::IOptimizedNetwork *){}};
454 {
455 ARMNN_SCOPED_HEAP_PROFILING("Optimizing");
telsoa01c577f2c2018-08-31 09:22:23 +0100456
457 armnn::OptimizerOptions options;
458 options.m_ReduceFp32ToFp16 = params.m_EnableFp16TurboMode;
459
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000460 optNet = armnn::Optimize(*network, params.m_ComputeDevices, m_Runtime->GetDeviceSpec(), options);
telsoa01c577f2c2018-08-31 09:22:23 +0100461 if (!optNet)
462 {
463 throw armnn::Exception("Optimize returned nullptr");
464 }
surmeh013537c2c2018-05-18 16:31:43 +0100465 }
telsoa014fcda012018-03-09 14:13:49 +0000466
surmeh013537c2c2018-05-18 16:31:43 +0100467 if (params.m_VisualizePostOptimizationModel)
468 {
469 boost::filesystem::path filename = params.m_ModelPath;
470 filename.replace_extension("dot");
471 std::fstream file(filename.c_str(),file.out);
472 optNet->SerializeToDot(file);
473 }
474
475 armnn::Status ret;
476 {
477 ARMNN_SCOPED_HEAP_PROFILING("LoadNetwork");
478 ret = m_Runtime->LoadNetwork(m_NetworkIdentifier, std::move(optNet));
479 }
480
telsoa014fcda012018-03-09 14:13:49 +0000481 if (ret == armnn::Status::Failure)
482 {
483 throw armnn::Exception("IRuntime::LoadNetwork failed");
484 }
485 }
486
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000487 void CheckInputIndexIsValid(unsigned int inputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000488 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000489 if (m_InputBindings.size() < inputIndex + 1)
490 {
491 throw armnn::Exception(boost::str(boost::format("Input index out of range: %1%") % inputIndex));
492 }
telsoa014fcda012018-03-09 14:13:49 +0000493 }
494
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000495 void CheckOutputIndexIsValid(unsigned int outputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000496 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000497 if (m_OutputBindings.size() < outputIndex + 1)
498 {
499 throw armnn::Exception(boost::str(boost::format("Output index out of range: %1%") % outputIndex));
500 }
501 }
502
503 unsigned int GetOutputSize(unsigned int outputIndex = 0u) const
504 {
505 CheckOutputIndexIsValid(outputIndex);
506 return m_OutputBindings[outputIndex].second.GetNumElements();
507 }
508
509 void Run(const std::vector<TContainer>& inputContainers, std::vector<TContainer>& outputContainers)
510 {
Ferran Balaguerc602f292019-02-08 17:09:55 +0000511 for (unsigned int i = 0; i < outputContainers.size(); ++i)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000512 {
513 const unsigned int expectedOutputDataSize = GetOutputSize(i);
Ferran Balaguerc602f292019-02-08 17:09:55 +0000514
515 boost::apply_visitor([expectedOutputDataSize, i](auto&& value)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000516 {
Ferran Balaguerc602f292019-02-08 17:09:55 +0000517 const unsigned int actualOutputDataSize = boost::numeric_cast<unsigned int>(value.size());
518 if (actualOutputDataSize < expectedOutputDataSize)
519 {
520 unsigned int outputIndex = boost::numeric_cast<unsigned int>(i);
521 throw armnn::Exception(
522 boost::str(boost::format("Not enough data for output #%1%: expected "
523 "%2% elements, got %3%") % outputIndex % expectedOutputDataSize % actualOutputDataSize));
524 }
525 },
526 outputContainers[i]);
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000527 }
telsoa01c577f2c2018-08-31 09:22:23 +0100528
529 std::shared_ptr<armnn::IProfiler> profiler = m_Runtime->GetProfiler(m_NetworkIdentifier);
530 if (profiler)
531 {
532 profiler->EnableProfiling(m_EnableProfiling);
533 }
534
telsoa014fcda012018-03-09 14:13:49 +0000535 armnn::Status ret = m_Runtime->EnqueueWorkload(m_NetworkIdentifier,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000536 MakeInputTensors(inputContainers),
537 MakeOutputTensors(outputContainers));
Sadik Armagan2b7a1582018-09-05 16:33:58 +0100538
539 // if profiling is enabled print out the results
540 if (profiler && profiler->IsProfilingEnabled())
541 {
542 profiler->Print(std::cout);
543 }
544
telsoa014fcda012018-03-09 14:13:49 +0000545 if (ret == armnn::Status::Failure)
546 {
547 throw armnn::Exception("IRuntime::EnqueueWorkload failed");
548 }
549 }
550
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000551 const BindingPointInfo& GetInputBindingInfo(unsigned int inputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100552 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000553 CheckInputIndexIsValid(inputIndex);
554 return m_InputBindings[inputIndex];
telsoa01c577f2c2018-08-31 09:22:23 +0100555 }
556
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000557 const std::vector<BindingPointInfo>& GetInputBindingInfos() const
telsoa01c577f2c2018-08-31 09:22:23 +0100558 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000559 return m_InputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100560 }
561
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000562 const BindingPointInfo& GetOutputBindingInfo(unsigned int outputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100563 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000564 CheckOutputIndexIsValid(outputIndex);
565 return m_OutputBindings[outputIndex];
566 }
567
568 const std::vector<BindingPointInfo>& GetOutputBindingInfos() const
569 {
570 return m_OutputBindings;
571 }
572
573 QuantizationParams GetQuantizationParams(unsigned int outputIndex = 0u) const
574 {
575 CheckOutputIndexIsValid(outputIndex);
576 return std::make_pair(m_OutputBindings[outputIndex].second.GetQuantizationScale(),
577 m_OutputBindings[outputIndex].second.GetQuantizationOffset());
578 }
579
580 std::vector<QuantizationParams> GetAllQuantizationParams() const
581 {
582 std::vector<QuantizationParams> quantizationParams;
583 for (unsigned int i = 0u; i < m_OutputBindings.size(); i++)
584 {
585 quantizationParams.push_back(GetQuantizationParams(i));
586 }
587 return quantizationParams;
telsoa01c577f2c2018-08-31 09:22:23 +0100588 }
589
telsoa014fcda012018-03-09 14:13:49 +0000590private:
telsoa01c577f2c2018-08-31 09:22:23 +0100591 armnn::NetworkId m_NetworkIdentifier;
592 std::shared_ptr<armnn::IRuntime> m_Runtime;
593
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000594 std::vector<InferenceModelInternal::BindingPointInfo> m_InputBindings;
595 std::vector<InferenceModelInternal::BindingPointInfo> m_OutputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100596 bool m_EnableProfiling;
597
telsoa014fcda012018-03-09 14:13:49 +0000598 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000599 armnn::InputTensors MakeInputTensors(const std::vector<TContainer>& inputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000600 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000601 return ::MakeInputTensors(m_InputBindings, inputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000602 }
603
604 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000605 armnn::OutputTensors MakeOutputTensors(std::vector<TContainer>& outputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000606 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000607 return ::MakeOutputTensors(m_OutputBindings, outputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000608 }
Ferran Balaguerc602f292019-02-08 17:09:55 +0000609};