blob: 0ede8983cf8acd4efb4fcf5c0ef13124d0425091 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#pragma once
David Beckf0b48452018-10-19 15:20:56 +01006#include <armnn/ArmNN.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +01007
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +00008#if defined(ARMNN_SERIALIZER)
Derek Lamberti0028d1b2019-02-20 13:57:42 +00009#include "armnnDeserializer/IDeserializer.hpp"
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000010#endif
telsoa01c577f2c2018-08-31 09:22:23 +010011#if defined(ARMNN_TF_LITE_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010012#include <armnnTfLiteParser/ITfLiteParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010013#endif
telsoa01c577f2c2018-08-31 09:22:23 +010014#if defined(ARMNN_ONNX_PARSER)
David Beckf0b48452018-10-19 15:20:56 +010015#include <armnnOnnxParser/IOnnxParser.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010016#endif
telsoa014fcda012018-03-09 14:13:49 +000017
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000018#include <HeapProfiling.hpp>
Jim Flynn2fd61002019-05-03 12:54:26 +010019#include <TensorIOUtils.hpp>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000020
David Beck1b61be52018-11-08 09:19:14 +000021#include <backendsCommon/BackendRegistry.hpp>
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010022
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000023#include <boost/algorithm/string/join.hpp>
surmeh013537c2c2018-05-18 16:31:43 +010024#include <boost/exception/exception.hpp>
25#include <boost/exception/diagnostic_information.hpp>
telsoa014fcda012018-03-09 14:13:49 +000026#include <boost/log/trivial.hpp>
27#include <boost/format.hpp>
28#include <boost/program_options.hpp>
surmeh013537c2c2018-05-18 16:31:43 +010029#include <boost/filesystem.hpp>
David Beckf0b48452018-10-19 15:20:56 +010030#include <boost/lexical_cast.hpp>
Ferran Balaguerc602f292019-02-08 17:09:55 +000031#include <boost/variant.hpp>
telsoa014fcda012018-03-09 14:13:49 +000032
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000033#include <algorithm>
James Conroy7b4886f2019-04-11 10:23:58 +010034#include <chrono>
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000035#include <iterator>
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010036#include <fstream>
telsoa014fcda012018-03-09 14:13:49 +000037#include <map>
38#include <string>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000039#include <vector>
telsoa01c577f2c2018-08-31 09:22:23 +010040#include <type_traits>
41
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010042namespace
43{
44
45inline bool CheckRequestedBackendsAreValid(const std::vector<armnn::BackendId>& backendIds,
46 armnn::Optional<std::string&> invalidBackendIds = armnn::EmptyOptional())
47{
48 if (backendIds.empty())
49 {
50 return false;
51 }
52
53 armnn::BackendIdSet validBackendIds = armnn::BackendRegistryInstance().GetBackendIds();
54
55 bool allValid = true;
56 for (const auto& backendId : backendIds)
57 {
58 if (std::find(validBackendIds.begin(), validBackendIds.end(), backendId) == validBackendIds.end())
59 {
60 allValid = false;
61 if (invalidBackendIds)
62 {
63 if (!invalidBackendIds.value().empty())
64 {
65 invalidBackendIds.value() += ", ";
66 }
67 invalidBackendIds.value() += backendId;
68 }
69 }
70 }
71 return allValid;
72}
73
74} // anonymous namespace
75
telsoa01c577f2c2018-08-31 09:22:23 +010076namespace InferenceModelInternal
77{
Jim Flynnb4d7eae2019-05-01 14:44:27 +010078using BindingPointInfo = armnn::BindingPointInfo;
telsoa01c577f2c2018-08-31 09:22:23 +010079
80using QuantizationParams = std::pair<float,int32_t>;
81
82struct Params
83{
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000084 std::string m_ModelPath;
85 std::vector<std::string> m_InputBindings;
86 std::vector<armnn::TensorShape> m_InputShapes;
87 std::vector<std::string> m_OutputBindings;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +000088 std::vector<armnn::BackendId> m_ComputeDevices;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010089 std::string m_DynamicBackendsPath;
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +000090 size_t m_SubgraphId;
91 bool m_IsModelBinary;
92 bool m_VisualizePostOptimizationModel;
93 bool m_EnableFp16TurboMode;
Matthew Jackson54658b92019-08-27 15:35:59 +010094 bool m_PrintIntermediateLayers;
telsoa01c577f2c2018-08-31 09:22:23 +010095
96 Params()
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010097 : m_ComputeDevices{}
telsoa01c577f2c2018-08-31 09:22:23 +010098 , m_SubgraphId(0)
99 , m_IsModelBinary(true)
100 , m_VisualizePostOptimizationModel(false)
101 , m_EnableFp16TurboMode(false)
Matthew Jackson54658b92019-08-27 15:35:59 +0100102 , m_PrintIntermediateLayers(false)
telsoa01c577f2c2018-08-31 09:22:23 +0100103 {}
104};
105
106} // namespace InferenceModelInternal
107
108template <typename IParser>
109struct CreateNetworkImpl
110{
111public:
112 using Params = InferenceModelInternal::Params;
telsoa01c577f2c2018-08-31 09:22:23 +0100113
114 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100115 std::vector<armnn::BindingPointInfo>& inputBindings,
116 std::vector<armnn::BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100117 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000118 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100119
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000120 // Create a network from a file on disk
121 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100122
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000123 std::map<std::string, armnn::TensorShape> inputShapes;
124 if (!params.m_InputShapes.empty())
125 {
126 const size_t numInputShapes = params.m_InputShapes.size();
127 const size_t numInputBindings = params.m_InputBindings.size();
128 if (numInputShapes < numInputBindings)
129 {
130 throw armnn::Exception(boost::str(boost::format(
131 "Not every input has its tensor shape specified: expected=%1%, got=%2%")
132 % numInputBindings % numInputShapes));
133 }
telsoa01c577f2c2018-08-31 09:22:23 +0100134
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000135 for (size_t i = 0; i < numInputShapes; i++)
136 {
137 inputShapes[params.m_InputBindings[i]] = params.m_InputShapes[i];
138 }
139 }
telsoa01c577f2c2018-08-31 09:22:23 +0100140
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000141 std::vector<std::string> requestedOutputs = params.m_OutputBindings;
142 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
143
144 {
145 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
146 // Handle text and binary input differently by calling the corresponding parser function
147 network = (params.m_IsModelBinary ?
148 parser->CreateNetworkFromBinaryFile(modelPath.c_str(), inputShapes, requestedOutputs) :
149 parser->CreateNetworkFromTextFile(modelPath.c_str(), inputShapes, requestedOutputs));
150 }
151
152 for (const std::string& inputLayerName : params.m_InputBindings)
153 {
154 inputBindings.push_back(parser->GetNetworkInputBindingInfo(inputLayerName));
155 }
156
157 for (const std::string& outputLayerName : params.m_OutputBindings)
158 {
159 outputBindings.push_back(parser->GetNetworkOutputBindingInfo(outputLayerName));
160 }
161
162 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100163 }
164};
165
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000166#if defined(ARMNN_SERIALIZER)
167template <>
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000168struct CreateNetworkImpl<armnnDeserializer::IDeserializer>
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000169{
170public:
Derek Lamberti0028d1b2019-02-20 13:57:42 +0000171 using IParser = armnnDeserializer::IDeserializer;
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000172 using Params = InferenceModelInternal::Params;
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000173
174 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100175 std::vector<armnn::BindingPointInfo>& inputBindings,
176 std::vector<armnn::BindingPointInfo>& outputBindings)
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000177 {
178 auto parser(IParser::Create());
179 BOOST_ASSERT(parser);
180
181 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
182
183 {
184 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
Derek Lamberti2b183fb2019-02-18 16:36:57 +0000185
186 boost::system::error_code errorCode;
187 boost::filesystem::path pathToFile(params.m_ModelPath);
188 if (!boost::filesystem::exists(pathToFile, errorCode))
189 {
190 throw armnn::FileNotFoundException(boost::str(
191 boost::format("Cannot find the file (%1%) errorCode: %2% %3%") %
192 params.m_ModelPath %
193 errorCode %
194 CHECK_LOCATION().AsString()));
195 }
196 std::ifstream file(params.m_ModelPath, std::ios::binary);
197
198 network = parser->CreateNetworkFromBinary(file);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000199 }
200
Derek Lambertiff05cc52019-04-26 13:05:17 +0100201 unsigned int subgraphId = boost::numeric_cast<unsigned int>(params.m_SubgraphId);
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000202
203 for (const std::string& inputLayerName : params.m_InputBindings)
204 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000205 armnnDeserializer::BindingPointInfo inputBinding =
Derek Lambertiff05cc52019-04-26 13:05:17 +0100206 parser->GetNetworkInputBindingInfo(subgraphId, inputLayerName);
Derek Lamberti8ddae332019-02-21 16:29:43 +0000207 inputBindings.push_back(std::make_pair(inputBinding.m_BindingId, inputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000208 }
209
210 for (const std::string& outputLayerName : params.m_OutputBindings)
211 {
Derek Lamberti8ddae332019-02-21 16:29:43 +0000212 armnnDeserializer::BindingPointInfo outputBinding =
Derek Lambertiff05cc52019-04-26 13:05:17 +0100213 parser->GetNetworkOutputBindingInfo(subgraphId, outputLayerName);
Derek Lamberti8ddae332019-02-21 16:29:43 +0000214 outputBindings.push_back(std::make_pair(outputBinding.m_BindingId, outputBinding.m_TensorInfo));
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +0000215 }
216
217 return network;
218 }
219};
220#endif
221
telsoa01c577f2c2018-08-31 09:22:23 +0100222#if defined(ARMNN_TF_LITE_PARSER)
223template <>
224struct CreateNetworkImpl<armnnTfLiteParser::ITfLiteParser>
225{
226public:
227 using IParser = armnnTfLiteParser::ITfLiteParser;
228 using Params = InferenceModelInternal::Params;
telsoa01c577f2c2018-08-31 09:22:23 +0100229
230 static armnn::INetworkPtr Create(const Params& params,
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100231 std::vector<armnn::BindingPointInfo>& inputBindings,
232 std::vector<armnn::BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100233 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000234 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100235
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000236 // Create a network from a file on disk
237 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100238
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000239 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100240
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000241 {
242 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
243 network = parser->CreateNetworkFromBinaryFile(modelPath.c_str());
244 }
telsoa01c577f2c2018-08-31 09:22:23 +0100245
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000246 for (const std::string& inputLayerName : params.m_InputBindings)
247 {
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100248 armnn::BindingPointInfo inputBinding =
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000249 parser->GetNetworkInputBindingInfo(params.m_SubgraphId, inputLayerName);
250 inputBindings.push_back(inputBinding);
251 }
252
253 for (const std::string& outputLayerName : params.m_OutputBindings)
254 {
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100255 armnn::BindingPointInfo outputBinding =
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000256 parser->GetNetworkOutputBindingInfo(params.m_SubgraphId, outputLayerName);
257 outputBindings.push_back(outputBinding);
258 }
259
260 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100261 }
262};
263#endif
264
265#if defined(ARMNN_ONNX_PARSER)
266template <>
267struct CreateNetworkImpl<armnnOnnxParser::IOnnxParser>
268{
269public:
270 using IParser = armnnOnnxParser::IOnnxParser;
271 using Params = InferenceModelInternal::Params;
272 using BindingPointInfo = InferenceModelInternal::BindingPointInfo;
273
274 static armnn::INetworkPtr Create(const Params& params,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000275 std::vector<BindingPointInfo>& inputBindings,
276 std::vector<BindingPointInfo>& outputBindings)
telsoa01c577f2c2018-08-31 09:22:23 +0100277 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000278 const std::string& modelPath = params.m_ModelPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100279
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000280 // Create a network from a file on disk
281 auto parser(IParser::Create());
telsoa01c577f2c2018-08-31 09:22:23 +0100282
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000283 armnn::INetworkPtr network{nullptr, [](armnn::INetwork *){}};
telsoa01c577f2c2018-08-31 09:22:23 +0100284
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000285 {
286 ARMNN_SCOPED_HEAP_PROFILING("Parsing");
287 network = (params.m_IsModelBinary ?
288 parser->CreateNetworkFromBinaryFile(modelPath.c_str()) :
289 parser->CreateNetworkFromTextFile(modelPath.c_str()));
290 }
telsoa01c577f2c2018-08-31 09:22:23 +0100291
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000292 for (const std::string& inputLayerName : params.m_InputBindings)
293 {
294 BindingPointInfo inputBinding = parser->GetNetworkInputBindingInfo(inputLayerName);
295 inputBindings.push_back(inputBinding);
296 }
297
298 for (const std::string& outputLayerName : params.m_OutputBindings)
299 {
300 BindingPointInfo outputBinding = parser->GetNetworkOutputBindingInfo(outputLayerName);
301 outputBindings.push_back(outputBinding);
302 }
303
304 return network;
telsoa01c577f2c2018-08-31 09:22:23 +0100305 }
306};
307#endif
telsoa014fcda012018-03-09 14:13:49 +0000308
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000309
telsoa014fcda012018-03-09 14:13:49 +0000310
311template <typename IParser, typename TDataType>
312class InferenceModel
313{
314public:
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000315 using DataType = TDataType;
316 using Params = InferenceModelInternal::Params;
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000317 using QuantizationParams = InferenceModelInternal::QuantizationParams;
Ferran Balaguerc602f292019-02-08 17:09:55 +0000318 using TContainer = boost::variant<std::vector<float>, std::vector<int>, std::vector<unsigned char>>;
telsoa014fcda012018-03-09 14:13:49 +0000319
320 struct CommandLineOptions
321 {
322 std::string m_ModelDir;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000323 std::vector<std::string> m_ComputeDevices;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100324 std::string m_DynamicBackendsPath;
surmeh013537c2c2018-05-18 16:31:43 +0100325 bool m_VisualizePostOptimizationModel;
telsoa01c577f2c2018-08-31 09:22:23 +0100326 bool m_EnableFp16TurboMode;
Pablo Tello507f39d2019-04-15 15:44:39 +0100327 std::string m_Labels;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000328
329 std::vector<armnn::BackendId> GetComputeDevicesAsBackendIds()
330 {
331 std::vector<armnn::BackendId> backendIds;
332 std::copy(m_ComputeDevices.begin(), m_ComputeDevices.end(), std::back_inserter(backendIds));
333 return backendIds;
334 }
telsoa014fcda012018-03-09 14:13:49 +0000335 };
336
337 static void AddCommandLineOptions(boost::program_options::options_description& desc, CommandLineOptions& options)
338 {
339 namespace po = boost::program_options;
340
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000341 const std::vector<std::string> defaultComputes = { "CpuAcc", "CpuRef" };
David Beckf0b48452018-10-19 15:20:56 +0100342
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100343 const std::string backendsMessage = "Which device to run layers on by default. Possible choices: "
344 + armnn::BackendRegistryInstance().GetBackendIdsAsString();
345
telsoa014fcda012018-03-09 14:13:49 +0000346 desc.add_options()
347 ("model-dir,m", po::value<std::string>(&options.m_ModelDir)->required(),
telsoa01c577f2c2018-08-31 09:22:23 +0100348 "Path to directory containing model files (.caffemodel/.prototxt/.tflite)")
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000349 ("compute,c", po::value<std::vector<std::string>>(&options.m_ComputeDevices)->
350 default_value(defaultComputes, boost::algorithm::join(defaultComputes, ", "))->
351 multitoken(), backendsMessage.c_str())
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100352 ("dynamic-backends-path,b", po::value(&options.m_DynamicBackendsPath),
353 "Path where to load any available dynamic backend from. "
354 "If left empty (the default), dynamic backends will not be used.")
Pablo Tello507f39d2019-04-15 15:44:39 +0100355 ("labels,l", po::value<std::string>(&options.m_Labels),
356 "Text file containing one image filename - correct label pair per line, "
357 "used to test the accuracy of the network.")
surmeh013537c2c2018-05-18 16:31:43 +0100358 ("visualize-optimized-model,v",
359 po::value<bool>(&options.m_VisualizePostOptimizationModel)->default_value(false),
360 "Produce a dot file useful for visualizing the graph post optimization."
telsoa01c577f2c2018-08-31 09:22:23 +0100361 "The file will have the same name as the model with the .dot extention.")
362 ("fp16-turbo-mode", po::value<bool>(&options.m_EnableFp16TurboMode)->default_value(false),
363 "If this option is enabled FP32 layers, weights and biases will be converted "
364 "to FP16 where the backend supports it.");
telsoa014fcda012018-03-09 14:13:49 +0000365 }
366
Matthew Bentham3e68b972019-04-09 13:10:46 +0100367 InferenceModel(const Params& params,
368 bool enableProfiling,
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100369 const std::string& dynamicBackendsPath,
Matthew Bentham3e68b972019-04-09 13:10:46 +0100370 const std::shared_ptr<armnn::IRuntime>& runtime = nullptr)
371 : m_EnableProfiling(enableProfiling)
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100372 , m_DynamicBackendsPath(dynamicBackendsPath)
telsoa014fcda012018-03-09 14:13:49 +0000373 {
telsoa01c577f2c2018-08-31 09:22:23 +0100374 if (runtime)
telsoa014fcda012018-03-09 14:13:49 +0000375 {
telsoa01c577f2c2018-08-31 09:22:23 +0100376 m_Runtime = runtime;
telsoa014fcda012018-03-09 14:13:49 +0000377 }
telsoa01c577f2c2018-08-31 09:22:23 +0100378 else
telsoa014fcda012018-03-09 14:13:49 +0000379 {
telsoa01c577f2c2018-08-31 09:22:23 +0100380 armnn::IRuntime::CreationOptions options;
Nina Drozd549ae372018-09-10 14:26:44 +0100381 options.m_EnableGpuProfiling = m_EnableProfiling;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100382 options.m_DynamicBackendsPath = m_DynamicBackendsPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100383 m_Runtime = std::move(armnn::IRuntime::Create(options));
surmeh013537c2c2018-05-18 16:31:43 +0100384 }
telsoa014fcda012018-03-09 14:13:49 +0000385
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100386 std::string invalidBackends;
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000387 if (!CheckRequestedBackendsAreValid(params.m_ComputeDevices, armnn::Optional<std::string&>(invalidBackends)))
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +0100388 {
389 throw armnn::Exception("Some backend IDs are invalid: " + invalidBackends);
390 }
391
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100392 armnn::INetworkPtr network = CreateNetworkImpl<IParser>::Create(params, m_InputBindings, m_OutputBindings);
telsoa014fcda012018-03-09 14:13:49 +0000393
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100394 armnn::IOptimizedNetworkPtr optNet{nullptr, [](armnn::IOptimizedNetwork*){}};
surmeh013537c2c2018-05-18 16:31:43 +0100395 {
396 ARMNN_SCOPED_HEAP_PROFILING("Optimizing");
telsoa01c577f2c2018-08-31 09:22:23 +0100397
398 armnn::OptimizerOptions options;
399 options.m_ReduceFp32ToFp16 = params.m_EnableFp16TurboMode;
Matthew Jackson54658b92019-08-27 15:35:59 +0100400 options.m_Debug = params.m_PrintIntermediateLayers;
telsoa01c577f2c2018-08-31 09:22:23 +0100401
Aron Virginas-Tar339bcae2019-01-31 16:44:26 +0000402 optNet = armnn::Optimize(*network, params.m_ComputeDevices, m_Runtime->GetDeviceSpec(), options);
telsoa01c577f2c2018-08-31 09:22:23 +0100403 if (!optNet)
404 {
405 throw armnn::Exception("Optimize returned nullptr");
406 }
surmeh013537c2c2018-05-18 16:31:43 +0100407 }
telsoa014fcda012018-03-09 14:13:49 +0000408
surmeh013537c2c2018-05-18 16:31:43 +0100409 if (params.m_VisualizePostOptimizationModel)
410 {
411 boost::filesystem::path filename = params.m_ModelPath;
412 filename.replace_extension("dot");
Rob Hughes9e10c2b2019-07-23 15:37:19 +0100413 std::fstream file(filename.c_str(), std::ios_base::out);
surmeh013537c2c2018-05-18 16:31:43 +0100414 optNet->SerializeToDot(file);
415 }
416
417 armnn::Status ret;
418 {
419 ARMNN_SCOPED_HEAP_PROFILING("LoadNetwork");
420 ret = m_Runtime->LoadNetwork(m_NetworkIdentifier, std::move(optNet));
421 }
422
telsoa014fcda012018-03-09 14:13:49 +0000423 if (ret == armnn::Status::Failure)
424 {
425 throw armnn::Exception("IRuntime::LoadNetwork failed");
426 }
427 }
428
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000429 void CheckInputIndexIsValid(unsigned int inputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000430 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000431 if (m_InputBindings.size() < inputIndex + 1)
432 {
433 throw armnn::Exception(boost::str(boost::format("Input index out of range: %1%") % inputIndex));
434 }
telsoa014fcda012018-03-09 14:13:49 +0000435 }
436
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000437 void CheckOutputIndexIsValid(unsigned int outputIndex) const
telsoa014fcda012018-03-09 14:13:49 +0000438 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000439 if (m_OutputBindings.size() < outputIndex + 1)
440 {
441 throw armnn::Exception(boost::str(boost::format("Output index out of range: %1%") % outputIndex));
442 }
443 }
444
445 unsigned int GetOutputSize(unsigned int outputIndex = 0u) const
446 {
447 CheckOutputIndexIsValid(outputIndex);
448 return m_OutputBindings[outputIndex].second.GetNumElements();
449 }
450
James Conroy7b4886f2019-04-11 10:23:58 +0100451 std::chrono::duration<double, std::milli> Run(
452 const std::vector<TContainer>& inputContainers,
453 std::vector<TContainer>& outputContainers)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000454 {
Ferran Balaguerc602f292019-02-08 17:09:55 +0000455 for (unsigned int i = 0; i < outputContainers.size(); ++i)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000456 {
457 const unsigned int expectedOutputDataSize = GetOutputSize(i);
Ferran Balaguerc602f292019-02-08 17:09:55 +0000458
459 boost::apply_visitor([expectedOutputDataSize, i](auto&& value)
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000460 {
Ferran Balaguerc602f292019-02-08 17:09:55 +0000461 const unsigned int actualOutputDataSize = boost::numeric_cast<unsigned int>(value.size());
462 if (actualOutputDataSize < expectedOutputDataSize)
463 {
464 unsigned int outputIndex = boost::numeric_cast<unsigned int>(i);
465 throw armnn::Exception(
466 boost::str(boost::format("Not enough data for output #%1%: expected "
467 "%2% elements, got %3%") % outputIndex % expectedOutputDataSize % actualOutputDataSize));
468 }
469 },
470 outputContainers[i]);
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000471 }
telsoa01c577f2c2018-08-31 09:22:23 +0100472
473 std::shared_ptr<armnn::IProfiler> profiler = m_Runtime->GetProfiler(m_NetworkIdentifier);
474 if (profiler)
475 {
476 profiler->EnableProfiling(m_EnableProfiling);
477 }
478
James Conroy7b4886f2019-04-11 10:23:58 +0100479 // Start timer to record inference time in EnqueueWorkload (in milliseconds)
480 const auto start_time = GetCurrentTime();
481
telsoa014fcda012018-03-09 14:13:49 +0000482 armnn::Status ret = m_Runtime->EnqueueWorkload(m_NetworkIdentifier,
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000483 MakeInputTensors(inputContainers),
484 MakeOutputTensors(outputContainers));
Sadik Armagan2b7a1582018-09-05 16:33:58 +0100485
James Conroy7b4886f2019-04-11 10:23:58 +0100486 const auto end_time = GetCurrentTime();
487
Sadik Armagan2b7a1582018-09-05 16:33:58 +0100488 // if profiling is enabled print out the results
489 if (profiler && profiler->IsProfilingEnabled())
490 {
491 profiler->Print(std::cout);
492 }
493
telsoa014fcda012018-03-09 14:13:49 +0000494 if (ret == armnn::Status::Failure)
495 {
496 throw armnn::Exception("IRuntime::EnqueueWorkload failed");
497 }
James Conroy7b4886f2019-04-11 10:23:58 +0100498 else
499 {
500 return std::chrono::duration<double, std::milli>(end_time - start_time);
501 }
telsoa014fcda012018-03-09 14:13:49 +0000502 }
503
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100504 const armnn::BindingPointInfo& GetInputBindingInfo(unsigned int inputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100505 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000506 CheckInputIndexIsValid(inputIndex);
507 return m_InputBindings[inputIndex];
telsoa01c577f2c2018-08-31 09:22:23 +0100508 }
509
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100510 const std::vector<armnn::BindingPointInfo>& GetInputBindingInfos() const
telsoa01c577f2c2018-08-31 09:22:23 +0100511 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000512 return m_InputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100513 }
514
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100515 const armnn::BindingPointInfo& GetOutputBindingInfo(unsigned int outputIndex = 0u) const
telsoa01c577f2c2018-08-31 09:22:23 +0100516 {
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000517 CheckOutputIndexIsValid(outputIndex);
518 return m_OutputBindings[outputIndex];
519 }
520
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100521 const std::vector<armnn::BindingPointInfo>& GetOutputBindingInfos() const
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000522 {
523 return m_OutputBindings;
524 }
525
526 QuantizationParams GetQuantizationParams(unsigned int outputIndex = 0u) const
527 {
528 CheckOutputIndexIsValid(outputIndex);
529 return std::make_pair(m_OutputBindings[outputIndex].second.GetQuantizationScale(),
530 m_OutputBindings[outputIndex].second.GetQuantizationOffset());
531 }
532
Narumol Prangnawarat4628d052019-02-25 17:26:05 +0000533 QuantizationParams GetInputQuantizationParams(unsigned int inputIndex = 0u) const
534 {
535 CheckInputIndexIsValid(inputIndex);
536 return std::make_pair(m_InputBindings[inputIndex].second.GetQuantizationScale(),
537 m_InputBindings[inputIndex].second.GetQuantizationOffset());
538 }
539
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000540 std::vector<QuantizationParams> GetAllQuantizationParams() const
541 {
542 std::vector<QuantizationParams> quantizationParams;
543 for (unsigned int i = 0u; i < m_OutputBindings.size(); i++)
544 {
545 quantizationParams.push_back(GetQuantizationParams(i));
546 }
547 return quantizationParams;
telsoa01c577f2c2018-08-31 09:22:23 +0100548 }
549
telsoa014fcda012018-03-09 14:13:49 +0000550private:
telsoa01c577f2c2018-08-31 09:22:23 +0100551 armnn::NetworkId m_NetworkIdentifier;
552 std::shared_ptr<armnn::IRuntime> m_Runtime;
553
Jim Flynnb4d7eae2019-05-01 14:44:27 +0100554 std::vector<armnn::BindingPointInfo> m_InputBindings;
555 std::vector<armnn::BindingPointInfo> m_OutputBindings;
telsoa01c577f2c2018-08-31 09:22:23 +0100556 bool m_EnableProfiling;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100557 std::string m_DynamicBackendsPath;
telsoa01c577f2c2018-08-31 09:22:23 +0100558
telsoa014fcda012018-03-09 14:13:49 +0000559 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000560 armnn::InputTensors MakeInputTensors(const std::vector<TContainer>& inputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000561 {
Jim Flynn2fd61002019-05-03 12:54:26 +0100562 return armnnUtils::MakeInputTensors(m_InputBindings, inputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000563 }
564
565 template<typename TContainer>
Aron Virginas-Tar7cf0eaa2019-01-24 17:05:36 +0000566 armnn::OutputTensors MakeOutputTensors(std::vector<TContainer>& outputDataContainers)
telsoa014fcda012018-03-09 14:13:49 +0000567 {
Jim Flynn2fd61002019-05-03 12:54:26 +0100568 return armnnUtils::MakeOutputTensors(m_OutputBindings, outputDataContainers);
telsoa014fcda012018-03-09 14:13:49 +0000569 }
James Conroy7b4886f2019-04-11 10:23:58 +0100570
571 std::chrono::high_resolution_clock::time_point GetCurrentTime()
572 {
573 return std::chrono::high_resolution_clock::now();
574 }
575
576 std::chrono::duration<double, std::milli> GetTimeDuration(
577 std::chrono::high_resolution_clock::time_point& start_time,
578 std::chrono::high_resolution_clock::time_point& end_time)
579 {
580 return std::chrono::duration<double, std::milli>(end_time - start_time);
581 }
582
Ferran Balaguerc602f292019-02-08 17:09:55 +0000583};