blob: f2763a72b734e1ddbde047437d12adc81fcfbf47 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Sadik Armagana9c2ce12020-07-14 10:02:22 +01002// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
telsoa01c577f2c2018-08-31 09:22:23 +01005
Francis Murtaghbee4bc92019-06-18 12:30:37 +01006#include "../NetworkExecutionUtils/NetworkExecutionUtils.hpp"
telsoa01c577f2c2018-08-31 09:22:23 +01007
James Conroy7b4886f2019-04-11 10:23:58 +01008// MAIN
telsoa01c577f2c2018-08-31 09:22:23 +01009int main(int argc, const char* argv[])
10{
11 // Configures logging for both the ARMNN library and this test program.
12#ifdef NDEBUG
13 armnn::LogSeverity level = armnn::LogSeverity::Info;
14#else
15 armnn::LogSeverity level = armnn::LogSeverity::Debug;
16#endif
17 armnn::ConfigureLogging(true, true, level);
telsoa01c577f2c2018-08-31 09:22:23 +010018
19 std::string testCasesFile;
20
21 std::string modelFormat;
22 std::string modelPath;
Ferran Balaguerc602f292019-02-08 17:09:55 +000023 std::string inputNames;
24 std::string inputTensorShapes;
25 std::string inputTensorDataFilePaths;
26 std::string outputNames;
27 std::string inputTypes;
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000028 std::string outputTypes;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010029 std::string dynamicBackendsPath;
Sadik Armagan77086282019-09-02 11:46:28 +010030 std::string outputTensorFiles;
telsoa01c577f2c2018-08-31 09:22:23 +010031
Jim Flynn4951b8c2019-10-03 10:04:30 -070032 // external profiling parameters
33 std::string outgoingCaptureFile;
34 std::string incomingCaptureFile;
35 uint32_t counterCapturePeriod;
Isabella Gottardia0687ee2020-03-11 18:04:20 +000036 std::string fileFormat;
Jim Flynn4951b8c2019-10-03 10:04:30 -070037
alered01a7227ac2020-05-07 14:58:29 +010038 size_t iterations = 1;
39 int tuningLevel = 0;
40 std::string tuningPath;
41
James Conroy7b4886f2019-04-11 10:23:58 +010042 double thresholdTime = 0.0;
43
telsoa01c577f2c2018-08-31 09:22:23 +010044 size_t subgraphId = 0;
45
Matthew Jackson07882f12019-09-05 15:55:55 +010046 const std::string backendsMessage = "REQUIRED: Which device to run layers on by default. Possible choices: "
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010047 + armnn::BackendRegistryInstance().GetBackendIdsAsString();
telsoa01c577f2c2018-08-31 09:22:23 +010048 po::options_description desc("Options");
49 try
50 {
51 desc.add_options()
52 ("help", "Display usage information")
Matthew Jackson07882f12019-09-05 15:55:55 +010053 ("compute,c", po::value<std::vector<std::string>>()->multitoken()->required(),
54 backendsMessage.c_str())
telsoa01c577f2c2018-08-31 09:22:23 +010055 ("test-cases,t", po::value(&testCasesFile), "Path to a CSV file containing test cases to run. "
56 "If set, further parameters -- with the exception of compute device and concurrency -- will be ignored, "
57 "as they are expected to be defined in the file for each test in particular.")
58 ("concurrent,n", po::bool_switch()->default_value(false),
59 "Whether or not the test cases should be executed in parallel")
Matteo Martincigh49124022019-01-11 13:25:59 +000060 ("model-format,f", po::value(&modelFormat)->required(),
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000061 "armnn-binary, caffe-binary, caffe-text, onnx-binary, onnx-text, tflite-binary, tensorflow-binary or "
62 "tensorflow-text.")
63 ("model-path,m", po::value(&modelPath)->required(), "Path to model file, e.g. .armnn, .caffemodel, "
64 ".prototxt, .tflite, .onnx")
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010065 ("dynamic-backends-path,b", po::value(&dynamicBackendsPath),
66 "Path where to load any available dynamic backend from. "
67 "If left empty (the default), dynamic backends will not be used.")
Ferran Balaguerc602f292019-02-08 17:09:55 +000068 ("input-name,i", po::value(&inputNames),
69 "Identifier of the input tensors in the network separated by comma.")
telsoa01c577f2c2018-08-31 09:22:23 +010070 ("subgraph-number,x", po::value<size_t>(&subgraphId)->default_value(0), "Id of the subgraph to be executed."
71 "Defaults to 0")
Ferran Balaguerc602f292019-02-08 17:09:55 +000072 ("input-tensor-shape,s", po::value(&inputTensorShapes),
Francis Murtagh1555cbd2019-10-08 14:47:46 +010073 "The shape of the input tensors in the network as a flat array of integers separated by comma."
74 "Several shapes can be passed by separating them with a colon (:)."
telsoa01c577f2c2018-08-31 09:22:23 +010075 "This parameter is optional, depending on the network.")
Aron Virginas-Tarc82c8732019-10-24 17:07:43 +010076 ("input-tensor-data,d", po::value(&inputTensorDataFilePaths)->default_value(""),
Ferran Balaguerc602f292019-02-08 17:09:55 +000077 "Path to files containing the input data as a flat array separated by whitespace. "
Aron Virginas-Tarc82c8732019-10-24 17:07:43 +010078 "Several paths can be passed by separating them with a comma. If not specified, the network will be run "
79 "with dummy data (useful for profiling).")
Ferran Balaguerc602f292019-02-08 17:09:55 +000080 ("input-type,y",po::value(&inputTypes), "The type of the input tensors in the network separated by comma. "
81 "If unset, defaults to \"float\" for all defined inputs. "
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000082 "Accepted values (float, int or qasymm8)")
Narumol Prangnawarat610256f2019-06-26 15:10:46 +010083 ("quantize-input,q",po::bool_switch()->default_value(false),
84 "If this option is enabled, all float inputs will be quantized to qasymm8. "
85 "If unset, default to not quantized. "
86 "Accepted values (true or false)")
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000087 ("output-type,z",po::value(&outputTypes),
88 "The type of the output tensors in the network separated by comma. "
89 "If unset, defaults to \"float\" for all defined outputs. "
90 "Accepted values (float, int or qasymm8).")
Georgios Pinitas50311ba2020-02-18 13:25:23 +000091 ("dequantize-output,l",po::bool_switch()->default_value(false),
92 "If this option is enabled, all quantized outputs will be dequantized to float. "
93 "If unset, default to not get dequantized. "
94 "Accepted values (true or false)")
Ferran Balaguerc602f292019-02-08 17:09:55 +000095 ("output-name,o", po::value(&outputNames),
96 "Identifier of the output tensors in the network separated by comma.")
Sadik Armagan77086282019-09-02 11:46:28 +010097 ("write-outputs-to-file,w", po::value(&outputTensorFiles),
98 "Comma-separated list of output file paths keyed with the binding-id of the output slot. "
99 "If left empty (the default), the output tensors will not be written to a file.")
telsoa01c577f2c2018-08-31 09:22:23 +0100100 ("event-based-profiling,e", po::bool_switch()->default_value(false),
Ruomei Yan2fcce082019-04-02 16:47:34 +0100101 "Enables built in profiler. If unset, defaults to off.")
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100102 ("visualize-optimized-model,v", po::bool_switch()->default_value(false),
103 "Enables built optimized model visualizer. If unset, defaults to off.")
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000104 ("bf16-turbo-mode", po::bool_switch()->default_value(false), "If this option is enabled, FP32 layers, "
105 "weights and biases will be converted to BFloat16 where the backend supports it")
Ruomei Yan2fcce082019-04-02 16:47:34 +0100106 ("fp16-turbo-mode,h", po::bool_switch()->default_value(false), "If this option is enabled, FP32 layers, "
James Conroy7b4886f2019-04-11 10:23:58 +0100107 "weights and biases will be converted to FP16 where the backend supports it")
108 ("threshold-time,r", po::value<double>(&thresholdTime)->default_value(0.0),
109 "Threshold time is the maximum allowed time for inference measured in milliseconds. If the actual "
110 "inference time is greater than the threshold time, the test will fail. By default, no threshold "
Matthew Jackson54658b92019-08-27 15:35:59 +0100111 "time is used.")
112 ("print-intermediate-layers,p", po::bool_switch()->default_value(false),
Jim Flynn4951b8c2019-10-03 10:04:30 -0700113 "If this option is enabled, the output of every graph layer will be printed.")
114 ("enable-external-profiling,a", po::bool_switch()->default_value(false),
115 "If enabled external profiling will be switched on")
Finn Williamsd7fcafa2020-04-23 17:55:18 +0100116 ("timeline-profiling", po::bool_switch()->default_value(false),
117 "If enabled timeline profiling will be switched on, requires external profiling")
Jim Flynn4951b8c2019-10-03 10:04:30 -0700118 ("outgoing-capture-file,j", po::value(&outgoingCaptureFile),
119 "If specified the outgoing external profiling packets will be captured in this binary file")
120 ("incoming-capture-file,k", po::value(&incomingCaptureFile),
121 "If specified the incoming external profiling packets will be captured in this binary file")
122 ("file-only-external-profiling,g", po::bool_switch()->default_value(false),
123 "If enabled then the 'file-only' test mode of external profiling will be enabled")
124 ("counter-capture-period,u", po::value<uint32_t>(&counterCapturePeriod)->default_value(150u),
Derek Lamberti132563c2019-12-02 16:06:40 +0000125 "If profiling is enabled in 'file-only' mode this is the capture period that will be used in the test")
Finn Williams50616012020-05-13 17:19:14 +0100126 ("file-format", po::value(&fileFormat)->default_value("binary"),
Isabella Gottardia0687ee2020-03-11 18:04:20 +0000127 "If profiling is enabled specifies the output file format")
alered01a7227ac2020-05-07 14:58:29 +0100128 ("iterations", po::value<size_t>(&iterations)->default_value(1),
129 "Number of iterations to run the network for, default is set to 1")
130 ("tuning-path", po::value(&tuningPath),
131 "Path to tuning file. Enables use of CL tuning")
132 ("tuning-level", po::value<int>(&tuningLevel)->default_value(0),
133 "Sets the tuning level which enables a tuning run which will update/create a tuning file. "
134 "Available options are: 1 (Rapid), 2 (Normal), 3 (Exhaustive). "
135 "Requires tuning-path to be set, default is set to 0 (No tuning run)")
Derek Lamberti132563c2019-12-02 16:06:40 +0000136 ("parse-unsupported", po::bool_switch()->default_value(false),
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100137 "Add unsupported operators as stand-in layers (where supported by parser)")
138 ("infer-output-shape", po::bool_switch()->default_value(false),
139 "Infers output tensor shape from input tensor shape and validate where applicable (where supported by "
140 "parser)");
telsoa01c577f2c2018-08-31 09:22:23 +0100141 }
142 catch (const std::exception& e)
143 {
144 // Coverity points out that default_value(...) can throw a bad_lexical_cast,
145 // and that desc.add_options() can throw boost::io::too_few_args.
146 // They really won't in any of these cases.
Narumol Prangnawaratac2770a2020-04-01 16:51:23 +0100147 ARMNN_ASSERT_MSG(false, "Caught unexpected exception");
Derek Lamberti08446972019-11-26 16:38:31 +0000148 ARMNN_LOG(fatal) << "Fatal internal error: " << e.what();
telsoa01c577f2c2018-08-31 09:22:23 +0100149 return EXIT_FAILURE;
150 }
151
152 // Parses the command-line.
153 po::variables_map vm;
154 try
155 {
156 po::store(po::parse_command_line(argc, argv, desc), vm);
157
158 if (CheckOption(vm, "help") || argc <= 1)
159 {
160 std::cout << "Executes a neural network model using the provided input tensor. " << std::endl;
161 std::cout << "Prints the resulting output tensor." << std::endl;
162 std::cout << std::endl;
163 std::cout << desc << std::endl;
164 return EXIT_SUCCESS;
165 }
166
167 po::notify(vm);
168 }
169 catch (const po::error& e)
170 {
171 std::cerr << e.what() << std::endl << std::endl;
172 std::cerr << desc << std::endl;
173 return EXIT_FAILURE;
174 }
175
176 // Get the value of the switch arguments.
177 bool concurrent = vm["concurrent"].as<bool>();
178 bool enableProfiling = vm["event-based-profiling"].as<bool>();
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100179 bool enableLayerDetails = vm["visualize-optimized-model"].as<bool>();
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000180 bool enableBf16TurboMode = vm["bf16-turbo-mode"].as<bool>();
Ruomei Yan2fcce082019-04-02 16:47:34 +0100181 bool enableFp16TurboMode = vm["fp16-turbo-mode"].as<bool>();
Narumol Prangnawarat610256f2019-06-26 15:10:46 +0100182 bool quantizeInput = vm["quantize-input"].as<bool>();
Georgios Pinitas50311ba2020-02-18 13:25:23 +0000183 bool dequantizeOutput = vm["dequantize-output"].as<bool>();
Matthew Jackson54658b92019-08-27 15:35:59 +0100184 bool printIntermediate = vm["print-intermediate-layers"].as<bool>();
Jim Flynn4951b8c2019-10-03 10:04:30 -0700185 bool enableExternalProfiling = vm["enable-external-profiling"].as<bool>();
186 bool fileOnlyExternalProfiling = vm["file-only-external-profiling"].as<bool>();
Derek Lamberti132563c2019-12-02 16:06:40 +0000187 bool parseUnsupported = vm["parse-unsupported"].as<bool>();
Finn Williamsd7fcafa2020-04-23 17:55:18 +0100188 bool timelineEnabled = vm["timeline-profiling"].as<bool>();
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100189 bool inferOutputShape = vm["infer-output-shape"].as<bool>();
Derek Lamberti132563c2019-12-02 16:06:40 +0000190
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000191 if (enableBf16TurboMode && enableFp16TurboMode)
192 {
193 ARMNN_LOG(fatal) << "BFloat16 and Float16 turbo mode cannot be enabled at the same time.";
194 return EXIT_FAILURE;
195 }
196
Finn Williamsd7fcafa2020-04-23 17:55:18 +0100197 // Create runtime
198 armnn::IRuntime::CreationOptions options;
199 options.m_EnableGpuProfiling = enableProfiling;
200 options.m_DynamicBackendsPath = dynamicBackendsPath;
201 options.m_ProfilingOptions.m_EnableProfiling = enableExternalProfiling;
202 options.m_ProfilingOptions.m_IncomingCaptureFile = incomingCaptureFile;
203 options.m_ProfilingOptions.m_OutgoingCaptureFile = outgoingCaptureFile;
204 options.m_ProfilingOptions.m_FileOnly = fileOnlyExternalProfiling;
205 options.m_ProfilingOptions.m_CapturePeriod = counterCapturePeriod;
206 options.m_ProfilingOptions.m_FileFormat = fileFormat;
207 options.m_ProfilingOptions.m_TimelineEnabled = timelineEnabled;
208
209 if (timelineEnabled && !enableExternalProfiling)
210 {
211 ARMNN_LOG(fatal) << "Timeline profiling requires external profiling to be turned on";
212 return EXIT_FAILURE;
213 }
telsoa01c577f2c2018-08-31 09:22:23 +0100214
215 // Check whether we have to load test cases from a file.
216 if (CheckOption(vm, "test-cases"))
217 {
218 // Check that the file exists.
Francis Murtagh532a29d2020-06-29 11:50:01 +0100219 if (!fs::exists(testCasesFile))
telsoa01c577f2c2018-08-31 09:22:23 +0100220 {
Derek Lamberti08446972019-11-26 16:38:31 +0000221 ARMNN_LOG(fatal) << "Given file \"" << testCasesFile << "\" does not exist";
telsoa01c577f2c2018-08-31 09:22:23 +0100222 return EXIT_FAILURE;
223 }
224
225 // Parse CSV file and extract test cases
226 armnnUtils::CsvReader reader;
227 std::vector<armnnUtils::CsvRow> testCases = reader.ParseFile(testCasesFile);
228
229 // Check that there is at least one test case to run
230 if (testCases.empty())
231 {
Derek Lamberti08446972019-11-26 16:38:31 +0000232 ARMNN_LOG(fatal) << "Given file \"" << testCasesFile << "\" has no test cases";
telsoa01c577f2c2018-08-31 09:22:23 +0100233 return EXIT_FAILURE;
234 }
telsoa01c577f2c2018-08-31 09:22:23 +0100235 // Create runtime
telsoa01c577f2c2018-08-31 09:22:23 +0100236 std::shared_ptr<armnn::IRuntime> runtime(armnn::IRuntime::Create(options));
237
238 const std::string executableName("ExecuteNetwork");
239
240 // Check whether we need to run the test cases concurrently
241 if (concurrent)
242 {
243 std::vector<std::future<int>> results;
244 results.reserve(testCases.size());
245
246 // Run each test case in its own thread
247 for (auto& testCase : testCases)
248 {
249 testCase.values.insert(testCase.values.begin(), executableName);
Nina Drozd549ae372018-09-10 14:26:44 +0100250 results.push_back(std::async(std::launch::async, RunCsvTest, std::cref(testCase), std::cref(runtime),
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000251 enableProfiling, enableFp16TurboMode, enableBf16TurboMode, thresholdTime,
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100252 printIntermediate, enableLayerDetails, parseUnsupported,
253 inferOutputShape));
telsoa01c577f2c2018-08-31 09:22:23 +0100254 }
255
256 // Check results
257 for (auto& result : results)
258 {
259 if (result.get() != EXIT_SUCCESS)
260 {
261 return EXIT_FAILURE;
262 }
263 }
264 }
265 else
266 {
267 // Run tests sequentially
268 for (auto& testCase : testCases)
269 {
270 testCase.values.insert(testCase.values.begin(), executableName);
Matthew Jackson54658b92019-08-27 15:35:59 +0100271 if (RunCsvTest(testCase, runtime, enableProfiling,
Narumol Prangnawaratd8cc8112020-03-24 13:54:05 +0000272 enableFp16TurboMode, enableBf16TurboMode, thresholdTime, printIntermediate,
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100273 enableLayerDetails, parseUnsupported, inferOutputShape) != EXIT_SUCCESS)
telsoa01c577f2c2018-08-31 09:22:23 +0100274 {
275 return EXIT_FAILURE;
276 }
277 }
278 }
279
280 return EXIT_SUCCESS;
281 }
282 else // Run single test
283 {
Aron Virginas-Tar382e21c2019-01-22 14:10:39 +0000284 // Get the preferred order of compute devices. If none are specified, default to using CpuRef
285 const std::string computeOption("compute");
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100286 std::vector<std::string> computeDevicesAsStrings =
287 CheckOption(vm, computeOption.c_str()) ?
288 vm[computeOption].as<std::vector<std::string>>() :
289 std::vector<std::string>();
Matteo Martincigh067112f2018-10-29 11:01:09 +0000290 std::vector<armnn::BackendId> computeDevices(computeDevicesAsStrings.begin(), computeDevicesAsStrings.end());
telsoa01c577f2c2018-08-31 09:22:23 +0100291
292 // Remove duplicates from the list of compute devices.
293 RemoveDuplicateDevices(computeDevices);
294
alered01a7227ac2020-05-07 14:58:29 +0100295#if defined(ARMCOMPUTECL_ENABLED)
296 std::shared_ptr<armnn::IGpuAccTunedParameters> tuned_params;
297
298 if (tuningPath != "")
299 {
300 if (tuningLevel != 0)
301 {
302 RunCLTuning(tuningPath, tuningLevel, modelFormat, inputTensorShapes, computeDevices,
303 dynamicBackendsPath, modelPath, inputNames, inputTensorDataFilePaths, inputTypes, quantizeInput,
304 outputTypes, outputNames, outputTensorFiles, dequantizeOutput, enableProfiling,
305 enableFp16TurboMode, enableBf16TurboMode, thresholdTime, printIntermediate, subgraphId,
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100306 enableLayerDetails, parseUnsupported, inferOutputShape);
alered01a7227ac2020-05-07 14:58:29 +0100307 }
308 ARMNN_LOG(info) << "Using tuning params: " << tuningPath << "\n";
309 options.m_BackendOptions.emplace_back(
310 armnn::BackendOptions
311 {
312 "GpuAcc",
313 {
314 {"TuningLevel", 0},
315 {"TuningFile", tuningPath.c_str()},
316 {"KernelProfilingEnabled", enableProfiling}
317 }
318 }
319 );
320 }
321#endif
telsoa01c577f2c2018-08-31 09:22:23 +0100322 try
323 {
324 CheckOptionDependencies(vm);
325 }
326 catch (const po::error& e)
327 {
328 std::cerr << e.what() << std::endl << std::endl;
329 std::cerr << desc << std::endl;
330 return EXIT_FAILURE;
331 }
Colm Donelanb682d842019-10-16 12:24:20 +0100332 // Create runtime
Colm Donelanb682d842019-10-16 12:24:20 +0100333 std::shared_ptr<armnn::IRuntime> runtime(armnn::IRuntime::Create(options));
Aron Virginas-Tarc82c8732019-10-24 17:07:43 +0100334
alered01a7227ac2020-05-07 14:58:29 +0100335 return RunTest(modelFormat, inputTensorShapes, computeDevices, dynamicBackendsPath, modelPath,
336 inputNames, inputTensorDataFilePaths, inputTypes, quantizeInput, outputTypes, outputNames,
337 outputTensorFiles, dequantizeOutput, enableProfiling, enableFp16TurboMode, enableBf16TurboMode,
Sadik Armagana9c2ce12020-07-14 10:02:22 +0100338 thresholdTime, printIntermediate, subgraphId, enableLayerDetails, parseUnsupported, inferOutputShape,
339 iterations, runtime);
telsoa014fcda012018-03-09 14:13:49 +0000340 }
341}