blob: d97a19928c5724ea97afebe4da3046b89f4f623a [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
telsoa01c577f2c2018-08-31 09:22:23 +01005
Francis Murtaghbee4bc92019-06-18 12:30:37 +01006#include "../NetworkExecutionUtils/NetworkExecutionUtils.hpp"
telsoa01c577f2c2018-08-31 09:22:23 +01007
James Conroy7b4886f2019-04-11 10:23:58 +01008// MAIN
telsoa01c577f2c2018-08-31 09:22:23 +01009int main(int argc, const char* argv[])
10{
11 // Configures logging for both the ARMNN library and this test program.
12#ifdef NDEBUG
13 armnn::LogSeverity level = armnn::LogSeverity::Info;
14#else
15 armnn::LogSeverity level = armnn::LogSeverity::Debug;
16#endif
17 armnn::ConfigureLogging(true, true, level);
18 armnnUtils::ConfigureLogging(boost::log::core::get().get(), true, true, level);
19
20 std::string testCasesFile;
21
22 std::string modelFormat;
23 std::string modelPath;
Ferran Balaguerc602f292019-02-08 17:09:55 +000024 std::string inputNames;
25 std::string inputTensorShapes;
26 std::string inputTensorDataFilePaths;
27 std::string outputNames;
28 std::string inputTypes;
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000029 std::string outputTypes;
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010030 std::string dynamicBackendsPath;
Sadik Armagan77086282019-09-02 11:46:28 +010031 std::string outputTensorFiles;
telsoa01c577f2c2018-08-31 09:22:23 +010032
James Conroy7b4886f2019-04-11 10:23:58 +010033 double thresholdTime = 0.0;
34
telsoa01c577f2c2018-08-31 09:22:23 +010035 size_t subgraphId = 0;
36
Matthew Jackson07882f12019-09-05 15:55:55 +010037 const std::string backendsMessage = "REQUIRED: Which device to run layers on by default. Possible choices: "
Aron Virginas-Tar5cc8e562018-10-23 15:14:46 +010038 + armnn::BackendRegistryInstance().GetBackendIdsAsString();
39
telsoa01c577f2c2018-08-31 09:22:23 +010040 po::options_description desc("Options");
41 try
42 {
43 desc.add_options()
44 ("help", "Display usage information")
Matthew Jackson07882f12019-09-05 15:55:55 +010045 ("compute,c", po::value<std::vector<std::string>>()->multitoken()->required(),
46 backendsMessage.c_str())
telsoa01c577f2c2018-08-31 09:22:23 +010047 ("test-cases,t", po::value(&testCasesFile), "Path to a CSV file containing test cases to run. "
48 "If set, further parameters -- with the exception of compute device and concurrency -- will be ignored, "
49 "as they are expected to be defined in the file for each test in particular.")
50 ("concurrent,n", po::bool_switch()->default_value(false),
51 "Whether or not the test cases should be executed in parallel")
Matteo Martincigh49124022019-01-11 13:25:59 +000052 ("model-format,f", po::value(&modelFormat)->required(),
Aron Virginas-Tar64e4ccb2019-02-12 11:27:53 +000053 "armnn-binary, caffe-binary, caffe-text, onnx-binary, onnx-text, tflite-binary, tensorflow-binary or "
54 "tensorflow-text.")
55 ("model-path,m", po::value(&modelPath)->required(), "Path to model file, e.g. .armnn, .caffemodel, "
56 ".prototxt, .tflite, .onnx")
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010057 ("dynamic-backends-path,b", po::value(&dynamicBackendsPath),
58 "Path where to load any available dynamic backend from. "
59 "If left empty (the default), dynamic backends will not be used.")
Ferran Balaguerc602f292019-02-08 17:09:55 +000060 ("input-name,i", po::value(&inputNames),
61 "Identifier of the input tensors in the network separated by comma.")
telsoa01c577f2c2018-08-31 09:22:23 +010062 ("subgraph-number,x", po::value<size_t>(&subgraphId)->default_value(0), "Id of the subgraph to be executed."
63 "Defaults to 0")
Ferran Balaguerc602f292019-02-08 17:09:55 +000064 ("input-tensor-shape,s", po::value(&inputTensorShapes),
65 "The shape of the input tensors in the network as a flat array of integers separated by comma. "
66 "Several shapes can be passed separating them by semicolon. "
telsoa01c577f2c2018-08-31 09:22:23 +010067 "This parameter is optional, depending on the network.")
Ferran Balaguerc602f292019-02-08 17:09:55 +000068 ("input-tensor-data,d", po::value(&inputTensorDataFilePaths),
69 "Path to files containing the input data as a flat array separated by whitespace. "
Matteo Martincigh00dda4a2019-08-14 11:42:30 +010070 "Several paths can be passed separating them by comma.")
Ferran Balaguerc602f292019-02-08 17:09:55 +000071 ("input-type,y",po::value(&inputTypes), "The type of the input tensors in the network separated by comma. "
72 "If unset, defaults to \"float\" for all defined inputs. "
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000073 "Accepted values (float, int or qasymm8)")
Narumol Prangnawarat610256f2019-06-26 15:10:46 +010074 ("quantize-input,q",po::bool_switch()->default_value(false),
75 "If this option is enabled, all float inputs will be quantized to qasymm8. "
76 "If unset, default to not quantized. "
77 "Accepted values (true or false)")
Éanna Ó Catháinb3d481a2019-02-26 11:26:24 +000078 ("output-type,z",po::value(&outputTypes),
79 "The type of the output tensors in the network separated by comma. "
80 "If unset, defaults to \"float\" for all defined outputs. "
81 "Accepted values (float, int or qasymm8).")
Ferran Balaguerc602f292019-02-08 17:09:55 +000082 ("output-name,o", po::value(&outputNames),
83 "Identifier of the output tensors in the network separated by comma.")
Sadik Armagan77086282019-09-02 11:46:28 +010084 ("write-outputs-to-file,w", po::value(&outputTensorFiles),
85 "Comma-separated list of output file paths keyed with the binding-id of the output slot. "
86 "If left empty (the default), the output tensors will not be written to a file.")
telsoa01c577f2c2018-08-31 09:22:23 +010087 ("event-based-profiling,e", po::bool_switch()->default_value(false),
Ruomei Yan2fcce082019-04-02 16:47:34 +010088 "Enables built in profiler. If unset, defaults to off.")
Andre Ghattas23ae2ea2019-08-07 12:18:38 +010089 ("visualize-optimized-model,v", po::bool_switch()->default_value(false),
90 "Enables built optimized model visualizer. If unset, defaults to off.")
Ruomei Yan2fcce082019-04-02 16:47:34 +010091 ("fp16-turbo-mode,h", po::bool_switch()->default_value(false), "If this option is enabled, FP32 layers, "
James Conroy7b4886f2019-04-11 10:23:58 +010092 "weights and biases will be converted to FP16 where the backend supports it")
93 ("threshold-time,r", po::value<double>(&thresholdTime)->default_value(0.0),
94 "Threshold time is the maximum allowed time for inference measured in milliseconds. If the actual "
95 "inference time is greater than the threshold time, the test will fail. By default, no threshold "
Matthew Jackson54658b92019-08-27 15:35:59 +010096 "time is used.")
97 ("print-intermediate-layers,p", po::bool_switch()->default_value(false),
98 "If this option is enabled, the output of every graph layer will be printed.");
telsoa01c577f2c2018-08-31 09:22:23 +010099 }
100 catch (const std::exception& e)
101 {
102 // Coverity points out that default_value(...) can throw a bad_lexical_cast,
103 // and that desc.add_options() can throw boost::io::too_few_args.
104 // They really won't in any of these cases.
105 BOOST_ASSERT_MSG(false, "Caught unexpected exception");
106 BOOST_LOG_TRIVIAL(fatal) << "Fatal internal error: " << e.what();
107 return EXIT_FAILURE;
108 }
109
110 // Parses the command-line.
111 po::variables_map vm;
112 try
113 {
114 po::store(po::parse_command_line(argc, argv, desc), vm);
115
116 if (CheckOption(vm, "help") || argc <= 1)
117 {
118 std::cout << "Executes a neural network model using the provided input tensor. " << std::endl;
119 std::cout << "Prints the resulting output tensor." << std::endl;
120 std::cout << std::endl;
121 std::cout << desc << std::endl;
122 return EXIT_SUCCESS;
123 }
124
125 po::notify(vm);
126 }
127 catch (const po::error& e)
128 {
129 std::cerr << e.what() << std::endl << std::endl;
130 std::cerr << desc << std::endl;
131 return EXIT_FAILURE;
132 }
133
134 // Get the value of the switch arguments.
135 bool concurrent = vm["concurrent"].as<bool>();
136 bool enableProfiling = vm["event-based-profiling"].as<bool>();
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100137 bool enableLayerDetails = vm["visualize-optimized-model"].as<bool>();
Ruomei Yan2fcce082019-04-02 16:47:34 +0100138 bool enableFp16TurboMode = vm["fp16-turbo-mode"].as<bool>();
Narumol Prangnawarat610256f2019-06-26 15:10:46 +0100139 bool quantizeInput = vm["quantize-input"].as<bool>();
Matthew Jackson54658b92019-08-27 15:35:59 +0100140 bool printIntermediate = vm["print-intermediate-layers"].as<bool>();
telsoa01c577f2c2018-08-31 09:22:23 +0100141
142 // Check whether we have to load test cases from a file.
143 if (CheckOption(vm, "test-cases"))
144 {
145 // Check that the file exists.
146 if (!boost::filesystem::exists(testCasesFile))
147 {
148 BOOST_LOG_TRIVIAL(fatal) << "Given file \"" << testCasesFile << "\" does not exist";
149 return EXIT_FAILURE;
150 }
151
152 // Parse CSV file and extract test cases
153 armnnUtils::CsvReader reader;
154 std::vector<armnnUtils::CsvRow> testCases = reader.ParseFile(testCasesFile);
155
156 // Check that there is at least one test case to run
157 if (testCases.empty())
158 {
159 BOOST_LOG_TRIVIAL(fatal) << "Given file \"" << testCasesFile << "\" has no test cases";
160 return EXIT_FAILURE;
161 }
162
163 // Create runtime
164 armnn::IRuntime::CreationOptions options;
Nina Drozd549ae372018-09-10 14:26:44 +0100165 options.m_EnableGpuProfiling = enableProfiling;
166
telsoa01c577f2c2018-08-31 09:22:23 +0100167 std::shared_ptr<armnn::IRuntime> runtime(armnn::IRuntime::Create(options));
168
169 const std::string executableName("ExecuteNetwork");
170
171 // Check whether we need to run the test cases concurrently
172 if (concurrent)
173 {
174 std::vector<std::future<int>> results;
175 results.reserve(testCases.size());
176
177 // Run each test case in its own thread
178 for (auto& testCase : testCases)
179 {
180 testCase.values.insert(testCase.values.begin(), executableName);
Nina Drozd549ae372018-09-10 14:26:44 +0100181 results.push_back(std::async(std::launch::async, RunCsvTest, std::cref(testCase), std::cref(runtime),
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100182 enableProfiling, enableFp16TurboMode, thresholdTime, printIntermediate,
183 enableLayerDetails));
telsoa01c577f2c2018-08-31 09:22:23 +0100184 }
185
186 // Check results
187 for (auto& result : results)
188 {
189 if (result.get() != EXIT_SUCCESS)
190 {
191 return EXIT_FAILURE;
192 }
193 }
194 }
195 else
196 {
197 // Run tests sequentially
198 for (auto& testCase : testCases)
199 {
200 testCase.values.insert(testCase.values.begin(), executableName);
Matthew Jackson54658b92019-08-27 15:35:59 +0100201 if (RunCsvTest(testCase, runtime, enableProfiling,
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100202 enableFp16TurboMode, thresholdTime, printIntermediate,
203 enableLayerDetails) != EXIT_SUCCESS)
telsoa01c577f2c2018-08-31 09:22:23 +0100204 {
205 return EXIT_FAILURE;
206 }
207 }
208 }
209
210 return EXIT_SUCCESS;
211 }
212 else // Run single test
213 {
Aron Virginas-Tar382e21c2019-01-22 14:10:39 +0000214 // Get the preferred order of compute devices. If none are specified, default to using CpuRef
215 const std::string computeOption("compute");
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100216 std::vector<std::string> computeDevicesAsStrings =
217 CheckOption(vm, computeOption.c_str()) ?
218 vm[computeOption].as<std::vector<std::string>>() :
219 std::vector<std::string>();
Matteo Martincigh067112f2018-10-29 11:01:09 +0000220 std::vector<armnn::BackendId> computeDevices(computeDevicesAsStrings.begin(), computeDevicesAsStrings.end());
telsoa01c577f2c2018-08-31 09:22:23 +0100221
222 // Remove duplicates from the list of compute devices.
223 RemoveDuplicateDevices(computeDevices);
224
telsoa01c577f2c2018-08-31 09:22:23 +0100225 try
226 {
227 CheckOptionDependencies(vm);
228 }
229 catch (const po::error& e)
230 {
231 std::cerr << e.what() << std::endl << std::endl;
232 std::cerr << desc << std::endl;
233 return EXIT_FAILURE;
234 }
235
Matteo Martincigh00dda4a2019-08-14 11:42:30 +0100236 return RunTest(modelFormat, inputTensorShapes, computeDevices, dynamicBackendsPath, modelPath, inputNames,
Narumol Prangnawarat610256f2019-06-26 15:10:46 +0100237 inputTensorDataFilePaths, inputTypes, quantizeInput, outputTypes, outputNames,
Sadik Armagan77086282019-09-02 11:46:28 +0100238 outputTensorFiles, enableProfiling, enableFp16TurboMode, thresholdTime, printIntermediate,
Andre Ghattas23ae2ea2019-08-07 12:18:38 +0100239 subgraphId, enableLayerDetails);
telsoa014fcda012018-03-09 14:13:49 +0000240 }
241}