blob: 0add6f594d560f0fa4ff51a6d4600a10a4b7cb08 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5
6#include "NeonLayerSupport.hpp"
7
telsoa014fcda012018-03-09 14:13:49 +00008#include <armnn/Descriptors.hpp>
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01009#include <armnn/InternalTypes.hpp>
10#include <armnn/LayerSupportCommon.hpp>
telsoa014fcda012018-03-09 14:13:49 +000011#include <armnn/Tensor.hpp>
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010012#include <armnn/Types.hpp>
telsoa014fcda012018-03-09 14:13:49 +000013
14#include <boost/core/ignore_unused.hpp>
15
16#ifdef ARMCOMPUTENEON_ENABLED
David Beck0dbe0ee2018-09-24 15:59:27 +010017#include "workloads/NeonAdditionFloatWorkload.hpp"
Nattapat Chaimanowongd4b70592018-10-12 11:21:49 +010018#include "workloads/NeonActivationWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010019#include "workloads/NeonBatchNormalizationFloatWorkload.hpp"
Nattapat Chaimanowong974b65f2018-10-15 15:07:34 +010020#include "workloads/NeonConvolution2dWorkload.hpp"
Nattapat Chaimanowong77140882018-10-17 11:12:19 +010021#include "workloads/NeonDepthwiseConvolutionWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010022#include "workloads/NeonL2NormalizationFloatWorkload.hpp"
23#include "workloads/NeonMultiplicationFloatWorkload.hpp"
24#include "workloads/NeonNormalizationFloatWorkload.hpp"
25#include "workloads/NeonFullyConnectedWorkload.hpp"
26#include "workloads/NeonPermuteWorkload.hpp"
Nattapat Chaimanowong5d2e7002018-10-12 16:03:56 +010027#include "workloads/NeonPooling2dWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010028#include "workloads/NeonSoftmaxBaseWorkload.hpp"
29#include "workloads/NeonSubtractionFloatWorkload.hpp"
telsoa014fcda012018-03-09 14:13:49 +000030#endif
31
32using namespace boost;
33
34namespace armnn
35{
telsoa014fcda012018-03-09 14:13:49 +000036
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010037namespace
arovir017ff76c52018-10-09 09:40:58 +010038{
telsoa014fcda012018-03-09 14:13:49 +000039
arovir01085f0a42018-10-08 14:48:19 +010040bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000041{
42#if ARMCOMPUTENEON_ENABLED
43 return true;
44#else
arovir01085f0a42018-10-08 14:48:19 +010045 if (reasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000046 {
arovir01085f0a42018-10-08 14:48:19 +010047 reasonIfUnsupported.value() = "The armnn library has been built without NEON support";
telsoa014fcda012018-03-09 14:13:49 +000048 }
49 return false;
50#endif
51}
52
telsoa01c577f2c2018-08-31 09:22:23 +010053template<typename FloatFunc, typename Uint8Func, typename ... Params>
arovir01085f0a42018-10-08 14:48:19 +010054bool IsSupportedForDataTypeNeon(Optional<std::string&> reasonIfUnsupported,
telsoa014fcda012018-03-09 14:13:49 +000055 DataType dataType,
telsoa01c577f2c2018-08-31 09:22:23 +010056 FloatFunc floatFuncPtr,
telsoa014fcda012018-03-09 14:13:49 +000057 Uint8Func uint8FuncPtr,
58 Params&&... params)
59{
60 return IsNeonBackendSupported(reasonIfUnsupported) &&
61 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
62 dataType,
63 floatFuncPtr,
telsoa01c577f2c2018-08-31 09:22:23 +010064 floatFuncPtr,
telsoa014fcda012018-03-09 14:13:49 +000065 uint8FuncPtr,
66 std::forward<Params>(params)...);
67}
68
69#if ARMCOMPUTENEON_ENABLED
70template<class FuncType, class... Args>
arovir01085f0a42018-10-08 14:48:19 +010071inline bool IsWorkloadSupported(FuncType& func, Optional<std::string&> reasonIfUnsupported, Args&&... args)
telsoa014fcda012018-03-09 14:13:49 +000072{
73 arm_compute::Status aclStatus = func(std::forward<Args>(args)...);
74 const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
75 if (!supported && reasonIfUnsupported)
76 {
arovir01085f0a42018-10-08 14:48:19 +010077 reasonIfUnsupported.value() = aclStatus.error_description();
telsoa014fcda012018-03-09 14:13:49 +000078 }
79 return supported;
80}
81
82#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
83 return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__);
84#else
85#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
86 return IsNeonBackendSupported(reasonIfUnsupported);
87#endif
88
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010089} // anonymous namespace
90
91bool NeonLayerSupport::IsActivationSupported(const TensorInfo& input,
92 const TensorInfo& output,
93 const ActivationDescriptor& descriptor,
94 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +000095{
96 ignore_unused(descriptor);
telsoa01c577f2c2018-08-31 09:22:23 +010097 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonActivationWorkloadValidate,
98 reasonIfUnsupported,
99 input,
100 output,
101 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000102}
103
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100104bool NeonLayerSupport::IsAdditionSupported(const TensorInfo& input0,
105 const TensorInfo& input1,
106 const TensorInfo& output,
107 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000108{
telsoa01c577f2c2018-08-31 09:22:23 +0100109 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonAdditionWorkloadValidate,
110 reasonIfUnsupported,
111 input0,
112 input1,
113 output);
telsoa014fcda012018-03-09 14:13:49 +0000114}
115
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100116bool NeonLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input,
117 const TensorInfo& output,
118 const TensorInfo& mean,
119 const TensorInfo& var,
120 const TensorInfo& beta,
121 const TensorInfo& gamma,
122 const BatchNormalizationDescriptor& descriptor,
123 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000124{
telsoa01c577f2c2018-08-31 09:22:23 +0100125 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonBatchNormalizationValidate,
126 reasonIfUnsupported,
127 input,
128 output,
129 mean,
130 var,
131 beta,
132 gamma,
133 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000134}
135
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100136bool NeonLayerSupport::IsConstantSupported(const TensorInfo& output,
137 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000138{
139 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
140 output.GetDataType(),
141 &TrueFunc<>,
142 &TrueFunc<>);
143}
144
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100145bool NeonLayerSupport::IsConvertFp16ToFp32Supported(const TensorInfo& input,
146 const TensorInfo& output,
147 Optional<std::string&> reasonIfUnsupported) const
148{
149 ignore_unused(input);
150 ignore_unused(output);
151 ignore_unused(reasonIfUnsupported);
152 return true;
153}
154
155bool NeonLayerSupport::IsConvertFp32ToFp16Supported(const TensorInfo& input,
156 const TensorInfo& output,
157 Optional<std::string&> reasonIfUnsupported) const
158{
159 ignore_unused(input);
160 ignore_unused(output);
161 ignore_unused(reasonIfUnsupported);
162 return true;
163}
164
165bool NeonLayerSupport::IsConvolution2dSupported(const TensorInfo& input,
166 const TensorInfo& output,
167 const Convolution2dDescriptor& descriptor,
168 const TensorInfo& weights,
169 const Optional<TensorInfo>& biases,
170 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000171{
surmeh013537c2c2018-05-18 16:31:43 +0100172 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvolution2dWorkloadValidate,
173 reasonIfUnsupported,
174 input,
175 output,
176 descriptor,
177 weights,
178 biases);
telsoa014fcda012018-03-09 14:13:49 +0000179}
180
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100181bool NeonLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input,
182 const TensorInfo& output,
183 const DepthwiseConvolution2dDescriptor& descriptor,
184 const TensorInfo& weights,
185 const Optional<TensorInfo>& biases,
186 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000187{
telsoa01c577f2c2018-08-31 09:22:23 +0100188 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDepthwiseConvolutionWorkloadValidate,
189 reasonIfUnsupported,
190 input,
191 output,
192 descriptor,
193 weights,
194 biases);
telsoa014fcda012018-03-09 14:13:49 +0000195}
196
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100197bool NeonLayerSupport::IsDivisionSupported(const TensorInfo& input0,
198 const TensorInfo& input1,
199 const TensorInfo& output,
200 Optional<std::string&> reasonIfUnsupported) const
Francis Murtaghe7a86a42018-08-29 12:42:10 +0100201{
arovir01085f0a42018-10-08 14:48:19 +0100202 ignore_unused(input0);
203 ignore_unused(input1);
204 ignore_unused(output);
205 ignore_unused(reasonIfUnsupported);
Francis Murtaghe7a86a42018-08-29 12:42:10 +0100206 return false;
207}
208
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100209bool NeonLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input,
210 const FakeQuantizationDescriptor& descriptor,
211 Optional<std::string&> reasonIfUnsupported) const
David Beckc2044fe2018-09-05 15:00:38 +0100212{
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100213 ignore_unused(input);
214 ignore_unused(descriptor);
215 ignore_unused(reasonIfUnsupported);
216 return false;
David Beckc2044fe2018-09-05 15:00:38 +0100217}
218
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100219bool NeonLayerSupport::IsFloorSupported(const TensorInfo& input,
220 const TensorInfo& output,
221 Optional<std::string&> reasonIfUnsupported) const
222{
223 ignore_unused(output);
224 return IsNeonBackendSupported(reasonIfUnsupported) &&
225 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
226 input.GetDataType(),
227 &FalseFuncF16<>,
228 &TrueFunc<>,
229 &FalseFuncU8<>);
230}
231
232bool NeonLayerSupport::IsFullyConnectedSupported(const TensorInfo& input,
233 const TensorInfo& output,
234 const TensorInfo& weights,
235 const TensorInfo& biases,
236 const FullyConnectedDescriptor& descriptor,
237 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000238{
telsoa01c577f2c2018-08-31 09:22:23 +0100239 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonFullyConnectedWorkloadValidate,
240 reasonIfUnsupported,
241 input,
242 output,
243 weights,
244 biases,
245 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000246}
247
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100248bool NeonLayerSupport::IsInputSupported(const TensorInfo& input,
249 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000250{
251 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
252 input.GetDataType(),
253 &TrueFunc<>,
254 &TrueFunc<>);
255}
256
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100257bool NeonLayerSupport::IsL2NormalizationSupported(const TensorInfo& input,
258 const TensorInfo& output,
259 const L2NormalizationDescriptor& descriptor,
260 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000261{
Matteo Martincighbcd3c852018-09-28 14:14:12 +0100262 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonL2NormalizationWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000263}
264
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100265bool NeonLayerSupport::IsLstmSupported(const TensorInfo& input,
266 const TensorInfo& outputStateIn,
267 const TensorInfo& cellStateIn,
268 const TensorInfo& scratchBuffer,
269 const TensorInfo& outputStateOut,
270 const TensorInfo& cellStateOut,
271 const TensorInfo& output,
272 const LstmDescriptor& descriptor,
273 const TensorInfo& inputToForgetWeights,
274 const TensorInfo& inputToCellWeights,
275 const TensorInfo& inputToOutputWeights,
276 const TensorInfo& recurrentToForgetWeights,
277 const TensorInfo& recurrentToCellWeights,
278 const TensorInfo& recurrentToOutputWeights,
279 const TensorInfo& forgetGateBias,
280 const TensorInfo& cellBias,
281 const TensorInfo& outputGateBias,
282 const TensorInfo* inputToInputWeights,
283 const TensorInfo* recurrentToInputWeights,
284 const TensorInfo* cellToInputWeights,
285 const TensorInfo* inputGateBias,
286 const TensorInfo* projectionWeights,
287 const TensorInfo* projectionBias,
288 const TensorInfo* cellToForgetWeights,
289 const TensorInfo* cellToOutputWeights,
290 Optional<std::string&> reasonIfUnsupported) const
telsoa01c577f2c2018-08-31 09:22:23 +0100291{
292 ignore_unused(input);
293 ignore_unused(outputStateIn);
294 ignore_unused(cellStateIn);
295 ignore_unused(scratchBuffer);
296 ignore_unused(outputStateOut);
297 ignore_unused(cellStateOut);
298 ignore_unused(output);
299 ignore_unused(descriptor);
300 ignore_unused(inputToForgetWeights);
301 ignore_unused(inputToCellWeights);
302 ignore_unused(inputToOutputWeights);
303 ignore_unused(recurrentToForgetWeights);
304 ignore_unused(recurrentToCellWeights);
305 ignore_unused(recurrentToOutputWeights);
306 ignore_unused(forgetGateBias);
307 ignore_unused(cellBias);
308 ignore_unused(outputGateBias);
309 ignore_unused(inputToInputWeights);
310 ignore_unused(recurrentToInputWeights);
311 ignore_unused(cellToInputWeights);
312 ignore_unused(inputGateBias);
313 ignore_unused(projectionWeights);
314 ignore_unused(projectionBias);
315 ignore_unused(cellToForgetWeights);
316 ignore_unused(cellToOutputWeights);
arovir01085f0a42018-10-08 14:48:19 +0100317 ignore_unused(reasonIfUnsupported);
telsoa01c577f2c2018-08-31 09:22:23 +0100318 return false;
319}
320
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100321bool NeonLayerSupport::IsMeanSupported(const TensorInfo& input,
322 const TensorInfo& output,
323 const MeanDescriptor& descriptor,
324 Optional<std::string&> reasonIfUnsupported) const
narpra0132b90462018-09-13 11:07:48 +0100325{
arovir01085f0a42018-10-08 14:48:19 +0100326 ignore_unused(input);
327 ignore_unused(output);
328 ignore_unused(descriptor);
329 ignore_unused(reasonIfUnsupported);
narpra0132b90462018-09-13 11:07:48 +0100330 return false;
331}
332
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100333bool NeonLayerSupport::IsMergerSupported(const std::vector<const TensorInfo*> inputs,
334 const OriginsDescriptor& descriptor,
335 Optional<std::string&> reasonIfUnsupported) const
336{
337 ignore_unused(descriptor);
338 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
339 inputs[0]->GetDataType(),
340 &TrueFunc<>,
341 &TrueFunc<>);
342}
343
344bool NeonLayerSupport::IsMultiplicationSupported(const TensorInfo& input0,
345 const TensorInfo& input1,
346 const TensorInfo& output,
347 Optional<std::string&> reasonIfUnsupported) const
348{
349 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMultiplicationWorkloadValidate,
350 reasonIfUnsupported,
351 input0,
352 input1,
353 output);
354}
355
356bool NeonLayerSupport::IsNormalizationSupported(const TensorInfo& input,
357 const TensorInfo& output,
358 const NormalizationDescriptor& descriptor,
359 Optional<std::string&> reasonIfUnsupported) const
360{
361 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonNormalizationWorkloadValidate,
362 reasonIfUnsupported,
363 input,
364 output,
365 descriptor);
366}
367
368bool NeonLayerSupport::IsOutputSupported(const TensorInfo& output,
369 Optional<std::string&> reasonIfUnsupported) const
370{
371 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
372 output.GetDataType(),
373 &TrueFunc<>,
374 &TrueFunc<>);
375}
376
377bool NeonLayerSupport::IsPadSupported(const TensorInfo& input,
378 const TensorInfo& output,
379 const PadDescriptor& descriptor,
380 Optional<std::string&> reasonIfUnsupported) const
Nina Drozd661dfa72018-10-02 11:14:17 +0100381{
arovir01085f0a42018-10-08 14:48:19 +0100382 ignore_unused(input);
383 ignore_unused(output);
384 ignore_unused(descriptor);
385 ignore_unused(reasonIfUnsupported);
Nina Drozd661dfa72018-10-02 11:14:17 +0100386 return false;
387}
388
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100389bool NeonLayerSupport::IsPermuteSupported(const TensorInfo& input,
390 const TensorInfo& output,
391 const PermuteDescriptor& descriptor,
392 Optional<std::string&> reasonIfUnsupported) const
393{
394 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPermuteWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000395}
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100396
397bool NeonLayerSupport::IsPooling2dSupported(const TensorInfo& input,
398 const TensorInfo& output,
399 const Pooling2dDescriptor& descriptor,
400 Optional<std::string&> reasonIfUnsupported) const
401{
402 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling2dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
403}
404
405bool NeonLayerSupport::IsReshapeSupported(const TensorInfo& input,
406 Optional<std::string&> reasonIfUnsupported) const
407{
408 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
409 input.GetDataType(),
410 &TrueFunc<>,
411 &TrueFunc<>);
412}
413
414bool NeonLayerSupport::IsResizeBilinearSupported(const TensorInfo& input,
415 Optional<std::string&> reasonIfUnsupported) const
416{
417 ignore_unused(input);
418 ignore_unused(reasonIfUnsupported);
419 return false;
420}
421
422bool NeonLayerSupport::IsSoftmaxSupported(const TensorInfo& input,
423 const TensorInfo& output,
424 const SoftmaxDescriptor& descriptor,
425 Optional<std::string&> reasonIfUnsupported) const
426{
427 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
428}
429
430bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input,
431 const ViewsDescriptor& descriptor,
432 Optional<std::string&> reasonIfUnsupported) const
433{
434 ignore_unused(descriptor);
435 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
436 input.GetDataType(),
437 &TrueFunc<>,
438 &TrueFunc<>);
439}
440
441bool NeonLayerSupport::IsSubtractionSupported(const TensorInfo& input0,
442 const TensorInfo& input1,
443 const TensorInfo& output,
444 Optional<std::string&> reasonIfUnsupported) const
445{
446 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSubtractionWorkloadValidate,
447 reasonIfUnsupported,
448 input0,
449 input1,
450 output);
451}
452
453bool IsNeonDirectConvolutionPreferred(const TensorInfo& weightInfo, const Convolution2dDescriptor& desc)
454{
455 // See arm_compute::NEDirectConvolutionLayer documentation for the supported cases,
456 // and complement with NEDirectConvolutionLayerKernel::configure() implementation.
457
458 // Only 1x1 is using direct convolution. Performance results and details are in:
459 // https://jira.arm.com/browse/IVGCVSW-1003
460 // Measurements were taken as of clframework: f105ab972135bcd21304883eff040d7e587099bc
461
462 const bool dataTypeSupported = (weightInfo.GetDataType() == armnn::DataType::Float32);
463
464 // Strides: 1|2|3
465 const bool strideSupported = (desc.m_StrideX == 1 || desc.m_StrideX == 2 || desc.m_StrideX == 3) &&
466 (desc.m_StrideY == 1 || desc.m_StrideY == 2 || desc.m_StrideY == 3);
467
468 auto paddingLargerThan = [](const Convolution2dDescriptor& conv2ddesc, unsigned int value)
469 {
470 return conv2ddesc.m_PadLeft > value || conv2ddesc.m_PadRight > value ||
471 conv2ddesc.m_PadTop > value || conv2ddesc.m_PadBottom > value;
472 };
473
474 // Supported sizes and padding.
475 const bool sizeAndPaddingSupported =
476 // Pad > 0 not supported for 1x1 weights.
477 (weightInfo.GetShape()[2] == 1 && weightInfo.GetShape()[3] == 1 && !paddingLargerThan(desc, 0u));
478
479 const bool preferDirectConvolution = dataTypeSupported &&
480 strideSupported &&
481 sizeAndPaddingSupported &&
482 // NEDirectConvolutionLayerKernel doesn't support NULL bias.
483 desc.m_BiasEnabled;
484 return preferDirectConvolution;
485}
486
487} // namespace armnn