blob: a4a6b67c0b12f250d997b430c1ee73598c2d05d6 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5
6#include "NeonLayerSupport.hpp"
David Beck3e9e1152018-10-17 14:17:50 +01007#include "NeonBackendId.hpp"
telsoa014fcda012018-03-09 14:13:49 +00008
telsoa014fcda012018-03-09 14:13:49 +00009#include <armnn/Descriptors.hpp>
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000010#include <InternalTypes.hpp>
11#include <LayerSupportCommon.hpp>
telsoa014fcda012018-03-09 14:13:49 +000012#include <armnn/Tensor.hpp>
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010013#include <armnn/Types.hpp>
telsoa014fcda012018-03-09 14:13:49 +000014
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000015#include <backendsCommon/LayerSupportRegistry.hpp>
David Beck3e9e1152018-10-17 14:17:50 +010016
telsoa014fcda012018-03-09 14:13:49 +000017#include <boost/core/ignore_unused.hpp>
18
19#ifdef ARMCOMPUTENEON_ENABLED
David Beck0dbe0ee2018-09-24 15:59:27 +010020#include "workloads/NeonAdditionFloatWorkload.hpp"
Nattapat Chaimanowongd4b70592018-10-12 11:21:49 +010021#include "workloads/NeonActivationWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010022#include "workloads/NeonBatchNormalizationFloatWorkload.hpp"
Nattapat Chaimanowong974b65f2018-10-15 15:07:34 +010023#include "workloads/NeonConvolution2dWorkload.hpp"
Nattapat Chaimanowong77140882018-10-17 11:12:19 +010024#include "workloads/NeonDepthwiseConvolutionWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010025#include "workloads/NeonL2NormalizationFloatWorkload.hpp"
26#include "workloads/NeonMultiplicationFloatWorkload.hpp"
27#include "workloads/NeonNormalizationFloatWorkload.hpp"
28#include "workloads/NeonFullyConnectedWorkload.hpp"
29#include "workloads/NeonPermuteWorkload.hpp"
Nattapat Chaimanowong5d2e7002018-10-12 16:03:56 +010030#include "workloads/NeonPooling2dWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010031#include "workloads/NeonSoftmaxBaseWorkload.hpp"
32#include "workloads/NeonSubtractionFloatWorkload.hpp"
telsoa014fcda012018-03-09 14:13:49 +000033#endif
34
35using namespace boost;
36
37namespace armnn
38{
telsoa014fcda012018-03-09 14:13:49 +000039
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010040namespace
arovir017ff76c52018-10-09 09:40:58 +010041{
telsoa014fcda012018-03-09 14:13:49 +000042
David Beck3e9e1152018-10-17 14:17:50 +010043ILayerSupportSharedPtr GetLayerSupportPointer()
44{
45 static ILayerSupportSharedPtr instance{new NeonLayerSupport};
46 return instance;
47}
48
49static StaticRegistryInitializer<LayerSupportRegistry> g_RegisterHelper{
50 LayerSupportRegistryInstance(),
51 NeonBackendId(),
David Beck9efb57d2018-11-05 13:40:33 +000052 []()
David Beck3e9e1152018-10-17 14:17:50 +010053 {
54 return GetLayerSupportPointer();
55 }
56};
57
arovir01085f0a42018-10-08 14:48:19 +010058bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000059{
60#if ARMCOMPUTENEON_ENABLED
61 return true;
62#else
arovir01085f0a42018-10-08 14:48:19 +010063 if (reasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000064 {
arovir01085f0a42018-10-08 14:48:19 +010065 reasonIfUnsupported.value() = "The armnn library has been built without NEON support";
telsoa014fcda012018-03-09 14:13:49 +000066 }
67 return false;
68#endif
69}
70
telsoa01c577f2c2018-08-31 09:22:23 +010071template<typename FloatFunc, typename Uint8Func, typename ... Params>
arovir01085f0a42018-10-08 14:48:19 +010072bool IsSupportedForDataTypeNeon(Optional<std::string&> reasonIfUnsupported,
telsoa014fcda012018-03-09 14:13:49 +000073 DataType dataType,
telsoa01c577f2c2018-08-31 09:22:23 +010074 FloatFunc floatFuncPtr,
telsoa014fcda012018-03-09 14:13:49 +000075 Uint8Func uint8FuncPtr,
76 Params&&... params)
77{
78 return IsNeonBackendSupported(reasonIfUnsupported) &&
79 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
80 dataType,
81 floatFuncPtr,
telsoa01c577f2c2018-08-31 09:22:23 +010082 floatFuncPtr,
telsoa014fcda012018-03-09 14:13:49 +000083 uint8FuncPtr,
84 std::forward<Params>(params)...);
85}
86
87#if ARMCOMPUTENEON_ENABLED
88template<class FuncType, class... Args>
arovir01085f0a42018-10-08 14:48:19 +010089inline bool IsWorkloadSupported(FuncType& func, Optional<std::string&> reasonIfUnsupported, Args&&... args)
telsoa014fcda012018-03-09 14:13:49 +000090{
91 arm_compute::Status aclStatus = func(std::forward<Args>(args)...);
92 const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
93 if (!supported && reasonIfUnsupported)
94 {
arovir01085f0a42018-10-08 14:48:19 +010095 reasonIfUnsupported.value() = aclStatus.error_description();
telsoa014fcda012018-03-09 14:13:49 +000096 }
97 return supported;
98}
99
100#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
101 return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__);
102#else
103#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
104 return IsNeonBackendSupported(reasonIfUnsupported);
105#endif
106
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100107} // anonymous namespace
108
109bool NeonLayerSupport::IsActivationSupported(const TensorInfo& input,
110 const TensorInfo& output,
111 const ActivationDescriptor& descriptor,
112 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000113{
114 ignore_unused(descriptor);
telsoa01c577f2c2018-08-31 09:22:23 +0100115 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonActivationWorkloadValidate,
116 reasonIfUnsupported,
117 input,
118 output,
119 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000120}
121
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100122bool NeonLayerSupport::IsAdditionSupported(const TensorInfo& input0,
123 const TensorInfo& input1,
124 const TensorInfo& output,
125 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000126{
telsoa01c577f2c2018-08-31 09:22:23 +0100127 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonAdditionWorkloadValidate,
128 reasonIfUnsupported,
129 input0,
130 input1,
131 output);
telsoa014fcda012018-03-09 14:13:49 +0000132}
133
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100134bool NeonLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input,
135 const TensorInfo& output,
136 const TensorInfo& mean,
137 const TensorInfo& var,
138 const TensorInfo& beta,
139 const TensorInfo& gamma,
140 const BatchNormalizationDescriptor& descriptor,
141 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000142{
telsoa01c577f2c2018-08-31 09:22:23 +0100143 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonBatchNormalizationValidate,
144 reasonIfUnsupported,
145 input,
146 output,
147 mean,
148 var,
149 beta,
150 gamma,
151 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000152}
153
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100154bool NeonLayerSupport::IsConstantSupported(const TensorInfo& output,
155 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000156{
157 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
158 output.GetDataType(),
159 &TrueFunc<>,
160 &TrueFunc<>);
161}
162
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100163bool NeonLayerSupport::IsConvertFp16ToFp32Supported(const TensorInfo& input,
164 const TensorInfo& output,
165 Optional<std::string&> reasonIfUnsupported) const
166{
167 ignore_unused(input);
168 ignore_unused(output);
169 ignore_unused(reasonIfUnsupported);
170 return true;
171}
172
173bool NeonLayerSupport::IsConvertFp32ToFp16Supported(const TensorInfo& input,
174 const TensorInfo& output,
175 Optional<std::string&> reasonIfUnsupported) const
176{
177 ignore_unused(input);
178 ignore_unused(output);
179 ignore_unused(reasonIfUnsupported);
180 return true;
181}
182
183bool NeonLayerSupport::IsConvolution2dSupported(const TensorInfo& input,
184 const TensorInfo& output,
185 const Convolution2dDescriptor& descriptor,
186 const TensorInfo& weights,
187 const Optional<TensorInfo>& biases,
188 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000189{
surmeh013537c2c2018-05-18 16:31:43 +0100190 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvolution2dWorkloadValidate,
191 reasonIfUnsupported,
192 input,
193 output,
194 descriptor,
195 weights,
196 biases);
telsoa014fcda012018-03-09 14:13:49 +0000197}
198
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100199bool NeonLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input,
200 const TensorInfo& output,
201 const DepthwiseConvolution2dDescriptor& descriptor,
202 const TensorInfo& weights,
203 const Optional<TensorInfo>& biases,
204 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000205{
telsoa01c577f2c2018-08-31 09:22:23 +0100206 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDepthwiseConvolutionWorkloadValidate,
207 reasonIfUnsupported,
208 input,
209 output,
210 descriptor,
211 weights,
212 biases);
telsoa014fcda012018-03-09 14:13:49 +0000213}
214
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100215bool NeonLayerSupport::IsDivisionSupported(const TensorInfo& input0,
216 const TensorInfo& input1,
217 const TensorInfo& output,
218 Optional<std::string&> reasonIfUnsupported) const
Francis Murtaghe7a86a42018-08-29 12:42:10 +0100219{
arovir01085f0a42018-10-08 14:48:19 +0100220 ignore_unused(input0);
221 ignore_unused(input1);
222 ignore_unused(output);
223 ignore_unused(reasonIfUnsupported);
Francis Murtaghe7a86a42018-08-29 12:42:10 +0100224 return false;
225}
226
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100227bool NeonLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input,
228 const FakeQuantizationDescriptor& descriptor,
229 Optional<std::string&> reasonIfUnsupported) const
David Beckc2044fe2018-09-05 15:00:38 +0100230{
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100231 ignore_unused(input);
232 ignore_unused(descriptor);
233 ignore_unused(reasonIfUnsupported);
234 return false;
David Beckc2044fe2018-09-05 15:00:38 +0100235}
236
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100237bool NeonLayerSupport::IsFloorSupported(const TensorInfo& input,
238 const TensorInfo& output,
239 Optional<std::string&> reasonIfUnsupported) const
240{
241 ignore_unused(output);
242 return IsNeonBackendSupported(reasonIfUnsupported) &&
243 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
244 input.GetDataType(),
245 &FalseFuncF16<>,
246 &TrueFunc<>,
247 &FalseFuncU8<>);
248}
249
250bool NeonLayerSupport::IsFullyConnectedSupported(const TensorInfo& input,
251 const TensorInfo& output,
252 const TensorInfo& weights,
253 const TensorInfo& biases,
254 const FullyConnectedDescriptor& descriptor,
255 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000256{
telsoa01c577f2c2018-08-31 09:22:23 +0100257 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonFullyConnectedWorkloadValidate,
258 reasonIfUnsupported,
259 input,
260 output,
261 weights,
262 biases,
263 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000264}
265
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100266bool NeonLayerSupport::IsInputSupported(const TensorInfo& input,
267 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000268{
269 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
270 input.GetDataType(),
271 &TrueFunc<>,
272 &TrueFunc<>);
273}
274
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100275bool NeonLayerSupport::IsL2NormalizationSupported(const TensorInfo& input,
276 const TensorInfo& output,
277 const L2NormalizationDescriptor& descriptor,
278 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000279{
Matteo Martincighbcd3c852018-09-28 14:14:12 +0100280 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonL2NormalizationWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000281}
282
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100283bool NeonLayerSupport::IsLstmSupported(const TensorInfo& input,
284 const TensorInfo& outputStateIn,
285 const TensorInfo& cellStateIn,
286 const TensorInfo& scratchBuffer,
287 const TensorInfo& outputStateOut,
288 const TensorInfo& cellStateOut,
289 const TensorInfo& output,
290 const LstmDescriptor& descriptor,
291 const TensorInfo& inputToForgetWeights,
292 const TensorInfo& inputToCellWeights,
293 const TensorInfo& inputToOutputWeights,
294 const TensorInfo& recurrentToForgetWeights,
295 const TensorInfo& recurrentToCellWeights,
296 const TensorInfo& recurrentToOutputWeights,
297 const TensorInfo& forgetGateBias,
298 const TensorInfo& cellBias,
299 const TensorInfo& outputGateBias,
300 const TensorInfo* inputToInputWeights,
301 const TensorInfo* recurrentToInputWeights,
302 const TensorInfo* cellToInputWeights,
303 const TensorInfo* inputGateBias,
304 const TensorInfo* projectionWeights,
305 const TensorInfo* projectionBias,
306 const TensorInfo* cellToForgetWeights,
307 const TensorInfo* cellToOutputWeights,
308 Optional<std::string&> reasonIfUnsupported) const
telsoa01c577f2c2018-08-31 09:22:23 +0100309{
310 ignore_unused(input);
311 ignore_unused(outputStateIn);
312 ignore_unused(cellStateIn);
313 ignore_unused(scratchBuffer);
314 ignore_unused(outputStateOut);
315 ignore_unused(cellStateOut);
316 ignore_unused(output);
317 ignore_unused(descriptor);
318 ignore_unused(inputToForgetWeights);
319 ignore_unused(inputToCellWeights);
320 ignore_unused(inputToOutputWeights);
321 ignore_unused(recurrentToForgetWeights);
322 ignore_unused(recurrentToCellWeights);
323 ignore_unused(recurrentToOutputWeights);
324 ignore_unused(forgetGateBias);
325 ignore_unused(cellBias);
326 ignore_unused(outputGateBias);
327 ignore_unused(inputToInputWeights);
328 ignore_unused(recurrentToInputWeights);
329 ignore_unused(cellToInputWeights);
330 ignore_unused(inputGateBias);
331 ignore_unused(projectionWeights);
332 ignore_unused(projectionBias);
333 ignore_unused(cellToForgetWeights);
334 ignore_unused(cellToOutputWeights);
arovir01085f0a42018-10-08 14:48:19 +0100335 ignore_unused(reasonIfUnsupported);
telsoa01c577f2c2018-08-31 09:22:23 +0100336 return false;
337}
338
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100339bool NeonLayerSupport::IsMeanSupported(const TensorInfo& input,
340 const TensorInfo& output,
341 const MeanDescriptor& descriptor,
342 Optional<std::string&> reasonIfUnsupported) const
narpra0132b90462018-09-13 11:07:48 +0100343{
arovir01085f0a42018-10-08 14:48:19 +0100344 ignore_unused(input);
345 ignore_unused(output);
346 ignore_unused(descriptor);
347 ignore_unused(reasonIfUnsupported);
narpra0132b90462018-09-13 11:07:48 +0100348 return false;
349}
350
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100351bool NeonLayerSupport::IsMergerSupported(const std::vector<const TensorInfo*> inputs,
352 const OriginsDescriptor& descriptor,
353 Optional<std::string&> reasonIfUnsupported) const
354{
355 ignore_unused(descriptor);
356 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
357 inputs[0]->GetDataType(),
358 &TrueFunc<>,
359 &TrueFunc<>);
360}
361
362bool NeonLayerSupport::IsMultiplicationSupported(const TensorInfo& input0,
363 const TensorInfo& input1,
364 const TensorInfo& output,
365 Optional<std::string&> reasonIfUnsupported) const
366{
367 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMultiplicationWorkloadValidate,
368 reasonIfUnsupported,
369 input0,
370 input1,
371 output);
372}
373
374bool NeonLayerSupport::IsNormalizationSupported(const TensorInfo& input,
375 const TensorInfo& output,
376 const NormalizationDescriptor& descriptor,
377 Optional<std::string&> reasonIfUnsupported) const
378{
379 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonNormalizationWorkloadValidate,
380 reasonIfUnsupported,
381 input,
382 output,
383 descriptor);
384}
385
386bool NeonLayerSupport::IsOutputSupported(const TensorInfo& output,
387 Optional<std::string&> reasonIfUnsupported) const
388{
389 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
390 output.GetDataType(),
391 &TrueFunc<>,
392 &TrueFunc<>);
393}
394
395bool NeonLayerSupport::IsPadSupported(const TensorInfo& input,
396 const TensorInfo& output,
397 const PadDescriptor& descriptor,
398 Optional<std::string&> reasonIfUnsupported) const
Nina Drozd661dfa72018-10-02 11:14:17 +0100399{
arovir01085f0a42018-10-08 14:48:19 +0100400 ignore_unused(input);
401 ignore_unused(output);
402 ignore_unused(descriptor);
403 ignore_unused(reasonIfUnsupported);
Nina Drozd661dfa72018-10-02 11:14:17 +0100404 return false;
405}
406
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100407bool NeonLayerSupport::IsPermuteSupported(const TensorInfo& input,
408 const TensorInfo& output,
409 const PermuteDescriptor& descriptor,
410 Optional<std::string&> reasonIfUnsupported) const
411{
412 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPermuteWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000413}
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100414
415bool NeonLayerSupport::IsPooling2dSupported(const TensorInfo& input,
416 const TensorInfo& output,
417 const Pooling2dDescriptor& descriptor,
418 Optional<std::string&> reasonIfUnsupported) const
419{
420 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling2dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
421}
422
423bool NeonLayerSupport::IsReshapeSupported(const TensorInfo& input,
424 Optional<std::string&> reasonIfUnsupported) const
425{
426 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
427 input.GetDataType(),
428 &TrueFunc<>,
429 &TrueFunc<>);
430}
431
432bool NeonLayerSupport::IsResizeBilinearSupported(const TensorInfo& input,
433 Optional<std::string&> reasonIfUnsupported) const
434{
435 ignore_unused(input);
436 ignore_unused(reasonIfUnsupported);
437 return false;
438}
439
440bool NeonLayerSupport::IsSoftmaxSupported(const TensorInfo& input,
441 const TensorInfo& output,
442 const SoftmaxDescriptor& descriptor,
443 Optional<std::string&> reasonIfUnsupported) const
444{
445 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
446}
447
448bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input,
449 const ViewsDescriptor& descriptor,
450 Optional<std::string&> reasonIfUnsupported) const
451{
452 ignore_unused(descriptor);
453 return IsSupportedForDataTypeNeon(reasonIfUnsupported,
454 input.GetDataType(),
455 &TrueFunc<>,
456 &TrueFunc<>);
457}
458
459bool NeonLayerSupport::IsSubtractionSupported(const TensorInfo& input0,
460 const TensorInfo& input1,
461 const TensorInfo& output,
462 Optional<std::string&> reasonIfUnsupported) const
463{
464 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSubtractionWorkloadValidate,
465 reasonIfUnsupported,
466 input0,
467 input1,
468 output);
469}
470
471bool IsNeonDirectConvolutionPreferred(const TensorInfo& weightInfo, const Convolution2dDescriptor& desc)
472{
473 // See arm_compute::NEDirectConvolutionLayer documentation for the supported cases,
474 // and complement with NEDirectConvolutionLayerKernel::configure() implementation.
475
476 // Only 1x1 is using direct convolution. Performance results and details are in:
477 // https://jira.arm.com/browse/IVGCVSW-1003
478 // Measurements were taken as of clframework: f105ab972135bcd21304883eff040d7e587099bc
479
480 const bool dataTypeSupported = (weightInfo.GetDataType() == armnn::DataType::Float32);
481
482 // Strides: 1|2|3
483 const bool strideSupported = (desc.m_StrideX == 1 || desc.m_StrideX == 2 || desc.m_StrideX == 3) &&
484 (desc.m_StrideY == 1 || desc.m_StrideY == 2 || desc.m_StrideY == 3);
485
486 auto paddingLargerThan = [](const Convolution2dDescriptor& conv2ddesc, unsigned int value)
487 {
488 return conv2ddesc.m_PadLeft > value || conv2ddesc.m_PadRight > value ||
489 conv2ddesc.m_PadTop > value || conv2ddesc.m_PadBottom > value;
490 };
491
492 // Supported sizes and padding.
493 const bool sizeAndPaddingSupported =
494 // Pad > 0 not supported for 1x1 weights.
495 (weightInfo.GetShape()[2] == 1 && weightInfo.GetShape()[3] == 1 && !paddingLargerThan(desc, 0u));
496
497 const bool preferDirectConvolution = dataTypeSupported &&
498 strideSupported &&
499 sizeAndPaddingSupported &&
500 // NEDirectConvolutionLayerKernel doesn't support NULL bias.
501 desc.m_BiasEnabled;
502 return preferDirectConvolution;
503}
504
505} // namespace armnn