blob: 0298c7c552d613f27b609927ea68adc96c0fb137 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
Francis Murtagh0d7963c2024-01-17 09:39:30 +00002// Copyright © 2017-2024 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5
6#include "NeonLayerSupport.hpp"
Sadik Armagan045f6be2020-09-10 13:37:32 +01007#include "NeonBackendModelContext.hpp"
telsoa014fcda012018-03-09 14:13:49 +00008
Derek Lambertic77874a2020-04-28 13:34:56 +01009#include <armnn/Exceptions.hpp>
telsoa014fcda012018-03-09 14:13:49 +000010#include <armnn/Tensor.hpp>
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010011#include <armnn/Types.hpp>
Matteo Martincighc601aa62019-10-29 15:03:22 +000012#include <armnn/BackendRegistry.hpp>
telsoa014fcda012018-03-09 14:13:49 +000013
Matteo Martincighc601aa62019-10-29 15:03:22 +000014#include <LayerSupportCommon.hpp>
Jan Eilers8eb25602020-03-09 12:13:48 +000015#include <armnn/utility/IgnoreUnused.hpp>
Francis Murtagh0d7963c2024-01-17 09:39:30 +000016#include <armnn/utility/NumericCast.hpp>
Sadik Armagan045f6be2020-09-10 13:37:32 +010017#include <armnn/utility/PolymorphicDowncast.hpp>
telsoa014fcda012018-03-09 14:13:49 +000018
Matteo Martincighd95e9062019-01-31 15:35:59 +000019#if defined(ARMCOMPUTENEON_ENABLED)
Narumol Prangnawarat15eb5832019-05-20 15:31:05 +010020#include <aclCommon/ArmComputeUtils.hpp>
Aron Virginas-Tar710f6642019-11-27 14:48:32 +000021#include <aclCommon/ArmComputeTensorUtils.hpp>
Teresa Charlin7db70892024-04-23 13:43:03 +010022#include <backendsCommon/WorkloadUtils.hpp>
Aron Virginas-Tar914e4db2019-09-09 13:36:45 +010023#include "workloads/NeonAbsWorkload.hpp"
Matthew Bentham955258d2018-12-10 10:48:52 +000024#include "workloads/NeonAdditionWorkload.hpp"
Nattapat Chaimanowongd4b70592018-10-12 11:21:49 +010025#include "workloads/NeonActivationWorkload.hpp"
James Conroyd47a0642019-09-17 14:22:06 +010026#include "workloads/NeonArgMinMaxWorkload.hpp"
Teresa Charlin0f86ecf2022-10-13 15:47:08 +010027#include "workloads/NeonBatchMatMulWorkload.hpp"
Matthew Benthamc48ac8c2018-12-12 16:15:59 +000028#include "workloads/NeonBatchNormalizationWorkload.hpp"
Mike Kelly56858022020-01-27 12:14:47 +000029#include "workloads/NeonBatchToSpaceNdWorkload.hpp"
Sadik Armagan48f011e2021-04-21 10:50:34 +010030#include "workloads/NeonCastWorkload.hpp"
Teresa Charline89dd692021-09-01 16:30:34 +010031#include "workloads/NeonChannelShuffleWorkload.hpp"
Teresa Charlincedd34f2020-03-30 11:17:30 +010032#include "workloads/NeonComparisonWorkload.hpp"
Teresa Charline89dd692021-09-01 16:30:34 +010033#include "workloads/NeonConcatWorkload.hpp"
Mike Kelly0886ac42020-04-27 09:55:40 +010034#include "workloads/NeonConstantWorkload.hpp"
Matthew Bentham34336f92023-04-27 12:13:50 +000035#include "workloads/NeonConvertFp16ToFp32Workload.hpp"
36#include "workloads/NeonConvertFp32ToFp16Workload.hpp"
Nattapat Chaimanowong974b65f2018-10-15 15:07:34 +010037#include "workloads/NeonConvolution2dWorkload.hpp"
Teresa Charlinec5f7d12021-10-22 17:15:00 +010038#include "workloads/NeonConvolution3dWorkload.hpp"
Aron Virginas-Tar2f00b742019-09-30 13:28:08 +010039#include "workloads/NeonDepthToSpaceWorkload.hpp"
Nattapat Chaimanowong77140882018-10-17 11:12:19 +010040#include "workloads/NeonDepthwiseConvolutionWorkload.hpp"
Narumol Prangnawarat01961a72019-05-30 16:47:12 +010041#include "workloads/NeonDequantizeWorkload.hpp"
Teresa Charlin9145e382023-08-17 18:44:58 +010042#include "workloads/NeonDivisionWorkload.hpp"
John Mcloughlin34c1c382023-05-17 15:08:36 +010043#include "workloads/NeonElementwiseBinaryWorkload.hpp"
Teresa Charline89dd692021-09-01 16:30:34 +010044#include "workloads/NeonExpWorkload.hpp"
Teresa Charlin9145e382023-08-17 18:44:58 +010045#include "workloads/NeonFullyConnectedWorkload.hpp"
46#include "workloads/NeonFusedWorkload.hpp"
47#include "workloads/NeonGatherWorkload.hpp"
48#include "workloads/NeonGatherNdWorkload.hpp"
Sadik Armagan0d4863d2019-10-09 14:26:32 +010049#include "workloads/NeonInstanceNormalizationWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010050#include "workloads/NeonL2NormalizationFloatWorkload.hpp"
Teresa Charlin50de4fa2021-05-31 18:47:33 +010051#include "workloads/NeonLogWorkload.hpp"
Keith Davis69e653f2020-07-02 11:49:26 +010052#include "workloads/NeonLogSoftmaxWorkload.hpp"
James Conroy177df1e2020-11-13 10:18:51 +000053#include "workloads/NeonLogicalAndWorkload.hpp"
54#include "workloads/NeonLogicalNotWorkload.hpp"
55#include "workloads/NeonLogicalOrWorkload.hpp"
Jan Eilersad5293a2019-07-08 09:57:55 +010056#include "workloads/NeonLstmFloatWorkload.hpp"
Nattapat Chaimanowong4e6597a2018-12-20 14:14:06 +000057#include "workloads/NeonMaximumWorkload.hpp"
Matthew Benthamfd899962018-12-31 15:49:42 +000058#include "workloads/NeonMeanWorkload.hpp"
Conor Kennedy54b21692019-01-09 07:57:38 +000059#include "workloads/NeonMinimumWorkload.hpp"
Conor Kennedyb99480b2019-03-08 08:24:41 +000060#include "workloads/NeonMultiplicationWorkload.hpp"
Sadik Armaganac472102020-03-24 09:54:36 +000061#include "workloads/NeonNegWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010062#include "workloads/NeonNormalizationFloatWorkload.hpp"
Éanna Ó Catháin12055742019-01-25 10:01:40 +000063#include "workloads/NeonPadWorkload.hpp"
David Beck0dbe0ee2018-09-24 15:59:27 +010064#include "workloads/NeonPermuteWorkload.hpp"
Nattapat Chaimanowong5d2e7002018-10-12 16:03:56 +010065#include "workloads/NeonPooling2dWorkload.hpp"
Ryan OShea19e79422022-05-04 00:38:03 +010066#include "workloads/NeonPooling3dWorkload.hpp"
Nikhil Raj9b461482019-07-03 15:58:31 +010067#include "workloads/NeonPreluWorkload.hpp"
James Conroycc340932020-05-12 18:08:52 +010068#include "workloads/NeonQLstmWorkload.hpp"
Sadik Armaganfabc2892019-05-31 09:05:11 +010069#include "workloads/NeonQuantizeWorkload.hpp"
Francis Murtagh4fc3c482019-08-02 13:20:54 +010070#include "workloads/NeonQuantizedLstmWorkload.hpp"
Sadik Armagana2747482021-02-09 10:28:54 +000071#include "workloads/NeonReduceWorkload.hpp"
Kevin Maya023c402019-12-12 17:28:05 +000072#include "workloads/NeonReshapeWorkload.hpp"
Ellen Norris-Thompson37e68682019-07-15 14:23:30 +010073#include "workloads/NeonResizeWorkload.hpp"
Tianle Cheng21a9f332023-11-09 13:56:53 +000074#include "workloads/NeonReverseV2Workload.hpp"
Aron Virginas-Tar0dd3b432019-09-10 13:55:09 +010075#include "workloads/NeonRsqrtWorkload.hpp"
Teresa Charlin50de4fa2021-05-31 18:47:33 +010076#include "workloads/NeonSinWorkload.hpp"
josh minor036f02d2019-11-15 14:53:22 -060077#include "workloads/NeonSliceWorkload.hpp"
Sadik Armaganbe88a572020-04-30 11:39:37 +010078#include "workloads/NeonSoftmaxWorkload.hpp"
Mike Kelly0be3a882020-01-24 11:27:50 +000079#include "workloads/NeonSpaceToBatchNdWorkload.hpp"
Ellen Norris-Thompson29794572019-06-26 16:40:36 +010080#include "workloads/NeonSpaceToDepthWorkload.hpp"
Narumol Prangnawarat15eb5832019-05-20 15:31:05 +010081#include "workloads/NeonSplitterWorkload.hpp"
Teresa Charlin06145cc2022-05-05 15:31:30 +010082#include "workloads/NeonSqrtWorkload.hpp"
Matthew Jackson87f65ea2019-08-01 10:01:34 +010083#include "workloads/NeonStackWorkload.hpp"
FinnWilliamsArm1fa19192019-08-02 17:26:31 +010084#include "workloads/NeonStridedSliceWorkload.hpp"
Conor Kennedyb99480b2019-03-08 08:24:41 +000085#include "workloads/NeonSubtractionWorkload.hpp"
David Monahan36e6eae2023-07-26 18:37:45 +010086#include "workloads/NeonTileWorkload.hpp"
Sadik Armagan581742d2019-08-12 14:11:37 +010087#include "workloads/NeonTransposeConvolution2dWorkload.hpp"
Mike Kellyc9ea45a2020-02-28 18:11:58 +000088#include "workloads/NeonTransposeWorkload.hpp"
Cathal Corbettfd5bec42022-03-03 15:13:23 +000089#include "workloads/NeonUnidirectionalSequenceLstmFloatWorkload.hpp"
Mike Kelly12994962022-04-21 11:57:09 +010090#include "workloads/NeonUnidirectionalSequenceLstmWorkload.hpp"
telsoa014fcda012018-03-09 14:13:49 +000091#endif
92
telsoa014fcda012018-03-09 14:13:49 +000093namespace armnn
94{
telsoa014fcda012018-03-09 14:13:49 +000095
Aron Virginas-Tarfc824312018-10-15 15:00:13 +010096namespace
arovir017ff76c52018-10-09 09:40:58 +010097{
telsoa014fcda012018-03-09 14:13:49 +000098
Cathal Corbett80f71a82022-12-20 18:25:40 +000099const TensorInfo OverrideDataType(const TensorInfo& info, Optional<DataType> type)
100{
101 if (!type)
102 {
103 return info;
104 }
105 return TensorInfo(info.GetShape(),
106 type.value(),
107 info.GetQuantizationScale(),
108 info.GetQuantizationOffset(),
109 info.IsConstant());
110}
111
Derek Lamberti901ea112019-12-10 22:07:09 +0000112template< typename ... Args>
113bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported, Args... args)
telsoa014fcda012018-03-09 14:13:49 +0000114{
Jan Eilers8eb25602020-03-09 12:13:48 +0000115 IgnoreUnused(reasonIfUnsupported, (args)...);
Matteo Martincighd95e9062019-01-31 15:35:59 +0000116#if defined(ARMCOMPUTENEON_ENABLED)
telsoa014fcda012018-03-09 14:13:49 +0000117 return true;
118#else
Derek Lamberti0790dce2019-04-15 18:37:35 +0100119 SetValueChecked(reasonIfUnsupported, "The armnn library has been built without NEON support");
telsoa014fcda012018-03-09 14:13:49 +0000120 return false;
121#endif
122}
123
telsoa01c577f2c2018-08-31 09:22:23 +0100124template<typename FloatFunc, typename Uint8Func, typename ... Params>
arovir01085f0a42018-10-08 14:48:19 +0100125bool IsSupportedForDataTypeNeon(Optional<std::string&> reasonIfUnsupported,
telsoa014fcda012018-03-09 14:13:49 +0000126 DataType dataType,
telsoa01c577f2c2018-08-31 09:22:23 +0100127 FloatFunc floatFuncPtr,
telsoa014fcda012018-03-09 14:13:49 +0000128 Uint8Func uint8FuncPtr,
129 Params&&... params)
130{
131 return IsNeonBackendSupported(reasonIfUnsupported) &&
132 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
Teresa Charlin9145e382023-08-17 18:44:58 +0100133 dataType,
134 floatFuncPtr,
135 floatFuncPtr,
136 uint8FuncPtr,
137 &FalseFunc<>,
138 &FalseFunc<>,
139 std::forward<Params>(params)...);
telsoa014fcda012018-03-09 14:13:49 +0000140}
141
Matteo Martincighd95e9062019-01-31 15:35:59 +0000142#if defined(ARMCOMPUTENEON_ENABLED)
telsoa014fcda012018-03-09 14:13:49 +0000143template<class FuncType, class... Args>
arovir01085f0a42018-10-08 14:48:19 +0100144inline bool IsWorkloadSupported(FuncType& func, Optional<std::string&> reasonIfUnsupported, Args&&... args)
telsoa014fcda012018-03-09 14:13:49 +0000145{
146 arm_compute::Status aclStatus = func(std::forward<Args>(args)...);
147 const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
148 if (!supported && reasonIfUnsupported)
149 {
arovir01085f0a42018-10-08 14:48:19 +0100150 reasonIfUnsupported.value() = aclStatus.error_description();
telsoa014fcda012018-03-09 14:13:49 +0000151 }
152 return supported;
153}
154
155#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
156 return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__);
157#else
158#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
Derek Lamberti901ea112019-12-10 22:07:09 +0000159 return IsNeonBackendSupported(reasonIfUnsupported, __VA_ARGS__);
telsoa014fcda012018-03-09 14:13:49 +0000160#endif
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100161} // anonymous namespace
162
Sadik Armagan045f6be2020-09-10 13:37:32 +0100163NeonLayerSupport::NeonLayerSupport(const IBackendInternal::IBackendSpecificModelContextPtr& modelContextPtr)
164 : m_ModelContextPtr(modelContextPtr)
165{
166}
167
168NeonLayerSupport::NeonLayerSupport()
169 : m_ModelContextPtr(nullptr)
170{
171}
172
Cathal Corbett80f71a82022-12-20 18:25:40 +0000173bool IsLayerTypeSupported(const LayerType& type,
174 const std::vector<TensorInfo>& infos,
175 const BaseDescriptor& descriptor,
176 const Optional<LstmInputParamsInfo>& lstmParamsInfo,
177 const Optional<QuantizedLstmInputParamsInfo>& quantizedLstmParamsInfo,
178 Optional<std::string&> reasonIfUnsupported,
179 const NeonLayerSupport& support)
Cathal Corbett34b429c2021-12-24 12:24:40 +0000180{
181 switch (type)
182 {
183 case LayerType::Activation:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000184 return support.IsActivationSupported(infos[0],
185 infos[1],
186 *(PolymorphicDowncast<const ActivationDescriptor*>(&descriptor)),
187 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000188 case LayerType::Addition:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000189 return support.IsAdditionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000190 case LayerType::ArgMinMax:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000191 return support.IsArgMinMaxSupported(infos[0],
192 infos[1],
193 *(PolymorphicDowncast<const ArgMinMaxDescriptor*>(&descriptor)),
194 reasonIfUnsupported);
Teresa Charlin0f86ecf2022-10-13 15:47:08 +0100195 case LayerType::BatchMatMul:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000196 return support.IsBatchMatMulSupported(infos[0],
197 infos[1],
198 infos[2],
199 *(PolymorphicDowncast<const BatchMatMulDescriptor*>(&descriptor)),
200 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000201 case LayerType::BatchNormalization:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000202 return support.IsBatchNormalizationSupported(infos[0],
203 infos[1],
204 infos[2],
205 infos[3],
206 infos[4],
207 infos[5],
208 *(PolymorphicDowncast<const
209 BatchNormalizationDescriptor*>(&descriptor)),
210 reasonIfUnsupported);
211 case LayerType::BatchToSpaceNd:
212 return support.IsBatchToSpaceNdSupported(infos[0],
213 infos[1],
214 *(PolymorphicDowncast<const
215 BatchToSpaceNdDescriptor*>(&descriptor)),
216 reasonIfUnsupported);
217 case LayerType::Cast:
218 return support.IsCastSupported(infos[0], infos[1], reasonIfUnsupported);
219 case LayerType::ChannelShuffle:
220 return support.IsChannelShuffleSupported(infos[0],
221 infos[1],
222 *(PolymorphicDowncast<const
223 ChannelShuffleDescriptor*>(&descriptor)),
224 reasonIfUnsupported);
225 case LayerType::Comparison:
226 return support.IsComparisonSupported(infos[0],
Cathal Corbett34b429c2021-12-24 12:24:40 +0000227 infos[1],
228 infos[2],
Cathal Corbett80f71a82022-12-20 18:25:40 +0000229 *(PolymorphicDowncast<const ComparisonDescriptor*>(&descriptor)),
Cathal Corbett34b429c2021-12-24 12:24:40 +0000230 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000231 case LayerType::Concat:
232 {
233 std::vector<const TensorInfo*> inputInfos;
234 for (uint32_t i = 0; i < (infos.size() - 1); i++)
235 {
236 inputInfos.push_back(&infos[i]);
237 }
Cathal Corbett80f71a82022-12-20 18:25:40 +0000238 return support.IsConcatSupported(inputInfos,
239 infos[infos.size() - 1],
240 *(PolymorphicDowncast<const OriginsDescriptor*>(&descriptor)),
241 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000242 }
243 case LayerType::Constant:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000244 return support.IsConstantSupported(infos[0], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000245 case LayerType::ConvertFp16ToFp32:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000246 return support.IsConvertFp16ToFp32Supported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000247 case LayerType::ConvertFp32ToFp16:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000248 return support.IsConvertFp32ToFp16Supported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000249 case LayerType::Convolution2d:
250 {
251 if (infos.size() != 4)
252 {
253 throw InvalidArgumentException("Invalid number of TransposeConvolution2d TensorInfos. "
254 "TensorInfos should be of format: {input, output, weights, biases}.");
255 }
256
257 auto desc = *(PolymorphicDowncast<const Convolution2dDescriptor*>(&descriptor));
258 if (infos[3] == TensorInfo())
259 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000260 return support.IsConvolution2dSupported(infos[0],
261 infos[1],
262 desc,
263 infos[2],
264 EmptyOptional(),
265 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000266 }
267 else
268 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000269 return support.IsConvolution2dSupported(infos[0],
270 infos[1],
271 desc,
272 infos[2],
273 infos[3],
274 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000275 }
276 }
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000277 case LayerType::Convolution3d:
278 {
279 if (infos.size() != 4)
280 {
281 throw InvalidArgumentException("Invalid number of Convolution3d TensorInfos. "
282 "TensorInfos should be of format: {input, output, weights, biases}.");
283 }
284
285 auto desc = *(PolymorphicDowncast<const Convolution3dDescriptor*>(&descriptor));
286 if (infos[3] == TensorInfo())
287 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000288 return support.IsConvolution3dSupported(infos[0],
289 infos[1],
290 desc,
291 infos[2],
292 EmptyOptional(),
293 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000294 }
295 else
296 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000297 return support.IsConvolution3dSupported(infos[0],
298 infos[1],
299 desc,
300 infos[2],
301 infos[3],
302 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000303 }
304 }
Cathal Corbett34b429c2021-12-24 12:24:40 +0000305 case LayerType::DepthToSpace:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000306 return support.IsDepthToSpaceSupported(infos[0],
307 infos[1],
308 *(PolymorphicDowncast<const DepthToSpaceDescriptor*>(&descriptor)),
309 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000310 case LayerType::DepthwiseConvolution2d:
311 {
312 if (infos.size() != 4)
313 {
314 throw InvalidArgumentException("Invalid number of DepthwiseConvolution2d TensorInfos. "
315 "TensorInfos should be of format: {input, output, weights, biases}.");
316 }
317
318 auto desc = *(PolymorphicDowncast<const DepthwiseConvolution2dDescriptor*>(&descriptor));
319 if (infos[3] == TensorInfo())
320 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000321 return support.IsDepthwiseConvolutionSupported(infos[0],
322 infos[1],
323 desc,
324 infos[2],
325 EmptyOptional(),
326 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000327 }
328 else
329 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000330 return support.IsDepthwiseConvolutionSupported(infos[0],
331 infos[1],
332 desc,
333 infos[2],
334 infos[3],
335 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000336 }
337 }
338 case LayerType::Dequantize:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000339 return support.IsDequantizeSupported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000340 case LayerType::DetectionPostProcess:
341 {
342 auto desc = *(PolymorphicDowncast<const DetectionPostProcessDescriptor*>(&descriptor));
Cathal Corbett80f71a82022-12-20 18:25:40 +0000343 return support.IsDetectionPostProcessSupported(infos[0],
344 infos[1],
345 infos[2],
346 infos[3],
347 infos[4],
348 infos[5],
349 infos[6],
350 desc,
351 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000352 }
Cathal Corbett34b429c2021-12-24 12:24:40 +0000353 case LayerType::Division:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000354 return support.IsDivisionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Mike Kelly3ec30772023-03-08 13:47:17 +0000355 case LayerType::ElementwiseBinary:
356 {
357 auto desc = *(PolymorphicDowncast<const ElementwiseBinaryDescriptor *>(&descriptor));
358
359 switch (desc.m_Operation)
360 {
361 case BinaryOperation::Add:
362 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonAdditionWorkloadValidate,
363 reasonIfUnsupported,
364 infos[0],
365 infos[1],
366 infos[2],
367 nullptr);
368 case BinaryOperation::Div:
369 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDivisionWorkloadValidate,
370 reasonIfUnsupported,
371 infos[0],
372 infos[1],
373 infos[2],
374 nullptr);
375 case BinaryOperation::Maximum:
376 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMaximumWorkloadValidate,
377 reasonIfUnsupported,
378 infos[0],
379 infos[1],
380 infos[2]);
381 case BinaryOperation::Minimum:
382 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMinimumWorkloadValidate,
383 reasonIfUnsupported,
384 infos[0],
385 infos[1],
386 infos[2]);
387 case BinaryOperation::Mul:
388 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMultiplicationWorkloadValidate,
389 reasonIfUnsupported,
390 infos[0],
391 infos[1],
392 infos[2],
393 nullptr);
John Mcloughlin34c1c382023-05-17 15:08:36 +0100394 case BinaryOperation::Power:
395 case BinaryOperation::SqDiff:
396 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonElementwiseBinaryWorkloadValidate,
397 reasonIfUnsupported,
398 infos[0],
399 infos[1],
400 infos[2],
401 desc,
402 nullptr);
Mike Kelly3ec30772023-03-08 13:47:17 +0000403 case BinaryOperation::Sub:
404 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSubtractionWorkloadValidate,
405 reasonIfUnsupported,
406 infos[0],
407 infos[1],
408 infos[2],
409 nullptr);
410 default:
411 return false;
412 }
413 }
Cathal Corbett34b429c2021-12-24 12:24:40 +0000414 case LayerType::ElementwiseUnary:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000415 return support.IsElementwiseUnarySupported(infos[0],
416 infos[1],
417 *(PolymorphicDowncast<const
418 ElementwiseUnaryDescriptor*>(&descriptor)),
419 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000420 case LayerType::Fill:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000421 return support.IsFillSupported(infos[0],
422 infos[1],
423 *(PolymorphicDowncast<const FillDescriptor*>(&descriptor)),
424 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000425 case LayerType::Floor:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000426 return support.IsFloorSupported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000427 case LayerType::FullyConnected:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000428 return support.IsFullyConnectedSupported(infos[0],
429 infos[1],
430 infos[2],
431 infos[3],
432 *(PolymorphicDowncast<const
433 FullyConnectedDescriptor*>(&descriptor)),
434 reasonIfUnsupported);
Teresa Charlin9145e382023-08-17 18:44:58 +0100435 case LayerType::Fused:
436 {
437 auto fusedDescriptor = *(PolymorphicDowncast<const FusedDescriptor*>(&descriptor));
438 if (fusedDescriptor.m_NumInputSlots + fusedDescriptor.m_NumOutputSlots != infos.size())
439 {
440 throw InvalidArgumentException("Invalid number of FusedLayer TensorInfos.");
441 }
442
Francis Murtagh0d7963c2024-01-17 09:39:30 +0000443 auto it = infos.begin() + numeric_cast<TensorInfo::DifferenceType>(fusedDescriptor.m_NumInputSlots);
444 std::vector<TensorInfo> inputInfos(infos.begin(), it);
445 std::vector<TensorInfo> outputInfos(it, infos.end());
Teresa Charlin9145e382023-08-17 18:44:58 +0100446
447 return support.IsFusedSupported({inputInfos.begin(), inputInfos.end()},
448 {outputInfos.begin(), outputInfos.end()},
449 fusedDescriptor,
450 reasonIfUnsupported);
451 }
Cathal Corbett80f71a82022-12-20 18:25:40 +0000452 case LayerType::Gather:
453 return support.IsGatherSupported(infos[0],
Cathal Corbett34b429c2021-12-24 12:24:40 +0000454 infos[1],
455 infos[2],
Cathal Corbett80f71a82022-12-20 18:25:40 +0000456 *(PolymorphicDowncast<const GatherDescriptor*>(&descriptor)),
Cathal Corbett34b429c2021-12-24 12:24:40 +0000457 reasonIfUnsupported);
Teresa Charlinbd22c7d2022-04-26 18:14:12 +0100458 case LayerType::GatherNd:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000459 return support.IsGatherNdSupported(infos[0],
460 infos[1],
461 infos[2],
462 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000463 case LayerType::Input:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000464 return support.IsInputSupported(infos[0], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000465 case LayerType::InstanceNormalization:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000466 return support.IsInstanceNormalizationSupported(infos[0],
467 infos[1],
468 *(PolymorphicDowncast<const
469 InstanceNormalizationDescriptor*>(&descriptor)),
470 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000471 case LayerType::L2Normalization:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000472 return support.IsL2NormalizationSupported(infos[0],
473 infos[1],
474 *(PolymorphicDowncast<const
475 L2NormalizationDescriptor*>(&descriptor)),
476 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000477 case LayerType::LogicalBinary:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000478 return support.IsLogicalBinarySupported(infos[0],
479 infos[1],
480 infos[2],
481 *(PolymorphicDowncast<const
482 LogicalBinaryDescriptor*>(&descriptor)),
483 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000484 case LayerType::LogSoftmax:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000485 return support.IsLogSoftmaxSupported(infos[0],
486 infos[1],
487 *(PolymorphicDowncast<const LogSoftmaxDescriptor*>(&descriptor)),
488 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000489 case LayerType::Lstm:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000490 return support.IsLstmSupported(infos[0],
491 infos[1],
492 infos[2],
493 infos[3],
494 infos[4],
495 infos[5],
496 infos[6],
497 *(PolymorphicDowncast<const LstmDescriptor*>(&descriptor)),
498 lstmParamsInfo.value(),
499 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000500 case LayerType::Map:
501 return true;
Cathal Corbett34b429c2021-12-24 12:24:40 +0000502 case LayerType::Maximum:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000503 return support.IsMaximumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000504 case LayerType::Mean:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000505 return support.IsMeanSupported(infos[0],
506 infos[1],
507 *(PolymorphicDowncast<const MeanDescriptor*>(&descriptor)),
508 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000509 case LayerType::MemCopy:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000510 return support.IsMemCopySupported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000511 case LayerType::MemImport:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000512 return support.IsMemImportSupported(infos[0], infos[1], reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000513 case LayerType::Merge:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000514 return support.IsMergeSupported(infos[0],
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000515 infos[1],
516 infos[2],
517 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000518 case LayerType::Minimum:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000519 return support.IsMinimumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000520 case LayerType::Multiplication:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000521 return support.IsMultiplicationSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000522 case LayerType::Normalization:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000523 return support.IsNormalizationSupported(infos[0],
524 infos[1],
525 *(PolymorphicDowncast<const
526 NormalizationDescriptor*>(&descriptor)),
527 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000528 case LayerType::Output:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000529 return support.IsOutputSupported(infos[0], reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000530 case LayerType::Pad:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000531 return support.IsPadSupported(infos[0],
532 infos[1],
533 *(PolymorphicDowncast<const PadDescriptor*>(&descriptor)),
534 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000535 case LayerType::Permute:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000536 return support.IsPermuteSupported(infos[0],
537 infos[1],
538 *(PolymorphicDowncast<const PermuteDescriptor*>(&descriptor)),
539 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000540 case LayerType::Pooling2d:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000541 return support.IsPooling2dSupported(infos[0],
542 infos[1],
543 *(PolymorphicDowncast<const Pooling2dDescriptor*>(&descriptor)),
544 reasonIfUnsupported);
Ryan OShea19e79422022-05-04 00:38:03 +0100545 case LayerType::Pooling3d:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000546 return support.IsPooling3dSupported(infos[0],
547 infos[1],
548 *(PolymorphicDowncast<const Pooling3dDescriptor*>(&descriptor)),
549 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000550 case LayerType::Prelu:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000551 return support.IsPreluSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000552 case LayerType::QLstm:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000553 return support.IsQLstmSupported(infos[0],
Cathal Corbett34b429c2021-12-24 12:24:40 +0000554 infos[1],
555 infos[2],
556 infos[3],
557 infos[4],
Cathal Corbett80f71a82022-12-20 18:25:40 +0000558 infos[5],
559 *(PolymorphicDowncast<const QLstmDescriptor*>(&descriptor)),
560 lstmParamsInfo.value(),
Cathal Corbett34b429c2021-12-24 12:24:40 +0000561 reasonIfUnsupported);
Cathal Corbett80f71a82022-12-20 18:25:40 +0000562 case LayerType::Quantize:
563 return support.IsQuantizeSupported(infos[0], infos[1], reasonIfUnsupported);
564 case LayerType::QuantizedLstm:
565 return support.IsQuantizedLstmSupported(infos[0],
566 infos[1],
567 infos[2],
568 infos[3],
569 infos[4],
570 quantizedLstmParamsInfo.value(),
571 reasonIfUnsupported);
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000572 case LayerType::Rank:
573 return true;
Cathal Corbett34b429c2021-12-24 12:24:40 +0000574 case LayerType::Reshape:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000575 return support.IsReshapeSupported(infos[0],
576 infos[1],
577 *(PolymorphicDowncast<const ReshapeDescriptor*>(&descriptor)),
578 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000579 case LayerType::Resize:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000580 return support.IsResizeSupported(infos[0],
Cathal Corbett34b429c2021-12-24 12:24:40 +0000581 infos[1],
Cathal Corbett80f71a82022-12-20 18:25:40 +0000582 *(PolymorphicDowncast<const ResizeDescriptor*>(&descriptor)),
Cathal Corbett34b429c2021-12-24 12:24:40 +0000583 reasonIfUnsupported);
Cathal Corbett80f71a82022-12-20 18:25:40 +0000584 case LayerType::Reduce:
585 return support.IsReduceSupported(infos[0],
586 infos[1],
587 *(PolymorphicDowncast<const ReduceDescriptor*>(&descriptor)),
588 reasonIfUnsupported);
Tianle Cheng21a9f332023-11-09 13:56:53 +0000589 case LayerType::ReverseV2:
590 return support.IsReverseV2Supported(infos[0],
591 infos[1],
592 infos[2],
593 reasonIfUnsupported);
Cathal Corbett80f71a82022-12-20 18:25:40 +0000594 case LayerType::Shape:
595 return support.IsShapeSupported(infos[0],
596 infos[1],
597 reasonIfUnsupported);
598 case LayerType::Slice:
599 return support.IsSliceSupported(infos[0],
600 infos[1],
601 *(PolymorphicDowncast<const SliceDescriptor*>(&descriptor)),
602 reasonIfUnsupported);
603 case LayerType::Softmax:
604 return support.IsSoftmaxSupported(infos[0],
605 infos[1],
606 *(PolymorphicDowncast<const SoftmaxDescriptor*>(&descriptor)),
607 reasonIfUnsupported);
608 case LayerType::SpaceToBatchNd:
609 return support.IsSpaceToBatchNdSupported(infos[0],
610 infos[1],
611 *(PolymorphicDowncast<const
612 SpaceToBatchNdDescriptor*>(&descriptor)),
613 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000614 case LayerType::SpaceToDepth:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000615 return support.IsSpaceToDepthSupported(infos[0],
616 infos[1],
617 *(PolymorphicDowncast<const SpaceToDepthDescriptor*>(&descriptor)),
618 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000619 case LayerType::Splitter:
620 {
621 std::vector<TensorInfo> outputInfos;
622 for (uint32_t i = 1; i < infos.size(); i++)
623 {
624 outputInfos.push_back(infos[i]);
625 }
Cathal Corbett80f71a82022-12-20 18:25:40 +0000626 return support.IsSplitterSupported(infos[0],
627 {outputInfos.begin(), outputInfos.end()},
628 *(PolymorphicDowncast<const ViewsDescriptor*>(&descriptor)),
629 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000630 }
631 case LayerType::Stack:
632 {
633 std::vector<const TensorInfo*> inputInfos;
634 for (uint32_t i = 0; i < infos.size() - 1; i++)
635 {
636 inputInfos.push_back(&infos[i]);
637 }
Cathal Corbett80f71a82022-12-20 18:25:40 +0000638 return support.IsStackSupported(inputInfos,
639 infos[infos.size() - 1],
640 *(PolymorphicDowncast<const StackDescriptor*>(&descriptor)),
641 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000642 }
643 case LayerType::StridedSlice:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000644 return support.IsStridedSliceSupported(infos[0],
645 infos[1],
646 *(PolymorphicDowncast<const StridedSliceDescriptor*>(&descriptor)),
647 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000648 case LayerType::Subtraction:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000649 return support.IsSubtractionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
David Monahan36e6eae2023-07-26 18:37:45 +0100650 case LayerType::Tile:
651 return support.IsTileSupported(infos[0],
652 infos[1],
653 *(PolymorphicDowncast<const TileDescriptor*>(&descriptor)),
654 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000655 case LayerType::Transpose:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000656 return support.IsTransposeSupported(infos[0],
657 infos[1],
658 *(PolymorphicDowncast<const TransposeDescriptor*>(&descriptor)),
659 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000660 case LayerType::TransposeConvolution2d:
661 {
662 if (infos.size() != 4)
663 {
664 throw InvalidArgumentException("Invalid number of TransposeConvolution2d TensorInfos. "
665 "TensorInfos should be of format: {input, output, weights, biases}.");
666 }
667
668 auto desc = *(PolymorphicDowncast<const TransposeConvolution2dDescriptor*>(&descriptor));
669 if (infos[3] == TensorInfo())
670 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000671 return support.IsTransposeConvolution2dSupported(infos[0],
672 infos[1],
673 desc,
674 infos[2],
675 EmptyOptional(),
676 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000677 }
678 else
679 {
Cathal Corbett80f71a82022-12-20 18:25:40 +0000680 return support.IsTransposeConvolution2dSupported(infos[0],
681 infos[1],
682 desc,
683 infos[2],
684 infos[3],
685 reasonIfUnsupported);
Cathal Corbett34b429c2021-12-24 12:24:40 +0000686 }
687 }
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000688 case LayerType::UnidirectionalSequenceLstm:
Cathal Corbett80f71a82022-12-20 18:25:40 +0000689 {
690 auto desc = *(PolymorphicDowncast<const UnidirectionalSequenceLstmDescriptor*>(&descriptor));
691 return support.IsUnidirectionalSequenceLstmSupported(infos[0],
692 infos[1],
693 infos[2],
694 infos[3],
695 infos[4],
696 infos[5],
697 desc,
698 lstmParamsInfo.value(),
699 reasonIfUnsupported);
700 }
Cathal Corbett34b429c2021-12-24 12:24:40 +0000701 case LayerType::Unmap:
702 return true;
Cathal Corbett34b429c2021-12-24 12:24:40 +0000703 default:
704 // layers not supported in neon by default:
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000705 // debug, fakequantization, precompiled,
Ryan OShea19e79422022-05-04 00:38:03 +0100706 // standin, switch
Cathal Corbett34b429c2021-12-24 12:24:40 +0000707 return false;
708 }
709}
710
Cathal Corbett80f71a82022-12-20 18:25:40 +0000711bool NeonLayerSupport::IsLayerSupported(const LayerType& type,
712 const std::vector<TensorInfo>& infos,
713 const BaseDescriptor& descriptor,
714 const Optional<LstmInputParamsInfo>& lstmParamsInfo,
715 const Optional<QuantizedLstmInputParamsInfo>& quantizedLstmParamsInfo,
716 Optional<std::string&> reasonIfUnsupported) const
717{
718 bool isSupported = IsLayerTypeSupported(type,
719 infos,
720 descriptor,
721 lstmParamsInfo,
722 quantizedLstmParamsInfo,
723 reasonIfUnsupported,
724 *this);
725
726 // For android-nn-driver and support library, to run FP16 operations on CpuAcc we need at least v8.2
727 // architecture. If the available architecture is older than v8.2, we can check if the operator is
728 // supported by changing operator inputs & outputs to be FP32.
729 // This does not change the operator datatype in the above parsers to be FP32. We are simply reporting
730 // to the parsers if the operator can supported in ArmNN. We will then re-enter ArmNN (Network.cpp)
731 // where we will recheck IsLayerSupported() on the FP16 datatype, update the operator to be FP32,
732 // and, insert convert layers around the FP32 operator.
733 if (reasonIfUnsupported.has_value())
734 {
735 std::string checkStr = "This CPU architecture does not support F16 data type, you need v8.2 or above";
736 if (!isSupported
737 && reasonIfUnsupported.value().find(checkStr) != std::string::npos)
738 {
739 std::vector<TensorInfo> newInfos;
740 for (auto info: infos)
741 {
742 newInfos.emplace_back(OverrideDataType(info, DataType::Float32));
743 }
744
745 std::string tmpString;
746 return IsLayerTypeSupported(type,
747 newInfos,
748 descriptor,
749 lstmParamsInfo,
750 quantizedLstmParamsInfo,
751 tmpString,
752 *this);
753 }
754 }
755
756 return isSupported;
757}
758
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100759bool NeonLayerSupport::IsActivationSupported(const TensorInfo& input,
760 const TensorInfo& output,
761 const ActivationDescriptor& descriptor,
762 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000763{
Jan Eilers8eb25602020-03-09 12:13:48 +0000764 IgnoreUnused(descriptor);
telsoa01c577f2c2018-08-31 09:22:23 +0100765 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonActivationWorkloadValidate,
766 reasonIfUnsupported,
767 input,
768 output,
769 descriptor);
telsoa014fcda012018-03-09 14:13:49 +0000770}
771
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100772bool NeonLayerSupport::IsAdditionSupported(const TensorInfo& input0,
773 const TensorInfo& input1,
774 const TensorInfo& output,
775 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000776{
telsoa01c577f2c2018-08-31 09:22:23 +0100777 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonAdditionWorkloadValidate,
778 reasonIfUnsupported,
779 input0,
780 input1,
Mike Kelly07810fc2020-11-12 10:58:48 +0000781 output,
782 nullptr);
telsoa014fcda012018-03-09 14:13:49 +0000783}
784
James Conroyd47a0642019-09-17 14:22:06 +0100785bool NeonLayerSupport::IsArgMinMaxSupported(const TensorInfo& input,
786 const TensorInfo& output,
787 const ArgMinMaxDescriptor& descriptor,
788 Optional<std::string&> reasonIfUnsupported) const
789{
790 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonArgMinMaxWorkloadValidate,
791 reasonIfUnsupported,
792 input,
793 output,
794 descriptor);
795}
796
Teresa Charlin0f86ecf2022-10-13 15:47:08 +0100797bool NeonLayerSupport::IsBatchMatMulSupported(const TensorInfo& inputX,
798 const TensorInfo& inputY,
799 const TensorInfo& output,
800 const BatchMatMulDescriptor& descriptor,
801 Optional<std::string&> reasonIfUnsupported) const
802{
Teresa Charlin1fe6c812022-11-01 15:59:50 +0000803 bool isFastMathEnabled = false;
804#if defined(ARMCOMPUTENEON_ENABLED)
805 if (m_ModelContextPtr)
806 {
807 if (m_ModelContextPtr.get() != nullptr)
808 {
809 auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
810 if (modelOptions)
811 {
812 isFastMathEnabled = modelOptions->IsFastMathEnabled();
813 }
814 }
815 }
816#endif
Teresa Charlin0f86ecf2022-10-13 15:47:08 +0100817 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonBatchMatMulValidate,
818 reasonIfUnsupported,
819 inputX,
820 inputY,
821 output,
Teresa Charlin1fe6c812022-11-01 15:59:50 +0000822 descriptor,
823 isFastMathEnabled,
824 nullptr);
Teresa Charlin0f86ecf2022-10-13 15:47:08 +0100825}
826
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100827bool NeonLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input,
828 const TensorInfo& output,
829 const TensorInfo& mean,
830 const TensorInfo& var,
831 const TensorInfo& beta,
832 const TensorInfo& gamma,
833 const BatchNormalizationDescriptor& descriptor,
834 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000835{
telsoa01c577f2c2018-08-31 09:22:23 +0100836 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonBatchNormalizationValidate,
837 reasonIfUnsupported,
838 input,
839 output,
840 mean,
841 var,
842 beta,
843 gamma,
Mike Kelly07810fc2020-11-12 10:58:48 +0000844 descriptor,
845 nullptr);
telsoa014fcda012018-03-09 14:13:49 +0000846}
847
Mike Kelly56858022020-01-27 12:14:47 +0000848bool NeonLayerSupport::IsBatchToSpaceNdSupported(const TensorInfo& input,
849 const TensorInfo& output,
850 const BatchToSpaceNdDescriptor& descriptor,
851 Optional<std::string&> reasonIfUnsupported) const
852{
853 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonBatchToSpaceNdWorkloadValidate,
854 reasonIfUnsupported,
855 input,
856 output,
857 descriptor);
858}
859
Sadik Armagan48f011e2021-04-21 10:50:34 +0100860bool NeonLayerSupport::IsCastSupported(const TensorInfo& input,
861 const TensorInfo& output,
862 Optional<std::string&> reasonIfUnsupported) const
863{
864 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonCastValidate,
865 reasonIfUnsupported,
866 input,
867 output);
868}
869
Teresa Charline89dd692021-09-01 16:30:34 +0100870bool NeonLayerSupport::IsChannelShuffleSupported(const TensorInfo& input,
871 const TensorInfo& output,
872 const ChannelShuffleDescriptor& descriptor,
873 Optional<std::string&> reasonIfUnsupported) const
874{
875 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonChannelShuffleValidate,
876 reasonIfUnsupported,
877 input,
878 output,
879 descriptor);
880}
881
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100882bool NeonLayerSupport::IsComparisonSupported(const TensorInfo& input0,
883 const TensorInfo& input1,
884 const TensorInfo& output,
885 const ComparisonDescriptor& descriptor,
886 Optional<std::string&> reasonIfUnsupported) const
887{
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100888
Teresa Charlincedd34f2020-03-30 11:17:30 +0100889 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonComparisonWorkloadValidate,
890 reasonIfUnsupported,
891 input0,
892 input1,
893 output,
894 descriptor);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100895}
896
Jim Flynn906f9462019-05-10 13:55:21 +0100897bool NeonLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inputs,
898 const TensorInfo& output,
Cathal Corbett34b429c2021-12-24 12:24:40 +0000899 const OriginsDescriptor& descriptor,
Jim Flynn906f9462019-05-10 13:55:21 +0100900 Optional<std::string&> reasonIfUnsupported) const
901{
Jim Flynne242f2d2019-05-22 14:24:13 +0100902 if (descriptor.GetNumDimensions() <= descriptor.GetConcatAxis())
903 {
904 SetValueChecked(reasonIfUnsupported, "Neon Concat: Concat axis > Number of dimensions.");
905 return false;
906 }
907
908 unsigned int concatInnerAxis = (descriptor.GetNumDimensions() - descriptor.GetConcatAxis()) - 1;
909 if(concatInnerAxis < 3) // Width, height, or channels
910 {
911 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConcatWorkloadValidate,
912 reasonIfUnsupported,
913 inputs,
914 output,
915 descriptor);
916 }
917 else if (concatInnerAxis == 3)
918 {
919 for (auto& input : inputs)
920 {
921 if (input && !output.IsTypeSpaceMatch(*input)) // Cannot use sub-tensors if the types are not same space
922 {
923 SetValueChecked(reasonIfUnsupported, "Neon Concat: Types and quantization parameters must match.");
924 return false;
925 }
926 }
927 return true; // Sub-tensors support concat along batch
928 }
929 else // > 4 dimensions not supported.
930 {
931 SetValueChecked(reasonIfUnsupported, "Neon Concat: Maximum of 4 dimensions supported.");
932 return false;
933 }
Jim Flynn906f9462019-05-10 13:55:21 +0100934}
935
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100936bool NeonLayerSupport::IsConstantSupported(const TensorInfo& output,
937 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000938{
Mike Kelly0886ac42020-04-27 09:55:40 +0100939 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConstantWorkloadValidate,
940 reasonIfUnsupported,
941 output);
telsoa014fcda012018-03-09 14:13:49 +0000942}
943
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100944bool NeonLayerSupport::IsConvertFp16ToFp32Supported(const TensorInfo& input,
945 const TensorInfo& output,
946 Optional<std::string&> reasonIfUnsupported) const
947{
Matthew Bentham34336f92023-04-27 12:13:50 +0000948 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvertFp16ToFp32WorkloadValidate,
949 reasonIfUnsupported,
950 input,
951 output);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100952}
953
954bool NeonLayerSupport::IsConvertFp32ToFp16Supported(const TensorInfo& input,
955 const TensorInfo& output,
956 Optional<std::string&> reasonIfUnsupported) const
957{
Matthew Bentham34336f92023-04-27 12:13:50 +0000958 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvertFp32ToFp16WorkloadValidate,
959 reasonIfUnsupported,
960 input,
961 output);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +0100962}
963
964bool NeonLayerSupport::IsConvolution2dSupported(const TensorInfo& input,
965 const TensorInfo& output,
966 const Convolution2dDescriptor& descriptor,
967 const TensorInfo& weights,
968 const Optional<TensorInfo>& biases,
969 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +0000970{
Sadik Armagan045f6be2020-09-10 13:37:32 +0100971 bool isFastMathEnabled = false;
972#if defined(ARMCOMPUTENEON_ENABLED)
973 if (m_ModelContextPtr)
974 {
975 if (m_ModelContextPtr.get() != nullptr)
976 {
Sadik Armagan04a72972020-09-14 15:44:18 +0100977 auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
Sadik Armagan045f6be2020-09-10 13:37:32 +0100978 if (modelOptions)
979 {
980 isFastMathEnabled = modelOptions->IsFastMathEnabled();
981 }
982 }
983 }
984#endif
985
surmeh013537c2c2018-05-18 16:31:43 +0100986 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvolution2dWorkloadValidate,
987 reasonIfUnsupported,
988 input,
989 output,
990 descriptor,
991 weights,
Sadik Armagan045f6be2020-09-10 13:37:32 +0100992 biases,
Mike Kelly07810fc2020-11-12 10:58:48 +0000993 isFastMathEnabled,
994 nullptr);
telsoa014fcda012018-03-09 14:13:49 +0000995}
996
Teresa Charlinec5f7d12021-10-22 17:15:00 +0100997bool NeonLayerSupport::IsConvolution3dSupported(const TensorInfo& input,
998 const TensorInfo& output,
999 const Convolution3dDescriptor& descriptor,
1000 const TensorInfo& weights,
1001 const Optional<TensorInfo>& biases,
1002 Optional<std::string&> reasonIfUnsupported) const
1003{
1004 bool isFastMathEnabled = false;
1005#if defined(ARMCOMPUTENEON_ENABLED)
1006 if (m_ModelContextPtr)
1007 {
1008 if (m_ModelContextPtr.get() != nullptr)
1009 {
1010 auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
1011 if (modelOptions)
1012 {
1013 isFastMathEnabled = modelOptions->IsFastMathEnabled();
1014 }
1015 }
1016 }
1017#endif
1018
1019 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonConvolution3dWorkloadValidate,
1020 reasonIfUnsupported,
1021 input,
1022 output,
1023 descriptor,
1024 weights,
1025 biases,
1026 isFastMathEnabled,
1027 nullptr);
1028}
1029
Aron Virginas-Tar2f00b742019-09-30 13:28:08 +01001030bool NeonLayerSupport::IsDepthToSpaceSupported(const TensorInfo& input,
1031 const TensorInfo& output,
1032 const DepthToSpaceDescriptor& descriptor,
1033 Optional<std::string&> reasonIfUnsupported) const
1034{
1035 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDepthToSpaceWorkloadValidate,
1036 reasonIfUnsupported,
1037 input,
1038 output,
1039 descriptor);
1040}
1041
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001042bool NeonLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input,
1043 const TensorInfo& output,
1044 const DepthwiseConvolution2dDescriptor& descriptor,
1045 const TensorInfo& weights,
1046 const Optional<TensorInfo>& biases,
1047 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +00001048{
telsoa01c577f2c2018-08-31 09:22:23 +01001049 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDepthwiseConvolutionWorkloadValidate,
1050 reasonIfUnsupported,
1051 input,
1052 output,
1053 descriptor,
1054 weights,
Mike Kelly07810fc2020-11-12 10:58:48 +00001055 biases,
1056 nullptr);
telsoa014fcda012018-03-09 14:13:49 +00001057}
1058
Narumol Prangnawarat01961a72019-05-30 16:47:12 +01001059bool NeonLayerSupport::IsDequantizeSupported(const TensorInfo& input,
1060 const TensorInfo& output,
1061 Optional<std::string&> reasonIfUnsupported) const
1062{
1063 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDequantizeWorkloadValidate,
1064 reasonIfUnsupported,
1065 input,
1066 output);
1067}
1068
Pablo Tellof0bd6832019-04-26 17:58:13 +01001069bool NeonLayerSupport::IsDilatedDepthwiseConvolutionSupported(const TensorInfo& input,
1070 const TensorInfo& output,
1071 const DepthwiseConvolution2dDescriptor& descriptor,
1072 const TensorInfo& weights,
1073 const Optional<TensorInfo>& biases,
1074 Optional<std::string&> reasonIfUnsupported) const
1075{
1076 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDepthwiseConvolutionWorkloadValidate,
1077 reasonIfUnsupported,
1078 input,
1079 output,
1080 descriptor,
1081 weights,
Mike Kelly07810fc2020-11-12 10:58:48 +00001082 biases,
1083 nullptr);
Pablo Tellof0bd6832019-04-26 17:58:13 +01001084}
1085
josh minor4a3c6102020-01-06 16:40:46 -06001086bool NeonLayerSupport::IsElementwiseUnarySupported(const TensorInfo& input,
1087 const TensorInfo& output,
1088 const ElementwiseUnaryDescriptor& descriptor,
1089 Optional<std::string&> reasonIfUnsupported) const
1090{
Derek Lambertic77874a2020-04-28 13:34:56 +01001091 switch(descriptor.m_Operation)
josh minor4a3c6102020-01-06 16:40:46 -06001092 {
Derek Lambertic77874a2020-04-28 13:34:56 +01001093 case UnaryOperation::Abs:
1094 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonAbsWorkloadValidate,
1095 reasonIfUnsupported,
1096 input,
1097 output);
1098 case UnaryOperation::Exp:
1099 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonExpWorkloadValidate,
1100 reasonIfUnsupported,
1101 input,
1102 output);
Teresa Charlin50de4fa2021-05-31 18:47:33 +01001103 case UnaryOperation::LogicalNot:
1104 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogicalNotWorkloadValidate,
1105 reasonIfUnsupported,
1106 input,
1107 output);
1108 case UnaryOperation::Log:
1109 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogWorkloadValidate,
1110 reasonIfUnsupported,
1111 input,
1112 output);
Derek Lambertic77874a2020-04-28 13:34:56 +01001113 case UnaryOperation::Neg:
1114 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonNegWorkloadValidate,
1115 reasonIfUnsupported,
1116 input,
1117 output);
1118 case UnaryOperation::Rsqrt:
1119 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonRsqrtWorkloadValidate,
1120 reasonIfUnsupported,
1121 input,
1122 output);
ryan.oshea3cff135b2021-10-07 15:28:14 +00001123 case UnaryOperation::Sin:
1124 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSinWorkloadValidate,
1125 reasonIfUnsupported,
1126 input,
1127 output);
Teresa Charlin06145cc2022-05-05 15:31:30 +01001128 case UnaryOperation::Sqrt:
1129 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSqrtWorkloadValidate,
1130 reasonIfUnsupported,
1131 input,
1132 output);
Derek Lambertic77874a2020-04-28 13:34:56 +01001133 default:
1134 return false;
josh minor4a3c6102020-01-06 16:40:46 -06001135 }
josh minor4a3c6102020-01-06 16:40:46 -06001136}
1137
Teresa Charlin4b10fef2020-07-29 09:36:41 +01001138bool NeonLayerSupport::IsFillSupported(const TensorInfo& input,
1139 const TensorInfo& output,
1140 const FillDescriptor& descriptor,
1141 Optional<std::string&> reasonIfUnsupported) const
Sadik Armagana792a052020-06-23 16:22:23 +01001142{
Teresa Charlin4b10fef2020-07-29 09:36:41 +01001143 armnn::IgnoreUnused(input);
1144 armnn::IgnoreUnused(output);
1145 armnn::IgnoreUnused(descriptor);
1146
1147 return IsNeonBackendSupported(reasonIfUnsupported);
Sadik Armagana792a052020-06-23 16:22:23 +01001148}
1149
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001150bool NeonLayerSupport::IsFloorSupported(const TensorInfo& input,
1151 const TensorInfo& output,
1152 Optional<std::string&> reasonIfUnsupported) const
1153{
Jan Eilers8eb25602020-03-09 12:13:48 +00001154 armnn::IgnoreUnused(output);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001155 return IsNeonBackendSupported(reasonIfUnsupported) &&
1156 IsSupportedForDataTypeGeneric(reasonIfUnsupported,
1157 input.GetDataType(),
1158 &FalseFuncF16<>,
1159 &TrueFunc<>,
narpra01db2b1602019-01-23 15:23:11 +00001160 &FalseFuncU8<>,
kevmay012b4d88e2019-01-24 14:05:09 +00001161 &FalseFuncI32<>,
1162 &FalseFuncU8<>);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001163}
1164
1165bool NeonLayerSupport::IsFullyConnectedSupported(const TensorInfo& input,
1166 const TensorInfo& output,
1167 const TensorInfo& weights,
1168 const TensorInfo& biases,
1169 const FullyConnectedDescriptor& descriptor,
1170 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +00001171{
telsoa01c577f2c2018-08-31 09:22:23 +01001172 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonFullyConnectedWorkloadValidate,
1173 reasonIfUnsupported,
1174 input,
1175 output,
1176 weights,
1177 biases,
Mike Kelly07810fc2020-11-12 10:58:48 +00001178 descriptor,
1179 nullptr);
telsoa014fcda012018-03-09 14:13:49 +00001180}
1181
Teresa Charlin9145e382023-08-17 18:44:58 +01001182bool NeonLayerSupport::IsFusedSupported(const std::vector<std::reference_wrapper<TensorInfo>>& inputs,
1183 const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
1184 const FusedDescriptor& descriptor,
1185 Optional<std::string&> reasonIfUnsupported) const
1186{
1187 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonFusedWorkloadValidate,
1188 reasonIfUnsupported,
1189 inputs,
1190 outputs,
1191 descriptor,
1192 nullptr);
1193}
1194
Teresa Charlinf540eb82020-04-10 19:24:55 +01001195bool NeonLayerSupport::IsGatherSupported(const TensorInfo& input0,
1196 const TensorInfo& input1,
1197 const TensorInfo& output,
Teresa Charlin52664732020-06-29 16:27:03 +01001198 const GatherDescriptor& descriptor,
Teresa Charlinf540eb82020-04-10 19:24:55 +01001199 Optional<std::string&> reasonIfUnsupported) const
1200{
1201 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonGatherWorkloadValidate,
1202 reasonIfUnsupported,
1203 input0,
1204 input1,
Teresa Charlin52664732020-06-29 16:27:03 +01001205 output,
1206 descriptor);
Teresa Charlinf540eb82020-04-10 19:24:55 +01001207}
1208
Teresa Charlinbd22c7d2022-04-26 18:14:12 +01001209bool NeonLayerSupport::IsGatherNdSupported(const TensorInfo& input0,
1210 const TensorInfo& input1,
1211 const TensorInfo& output,
1212 Optional<std::string&> reasonIfUnsupported) const
1213{
1214 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonGatherNdWorkloadValidate,
1215 reasonIfUnsupported,
1216 input0,
1217 input1,
1218 output);
1219}
1220
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001221bool NeonLayerSupport::IsInputSupported(const TensorInfo& input,
1222 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +00001223{
Derek Lamberti901ea112019-12-10 22:07:09 +00001224 return IsNeonBackendSupported(reasonIfUnsupported, input);
telsoa014fcda012018-03-09 14:13:49 +00001225}
1226
Sadik Armagan0d4863d2019-10-09 14:26:32 +01001227bool NeonLayerSupport::IsInstanceNormalizationSupported(const TensorInfo& input,
1228 const TensorInfo& output,
1229 const InstanceNormalizationDescriptor& descriptor,
1230 Optional<std::string&> reasonIfUnsupported) const
1231{
1232 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonInstanceNormalizationWorkloadValidate,
1233 reasonIfUnsupported,
1234 input,
1235 output,
1236 descriptor);
1237}
1238
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001239bool NeonLayerSupport::IsL2NormalizationSupported(const TensorInfo& input,
1240 const TensorInfo& output,
1241 const L2NormalizationDescriptor& descriptor,
1242 Optional<std::string&> reasonIfUnsupported) const
telsoa014fcda012018-03-09 14:13:49 +00001243{
Matteo Martincighbcd3c852018-09-28 14:14:12 +01001244 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonL2NormalizationWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +00001245}
1246
James Conroy177df1e2020-11-13 10:18:51 +00001247bool NeonLayerSupport::IsLogicalBinarySupported(const TensorInfo& input0,
1248 const TensorInfo& input1,
1249 const TensorInfo& output,
1250 const LogicalBinaryDescriptor& descriptor,
1251 Optional<std::string&> reasonIfUnsupported) const
1252{
1253 switch(descriptor.m_Operation)
1254 {
1255 case LogicalBinaryOperation::LogicalAnd:
1256 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogicalAndWorkloadValidate,
1257 reasonIfUnsupported,
1258 input0,
1259 input1,
1260 output);
1261 case LogicalBinaryOperation::LogicalOr:
1262 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogicalOrWorkloadValidate,
1263 reasonIfUnsupported,
1264 input0,
1265 input1,
1266 output);
1267 default:
1268 return false;
1269 }
1270}
1271
Keith Davis69e653f2020-07-02 11:49:26 +01001272bool NeonLayerSupport::IsLogSoftmaxSupported(const TensorInfo& input,
1273 const TensorInfo& output,
1274 const LogSoftmaxDescriptor& descriptor,
1275 Optional<std::string&> reasonIfUnsupported) const
1276{
1277 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1278}
1279
Jan Eilersad5293a2019-07-08 09:57:55 +01001280bool NeonLayerSupport::IsLstmSupported(const TensorInfo& input,
1281 const TensorInfo& outputStateIn,
1282 const TensorInfo& cellStateIn,
1283 const TensorInfo& scratchBuffer,
1284 const TensorInfo& outputStateOut,
1285 const TensorInfo& cellStateOut,
1286 const TensorInfo& output,
1287 const LstmDescriptor& descriptor,
1288 const LstmInputParamsInfo& paramsInfo,
1289 Optional<std::string&> reasonIfUnsupported) const
1290{
1291 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLstmFloatWorkloadValidate,
1292 reasonIfUnsupported,
1293 input,
1294 outputStateIn,
1295 cellStateIn,
1296 scratchBuffer,
1297 outputStateOut,
1298 cellStateOut,
1299 output,
1300 descriptor,
1301 paramsInfo);
1302}
1303
Nattapat Chaimanowong4e6597a2018-12-20 14:14:06 +00001304bool NeonLayerSupport::IsMaximumSupported(const TensorInfo& input0,
1305 const TensorInfo& input1,
1306 const TensorInfo& output,
1307 Optional<std::string&> reasonIfUnsupported) const
1308{
1309 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMaximumWorkloadValidate,
1310 reasonIfUnsupported,
1311 input0,
1312 input1,
1313 output);
1314}
1315
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001316bool NeonLayerSupport::IsMeanSupported(const TensorInfo& input,
1317 const TensorInfo& output,
1318 const MeanDescriptor& descriptor,
1319 Optional<std::string&> reasonIfUnsupported) const
narpra0132b90462018-09-13 11:07:48 +01001320{
Matthew Benthamfd899962018-12-31 15:49:42 +00001321 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMeanWorkloadValidate,
1322 reasonIfUnsupported,
1323 input,
1324 output,
1325 descriptor);
narpra0132b90462018-09-13 11:07:48 +01001326}
1327
Conor Kennedy54b21692019-01-09 07:57:38 +00001328bool NeonLayerSupport::IsMinimumSupported(const TensorInfo& input0,
1329 const TensorInfo& input1,
1330 const TensorInfo& output,
1331 Optional<std::string&> reasonIfUnsupported) const
1332{
1333 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMinimumWorkloadValidate,
1334 reasonIfUnsupported,
1335 input0,
1336 input1,
1337 output);
1338}
1339
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001340bool NeonLayerSupport::IsMultiplicationSupported(const TensorInfo& input0,
1341 const TensorInfo& input1,
1342 const TensorInfo& output,
1343 Optional<std::string&> reasonIfUnsupported) const
1344{
1345 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonMultiplicationWorkloadValidate,
1346 reasonIfUnsupported,
1347 input0,
1348 input1,
Mike Kelly07810fc2020-11-12 10:58:48 +00001349 output,
1350 nullptr);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001351}
1352
Pablo Telloe61f0712020-01-23 10:37:17 +00001353bool NeonLayerSupport::IsDivisionSupported(const TensorInfo& input0,
1354 const TensorInfo& input1,
1355 const TensorInfo& output,
1356 Optional<std::string&> reasonIfUnsupported) const
1357{
1358 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDivisionWorkloadValidate,
1359 reasonIfUnsupported,
1360 input0,
1361 input1,
Mike Kelly07810fc2020-11-12 10:58:48 +00001362 output,
1363 nullptr);
Pablo Telloe61f0712020-01-23 10:37:17 +00001364}
1365
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001366bool NeonLayerSupport::IsNormalizationSupported(const TensorInfo& input,
1367 const TensorInfo& output,
1368 const NormalizationDescriptor& descriptor,
1369 Optional<std::string&> reasonIfUnsupported) const
1370{
1371 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonNormalizationWorkloadValidate,
1372 reasonIfUnsupported,
1373 input,
1374 output,
1375 descriptor);
1376}
1377
1378bool NeonLayerSupport::IsOutputSupported(const TensorInfo& output,
1379 Optional<std::string&> reasonIfUnsupported) const
1380{
Derek Lamberti901ea112019-12-10 22:07:09 +00001381 return IsNeonBackendSupported(reasonIfUnsupported, output);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001382}
1383
Éanna Ó Catháin12055742019-01-25 10:01:40 +00001384bool NeonLayerSupport::IsPadSupported(const TensorInfo& input,
1385 const TensorInfo& output,
1386 const PadDescriptor& descriptor,
1387 Optional<std::string&> reasonIfUnsupported) const
1388{
1389 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPadWorkloadValidate,
1390 reasonIfUnsupported,
1391 input,
1392 output,
1393 descriptor);
1394}
1395
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001396bool NeonLayerSupport::IsPermuteSupported(const TensorInfo& input,
1397 const TensorInfo& output,
1398 const PermuteDescriptor& descriptor,
1399 Optional<std::string&> reasonIfUnsupported) const
1400{
1401 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPermuteWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
telsoa014fcda012018-03-09 14:13:49 +00001402}
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001403
1404bool NeonLayerSupport::IsPooling2dSupported(const TensorInfo& input,
1405 const TensorInfo& output,
1406 const Pooling2dDescriptor& descriptor,
1407 Optional<std::string&> reasonIfUnsupported) const
1408{
1409 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling2dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1410}
1411
Ryan OShea19e79422022-05-04 00:38:03 +01001412bool NeonLayerSupport::IsPooling3dSupported(const TensorInfo& input,
1413 const TensorInfo& output,
1414 const Pooling3dDescriptor& descriptor,
1415 Optional<std::string&> reasonIfUnsupported) const
1416{
1417 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling3dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1418}
1419
Nikhil Raj9b461482019-07-03 15:58:31 +01001420bool NeonLayerSupport::IsPreluSupported(const armnn::TensorInfo &input,
1421 const armnn::TensorInfo &alpha,
1422 const armnn::TensorInfo &output,
1423 armnn::Optional<std::string &> reasonIfUnsupported) const
1424{
1425 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPreluWorkloadValidate, reasonIfUnsupported, input, alpha, output);
1426}
1427
James Conroycc340932020-05-12 18:08:52 +01001428bool NeonLayerSupport::IsQLstmSupported(const TensorInfo& input,
1429 const TensorInfo& previousOutputIn,
1430 const TensorInfo& previousCellStateIn,
1431 const TensorInfo& outputStateOut,
1432 const TensorInfo& cellStateOut,
1433 const TensorInfo& output,
1434 const QLstmDescriptor& descriptor,
1435 const LstmInputParamsInfo& paramsInfo,
1436 Optional<std::string&> reasonIfUnsupported) const
1437{
1438 // Check required here in order to pass IsLayerSupported for datatypes tests
1439 if (input.GetDataType() == armnn::DataType::QAsymmS8 &&
1440 previousOutputIn.GetDataType() == armnn::DataType::QAsymmS8 &&
1441 previousCellStateIn.GetDataType() == armnn::DataType::QSymmS16 &&
1442 outputStateOut.GetDataType() == armnn::DataType::QAsymmS8 &&
1443 cellStateOut.GetDataType() == armnn::DataType::QSymmS16 &&
1444 output.GetDataType() == armnn::DataType::QAsymmS8)
1445 {
1446 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonQLstmWorkloadValidate,
1447 reasonIfUnsupported,
1448 input,
1449 previousCellStateIn,
1450 previousOutputIn,
1451 cellStateOut,
1452 outputStateOut,
1453 output,
1454 descriptor,
1455 paramsInfo);
1456 }
1457 else
1458 {
1459 return false;
1460 }
1461}
1462
Sadik Armaganfabc2892019-05-31 09:05:11 +01001463bool NeonLayerSupport::IsQuantizeSupported(const TensorInfo& input,
1464 const TensorInfo& output,
1465 Optional<std::string&> reasonIfUnsupported) const
1466{
1467 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonQuantizeWorkloadValidate,
1468 reasonIfUnsupported,
1469 input,
1470 output);
1471}
1472
Francis Murtagh4fc3c482019-08-02 13:20:54 +01001473bool NeonLayerSupport::IsQuantizedLstmSupported(const TensorInfo& input,
1474 const TensorInfo& cellStateIn,
1475 const TensorInfo& outputStateIn,
1476 const TensorInfo& cellStateOut,
1477 const TensorInfo& outputStateOut,
1478 const QuantizedLstmInputParamsInfo& paramsInfo,
1479 Optional<std::string&> reasonIfUnsupported) const
1480{
1481 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonQuantizedLstmWorkloadValidate,
1482 reasonIfUnsupported,
1483 input,
1484 cellStateIn,
1485 outputStateIn,
1486 cellStateOut,
1487 outputStateOut,
1488 paramsInfo);
1489}
1490
Sadik Armagana2747482021-02-09 10:28:54 +00001491bool NeonLayerSupport::IsReduceSupported(const TensorInfo& input,
1492 const TensorInfo& output,
1493 const ReduceDescriptor& descriptor,
1494 Optional<std::string&> reasonIfUnsupported) const
1495{
1496 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonReduceWorkloadValidate,
1497 reasonIfUnsupported,
1498 input,
1499 output,
1500 descriptor);
1501}
1502
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001503bool NeonLayerSupport::IsReshapeSupported(const TensorInfo& input,
Kevin Maya023c402019-12-12 17:28:05 +00001504 const TensorInfo& output,
Matteo Martincigh992d6dc2019-01-10 17:34:20 +00001505 const ReshapeDescriptor& descriptor,
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001506 Optional<std::string&> reasonIfUnsupported) const
1507{
Jan Eilers8eb25602020-03-09 12:13:48 +00001508 armnn::IgnoreUnused(descriptor);
Kevin Maya023c402019-12-12 17:28:05 +00001509 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonReshapeWorkloadValidate,
1510 reasonIfUnsupported,
1511 input,
1512 output);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001513}
1514
Aron Virginas-Tar169d2f12019-07-01 19:01:44 +01001515bool NeonLayerSupport::IsResizeSupported(const TensorInfo& input,
1516 const TensorInfo& output,
1517 const ResizeDescriptor& descriptor,
1518 Optional<std::string&> reasonIfUnsupported) const
1519{
Ellen Norris-Thompson37e68682019-07-15 14:23:30 +01001520 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonResizeWorkloadValidate,
1521 reasonIfUnsupported,
1522 input,
1523 output,
1524 descriptor);
Aron Virginas-Tar169d2f12019-07-01 19:01:44 +01001525}
1526
Tianle Cheng21a9f332023-11-09 13:56:53 +00001527bool NeonLayerSupport::IsReverseV2Supported(const armnn::TensorInfo &input,
1528 const armnn::TensorInfo &axis,
1529 const armnn::TensorInfo &output,
1530 Optional<std::string &> reasonIfUnsupported) const
1531{
1532 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonReverseV2WorkloadValidate,
1533 reasonIfUnsupported,
1534 input,
1535 axis,
1536 output);
1537}
1538
josh minor036f02d2019-11-15 14:53:22 -06001539bool NeonLayerSupport::IsSliceSupported(const TensorInfo& input,
1540 const TensorInfo& output,
1541 const SliceDescriptor& descriptor,
1542 Optional<std::string&> reasonIfUnsupported) const
1543{
1544 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSliceWorkloadValidate,
1545 reasonIfUnsupported,
1546 input,
1547 output,
1548 descriptor);
1549}
1550
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001551bool NeonLayerSupport::IsSoftmaxSupported(const TensorInfo& input,
1552 const TensorInfo& output,
1553 const SoftmaxDescriptor& descriptor,
1554 Optional<std::string&> reasonIfUnsupported) const
1555{
1556 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1557}
1558
Mike Kelly0be3a882020-01-24 11:27:50 +00001559bool NeonLayerSupport::IsSpaceToBatchNdSupported(const TensorInfo& input,
1560 const TensorInfo& output,
1561 const SpaceToBatchNdDescriptor& descriptor,
1562 Optional<std::string&> reasonIfUnsupported) const
1563{
1564 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSpaceToBatchNdWorkloadValidate,
1565 reasonIfUnsupported,
1566 input,
1567 output,
1568 descriptor);
1569}
1570
Ellen Norris-Thompson29794572019-06-26 16:40:36 +01001571bool NeonLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input,
Mike Kelly0be3a882020-01-24 11:27:50 +00001572 const TensorInfo& output,
1573 const SpaceToDepthDescriptor& descriptor,
1574 Optional<std::string&> reasonIfUnsupported) const
Ellen Norris-Thompson29794572019-06-26 16:40:36 +01001575{
1576 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSpaceToDepthWorkloadValidate,
1577 reasonIfUnsupported,
1578 input,
1579 output,
1580 descriptor);
1581}
1582
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001583bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input,
Narumol Prangnawarat15eb5832019-05-20 15:31:05 +01001584 const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
1585 const ViewsDescriptor& descriptor,
1586 Optional<std::string&> reasonIfUnsupported) const
1587{
1588#if defined(ARMCOMPUTENEON_ENABLED)
1589 // Split along the last dimension, cannot use sub-tensors
1590 // as width and height of the sub-tensors do not match
1591 // the width and height of the parent tensor
1592 // in case of input with more than 2D.
1593 std::set<unsigned int> splitAxis = ComputeSplitAxis(descriptor, input.GetShape());
1594 if (descriptor.GetNumDimensions() > 2 && splitAxis.size() == 1 &&
1595 *splitAxis.begin() == descriptor.GetNumDimensions() - 1 )
1596 {
1597 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSplitterWorkloadValidate,
1598 reasonIfUnsupported,
1599 input,
1600 outputs,
1601 *splitAxis.begin());
1602 }
1603#endif
Jan Eilers8eb25602020-03-09 12:13:48 +00001604 IgnoreUnused(descriptor);
Narumol Prangnawarat15eb5832019-05-20 15:31:05 +01001605 for (auto output : outputs)
1606 {
1607 if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space
1608 {
1609 SetValueChecked(reasonIfUnsupported, "Neon Splitter: Types and quantization parameters must match.");
1610 return false;
1611 }
1612 }
1613 return true;
1614}
1615
Matthew Jackson87f65ea2019-08-01 10:01:34 +01001616bool NeonLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inputs,
1617 const TensorInfo& output,
1618 const StackDescriptor& descriptor,
1619 Optional<std::string&> reasonIfUnsupported) const
1620{
1621 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonStackWorkloadValidate,
1622 reasonIfUnsupported,
1623 inputs,
1624 output,
1625 descriptor);
1626}
1627
FinnWilliamsArm1fa19192019-08-02 17:26:31 +01001628bool NeonLayerSupport::IsStridedSliceSupported(const TensorInfo& input,
1629 const TensorInfo& output,
1630 const StridedSliceDescriptor& descriptor,
1631 Optional<std::string&> reasonIfUnsupported) const
1632{
1633 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonStridedSliceWorkloadValidate,
1634 reasonIfUnsupported,
1635 input,
1636 output,
1637 descriptor);
1638}
1639
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001640bool NeonLayerSupport::IsSubtractionSupported(const TensorInfo& input0,
1641 const TensorInfo& input1,
1642 const TensorInfo& output,
1643 Optional<std::string&> reasonIfUnsupported) const
1644{
1645 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSubtractionWorkloadValidate,
1646 reasonIfUnsupported,
1647 input0,
1648 input1,
Mike Kelly07810fc2020-11-12 10:58:48 +00001649 output,
1650 nullptr);
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001651}
1652
David Monahan36e6eae2023-07-26 18:37:45 +01001653bool NeonLayerSupport::IsTileSupported(const TensorInfo& input,
1654 const TensorInfo& output,
1655 const TileDescriptor& descriptor,
1656 Optional<std::string&> reasonIfUnsupported) const
1657{
1658 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonTileWorkloadValidate,
1659 reasonIfUnsupported,
1660 input,
1661 output,
1662 descriptor);
1663}
1664
Sadik Armagan581742d2019-08-12 14:11:37 +01001665bool NeonLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input,
1666 const TensorInfo& output,
1667 const TransposeConvolution2dDescriptor& descriptor,
1668 const TensorInfo& weights,
1669 const Optional<TensorInfo>& biases,
1670 Optional<std::string&> reasonIfUnsupported) const
1671{
1672 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonTransposeConvolution2dWorkloadValidate,
1673 reasonIfUnsupported,
1674 input,
1675 output,
1676 descriptor,
1677 weights,
1678 biases);
1679}
1680
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001681bool NeonLayerSupport::IsTransposeSupported(const TensorInfo& input,
1682 const TensorInfo& output,
1683 const TransposeDescriptor& descriptor,
1684 Optional<std::string&> reasonIfUnsupported) const
1685{
1686 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonTransposeWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1687}
1688
Cathal Corbettfd5bec42022-03-03 15:13:23 +00001689bool NeonLayerSupport::IsUnidirectionalSequenceLstmSupported(const TensorInfo& input,
1690 const TensorInfo& outputStateIn,
1691 const TensorInfo& cellStateIn,
Mike Kelly12994962022-04-21 11:57:09 +01001692 const TensorInfo& outputStateOut,
1693 const TensorInfo& cellStateOut,
Cathal Corbettfd5bec42022-03-03 15:13:23 +00001694 const TensorInfo& output,
Cathal Corbettfd5bec42022-03-03 15:13:23 +00001695 const UnidirectionalSequenceLstmDescriptor& descriptor,
1696 const LstmInputParamsInfo& paramsInfo,
1697 Optional<std::string&> reasonIfUnsupported) const
1698{
Mike Kelly12994962022-04-21 11:57:09 +01001699 if (input.GetDataType() == armnn::DataType::QAsymmS8 &&
1700 outputStateIn.GetDataType() == armnn::DataType::QAsymmS8 &&
1701 cellStateIn.GetDataType() == armnn::DataType::QSymmS16 &&
1702 outputStateOut.GetDataType() == armnn::DataType::QAsymmS8 &&
1703 cellStateOut.GetDataType() == armnn::DataType::QSymmS16 &&
1704 output.GetDataType() == armnn::DataType::QAsymmS8)
1705 {
1706 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonUnidirectionalSequenceLstmWorkloadValidate,
1707 reasonIfUnsupported,
1708 input,
1709 outputStateIn,
1710 cellStateIn,
1711 outputStateOut,
1712 cellStateOut,
1713 output,
1714 descriptor,
1715 paramsInfo);
1716 }
1717 else
1718 {
1719 FORWARD_WORKLOAD_VALIDATE_FUNC(NeonUnidirectionalSequenceLstmFloatWorkloadValidate,
1720 reasonIfUnsupported,
1721 input,
1722 outputStateIn,
1723 cellStateIn,
1724 outputStateOut,
1725 cellStateOut,
1726 output,
1727 descriptor,
1728 paramsInfo);
1729 }
Cathal Corbettfd5bec42022-03-03 15:13:23 +00001730}
1731
Aron Virginas-Tarfc824312018-10-15 15:00:13 +01001732} // namespace armnn