blob: 7d5f74212682bb0babc27eb1cc742637358122f4 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Mike Kelly3ec30772023-03-08 13:47:17 +00002// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
Matthew Bentham4cefc412019-06-18 16:14:34 +01005#include <Layer.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +00006#include <armnn/backends/MemCopyWorkload.hpp>
Derek Lambertif674aa02019-08-01 15:56:25 +01007#include <backendsCommon/MemImportWorkload.hpp>
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +00008#include <backendsCommon/MakeWorkloadHelper.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +00009#include <armnn/backends/TensorHandle.hpp>
telsoa014fcda012018-03-09 14:13:49 +000010#include "RefWorkloadFactory.hpp"
David Beck79141b92018-10-23 16:09:36 +010011#include "RefBackendId.hpp"
David Beckb4540be2018-09-24 13:18:27 +010012#include "workloads/RefWorkloads.hpp"
Matthew Bentham4cefc412019-06-18 16:14:34 +010013#include "RefTensorHandle.hpp"
telsoa014fcda012018-03-09 14:13:49 +000014
telsoa014fcda012018-03-09 14:13:49 +000015
16namespace armnn
17{
18
David Beck79141b92018-10-23 16:09:36 +010019namespace
20{
21static const BackendId s_Id{RefBackendId()};
22}
telsoa014fcda012018-03-09 14:13:49 +000023template <typename F32Workload, typename U8Workload, typename QueueDescriptorType>
24std::unique_ptr<IWorkload> RefWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
Aron Virginas-Tare662a942019-10-14 15:12:00 +010025 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +000026{
Keith Davis5204aa82020-01-27 15:24:59 +000027 return MakeWorkloadHelper<NullWorkload, F32Workload, U8Workload, NullWorkload, NullWorkload, NullWorkload>
28 (descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +000029}
30
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010031template <DataType ArmnnType>
32bool IsDataType(const WorkloadInfo& info)
Jim Flynn82fbe7c2019-04-02 15:19:08 +010033{
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010034 auto checkType = [](const TensorInfo& tensorInfo) {return tensorInfo.GetDataType() == ArmnnType;};
35 auto it = std::find_if(std::begin(info.m_InputTensorInfos), std::end(info.m_InputTensorInfos), checkType);
Jim Flynn82fbe7c2019-04-02 15:19:08 +010036 if (it != std::end(info.m_InputTensorInfos))
37 {
38 return true;
39 }
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010040 it = std::find_if(std::begin(info.m_OutputTensorInfos), std::end(info.m_OutputTensorInfos), checkType);
Jim Flynn82fbe7c2019-04-02 15:19:08 +010041 if (it != std::end(info.m_OutputTensorInfos))
42 {
43 return true;
44 }
45 return false;
46}
47
Keith Davis0c2eeac2020-02-11 16:51:50 +000048bool IsSigned32(const WorkloadInfo& info)
49{
50 return IsDataType<DataType::Signed32>(info);
51}
52
Narumol Prangnawarat44179c32020-03-11 14:51:27 +000053bool IsBFloat16(const WorkloadInfo& info)
54{
55 return IsDataType<DataType::BFloat16>(info);
56}
57
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010058bool IsFloat16(const WorkloadInfo& info)
59{
60 return IsDataType<DataType::Float16>(info);
61}
62
Keith Davis0c2eeac2020-02-11 16:51:50 +000063bool IsQSymmS16(const WorkloadInfo& info)
nikraj0199a66312019-06-06 10:31:27 +010064{
Derek Lambertif90c56d2020-01-10 17:14:08 +000065 return IsDataType<DataType::QSymmS16>(info);
nikraj0199a66312019-06-06 10:31:27 +010066}
67
Keith Davis0c2eeac2020-02-11 16:51:50 +000068bool IsQSymmS8(const WorkloadInfo& info)
Keith Davis5204aa82020-01-27 15:24:59 +000069{
70 return IsDataType<DataType::QSymmS8>(info);
71}
72
Keith Davis67e6c542020-02-19 10:08:33 +000073bool IsQAsymmS8(const WorkloadInfo& info)
74{
75 return IsDataType<DataType::QAsymmS8>(info);
76}
77
78bool IsQAsymmU8(const WorkloadInfo& info)
79{
80 return IsDataType<DataType::QAsymmU8>(info);
81}
82
Matthew Bentham7c1603a2019-06-21 17:22:23 +010083RefWorkloadFactory::RefWorkloadFactory(const std::shared_ptr<RefMemoryManager>& memoryManager)
84 : m_MemoryManager(memoryManager)
85{
86}
87
telsoa01c577f2c2018-08-31 09:22:23 +010088RefWorkloadFactory::RefWorkloadFactory()
Matthew Bentham7c1603a2019-06-21 17:22:23 +010089 : m_MemoryManager(new RefMemoryManager())
telsoa014fcda012018-03-09 14:13:49 +000090{
91}
92
David Beck79141b92018-10-23 16:09:36 +010093const BackendId& RefWorkloadFactory::GetBackendId() const
94{
95 return s_Id;
96}
97
David Beck29c75de2018-10-23 13:35:58 +010098bool RefWorkloadFactory::IsLayerSupported(const Layer& layer,
99 Optional<DataType> dataType,
telsoa01c577f2c2018-08-31 09:22:23 +0100100 std::string& outReasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +0000101{
David Beck79141b92018-10-23 16:09:36 +0100102 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
telsoa014fcda012018-03-09 14:13:49 +0000103}
104
Sadik Armagan04a72972020-09-14 15:44:18 +0100105bool RefWorkloadFactory::IsLayerSupported(const IConnectableLayer& layer,
106 Optional<DataType> dataType,
107 std::string& outReasonIfUnsupported,
108 const ModelOptions& modelOptions)
109{
110 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
111}
112
David Monahan3fb7e102019-08-20 11:25:29 +0100113std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
Derek Lamberti901ea112019-12-10 22:07:09 +0000114 const bool isMemoryManaged) const
telsoa014fcda012018-03-09 14:13:49 +0000115{
Finn Williamsb1aad422021-10-28 19:07:32 +0100116 if (isMemoryManaged)
117 {
118 return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
119 }
120 else
121 {
Matthew Benthamc30abd82022-11-23 12:11:32 +0000122 return std::make_unique<RefTensorHandle>(tensorInfo);
Finn Williamsb1aad422021-10-28 19:07:32 +0100123 }
telsoa014fcda012018-03-09 14:13:49 +0000124}
125
Francis Murtagh351d13d2018-09-24 15:01:18 +0100126std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
David Monahan3fb7e102019-08-20 11:25:29 +0100127 DataLayout dataLayout,
Derek Lamberti901ea112019-12-10 22:07:09 +0000128 const bool isMemoryManaged) const
Francis Murtagh351d13d2018-09-24 15:01:18 +0100129{
David Monahan3fb7e102019-08-20 11:25:29 +0100130 // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
131 // to unmanaged memory. This also ensures memory alignment.
Jan Eilers8eb25602020-03-09 12:13:48 +0000132 IgnoreUnused(isMemoryManaged, dataLayout);
Finn Williamsb1aad422021-10-28 19:07:32 +0100133
134 if (isMemoryManaged)
135 {
136 return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
137 }
138 else
139 {
Matthew Benthamc30abd82022-11-23 12:11:32 +0000140 return std::make_unique<RefTensorHandle>(tensorInfo);
Finn Williamsb1aad422021-10-28 19:07:32 +0100141 }
Francis Murtagh351d13d2018-09-24 15:01:18 +0100142}
143
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000144std::unique_ptr<IWorkload> RefWorkloadFactory::CreateWorkload(LayerType type,
145 const QueueDescriptor& descriptor,
146 const WorkloadInfo& info) const
147{
148 switch(type)
149 {
150 case LayerType::Activation :
151 {
152 auto activationQueueDescriptor = PolymorphicDowncast<const ActivationQueueDescriptor*>(&descriptor);
153 return std::make_unique<RefActivationWorkload>(*activationQueueDescriptor, info);
154 }
155 case LayerType::Addition :
156 {
157 auto additionQueueDescriptor = PolymorphicDowncast<const AdditionQueueDescriptor*>(&descriptor);
158
159 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
160 {
161 return std::make_unique<RefAdditionWorkload<int32_t>>(*additionQueueDescriptor, info);
162 }
163 else
164 {
165 return std::make_unique<RefAdditionWorkload<float>>(*additionQueueDescriptor, info);
166 }
167 }
168 case LayerType::ArgMinMax :
169 {
170 auto argMinMaxQueueDescriptor = PolymorphicDowncast<const ArgMinMaxQueueDescriptor*>(&descriptor);
171 return std::make_unique<RefArgMinMaxWorkload>(*argMinMaxQueueDescriptor, info);
172 }
Samuel Yap6b478092022-07-06 15:36:03 +0100173 case LayerType::BatchMatMul:
174 {
175 auto batchMatMulQueueDescriptor = PolymorphicDowncast<const BatchMatMulQueueDescriptor*>(&descriptor);
176 return std::make_unique<RefBatchMatMulWorkload>(*batchMatMulQueueDescriptor, info);
177 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000178 case LayerType::BatchNormalization :
179 {
180 auto batchNormQueueDescriptor = PolymorphicDowncast<const BatchNormalizationQueueDescriptor*>(&descriptor);
181 return std::make_unique<RefBatchNormalizationWorkload>(*batchNormQueueDescriptor, info);
182 }
183 case LayerType::BatchToSpaceNd :
184 {
185 auto batchToSpaceNdQueueDescriptor
186 = PolymorphicDowncast<const BatchToSpaceNdQueueDescriptor*>(&descriptor);
187 return std::make_unique<RefBatchToSpaceNdWorkload>(*batchToSpaceNdQueueDescriptor, info);
188 }
189 case LayerType::Cast :
190 {
191 auto castQueueDescriptor = PolymorphicDowncast<const CastQueueDescriptor*>(&descriptor);
192 return std::make_unique<RefCastWorkload>(*castQueueDescriptor, info);
193 }
194 case LayerType::ChannelShuffle :
195 {
196 auto channelShuffleQueueDescriptor
197 = PolymorphicDowncast<const ChannelShuffleQueueDescriptor*>(&descriptor);
198 return std::make_unique<RefChannelShuffleWorkload>(*channelShuffleQueueDescriptor, info);
199 }
200 case LayerType::Comparison :
201 {
202 auto comparisonQueueDescriptor = PolymorphicDowncast<const ComparisonQueueDescriptor*>(&descriptor);
203 return std::make_unique<RefComparisonWorkload>(*comparisonQueueDescriptor, info);
204 }
205 case LayerType::Concat :
206 {
207 auto concatQueueDescriptor = PolymorphicDowncast<const ConcatQueueDescriptor*>(&descriptor);
208 return std::make_unique<RefConcatWorkload>(*concatQueueDescriptor, info);
209 }
210 case LayerType::Constant :
211 {
212 auto constantQueueDescriptor = PolymorphicDowncast<const ConstantQueueDescriptor*>(&descriptor);
213 return std::make_unique<RefConstantWorkload>(*constantQueueDescriptor, info);
214 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000215 case LayerType::ConvertFp16ToFp32:
216 {
217 auto convertFp16ToFp32QueueDescriptor
218 = PolymorphicDowncast<const ConvertFp16ToFp32QueueDescriptor*>(&descriptor);
219 return std::make_unique<RefConvertFp16ToFp32Workload>(*convertFp16ToFp32QueueDescriptor, info);
220 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000221 case LayerType::ConvertFp32ToFp16:
222 {
223 auto convertFp32ToFp16QueueDescriptor
224 = PolymorphicDowncast<const ConvertFp32ToFp16QueueDescriptor*>(&descriptor);
225 return std::make_unique<RefConvertFp32ToFp16Workload>(*convertFp32ToFp16QueueDescriptor, info);
226 }
227 case LayerType::Convolution2d:
228 {
229 auto convolution2dQueueDescriptor = PolymorphicDowncast<const Convolution2dQueueDescriptor*>(&descriptor);
230 return std::make_unique<RefConvolution2dWorkload>(*convolution2dQueueDescriptor, info);
231 }
232 case LayerType::Convolution3d:
233 {
234 auto convolution3dQueueDescriptor = PolymorphicDowncast<const Convolution3dQueueDescriptor*>(&descriptor);
235 return std::make_unique<RefConvolution3dWorkload>(*convolution3dQueueDescriptor, info);
236 }
237 case LayerType::Debug:
238 {
239 auto debugQueueDescriptor = PolymorphicDowncast<const DebugQueueDescriptor*>(&descriptor);
240 if (IsBFloat16(info))
241 {
242 return std::make_unique<RefDebugBFloat16Workload>(*debugQueueDescriptor, info);
243 }
244 if (IsFloat16(info))
245 {
246 return std::make_unique<RefDebugFloat16Workload>(*debugQueueDescriptor, info);
247 }
248 if (IsQSymmS16(info))
249 {
250 return std::make_unique<RefDebugQSymmS16Workload>(*debugQueueDescriptor, info);
251 }
252 if (IsQSymmS8(info))
253 {
254 return std::make_unique<RefDebugQSymmS8Workload>(*debugQueueDescriptor, info);
255 }
256 if (IsQAsymmU8(info))
257 {
258 return std::make_unique<RefDebugQAsymmU8Workload>(*debugQueueDescriptor, info);
259 }
260 if (IsQAsymmS8(info))
261 {
262 return std::make_unique<RefDebugQAsymmS8Workload>(*debugQueueDescriptor, info);
263 }
264 if (IsSigned32(info))
265 {
266 return std::make_unique<RefDebugSigned32Workload>(*debugQueueDescriptor, info);
267 }
268
269 return MakeWorkload<RefDebugFloat32Workload, RefDebugQAsymmU8Workload>(*debugQueueDescriptor, info);
270 }
271 case LayerType::DepthToSpace:
272 {
273 auto depthToSpaceQueueDescriptor = PolymorphicDowncast<const DepthToSpaceQueueDescriptor*>(&descriptor);
274 return std::make_unique<RefDepthToSpaceWorkload>(*depthToSpaceQueueDescriptor, info);
275 }
276 case LayerType::DepthwiseConvolution2d:
277 {
278 auto depthwiseConvolution2DQueueDescriptor
279 = PolymorphicDowncast<const DepthwiseConvolution2dQueueDescriptor*>(&descriptor);
280 return std::make_unique<RefDepthwiseConvolution2dWorkload>(*depthwiseConvolution2DQueueDescriptor, info);
281 }
282 case LayerType::Dequantize:
283 {
284 auto dequantizeQueueDescriptor = PolymorphicDowncast<const DequantizeQueueDescriptor*>(&descriptor);
285 return std::make_unique<RefDequantizeWorkload>(*dequantizeQueueDescriptor, info);
286 }
287 case LayerType::DetectionPostProcess:
288 {
289 auto detectionPostProcessQueueDescriptor
290 = PolymorphicDowncast<const DetectionPostProcessQueueDescriptor*>(&descriptor);
291 return std::make_unique<RefDetectionPostProcessWorkload>(*detectionPostProcessQueueDescriptor, info);
292 }
293 case LayerType::Division:
294 {
295 auto divisionQueueDescriptor = PolymorphicDowncast<const DivisionQueueDescriptor*>(&descriptor);
296 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
297 {
298 return std::make_unique<RefDivisionWorkload<int32_t>>(*divisionQueueDescriptor, info);
299 }
300 else
301 {
302 return std::make_unique<RefDivisionWorkload<float>>(*divisionQueueDescriptor, info);
303 }
304 }
Mike Kelly3ec30772023-03-08 13:47:17 +0000305 case LayerType::ElementwiseBinary:
306 {
307 auto elementwiseBinaryQueueDescriptor
308 = PolymorphicDowncast<const ElementwiseBinaryQueueDescriptor*>(&descriptor);
309 return std::make_unique<RefElementwiseBinaryWorkload>(*elementwiseBinaryQueueDescriptor, info);
310 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000311 case LayerType::ElementwiseUnary:
312 {
313 auto elementwiseUnaryQueueDescriptor
314 = PolymorphicDowncast<const ElementwiseUnaryQueueDescriptor*>(&descriptor);
315 if ((*elementwiseUnaryQueueDescriptor).m_Parameters.m_Operation == UnaryOperation::LogicalNot)
316 {
317 return std::make_unique<RefLogicalUnaryWorkload>(*elementwiseUnaryQueueDescriptor, info);
318 }
319 return std::make_unique<RefElementwiseUnaryWorkload>(*elementwiseUnaryQueueDescriptor, info);
320 }
321 case LayerType::FakeQuantization:
322 {
323 auto fakeQuantizationQueueDescriptor
324 = PolymorphicDowncast<const FakeQuantizationQueueDescriptor*>(&descriptor);
325 return std::make_unique<RefFakeQuantizationFloat32Workload>(*fakeQuantizationQueueDescriptor, info);
326 }
327 case LayerType::Fill:
328 {
329 auto fillQueueDescriptor = PolymorphicDowncast<const FillQueueDescriptor*>(&descriptor);
330 return std::make_unique<RefFillWorkload>(*fillQueueDescriptor, info);
331 }
332 case LayerType::Floor:
333 {
334 auto floorQueueDescriptor = PolymorphicDowncast<const FloorQueueDescriptor*>(&descriptor);
335 if(IsQuantizedType(info.m_InputTensorInfos[0].GetDataType()))
336 {
337 return nullptr;
338 }
339 else
340 {
341 return std::make_unique<RefFloorWorkload>(*floorQueueDescriptor, info);
342 }
343 }
344 case LayerType::FullyConnected:
345 {
346 auto fullyConnectedQueueDescriptor
347 = PolymorphicDowncast<const FullyConnectedQueueDescriptor*>(&descriptor);
348 return std::make_unique<RefFullyConnectedWorkload>(*fullyConnectedQueueDescriptor, info);
349 }
350 case LayerType::Gather:
351 {
352 auto gatherQueueDescriptor = PolymorphicDowncast<const GatherQueueDescriptor*>(&descriptor);
353 return std::make_unique<RefGatherWorkload>(*gatherQueueDescriptor, info);
354 }
Teresa Charlinb2d3ec52022-04-12 22:07:09 +0100355 case LayerType::GatherNd:
356 {
357 auto gatherNdQueueDescriptor = PolymorphicDowncast<const GatherNdQueueDescriptor*>(&descriptor);
358 return std::make_unique<RefGatherNdWorkload>(*gatherNdQueueDescriptor, info);
359 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000360 case LayerType::Input:
361 {
362 auto inputQueueDescriptor = PolymorphicDowncast<const InputQueueDescriptor*>(&descriptor);
363 if (info.m_InputTensorInfos.empty() )
364 {
365 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Input cannot be zero length");
366 }
367 if (info.m_OutputTensorInfos.empty())
368 {
369 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Output cannot be zero length");
370 }
371
372 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
373 {
374 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: "
375 "data input and output differ in byte count.");
376 }
377
378 return std::make_unique<CopyMemGenericWorkload>(*inputQueueDescriptor, info);
379 }
380 case LayerType::InstanceNormalization:
381 {
382 auto instanceNormalizationQueueDescriptor
383 = PolymorphicDowncast<const InstanceNormalizationQueueDescriptor*>(&descriptor);
384 return std::make_unique<RefInstanceNormalizationWorkload>(*instanceNormalizationQueueDescriptor, info);
385 }
386 case LayerType::L2Normalization:
387 {
388 auto l2NormalizationQueueDescriptor
389 = PolymorphicDowncast<const L2NormalizationQueueDescriptor*>(&descriptor);
390 return std::make_unique<RefL2NormalizationWorkload>(*l2NormalizationQueueDescriptor, info);
391 }
392 case LayerType::LogicalBinary:
393 {
394 auto logicalBinaryQueueDescriptor = PolymorphicDowncast<const LogicalBinaryQueueDescriptor*>(&descriptor);
395 return std::make_unique<RefLogicalBinaryWorkload>(*logicalBinaryQueueDescriptor, info);
396 }
397 case LayerType::LogSoftmax:
398 {
399 auto logSoftmaxQueueDescriptor = PolymorphicDowncast<const LogSoftmaxQueueDescriptor*>(&descriptor);
400 return std::make_unique<RefLogSoftmaxWorkload>(*logSoftmaxQueueDescriptor, info);
401 }
402 case LayerType::Lstm:
403 {
404 auto lstmQueueDescriptor = PolymorphicDowncast<const LstmQueueDescriptor*>(&descriptor);
405 return std::make_unique<RefLstmWorkload>(*lstmQueueDescriptor, info);
406 }
407 case LayerType::Maximum:
408 {
409 auto maximumQueueDescriptor = PolymorphicDowncast<const MaximumQueueDescriptor*>(&descriptor);
410 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
411 {
412 return std::make_unique<RefMaximumWorkload<int32_t>>(*maximumQueueDescriptor, info);
413 }
414 else
415 {
416 return std::make_unique<RefMaximumWorkload<float>>(*maximumQueueDescriptor, info);
417 }
418 }
419 case LayerType::Mean:
420 {
421 auto meanQueueDescriptor = PolymorphicDowncast<const MeanQueueDescriptor*>(&descriptor);
422 return std::make_unique<RefMeanWorkload>(*meanQueueDescriptor, info);
423 }
424 case LayerType::MemCopy:
425 {
426 auto memCopyQueueDescriptor = PolymorphicDowncast<const MemCopyQueueDescriptor*>(&descriptor);
427 if (descriptor.m_Inputs.empty())
428 {
429 throw InvalidArgumentException("RefWorkloadFactory: CreateMemCopy() expected an input tensor.");
430 }
431 return std::make_unique<CopyMemGenericWorkload>(*memCopyQueueDescriptor, info);
432 }
433 case LayerType::MemImport:
434 {
435 auto memImportQueueDescriptor = PolymorphicDowncast<const MemImportQueueDescriptor*>(&descriptor);
436 if (descriptor.m_Inputs.empty())
437 {
438 throw InvalidArgumentException("RefWorkloadFactory: CreateMemImport() expected an input tensor.");
439 }
440 return std::make_unique<ImportMemGenericWorkload>(*memImportQueueDescriptor, info);
441 }
442 case LayerType::Minimum:
443 {
444 auto minimumQueueDescriptor = PolymorphicDowncast<const MinimumQueueDescriptor*>(&descriptor);
445 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
446 {
447 return std::make_unique<RefMinimumWorkload<int32_t>>(*minimumQueueDescriptor, info);
448 }
449 else
450 {
451 return std::make_unique<RefMinimumWorkload<float>>(*minimumQueueDescriptor, info);
452 }
453 }
454 case LayerType::Multiplication:
455 {
456 auto multiplicationQueueDescriptor
457 = PolymorphicDowncast<const MultiplicationQueueDescriptor*>(&descriptor);
458 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
459 {
460 return std::make_unique<RefMultiplicationWorkload<int32_t>>(*multiplicationQueueDescriptor, info);
461 }
462 else
463 {
464 return std::make_unique<RefMultiplicationWorkload<float>>(*multiplicationQueueDescriptor, info);
465 }
466 }
467 case LayerType::Normalization:
468 {
469 auto normalizationQueueDescriptor = PolymorphicDowncast<const NormalizationQueueDescriptor*>(&descriptor);
470 return std::make_unique<RefNormalizationWorkload>(*normalizationQueueDescriptor, info);
471 }
472 case LayerType::Output:
473 {
474 auto outputQueueDescriptor = PolymorphicDowncast<const OutputQueueDescriptor*>(&descriptor);
475 if (info.m_InputTensorInfos.empty() )
476 {
477 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Input cannot be zero length");
478 }
479 if (info.m_OutputTensorInfos.empty())
480 {
481 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Output cannot be zero length");
482 }
483 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
484 {
485 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: data input and output "
486 "differ in byte count.");
487 }
488
489 return std::make_unique<CopyMemGenericWorkload>(*outputQueueDescriptor, info);
490 }
491 case LayerType::Pad:
492 {
493 auto padQueueDescriptor = PolymorphicDowncast<const PadQueueDescriptor*>(&descriptor);
494 return std::make_unique<RefPadWorkload>(*padQueueDescriptor, info);
495 }
496 case LayerType::Permute:
497 {
498 auto permuteQueueDescriptor = PolymorphicDowncast<const PermuteQueueDescriptor*>(&descriptor);
499 if (IsQSymmS16(info))
500 {
501 return std::make_unique<RefPermuteQSymm16Workload>(*permuteQueueDescriptor, info);
502 }
503 else if (IsBFloat16(info))
504 {
505 return std::make_unique<RefPermuteBFloat16Workload>(*permuteQueueDescriptor, info);
506 }
507 else if (IsQAsymmS8(info))
508 {
509 return std::make_unique<RefPermuteQAsymmS8Workload>(*permuteQueueDescriptor, info);
510 }
511 return MakeWorkloadHelper<RefPermuteFloat16Workload, RefPermuteFloat32Workload, RefPermuteQAsymm8Workload,
512 NullWorkload, NullWorkload, NullWorkload>(*permuteQueueDescriptor, info);
513 }
514 case LayerType::Pooling2d:
515 {
516 auto pooling2dQueueDescriptor = PolymorphicDowncast<const Pooling2dQueueDescriptor*>(&descriptor);
517 return std::make_unique<RefPooling2dWorkload>(*pooling2dQueueDescriptor, info);
518 }
519 case LayerType::Pooling3d:
520 {
521 auto pooling3dQueueDescriptor = PolymorphicDowncast<const Pooling3dQueueDescriptor*>(&descriptor);
522 return std::make_unique<RefPooling3dWorkload>(*pooling3dQueueDescriptor, info);
523 }
524 case LayerType::PreCompiled:
525 {
526 return nullptr;
527 }
528 case LayerType::Prelu:
529 {
530 auto preluQueueDescriptor = PolymorphicDowncast<const PreluQueueDescriptor*>(&descriptor);
531 return std::make_unique<RefPreluWorkload>(*preluQueueDescriptor, info);
532 }
533 case LayerType::QLstm:
534 {
535 auto qlstmQueueDescriptor = PolymorphicDowncast<const QLstmQueueDescriptor*>(&descriptor);
536 return std::make_unique<RefQLstmWorkload>(*qlstmQueueDescriptor, info);
537 }
538 case LayerType::Quantize:
539 {
540 auto quantizeQueueDescriptor = PolymorphicDowncast<const QuantizeQueueDescriptor*>(&descriptor);
541 return std::make_unique<RefQuantizeWorkload>(*quantizeQueueDescriptor, info);
542 }
543 case LayerType::Rank:
544 {
545 auto rankQueueDescriptor = PolymorphicDowncast<const RankQueueDescriptor*>(&descriptor);
546 return std::make_unique<RefRankWorkload>(*rankQueueDescriptor, info);
547 }
548 case LayerType::Reduce:
549 {
550 auto reduceQueueDescriptor = PolymorphicDowncast<const ReduceQueueDescriptor*>(&descriptor);
551 return std::make_unique<RefReduceWorkload>(*reduceQueueDescriptor, info);
552 }
553 case LayerType::Reshape:
554 {
555 auto reshapeQueueDescriptor = PolymorphicDowncast<const ReshapeQueueDescriptor*>(&descriptor);
556 return std::make_unique<RefReshapeWorkload>(*reshapeQueueDescriptor, info);
557 }
558 case LayerType::Resize:
559 {
560 auto resizeQueueDescriptor = PolymorphicDowncast<const ResizeQueueDescriptor*>(&descriptor);
561 return std::make_unique<RefResizeWorkload>(*resizeQueueDescriptor, info);
562 }
Tianle Cheng988354d2023-06-28 13:20:47 +0100563 case LayerType::ReverseV2:
564 {
565 auto reverseV2QueueDescriptor = PolymorphicDowncast<const ReverseV2QueueDescriptor*>(&descriptor);
566 return std::make_unique<RefReverseV2Workload>(*reverseV2QueueDescriptor, info);
567 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000568 case LayerType::Shape:
569 {
570 auto shapeQueueDescriptor = PolymorphicDowncast<const ShapeQueueDescriptor*>(&descriptor);
571 return std::make_unique<RefShapeWorkload>(*shapeQueueDescriptor, info);
572 }
573 case LayerType::Slice:
574 {
575 auto sliceQueueDescriptor = PolymorphicDowncast<const SliceQueueDescriptor*>(&descriptor);
576 return std::make_unique<RefSliceWorkload>(*sliceQueueDescriptor, info);
577 }
578 case LayerType::Softmax:
579 {
580 auto softmaxQueueDescriptor = PolymorphicDowncast<const SoftmaxQueueDescriptor*>(&descriptor);
581 return std::make_unique<RefSoftmaxWorkload>(*softmaxQueueDescriptor, info);
582 }
583 case LayerType::SpaceToBatchNd:
584 {
585 auto spaceToBatchNdQueueDescriptor
586 = PolymorphicDowncast<const SpaceToBatchNdQueueDescriptor*>(&descriptor);
587 return std::make_unique<RefSpaceToBatchNdWorkload>(*spaceToBatchNdQueueDescriptor, info);
588 }
589 case LayerType::SpaceToDepth:
590 {
591 auto spaceToDepthQueueDescriptor = PolymorphicDowncast<const SpaceToDepthQueueDescriptor*>(&descriptor);
592 return std::make_unique<RefSpaceToDepthWorkload>(*spaceToDepthQueueDescriptor, info);
593 }
594 case LayerType::Splitter:
595 {
596 auto splitterQueueDescriptor = PolymorphicDowncast<const SplitterQueueDescriptor*>(&descriptor);
597 return std::make_unique<RefSplitterWorkload>(*splitterQueueDescriptor, info);
598 }
599 case LayerType::Stack:
600 {
601 auto stackQueueDescriptor = PolymorphicDowncast<const StackQueueDescriptor*>(&descriptor);
602 return std::make_unique<RefStackWorkload>(*stackQueueDescriptor, info);
603 }
604 case LayerType::StridedSlice:
605 {
606 auto stridedSliceQueueDescriptor = PolymorphicDowncast<const StridedSliceQueueDescriptor*>(&descriptor);
607 return std::make_unique<RefStridedSliceWorkload>(*stridedSliceQueueDescriptor, info);
608 }
609 case LayerType::Subtraction:
610 {
611 auto subtractionQueueDescriptor = PolymorphicDowncast<const SubtractionQueueDescriptor*>(&descriptor);
612 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
613 {
614 return std::make_unique<RefSubtractionWorkload<int32_t>>(*subtractionQueueDescriptor, info);
615 }
616 else
617 {
618 return std::make_unique<RefSubtractionWorkload<float>>(*subtractionQueueDescriptor, info);
619 }
620 }
621 case LayerType::Transpose:
622 {
623 auto transposeQueueDescriptor = PolymorphicDowncast<const TransposeQueueDescriptor*>(&descriptor);
624 if (IsQSymmS16(info))
625 {
626 return std::make_unique<RefTransposeQSymm16Workload>(*transposeQueueDescriptor, info);
627 }
628 else if (IsBFloat16(info))
629 {
630 return std::make_unique<RefTransposeBFloat16Workload>(*transposeQueueDescriptor, info);
631 }
632 else if (IsQAsymmS8(info))
633 {
634 return std::make_unique<RefTransposeQAsymmS8Workload>(*transposeQueueDescriptor, info);
635 }
636 return MakeWorkloadHelper<RefTransposeFloat16Workload, RefTransposeFloat32Workload,
637 RefTransposeQAsymm8Workload, NullWorkload, NullWorkload, NullWorkload>
638 (*transposeQueueDescriptor, info);
639 }
640 case LayerType::TransposeConvolution2d:
641 {
642 auto transposeConvolution2dQueueDescriptor
643 = PolymorphicDowncast<const TransposeConvolution2dQueueDescriptor*>(&descriptor);
644 return std::make_unique<RefTransposeConvolution2dWorkload>(*transposeConvolution2dQueueDescriptor, info);
645 }
646 case LayerType::UnidirectionalSequenceLstm:
647 {
648 auto unidirectionalSequenceLstmQueueDescriptor
649 = PolymorphicDowncast<const UnidirectionalSequenceLstmQueueDescriptor*>(&descriptor);
650 return std::make_unique<RefUnidirectionalSequenceLstmWorkload>(*unidirectionalSequenceLstmQueueDescriptor,
651 info);
652 }
653 default:
654 return nullptr;
655 }
656}
657
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100658std::unique_ptr<IWorkload> RefWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
659 const WorkloadInfo& info) const
660{
661 return std::make_unique<RefActivationWorkload>(descriptor, info);
662}
663
664std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
665 const WorkloadInfo& info) const
666{
Finn Williamscbd2c232020-06-22 15:58:32 +0100667 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
668 {
669 return std::make_unique<RefAdditionWorkload<int32_t>>(descriptor, info);
670 }
671 else
672 {
673 return std::make_unique<RefAdditionWorkload<float>>(descriptor, info);
674 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100675}
676
677std::unique_ptr<IWorkload> RefWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
678 const WorkloadInfo& info) const
679{
680 return std::make_unique<RefArgMinMaxWorkload>(descriptor, info);
681}
682
683std::unique_ptr<IWorkload> RefWorkloadFactory::CreateBatchNormalization(
684 const BatchNormalizationQueueDescriptor& descriptor,
685 const WorkloadInfo& info) const
686{
687 return std::make_unique<RefBatchNormalizationWorkload>(descriptor, info);
688}
689
690std::unique_ptr<IWorkload> RefWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
691 const WorkloadInfo& info) const
692{
693 return std::make_unique<RefBatchToSpaceNdWorkload>(descriptor, info);
694}
695
mathad01b392e982021-04-07 12:07:30 +0100696std::unique_ptr<IWorkload> RefWorkloadFactory::CreateCast(const CastQueueDescriptor& descriptor,
697 const WorkloadInfo& info) const
698{
699 return std::make_unique<RefCastWorkload>(descriptor, info);
700}
701
Simon Obute51f67772021-09-03 15:50:13 +0100702std::unique_ptr<IWorkload> RefWorkloadFactory::CreateChannelShuffle(const ChannelShuffleQueueDescriptor &descriptor,
703 const WorkloadInfo &info) const
704{
705 return std::make_unique<RefChannelShuffleWorkload>(descriptor,info);
706}
707
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100708std::unique_ptr<IWorkload> RefWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
709 const WorkloadInfo& info) const
710{
711 return std::make_unique<RefComparisonWorkload>(descriptor, info);
712}
713
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100714std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
715 const WorkloadInfo& info) const
716{
717 return std::make_unique<RefConcatWorkload>(descriptor, info);
718}
719
720std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
721 const WorkloadInfo& info) const
722{
723 return std::make_unique<RefConstantWorkload>(descriptor, info);
724}
725
726std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertFp16ToFp32(
727 const ConvertFp16ToFp32QueueDescriptor& descriptor,
728 const WorkloadInfo& info) const
729{
730 return std::make_unique<RefConvertFp16ToFp32Workload>(descriptor, info);
731}
732
733std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertFp32ToFp16(
734 const ConvertFp32ToFp16QueueDescriptor& descriptor,
735 const WorkloadInfo& info) const
736{
737 return std::make_unique<RefConvertFp32ToFp16Workload>(descriptor, info);
738}
739
740std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
741 const WorkloadInfo& info) const
742{
743 return std::make_unique<RefConvolution2dWorkload>(descriptor, info);
744}
745
Matthew Sloyanb63a3112021-09-08 13:05:51 +0100746std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvolution3d(const Convolution3dQueueDescriptor& descriptor,
747 const WorkloadInfo& info) const
748{
749 return std::make_unique<RefConvolution3dWorkload>(descriptor, info);
750}
751
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100752std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
753 const WorkloadInfo& info) const
754{
Narumol Prangnawarat403a1852020-03-12 14:24:13 +0000755 if (IsBFloat16(info))
756 {
757 return std::make_unique<RefDebugBFloat16Workload>(descriptor, info);
758 }
Aron Virginas-Tardb1a2832019-11-12 16:15:11 +0000759 if (IsFloat16(info))
760 {
761 return std::make_unique<RefDebugFloat16Workload>(descriptor, info);
762 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000763 if (IsQSymmS16(info))
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100764 {
Keith Davis0c2eeac2020-02-11 16:51:50 +0000765 return std::make_unique<RefDebugQSymmS16Workload>(descriptor, info);
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100766 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000767 if (IsQSymmS8(info))
Keith Davis5204aa82020-01-27 15:24:59 +0000768 {
Keith Davis0c2eeac2020-02-11 16:51:50 +0000769 return std::make_unique<RefDebugQSymmS8Workload>(descriptor, info);
Keith Davis5204aa82020-01-27 15:24:59 +0000770 }
Keith Davis67e6c542020-02-19 10:08:33 +0000771 if (IsQAsymmU8(info))
772 {
773 return std::make_unique<RefDebugQAsymmU8Workload>(descriptor, info);
774 }
775 if (IsQAsymmS8(info))
776 {
777 return std::make_unique<RefDebugQAsymmS8Workload>(descriptor, info);
778 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000779 if (IsSigned32(info))
Narumol Prangnawaratd2d917d2020-01-09 10:16:39 +0000780 {
781 return std::make_unique<RefDebugSigned32Workload>(descriptor, info);
782 }
Aron Virginas-Tardb1a2832019-11-12 16:15:11 +0000783
Keith Davis0c2eeac2020-02-11 16:51:50 +0000784 return MakeWorkload<RefDebugFloat32Workload, RefDebugQAsymmU8Workload>(descriptor, info);
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100785}
786
787std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
788 const WorkloadInfo& info) const
789{
790 return std::make_unique<RefDepthToSpaceWorkload>(descriptor, info);
791}
792
793std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDepthwiseConvolution2d(
794 const DepthwiseConvolution2dQueueDescriptor& descriptor,
795 const WorkloadInfo& info) const
796{
797 return std::make_unique<RefDepthwiseConvolution2dWorkload>(descriptor, info);
798}
799
800std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
801 const WorkloadInfo& info) const
802{
803 return std::make_unique<RefDequantizeWorkload>(descriptor, info);
804}
805
806std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDetectionPostProcess(
807 const DetectionPostProcessQueueDescriptor& descriptor,
808 const WorkloadInfo& info) const
809{
810 return std::make_unique<RefDetectionPostProcessWorkload>(descriptor, info);
811}
812
813std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
814 const WorkloadInfo& info) const
815{
Finn Williamscbd2c232020-06-22 15:58:32 +0100816 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
817 {
818 return std::make_unique<RefDivisionWorkload<int32_t>>(descriptor, info);
819 }
820 else
821 {
822 return std::make_unique<RefDivisionWorkload<float>>(descriptor, info);
823 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100824}
825
josh minor4a3c6102020-01-06 16:40:46 -0600826std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
827 const WorkloadInfo& info) const
828{
Narumol Prangnawarat0c95f4c2020-11-18 16:52:07 +0000829 if (descriptor.m_Parameters.m_Operation == UnaryOperation::LogicalNot)
830 {
831 return std::make_unique<RefLogicalUnaryWorkload>(descriptor, info);
832 }
josh minor4a3c6102020-01-06 16:40:46 -0600833 return std::make_unique<RefElementwiseUnaryWorkload>(descriptor, info);
834}
835
Ryan OSheaf4bfa6a2020-06-10 11:33:37 +0100836std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor,
837 const WorkloadInfo& info) const
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100838{
839 return MakeWorkload<RefFakeQuantizationFloat32Workload, NullWorkload>(descriptor, info);
840}
841
Ryan OSheaf4bfa6a2020-06-10 11:33:37 +0100842std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
843 const WorkloadInfo& info) const
844{
845 return std::make_unique<RefFillWorkload>(descriptor, info);
846}
847
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100848std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
849 const WorkloadInfo& info) const
850{
Francis Murtaghe8ac1332020-07-30 18:03:40 +0100851 if(IsQuantizedType(info.m_InputTensorInfos[0].GetDataType()))
852 {
853 return nullptr;
854 }
855 else
856 {
857 return std::make_unique<RefFloorWorkload>(descriptor, info);
858 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100859}
860
861std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFullyConnected(
862 const FullyConnectedQueueDescriptor& descriptor,
863 const WorkloadInfo& info) const
864{
865 return std::make_unique<RefFullyConnectedWorkload>(descriptor, info);
866}
867
868std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
869 const WorkloadInfo& info) const
870{
871 return std::make_unique<RefGatherWorkload>(descriptor, info);
872}
873
telsoa014fcda012018-03-09 14:13:49 +0000874std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
875 const WorkloadInfo& info) const
876{
877 if (info.m_InputTensorInfos.empty() )
878 {
879 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Input cannot be zero length");
880 }
881 if (info.m_OutputTensorInfos.empty())
882 {
883 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Output cannot be zero length");
884 }
885
886 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
887 {
888 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: data input and output differ in byte count.");
889 }
890
Narumol Prangnawaratb6441e42019-06-04 11:22:00 +0100891 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000892}
893
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100894std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInstanceNormalization(
895 const InstanceNormalizationQueueDescriptor& descriptor,
896 const WorkloadInfo& info) const
897{
898 return std::make_unique<RefInstanceNormalizationWorkload>(descriptor, info);
899}
900
901std::unique_ptr<IWorkload> RefWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
902 const WorkloadInfo& info) const
903{
904 return std::make_unique<RefL2NormalizationWorkload>(descriptor, info);
905}
906
James Conroyaba90cd2020-11-06 16:28:18 +0000907std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
908 const WorkloadInfo& info) const
909{
910 return std::make_unique<RefLogicalBinaryWorkload>(descriptor, info);
911}
912
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100913std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
914 const WorkloadInfo& info) const
915{
916 return std::make_unique<RefLogSoftmaxWorkload>(descriptor, info);
917}
918
919std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
920 const WorkloadInfo& info) const
921{
922 return std::make_unique<RefLstmWorkload>(descriptor, info);
923}
924
925std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
926 const WorkloadInfo& info) const
927{
Finn Williamscbd2c232020-06-22 15:58:32 +0100928 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
929 {
930 return std::make_unique<RefMaximumWorkload<int32_t>>(descriptor, info);
931 }
932 else
933 {
934 return std::make_unique<RefMaximumWorkload<float>>(descriptor, info);
935 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100936}
937
938std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
939 const WorkloadInfo& info) const
940{
941 return std::make_unique<RefMeanWorkload>(descriptor, info);
942}
943
944std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
945 const WorkloadInfo& info) const
946{
947 if (descriptor.m_Inputs.empty())
948 {
949 throw InvalidArgumentException("RefWorkloadFactory: CreateMemCopy() expected an input tensor.");
950 }
951 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
952}
953
954std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
955 const WorkloadInfo& info) const
956{
957 if (descriptor.m_Inputs.empty())
958 {
959 throw InvalidArgumentException("RefWorkloadFactory: CreateMemImport() expected an input tensor.");
960 }
961 return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
962}
963
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100964std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
965 const WorkloadInfo& info) const
966{
Finn Williamscbd2c232020-06-22 15:58:32 +0100967 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
968 {
969 return std::make_unique<RefMinimumWorkload<int32_t>>(descriptor, info);
970 }
971 else
972 {
973 return std::make_unique<RefMinimumWorkload<float>>(descriptor, info);
974 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100975}
976
977std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
978 const WorkloadInfo& info) const
979{
Finn Williamscbd2c232020-06-22 15:58:32 +0100980 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
981 {
982 return std::make_unique<RefMultiplicationWorkload<int32_t>>(descriptor, info);
983 }
984 else
985 {
986 return std::make_unique<RefMultiplicationWorkload<float>>(descriptor, info);
987 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100988}
989
990std::unique_ptr<IWorkload> RefWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
991 const WorkloadInfo& info) const
992{
993 return std::make_unique<RefNormalizationWorkload>(descriptor, info);
994}
995
telsoa014fcda012018-03-09 14:13:49 +0000996std::unique_ptr<IWorkload> RefWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
997 const WorkloadInfo& info) const
998{
999 if (info.m_InputTensorInfos.empty() )
1000 {
1001 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Input cannot be zero length");
1002 }
1003 if (info.m_OutputTensorInfos.empty())
1004 {
1005 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Output cannot be zero length");
1006 }
1007 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
1008 {
1009 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: data input and output differ in byte count.");
1010 }
1011
Narumol Prangnawaratb6441e42019-06-04 11:22:00 +01001012 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001013}
1014
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001015std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
1016 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001017{
Sadik Armagan041b3c02020-06-04 10:32:18 +01001018 return std::make_unique<RefPadWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001019}
1020
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001021std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
1022 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001023{
Keith Davis0c2eeac2020-02-11 16:51:50 +00001024 if (IsQSymmS16(info))
Narumol Prangnawarat86bb4e12019-07-08 11:36:05 +01001025 {
1026 return std::make_unique<RefPermuteQSymm16Workload>(descriptor, info);
1027 }
Narumol Prangnawarat44179c32020-03-11 14:51:27 +00001028 else if (IsBFloat16(info))
1029 {
1030 return std::make_unique<RefPermuteBFloat16Workload>(descriptor, info);
1031 }
Sadik Armagan303980c2020-04-17 12:45:14 +01001032 else if (IsQAsymmS8(info))
1033 {
1034 return std::make_unique<RefPermuteQAsymmS8Workload>(descriptor, info);
1035 }
Narumol Prangnawarat86bb4e12019-07-08 11:36:05 +01001036 return MakeWorkloadHelper<RefPermuteFloat16Workload, RefPermuteFloat32Workload, RefPermuteQAsymm8Workload,
Keith Davis5204aa82020-01-27 15:24:59 +00001037 NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001038}
1039
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001040std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
1041 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001042{
Teresa Charlina3b20472019-06-06 11:12:32 +01001043 return std::make_unique<RefPooling2dWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001044}
1045
Tamás Nyíri7b885b32021-10-26 14:47:57 +01001046std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePooling3d(const Pooling3dQueueDescriptor& descriptor,
1047 const WorkloadInfo& info) const
1048{
1049 return std::make_unique<RefPooling3dWorkload>(descriptor, info);
1050}
1051
Derek Lamberti901ea112019-12-10 22:07:09 +00001052std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/,
1053 const WorkloadInfo& /*info*/) const
telsoa014fcda012018-03-09 14:13:49 +00001054{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001055 return nullptr;
telsoa014fcda012018-03-09 14:13:49 +00001056}
1057
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001058std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePrelu(const PreluQueueDescriptor& descriptor,
1059 const WorkloadInfo& info) const
Aron Virginas-Tar73f66422019-09-23 19:11:59 +01001060{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001061 return std::make_unique<RefPreluWorkload>(descriptor, info);
Aron Virginas-Tar73f66422019-09-23 19:11:59 +01001062}
1063
James Conroy4f1f8992020-04-29 20:01:10 +01001064std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
1065 const WorkloadInfo& info) const
1066{
1067 return std::make_unique<RefQLstmWorkload>(descriptor, info);
1068}
1069
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001070std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
1071 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001072{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001073 return std::make_unique<RefQuantizeWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001074}
1075
Finn Williams2605b232020-06-10 15:53:46 +01001076std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRank(const RankQueueDescriptor& descriptor,
1077 const WorkloadInfo& info) const
1078{
1079 return std::make_unique<RefRankWorkload>(descriptor, info);
1080}
1081
Sadik Armagan0c3ea5b2021-02-03 09:29:30 +00001082std::unique_ptr<IWorkload> RefWorkloadFactory::CreateReduce(const ReduceQueueDescriptor& descriptor,
1083 const WorkloadInfo& info) const
1084{
1085 return std::make_unique<RefReduceWorkload>(descriptor, info);
1086}
1087
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001088std::unique_ptr<IWorkload> RefWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
1089 const WorkloadInfo& info) const
Narumol Prangnawarat94dd5d82019-01-23 18:06:26 +00001090{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001091 return std::make_unique<RefReshapeWorkload>(descriptor, info);
Derek Lambertif674aa02019-08-01 15:56:25 +01001092}
1093
Teresa Charlin970f43b2019-07-01 13:51:07 +01001094std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
1095 const WorkloadInfo& info) const
1096{
Teresa Charlin970f43b2019-07-01 13:51:07 +01001097 return std::make_unique<RefResizeWorkload>(descriptor, info);
1098}
1099
Keith Davis3ae3f972021-05-21 16:33:48 +01001100std::unique_ptr<IWorkload> RefWorkloadFactory::CreateShape(const ShapeQueueDescriptor& descriptor,
1101 const WorkloadInfo& info) const
1102{
1103 return std::make_unique<RefShapeWorkload>(descriptor, info);
1104}
1105
Aron Virginas-Tar92b9f872019-09-17 17:27:04 +01001106std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
1107 const WorkloadInfo& info) const
1108{
1109 return std::make_unique<RefSliceWorkload>(descriptor, info);
1110}
1111
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001112std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
1113 const WorkloadInfo& info) const
Kevin May09ca49c2019-10-09 12:37:34 +01001114{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001115 return std::make_unique<RefSoftmaxWorkload>(descriptor, info);
1116}
1117
1118std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
1119 const WorkloadInfo& info) const
1120{
1121 return std::make_unique<RefSpaceToBatchNdWorkload>(descriptor, info);
1122}
1123
1124std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
1125 const WorkloadInfo& info) const
1126{
1127 return std::make_unique<RefSpaceToDepthWorkload>(descriptor, info);
1128}
1129
1130std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
1131 const WorkloadInfo& info) const
1132{
1133 return std::make_unique<RefSplitterWorkload>(descriptor, info);
1134}
1135
1136std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
1137 const WorkloadInfo& info) const
1138{
1139 return std::make_unique<RefStackWorkload>(descriptor, info);
1140}
1141
1142std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
1143 const WorkloadInfo& info) const
1144{
1145 return std::make_unique<RefStridedSliceWorkload>(descriptor, info);
1146}
1147
1148std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
1149 const WorkloadInfo& info) const
1150{
Finn Williamscbd2c232020-06-22 15:58:32 +01001151 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
1152 {
1153 return std::make_unique<RefSubtractionWorkload<int32_t>>(descriptor, info);
1154 }
1155 else
1156 {
1157 return std::make_unique<RefSubtractionWorkload<float>>(descriptor, info);
1158 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001159}
1160
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001161std::unique_ptr<IWorkload> RefWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
1162 const WorkloadInfo& info) const
1163{
1164 if (IsQSymmS16(info))
1165 {
1166 return std::make_unique<RefTransposeQSymm16Workload>(descriptor, info);
1167 }
Narumol Prangnawarat44179c32020-03-11 14:51:27 +00001168 else if (IsBFloat16(info))
1169 {
1170 return std::make_unique<RefTransposeBFloat16Workload>(descriptor, info);
1171 }
Sadik Armagan303980c2020-04-17 12:45:14 +01001172 else if (IsQAsymmS8(info))
1173 {
1174 return std::make_unique<RefTransposeQAsymmS8Workload>(descriptor, info);
1175 }
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001176 return MakeWorkloadHelper<RefTransposeFloat16Workload, RefTransposeFloat32Workload, RefTransposeQAsymm8Workload,
1177 NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
1178}
1179
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001180std::unique_ptr<IWorkload> RefWorkloadFactory::CreateTransposeConvolution2d(
1181 const TransposeConvolution2dQueueDescriptor& descriptor,
1182 const WorkloadInfo& info) const
1183{
1184 return std::make_unique<RefTransposeConvolution2dWorkload>(descriptor, info);
Kevin May09ca49c2019-10-09 12:37:34 +01001185}
1186
Narumol Prangnawarate5339e72021-07-28 17:33:28 +01001187std::unique_ptr<IWorkload> RefWorkloadFactory::CreateUnidirectionalSequenceLstm(
1188 const UnidirectionalSequenceLstmQueueDescriptor& descriptor,
1189 const WorkloadInfo& info) const
1190{
1191 return std::make_unique<RefUnidirectionalSequenceLstmWorkload>(descriptor, info);;
1192}
1193
Matteo Martincigh49124022019-01-11 13:25:59 +00001194} // namespace armnn