blob: 093d0d5e20af74740e82884469bdeebaac33dce6 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Teresa Charlin611c7fb2022-01-07 09:47:29 +00002// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
Matthew Bentham4cefc412019-06-18 16:14:34 +01005#include <Layer.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +00006#include <armnn/backends/MemCopyWorkload.hpp>
Derek Lambertif674aa02019-08-01 15:56:25 +01007#include <backendsCommon/MemImportWorkload.hpp>
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +00008#include <backendsCommon/MakeWorkloadHelper.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +00009#include <armnn/backends/TensorHandle.hpp>
telsoa014fcda012018-03-09 14:13:49 +000010#include "RefWorkloadFactory.hpp"
David Beck79141b92018-10-23 16:09:36 +010011#include "RefBackendId.hpp"
David Beckb4540be2018-09-24 13:18:27 +010012#include "workloads/RefWorkloads.hpp"
Matthew Bentham4cefc412019-06-18 16:14:34 +010013#include "RefTensorHandle.hpp"
telsoa014fcda012018-03-09 14:13:49 +000014
telsoa014fcda012018-03-09 14:13:49 +000015
16namespace armnn
17{
18
David Beck79141b92018-10-23 16:09:36 +010019namespace
20{
21static const BackendId s_Id{RefBackendId()};
22}
telsoa014fcda012018-03-09 14:13:49 +000023template <typename F32Workload, typename U8Workload, typename QueueDescriptorType>
24std::unique_ptr<IWorkload> RefWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
Aron Virginas-Tare662a942019-10-14 15:12:00 +010025 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +000026{
Keith Davis5204aa82020-01-27 15:24:59 +000027 return MakeWorkloadHelper<NullWorkload, F32Workload, U8Workload, NullWorkload, NullWorkload, NullWorkload>
28 (descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +000029}
30
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010031template <DataType ArmnnType>
32bool IsDataType(const WorkloadInfo& info)
Jim Flynn82fbe7c2019-04-02 15:19:08 +010033{
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010034 auto checkType = [](const TensorInfo& tensorInfo) {return tensorInfo.GetDataType() == ArmnnType;};
35 auto it = std::find_if(std::begin(info.m_InputTensorInfos), std::end(info.m_InputTensorInfos), checkType);
Jim Flynn82fbe7c2019-04-02 15:19:08 +010036 if (it != std::end(info.m_InputTensorInfos))
37 {
38 return true;
39 }
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010040 it = std::find_if(std::begin(info.m_OutputTensorInfos), std::end(info.m_OutputTensorInfos), checkType);
Jim Flynn82fbe7c2019-04-02 15:19:08 +010041 if (it != std::end(info.m_OutputTensorInfos))
42 {
43 return true;
44 }
45 return false;
46}
47
Keith Davis0c2eeac2020-02-11 16:51:50 +000048bool IsSigned32(const WorkloadInfo& info)
49{
50 return IsDataType<DataType::Signed32>(info);
51}
52
Narumol Prangnawarat44179c32020-03-11 14:51:27 +000053bool IsBFloat16(const WorkloadInfo& info)
54{
55 return IsDataType<DataType::BFloat16>(info);
56}
57
Ferran Balaguerd73d14f2019-06-10 10:29:54 +010058bool IsFloat16(const WorkloadInfo& info)
59{
60 return IsDataType<DataType::Float16>(info);
61}
62
Keith Davis0c2eeac2020-02-11 16:51:50 +000063bool IsQSymmS16(const WorkloadInfo& info)
nikraj0199a66312019-06-06 10:31:27 +010064{
Derek Lambertif90c56d2020-01-10 17:14:08 +000065 return IsDataType<DataType::QSymmS16>(info);
nikraj0199a66312019-06-06 10:31:27 +010066}
67
Keith Davis0c2eeac2020-02-11 16:51:50 +000068bool IsQSymmS8(const WorkloadInfo& info)
Keith Davis5204aa82020-01-27 15:24:59 +000069{
70 return IsDataType<DataType::QSymmS8>(info);
71}
72
Keith Davis67e6c542020-02-19 10:08:33 +000073bool IsQAsymmS8(const WorkloadInfo& info)
74{
75 return IsDataType<DataType::QAsymmS8>(info);
76}
77
78bool IsQAsymmU8(const WorkloadInfo& info)
79{
80 return IsDataType<DataType::QAsymmU8>(info);
81}
82
Matthew Bentham7c1603a2019-06-21 17:22:23 +010083RefWorkloadFactory::RefWorkloadFactory(const std::shared_ptr<RefMemoryManager>& memoryManager)
84 : m_MemoryManager(memoryManager)
85{
86}
87
telsoa01c577f2c2018-08-31 09:22:23 +010088RefWorkloadFactory::RefWorkloadFactory()
Matthew Bentham7c1603a2019-06-21 17:22:23 +010089 : m_MemoryManager(new RefMemoryManager())
telsoa014fcda012018-03-09 14:13:49 +000090{
91}
92
David Beck79141b92018-10-23 16:09:36 +010093const BackendId& RefWorkloadFactory::GetBackendId() const
94{
95 return s_Id;
96}
97
David Beck29c75de2018-10-23 13:35:58 +010098bool RefWorkloadFactory::IsLayerSupported(const Layer& layer,
99 Optional<DataType> dataType,
telsoa01c577f2c2018-08-31 09:22:23 +0100100 std::string& outReasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +0000101{
David Beck79141b92018-10-23 16:09:36 +0100102 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
telsoa014fcda012018-03-09 14:13:49 +0000103}
104
Sadik Armagan04a72972020-09-14 15:44:18 +0100105bool RefWorkloadFactory::IsLayerSupported(const IConnectableLayer& layer,
106 Optional<DataType> dataType,
107 std::string& outReasonIfUnsupported,
108 const ModelOptions& modelOptions)
109{
110 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
111}
112
David Monahan3fb7e102019-08-20 11:25:29 +0100113std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
Derek Lamberti901ea112019-12-10 22:07:09 +0000114 const bool isMemoryManaged) const
telsoa014fcda012018-03-09 14:13:49 +0000115{
Finn Williamsb1aad422021-10-28 19:07:32 +0100116 if (isMemoryManaged)
117 {
118 return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
119 }
120 else
121 {
122 return std::make_unique<RefTensorHandle>(tensorInfo, static_cast<unsigned int>(MemorySource::Malloc));
123 }
telsoa014fcda012018-03-09 14:13:49 +0000124}
125
Francis Murtagh351d13d2018-09-24 15:01:18 +0100126std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
David Monahan3fb7e102019-08-20 11:25:29 +0100127 DataLayout dataLayout,
Derek Lamberti901ea112019-12-10 22:07:09 +0000128 const bool isMemoryManaged) const
Francis Murtagh351d13d2018-09-24 15:01:18 +0100129{
David Monahan3fb7e102019-08-20 11:25:29 +0100130 // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
131 // to unmanaged memory. This also ensures memory alignment.
Jan Eilers8eb25602020-03-09 12:13:48 +0000132 IgnoreUnused(isMemoryManaged, dataLayout);
Finn Williamsb1aad422021-10-28 19:07:32 +0100133
134 if (isMemoryManaged)
135 {
136 return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
137 }
138 else
139 {
140 return std::make_unique<RefTensorHandle>(tensorInfo, static_cast<unsigned int>(MemorySource::Malloc));
141 }
Francis Murtagh351d13d2018-09-24 15:01:18 +0100142}
143
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000144std::unique_ptr<IWorkload> RefWorkloadFactory::CreateWorkload(LayerType type,
145 const QueueDescriptor& descriptor,
146 const WorkloadInfo& info) const
147{
148 switch(type)
149 {
150 case LayerType::Activation :
151 {
152 auto activationQueueDescriptor = PolymorphicDowncast<const ActivationQueueDescriptor*>(&descriptor);
153 return std::make_unique<RefActivationWorkload>(*activationQueueDescriptor, info);
154 }
155 case LayerType::Addition :
156 {
157 auto additionQueueDescriptor = PolymorphicDowncast<const AdditionQueueDescriptor*>(&descriptor);
158
159 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
160 {
161 return std::make_unique<RefAdditionWorkload<int32_t>>(*additionQueueDescriptor, info);
162 }
163 else
164 {
165 return std::make_unique<RefAdditionWorkload<float>>(*additionQueueDescriptor, info);
166 }
167 }
168 case LayerType::ArgMinMax :
169 {
170 auto argMinMaxQueueDescriptor = PolymorphicDowncast<const ArgMinMaxQueueDescriptor*>(&descriptor);
171 return std::make_unique<RefArgMinMaxWorkload>(*argMinMaxQueueDescriptor, info);
172 }
Samuel Yap6b478092022-07-06 15:36:03 +0100173 case LayerType::BatchMatMul:
174 {
175 auto batchMatMulQueueDescriptor = PolymorphicDowncast<const BatchMatMulQueueDescriptor*>(&descriptor);
176 return std::make_unique<RefBatchMatMulWorkload>(*batchMatMulQueueDescriptor, info);
177 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000178 case LayerType::BatchNormalization :
179 {
180 auto batchNormQueueDescriptor = PolymorphicDowncast<const BatchNormalizationQueueDescriptor*>(&descriptor);
181 return std::make_unique<RefBatchNormalizationWorkload>(*batchNormQueueDescriptor, info);
182 }
183 case LayerType::BatchToSpaceNd :
184 {
185 auto batchToSpaceNdQueueDescriptor
186 = PolymorphicDowncast<const BatchToSpaceNdQueueDescriptor*>(&descriptor);
187 return std::make_unique<RefBatchToSpaceNdWorkload>(*batchToSpaceNdQueueDescriptor, info);
188 }
189 case LayerType::Cast :
190 {
191 auto castQueueDescriptor = PolymorphicDowncast<const CastQueueDescriptor*>(&descriptor);
192 return std::make_unique<RefCastWorkload>(*castQueueDescriptor, info);
193 }
194 case LayerType::ChannelShuffle :
195 {
196 auto channelShuffleQueueDescriptor
197 = PolymorphicDowncast<const ChannelShuffleQueueDescriptor*>(&descriptor);
198 return std::make_unique<RefChannelShuffleWorkload>(*channelShuffleQueueDescriptor, info);
199 }
200 case LayerType::Comparison :
201 {
202 auto comparisonQueueDescriptor = PolymorphicDowncast<const ComparisonQueueDescriptor*>(&descriptor);
203 return std::make_unique<RefComparisonWorkload>(*comparisonQueueDescriptor, info);
204 }
205 case LayerType::Concat :
206 {
207 auto concatQueueDescriptor = PolymorphicDowncast<const ConcatQueueDescriptor*>(&descriptor);
208 return std::make_unique<RefConcatWorkload>(*concatQueueDescriptor, info);
209 }
210 case LayerType::Constant :
211 {
212 auto constantQueueDescriptor = PolymorphicDowncast<const ConstantQueueDescriptor*>(&descriptor);
213 return std::make_unique<RefConstantWorkload>(*constantQueueDescriptor, info);
214 }
215 case LayerType::ConvertBf16ToFp32 :
216 {
217 auto convertBf16ToFp32QueueDescriptor
218 = PolymorphicDowncast<const ConvertBf16ToFp32QueueDescriptor*>(&descriptor);
219 return std::make_unique<RefConvertBf16ToFp32Workload>(*convertBf16ToFp32QueueDescriptor, info);
220 }
221 case LayerType::ConvertFp16ToFp32:
222 {
223 auto convertFp16ToFp32QueueDescriptor
224 = PolymorphicDowncast<const ConvertFp16ToFp32QueueDescriptor*>(&descriptor);
225 return std::make_unique<RefConvertFp16ToFp32Workload>(*convertFp16ToFp32QueueDescriptor, info);
226 }
227 case LayerType::ConvertFp32ToBf16:
228 {
229 auto convertFp32ToBf16QueueDescriptor
230 = PolymorphicDowncast<const ConvertFp32ToBf16QueueDescriptor*>(&descriptor);
231 return std::make_unique<RefConvertFp32ToBf16Workload>(*convertFp32ToBf16QueueDescriptor, info);
232 }
233 case LayerType::ConvertFp32ToFp16:
234 {
235 auto convertFp32ToFp16QueueDescriptor
236 = PolymorphicDowncast<const ConvertFp32ToFp16QueueDescriptor*>(&descriptor);
237 return std::make_unique<RefConvertFp32ToFp16Workload>(*convertFp32ToFp16QueueDescriptor, info);
238 }
239 case LayerType::Convolution2d:
240 {
241 auto convolution2dQueueDescriptor = PolymorphicDowncast<const Convolution2dQueueDescriptor*>(&descriptor);
242 return std::make_unique<RefConvolution2dWorkload>(*convolution2dQueueDescriptor, info);
243 }
244 case LayerType::Convolution3d:
245 {
246 auto convolution3dQueueDescriptor = PolymorphicDowncast<const Convolution3dQueueDescriptor*>(&descriptor);
247 return std::make_unique<RefConvolution3dWorkload>(*convolution3dQueueDescriptor, info);
248 }
249 case LayerType::Debug:
250 {
251 auto debugQueueDescriptor = PolymorphicDowncast<const DebugQueueDescriptor*>(&descriptor);
252 if (IsBFloat16(info))
253 {
254 return std::make_unique<RefDebugBFloat16Workload>(*debugQueueDescriptor, info);
255 }
256 if (IsFloat16(info))
257 {
258 return std::make_unique<RefDebugFloat16Workload>(*debugQueueDescriptor, info);
259 }
260 if (IsQSymmS16(info))
261 {
262 return std::make_unique<RefDebugQSymmS16Workload>(*debugQueueDescriptor, info);
263 }
264 if (IsQSymmS8(info))
265 {
266 return std::make_unique<RefDebugQSymmS8Workload>(*debugQueueDescriptor, info);
267 }
268 if (IsQAsymmU8(info))
269 {
270 return std::make_unique<RefDebugQAsymmU8Workload>(*debugQueueDescriptor, info);
271 }
272 if (IsQAsymmS8(info))
273 {
274 return std::make_unique<RefDebugQAsymmS8Workload>(*debugQueueDescriptor, info);
275 }
276 if (IsSigned32(info))
277 {
278 return std::make_unique<RefDebugSigned32Workload>(*debugQueueDescriptor, info);
279 }
280
281 return MakeWorkload<RefDebugFloat32Workload, RefDebugQAsymmU8Workload>(*debugQueueDescriptor, info);
282 }
283 case LayerType::DepthToSpace:
284 {
285 auto depthToSpaceQueueDescriptor = PolymorphicDowncast<const DepthToSpaceQueueDescriptor*>(&descriptor);
286 return std::make_unique<RefDepthToSpaceWorkload>(*depthToSpaceQueueDescriptor, info);
287 }
288 case LayerType::DepthwiseConvolution2d:
289 {
290 auto depthwiseConvolution2DQueueDescriptor
291 = PolymorphicDowncast<const DepthwiseConvolution2dQueueDescriptor*>(&descriptor);
292 return std::make_unique<RefDepthwiseConvolution2dWorkload>(*depthwiseConvolution2DQueueDescriptor, info);
293 }
294 case LayerType::Dequantize:
295 {
296 auto dequantizeQueueDescriptor = PolymorphicDowncast<const DequantizeQueueDescriptor*>(&descriptor);
297 return std::make_unique<RefDequantizeWorkload>(*dequantizeQueueDescriptor, info);
298 }
299 case LayerType::DetectionPostProcess:
300 {
301 auto detectionPostProcessQueueDescriptor
302 = PolymorphicDowncast<const DetectionPostProcessQueueDescriptor*>(&descriptor);
303 return std::make_unique<RefDetectionPostProcessWorkload>(*detectionPostProcessQueueDescriptor, info);
304 }
305 case LayerType::Division:
306 {
307 auto divisionQueueDescriptor = PolymorphicDowncast<const DivisionQueueDescriptor*>(&descriptor);
308 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
309 {
310 return std::make_unique<RefDivisionWorkload<int32_t>>(*divisionQueueDescriptor, info);
311 }
312 else
313 {
314 return std::make_unique<RefDivisionWorkload<float>>(*divisionQueueDescriptor, info);
315 }
316 }
317 case LayerType::ElementwiseUnary:
318 {
319 auto elementwiseUnaryQueueDescriptor
320 = PolymorphicDowncast<const ElementwiseUnaryQueueDescriptor*>(&descriptor);
321 if ((*elementwiseUnaryQueueDescriptor).m_Parameters.m_Operation == UnaryOperation::LogicalNot)
322 {
323 return std::make_unique<RefLogicalUnaryWorkload>(*elementwiseUnaryQueueDescriptor, info);
324 }
325 return std::make_unique<RefElementwiseUnaryWorkload>(*elementwiseUnaryQueueDescriptor, info);
326 }
327 case LayerType::FakeQuantization:
328 {
329 auto fakeQuantizationQueueDescriptor
330 = PolymorphicDowncast<const FakeQuantizationQueueDescriptor*>(&descriptor);
331 return std::make_unique<RefFakeQuantizationFloat32Workload>(*fakeQuantizationQueueDescriptor, info);
332 }
333 case LayerType::Fill:
334 {
335 auto fillQueueDescriptor = PolymorphicDowncast<const FillQueueDescriptor*>(&descriptor);
336 return std::make_unique<RefFillWorkload>(*fillQueueDescriptor, info);
337 }
338 case LayerType::Floor:
339 {
340 auto floorQueueDescriptor = PolymorphicDowncast<const FloorQueueDescriptor*>(&descriptor);
341 if(IsQuantizedType(info.m_InputTensorInfos[0].GetDataType()))
342 {
343 return nullptr;
344 }
345 else
346 {
347 return std::make_unique<RefFloorWorkload>(*floorQueueDescriptor, info);
348 }
349 }
350 case LayerType::FullyConnected:
351 {
352 auto fullyConnectedQueueDescriptor
353 = PolymorphicDowncast<const FullyConnectedQueueDescriptor*>(&descriptor);
354 return std::make_unique<RefFullyConnectedWorkload>(*fullyConnectedQueueDescriptor, info);
355 }
356 case LayerType::Gather:
357 {
358 auto gatherQueueDescriptor = PolymorphicDowncast<const GatherQueueDescriptor*>(&descriptor);
359 return std::make_unique<RefGatherWorkload>(*gatherQueueDescriptor, info);
360 }
Teresa Charlinb2d3ec52022-04-12 22:07:09 +0100361 case LayerType::GatherNd:
362 {
363 auto gatherNdQueueDescriptor = PolymorphicDowncast<const GatherNdQueueDescriptor*>(&descriptor);
364 return std::make_unique<RefGatherNdWorkload>(*gatherNdQueueDescriptor, info);
365 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000366 case LayerType::Input:
367 {
368 auto inputQueueDescriptor = PolymorphicDowncast<const InputQueueDescriptor*>(&descriptor);
369 if (info.m_InputTensorInfos.empty() )
370 {
371 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Input cannot be zero length");
372 }
373 if (info.m_OutputTensorInfos.empty())
374 {
375 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Output cannot be zero length");
376 }
377
378 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
379 {
380 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: "
381 "data input and output differ in byte count.");
382 }
383
384 return std::make_unique<CopyMemGenericWorkload>(*inputQueueDescriptor, info);
385 }
386 case LayerType::InstanceNormalization:
387 {
388 auto instanceNormalizationQueueDescriptor
389 = PolymorphicDowncast<const InstanceNormalizationQueueDescriptor*>(&descriptor);
390 return std::make_unique<RefInstanceNormalizationWorkload>(*instanceNormalizationQueueDescriptor, info);
391 }
392 case LayerType::L2Normalization:
393 {
394 auto l2NormalizationQueueDescriptor
395 = PolymorphicDowncast<const L2NormalizationQueueDescriptor*>(&descriptor);
396 return std::make_unique<RefL2NormalizationWorkload>(*l2NormalizationQueueDescriptor, info);
397 }
398 case LayerType::LogicalBinary:
399 {
400 auto logicalBinaryQueueDescriptor = PolymorphicDowncast<const LogicalBinaryQueueDescriptor*>(&descriptor);
401 return std::make_unique<RefLogicalBinaryWorkload>(*logicalBinaryQueueDescriptor, info);
402 }
403 case LayerType::LogSoftmax:
404 {
405 auto logSoftmaxQueueDescriptor = PolymorphicDowncast<const LogSoftmaxQueueDescriptor*>(&descriptor);
406 return std::make_unique<RefLogSoftmaxWorkload>(*logSoftmaxQueueDescriptor, info);
407 }
408 case LayerType::Lstm:
409 {
410 auto lstmQueueDescriptor = PolymorphicDowncast<const LstmQueueDescriptor*>(&descriptor);
411 return std::make_unique<RefLstmWorkload>(*lstmQueueDescriptor, info);
412 }
413 case LayerType::Maximum:
414 {
415 auto maximumQueueDescriptor = PolymorphicDowncast<const MaximumQueueDescriptor*>(&descriptor);
416 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
417 {
418 return std::make_unique<RefMaximumWorkload<int32_t>>(*maximumQueueDescriptor, info);
419 }
420 else
421 {
422 return std::make_unique<RefMaximumWorkload<float>>(*maximumQueueDescriptor, info);
423 }
424 }
425 case LayerType::Mean:
426 {
427 auto meanQueueDescriptor = PolymorphicDowncast<const MeanQueueDescriptor*>(&descriptor);
428 return std::make_unique<RefMeanWorkload>(*meanQueueDescriptor, info);
429 }
430 case LayerType::MemCopy:
431 {
432 auto memCopyQueueDescriptor = PolymorphicDowncast<const MemCopyQueueDescriptor*>(&descriptor);
433 if (descriptor.m_Inputs.empty())
434 {
435 throw InvalidArgumentException("RefWorkloadFactory: CreateMemCopy() expected an input tensor.");
436 }
437 return std::make_unique<CopyMemGenericWorkload>(*memCopyQueueDescriptor, info);
438 }
439 case LayerType::MemImport:
440 {
441 auto memImportQueueDescriptor = PolymorphicDowncast<const MemImportQueueDescriptor*>(&descriptor);
442 if (descriptor.m_Inputs.empty())
443 {
444 throw InvalidArgumentException("RefWorkloadFactory: CreateMemImport() expected an input tensor.");
445 }
446 return std::make_unique<ImportMemGenericWorkload>(*memImportQueueDescriptor, info);
447 }
448 case LayerType::Minimum:
449 {
450 auto minimumQueueDescriptor = PolymorphicDowncast<const MinimumQueueDescriptor*>(&descriptor);
451 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
452 {
453 return std::make_unique<RefMinimumWorkload<int32_t>>(*minimumQueueDescriptor, info);
454 }
455 else
456 {
457 return std::make_unique<RefMinimumWorkload<float>>(*minimumQueueDescriptor, info);
458 }
459 }
460 case LayerType::Multiplication:
461 {
462 auto multiplicationQueueDescriptor
463 = PolymorphicDowncast<const MultiplicationQueueDescriptor*>(&descriptor);
464 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
465 {
466 return std::make_unique<RefMultiplicationWorkload<int32_t>>(*multiplicationQueueDescriptor, info);
467 }
468 else
469 {
470 return std::make_unique<RefMultiplicationWorkload<float>>(*multiplicationQueueDescriptor, info);
471 }
472 }
473 case LayerType::Normalization:
474 {
475 auto normalizationQueueDescriptor = PolymorphicDowncast<const NormalizationQueueDescriptor*>(&descriptor);
476 return std::make_unique<RefNormalizationWorkload>(*normalizationQueueDescriptor, info);
477 }
478 case LayerType::Output:
479 {
480 auto outputQueueDescriptor = PolymorphicDowncast<const OutputQueueDescriptor*>(&descriptor);
481 if (info.m_InputTensorInfos.empty() )
482 {
483 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Input cannot be zero length");
484 }
485 if (info.m_OutputTensorInfos.empty())
486 {
487 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Output cannot be zero length");
488 }
489 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
490 {
491 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: data input and output "
492 "differ in byte count.");
493 }
494
495 return std::make_unique<CopyMemGenericWorkload>(*outputQueueDescriptor, info);
496 }
497 case LayerType::Pad:
498 {
499 auto padQueueDescriptor = PolymorphicDowncast<const PadQueueDescriptor*>(&descriptor);
500 return std::make_unique<RefPadWorkload>(*padQueueDescriptor, info);
501 }
502 case LayerType::Permute:
503 {
504 auto permuteQueueDescriptor = PolymorphicDowncast<const PermuteQueueDescriptor*>(&descriptor);
505 if (IsQSymmS16(info))
506 {
507 return std::make_unique<RefPermuteQSymm16Workload>(*permuteQueueDescriptor, info);
508 }
509 else if (IsBFloat16(info))
510 {
511 return std::make_unique<RefPermuteBFloat16Workload>(*permuteQueueDescriptor, info);
512 }
513 else if (IsQAsymmS8(info))
514 {
515 return std::make_unique<RefPermuteQAsymmS8Workload>(*permuteQueueDescriptor, info);
516 }
517 return MakeWorkloadHelper<RefPermuteFloat16Workload, RefPermuteFloat32Workload, RefPermuteQAsymm8Workload,
518 NullWorkload, NullWorkload, NullWorkload>(*permuteQueueDescriptor, info);
519 }
520 case LayerType::Pooling2d:
521 {
522 auto pooling2dQueueDescriptor = PolymorphicDowncast<const Pooling2dQueueDescriptor*>(&descriptor);
523 return std::make_unique<RefPooling2dWorkload>(*pooling2dQueueDescriptor, info);
524 }
525 case LayerType::Pooling3d:
526 {
527 auto pooling3dQueueDescriptor = PolymorphicDowncast<const Pooling3dQueueDescriptor*>(&descriptor);
528 return std::make_unique<RefPooling3dWorkload>(*pooling3dQueueDescriptor, info);
529 }
530 case LayerType::PreCompiled:
531 {
532 return nullptr;
533 }
534 case LayerType::Prelu:
535 {
536 auto preluQueueDescriptor = PolymorphicDowncast<const PreluQueueDescriptor*>(&descriptor);
537 return std::make_unique<RefPreluWorkload>(*preluQueueDescriptor, info);
538 }
539 case LayerType::QLstm:
540 {
541 auto qlstmQueueDescriptor = PolymorphicDowncast<const QLstmQueueDescriptor*>(&descriptor);
542 return std::make_unique<RefQLstmWorkload>(*qlstmQueueDescriptor, info);
543 }
544 case LayerType::Quantize:
545 {
546 auto quantizeQueueDescriptor = PolymorphicDowncast<const QuantizeQueueDescriptor*>(&descriptor);
547 return std::make_unique<RefQuantizeWorkload>(*quantizeQueueDescriptor, info);
548 }
549 case LayerType::Rank:
550 {
551 auto rankQueueDescriptor = PolymorphicDowncast<const RankQueueDescriptor*>(&descriptor);
552 return std::make_unique<RefRankWorkload>(*rankQueueDescriptor, info);
553 }
554 case LayerType::Reduce:
555 {
556 auto reduceQueueDescriptor = PolymorphicDowncast<const ReduceQueueDescriptor*>(&descriptor);
557 return std::make_unique<RefReduceWorkload>(*reduceQueueDescriptor, info);
558 }
559 case LayerType::Reshape:
560 {
561 auto reshapeQueueDescriptor = PolymorphicDowncast<const ReshapeQueueDescriptor*>(&descriptor);
562 return std::make_unique<RefReshapeWorkload>(*reshapeQueueDescriptor, info);
563 }
564 case LayerType::Resize:
565 {
566 auto resizeQueueDescriptor = PolymorphicDowncast<const ResizeQueueDescriptor*>(&descriptor);
567 return std::make_unique<RefResizeWorkload>(*resizeQueueDescriptor, info);
568 }
569 case LayerType::Shape:
570 {
571 auto shapeQueueDescriptor = PolymorphicDowncast<const ShapeQueueDescriptor*>(&descriptor);
572 return std::make_unique<RefShapeWorkload>(*shapeQueueDescriptor, info);
573 }
574 case LayerType::Slice:
575 {
576 auto sliceQueueDescriptor = PolymorphicDowncast<const SliceQueueDescriptor*>(&descriptor);
577 return std::make_unique<RefSliceWorkload>(*sliceQueueDescriptor, info);
578 }
579 case LayerType::Softmax:
580 {
581 auto softmaxQueueDescriptor = PolymorphicDowncast<const SoftmaxQueueDescriptor*>(&descriptor);
582 return std::make_unique<RefSoftmaxWorkload>(*softmaxQueueDescriptor, info);
583 }
584 case LayerType::SpaceToBatchNd:
585 {
586 auto spaceToBatchNdQueueDescriptor
587 = PolymorphicDowncast<const SpaceToBatchNdQueueDescriptor*>(&descriptor);
588 return std::make_unique<RefSpaceToBatchNdWorkload>(*spaceToBatchNdQueueDescriptor, info);
589 }
590 case LayerType::SpaceToDepth:
591 {
592 auto spaceToDepthQueueDescriptor = PolymorphicDowncast<const SpaceToDepthQueueDescriptor*>(&descriptor);
593 return std::make_unique<RefSpaceToDepthWorkload>(*spaceToDepthQueueDescriptor, info);
594 }
595 case LayerType::Splitter:
596 {
597 auto splitterQueueDescriptor = PolymorphicDowncast<const SplitterQueueDescriptor*>(&descriptor);
598 return std::make_unique<RefSplitterWorkload>(*splitterQueueDescriptor, info);
599 }
600 case LayerType::Stack:
601 {
602 auto stackQueueDescriptor = PolymorphicDowncast<const StackQueueDescriptor*>(&descriptor);
603 return std::make_unique<RefStackWorkload>(*stackQueueDescriptor, info);
604 }
605 case LayerType::StridedSlice:
606 {
607 auto stridedSliceQueueDescriptor = PolymorphicDowncast<const StridedSliceQueueDescriptor*>(&descriptor);
608 return std::make_unique<RefStridedSliceWorkload>(*stridedSliceQueueDescriptor, info);
609 }
610 case LayerType::Subtraction:
611 {
612 auto subtractionQueueDescriptor = PolymorphicDowncast<const SubtractionQueueDescriptor*>(&descriptor);
613 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
614 {
615 return std::make_unique<RefSubtractionWorkload<int32_t>>(*subtractionQueueDescriptor, info);
616 }
617 else
618 {
619 return std::make_unique<RefSubtractionWorkload<float>>(*subtractionQueueDescriptor, info);
620 }
621 }
622 case LayerType::Transpose:
623 {
624 auto transposeQueueDescriptor = PolymorphicDowncast<const TransposeQueueDescriptor*>(&descriptor);
625 if (IsQSymmS16(info))
626 {
627 return std::make_unique<RefTransposeQSymm16Workload>(*transposeQueueDescriptor, info);
628 }
629 else if (IsBFloat16(info))
630 {
631 return std::make_unique<RefTransposeBFloat16Workload>(*transposeQueueDescriptor, info);
632 }
633 else if (IsQAsymmS8(info))
634 {
635 return std::make_unique<RefTransposeQAsymmS8Workload>(*transposeQueueDescriptor, info);
636 }
637 return MakeWorkloadHelper<RefTransposeFloat16Workload, RefTransposeFloat32Workload,
638 RefTransposeQAsymm8Workload, NullWorkload, NullWorkload, NullWorkload>
639 (*transposeQueueDescriptor, info);
640 }
641 case LayerType::TransposeConvolution2d:
642 {
643 auto transposeConvolution2dQueueDescriptor
644 = PolymorphicDowncast<const TransposeConvolution2dQueueDescriptor*>(&descriptor);
645 return std::make_unique<RefTransposeConvolution2dWorkload>(*transposeConvolution2dQueueDescriptor, info);
646 }
647 case LayerType::UnidirectionalSequenceLstm:
648 {
649 auto unidirectionalSequenceLstmQueueDescriptor
650 = PolymorphicDowncast<const UnidirectionalSequenceLstmQueueDescriptor*>(&descriptor);
651 return std::make_unique<RefUnidirectionalSequenceLstmWorkload>(*unidirectionalSequenceLstmQueueDescriptor,
652 info);
653 }
654 default:
655 return nullptr;
656 }
657}
658
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100659std::unique_ptr<IWorkload> RefWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
660 const WorkloadInfo& info) const
661{
662 return std::make_unique<RefActivationWorkload>(descriptor, info);
663}
664
665std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
666 const WorkloadInfo& info) const
667{
Finn Williamscbd2c232020-06-22 15:58:32 +0100668 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
669 {
670 return std::make_unique<RefAdditionWorkload<int32_t>>(descriptor, info);
671 }
672 else
673 {
674 return std::make_unique<RefAdditionWorkload<float>>(descriptor, info);
675 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100676}
677
678std::unique_ptr<IWorkload> RefWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
679 const WorkloadInfo& info) const
680{
681 return std::make_unique<RefArgMinMaxWorkload>(descriptor, info);
682}
683
684std::unique_ptr<IWorkload> RefWorkloadFactory::CreateBatchNormalization(
685 const BatchNormalizationQueueDescriptor& descriptor,
686 const WorkloadInfo& info) const
687{
688 return std::make_unique<RefBatchNormalizationWorkload>(descriptor, info);
689}
690
691std::unique_ptr<IWorkload> RefWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
692 const WorkloadInfo& info) const
693{
694 return std::make_unique<RefBatchToSpaceNdWorkload>(descriptor, info);
695}
696
mathad01b392e982021-04-07 12:07:30 +0100697std::unique_ptr<IWorkload> RefWorkloadFactory::CreateCast(const CastQueueDescriptor& descriptor,
698 const WorkloadInfo& info) const
699{
700 return std::make_unique<RefCastWorkload>(descriptor, info);
701}
702
Simon Obute51f67772021-09-03 15:50:13 +0100703std::unique_ptr<IWorkload> RefWorkloadFactory::CreateChannelShuffle(const ChannelShuffleQueueDescriptor &descriptor,
704 const WorkloadInfo &info) const
705{
706 return std::make_unique<RefChannelShuffleWorkload>(descriptor,info);
707}
708
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100709std::unique_ptr<IWorkload> RefWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
710 const WorkloadInfo& info) const
711{
712 return std::make_unique<RefComparisonWorkload>(descriptor, info);
713}
714
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100715std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
716 const WorkloadInfo& info) const
717{
718 return std::make_unique<RefConcatWorkload>(descriptor, info);
719}
720
721std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
722 const WorkloadInfo& info) const
723{
724 return std::make_unique<RefConstantWorkload>(descriptor, info);
725}
726
Narumol Prangnawarat7ddbbae2020-03-13 10:26:05 +0000727std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertBf16ToFp32(
728 const ConvertBf16ToFp32QueueDescriptor& descriptor,
729 const WorkloadInfo& info) const
730{
731 return std::make_unique<RefConvertBf16ToFp32Workload>(descriptor, info);
732}
733
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100734std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertFp16ToFp32(
735 const ConvertFp16ToFp32QueueDescriptor& descriptor,
736 const WorkloadInfo& info) const
737{
738 return std::make_unique<RefConvertFp16ToFp32Workload>(descriptor, info);
739}
740
Narumol Prangnawaratea54a012020-03-16 16:36:10 +0000741std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertFp32ToBf16(
742 const ConvertFp32ToBf16QueueDescriptor& descriptor,
743 const WorkloadInfo& info) const
744{
745 return std::make_unique<RefConvertFp32ToBf16Workload>(descriptor, info);
746}
747
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100748std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvertFp32ToFp16(
749 const ConvertFp32ToFp16QueueDescriptor& descriptor,
750 const WorkloadInfo& info) const
751{
752 return std::make_unique<RefConvertFp32ToFp16Workload>(descriptor, info);
753}
754
755std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
756 const WorkloadInfo& info) const
757{
758 return std::make_unique<RefConvolution2dWorkload>(descriptor, info);
759}
760
Matthew Sloyanb63a3112021-09-08 13:05:51 +0100761std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvolution3d(const Convolution3dQueueDescriptor& descriptor,
762 const WorkloadInfo& info) const
763{
764 return std::make_unique<RefConvolution3dWorkload>(descriptor, info);
765}
766
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100767std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
768 const WorkloadInfo& info) const
769{
Narumol Prangnawarat403a1852020-03-12 14:24:13 +0000770 if (IsBFloat16(info))
771 {
772 return std::make_unique<RefDebugBFloat16Workload>(descriptor, info);
773 }
Aron Virginas-Tardb1a2832019-11-12 16:15:11 +0000774 if (IsFloat16(info))
775 {
776 return std::make_unique<RefDebugFloat16Workload>(descriptor, info);
777 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000778 if (IsQSymmS16(info))
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100779 {
Keith Davis0c2eeac2020-02-11 16:51:50 +0000780 return std::make_unique<RefDebugQSymmS16Workload>(descriptor, info);
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100781 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000782 if (IsQSymmS8(info))
Keith Davis5204aa82020-01-27 15:24:59 +0000783 {
Keith Davis0c2eeac2020-02-11 16:51:50 +0000784 return std::make_unique<RefDebugQSymmS8Workload>(descriptor, info);
Keith Davis5204aa82020-01-27 15:24:59 +0000785 }
Keith Davis67e6c542020-02-19 10:08:33 +0000786 if (IsQAsymmU8(info))
787 {
788 return std::make_unique<RefDebugQAsymmU8Workload>(descriptor, info);
789 }
790 if (IsQAsymmS8(info))
791 {
792 return std::make_unique<RefDebugQAsymmS8Workload>(descriptor, info);
793 }
Keith Davis0c2eeac2020-02-11 16:51:50 +0000794 if (IsSigned32(info))
Narumol Prangnawaratd2d917d2020-01-09 10:16:39 +0000795 {
796 return std::make_unique<RefDebugSigned32Workload>(descriptor, info);
797 }
Aron Virginas-Tardb1a2832019-11-12 16:15:11 +0000798
Keith Davis0c2eeac2020-02-11 16:51:50 +0000799 return MakeWorkload<RefDebugFloat32Workload, RefDebugQAsymmU8Workload>(descriptor, info);
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100800}
801
802std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
803 const WorkloadInfo& info) const
804{
805 return std::make_unique<RefDepthToSpaceWorkload>(descriptor, info);
806}
807
808std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDepthwiseConvolution2d(
809 const DepthwiseConvolution2dQueueDescriptor& descriptor,
810 const WorkloadInfo& info) const
811{
812 return std::make_unique<RefDepthwiseConvolution2dWorkload>(descriptor, info);
813}
814
815std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
816 const WorkloadInfo& info) const
817{
818 return std::make_unique<RefDequantizeWorkload>(descriptor, info);
819}
820
821std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDetectionPostProcess(
822 const DetectionPostProcessQueueDescriptor& descriptor,
823 const WorkloadInfo& info) const
824{
825 return std::make_unique<RefDetectionPostProcessWorkload>(descriptor, info);
826}
827
828std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
829 const WorkloadInfo& info) const
830{
Finn Williamscbd2c232020-06-22 15:58:32 +0100831 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
832 {
833 return std::make_unique<RefDivisionWorkload<int32_t>>(descriptor, info);
834 }
835 else
836 {
837 return std::make_unique<RefDivisionWorkload<float>>(descriptor, info);
838 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100839}
840
josh minor4a3c6102020-01-06 16:40:46 -0600841std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
842 const WorkloadInfo& info) const
843{
Narumol Prangnawarat0c95f4c2020-11-18 16:52:07 +0000844 if (descriptor.m_Parameters.m_Operation == UnaryOperation::LogicalNot)
845 {
846 return std::make_unique<RefLogicalUnaryWorkload>(descriptor, info);
847 }
josh minor4a3c6102020-01-06 16:40:46 -0600848 return std::make_unique<RefElementwiseUnaryWorkload>(descriptor, info);
849}
850
Ryan OSheaf4bfa6a2020-06-10 11:33:37 +0100851std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor,
852 const WorkloadInfo& info) const
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100853{
854 return MakeWorkload<RefFakeQuantizationFloat32Workload, NullWorkload>(descriptor, info);
855}
856
Ryan OSheaf4bfa6a2020-06-10 11:33:37 +0100857std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
858 const WorkloadInfo& info) const
859{
860 return std::make_unique<RefFillWorkload>(descriptor, info);
861}
862
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100863std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
864 const WorkloadInfo& info) const
865{
Francis Murtaghe8ac1332020-07-30 18:03:40 +0100866 if(IsQuantizedType(info.m_InputTensorInfos[0].GetDataType()))
867 {
868 return nullptr;
869 }
870 else
871 {
872 return std::make_unique<RefFloorWorkload>(descriptor, info);
873 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100874}
875
876std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFullyConnected(
877 const FullyConnectedQueueDescriptor& descriptor,
878 const WorkloadInfo& info) const
879{
880 return std::make_unique<RefFullyConnectedWorkload>(descriptor, info);
881}
882
883std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
884 const WorkloadInfo& info) const
885{
886 return std::make_unique<RefGatherWorkload>(descriptor, info);
887}
888
telsoa014fcda012018-03-09 14:13:49 +0000889std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
890 const WorkloadInfo& info) const
891{
892 if (info.m_InputTensorInfos.empty() )
893 {
894 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Input cannot be zero length");
895 }
896 if (info.m_OutputTensorInfos.empty())
897 {
898 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Output cannot be zero length");
899 }
900
901 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
902 {
903 throw InvalidArgumentException("RefWorkloadFactory::CreateInput: data input and output differ in byte count.");
904 }
905
Narumol Prangnawaratb6441e42019-06-04 11:22:00 +0100906 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000907}
908
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100909std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInstanceNormalization(
910 const InstanceNormalizationQueueDescriptor& descriptor,
911 const WorkloadInfo& info) const
912{
913 return std::make_unique<RefInstanceNormalizationWorkload>(descriptor, info);
914}
915
916std::unique_ptr<IWorkload> RefWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
917 const WorkloadInfo& info) const
918{
919 return std::make_unique<RefL2NormalizationWorkload>(descriptor, info);
920}
921
James Conroyaba90cd2020-11-06 16:28:18 +0000922std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
923 const WorkloadInfo& info) const
924{
925 return std::make_unique<RefLogicalBinaryWorkload>(descriptor, info);
926}
927
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100928std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
929 const WorkloadInfo& info) const
930{
931 return std::make_unique<RefLogSoftmaxWorkload>(descriptor, info);
932}
933
934std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
935 const WorkloadInfo& info) const
936{
937 return std::make_unique<RefLstmWorkload>(descriptor, info);
938}
939
940std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
941 const WorkloadInfo& info) const
942{
Finn Williamscbd2c232020-06-22 15:58:32 +0100943 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
944 {
945 return std::make_unique<RefMaximumWorkload<int32_t>>(descriptor, info);
946 }
947 else
948 {
949 return std::make_unique<RefMaximumWorkload<float>>(descriptor, info);
950 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100951}
952
953std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
954 const WorkloadInfo& info) const
955{
956 return std::make_unique<RefMeanWorkload>(descriptor, info);
957}
958
959std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
960 const WorkloadInfo& info) const
961{
962 if (descriptor.m_Inputs.empty())
963 {
964 throw InvalidArgumentException("RefWorkloadFactory: CreateMemCopy() expected an input tensor.");
965 }
966 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
967}
968
969std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
970 const WorkloadInfo& info) const
971{
972 if (descriptor.m_Inputs.empty())
973 {
974 throw InvalidArgumentException("RefWorkloadFactory: CreateMemImport() expected an input tensor.");
975 }
976 return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
977}
978
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100979std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
980 const WorkloadInfo& info) const
981{
Finn Williamscbd2c232020-06-22 15:58:32 +0100982 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
983 {
984 return std::make_unique<RefMinimumWorkload<int32_t>>(descriptor, info);
985 }
986 else
987 {
988 return std::make_unique<RefMinimumWorkload<float>>(descriptor, info);
989 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +0100990}
991
992std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
993 const WorkloadInfo& info) const
994{
Finn Williamscbd2c232020-06-22 15:58:32 +0100995 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
996 {
997 return std::make_unique<RefMultiplicationWorkload<int32_t>>(descriptor, info);
998 }
999 else
1000 {
1001 return std::make_unique<RefMultiplicationWorkload<float>>(descriptor, info);
1002 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001003}
1004
1005std::unique_ptr<IWorkload> RefWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
1006 const WorkloadInfo& info) const
1007{
1008 return std::make_unique<RefNormalizationWorkload>(descriptor, info);
1009}
1010
telsoa014fcda012018-03-09 14:13:49 +00001011std::unique_ptr<IWorkload> RefWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
1012 const WorkloadInfo& info) const
1013{
1014 if (info.m_InputTensorInfos.empty() )
1015 {
1016 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Input cannot be zero length");
1017 }
1018 if (info.m_OutputTensorInfos.empty())
1019 {
1020 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Output cannot be zero length");
1021 }
1022 if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
1023 {
1024 throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: data input and output differ in byte count.");
1025 }
1026
Narumol Prangnawaratb6441e42019-06-04 11:22:00 +01001027 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001028}
1029
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001030std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
1031 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001032{
Sadik Armagan041b3c02020-06-04 10:32:18 +01001033 return std::make_unique<RefPadWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001034}
1035
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001036std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
1037 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001038{
Keith Davis0c2eeac2020-02-11 16:51:50 +00001039 if (IsQSymmS16(info))
Narumol Prangnawarat86bb4e12019-07-08 11:36:05 +01001040 {
1041 return std::make_unique<RefPermuteQSymm16Workload>(descriptor, info);
1042 }
Narumol Prangnawarat44179c32020-03-11 14:51:27 +00001043 else if (IsBFloat16(info))
1044 {
1045 return std::make_unique<RefPermuteBFloat16Workload>(descriptor, info);
1046 }
Sadik Armagan303980c2020-04-17 12:45:14 +01001047 else if (IsQAsymmS8(info))
1048 {
1049 return std::make_unique<RefPermuteQAsymmS8Workload>(descriptor, info);
1050 }
Narumol Prangnawarat86bb4e12019-07-08 11:36:05 +01001051 return MakeWorkloadHelper<RefPermuteFloat16Workload, RefPermuteFloat32Workload, RefPermuteQAsymm8Workload,
Keith Davis5204aa82020-01-27 15:24:59 +00001052 NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001053}
1054
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001055std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
1056 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001057{
Teresa Charlina3b20472019-06-06 11:12:32 +01001058 return std::make_unique<RefPooling2dWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001059}
1060
Tamás Nyíri7b885b32021-10-26 14:47:57 +01001061std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePooling3d(const Pooling3dQueueDescriptor& descriptor,
1062 const WorkloadInfo& info) const
1063{
1064 return std::make_unique<RefPooling3dWorkload>(descriptor, info);
1065}
1066
Derek Lamberti901ea112019-12-10 22:07:09 +00001067std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/,
1068 const WorkloadInfo& /*info*/) const
telsoa014fcda012018-03-09 14:13:49 +00001069{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001070 return nullptr;
telsoa014fcda012018-03-09 14:13:49 +00001071}
1072
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001073std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePrelu(const PreluQueueDescriptor& descriptor,
1074 const WorkloadInfo& info) const
Aron Virginas-Tar73f66422019-09-23 19:11:59 +01001075{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001076 return std::make_unique<RefPreluWorkload>(descriptor, info);
Aron Virginas-Tar73f66422019-09-23 19:11:59 +01001077}
1078
James Conroy4f1f8992020-04-29 20:01:10 +01001079std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
1080 const WorkloadInfo& info) const
1081{
1082 return std::make_unique<RefQLstmWorkload>(descriptor, info);
1083}
1084
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001085std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
1086 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +00001087{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001088 return std::make_unique<RefQuantizeWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +00001089}
1090
Finn Williams2605b232020-06-10 15:53:46 +01001091std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRank(const RankQueueDescriptor& descriptor,
1092 const WorkloadInfo& info) const
1093{
1094 return std::make_unique<RefRankWorkload>(descriptor, info);
1095}
1096
Sadik Armagan0c3ea5b2021-02-03 09:29:30 +00001097std::unique_ptr<IWorkload> RefWorkloadFactory::CreateReduce(const ReduceQueueDescriptor& descriptor,
1098 const WorkloadInfo& info) const
1099{
1100 return std::make_unique<RefReduceWorkload>(descriptor, info);
1101}
1102
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001103std::unique_ptr<IWorkload> RefWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
1104 const WorkloadInfo& info) const
Narumol Prangnawarat94dd5d82019-01-23 18:06:26 +00001105{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001106 return std::make_unique<RefReshapeWorkload>(descriptor, info);
Derek Lambertif674aa02019-08-01 15:56:25 +01001107}
1108
Teresa Charlin970f43b2019-07-01 13:51:07 +01001109std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
1110 const WorkloadInfo& info) const
1111{
Teresa Charlin970f43b2019-07-01 13:51:07 +01001112 return std::make_unique<RefResizeWorkload>(descriptor, info);
1113}
1114
Keith Davis3ae3f972021-05-21 16:33:48 +01001115std::unique_ptr<IWorkload> RefWorkloadFactory::CreateShape(const ShapeQueueDescriptor& descriptor,
1116 const WorkloadInfo& info) const
1117{
1118 return std::make_unique<RefShapeWorkload>(descriptor, info);
1119}
1120
Aron Virginas-Tar92b9f872019-09-17 17:27:04 +01001121std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
1122 const WorkloadInfo& info) const
1123{
1124 return std::make_unique<RefSliceWorkload>(descriptor, info);
1125}
1126
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001127std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
1128 const WorkloadInfo& info) const
Kevin May09ca49c2019-10-09 12:37:34 +01001129{
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001130 return std::make_unique<RefSoftmaxWorkload>(descriptor, info);
1131}
1132
1133std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
1134 const WorkloadInfo& info) const
1135{
1136 return std::make_unique<RefSpaceToBatchNdWorkload>(descriptor, info);
1137}
1138
1139std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
1140 const WorkloadInfo& info) const
1141{
1142 return std::make_unique<RefSpaceToDepthWorkload>(descriptor, info);
1143}
1144
1145std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
1146 const WorkloadInfo& info) const
1147{
1148 return std::make_unique<RefSplitterWorkload>(descriptor, info);
1149}
1150
1151std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
1152 const WorkloadInfo& info) const
1153{
1154 return std::make_unique<RefStackWorkload>(descriptor, info);
1155}
1156
1157std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
1158 const WorkloadInfo& info) const
1159{
1160 return std::make_unique<RefStridedSliceWorkload>(descriptor, info);
1161}
1162
1163std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
1164 const WorkloadInfo& info) const
1165{
Finn Williamscbd2c232020-06-22 15:58:32 +01001166 if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
1167 {
1168 return std::make_unique<RefSubtractionWorkload<int32_t>>(descriptor, info);
1169 }
1170 else
1171 {
1172 return std::make_unique<RefSubtractionWorkload<float>>(descriptor, info);
1173 }
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001174}
1175
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001176std::unique_ptr<IWorkload> RefWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
1177 const WorkloadInfo& info) const
1178{
1179 if (IsQSymmS16(info))
1180 {
1181 return std::make_unique<RefTransposeQSymm16Workload>(descriptor, info);
1182 }
Narumol Prangnawarat44179c32020-03-11 14:51:27 +00001183 else if (IsBFloat16(info))
1184 {
1185 return std::make_unique<RefTransposeBFloat16Workload>(descriptor, info);
1186 }
Sadik Armagan303980c2020-04-17 12:45:14 +01001187 else if (IsQAsymmS8(info))
1188 {
1189 return std::make_unique<RefTransposeQAsymmS8Workload>(descriptor, info);
1190 }
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001191 return MakeWorkloadHelper<RefTransposeFloat16Workload, RefTransposeFloat32Workload, RefTransposeQAsymm8Workload,
1192 NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
1193}
1194
Aron Virginas-Tare662a942019-10-14 15:12:00 +01001195std::unique_ptr<IWorkload> RefWorkloadFactory::CreateTransposeConvolution2d(
1196 const TransposeConvolution2dQueueDescriptor& descriptor,
1197 const WorkloadInfo& info) const
1198{
1199 return std::make_unique<RefTransposeConvolution2dWorkload>(descriptor, info);
Kevin May09ca49c2019-10-09 12:37:34 +01001200}
1201
Narumol Prangnawarate5339e72021-07-28 17:33:28 +01001202std::unique_ptr<IWorkload> RefWorkloadFactory::CreateUnidirectionalSequenceLstm(
1203 const UnidirectionalSequenceLstmQueueDescriptor& descriptor,
1204 const WorkloadInfo& info) const
1205{
1206 return std::make_unique<RefUnidirectionalSequenceLstmWorkload>(descriptor, info);;
1207}
1208
Matteo Martincigh49124022019-01-11 13:25:59 +00001209} // namespace armnn