blob: 41b779f64a7605ac084f00c2d993350680938b34 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Teresa Charlin8398edc2020-07-20 14:23:02 +01002// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#include "ClWorkloadFactory.hpp"
David Beck79141b92018-10-23 16:09:36 +01006#include "ClBackendId.hpp"
Sadik Armagan04a72972020-09-14 15:44:18 +01007#include "ClBackendModelContext.hpp"
telsoa014fcda012018-03-09 14:13:49 +00008
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +00009#include <Layer.hpp>
10
David Beck0dbe0ee2018-09-24 15:59:27 +010011#include <armnn/Exceptions.hpp>
12#include <armnn/Utils.hpp>
Jan Eilers8eb25602020-03-09 12:13:48 +000013#include <armnn/utility/IgnoreUnused.hpp>
Matthew Sloyan171214c2020-09-09 09:07:37 +010014#include <armnn/utility/NumericCast.hpp>
Jan Eilersbb446e52020-04-02 13:56:54 +010015#include <armnn/utility/PolymorphicDowncast.hpp>
telsoa014fcda012018-03-09 14:13:49 +000016
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000017#include <backendsCommon/CpuTensorHandle.hpp>
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000018#include <backendsCommon/MakeWorkloadHelper.hpp>
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000019#include <backendsCommon/MemCopyWorkload.hpp>
Derek Lambertif674aa02019-08-01 15:56:25 +010020#include <backendsCommon/MemImportWorkload.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010021
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000022#include <cl/ClTensorHandle.hpp>
23#include <cl/workloads/ClWorkloads.hpp>
24#include <cl/workloads/ClWorkloadUtils.hpp>
telsoa014fcda012018-03-09 14:13:49 +000025
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000026#include <arm_compute/core/CL/CLKernelLibrary.h>
27#include <arm_compute/runtime/CL/CLBufferAllocator.h>
28#include <arm_compute/runtime/CL/CLScheduler.h>
telsoa014fcda012018-03-09 14:13:49 +000029
Sadik Armagandea8fb62020-11-26 10:38:11 +000030#include <Filesystem.hpp>
31
telsoa014fcda012018-03-09 14:13:49 +000032namespace armnn
33{
34
David Beck79141b92018-10-23 16:09:36 +010035namespace
36{
37static const BackendId s_Id{ClBackendId()};
38}
39
telsoa01c577f2c2018-08-31 09:22:23 +010040bool ClWorkloadFactory::IsLayerSupported(const Layer& layer,
David Beck29c75de2018-10-23 13:35:58 +010041 Optional<DataType> dataType,
telsoa01c577f2c2018-08-31 09:22:23 +010042 std::string& outReasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000043{
David Beck79141b92018-10-23 16:09:36 +010044 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
45}
46
Sadik Armagan04a72972020-09-14 15:44:18 +010047bool ClWorkloadFactory::IsLayerSupported(const IConnectableLayer& layer,
48 Optional<DataType> dataType,
49 std::string& outReasonIfUnsupported,
50 const ModelOptions& modelOptions)
51{
52 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
53}
54
David Beck79141b92018-10-23 16:09:36 +010055const BackendId& ClWorkloadFactory::GetBackendId() const
56{
57 return s_Id;
telsoa014fcda012018-03-09 14:13:49 +000058}
59
Sadik Armagandea8fb62020-11-26 10:38:11 +000060void ClWorkloadFactory::AfterWorkloadsCreated()
61{
62 if(m_ModelContextPtr)
63 {
64 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
65 if (modelOptions->SaveCachedNetwork())
66 {
67 // Save map to a filepath provided in ModelOptions
68 auto filePath = modelOptions->GetCachedNetworkFilePath();
69 if (filePath != "" && fs::exists(filePath) && fs::is_regular_file(filePath))
70 {
71 /// Saving will be implemented within IVGCVSW-5483 story.
72 }
73 }
74 }
75}
76
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +010077template <typename FloatWorkload, typename Uint8Workload, typename QueueDescriptorType, typename... Args>
78std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
79 const WorkloadInfo& info,
80 Args&&... args)
81{
82 try
83 {
84 return MakeWorkloadHelper<FloatWorkload, Uint8Workload>(descriptor, info, std::forward<Args>(args)...);
85 }
86 catch (const cl::Error& clError)
87 {
88 throw WrapClError(clError, CHECK_LOCATION());
89 }
90}
91
92template <typename Workload, typename QueueDescriptorType, typename... Args>
93std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
94 const WorkloadInfo& info,
95 Args&&... args)
96{
97 try
98 {
99 return std::make_unique<Workload>(descriptor, info, std::forward<Args>(args)...);
100 }
101 catch (const cl::Error& clError)
102 {
103 throw WrapClError(clError, CHECK_LOCATION());
104 }
105}
106
Sadik Armagandea8fb62020-11-26 10:38:11 +0000107void ClWorkloadFactory::InitializeCLCompileContext()
108{
109 // Initialize our m_CLCompileContext using default device and context
110 cl::Device device = cl::Device::getDefault();
111 cl::Context context = cl::Context(device);
112
113 m_CLCompileContext = arm_compute::CLCompileContext(context, device);
114
115 if (m_ModelContextPtr)
116 {
117 // Load saved programs if the user has set a filepath
118 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
119 auto filePath = modelOptions->GetCachedNetworkFilePath();
120 if (filePath != ""
121 && fs::exists(filePath)
122 && fs::is_regular_file(filePath)
123 && !(modelOptions->SaveCachedNetwork()))
124 {
125 /// Loading will be implemented within IVGCVSW-5483 story.
126 }
127 }
128}
129
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000130ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager)
Sadik Armagan04a72972020-09-14 15:44:18 +0100131 : m_MemoryManager(memoryManager), m_ModelContextPtr(IBackendInternal::IBackendSpecificModelContextPtr{})
132{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000133 InitializeCLCompileContext();
Sadik Armagan04a72972020-09-14 15:44:18 +0100134}
135
136ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager,
137 const IBackendInternal::IBackendSpecificModelContextPtr& modelContextPtr)
138 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
telsoa014fcda012018-03-09 14:13:49 +0000139{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000140 InitializeCLCompileContext();
telsoa014fcda012018-03-09 14:13:49 +0000141}
142
David Monahan3fb7e102019-08-20 11:25:29 +0100143std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
144 const bool IsMemoryManaged) const
telsoa014fcda012018-03-09 14:13:49 +0000145{
Jan Eilers8eb25602020-03-09 12:13:48 +0000146 IgnoreUnused(IsMemoryManaged);
telsoa01c577f2c2018-08-31 09:22:23 +0100147 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000148 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
telsoa01c577f2c2018-08-31 09:22:23 +0100149
150 return tensorHandle;
telsoa014fcda012018-03-09 14:13:49 +0000151}
152
Francis Murtagh351d13d2018-09-24 15:01:18 +0100153std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
David Monahan3fb7e102019-08-20 11:25:29 +0100154 DataLayout dataLayout,
155 const bool IsMemoryManaged) const
Francis Murtagh351d13d2018-09-24 15:01:18 +0100156{
Jan Eilers8eb25602020-03-09 12:13:48 +0000157 IgnoreUnused(IsMemoryManaged);
Francis Murtagh351d13d2018-09-24 15:01:18 +0100158 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000159 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
Francis Murtagh351d13d2018-09-24 15:01:18 +0100160
161 return tensorHandle;
162}
163
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100164std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateSubTensorHandle(ITensorHandle& parent,
165 TensorShape const& subTensorShape,
telsoa014fcda012018-03-09 14:13:49 +0000166 unsigned int const* subTensorOrigin) const
167{
telsoa014fcda012018-03-09 14:13:49 +0000168 arm_compute::Coordinates coords;
169 arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
170
171 coords.set_num_dimensions(subTensorShape.GetNumDimensions());
172 for (unsigned int i = 0; i < subTensorShape.GetNumDimensions(); i++)
173 {
telsoa01c577f2c2018-08-31 09:22:23 +0100174 // Arm compute indexes tensor coords in reverse order.
telsoa014fcda012018-03-09 14:13:49 +0000175 unsigned int revertedIndex = subTensorShape.GetNumDimensions() - i - 1;
Matthew Sloyan171214c2020-09-09 09:07:37 +0100176 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
telsoa014fcda012018-03-09 14:13:49 +0000177 }
178
Derek Lamberti0790dce2019-04-15 18:37:35 +0100179 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.GetShape());
180 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
181 {
182 return nullptr;
183 }
184
telsoa01c577f2c2018-08-31 09:22:23 +0100185 return std::make_unique<ClSubTensorHandle>(
Jan Eilersbb446e52020-04-02 13:56:54 +0100186 PolymorphicDowncast<IClTensorHandle*>(&parent), shape, coords);
telsoa014fcda012018-03-09 14:13:49 +0000187}
188
Aron Virginas-Tar82046942019-09-09 15:18:29 +0100189std::unique_ptr<IWorkload> ClWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100190 const WorkloadInfo& info) const
Aron Virginas-Tar82046942019-09-09 15:18:29 +0100191{
Jan Eilers8eb25602020-03-09 12:13:48 +0000192 IgnoreUnused(descriptor);
josh minor4a3c6102020-01-06 16:40:46 -0600193
194 ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
195 elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Abs);
196
197 return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
Aron Virginas-Tar82046942019-09-09 15:18:29 +0100198}
199
telsoa014fcda012018-03-09 14:13:49 +0000200std::unique_ptr<IWorkload> ClWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100201 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000202{
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100203 return MakeWorkload<ClActivationWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000204}
205
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100206std::unique_ptr<IWorkload> ClWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
207 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000208{
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100209 return MakeWorkload<ClAdditionWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000210}
211
James Conroy2dc05722019-09-19 17:00:31 +0100212std::unique_ptr<IWorkload> ClWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
213 const WorkloadInfo& info) const
214{
215 return std::make_unique<ClArgMinMaxWorkload>(descriptor, info);
216}
217
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100218std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchNormalization(
219 const BatchNormalizationQueueDescriptor& descriptor,
220 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000221{
arovir019e53a352018-08-31 15:26:35 +0100222 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000223}
224
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100225std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
telsoa014fcda012018-03-09 14:13:49 +0000226 const WorkloadInfo& info) const
227{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100228 return MakeWorkload<ClBatchToSpaceNdWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000229}
230
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100231std::unique_ptr<IWorkload> ClWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
232 const WorkloadInfo& info) const
233{
Teresa Charlin2b030d92020-03-27 16:40:56 +0000234 return MakeWorkload<ClComparisonWorkload>(descriptor, info);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100235}
236
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100237std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
238 const WorkloadInfo& info) const
Jim Flynn4ed6c832019-05-20 11:02:46 +0100239{
240 return MakeWorkload<ClConcatWorkload>(descriptor, info);
241}
242
telsoa014fcda012018-03-09 14:13:49 +0000243std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100244 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000245{
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100246 return MakeWorkload<ClConstantWorkload>(descriptor, info);
telsoa014fcda012018-03-09 14:13:49 +0000247}
248
telsoa01c577f2c2018-08-31 09:22:23 +0100249std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp16ToFp32(
250 const ConvertFp16ToFp32QueueDescriptor& descriptor,
251 const WorkloadInfo& info) const
252{
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100253 return MakeWorkload<ClConvertFp16ToFp32Workload>(descriptor, info);
telsoa01c577f2c2018-08-31 09:22:23 +0100254}
255
256std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp32ToFp16(
257 const ConvertFp32ToFp16QueueDescriptor& descriptor,
258 const WorkloadInfo& info) const
259{
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100260 return MakeWorkload<ClConvertFp32ToFp16Workload>(descriptor, info);
telsoa01c577f2c2018-08-31 09:22:23 +0100261}
262
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100263std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
264 const WorkloadInfo& info) const
265{
Sadik Armagan04a72972020-09-14 15:44:18 +0100266 bool isFastMathEnabled = false;
267 if (m_ModelContextPtr)
268 {
269 if (m_ModelContextPtr.get() != nullptr)
270 {
271 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
272 if (modelOptions)
273 {
274 isFastMathEnabled = modelOptions->IsFastMathEnabled();
275 }
276 }
277 }
278 return MakeWorkload<ClConvolution2dWorkload>(descriptor,
279 info,
280 m_MemoryManager->GetIntraLayerManager(),
281 isFastMathEnabled);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100282}
283
284std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
285 const WorkloadInfo& info) const
286{
287 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info);
288}
289
290std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
291 const WorkloadInfo& info) const
292{
293 return MakeWorkload<ClDepthToSpaceWorkload>(descriptor, info);
294}
295
296std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthwiseConvolution2d(
297 const DepthwiseConvolution2dQueueDescriptor& descriptor,
298 const WorkloadInfo& info) const
299{
300 return MakeWorkload<ClDepthwiseConvolutionWorkload>(descriptor, info);
301}
302
303std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
304 const WorkloadInfo& info) const
305{
306 return MakeWorkload<ClDequantizeWorkload>(descriptor, info);
307}
308
309std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDetectionPostProcess(
310 const DetectionPostProcessQueueDescriptor& descriptor,
311 const WorkloadInfo& info) const
312{
313 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info);
314}
315
316std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
317 const WorkloadInfo& info) const
318{
319 return MakeWorkload<ClDivisionFloatWorkload, NullWorkload>(descriptor, info);
320}
321
josh minor4a3c6102020-01-06 16:40:46 -0600322std::unique_ptr<IWorkload> ClWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
323 const WorkloadInfo& info) const
324{
Sadik Armagan9fabf432020-05-27 13:40:58 +0100325 switch(descriptor.m_Parameters.m_Operation)
josh minor4a3c6102020-01-06 16:40:46 -0600326 {
Sadik Armagan9fabf432020-05-27 13:40:58 +0100327 case UnaryOperation::Abs:
James Conroyfe3ec942020-11-18 14:20:53 +0000328 {
329 AbsQueueDescriptor absQueueDescriptor;
330 absQueueDescriptor.m_Inputs = descriptor.m_Inputs;
331 absQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600332
James Conroyfe3ec942020-11-18 14:20:53 +0000333 return std::make_unique<ClAbsWorkload>(absQueueDescriptor, info);
334 }
Sadik Armagan9fabf432020-05-27 13:40:58 +0100335 case UnaryOperation::Exp:
336 return std::make_unique<ClExpWorkload>(descriptor, info);
337 case UnaryOperation::Neg:
338 return std::make_unique<ClNegWorkload>(descriptor, info);
339 case UnaryOperation::Rsqrt:
James Conroyfe3ec942020-11-18 14:20:53 +0000340 {
341 RsqrtQueueDescriptor rsqrtQueueDescriptor;
342 rsqrtQueueDescriptor.m_Inputs = descriptor.m_Inputs;
343 rsqrtQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600344
James Conroyfe3ec942020-11-18 14:20:53 +0000345 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor, info);
346 }
347 case UnaryOperation::LogicalNot:
348 return std::make_unique<ClLogicalNotWorkload>(descriptor, info);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100349 default:
350 return nullptr;
josh minor4a3c6102020-01-06 16:40:46 -0600351 }
josh minor4a3c6102020-01-06 16:40:46 -0600352}
353
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100354std::unique_ptr<IWorkload> ClWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
355 const WorkloadInfo& info) const
356{
Jan Eilers8eb25602020-03-09 12:13:48 +0000357 IgnoreUnused(descriptor);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100358
359 ComparisonQueueDescriptor comparisonDescriptor;
360 comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Equal);
361
362 return CreateComparison(comparisonDescriptor, info);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100363}
364
Sadik Armagan66aecb02020-06-24 11:42:20 +0100365std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
366 const WorkloadInfo& info) const
367{
368 return std::make_unique<ClFillWorkload>(descriptor, info);
369}
370
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100371std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
372 const WorkloadInfo& info) const
373{
374 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor, info);
375}
376
377std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFullyConnected(const FullyConnectedQueueDescriptor& descriptor,
378 const WorkloadInfo& info) const
379{
380 return MakeWorkload<ClFullyConnectedWorkload>(descriptor, info, m_MemoryManager->GetIntraLayerManager());
381}
382
383std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
384 const WorkloadInfo& info) const
385{
Teresa Charlin9ad2e5b2020-04-10 22:34:48 +0100386 return MakeWorkload<ClGatherWorkload>(descriptor, info);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100387}
388
389std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
390 const WorkloadInfo& info) const
391{
Jan Eilers8eb25602020-03-09 12:13:48 +0000392 IgnoreUnused(descriptor);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100393
394 ComparisonQueueDescriptor comparisonDescriptor;
395 comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Greater);
396
397 return CreateComparison(comparisonDescriptor, info);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100398}
399
400std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
401 const WorkloadInfo& info) const
402{
403 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
404}
405
406std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInstanceNormalization(
407 const InstanceNormalizationQueueDescriptor& descriptor,
408 const WorkloadInfo& info) const
409{
410 return MakeWorkload<ClInstanceNormalizationWorkload>(descriptor, info);
411}
412
413std::unique_ptr<IWorkload> ClWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
414 const WorkloadInfo& info) const
415{
416 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor, info);
417}
418
James Conroyfe3ec942020-11-18 14:20:53 +0000419std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
420 const WorkloadInfo& info) const
421{
422 switch(descriptor.m_Parameters.m_Operation)
423 {
424 case LogicalBinaryOperation::LogicalAnd:
425 return std::make_unique<ClLogicalAndWorkload>(descriptor, info);
426 case LogicalBinaryOperation::LogicalOr:
427 return std::make_unique<ClLogicalOrWorkload>(descriptor, info);
428 default:
429 return nullptr;
430 }
431}
432
Teresa Charlin8398edc2020-07-20 14:23:02 +0100433std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
434 const WorkloadInfo& info) const
435{
436 return MakeWorkload<ClLogSoftmaxWorkload>(descriptor, info, m_MemoryManager->GetIntraLayerManager());
437}
438
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100439std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
440 const WorkloadInfo& info) const
441{
442 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor, info);
443}
444
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +0000445std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
446 const WorkloadInfo& info) const
447{
keidav01a959ee52018-12-19 10:04:58 +0000448 return MakeWorkload<ClMaximumWorkload>(descriptor, info);
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +0000449}
450
narpra01a6bf9122018-09-10 09:50:09 +0100451std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
452 const WorkloadInfo& info) const
453{
Aron Virginas-Tar39f44af2019-09-10 13:32:42 +0100454 return MakeWorkload<ClMeanWorkload>(descriptor, info);
narpra01a6bf9122018-09-10 09:50:09 +0100455}
456
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100457std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
458 const WorkloadInfo& info) const
jimfly012c9322a2018-09-19 10:59:49 +0100459{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100460 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
461 {
462 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload");
463 }
464
465 return MakeWorkload<CopyMemGenericWorkload>(descriptor, info);
jimfly012c9322a2018-09-19 10:59:49 +0100466}
467
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100468std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
469 const WorkloadInfo& info) const
FrancisMurtagh20995952018-12-17 12:11:36 +0000470{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100471 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
472 {
473 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemImport workload");
474 }
475
476 return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
FrancisMurtagh20995952018-12-17 12:11:36 +0000477}
478
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100479std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor,
480 const WorkloadInfo& info) const
Éanna Ó Catháin4e1e1362018-11-12 11:36:34 +0000481{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100482 return CreateConcat(descriptor, info);
Conor Kennedy430b5d82018-11-14 15:28:28 +0000483}
484
kevmay0190539692018-11-29 08:40:19 +0000485std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
486 const WorkloadInfo& info) const
487{
saoste019292aa32019-01-08 13:55:59 +0000488 return MakeWorkload<ClMinimumWorkload>(descriptor, info);
kevmay0190539692018-11-29 08:40:19 +0000489}
490
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100491std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
492 const WorkloadInfo& info) const
493{
494 return MakeWorkload<ClMultiplicationWorkload>(descriptor, info);
495}
496
497std::unique_ptr<IWorkload> ClWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
498 const WorkloadInfo& info) const
499{
500 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor, info);
501}
502
503std::unique_ptr<IWorkload> ClWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
504 const WorkloadInfo& info) const
505{
506 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
507}
508
509std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
510 const WorkloadInfo& info) const
511{
512 return MakeWorkload<ClPadWorkload>(descriptor, info);
513}
514
515std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
Matteo Martincigh59a950c2018-12-13 12:48:25 +0000516 const WorkloadInfo& info) const
517{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100518 return MakeWorkload<ClPermuteWorkload>(descriptor, info);
Matteo Martincigh59a950c2018-12-13 12:48:25 +0000519}
520
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100521std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
522 const WorkloadInfo& info) const
Nattapat Chaimanowonga9a1cf12018-12-03 16:06:49 +0000523{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100524 return MakeWorkload<ClPooling2dWorkload>(descriptor, info);
Mohamed Nour Abouelseouda1d3c6a2018-12-27 12:39:16 +0000525}
526
Matteo Martincigh49124022019-01-11 13:25:59 +0000527std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor,
528 const WorkloadInfo& info) const
529{
530 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info);
531}
532
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100533std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePrelu(const PreluQueueDescriptor &descriptor,
534 const WorkloadInfo &info) const
narpra01b89b05f2019-01-16 09:53:09 +0000535{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100536 return MakeWorkload<ClPreluWorkload>(descriptor, info);
narpra01b89b05f2019-01-16 09:53:09 +0000537}
538
Ryan OShea2323af42020-05-13 16:36:19 +0100539std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
540 const WorkloadInfo& info) const
541{
542 return std::make_unique<ClQLstmWorkload>(descriptor, info);
543}
544
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100545std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
546 const WorkloadInfo& info) const
Aron Virginas-Tar7a3e2fe2019-06-27 18:54:47 +0100547{
Keith Davisa8565012020-02-14 12:22:40 +0000548 return MakeWorkload<ClQuantizeWorkload>(descriptor, info);
James Conroyd2aa85e2019-07-01 17:12:40 +0100549}
550
Ferran Balaguer737d9ff2019-08-01 09:58:08 +0100551std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& descriptor,
552 const WorkloadInfo& info) const
553{
554 return MakeWorkload<ClQuantizedLstmWorkload>(descriptor, info);
555}
556
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100557std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
558 const WorkloadInfo& info) const
559{
560 return MakeWorkload<ClReshapeWorkload>(descriptor, info);
561}
562
563std::unique_ptr<IWorkload> ClWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
564 const WorkloadInfo& info) const
565{
566 return MakeWorkload<ClResizeWorkload>(descriptor, info);
567}
568
569std::unique_ptr<IWorkload> ClWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor,
570 const WorkloadInfo& info) const
571{
572 ResizeQueueDescriptor resizeDescriptor;
573 resizeDescriptor.m_Inputs = descriptor.m_Inputs;
574 resizeDescriptor.m_Outputs = descriptor.m_Outputs;
575
576 resizeDescriptor.m_Parameters.m_Method = ResizeMethod::Bilinear;
577 resizeDescriptor.m_Parameters.m_DataLayout = descriptor.m_Parameters.m_DataLayout;
578 resizeDescriptor.m_Parameters.m_TargetHeight = descriptor.m_Parameters.m_TargetHeight;
579 resizeDescriptor.m_Parameters.m_TargetWidth = descriptor.m_Parameters.m_TargetWidth;
580
581 return CreateResize(resizeDescriptor, info);
582}
583
584std::unique_ptr<IWorkload> ClWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
585 const WorkloadInfo& info) const
586{
Jan Eilers8eb25602020-03-09 12:13:48 +0000587 IgnoreUnused(descriptor);
josh minor4a3c6102020-01-06 16:40:46 -0600588
589 ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
590 elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Rsqrt);
591
592 return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100593}
594
Aron Virginas-Tar94c4fef2019-11-25 15:37:08 +0000595std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
596 const WorkloadInfo& info) const
597{
598 return MakeWorkload<ClSliceWorkload>(descriptor, info);
599}
600
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100601std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
602 const WorkloadInfo& info) const
603{
Teresa Charlinc1f6b092020-05-11 16:10:38 +0100604 return std::make_unique<ClSoftmaxWorkload>(descriptor, info, m_MemoryManager->GetIntraLayerManager());
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100605}
606
607std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
608 const WorkloadInfo& info) const
609{
610 return MakeWorkload<ClSpaceToBatchNdWorkload>(descriptor, info);
611}
612
613std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
614 const WorkloadInfo& info) const
615{
616 return MakeWorkload<ClSpaceToDepthWorkload>(descriptor, info);
617}
618
619std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
620 const WorkloadInfo& info) const
621{
622 return MakeWorkload<ClSplitterWorkload>(descriptor, info);
623}
624
Matthew Jacksond5166102019-07-31 14:06:28 +0100625std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
626 const WorkloadInfo& info) const
627{
628 return MakeWorkload<ClStackWorkload>(descriptor, info);
629}
630
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100631std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
Aron Virginas-Tarb2801962019-09-30 11:24:53 +0100632 const WorkloadInfo& info) const
633{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100634 return MakeWorkload<ClStridedSliceWorkload>(descriptor, info);
635}
636
637std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
638 const WorkloadInfo& info) const
639{
640 return MakeWorkload<ClSubtractionWorkload>(descriptor, info);
641}
642
Mike Kellyc9ea45a2020-02-28 18:11:58 +0000643std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
644 const WorkloadInfo& info) const
645{
646 return MakeWorkload<ClTransposeWorkload>(descriptor, info);
647}
648
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100649std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTransposeConvolution2d(
650 const TransposeConvolution2dQueueDescriptor& descriptor,
651 const WorkloadInfo& info) const
652{
653 return MakeWorkload<ClTransposeConvolution2dWorkload>(descriptor, info, m_MemoryManager->GetIntraLayerManager());
Aron Virginas-Tarb2801962019-09-30 11:24:53 +0100654}
655
telsoa014fcda012018-03-09 14:13:49 +0000656} // namespace armnn