blob: 493080f7af9ac74407f241a65d7397ebfbbfbf04 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Mike Kelly3ec30772023-03-08 13:47:17 +00002// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#include "ClWorkloadFactory.hpp"
David Beck79141b92018-10-23 16:09:36 +01006#include "ClBackendId.hpp"
Sadik Armagan04a72972020-09-14 15:44:18 +01007#include "ClBackendModelContext.hpp"
Matthew Sloyan80fbcd52021-01-07 13:28:47 +00008#include "ClContextDeserializer.hpp"
9#include "ClContextSerializer.hpp"
telsoa014fcda012018-03-09 14:13:49 +000010
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000011#include <Layer.hpp>
12
David Beck0dbe0ee2018-09-24 15:59:27 +010013#include <armnn/Exceptions.hpp>
Sadik Armagan6b9eba22021-11-01 07:27:50 +000014#include <armnn/Logging.hpp>
David Beck0dbe0ee2018-09-24 15:59:27 +010015#include <armnn/Utils.hpp>
Jan Eilers8eb25602020-03-09 12:13:48 +000016#include <armnn/utility/IgnoreUnused.hpp>
Matthew Sloyan171214c2020-09-09 09:07:37 +010017#include <armnn/utility/NumericCast.hpp>
Jan Eilersbb446e52020-04-02 13:56:54 +010018#include <armnn/utility/PolymorphicDowncast.hpp>
telsoa014fcda012018-03-09 14:13:49 +000019
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000020#include <backendsCommon/MakeWorkloadHelper.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +000021#include <armnn/backends/MemCopyWorkload.hpp>
Derek Lambertif674aa02019-08-01 15:56:25 +010022#include <backendsCommon/MemImportWorkload.hpp>
Colm Donelan0c479742021-12-10 12:43:54 +000023#include <armnn/backends/TensorHandle.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010024
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000025#include <cl/ClTensorHandle.hpp>
26#include <cl/workloads/ClWorkloads.hpp>
27#include <cl/workloads/ClWorkloadUtils.hpp>
telsoa014fcda012018-03-09 14:13:49 +000028
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000029#include <arm_compute/core/CL/CLKernelLibrary.h>
30#include <arm_compute/runtime/CL/CLBufferAllocator.h>
31#include <arm_compute/runtime/CL/CLScheduler.h>
telsoa014fcda012018-03-09 14:13:49 +000032
Rob Hughes9542f902021-07-14 09:48:54 +010033#include <armnnUtils/Filesystem.hpp>
Matthew Sloyan80fbcd52021-01-07 13:28:47 +000034#include <fstream>
Sadik Armagandea8fb62020-11-26 10:38:11 +000035
Sadik Armaganb7851f92021-10-06 16:37:02 +010036#include <sys/stat.h>
37
telsoa014fcda012018-03-09 14:13:49 +000038namespace armnn
39{
40
David Beck79141b92018-10-23 16:09:36 +010041namespace
42{
43static const BackendId s_Id{ClBackendId()};
44}
45
telsoa01c577f2c2018-08-31 09:22:23 +010046bool ClWorkloadFactory::IsLayerSupported(const Layer& layer,
David Beck29c75de2018-10-23 13:35:58 +010047 Optional<DataType> dataType,
telsoa01c577f2c2018-08-31 09:22:23 +010048 std::string& outReasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000049{
David Beck79141b92018-10-23 16:09:36 +010050 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
51}
52
Sadik Armagan04a72972020-09-14 15:44:18 +010053bool ClWorkloadFactory::IsLayerSupported(const IConnectableLayer& layer,
54 Optional<DataType> dataType,
55 std::string& outReasonIfUnsupported,
56 const ModelOptions& modelOptions)
57{
58 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
59}
60
David Beck79141b92018-10-23 16:09:36 +010061const BackendId& ClWorkloadFactory::GetBackendId() const
62{
63 return s_Id;
telsoa014fcda012018-03-09 14:13:49 +000064}
65
Sadik Armagandea8fb62020-11-26 10:38:11 +000066void ClWorkloadFactory::AfterWorkloadsCreated()
67{
68 if(m_ModelContextPtr)
69 {
70 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
71 if (modelOptions->SaveCachedNetwork())
72 {
Sadik Armaganb7851f92021-10-06 16:37:02 +010073 ClContextSerializer serializer;
74 serializer.Serialize(m_CLCompileContext);
75 auto cachedFd = modelOptions->GetCachedFileDescriptor();
76 if (cachedFd != -1)
77 {
78 std::vector<uint8_t> compiledContextData;
79 std::stringstream stream;
80 bool serialized = serializer.SaveSerializedToStream(stream);
81 if (serialized)
82 {
83 std::string const serializedString{stream.str()};
84 std::copy(serializedString.begin(),
85 serializedString.end(),
86 std::back_inserter(compiledContextData));
Sadik Armagan6b9eba22021-11-01 07:27:50 +000087 auto success = write(cachedFd, compiledContextData.data(), compiledContextData.size());
88 if (success == -1)
89 {
90 ARMNN_LOG(info) << "ClWorkloadFactory:: Could not cache the compiled context!";
91 }
Sadik Armaganb7851f92021-10-06 16:37:02 +010092 }
93 }
94
Sadik Armagandea8fb62020-11-26 10:38:11 +000095 // Save map to a filepath provided in ModelOptions
96 auto filePath = modelOptions->GetCachedNetworkFilePath();
97 if (filePath != "" && fs::exists(filePath) && fs::is_regular_file(filePath))
98 {
Matthew Sloyan80fbcd52021-01-07 13:28:47 +000099 // Serialize ClContext to the file specified
Matthew Sloyan80fbcd52021-01-07 13:28:47 +0000100 std::ofstream file(filePath, std::ios::out | std::ios::binary);
101 serializer.SaveSerializedToStream(file);
Sadik Armagandea8fb62020-11-26 10:38:11 +0000102 }
103 }
104 }
105}
106
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +0100107template <typename FloatWorkload, typename Uint8Workload, typename QueueDescriptorType, typename... Args>
108std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
109 const WorkloadInfo& info,
110 Args&&... args)
111{
112 try
113 {
114 return MakeWorkloadHelper<FloatWorkload, Uint8Workload>(descriptor, info, std::forward<Args>(args)...);
115 }
116 catch (const cl::Error& clError)
117 {
118 throw WrapClError(clError, CHECK_LOCATION());
119 }
120}
121
122template <typename Workload, typename QueueDescriptorType, typename... Args>
123std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
124 const WorkloadInfo& info,
125 Args&&... args)
126{
127 try
128 {
129 return std::make_unique<Workload>(descriptor, info, std::forward<Args>(args)...);
130 }
131 catch (const cl::Error& clError)
132 {
133 throw WrapClError(clError, CHECK_LOCATION());
134 }
135}
136
Sadik Armagandea8fb62020-11-26 10:38:11 +0000137void ClWorkloadFactory::InitializeCLCompileContext()
138{
139 // Initialize our m_CLCompileContext using default device and context
Sadik Armagane9444752020-12-02 11:28:58 +0000140 auto context = arm_compute::CLKernelLibrary::get().context();
141 auto device = arm_compute::CLKernelLibrary::get().get_device();
Sadik Armagandea8fb62020-11-26 10:38:11 +0000142 m_CLCompileContext = arm_compute::CLCompileContext(context, device);
143
144 if (m_ModelContextPtr)
145 {
146 // Load saved programs if the user has set a filepath
147 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
148 auto filePath = modelOptions->GetCachedNetworkFilePath();
Sadik Armaganb7851f92021-10-06 16:37:02 +0100149 if (!(modelOptions->SaveCachedNetwork()))
Sadik Armagandea8fb62020-11-26 10:38:11 +0000150 {
Matthew Sloyan80fbcd52021-01-07 13:28:47 +0000151 ClContextDeserializer deserializer;
Sadik Armaganb7851f92021-10-06 16:37:02 +0100152 auto cachedFd = modelOptions->GetCachedFileDescriptor();
153 if (cachedFd != -1)
154 {
155 struct stat statBuffer;
156 if (fstat(cachedFd, &statBuffer) == 0)
157 {
158 long dataSize = static_cast<long>(statBuffer.st_size);
159 if( dataSize > 0)
160 {
161 auto offset = lseek(cachedFd, 0, SEEK_CUR);
162 if (offset == 0)
163 {
164 std::vector <uint8_t> compiledContextData(static_cast<unsigned int>(dataSize));
Sadik Armagan6b9eba22021-11-01 07:27:50 +0000165 auto success = pread(cachedFd, compiledContextData.data(), compiledContextData.size(), 0);
166 if (success != -1)
167 {
168 deserializer.DeserializeFromBinary(m_CLCompileContext,
169 context,
170 device,
171 compiledContextData);
172 }
Sadik Armaganb7851f92021-10-06 16:37:02 +0100173 }
174 }
175
176 }
177 }
178
179 if (filePath != "" && fs::exists(filePath) && fs::is_regular_file(filePath))
180 {
181 // Deserialize binary file and load into m_CLCompileContext
182 deserializer.Deserialize(m_CLCompileContext, context, device, filePath);
183 }
Sadik Armagandea8fb62020-11-26 10:38:11 +0000184 }
185 }
186}
187
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000188ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager)
Sadik Armagan04a72972020-09-14 15:44:18 +0100189 : m_MemoryManager(memoryManager), m_ModelContextPtr(IBackendInternal::IBackendSpecificModelContextPtr{})
190{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000191 InitializeCLCompileContext();
Sadik Armagan04a72972020-09-14 15:44:18 +0100192}
193
194ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager,
195 const IBackendInternal::IBackendSpecificModelContextPtr& modelContextPtr)
196 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
telsoa014fcda012018-03-09 14:13:49 +0000197{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000198 InitializeCLCompileContext();
telsoa014fcda012018-03-09 14:13:49 +0000199}
200
David Monahan3fb7e102019-08-20 11:25:29 +0100201std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
202 const bool IsMemoryManaged) const
telsoa014fcda012018-03-09 14:13:49 +0000203{
Jan Eilers8eb25602020-03-09 12:13:48 +0000204 IgnoreUnused(IsMemoryManaged);
telsoa01c577f2c2018-08-31 09:22:23 +0100205 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000206 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
telsoa01c577f2c2018-08-31 09:22:23 +0100207
208 return tensorHandle;
telsoa014fcda012018-03-09 14:13:49 +0000209}
210
Francis Murtagh351d13d2018-09-24 15:01:18 +0100211std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
David Monahan3fb7e102019-08-20 11:25:29 +0100212 DataLayout dataLayout,
213 const bool IsMemoryManaged) const
Francis Murtagh351d13d2018-09-24 15:01:18 +0100214{
Jan Eilers8eb25602020-03-09 12:13:48 +0000215 IgnoreUnused(IsMemoryManaged);
Francis Murtagh351d13d2018-09-24 15:01:18 +0100216 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000217 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
Francis Murtagh351d13d2018-09-24 15:01:18 +0100218
219 return tensorHandle;
220}
221
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100222std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateSubTensorHandle(ITensorHandle& parent,
223 TensorShape const& subTensorShape,
telsoa014fcda012018-03-09 14:13:49 +0000224 unsigned int const* subTensorOrigin) const
225{
telsoa014fcda012018-03-09 14:13:49 +0000226 arm_compute::Coordinates coords;
227 arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
228
229 coords.set_num_dimensions(subTensorShape.GetNumDimensions());
230 for (unsigned int i = 0; i < subTensorShape.GetNumDimensions(); i++)
231 {
telsoa01c577f2c2018-08-31 09:22:23 +0100232 // Arm compute indexes tensor coords in reverse order.
telsoa014fcda012018-03-09 14:13:49 +0000233 unsigned int revertedIndex = subTensorShape.GetNumDimensions() - i - 1;
Matthew Sloyan171214c2020-09-09 09:07:37 +0100234 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
telsoa014fcda012018-03-09 14:13:49 +0000235 }
236
Derek Lamberti0790dce2019-04-15 18:37:35 +0100237 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.GetShape());
238 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
239 {
240 return nullptr;
241 }
242
telsoa01c577f2c2018-08-31 09:22:23 +0100243 return std::make_unique<ClSubTensorHandle>(
Jan Eilersbb446e52020-04-02 13:56:54 +0100244 PolymorphicDowncast<IClTensorHandle*>(&parent), shape, coords);
telsoa014fcda012018-03-09 14:13:49 +0000245}
246
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000247std::unique_ptr<IWorkload> ClWorkloadFactory::CreateWorkload(LayerType type,
248 const QueueDescriptor& descriptor,
249 const WorkloadInfo& info) const
250{
251 switch(type)
252 {
253 case LayerType::Activation :
254 {
255 auto activationQueueDescriptor = PolymorphicDowncast<const ActivationQueueDescriptor*>(&descriptor);
256 return MakeWorkload<ClActivationWorkload>(*activationQueueDescriptor, info, m_CLCompileContext);
257 }
258 case LayerType::Addition :
259 {
260 auto additionQueueDescriptor = PolymorphicDowncast<const AdditionQueueDescriptor*>(&descriptor);
261 return MakeWorkload<ClAdditionWorkload>(*additionQueueDescriptor, info, m_CLCompileContext);
262 }
263 case LayerType::ArgMinMax :
264 {
265 auto argMinMaxQueueDescriptor = PolymorphicDowncast<const ArgMinMaxQueueDescriptor*>(&descriptor);
Cathal Corbettea660e12022-06-08 17:36:13 +0100266 return MakeWorkload<ClArgMinMaxWorkload>(*argMinMaxQueueDescriptor, info, m_CLCompileContext);
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000267 }
Teresa Charlin94916a52022-10-19 08:48:07 +0100268 case LayerType::BatchMatMul :
269 {
270 auto batchMatMulQueueDescriptor = PolymorphicDowncast<const BatchMatMulQueueDescriptor*>(&descriptor);
271 return std::make_unique<ClBatchMatMulWorkload>(*batchMatMulQueueDescriptor, info, m_CLCompileContext);
272 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000273 case LayerType::BatchNormalization :
274 {
275 auto batchNormalizationQueueDescriptor
276 = PolymorphicDowncast<const BatchNormalizationQueueDescriptor*>(&descriptor);
277 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>
278 (*batchNormalizationQueueDescriptor, info, m_CLCompileContext);
279 }
280 case LayerType::BatchToSpaceNd :
281 {
282 auto batchToSpaceNdQueueDescriptor
283 = PolymorphicDowncast<const BatchToSpaceNdQueueDescriptor*>(&descriptor);
284 return MakeWorkload<ClBatchToSpaceNdWorkload>(*batchToSpaceNdQueueDescriptor, info, m_CLCompileContext);
285 }
286 case LayerType::Cast :
287 {
288 auto castQueueDescriptor = PolymorphicDowncast<const CastQueueDescriptor*>(&descriptor);
289 return MakeWorkload<ClCastWorkload>(*castQueueDescriptor, info, m_CLCompileContext);
290 }
291 case LayerType::ChannelShuffle :
292 {
293 auto channelShuffleQueueDescriptor
294 = PolymorphicDowncast<const ChannelShuffleQueueDescriptor*>(&descriptor);
295 return MakeWorkload<ClChannelShuffleWorkload>(*channelShuffleQueueDescriptor, info, m_CLCompileContext);
296 }
297 case LayerType::Comparison :
298 {
299 auto comparisonQueueDescriptor = PolymorphicDowncast<const ComparisonQueueDescriptor*>(&descriptor);
300 return MakeWorkload<ClComparisonWorkload>(*comparisonQueueDescriptor, info, m_CLCompileContext);
301 }
302 case LayerType::Concat :
303 {
304 auto concatQueueDescriptor = PolymorphicDowncast<const ConcatQueueDescriptor*>(&descriptor);
305 return MakeWorkload<ClConcatWorkload>(*concatQueueDescriptor, info, m_CLCompileContext);
306 }
307 case LayerType::Constant :
308 {
309 auto constantQueueDescriptor = PolymorphicDowncast<const ConstantQueueDescriptor*>(&descriptor);
310 return MakeWorkload<ClConstantWorkload>(*constantQueueDescriptor, info, m_CLCompileContext);
311 }
312 case LayerType::ConvertFp16ToFp32 :
313 {
314 auto convertFp16ToFp32QueueDescriptor
315 = PolymorphicDowncast<const ConvertFp16ToFp32QueueDescriptor*>(&descriptor);
316 return MakeWorkload<ClConvertFp16ToFp32Workload>(*convertFp16ToFp32QueueDescriptor,
317 info,
318 m_CLCompileContext);
319 }
320 case LayerType::ConvertFp32ToFp16 :
321 {
322 auto convertFp32ToFp16QueueDescriptor
323 = PolymorphicDowncast<const ConvertFp32ToFp16QueueDescriptor*>(&descriptor);
324 return MakeWorkload<ClConvertFp32ToFp16Workload>(*convertFp32ToFp16QueueDescriptor,
325 info,
326 m_CLCompileContext);
327 }
328 case LayerType::Convolution2d :
329 {
330 auto convolution2dQueueDescriptor = PolymorphicDowncast<const Convolution2dQueueDescriptor*>(&descriptor);
331
332 bool isFastMathEnabled = false;
333 if (m_ModelContextPtr)
334 {
335 if (m_ModelContextPtr.get() != nullptr)
336 {
337 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
338 if (modelOptions)
339 {
340 isFastMathEnabled = modelOptions->IsFastMathEnabled();
341 }
342 }
343 }
344 return MakeWorkload<ClConvolution2dWorkload>(*convolution2dQueueDescriptor,
345 info,
346 m_MemoryManager->GetIntraLayerManager(),
347 m_CLCompileContext,
348 isFastMathEnabled);
349 }
350 case LayerType::Convolution3d :
351 {
352 auto convolution3dQueueDescriptor = PolymorphicDowncast<const Convolution3dQueueDescriptor*>(&descriptor);
353
354 bool isFastMathEnabled = false;
355 if (m_ModelContextPtr)
356 {
357 if (m_ModelContextPtr.get() != nullptr)
358 {
359 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
360 if (modelOptions)
361 {
362 isFastMathEnabled = modelOptions->IsFastMathEnabled();
363 }
364 }
365 }
366 return MakeWorkload<ClConvolution3dWorkload>(*convolution3dQueueDescriptor,
367 info,
368 m_MemoryManager->GetIntraLayerManager(),
369 m_CLCompileContext,
370 isFastMathEnabled);
371 }
372 case LayerType::Debug :
373 {
374 auto debugQueueDescriptor = PolymorphicDowncast<const DebugQueueDescriptor*>(&descriptor);
375 return MakeWorkload<NullWorkload, NullWorkload>(*debugQueueDescriptor, info, m_CLCompileContext);
376 }
377 case LayerType::DepthToSpace :
378 {
379 auto depthToSpaceQueueDescriptor = PolymorphicDowncast<const DepthToSpaceQueueDescriptor*>(&descriptor);
380 return MakeWorkload<ClDepthToSpaceWorkload>(*depthToSpaceQueueDescriptor, info, m_CLCompileContext);
381 }
382 case LayerType::DepthwiseConvolution2d :
383 {
384 auto depthwiseConvolution2dQueueDescriptor
385 = PolymorphicDowncast<const DepthwiseConvolution2dQueueDescriptor*>(&descriptor);
386 return MakeWorkload<ClDepthwiseConvolutionWorkload>(*depthwiseConvolution2dQueueDescriptor,
387 info,
388 m_CLCompileContext);
389 }
390 case LayerType::Dequantize :
391 {
392 auto dequantizeQueueDescriptor = PolymorphicDowncast<const DequantizeQueueDescriptor*>(&descriptor);
393 return MakeWorkload<ClDequantizeWorkload>(*dequantizeQueueDescriptor, info, m_CLCompileContext);
394 }
395 case LayerType::DetectionPostProcess :
396 {
397 auto detectionPostProcessQueueDescriptor
398 = PolymorphicDowncast<const DetectionPostProcessQueueDescriptor*>(&descriptor);
399 return MakeWorkload<NullWorkload, NullWorkload>(*detectionPostProcessQueueDescriptor,
400 info,
401 m_CLCompileContext);
402 }
403 case LayerType::Division :
404 {
405 auto divisionQueueDescriptor = PolymorphicDowncast<const DivisionQueueDescriptor*>(&descriptor);
406 return std::make_unique<ClDivisionWorkload>(*divisionQueueDescriptor, info, m_CLCompileContext);
407 }
Mike Kelly3ec30772023-03-08 13:47:17 +0000408 case LayerType::ElementwiseBinary :
409 {
410 auto elementwiseBinaryQueueDescriptor
411 = PolymorphicDowncast<const ElementwiseBinaryQueueDescriptor*>(&descriptor);
412
413 switch (elementwiseBinaryQueueDescriptor->m_Parameters.m_Operation)
414 {
415 case BinaryOperation::Add:
416 {
417 AdditionQueueDescriptor additionQueueDescriptor;
418 additionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
419 additionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
420 additionQueueDescriptor.m_AdditionalInfoObject =
421 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
422 return std::make_unique<ClAdditionWorkload>(additionQueueDescriptor, info, m_CLCompileContext);
423 }
424 case BinaryOperation::Div:
425 {
426 DivisionQueueDescriptor divisionQueueDescriptor;
427 divisionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
428 divisionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
429 divisionQueueDescriptor.m_AdditionalInfoObject =
430 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
431 return std::make_unique<ClDivisionWorkload>(divisionQueueDescriptor, info, m_CLCompileContext);
432 }
433 case BinaryOperation::Maximum:
434 {
435 MaximumQueueDescriptor maximumQueueDescriptor;
436 maximumQueueDescriptor.m_Inputs = descriptor.m_Inputs;
437 maximumQueueDescriptor.m_Outputs = descriptor.m_Outputs;
438 maximumQueueDescriptor.m_AdditionalInfoObject =
439 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
440 return std::make_unique<ClMaximumWorkload>(maximumQueueDescriptor, info, m_CLCompileContext);
441 }
442 case BinaryOperation::Minimum:
443 {
444 MinimumQueueDescriptor minimumQueueDescriptor;
445 minimumQueueDescriptor.m_Inputs = descriptor.m_Inputs;
446 minimumQueueDescriptor.m_Outputs = descriptor.m_Outputs;
447 minimumQueueDescriptor.m_AdditionalInfoObject =
448 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
449 return std::make_unique<ClMinimumWorkload>(minimumQueueDescriptor, info, m_CLCompileContext);
450 }
451 case BinaryOperation::Mul:
452 {
453 MultiplicationQueueDescriptor multiplicationQueueDescriptor;
454 multiplicationQueueDescriptor.m_Inputs = descriptor.m_Inputs;
455 multiplicationQueueDescriptor.m_Outputs = descriptor.m_Outputs;
456 multiplicationQueueDescriptor.m_AdditionalInfoObject =
457 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
458 return std::make_unique<ClMultiplicationWorkload>(multiplicationQueueDescriptor,
459 info,
460 m_CLCompileContext);
461 }
John Mcloughlin34c1c382023-05-17 15:08:36 +0100462 case BinaryOperation::Power:
463 case BinaryOperation::SqDiff:
464 {
465 return std::make_unique<ClElementwiseBinaryWorkload>(*elementwiseBinaryQueueDescriptor,
466 info,
467 m_CLCompileContext);
468 }
Mike Kelly3ec30772023-03-08 13:47:17 +0000469 case BinaryOperation::Sub:
470 {
471 SubtractionQueueDescriptor subtractionQueueDescriptor;
472 subtractionQueueDescriptor.m_Inputs = descriptor.m_Inputs;
473 subtractionQueueDescriptor.m_Outputs = descriptor.m_Outputs;
474 subtractionQueueDescriptor.m_AdditionalInfoObject =
475 elementwiseBinaryQueueDescriptor->m_AdditionalInfoObject;
476 return std::make_unique<ClSubtractionWorkload>(subtractionQueueDescriptor,
477 info,
478 m_CLCompileContext);
479 }
480 default:
481 return nullptr;
482 }
483 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000484 case LayerType::ElementwiseUnary :
485 {
486 auto elementwiseUnaryQueueDescriptor
487 = PolymorphicDowncast<const ElementwiseUnaryQueueDescriptor*>(&descriptor);
488
489 switch(elementwiseUnaryQueueDescriptor->m_Parameters.m_Operation)
490 {
491 case UnaryOperation::Abs:
492 {
493 AbsQueueDescriptor absQueueDescriptor;
494 absQueueDescriptor.m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
495 absQueueDescriptor.m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
496
497 return std::make_unique<ClAbsWorkload>(absQueueDescriptor, info, m_CLCompileContext);
498 }
499 case UnaryOperation::Exp:
500 return std::make_unique<ClExpWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
501 case UnaryOperation::Log:
502 return std::make_unique<ClLogWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
503 case UnaryOperation::LogicalNot:
504 return std::make_unique<ClLogicalNotWorkload>(*elementwiseUnaryQueueDescriptor,
505 info,
506 m_CLCompileContext);
507 case UnaryOperation::Neg:
508 return std::make_unique<ClNegWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
509 case UnaryOperation::Rsqrt:
510 {
511 RsqrtQueueDescriptor rsqrtQueueDescriptor;
512 rsqrtQueueDescriptor.m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
513 rsqrtQueueDescriptor.m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
514
515 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor, info, m_CLCompileContext);
516 }
517 case UnaryOperation::Sin:
518 return std::make_unique<ClSinWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
Teresa Charlin03027232022-05-09 17:27:08 +0100519 case UnaryOperation::Sqrt:
520 return std::make_unique<ClSqrtWorkload>(*elementwiseUnaryQueueDescriptor, info, m_CLCompileContext);
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000521 default:
522 return nullptr;
523 }
524 }
525 case LayerType::Fill :
526 {
527 auto fillQueueDescriptor = PolymorphicDowncast<const FillQueueDescriptor*>(&descriptor);
528 return std::make_unique<ClFillWorkload>(*fillQueueDescriptor, info, m_CLCompileContext);
529 }
530 case LayerType::Floor :
531 {
532 auto floorQueueDescriptor = PolymorphicDowncast<const FloorQueueDescriptor*>(&descriptor);
533 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(*floorQueueDescriptor, info, m_CLCompileContext);
534 }
535 case LayerType::FullyConnected :
536 {
537 auto fullyConnectedQueueDescriptor
538 = PolymorphicDowncast<const FullyConnectedQueueDescriptor*>(&descriptor);
539 return MakeWorkload<ClFullyConnectedWorkload>(*fullyConnectedQueueDescriptor,
540 info,
541 m_MemoryManager->GetIntraLayerManager(),
542 m_CLCompileContext);
543 }
544 case LayerType::Gather :
545 {
546 auto gatherQueueDescriptor = PolymorphicDowncast<const GatherQueueDescriptor*>(&descriptor);
547 return MakeWorkload<ClGatherWorkload>(*gatherQueueDescriptor, info, m_CLCompileContext);
548 }
Teresa Charlin989e2f62022-04-27 16:26:11 +0100549 case LayerType::GatherNd :
550 {
551 auto gatherNdQueueDescriptor = PolymorphicDowncast<const GatherNdQueueDescriptor*>(&descriptor);
552 return MakeWorkload<ClGatherNdWorkload>(*gatherNdQueueDescriptor, info, m_CLCompileContext);
553 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000554 case LayerType::Input :
555 {
556 auto inputQueueDescriptor = PolymorphicDowncast<const InputQueueDescriptor*>(&descriptor);
557 return std::make_unique<CopyMemGenericWorkload>(*inputQueueDescriptor, info);
558 }
559 case LayerType::InstanceNormalization :
560 {
561 auto instanceNormalizationQueueDescriptor
562 = PolymorphicDowncast<const InstanceNormalizationQueueDescriptor*>(&descriptor);
563 return MakeWorkload<ClInstanceNormalizationWorkload>(*instanceNormalizationQueueDescriptor,
564 info,
565 m_CLCompileContext);
566 }
567 case LayerType::L2Normalization :
568 {
569 auto l2NormalizationQueueDescriptor
570 = PolymorphicDowncast<const L2NormalizationQueueDescriptor*>(&descriptor);
571 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(*l2NormalizationQueueDescriptor,
572 info,
573 m_CLCompileContext);
574 }
575 case LayerType::LogicalBinary :
576 {
577 auto logicalBinaryQueueDescriptor = PolymorphicDowncast<const LogicalBinaryQueueDescriptor*>(&descriptor);
578
579 switch(logicalBinaryQueueDescriptor->m_Parameters.m_Operation)
580 {
581 case LogicalBinaryOperation::LogicalAnd:
582 return std::make_unique<ClLogicalAndWorkload>(*logicalBinaryQueueDescriptor,
583 info,
584 m_CLCompileContext);
585 case LogicalBinaryOperation::LogicalOr:
586 return std::make_unique<ClLogicalOrWorkload>(*logicalBinaryQueueDescriptor,
587 info,
588 m_CLCompileContext);
589 default:
590 return nullptr;
591 }
592 }
593 case LayerType::LogSoftmax :
594 {
595 auto logSoftmaxQueueDescriptor = PolymorphicDowncast<const LogSoftmaxQueueDescriptor*>(&descriptor);
596
597 return MakeWorkload<ClLogSoftmaxWorkload>(*logSoftmaxQueueDescriptor,
598 info,
599 m_MemoryManager->GetIntraLayerManager(),
600 m_CLCompileContext);
601 }
602 case LayerType::Lstm :
603 {
604 auto lstmQueueDescriptor = PolymorphicDowncast<const LstmQueueDescriptor*>(&descriptor);
605 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(*lstmQueueDescriptor, info, m_CLCompileContext);
606 }
607 case LayerType::Maximum :
608 {
609 auto maximumQueueDescriptor = PolymorphicDowncast<const MaximumQueueDescriptor*>(&descriptor);
610 return MakeWorkload<ClMaximumWorkload>(*maximumQueueDescriptor, info, m_CLCompileContext);
611 }
612 case LayerType::Mean :
613 {
614 auto meanQueueDescriptor = PolymorphicDowncast<const MeanQueueDescriptor*>(&descriptor);
615 return MakeWorkload<ClMeanWorkload>(*meanQueueDescriptor, info, m_CLCompileContext);
616 }
617 case LayerType::MemCopy :
618 {
619 auto memCopyQueueDescriptor = PolymorphicDowncast<const MemCopyQueueDescriptor*>(&descriptor);
620 if (memCopyQueueDescriptor->m_Inputs.empty() || !memCopyQueueDescriptor->m_Inputs[0])
621 {
622 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload");
623 }
624 return MakeWorkload<CopyMemGenericWorkload>(*memCopyQueueDescriptor, info);
625 }
626 case LayerType::MemImport :
627 {
628 auto memImportQueueDescriptor = PolymorphicDowncast<const MemImportQueueDescriptor*>(&descriptor);
629 if (memImportQueueDescriptor->m_Inputs.empty() || !memImportQueueDescriptor->m_Inputs[0])
630 {
631 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemImport workload");
632 }
633 return std::make_unique<ImportMemGenericWorkload>(*memImportQueueDescriptor, info);
634 }
635 case LayerType::Minimum :
636 {
637 auto minimumQueueDescriptor = PolymorphicDowncast<const MinimumQueueDescriptor*>(&descriptor);
638 return MakeWorkload<ClMinimumWorkload>(*minimumQueueDescriptor, info, m_CLCompileContext);
639 }
640 case LayerType::Multiplication :
641 {
642 auto multiplicationQueueDescriptor = PolymorphicDowncast<const MultiplicationQueueDescriptor*>(&descriptor);
643 return MakeWorkload<ClMultiplicationWorkload>(*multiplicationQueueDescriptor, info, m_CLCompileContext);
644 }
645 case LayerType::Normalization :
646 {
647 auto normalizationQueueDescriptor = PolymorphicDowncast<const NormalizationQueueDescriptor*>(&descriptor);
648 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(*normalizationQueueDescriptor,
649 info,
650 m_CLCompileContext);
651 }
652 case LayerType::Output :
653 {
654 auto outputQueueDescriptor = PolymorphicDowncast<const OutputQueueDescriptor*>(&descriptor);
655 return std::make_unique<CopyMemGenericWorkload>(*outputQueueDescriptor, info);
656 }
657 case LayerType::Pad :
658 {
659 auto padQueueDescriptor = PolymorphicDowncast<const PadQueueDescriptor*>(&descriptor);
660 return MakeWorkload<ClPadWorkload>(*padQueueDescriptor, info, m_CLCompileContext);
661 }
662 case LayerType::Permute :
663 {
664 auto permuteQueueDescriptor = PolymorphicDowncast<const PermuteQueueDescriptor*>(&descriptor);
665 return MakeWorkload<ClPermuteWorkload>(*permuteQueueDescriptor, info, m_CLCompileContext);
666 }
667 case LayerType::Pooling2d :
668 {
669 auto pooling2dQueueDescriptor = PolymorphicDowncast<const Pooling2dQueueDescriptor*>(&descriptor);
670 return MakeWorkload<ClPooling2dWorkload>(*pooling2dQueueDescriptor, info, m_CLCompileContext);
671 }
Ryan OSheabab8fa92022-03-09 10:29:02 +0000672 case LayerType::Pooling3d :
673 {
674 auto pooling3dQueueDescriptor = PolymorphicDowncast<const Pooling3dQueueDescriptor*>(&descriptor);
675 return MakeWorkload<ClPooling3dWorkload>(*pooling3dQueueDescriptor, info, m_CLCompileContext);
676 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000677 case LayerType::PreCompiled :
678 {
679 auto preCompiledQueueDescriptor = PolymorphicDowncast<const PreCompiledQueueDescriptor*>(&descriptor);
680 return MakeWorkload<NullWorkload, NullWorkload>(*preCompiledQueueDescriptor, info, m_CLCompileContext);
681 }
682 case LayerType::Prelu :
683 {
684 auto preluQueueDescriptor = PolymorphicDowncast<const PreluQueueDescriptor*>(&descriptor);
685 return MakeWorkload<ClPreluWorkload>(*preluQueueDescriptor, info, m_CLCompileContext);
686 }
687 case LayerType::QLstm :
688 {
689 auto qLstmQueueDescriptor = PolymorphicDowncast<const QLstmQueueDescriptor*>(&descriptor);
690 return std::make_unique<ClQLstmWorkload>(*qLstmQueueDescriptor, info, m_CLCompileContext);
691 }
692 case LayerType::Quantize :
693 {
694 auto quantizeQueueDescriptor = PolymorphicDowncast<const QuantizeQueueDescriptor*>(&descriptor);
695 return MakeWorkload<ClQuantizeWorkload>(*quantizeQueueDescriptor, info, m_CLCompileContext);
696 }
697 case LayerType::QuantizedLstm :
698 {
699 auto quantizedLstmQueueDescriptor = PolymorphicDowncast<const QuantizedLstmQueueDescriptor*>(&descriptor);
700 return MakeWorkload<ClQuantizedLstmWorkload>(*quantizedLstmQueueDescriptor, info, m_CLCompileContext);
701 }
702 case LayerType::Rank :
703 {
704 auto rankQueueDescriptor = PolymorphicDowncast<const RankQueueDescriptor*>(&descriptor);
705 return std::make_unique<ClRankWorkload>(*rankQueueDescriptor, info);
706 }
707 case LayerType::Reduce :
708 {
709 auto reduceQueueDescriptor = PolymorphicDowncast<const ReduceQueueDescriptor*>(&descriptor);
710 return std::make_unique<ClReduceWorkload>(*reduceQueueDescriptor, info);
711 }
712 case LayerType::Reshape :
713 {
714 auto reshapeQueueDescriptor = PolymorphicDowncast<const ReshapeQueueDescriptor*>(&descriptor);
715 return MakeWorkload<ClReshapeWorkload>(*reshapeQueueDescriptor, info, m_CLCompileContext);
716 }
717 case LayerType::Resize :
718 {
719 auto resizeQueueDescriptor = PolymorphicDowncast<const ResizeQueueDescriptor*>(&descriptor);
720 return MakeWorkload<ClResizeWorkload>(*resizeQueueDescriptor, info, m_CLCompileContext);
721 }
722 case LayerType::Slice :
723 {
724 auto sliceQueueDescriptor = PolymorphicDowncast<const SliceQueueDescriptor*>(&descriptor);
725 return MakeWorkload<ClSliceWorkload>(*sliceQueueDescriptor, info, m_CLCompileContext);
726 }
727 case LayerType::Softmax :
728 {
729 auto softmaxQueueDescriptor = PolymorphicDowncast<const SoftmaxQueueDescriptor*>(&descriptor);
730 return std::make_unique<ClSoftmaxWorkload>(*softmaxQueueDescriptor,
731 info,
732 m_MemoryManager->GetIntraLayerManager(),
733 m_CLCompileContext);
734 }
735 case LayerType::SpaceToBatchNd :
736 {
737 auto spaceToBatchNdQueueDescriptor
738 = PolymorphicDowncast<const SpaceToBatchNdQueueDescriptor*>(&descriptor);
739 return MakeWorkload<ClSpaceToBatchNdWorkload>(*spaceToBatchNdQueueDescriptor, info, m_CLCompileContext);
740 }
741 case LayerType::SpaceToDepth :
742 {
743 auto spaceToDepthQueueDescriptor = PolymorphicDowncast<const SpaceToDepthQueueDescriptor*>(&descriptor);
744 return MakeWorkload<ClSpaceToDepthWorkload>(*spaceToDepthQueueDescriptor, info, m_CLCompileContext);
745 }
746 case LayerType::Splitter :
747 {
748 auto splitterQueueDescriptor = PolymorphicDowncast<const SplitterQueueDescriptor*>(&descriptor);
749 return MakeWorkload<ClSplitterWorkload>(*splitterQueueDescriptor, info, m_CLCompileContext);
750 }
751 case LayerType::Stack :
752 {
753 auto stackQueueDescriptor = PolymorphicDowncast<const StackQueueDescriptor*>(&descriptor);
754 return MakeWorkload<ClStackWorkload>(*stackQueueDescriptor, info, m_CLCompileContext);
755 }
756 case LayerType::StridedSlice :
757 {
758 auto stridedSliceQueueDescriptor = PolymorphicDowncast<const StridedSliceQueueDescriptor*>(&descriptor);
759 return MakeWorkload<ClStridedSliceWorkload>(*stridedSliceQueueDescriptor, info, m_CLCompileContext);
760 }
761 case LayerType::Subtraction :
762 {
763 auto subtractionQueueDescriptor = PolymorphicDowncast<const SubtractionQueueDescriptor*>(&descriptor);
764 return MakeWorkload<ClSubtractionWorkload>(*subtractionQueueDescriptor, info, m_CLCompileContext);
765 }
766 case LayerType::Transpose :
767 {
768 auto transposeQueueDescriptor = PolymorphicDowncast<const TransposeQueueDescriptor*>(&descriptor);
769 return MakeWorkload<ClTransposeWorkload>(*transposeQueueDescriptor, info, m_CLCompileContext);
770 }
771 case LayerType::TransposeConvolution2d :
772 {
773 auto transposeConvolution2dQueueDescriptor
774 = PolymorphicDowncast<const TransposeConvolution2dQueueDescriptor*>(&descriptor);
775 return MakeWorkload<ClTransposeConvolution2dWorkload>(*transposeConvolution2dQueueDescriptor,
776 info,
777 m_MemoryManager->GetIntraLayerManager(),
778 m_CLCompileContext);
779 }
Cathal Corbett4952a3e2022-03-03 15:14:18 +0000780 case LayerType::UnidirectionalSequenceLstm :
781 {
782 auto desc = PolymorphicDowncast<const UnidirectionalSequenceLstmQueueDescriptor*>(&descriptor);
783 return MakeWorkloadHelper<ClUnidirectionalSequenceLstmFloatWorkload, NullWorkload>(*desc,
784 info,
785 m_CLCompileContext);
786 }
Teresa Charlin611c7fb2022-01-07 09:47:29 +0000787 default:
788 return nullptr;
789 }
790}
791
telsoa014fcda012018-03-09 14:13:49 +0000792std::unique_ptr<IWorkload> ClWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100793 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000794{
Sadik Armagane9444752020-12-02 11:28:58 +0000795 return MakeWorkload<ClActivationWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000796}
797
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100798std::unique_ptr<IWorkload> ClWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
799 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000800{
Sadik Armagane9444752020-12-02 11:28:58 +0000801 return MakeWorkload<ClAdditionWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000802}
803
James Conroy2dc05722019-09-19 17:00:31 +0100804std::unique_ptr<IWorkload> ClWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
805 const WorkloadInfo& info) const
806{
Sadik Armagane9444752020-12-02 11:28:58 +0000807 return std::make_unique<ClArgMinMaxWorkload>(descriptor, info, m_CLCompileContext);
James Conroy2dc05722019-09-19 17:00:31 +0100808}
809
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100810std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchNormalization(
811 const BatchNormalizationQueueDescriptor& descriptor,
812 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000813{
Sadik Armagane9444752020-12-02 11:28:58 +0000814 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000815}
816
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100817std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
telsoa014fcda012018-03-09 14:13:49 +0000818 const WorkloadInfo& info) const
819{
Sadik Armagane9444752020-12-02 11:28:58 +0000820 return MakeWorkload<ClBatchToSpaceNdWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000821}
822
Sadik Armaganf40d6d42021-04-22 09:12:11 +0100823std::unique_ptr<IWorkload> ClWorkloadFactory::CreateCast(const CastQueueDescriptor& descriptor,
824 const WorkloadInfo& info) const
825{
826 return MakeWorkload<ClCastWorkload>(descriptor, info, m_CLCompileContext);
827}
828
Teresa Charlin1222dbd2021-09-02 13:58:52 +0100829std::unique_ptr<IWorkload> ClWorkloadFactory::CreateChannelShuffle(const ChannelShuffleQueueDescriptor& descriptor,
830 const WorkloadInfo& info) const
831{
832 return MakeWorkload<ClChannelShuffleWorkload>(descriptor, info, m_CLCompileContext);
833}
834
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100835std::unique_ptr<IWorkload> ClWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
836 const WorkloadInfo& info) const
837{
Sadik Armagane9444752020-12-02 11:28:58 +0000838 return MakeWorkload<ClComparisonWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100839}
840
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100841std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
842 const WorkloadInfo& info) const
Jim Flynn4ed6c832019-05-20 11:02:46 +0100843{
Sadik Armagane9444752020-12-02 11:28:58 +0000844 return MakeWorkload<ClConcatWorkload>(descriptor, info, m_CLCompileContext);
Jim Flynn4ed6c832019-05-20 11:02:46 +0100845}
846
telsoa014fcda012018-03-09 14:13:49 +0000847std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100848 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000849{
Sadik Armagane9444752020-12-02 11:28:58 +0000850 return MakeWorkload<ClConstantWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000851}
852
telsoa01c577f2c2018-08-31 09:22:23 +0100853std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp16ToFp32(
854 const ConvertFp16ToFp32QueueDescriptor& descriptor,
855 const WorkloadInfo& info) const
856{
Sadik Armagane9444752020-12-02 11:28:58 +0000857 return MakeWorkload<ClConvertFp16ToFp32Workload>(descriptor, info, m_CLCompileContext);
telsoa01c577f2c2018-08-31 09:22:23 +0100858}
859
860std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp32ToFp16(
861 const ConvertFp32ToFp16QueueDescriptor& descriptor,
862 const WorkloadInfo& info) const
863{
Sadik Armagane9444752020-12-02 11:28:58 +0000864 return MakeWorkload<ClConvertFp32ToFp16Workload>(descriptor, info, m_CLCompileContext);
telsoa01c577f2c2018-08-31 09:22:23 +0100865}
866
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100867std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
868 const WorkloadInfo& info) const
869{
Sadik Armagan04a72972020-09-14 15:44:18 +0100870 bool isFastMathEnabled = false;
871 if (m_ModelContextPtr)
872 {
873 if (m_ModelContextPtr.get() != nullptr)
874 {
875 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
876 if (modelOptions)
877 {
878 isFastMathEnabled = modelOptions->IsFastMathEnabled();
879 }
880 }
881 }
882 return MakeWorkload<ClConvolution2dWorkload>(descriptor,
883 info,
884 m_MemoryManager->GetIntraLayerManager(),
Sadik Armagane9444752020-12-02 11:28:58 +0000885 m_CLCompileContext,
Sadik Armagan04a72972020-09-14 15:44:18 +0100886 isFastMathEnabled);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100887}
888
Teresa Charlin615ad6c2021-10-26 12:22:20 +0100889std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvolution3d(const Convolution3dQueueDescriptor& descriptor,
890 const WorkloadInfo& info) const
891{
892 bool isFastMathEnabled = false;
893 if (m_ModelContextPtr)
894 {
895 if (m_ModelContextPtr.get() != nullptr)
896 {
897 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
898 if (modelOptions)
899 {
900 isFastMathEnabled = modelOptions->IsFastMathEnabled();
901 }
902 }
903 }
904 return MakeWorkload<ClConvolution3dWorkload>(descriptor,
905 info,
906 m_MemoryManager->GetIntraLayerManager(),
907 m_CLCompileContext,
908 isFastMathEnabled);
909}
910
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100911std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
912 const WorkloadInfo& info) const
913{
Sadik Armagane9444752020-12-02 11:28:58 +0000914 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100915}
916
917std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
918 const WorkloadInfo& info) const
919{
Sadik Armagane9444752020-12-02 11:28:58 +0000920 return MakeWorkload<ClDepthToSpaceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100921}
922
923std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthwiseConvolution2d(
924 const DepthwiseConvolution2dQueueDescriptor& descriptor,
925 const WorkloadInfo& info) const
926{
Sadik Armagane9444752020-12-02 11:28:58 +0000927 return MakeWorkload<ClDepthwiseConvolutionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100928}
929
930std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
931 const WorkloadInfo& info) const
932{
Sadik Armagane9444752020-12-02 11:28:58 +0000933 return MakeWorkload<ClDequantizeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100934}
935
936std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDetectionPostProcess(
937 const DetectionPostProcessQueueDescriptor& descriptor,
938 const WorkloadInfo& info) const
939{
Sadik Armagane9444752020-12-02 11:28:58 +0000940 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100941}
942
943std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
944 const WorkloadInfo& info) const
945{
Teresa Charline11e63d2021-04-21 12:56:45 +0100946 return std::make_unique<ClDivisionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100947}
948
josh minor4a3c6102020-01-06 16:40:46 -0600949std::unique_ptr<IWorkload> ClWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
950 const WorkloadInfo& info) const
951{
Sadik Armagan9fabf432020-05-27 13:40:58 +0100952 switch(descriptor.m_Parameters.m_Operation)
josh minor4a3c6102020-01-06 16:40:46 -0600953 {
Sadik Armagan9fabf432020-05-27 13:40:58 +0100954 case UnaryOperation::Abs:
James Conroyfe3ec942020-11-18 14:20:53 +0000955 {
956 AbsQueueDescriptor absQueueDescriptor;
957 absQueueDescriptor.m_Inputs = descriptor.m_Inputs;
958 absQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600959
Sadik Armagane9444752020-12-02 11:28:58 +0000960 return std::make_unique<ClAbsWorkload>(absQueueDescriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000961 }
Sadik Armagan9fabf432020-05-27 13:40:58 +0100962 case UnaryOperation::Exp:
Sadik Armagane9444752020-12-02 11:28:58 +0000963 return std::make_unique<ClExpWorkload>(descriptor, info, m_CLCompileContext);
Teresa Charlin50de4fa2021-05-31 18:47:33 +0100964 case UnaryOperation::Log:
965 return std::make_unique<ClLogWorkload>(descriptor, info, m_CLCompileContext);
966 case UnaryOperation::LogicalNot:
967 return std::make_unique<ClLogicalNotWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100968 case UnaryOperation::Neg:
Sadik Armagane9444752020-12-02 11:28:58 +0000969 return std::make_unique<ClNegWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100970 case UnaryOperation::Rsqrt:
James Conroyfe3ec942020-11-18 14:20:53 +0000971 {
972 RsqrtQueueDescriptor rsqrtQueueDescriptor;
973 rsqrtQueueDescriptor.m_Inputs = descriptor.m_Inputs;
974 rsqrtQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600975
Sadik Armagane9444752020-12-02 11:28:58 +0000976 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000977 }
Teresa Charlin50de4fa2021-05-31 18:47:33 +0100978 case UnaryOperation::Sin:
979 return std::make_unique<ClSinWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100980 default:
981 return nullptr;
josh minor4a3c6102020-01-06 16:40:46 -0600982 }
josh minor4a3c6102020-01-06 16:40:46 -0600983}
984
Sadik Armagan66aecb02020-06-24 11:42:20 +0100985std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
986 const WorkloadInfo& info) const
987{
Sadik Armagane9444752020-12-02 11:28:58 +0000988 return std::make_unique<ClFillWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan66aecb02020-06-24 11:42:20 +0100989}
990
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100991std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
992 const WorkloadInfo& info) const
993{
Sadik Armagane9444752020-12-02 11:28:58 +0000994 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100995}
996
997std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFullyConnected(const FullyConnectedQueueDescriptor& descriptor,
998 const WorkloadInfo& info) const
999{
Sadik Armagane9444752020-12-02 11:28:58 +00001000 return MakeWorkload<ClFullyConnectedWorkload>(descriptor,
1001 info,
1002 m_MemoryManager->GetIntraLayerManager(),
1003 m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001004}
1005
1006std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
1007 const WorkloadInfo& info) const
1008{
Sadik Armagane9444752020-12-02 11:28:58 +00001009 return MakeWorkload<ClGatherWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001010}
1011
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001012std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
1013 const WorkloadInfo& info) const
1014{
1015 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
1016}
1017
1018std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInstanceNormalization(
1019 const InstanceNormalizationQueueDescriptor& descriptor,
1020 const WorkloadInfo& info) const
1021{
Sadik Armagane9444752020-12-02 11:28:58 +00001022 return MakeWorkload<ClInstanceNormalizationWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001023}
1024
1025std::unique_ptr<IWorkload> ClWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
1026 const WorkloadInfo& info) const
1027{
Sadik Armagane9444752020-12-02 11:28:58 +00001028 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001029}
1030
James Conroyfe3ec942020-11-18 14:20:53 +00001031std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
1032 const WorkloadInfo& info) const
1033{
1034 switch(descriptor.m_Parameters.m_Operation)
1035 {
1036 case LogicalBinaryOperation::LogicalAnd:
Sadik Armagane9444752020-12-02 11:28:58 +00001037 return std::make_unique<ClLogicalAndWorkload>(descriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +00001038 case LogicalBinaryOperation::LogicalOr:
Sadik Armagane9444752020-12-02 11:28:58 +00001039 return std::make_unique<ClLogicalOrWorkload>(descriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +00001040 default:
1041 return nullptr;
1042 }
1043}
1044
Teresa Charlin8398edc2020-07-20 14:23:02 +01001045std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
1046 const WorkloadInfo& info) const
1047{
Sadik Armagane9444752020-12-02 11:28:58 +00001048 return MakeWorkload<ClLogSoftmaxWorkload>(descriptor,
1049 info,
1050 m_MemoryManager->GetIntraLayerManager(),
1051 m_CLCompileContext);
Teresa Charlin8398edc2020-07-20 14:23:02 +01001052}
1053
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001054std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
1055 const WorkloadInfo& info) const
1056{
Sadik Armagane9444752020-12-02 11:28:58 +00001057 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001058}
1059
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +00001060std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
1061 const WorkloadInfo& info) const
1062{
Sadik Armagane9444752020-12-02 11:28:58 +00001063 return MakeWorkload<ClMaximumWorkload>(descriptor, info, m_CLCompileContext);
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +00001064}
1065
narpra01a6bf9122018-09-10 09:50:09 +01001066std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
1067 const WorkloadInfo& info) const
1068{
Sadik Armagane9444752020-12-02 11:28:58 +00001069 return MakeWorkload<ClMeanWorkload>(descriptor, info, m_CLCompileContext);
narpra01a6bf9122018-09-10 09:50:09 +01001070}
1071
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001072std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
1073 const WorkloadInfo& info) const
jimfly012c9322a2018-09-19 10:59:49 +01001074{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001075 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
1076 {
1077 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload");
1078 }
1079
1080 return MakeWorkload<CopyMemGenericWorkload>(descriptor, info);
jimfly012c9322a2018-09-19 10:59:49 +01001081}
1082
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001083std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
1084 const WorkloadInfo& info) const
FrancisMurtagh20995952018-12-17 12:11:36 +00001085{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001086 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
1087 {
1088 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemImport workload");
1089 }
1090
1091 return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
FrancisMurtagh20995952018-12-17 12:11:36 +00001092}
1093
kevmay0190539692018-11-29 08:40:19 +00001094std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
1095 const WorkloadInfo& info) const
1096{
Sadik Armagane9444752020-12-02 11:28:58 +00001097 return MakeWorkload<ClMinimumWorkload>(descriptor, info, m_CLCompileContext);
kevmay0190539692018-11-29 08:40:19 +00001098}
1099
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001100std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
1101 const WorkloadInfo& info) const
1102{
Sadik Armagane9444752020-12-02 11:28:58 +00001103 return MakeWorkload<ClMultiplicationWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001104}
1105
1106std::unique_ptr<IWorkload> ClWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
1107 const WorkloadInfo& info) const
1108{
Sadik Armagane9444752020-12-02 11:28:58 +00001109 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001110}
1111
1112std::unique_ptr<IWorkload> ClWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
1113 const WorkloadInfo& info) const
1114{
1115 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
1116}
1117
1118std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
1119 const WorkloadInfo& info) const
1120{
Sadik Armagane9444752020-12-02 11:28:58 +00001121 return MakeWorkload<ClPadWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001122}
1123
1124std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
Matteo Martincigh59a950c2018-12-13 12:48:25 +00001125 const WorkloadInfo& info) const
1126{
Sadik Armagane9444752020-12-02 11:28:58 +00001127 return MakeWorkload<ClPermuteWorkload>(descriptor, info, m_CLCompileContext);
Matteo Martincigh59a950c2018-12-13 12:48:25 +00001128}
1129
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001130std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
1131 const WorkloadInfo& info) const
Nattapat Chaimanowonga9a1cf12018-12-03 16:06:49 +00001132{
Sadik Armagane9444752020-12-02 11:28:58 +00001133 return MakeWorkload<ClPooling2dWorkload>(descriptor, info, m_CLCompileContext);
Mohamed Nour Abouelseouda1d3c6a2018-12-27 12:39:16 +00001134}
1135
Matteo Martincigh49124022019-01-11 13:25:59 +00001136std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor,
1137 const WorkloadInfo& info) const
1138{
Sadik Armagane9444752020-12-02 11:28:58 +00001139 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Matteo Martincigh49124022019-01-11 13:25:59 +00001140}
1141
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001142std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePrelu(const PreluQueueDescriptor &descriptor,
1143 const WorkloadInfo &info) const
narpra01b89b05f2019-01-16 09:53:09 +00001144{
Sadik Armagane9444752020-12-02 11:28:58 +00001145 return MakeWorkload<ClPreluWorkload>(descriptor, info, m_CLCompileContext);
narpra01b89b05f2019-01-16 09:53:09 +00001146}
1147
Ryan OShea2323af42020-05-13 16:36:19 +01001148std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
1149 const WorkloadInfo& info) const
1150{
Sadik Armagane9444752020-12-02 11:28:58 +00001151 return std::make_unique<ClQLstmWorkload>(descriptor, info, m_CLCompileContext);
Ryan OShea2323af42020-05-13 16:36:19 +01001152}
1153
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001154std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
1155 const WorkloadInfo& info) const
Aron Virginas-Tar7a3e2fe2019-06-27 18:54:47 +01001156{
Sadik Armagane9444752020-12-02 11:28:58 +00001157 return MakeWorkload<ClQuantizeWorkload>(descriptor, info, m_CLCompileContext);
James Conroyd2aa85e2019-07-01 17:12:40 +01001158}
1159
Ferran Balaguer737d9ff2019-08-01 09:58:08 +01001160std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& descriptor,
1161 const WorkloadInfo& info) const
1162{
Sadik Armagane9444752020-12-02 11:28:58 +00001163 return MakeWorkload<ClQuantizedLstmWorkload>(descriptor, info, m_CLCompileContext);
Ferran Balaguer737d9ff2019-08-01 09:58:08 +01001164}
1165
David Monahanc11ba462020-12-03 11:09:46 +00001166std::unique_ptr<IWorkload> ClWorkloadFactory::CreateRank(const RankQueueDescriptor& descriptor,
1167 const WorkloadInfo& info) const
1168{
1169 return std::make_unique<ClRankWorkload>(descriptor, info);
1170}
1171
Sadik Armagana2747482021-02-09 10:28:54 +00001172std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReduce(const ReduceQueueDescriptor& descriptor,
1173 const WorkloadInfo& info) const
1174{
1175 return std::make_unique<ClReduceWorkload>(descriptor, info);
1176}
1177
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001178std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
1179 const WorkloadInfo& info) const
1180{
Sadik Armagane9444752020-12-02 11:28:58 +00001181 return MakeWorkload<ClReshapeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001182}
1183
1184std::unique_ptr<IWorkload> ClWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
1185 const WorkloadInfo& info) const
1186{
Sadik Armagane9444752020-12-02 11:28:58 +00001187 return MakeWorkload<ClResizeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001188}
1189
Aron Virginas-Tar94c4fef2019-11-25 15:37:08 +00001190std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
1191 const WorkloadInfo& info) const
1192{
Sadik Armagane9444752020-12-02 11:28:58 +00001193 return MakeWorkload<ClSliceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar94c4fef2019-11-25 15:37:08 +00001194}
1195
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001196std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
1197 const WorkloadInfo& info) const
1198{
Sadik Armagane9444752020-12-02 11:28:58 +00001199 return std::make_unique<ClSoftmaxWorkload>(descriptor,
1200 info,
1201 m_MemoryManager->GetIntraLayerManager(),
1202 m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001203}
1204
1205std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
1206 const WorkloadInfo& info) const
1207{
Sadik Armagane9444752020-12-02 11:28:58 +00001208 return MakeWorkload<ClSpaceToBatchNdWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001209}
1210
1211std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
1212 const WorkloadInfo& info) const
1213{
Sadik Armagane9444752020-12-02 11:28:58 +00001214 return MakeWorkload<ClSpaceToDepthWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001215}
1216
1217std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
1218 const WorkloadInfo& info) const
1219{
Sadik Armagane9444752020-12-02 11:28:58 +00001220 return MakeWorkload<ClSplitterWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001221}
1222
Matthew Jacksond5166102019-07-31 14:06:28 +01001223std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
1224 const WorkloadInfo& info) const
1225{
Sadik Armagane9444752020-12-02 11:28:58 +00001226 return MakeWorkload<ClStackWorkload>(descriptor, info, m_CLCompileContext);
Matthew Jacksond5166102019-07-31 14:06:28 +01001227}
1228
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001229std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
Aron Virginas-Tarb2801962019-09-30 11:24:53 +01001230 const WorkloadInfo& info) const
1231{
Sadik Armagane9444752020-12-02 11:28:58 +00001232 return MakeWorkload<ClStridedSliceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001233}
1234
1235std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
1236 const WorkloadInfo& info) const
1237{
Sadik Armagane9444752020-12-02 11:28:58 +00001238 return MakeWorkload<ClSubtractionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001239}
1240
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001241std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
1242 const WorkloadInfo& info) const
1243{
Sadik Armagane9444752020-12-02 11:28:58 +00001244 return MakeWorkload<ClTransposeWorkload>(descriptor, info, m_CLCompileContext);
Mike Kellyc9ea45a2020-02-28 18:11:58 +00001245}
1246
Aron Virginas-Tar8168f402019-10-04 13:10:16 +01001247std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTransposeConvolution2d(
1248 const TransposeConvolution2dQueueDescriptor& descriptor,
1249 const WorkloadInfo& info) const
1250{
Sadik Armagane9444752020-12-02 11:28:58 +00001251 return MakeWorkload<ClTransposeConvolution2dWorkload>(descriptor,
1252 info,
1253 m_MemoryManager->GetIntraLayerManager(),
1254 m_CLCompileContext);
Aron Virginas-Tarb2801962019-09-30 11:24:53 +01001255}
1256
telsoa014fcda012018-03-09 14:13:49 +00001257} // namespace armnn