blob: 3400799f4583f7d6d2df9f341b7ca7d7728c6b68 [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
Teresa Charlin8398edc2020-07-20 14:23:02 +01002// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#include "ClWorkloadFactory.hpp"
David Beck79141b92018-10-23 16:09:36 +01006#include "ClBackendId.hpp"
Sadik Armagan04a72972020-09-14 15:44:18 +01007#include "ClBackendModelContext.hpp"
Matthew Sloyan80fbcd52021-01-07 13:28:47 +00008#include "ClContextDeserializer.hpp"
9#include "ClContextSerializer.hpp"
telsoa014fcda012018-03-09 14:13:49 +000010
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000011#include <Layer.hpp>
12
David Beck0dbe0ee2018-09-24 15:59:27 +010013#include <armnn/Exceptions.hpp>
14#include <armnn/Utils.hpp>
Jan Eilers8eb25602020-03-09 12:13:48 +000015#include <armnn/utility/IgnoreUnused.hpp>
Matthew Sloyan171214c2020-09-09 09:07:37 +010016#include <armnn/utility/NumericCast.hpp>
Jan Eilersbb446e52020-04-02 13:56:54 +010017#include <armnn/utility/PolymorphicDowncast.hpp>
telsoa014fcda012018-03-09 14:13:49 +000018
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000019#include <backendsCommon/MakeWorkloadHelper.hpp>
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000020#include <backendsCommon/MemCopyWorkload.hpp>
Derek Lambertif674aa02019-08-01 15:56:25 +010021#include <backendsCommon/MemImportWorkload.hpp>
James Conroy1f58f032021-04-27 17:13:27 +010022#include <backendsCommon/TensorHandle.hpp>
telsoa01c577f2c2018-08-31 09:22:23 +010023
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +000024#include <cl/ClTensorHandle.hpp>
25#include <cl/workloads/ClWorkloads.hpp>
26#include <cl/workloads/ClWorkloadUtils.hpp>
telsoa014fcda012018-03-09 14:13:49 +000027
Aron Virginas-Tar5caf9072018-11-14 18:35:18 +000028#include <arm_compute/core/CL/CLKernelLibrary.h>
29#include <arm_compute/runtime/CL/CLBufferAllocator.h>
30#include <arm_compute/runtime/CL/CLScheduler.h>
telsoa014fcda012018-03-09 14:13:49 +000031
Rob Hughes9542f902021-07-14 09:48:54 +010032#include <armnnUtils/Filesystem.hpp>
Matthew Sloyan80fbcd52021-01-07 13:28:47 +000033#include <fstream>
Sadik Armagandea8fb62020-11-26 10:38:11 +000034
telsoa014fcda012018-03-09 14:13:49 +000035namespace armnn
36{
37
David Beck79141b92018-10-23 16:09:36 +010038namespace
39{
40static const BackendId s_Id{ClBackendId()};
41}
42
telsoa01c577f2c2018-08-31 09:22:23 +010043bool ClWorkloadFactory::IsLayerSupported(const Layer& layer,
David Beck29c75de2018-10-23 13:35:58 +010044 Optional<DataType> dataType,
telsoa01c577f2c2018-08-31 09:22:23 +010045 std::string& outReasonIfUnsupported)
telsoa014fcda012018-03-09 14:13:49 +000046{
David Beck79141b92018-10-23 16:09:36 +010047 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
48}
49
Sadik Armagan04a72972020-09-14 15:44:18 +010050bool ClWorkloadFactory::IsLayerSupported(const IConnectableLayer& layer,
51 Optional<DataType> dataType,
52 std::string& outReasonIfUnsupported,
53 const ModelOptions& modelOptions)
54{
55 return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
56}
57
David Beck79141b92018-10-23 16:09:36 +010058const BackendId& ClWorkloadFactory::GetBackendId() const
59{
60 return s_Id;
telsoa014fcda012018-03-09 14:13:49 +000061}
62
Sadik Armagandea8fb62020-11-26 10:38:11 +000063void ClWorkloadFactory::AfterWorkloadsCreated()
64{
65 if(m_ModelContextPtr)
66 {
67 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
68 if (modelOptions->SaveCachedNetwork())
69 {
70 // Save map to a filepath provided in ModelOptions
71 auto filePath = modelOptions->GetCachedNetworkFilePath();
72 if (filePath != "" && fs::exists(filePath) && fs::is_regular_file(filePath))
73 {
Matthew Sloyan80fbcd52021-01-07 13:28:47 +000074 // Serialize ClContext to the file specified
75 ClContextSerializer serializer;
76 serializer.Serialize(m_CLCompileContext);
77 std::ofstream file(filePath, std::ios::out | std::ios::binary);
78 serializer.SaveSerializedToStream(file);
Sadik Armagandea8fb62020-11-26 10:38:11 +000079 }
80 }
81 }
82}
83
Aron Virginas-Tara8e06ed2018-10-19 16:46:15 +010084template <typename FloatWorkload, typename Uint8Workload, typename QueueDescriptorType, typename... Args>
85std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
86 const WorkloadInfo& info,
87 Args&&... args)
88{
89 try
90 {
91 return MakeWorkloadHelper<FloatWorkload, Uint8Workload>(descriptor, info, std::forward<Args>(args)...);
92 }
93 catch (const cl::Error& clError)
94 {
95 throw WrapClError(clError, CHECK_LOCATION());
96 }
97}
98
99template <typename Workload, typename QueueDescriptorType, typename... Args>
100std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
101 const WorkloadInfo& info,
102 Args&&... args)
103{
104 try
105 {
106 return std::make_unique<Workload>(descriptor, info, std::forward<Args>(args)...);
107 }
108 catch (const cl::Error& clError)
109 {
110 throw WrapClError(clError, CHECK_LOCATION());
111 }
112}
113
Sadik Armagandea8fb62020-11-26 10:38:11 +0000114void ClWorkloadFactory::InitializeCLCompileContext()
115{
116 // Initialize our m_CLCompileContext using default device and context
Sadik Armagane9444752020-12-02 11:28:58 +0000117 auto context = arm_compute::CLKernelLibrary::get().context();
118 auto device = arm_compute::CLKernelLibrary::get().get_device();
Sadik Armagandea8fb62020-11-26 10:38:11 +0000119 m_CLCompileContext = arm_compute::CLCompileContext(context, device);
120
121 if (m_ModelContextPtr)
122 {
123 // Load saved programs if the user has set a filepath
124 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
125 auto filePath = modelOptions->GetCachedNetworkFilePath();
126 if (filePath != ""
127 && fs::exists(filePath)
128 && fs::is_regular_file(filePath)
129 && !(modelOptions->SaveCachedNetwork()))
130 {
Matthew Sloyan80fbcd52021-01-07 13:28:47 +0000131 // Deserialize binary file and load into m_CLCompileContext
132 ClContextDeserializer deserializer;
133 deserializer.Deserialize(m_CLCompileContext, context, device, filePath);
Sadik Armagandea8fb62020-11-26 10:38:11 +0000134 }
135 }
136}
137
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000138ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager)
Sadik Armagan04a72972020-09-14 15:44:18 +0100139 : m_MemoryManager(memoryManager), m_ModelContextPtr(IBackendInternal::IBackendSpecificModelContextPtr{})
140{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000141 InitializeCLCompileContext();
Sadik Armagan04a72972020-09-14 15:44:18 +0100142}
143
144ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& memoryManager,
145 const IBackendInternal::IBackendSpecificModelContextPtr& modelContextPtr)
146 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
telsoa014fcda012018-03-09 14:13:49 +0000147{
Sadik Armagandea8fb62020-11-26 10:38:11 +0000148 InitializeCLCompileContext();
telsoa014fcda012018-03-09 14:13:49 +0000149}
150
David Monahan3fb7e102019-08-20 11:25:29 +0100151std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
152 const bool IsMemoryManaged) const
telsoa014fcda012018-03-09 14:13:49 +0000153{
Jan Eilers8eb25602020-03-09 12:13:48 +0000154 IgnoreUnused(IsMemoryManaged);
telsoa01c577f2c2018-08-31 09:22:23 +0100155 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000156 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
telsoa01c577f2c2018-08-31 09:22:23 +0100157
158 return tensorHandle;
telsoa014fcda012018-03-09 14:13:49 +0000159}
160
Francis Murtagh351d13d2018-09-24 15:01:18 +0100161std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
David Monahan3fb7e102019-08-20 11:25:29 +0100162 DataLayout dataLayout,
163 const bool IsMemoryManaged) const
Francis Murtagh351d13d2018-09-24 15:01:18 +0100164{
Jan Eilers8eb25602020-03-09 12:13:48 +0000165 IgnoreUnused(IsMemoryManaged);
Francis Murtagh351d13d2018-09-24 15:01:18 +0100166 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout);
Aron Virginas-Tar56055192018-11-12 18:10:43 +0000167 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
Francis Murtagh351d13d2018-09-24 15:01:18 +0100168
169 return tensorHandle;
170}
171
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100172std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateSubTensorHandle(ITensorHandle& parent,
173 TensorShape const& subTensorShape,
telsoa014fcda012018-03-09 14:13:49 +0000174 unsigned int const* subTensorOrigin) const
175{
telsoa014fcda012018-03-09 14:13:49 +0000176 arm_compute::Coordinates coords;
177 arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
178
179 coords.set_num_dimensions(subTensorShape.GetNumDimensions());
180 for (unsigned int i = 0; i < subTensorShape.GetNumDimensions(); i++)
181 {
telsoa01c577f2c2018-08-31 09:22:23 +0100182 // Arm compute indexes tensor coords in reverse order.
telsoa014fcda012018-03-09 14:13:49 +0000183 unsigned int revertedIndex = subTensorShape.GetNumDimensions() - i - 1;
Matthew Sloyan171214c2020-09-09 09:07:37 +0100184 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
telsoa014fcda012018-03-09 14:13:49 +0000185 }
186
Derek Lamberti0790dce2019-04-15 18:37:35 +0100187 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.GetShape());
188 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
189 {
190 return nullptr;
191 }
192
telsoa01c577f2c2018-08-31 09:22:23 +0100193 return std::make_unique<ClSubTensorHandle>(
Jan Eilersbb446e52020-04-02 13:56:54 +0100194 PolymorphicDowncast<IClTensorHandle*>(&parent), shape, coords);
telsoa014fcda012018-03-09 14:13:49 +0000195}
196
telsoa014fcda012018-03-09 14:13:49 +0000197std::unique_ptr<IWorkload> ClWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100198 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000199{
Sadik Armagane9444752020-12-02 11:28:58 +0000200 return MakeWorkload<ClActivationWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000201}
202
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100203std::unique_ptr<IWorkload> ClWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
204 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000205{
Sadik Armagane9444752020-12-02 11:28:58 +0000206 return MakeWorkload<ClAdditionWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000207}
208
James Conroy2dc05722019-09-19 17:00:31 +0100209std::unique_ptr<IWorkload> ClWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
210 const WorkloadInfo& info) const
211{
Sadik Armagane9444752020-12-02 11:28:58 +0000212 return std::make_unique<ClArgMinMaxWorkload>(descriptor, info, m_CLCompileContext);
James Conroy2dc05722019-09-19 17:00:31 +0100213}
214
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100215std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchNormalization(
216 const BatchNormalizationQueueDescriptor& descriptor,
217 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000218{
Sadik Armagane9444752020-12-02 11:28:58 +0000219 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000220}
221
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100222std::unique_ptr<IWorkload> ClWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
telsoa014fcda012018-03-09 14:13:49 +0000223 const WorkloadInfo& info) const
224{
Sadik Armagane9444752020-12-02 11:28:58 +0000225 return MakeWorkload<ClBatchToSpaceNdWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000226}
227
Sadik Armaganf40d6d42021-04-22 09:12:11 +0100228std::unique_ptr<IWorkload> ClWorkloadFactory::CreateCast(const CastQueueDescriptor& descriptor,
229 const WorkloadInfo& info) const
230{
231 return MakeWorkload<ClCastWorkload>(descriptor, info, m_CLCompileContext);
232}
233
Teresa Charlin1222dbd2021-09-02 13:58:52 +0100234std::unique_ptr<IWorkload> ClWorkloadFactory::CreateChannelShuffle(const ChannelShuffleQueueDescriptor& descriptor,
235 const WorkloadInfo& info) const
236{
237 return MakeWorkload<ClChannelShuffleWorkload>(descriptor, info, m_CLCompileContext);
238}
239
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100240std::unique_ptr<IWorkload> ClWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
241 const WorkloadInfo& info) const
242{
Sadik Armagane9444752020-12-02 11:28:58 +0000243 return MakeWorkload<ClComparisonWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar77bfb5e2019-10-16 17:45:38 +0100244}
245
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100246std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
247 const WorkloadInfo& info) const
Jim Flynn4ed6c832019-05-20 11:02:46 +0100248{
Sadik Armagane9444752020-12-02 11:28:58 +0000249 return MakeWorkload<ClConcatWorkload>(descriptor, info, m_CLCompileContext);
Jim Flynn4ed6c832019-05-20 11:02:46 +0100250}
251
telsoa014fcda012018-03-09 14:13:49 +0000252std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100253 const WorkloadInfo& info) const
telsoa014fcda012018-03-09 14:13:49 +0000254{
Sadik Armagane9444752020-12-02 11:28:58 +0000255 return MakeWorkload<ClConstantWorkload>(descriptor, info, m_CLCompileContext);
telsoa014fcda012018-03-09 14:13:49 +0000256}
257
telsoa01c577f2c2018-08-31 09:22:23 +0100258std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp16ToFp32(
259 const ConvertFp16ToFp32QueueDescriptor& descriptor,
260 const WorkloadInfo& info) const
261{
Sadik Armagane9444752020-12-02 11:28:58 +0000262 return MakeWorkload<ClConvertFp16ToFp32Workload>(descriptor, info, m_CLCompileContext);
telsoa01c577f2c2018-08-31 09:22:23 +0100263}
264
265std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp32ToFp16(
266 const ConvertFp32ToFp16QueueDescriptor& descriptor,
267 const WorkloadInfo& info) const
268{
Sadik Armagane9444752020-12-02 11:28:58 +0000269 return MakeWorkload<ClConvertFp32ToFp16Workload>(descriptor, info, m_CLCompileContext);
telsoa01c577f2c2018-08-31 09:22:23 +0100270}
271
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100272std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
273 const WorkloadInfo& info) const
274{
Sadik Armagan04a72972020-09-14 15:44:18 +0100275 bool isFastMathEnabled = false;
276 if (m_ModelContextPtr)
277 {
278 if (m_ModelContextPtr.get() != nullptr)
279 {
280 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get());
281 if (modelOptions)
282 {
283 isFastMathEnabled = modelOptions->IsFastMathEnabled();
284 }
285 }
286 }
287 return MakeWorkload<ClConvolution2dWorkload>(descriptor,
288 info,
289 m_MemoryManager->GetIntraLayerManager(),
Sadik Armagane9444752020-12-02 11:28:58 +0000290 m_CLCompileContext,
Sadik Armagan04a72972020-09-14 15:44:18 +0100291 isFastMathEnabled);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100292}
293
294std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
295 const WorkloadInfo& info) const
296{
Sadik Armagane9444752020-12-02 11:28:58 +0000297 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100298}
299
300std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
301 const WorkloadInfo& info) const
302{
Sadik Armagane9444752020-12-02 11:28:58 +0000303 return MakeWorkload<ClDepthToSpaceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100304}
305
306std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthwiseConvolution2d(
307 const DepthwiseConvolution2dQueueDescriptor& descriptor,
308 const WorkloadInfo& info) const
309{
Sadik Armagane9444752020-12-02 11:28:58 +0000310 return MakeWorkload<ClDepthwiseConvolutionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100311}
312
313std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
314 const WorkloadInfo& info) const
315{
Sadik Armagane9444752020-12-02 11:28:58 +0000316 return MakeWorkload<ClDequantizeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100317}
318
319std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDetectionPostProcess(
320 const DetectionPostProcessQueueDescriptor& descriptor,
321 const WorkloadInfo& info) const
322{
Sadik Armagane9444752020-12-02 11:28:58 +0000323 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100324}
325
326std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
327 const WorkloadInfo& info) const
328{
Teresa Charline11e63d2021-04-21 12:56:45 +0100329 return std::make_unique<ClDivisionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100330}
331
josh minor4a3c6102020-01-06 16:40:46 -0600332std::unique_ptr<IWorkload> ClWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
333 const WorkloadInfo& info) const
334{
Sadik Armagan9fabf432020-05-27 13:40:58 +0100335 switch(descriptor.m_Parameters.m_Operation)
josh minor4a3c6102020-01-06 16:40:46 -0600336 {
Sadik Armagan9fabf432020-05-27 13:40:58 +0100337 case UnaryOperation::Abs:
James Conroyfe3ec942020-11-18 14:20:53 +0000338 {
339 AbsQueueDescriptor absQueueDescriptor;
340 absQueueDescriptor.m_Inputs = descriptor.m_Inputs;
341 absQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600342
Sadik Armagane9444752020-12-02 11:28:58 +0000343 return std::make_unique<ClAbsWorkload>(absQueueDescriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000344 }
Sadik Armagan9fabf432020-05-27 13:40:58 +0100345 case UnaryOperation::Exp:
Sadik Armagane9444752020-12-02 11:28:58 +0000346 return std::make_unique<ClExpWorkload>(descriptor, info, m_CLCompileContext);
Teresa Charlin50de4fa2021-05-31 18:47:33 +0100347 case UnaryOperation::Log:
348 return std::make_unique<ClLogWorkload>(descriptor, info, m_CLCompileContext);
349 case UnaryOperation::LogicalNot:
350 return std::make_unique<ClLogicalNotWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100351 case UnaryOperation::Neg:
Sadik Armagane9444752020-12-02 11:28:58 +0000352 return std::make_unique<ClNegWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100353 case UnaryOperation::Rsqrt:
James Conroyfe3ec942020-11-18 14:20:53 +0000354 {
355 RsqrtQueueDescriptor rsqrtQueueDescriptor;
356 rsqrtQueueDescriptor.m_Inputs = descriptor.m_Inputs;
357 rsqrtQueueDescriptor.m_Outputs = descriptor.m_Outputs;
josh minor4a3c6102020-01-06 16:40:46 -0600358
Sadik Armagane9444752020-12-02 11:28:58 +0000359 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000360 }
Teresa Charlin50de4fa2021-05-31 18:47:33 +0100361 case UnaryOperation::Sin:
362 return std::make_unique<ClSinWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan9fabf432020-05-27 13:40:58 +0100363 default:
364 return nullptr;
josh minor4a3c6102020-01-06 16:40:46 -0600365 }
josh minor4a3c6102020-01-06 16:40:46 -0600366}
367
Sadik Armagan66aecb02020-06-24 11:42:20 +0100368std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
369 const WorkloadInfo& info) const
370{
Sadik Armagane9444752020-12-02 11:28:58 +0000371 return std::make_unique<ClFillWorkload>(descriptor, info, m_CLCompileContext);
Sadik Armagan66aecb02020-06-24 11:42:20 +0100372}
373
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100374std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
375 const WorkloadInfo& info) const
376{
Sadik Armagane9444752020-12-02 11:28:58 +0000377 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100378}
379
380std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFullyConnected(const FullyConnectedQueueDescriptor& descriptor,
381 const WorkloadInfo& info) const
382{
Sadik Armagane9444752020-12-02 11:28:58 +0000383 return MakeWorkload<ClFullyConnectedWorkload>(descriptor,
384 info,
385 m_MemoryManager->GetIntraLayerManager(),
386 m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100387}
388
389std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
390 const WorkloadInfo& info) const
391{
Sadik Armagane9444752020-12-02 11:28:58 +0000392 return MakeWorkload<ClGatherWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100393}
394
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100395std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
396 const WorkloadInfo& info) const
397{
398 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
399}
400
401std::unique_ptr<IWorkload> ClWorkloadFactory::CreateInstanceNormalization(
402 const InstanceNormalizationQueueDescriptor& descriptor,
403 const WorkloadInfo& info) const
404{
Sadik Armagane9444752020-12-02 11:28:58 +0000405 return MakeWorkload<ClInstanceNormalizationWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100406}
407
408std::unique_ptr<IWorkload> ClWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
409 const WorkloadInfo& info) const
410{
Sadik Armagane9444752020-12-02 11:28:58 +0000411 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100412}
413
James Conroyfe3ec942020-11-18 14:20:53 +0000414std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
415 const WorkloadInfo& info) const
416{
417 switch(descriptor.m_Parameters.m_Operation)
418 {
419 case LogicalBinaryOperation::LogicalAnd:
Sadik Armagane9444752020-12-02 11:28:58 +0000420 return std::make_unique<ClLogicalAndWorkload>(descriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000421 case LogicalBinaryOperation::LogicalOr:
Sadik Armagane9444752020-12-02 11:28:58 +0000422 return std::make_unique<ClLogicalOrWorkload>(descriptor, info, m_CLCompileContext);
James Conroyfe3ec942020-11-18 14:20:53 +0000423 default:
424 return nullptr;
425 }
426}
427
Teresa Charlin8398edc2020-07-20 14:23:02 +0100428std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
429 const WorkloadInfo& info) const
430{
Sadik Armagane9444752020-12-02 11:28:58 +0000431 return MakeWorkload<ClLogSoftmaxWorkload>(descriptor,
432 info,
433 m_MemoryManager->GetIntraLayerManager(),
434 m_CLCompileContext);
Teresa Charlin8398edc2020-07-20 14:23:02 +0100435}
436
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100437std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
438 const WorkloadInfo& info) const
439{
Sadik Armagane9444752020-12-02 11:28:58 +0000440 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100441}
442
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +0000443std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
444 const WorkloadInfo& info) const
445{
Sadik Armagane9444752020-12-02 11:28:58 +0000446 return MakeWorkload<ClMaximumWorkload>(descriptor, info, m_CLCompileContext);
Nattapat Chaimanowong5a4304a2018-11-28 10:44:37 +0000447}
448
narpra01a6bf9122018-09-10 09:50:09 +0100449std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
450 const WorkloadInfo& info) const
451{
Sadik Armagane9444752020-12-02 11:28:58 +0000452 return MakeWorkload<ClMeanWorkload>(descriptor, info, m_CLCompileContext);
narpra01a6bf9122018-09-10 09:50:09 +0100453}
454
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100455std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
456 const WorkloadInfo& info) const
jimfly012c9322a2018-09-19 10:59:49 +0100457{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100458 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
459 {
460 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemCopy workload");
461 }
462
463 return MakeWorkload<CopyMemGenericWorkload>(descriptor, info);
jimfly012c9322a2018-09-19 10:59:49 +0100464}
465
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100466std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
467 const WorkloadInfo& info) const
FrancisMurtagh20995952018-12-17 12:11:36 +0000468{
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100469 if (descriptor.m_Inputs.empty() || !descriptor.m_Inputs[0])
470 {
471 throw InvalidArgumentException("ClWorkloadFactory: Invalid null input for MemImport workload");
472 }
473
474 return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
FrancisMurtagh20995952018-12-17 12:11:36 +0000475}
476
kevmay0190539692018-11-29 08:40:19 +0000477std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
478 const WorkloadInfo& info) const
479{
Sadik Armagane9444752020-12-02 11:28:58 +0000480 return MakeWorkload<ClMinimumWorkload>(descriptor, info, m_CLCompileContext);
kevmay0190539692018-11-29 08:40:19 +0000481}
482
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100483std::unique_ptr<IWorkload> ClWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
484 const WorkloadInfo& info) const
485{
Sadik Armagane9444752020-12-02 11:28:58 +0000486 return MakeWorkload<ClMultiplicationWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100487}
488
489std::unique_ptr<IWorkload> ClWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
490 const WorkloadInfo& info) const
491{
Sadik Armagane9444752020-12-02 11:28:58 +0000492 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100493}
494
495std::unique_ptr<IWorkload> ClWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
496 const WorkloadInfo& info) const
497{
498 return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
499}
500
501std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
502 const WorkloadInfo& info) const
503{
Sadik Armagane9444752020-12-02 11:28:58 +0000504 return MakeWorkload<ClPadWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100505}
506
507std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
Matteo Martincigh59a950c2018-12-13 12:48:25 +0000508 const WorkloadInfo& info) const
509{
Sadik Armagane9444752020-12-02 11:28:58 +0000510 return MakeWorkload<ClPermuteWorkload>(descriptor, info, m_CLCompileContext);
Matteo Martincigh59a950c2018-12-13 12:48:25 +0000511}
512
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100513std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
514 const WorkloadInfo& info) const
Nattapat Chaimanowonga9a1cf12018-12-03 16:06:49 +0000515{
Sadik Armagane9444752020-12-02 11:28:58 +0000516 return MakeWorkload<ClPooling2dWorkload>(descriptor, info, m_CLCompileContext);
Mohamed Nour Abouelseouda1d3c6a2018-12-27 12:39:16 +0000517}
518
Matteo Martincigh49124022019-01-11 13:25:59 +0000519std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor,
520 const WorkloadInfo& info) const
521{
Sadik Armagane9444752020-12-02 11:28:58 +0000522 return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info, m_CLCompileContext);
Matteo Martincigh49124022019-01-11 13:25:59 +0000523}
524
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100525std::unique_ptr<IWorkload> ClWorkloadFactory::CreatePrelu(const PreluQueueDescriptor &descriptor,
526 const WorkloadInfo &info) const
narpra01b89b05f2019-01-16 09:53:09 +0000527{
Sadik Armagane9444752020-12-02 11:28:58 +0000528 return MakeWorkload<ClPreluWorkload>(descriptor, info, m_CLCompileContext);
narpra01b89b05f2019-01-16 09:53:09 +0000529}
530
Ryan OShea2323af42020-05-13 16:36:19 +0100531std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
532 const WorkloadInfo& info) const
533{
Sadik Armagane9444752020-12-02 11:28:58 +0000534 return std::make_unique<ClQLstmWorkload>(descriptor, info, m_CLCompileContext);
Ryan OShea2323af42020-05-13 16:36:19 +0100535}
536
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100537std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
538 const WorkloadInfo& info) const
Aron Virginas-Tar7a3e2fe2019-06-27 18:54:47 +0100539{
Sadik Armagane9444752020-12-02 11:28:58 +0000540 return MakeWorkload<ClQuantizeWorkload>(descriptor, info, m_CLCompileContext);
James Conroyd2aa85e2019-07-01 17:12:40 +0100541}
542
Ferran Balaguer737d9ff2019-08-01 09:58:08 +0100543std::unique_ptr<IWorkload> ClWorkloadFactory::CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& descriptor,
544 const WorkloadInfo& info) const
545{
Sadik Armagane9444752020-12-02 11:28:58 +0000546 return MakeWorkload<ClQuantizedLstmWorkload>(descriptor, info, m_CLCompileContext);
Ferran Balaguer737d9ff2019-08-01 09:58:08 +0100547}
548
David Monahanc11ba462020-12-03 11:09:46 +0000549std::unique_ptr<IWorkload> ClWorkloadFactory::CreateRank(const RankQueueDescriptor& descriptor,
550 const WorkloadInfo& info) const
551{
552 return std::make_unique<ClRankWorkload>(descriptor, info);
553}
554
Sadik Armagana2747482021-02-09 10:28:54 +0000555std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReduce(const ReduceQueueDescriptor& descriptor,
556 const WorkloadInfo& info) const
557{
558 return std::make_unique<ClReduceWorkload>(descriptor, info);
559}
560
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100561std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
562 const WorkloadInfo& info) const
563{
Sadik Armagane9444752020-12-02 11:28:58 +0000564 return MakeWorkload<ClReshapeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100565}
566
567std::unique_ptr<IWorkload> ClWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
568 const WorkloadInfo& info) const
569{
Sadik Armagane9444752020-12-02 11:28:58 +0000570 return MakeWorkload<ClResizeWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100571}
572
Aron Virginas-Tar94c4fef2019-11-25 15:37:08 +0000573std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
574 const WorkloadInfo& info) const
575{
Sadik Armagane9444752020-12-02 11:28:58 +0000576 return MakeWorkload<ClSliceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar94c4fef2019-11-25 15:37:08 +0000577}
578
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100579std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
580 const WorkloadInfo& info) const
581{
Sadik Armagane9444752020-12-02 11:28:58 +0000582 return std::make_unique<ClSoftmaxWorkload>(descriptor,
583 info,
584 m_MemoryManager->GetIntraLayerManager(),
585 m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100586}
587
588std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
589 const WorkloadInfo& info) const
590{
Sadik Armagane9444752020-12-02 11:28:58 +0000591 return MakeWorkload<ClSpaceToBatchNdWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100592}
593
594std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
595 const WorkloadInfo& info) const
596{
Sadik Armagane9444752020-12-02 11:28:58 +0000597 return MakeWorkload<ClSpaceToDepthWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100598}
599
600std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
601 const WorkloadInfo& info) const
602{
Sadik Armagane9444752020-12-02 11:28:58 +0000603 return MakeWorkload<ClSplitterWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100604}
605
Matthew Jacksond5166102019-07-31 14:06:28 +0100606std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
607 const WorkloadInfo& info) const
608{
Sadik Armagane9444752020-12-02 11:28:58 +0000609 return MakeWorkload<ClStackWorkload>(descriptor, info, m_CLCompileContext);
Matthew Jacksond5166102019-07-31 14:06:28 +0100610}
611
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100612std::unique_ptr<IWorkload> ClWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
Aron Virginas-Tarb2801962019-09-30 11:24:53 +0100613 const WorkloadInfo& info) const
614{
Sadik Armagane9444752020-12-02 11:28:58 +0000615 return MakeWorkload<ClStridedSliceWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100616}
617
618std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
619 const WorkloadInfo& info) const
620{
Sadik Armagane9444752020-12-02 11:28:58 +0000621 return MakeWorkload<ClSubtractionWorkload>(descriptor, info, m_CLCompileContext);
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100622}
623
Mike Kellyc9ea45a2020-02-28 18:11:58 +0000624std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
625 const WorkloadInfo& info) const
626{
Sadik Armagane9444752020-12-02 11:28:58 +0000627 return MakeWorkload<ClTransposeWorkload>(descriptor, info, m_CLCompileContext);
Mike Kellyc9ea45a2020-02-28 18:11:58 +0000628}
629
Aron Virginas-Tar8168f402019-10-04 13:10:16 +0100630std::unique_ptr<IWorkload> ClWorkloadFactory::CreateTransposeConvolution2d(
631 const TransposeConvolution2dQueueDescriptor& descriptor,
632 const WorkloadInfo& info) const
633{
Sadik Armagane9444752020-12-02 11:28:58 +0000634 return MakeWorkload<ClTransposeConvolution2dWorkload>(descriptor,
635 info,
636 m_MemoryManager->GetIntraLayerManager(),
637 m_CLCompileContext);
Aron Virginas-Tarb2801962019-09-30 11:24:53 +0100638}
639
telsoa014fcda012018-03-09 14:13:49 +0000640} // namespace armnn