blob: 98b5adafbc4025058db587318e270f93d1d9874d [file] [log] [blame]
Laurent Carlier749294b2020-06-01 09:03:17 +01001//
telsoa014fcda012018-03-09 14:13:49 +00002// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +00005#include <aclCommon/ArmComputeTensorUtils.hpp>
6#include <aclCommon/ArmComputeUtils.hpp>
telsoa014fcda012018-03-09 14:13:49 +00007
Francis Murtagh351d13d2018-09-24 15:01:18 +01008#include "armnn/Exceptions.hpp"
telsoa014fcda012018-03-09 14:13:49 +00009#include <armnn/Descriptors.hpp>
10
11namespace armnn
12{
13namespace armcomputetensorutils
14{
15
Derek Lambertid466a542020-01-22 15:37:29 +000016arm_compute::DataType GetArmComputeDataType(armnn::DataType dataType, bool multiScales)
telsoa014fcda012018-03-09 14:13:49 +000017{
18 switch(dataType)
19 {
Narumol Prangnawarat250d3922020-03-30 16:11:04 +010020 case armnn::DataType::BFloat16:
21 return arm_compute::DataType::BFLOAT16;
Mike Kelly130ec602019-11-08 12:08:35 +000022 case armnn::DataType::Boolean:
23 return arm_compute::DataType::U8;
telsoa01c577f2c2018-08-31 09:22:23 +010024 case armnn::DataType::Float16:
25 return arm_compute::DataType::F16;
telsoa014fcda012018-03-09 14:13:49 +000026 case armnn::DataType::Float32:
telsoa014fcda012018-03-09 14:13:49 +000027 return arm_compute::DataType::F32;
Ryan OShea9add1202020-02-07 10:06:33 +000028 case armnn::DataType::QAsymmS8:
29 return arm_compute::DataType::QASYMM8_SIGNED;
Derek Lambertif90c56d2020-01-10 17:14:08 +000030 case armnn::DataType::QAsymmU8:
telsoa014fcda012018-03-09 14:13:49 +000031 return arm_compute::DataType::QASYMM8;
Derek Lambertif90c56d2020-01-10 17:14:08 +000032 case armnn::DataType::QSymmS16:
Aron Virginas-Tar7a3e2fe2019-06-27 18:54:47 +010033 return arm_compute::DataType::QSYMM16;
Inki Daed4619e22020-09-10 15:33:54 +090034 case armnn::DataType::Signed64:
35 return arm_compute::DataType::S64;
Finn Williamsfd271062019-12-04 14:27:27 +000036 case armnn::DataType::QSymmS8:
Derek Lambertid466a542020-01-22 15:37:29 +000037 {
38 return multiScales ? arm_compute::DataType::QSYMM8_PER_CHANNEL : arm_compute::DataType::QSYMM8;
39 }
40 ARMNN_NO_DEPRECATE_WARN_BEGIN
Mike Kelly130ec602019-11-08 12:08:35 +000041 case armnn::DataType::QuantizedSymm8PerAxis:
42 return arm_compute::DataType::QSYMM8_PER_CHANNEL;
Derek Lambertid466a542020-01-22 15:37:29 +000043 ARMNN_NO_DEPRECATE_WARN_END
telsoa014fcda012018-03-09 14:13:49 +000044 case armnn::DataType::Signed32:
telsoa014fcda012018-03-09 14:13:49 +000045 return arm_compute::DataType::S32;
telsoa014fcda012018-03-09 14:13:49 +000046 default:
Narumol Prangnawaratac2770a2020-04-01 16:51:23 +010047 ARMNN_ASSERT_MSG(false, "Unknown data type");
telsoa014fcda012018-03-09 14:13:49 +000048 return arm_compute::DataType::UNKNOWN;
telsoa014fcda012018-03-09 14:13:49 +000049 }
50}
51
Matthew Benthamfd899962018-12-31 15:49:42 +000052arm_compute::Coordinates BuildArmComputeReductionCoordinates(size_t inputDimensions,
53 unsigned int originalInputRank,
54 const std::vector<unsigned int>& armnnAxes)
55{
56 arm_compute::Coordinates outAclCoords;
57
58 if (armnnAxes.empty())
59 {
60 // If no reduction axes were provided, then the input must be reduced along all dimensions.
61 // Since Compute Library does not accept an empty vector as the reduction dimensions, we then
62 // manually create a vector including all the input dimensions (in reversed order) as:
63 //
64 // { inputDimensions - 1, inputDimensions - 2, ..., 1, 0 }
65 //
66 outAclCoords.set_num_dimensions(inputDimensions);
67 std::generate(outAclCoords.begin(), outAclCoords.end(), [d = inputDimensions - 1] () mutable { return d--; });
68 }
69 else
70 {
71 // Create a vector of reduction dimensions (in reversed order) with the given reduction axes.
72 //
73 // Adjust the given reduction axes according to the original rank of the input tensor (before ACL applied any
74 // dimension correction).
75 // For example, if the input tensor originally had 4 dimensions, and one of the reduction axes was 2, then the
76 // new value for that reduction axis should be 1.
77 //
78 // Example:
79 // ArmNN input shape = { 1, 1, 3, 2 } -> ACL input shape = { 2, 3 }
80 // ArmNN reduction axis = { 2 } -> ACL reduction axis = { 1 }
81 // ArmNN reduction axis = { 3 } -> ACL reduction axis = { 0 }
82 //
83 // The transformation: ACL reduction axis index = original rank - ArmNN reduction axis index - 1
84 //
85 outAclCoords.set_num_dimensions(armnnAxes.size());
86 std::transform(armnnAxes.begin(), armnnAxes.end(),
87 outAclCoords.begin(),
88 [originalInputRank](unsigned int i){ return originalInputRank - i - 1; });
89 }
90
91 return outAclCoords;
92}
93
telsoa014fcda012018-03-09 14:13:49 +000094arm_compute::TensorShape BuildArmComputeTensorShape(const armnn::TensorShape& tensorShape)
95{
96 arm_compute::TensorShape shape;
97
telsoa01c577f2c2018-08-31 09:22:23 +010098 // armnn tensors are (batch, channels, height, width).
99 // arm_compute tensors are (width, height, channels, batch).
telsoa014fcda012018-03-09 14:13:49 +0000100 for (unsigned int i = 0; i < tensorShape.GetNumDimensions(); i++)
101 {
telsoa01c577f2c2018-08-31 09:22:23 +0100102 // Note that our dimensions are stored in the opposite order to ACL's.
Matthew Bentham89105282018-11-20 14:33:33 +0000103 shape.set(tensorShape.GetNumDimensions() - i - 1, tensorShape[i], false);
telsoa014fcda012018-03-09 14:13:49 +0000104
105 // TensorShape::set() flattens leading ones, so that batch size 1 cannot happen.
telsoa01c577f2c2018-08-31 09:22:23 +0100106 // arm_compute tensors expect this.
telsoa014fcda012018-03-09 14:13:49 +0000107 }
108
109 // prevent arm_compute issue where tensor is flattened to nothing
110 if (shape.num_dimensions() == 0)
111 {
112 shape.set_num_dimensions(1);
113 }
114
115 return shape;
116}
117
118// Utility function used to build a TensorInfo object, that can be used to initialise
119// ARM Compute Tensor and CLTensor allocators.
120arm_compute::TensorInfo BuildArmComputeTensorInfo(const armnn::TensorInfo& tensorInfo)
121{
Derek Lambertid466a542020-01-22 15:37:29 +0000122 bool multiScales = tensorInfo.HasMultipleQuantizationScales();
telsoa014fcda012018-03-09 14:13:49 +0000123 const arm_compute::TensorShape aclTensorShape = BuildArmComputeTensorShape(tensorInfo.GetShape());
Derek Lambertid466a542020-01-22 15:37:29 +0000124 const arm_compute::DataType aclDataType = GetArmComputeDataType(tensorInfo.GetDataType(), multiScales);
Aron Virginas-Tar13b653f2019-11-01 11:40:39 +0000125
Derek Lambertid466a542020-01-22 15:37:29 +0000126 const arm_compute::QuantizationInfo aclQuantizationInfo = multiScales ?
Aron Virginas-Tar13b653f2019-11-01 11:40:39 +0000127 arm_compute::QuantizationInfo(tensorInfo.GetQuantizationScales()) :
128 arm_compute::QuantizationInfo(tensorInfo.GetQuantizationScale(), tensorInfo.GetQuantizationOffset());
telsoa014fcda012018-03-09 14:13:49 +0000129
130 return arm_compute::TensorInfo(aclTensorShape, 1, aclDataType, aclQuantizationInfo);
131}
132
Francis Murtagh351d13d2018-09-24 15:01:18 +0100133arm_compute::TensorInfo BuildArmComputeTensorInfo(const armnn::TensorInfo& tensorInfo,
134 armnn::DataLayout dataLayout)
135{
Aron Virginas-Tar13b653f2019-11-01 11:40:39 +0000136 arm_compute::TensorInfo aclTensorInfo = BuildArmComputeTensorInfo(tensorInfo);
137 aclTensorInfo.set_data_layout(ConvertDataLayout(dataLayout));
Francis Murtagh351d13d2018-09-24 15:01:18 +0100138
Aron Virginas-Tar13b653f2019-11-01 11:40:39 +0000139 return aclTensorInfo;
Francis Murtagh351d13d2018-09-24 15:01:18 +0100140}
141
Matteo Martincigh747ef822018-12-18 09:26:39 +0000142arm_compute::DataLayout ConvertDataLayout(armnn::DataLayout dataLayout)
143{
144 switch(dataLayout)
145 {
146 case armnn::DataLayout::NHWC : return arm_compute::DataLayout::NHWC;
147
148 case armnn::DataLayout::NCHW : return arm_compute::DataLayout::NCHW;
149
150 default: throw InvalidArgumentException("Unknown armnn::DataLayout: [" +
151 std::to_string(static_cast<int>(dataLayout)) + "]");
152 }
153}
154
Sadik Armagana3600ba2019-10-10 10:43:20 +0100155arm_compute::PoolingLayerInfo BuildArmComputePoolingLayerInfo(const Pooling2dDescriptor& descriptor,
156 bool fpMixedPrecision)
telsoa014fcda012018-03-09 14:13:49 +0000157{
158 using arm_compute::PoolingType;
159 using arm_compute::DimensionRoundingType;
160 using arm_compute::PadStrideInfo;
161 using arm_compute::PoolingLayerInfo;
surmeh01bceff2f2018-03-29 16:29:27 +0100162 using arm_compute::Size2D;
Teresa Charlinc809a292020-01-31 10:21:44 +0000163 using arm_compute::DataLayout;
telsoa014fcda012018-03-09 14:13:49 +0000164
telsoa01c577f2c2018-08-31 09:22:23 +0100165 // Resolve ARM Compute layer parameters.
telsoa014fcda012018-03-09 14:13:49 +0000166 const PoolingType poolingType = ConvertPoolingAlgorithmToAclPoolingType(descriptor.m_PoolType);
telsoa01c577f2c2018-08-31 09:22:23 +0100167
Teresa Charlinc809a292020-01-31 10:21:44 +0000168 const DataLayout dataLayout = ConvertDataLayout(descriptor.m_DataLayout);
169
telsoa01c577f2c2018-08-31 09:22:23 +0100170 bool isGlobalPooling = (descriptor.m_StrideX==0 && descriptor.m_StrideY==0);
171 //use specific constructor if global pooling
172 if(isGlobalPooling)
173 {
Teresa Charlinc809a292020-01-31 10:21:44 +0000174 return arm_compute::PoolingLayerInfo(poolingType, dataLayout);
telsoa01c577f2c2018-08-31 09:22:23 +0100175 }
176
telsoa014fcda012018-03-09 14:13:49 +0000177 const DimensionRoundingType rounding = ConvertOutputShapeRoundingToAclDimensionRoundingType(
178 descriptor.m_OutputShapeRounding);
telsoa014fcda012018-03-09 14:13:49 +0000179 const PadStrideInfo padStrideInfo(descriptor.m_StrideX,
180 descriptor.m_StrideY,
181 descriptor.m_PadLeft,
182 descriptor.m_PadRight,
183 descriptor.m_PadTop,
184 descriptor.m_PadBottom,
185 rounding);
186
187 const bool excludePadding = (descriptor.m_PaddingMethod == PaddingMethod::Exclude);
188
surmeh01bceff2f2018-03-29 16:29:27 +0100189 const Size2D poolSize(descriptor.m_PoolWidth, descriptor.m_PoolHeight);
190
Teresa Charlinc809a292020-01-31 10:21:44 +0000191 return arm_compute::PoolingLayerInfo(poolingType, poolSize, dataLayout, padStrideInfo, excludePadding,
192 fpMixedPrecision);
telsoa014fcda012018-03-09 14:13:49 +0000193}
194
195arm_compute::NormalizationLayerInfo BuildArmComputeNormalizationLayerInfo(const NormalizationDescriptor& descriptor)
196{
197 const arm_compute::NormType normType =
198 ConvertNormalizationAlgorithmChannelToAclNormType(descriptor.m_NormChannelType);
199 return arm_compute::NormalizationLayerInfo(normType,
200 descriptor.m_NormSize,
201 descriptor.m_Alpha,
202 descriptor.m_Beta,
203 descriptor.m_K,
204 false);
205}
206
207arm_compute::PermutationVector BuildArmComputePermutationVector(const armnn::PermutationVector& perm)
208{
209 arm_compute::PermutationVector aclPerm;
210
211 unsigned int start = 0;
surmeh01bceff2f2018-03-29 16:29:27 +0100212 while ((start < perm.GetSize()) && (start == perm[start]))
telsoa014fcda012018-03-09 14:13:49 +0000213 {
214 ++start;
215 }
216
217 for (unsigned int i = start; i < perm.GetSize(); ++i)
218 {
219 aclPerm.set(i - start, perm[i] - start);
220 }
Mike Kellyc9ea45a2020-02-28 18:11:58 +0000221 return aclPerm;
222}
telsoa014fcda012018-03-09 14:13:49 +0000223
Mike Kellyc9ea45a2020-02-28 18:11:58 +0000224arm_compute::PermutationVector BuildArmComputeTransposeVector(const armnn::PermutationVector& perm)
225{
226 arm_compute::PermutationVector aclPerm;
227 std::map<unsigned int, unsigned int> permuteMappings;
228 for (unsigned int i = 0; i < perm.GetSize(); ++i)
229 {
230 permuteMappings[perm[i]] = i;
231 }
232
233 std::vector<unsigned int> permuteVector;
234 for (unsigned int i = 0; i < perm.GetSize(); ++i)
235 {
236 permuteVector.push_back(permuteMappings.at(i));
237 }
238
239 unsigned int start = 0;
240 while ((start < perm.GetSize()) && (start == permuteVector[start]))
241 {
242 ++start;
243 }
244
245 for (unsigned int i = start; i < perm.GetSize(); ++i)
246 {
247 aclPerm.set(i - start, permuteVector[i] - start);
248 }
telsoa014fcda012018-03-09 14:13:49 +0000249 return aclPerm;
250}
251
Sadik Armaganf4464322018-12-20 16:19:12 +0000252arm_compute::Size2D BuildArmComputeSize2D(const unsigned int width, const unsigned int height)
253{
254 return arm_compute::Size2D(width, height);
255}
256
Mike Kelly0a08ec62019-07-25 08:39:31 +0100257arm_compute::PixelValue GetPixelValue(arm_compute::ITensor& input, float pixelValue)
258{
259 switch (input.info()->data_type())
260 {
Mike Kelly0a08ec62019-07-25 08:39:31 +0100261 case arm_compute::DataType::F16:
262 return arm_compute::PixelValue(static_cast<Half>(pixelValue));
263 case arm_compute::DataType::F32:
264 return arm_compute::PixelValue(pixelValue);
Mike Kelly130ec602019-11-08 12:08:35 +0000265 case arm_compute::DataType::QASYMM8:
266 return arm_compute::PixelValue(static_cast<uint8_t>(pixelValue));
267 case arm_compute::DataType::QSYMM16:
268 return arm_compute::PixelValue(static_cast<int16_t>(pixelValue));
Sadik Armagane5d0b932020-04-09 15:48:44 +0100269 case arm_compute::DataType::QASYMM8_SIGNED:
Mike Kelly130ec602019-11-08 12:08:35 +0000270 case arm_compute::DataType::QSYMM8_PER_CHANNEL:
271 return arm_compute::PixelValue(static_cast<int8_t>(pixelValue));
Sadik Armagana792a052020-06-23 16:22:23 +0100272 case arm_compute::DataType::S32:
273 return arm_compute::PixelValue(static_cast<int32_t>(pixelValue));
Mike Kelly0a08ec62019-07-25 08:39:31 +0100274 default:
275 throw InvalidArgumentException("Unsupported DataType: [" +
276 std::to_string(static_cast<int>(input.info()->data_type())) + "]");
277 }
278}
279
telsoa014fcda012018-03-09 14:13:49 +0000280} // namespace armcomputetensorutils
281} // namespace armnn