blob: 3f0fe842aa7e391db57675b8f778fd5bb0bba31f [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
David Beckecb56cd2018-09-05 12:52:57 +01003// SPDX-License-Identifier: MIT
telsoa014fcda012018-03-09 14:13:49 +00004//
5#pragma once
6
Aron Virginas-Tarc9cc8042018-11-01 16:15:57 +00007#include <backendsCommon/Workload.hpp>
8#include <aclCommon/ArmComputeTensorUtils.hpp>
9#include <neon/NeonTensorHandle.hpp>
10#include <neon/NeonTimer.hpp>
11#include <backendsCommon/CpuTensorHandle.hpp>
telsoa014fcda012018-03-09 14:13:49 +000012
Derek Lambertid466a542020-01-22 15:37:29 +000013#include <armnn/Utils.hpp>
14
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +010015#include <Half.hpp>
telsoa014fcda012018-03-09 14:13:49 +000016
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +010017#define ARMNN_SCOPED_PROFILING_EVENT_NEON(name) \
telsoa01c577f2c2018-08-31 09:22:23 +010018 ARMNN_SCOPED_PROFILING_EVENT_WITH_INSTRUMENTS(armnn::Compute::CpuAcc, \
19 name, \
Nina Drozd69851b52018-09-21 18:42:09 +010020 armnn::NeonTimer(), \
21 armnn::WallClockTimer())
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +010022
23using namespace armnn::armcomputetensorutils;
24
25namespace armnn
26{
27
28template <typename T>
29void CopyArmComputeTensorData(arm_compute::Tensor& dstTensor, const T* srcData)
30{
31 InitialiseArmComputeTensorEmpty(dstTensor);
32 CopyArmComputeITensorData(srcData, dstTensor);
33}
34
35inline void InitializeArmComputeTensorData(arm_compute::Tensor& tensor,
36 const ConstCpuTensorHandle* handle)
37{
38 BOOST_ASSERT(handle);
39
40 switch(handle->GetTensorInfo().GetDataType())
41 {
42 case DataType::Float16:
43 CopyArmComputeTensorData(tensor, handle->GetConstTensor<armnn::Half>());
44 break;
45 case DataType::Float32:
46 CopyArmComputeTensorData(tensor, handle->GetConstTensor<float>());
47 break;
Derek Lambertif90c56d2020-01-10 17:14:08 +000048 case DataType::QAsymmU8:
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +010049 CopyArmComputeTensorData(tensor, handle->GetConstTensor<uint8_t>());
50 break;
Derek Lambertid466a542020-01-22 15:37:29 +000051 ARMNN_NO_DEPRECATE_WARN_BEGIN
Aron Virginas-Tar21fc28b2019-11-26 14:04:54 +000052 case DataType::QuantizedSymm8PerAxis:
Derek Lambertid466a542020-01-22 15:37:29 +000053 ARMNN_FALLTHROUGH;
54 case DataType::QSymmS8:
Aron Virginas-Tar21fc28b2019-11-26 14:04:54 +000055 CopyArmComputeTensorData(tensor, handle->GetConstTensor<int8_t>());
56 break;
Derek Lambertid466a542020-01-22 15:37:29 +000057 ARMNN_NO_DEPRECATE_WARN_END
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +010058 case DataType::Signed32:
59 CopyArmComputeTensorData(tensor, handle->GetConstTensor<int32_t>());
60 break;
61 default:
62 BOOST_ASSERT_MSG(false, "Unexpected tensor type.");
63 }
64};
65
FinnWilliamsArm1fa19192019-08-02 17:26:31 +010066inline auto SetNeonStridedSliceData(const std::vector<int>& m_begin,
67 const std::vector<int>& m_end,
68 const std::vector<int>& m_stride)
69{
70 arm_compute::Coordinates starts;
71 arm_compute::Coordinates ends;
72 arm_compute::Coordinates strides;
73
74 unsigned int num_dims = static_cast<unsigned int>(m_begin.size());
75
76 for (unsigned int i = 0; i < num_dims; i++)
77 {
78 unsigned int revertedIndex = num_dims - i - 1;
79
80 starts.set(i, static_cast<int>(m_begin[revertedIndex]));
81 ends.set(i, static_cast<int>(m_end[revertedIndex]));
82 strides.set(i, static_cast<int>(m_stride[revertedIndex]));
83 }
84
85 return std::make_tuple(starts, ends, strides);
86}
87
josh minor036f02d2019-11-15 14:53:22 -060088inline auto SetNeonSliceData(const std::vector<unsigned int>& m_begin,
89 const std::vector<unsigned int>& m_size)
90{
91 // This function must translate the size vector given to an end vector
92 // expected by the ACL NESlice workload
93 arm_compute::Coordinates starts;
94 arm_compute::Coordinates ends;
95
96 unsigned int num_dims = static_cast<unsigned int>(m_begin.size());
97
98 // For strided slices, we have the relationship size = (end - begin) / stride
99 // For slice, we assume stride to be a vector of all ones, yielding the formula
100 // size = (end - begin) therefore we know end = size + begin
101 for (unsigned int i = 0; i < num_dims; i++)
102 {
103 unsigned int revertedIndex = num_dims - i - 1;
104
105 starts.set(i, static_cast<int>(m_begin[revertedIndex]));
106 ends.set(i, static_cast<int>(m_begin[revertedIndex] + m_size[revertedIndex]));
107 }
108
109 return std::make_tuple(starts, ends);
110}
111
Nattapat Chaimanowong177d8d22018-10-16 13:21:27 +0100112} //namespace armnn