blob: 20936a2760a77a1e6baf7d456d2fc27e285a5ee1 [file] [log] [blame]
telsoa014fcda012018-03-09 14:13:49 +00001//
2// Copyright © 2017 Arm Ltd. All rights reserved.
3// See LICENSE file in the project root for full license information.
4//
5
6#include "NeonNormalizationFloat32Workload.hpp"
7#include "backends/NeonLayerSupport.hpp"
8#include "backends/ArmComputeUtils.hpp"
telsoa01c577f2c2018-08-31 09:22:23 +01009#include "backends/ArmComputeTensorUtils.hpp"
telsoa014fcda012018-03-09 14:13:49 +000010
11namespace armnn
12{
13
telsoa01c577f2c2018-08-31 09:22:23 +010014arm_compute::Status NeonNormalizationWorkloadValidate(const TensorInfo& input,
15 const TensorInfo& output,
16 const NormalizationDescriptor& descriptor)
17{
18 const arm_compute::TensorInfo aclInput = armcomputetensorutils::BuildArmComputeTensorInfo(input);
19 const arm_compute::TensorInfo aclOutput = armcomputetensorutils::BuildArmComputeTensorInfo(output);
20
21 arm_compute::NormalizationLayerInfo normalizationInfo =
22 armcomputetensorutils::BuildArmComputeNormalizationLayerInfo(descriptor);
23
24 return arm_compute::NENormalizationLayer::validate(&aclInput, &aclOutput, normalizationInfo);
25}
26
telsoa014fcda012018-03-09 14:13:49 +000027NeonNormalizationFloat32Workload::NeonNormalizationFloat32Workload(const NormalizationQueueDescriptor& descriptor,
telsoa01c577f2c2018-08-31 09:22:23 +010028 const WorkloadInfo& info,
29 std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
30 : FloatWorkload<NormalizationQueueDescriptor>(descriptor, info)
surmeh013537c2c2018-05-18 16:31:43 +010031 , m_NormalizationLayer(memoryManager)
telsoa014fcda012018-03-09 14:13:49 +000032{
33 m_Data.ValidateInputsOutputs("NeonNormalizationFloat32Workload", 1, 1);
34 std::string reasonIfUnsupported;
35 if (!IsNeonNormalizationDescParamsSupported(&reasonIfUnsupported, m_Data.m_Parameters))
36 {
37 throw UnimplementedException(reasonIfUnsupported);
38 }
39
telsoa01c577f2c2018-08-31 09:22:23 +010040 // Input and output tensors have to have the same dimensionality.
telsoa014fcda012018-03-09 14:13:49 +000041 if (info.m_InputTensorInfos[0].GetShape()[1] != info.m_OutputTensorInfos[0].GetShape()[1]
42 || info.m_InputTensorInfos[0].GetShape()[0] != info.m_OutputTensorInfos[0].GetShape()[0]
43 || info.m_InputTensorInfos[0].GetShape()[3] != info.m_OutputTensorInfos[0].GetShape()[3]
44 || info.m_InputTensorInfos[0].GetShape()[2] != info.m_OutputTensorInfos[0].GetShape()[2])
45 {
46 throw InvalidArgumentException("Normalization requires input and output tensors to have equal dimensionality.");
47 }
48
49 arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
50 arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
51
52 const arm_compute::NormType normType =
53 ConvertNormalizationAlgorithmChannelToAclNormType(m_Data.m_Parameters.m_NormChannelType);
54 arm_compute::NormalizationLayerInfo normalizationInfo(normType,
55 m_Data.m_Parameters.m_NormSize,
56 m_Data.m_Parameters.m_Alpha,
57 m_Data.m_Parameters.m_Beta,
58 m_Data.m_Parameters.m_K,
59 false);
60
61 m_NormalizationLayer.configure(&input, &output, normalizationInfo);
62}
63
64void NeonNormalizationFloat32Workload::Execute() const
65{
telsoa01c577f2c2018-08-31 09:22:23 +010066 ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonNormalizationFloat32Workload_Execute");
telsoa014fcda012018-03-09 14:13:49 +000067 m_NormalizationLayer.run();
68}
69
70} //namespace armnn