| // |
| // Copyright © 2019,2021-2023 Arm Ltd and Contributors. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| |
| #include "RefLstmWorkload.hpp" |
| #include "Activation.hpp" |
| #include "Encoders.hpp" |
| #include "Decoders.hpp" |
| #include "Lstm.hpp" |
| #include "LstmUtils.hpp" |
| #include "RefWorkloadUtils.hpp" |
| |
| namespace armnn |
| { |
| |
| RefLstmWorkload::RefLstmWorkload(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info) |
| : RefBaseWorkload<LstmQueueDescriptor>(descriptor, info) |
| , m_InputToInputWeightsTensor (AssignScopedTensorHandle(descriptor.m_InputToInputWeights)) |
| , m_InputToForgetWeightsTensor (AssignScopedTensorHandle(descriptor.m_InputToForgetWeights)) |
| , m_InputToCellWeightsTensor (AssignScopedTensorHandle(descriptor.m_InputToCellWeights)) |
| , m_InputToOutputWeightsTensor (AssignScopedTensorHandle(descriptor.m_InputToOutputWeights)) |
| , m_RecurrentToInputWeightsTensor (AssignScopedTensorHandle(descriptor.m_RecurrentToInputWeights)) |
| , m_RecurrentToForgetWeightsTensor(AssignScopedTensorHandle(descriptor.m_RecurrentToForgetWeights)) |
| , m_RecurrentToCellWeightsTensor (AssignScopedTensorHandle(descriptor.m_RecurrentToCellWeights)) |
| , m_RecurrentToOutputWeightsTensor(AssignScopedTensorHandle(descriptor.m_RecurrentToOutputWeights)) |
| , m_CellToInputWeightsTensor (AssignScopedTensorHandle(descriptor.m_CellToInputWeights)) |
| , m_CellToForgetWeightsTensor (AssignScopedTensorHandle(descriptor.m_CellToForgetWeights)) |
| , m_CellToOutputWeightsTensor (AssignScopedTensorHandle(descriptor.m_CellToOutputWeights)) |
| , m_InputGateBiasTensor (AssignScopedTensorHandle(descriptor.m_InputGateBias)) |
| , m_ForgetGateBiasTensor (AssignScopedTensorHandle(descriptor.m_ForgetGateBias)) |
| , m_CellBiasTensor (AssignScopedTensorHandle(descriptor.m_CellBias)) |
| , m_OutputGateBiasTensor (AssignScopedTensorHandle(descriptor.m_OutputGateBias)) |
| , m_ProjectionWeightsTensor (AssignScopedTensorHandle(descriptor.m_ProjectionWeights)) |
| , m_ProjectionBiasTensor (AssignScopedTensorHandle(descriptor.m_ProjectionBias)) |
| , m_InputLayerNormWeights (AssignScopedTensorHandle(descriptor.m_InputLayerNormWeights)) |
| , m_ForgetLayerNormWeights (AssignScopedTensorHandle(descriptor.m_ForgetLayerNormWeights)) |
| , m_CellLayerNormWeights (AssignScopedTensorHandle(descriptor.m_CellLayerNormWeights)) |
| , m_OutputLayerNormWeights (AssignScopedTensorHandle(descriptor.m_OutputLayerNormWeights)) |
| {} |
| |
| void RefLstmWorkload::Execute() const |
| { |
| Execute(m_Data.m_Inputs, m_Data.m_Outputs); |
| } |
| |
| void RefLstmWorkload::ExecuteAsync(ExecutionData& executionData) |
| { |
| WorkingMemDescriptor* workingMemDescriptor = static_cast<WorkingMemDescriptor*>(executionData.m_Data); |
| Execute(workingMemDescriptor->m_Inputs, workingMemDescriptor->m_Outputs); |
| } |
| |
| void RefLstmWorkload::Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const |
| { |
| ARMNN_SCOPED_PROFILING_EVENT_REF_NAME_GUID("RefLstmWorkload_Execute"); |
| |
| // This is a porting of the LSTM::Eval() method in the Android code base |
| // Refer to: android/frameworks/ml/nn/common/operations/LSTM.cpp |
| |
| const TensorInfo& inputInfo = GetTensorInfo(inputs[0]); |
| const TensorInfo& outputInfo = GetTensorInfo(outputs[0]); |
| |
| const TensorShape& inputShape = inputInfo.GetShape(); |
| |
| std::unique_ptr<Encoder<float>> outputStateOut = MakeEncoder<float>(outputInfo, outputs[1]->Map()); |
| std::unique_ptr<Encoder<float>> cellStateOut = MakeEncoder<float>(outputInfo, outputs[2]->Map()); |
| std::unique_ptr<Encoder<float>> output = MakeEncoder<float>(outputInfo, outputs[3]->Map()); |
| |
| std::unique_ptr<Decoder<float>> cellStateOutDecoder = MakeDecoder<float>(outputInfo, outputs[2]->Map()); |
| std::unique_ptr<Decoder<float>> outputDecoder = MakeDecoder<float>(outputInfo, outputs[3]->Map()); |
| |
| std::unique_ptr<Decoder<float>> inputData = MakeDecoder<float>(inputInfo, inputs[0]->Map()); |
| std::unique_ptr<Decoder<float>> outputStateIn = MakeDecoder<float>(inputInfo, inputs[1]->Map()); |
| std::unique_ptr<Decoder<float>> cellStateIn = MakeDecoder<float>(inputInfo, inputs[2]->Map()); |
| |
| const uint32_t nBatch = inputShape[0]; |
| const uint32_t nCell = m_InputToOutputWeightsTensor->GetShape()[0]; |
| |
| const bool useCifg = m_Data.m_Parameters.m_CifgEnabled; |
| const bool usePeephole = m_Data.m_Parameters.m_PeepholeEnabled; |
| const bool useLayerNorm = m_Data.m_Parameters.m_LayerNormEnabled; |
| |
| // Index the scratch buffers pointers to the global scratch buffer. |
| std::unique_ptr<Encoder<float>> inputGateScratch = MakeEncoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Encoder<float>> cellScratch = MakeEncoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Encoder<float>> forgetGateScratch = MakeEncoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Encoder<float>> outputGateScratch = MakeEncoder<float>(outputInfo, outputs[0]->Map()); |
| |
| std::unique_ptr<Decoder<float>> inputGateScratchDecoder = |
| MakeDecoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Decoder<float>> cellScratchDecoder = |
| MakeDecoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Decoder<float>> forgetGateScratchDecoder = |
| MakeDecoder<float>(outputInfo, outputs[0]->Map()); |
| std::unique_ptr<Decoder<float>> outputGateScratchDecoder = |
| MakeDecoder<float>(outputInfo, outputs[0]->Map()); |
| |
| if (useCifg) |
| { |
| *cellScratch += (0 * nCell * nBatch); |
| *forgetGateScratch += (1 * nCell * nBatch); |
| *outputGateScratch += (2 * nCell * nBatch); |
| |
| *cellScratchDecoder += (0 * nCell * nBatch); |
| *forgetGateScratchDecoder += (1 * nCell * nBatch); |
| *outputGateScratchDecoder += (2 * nCell * nBatch); |
| } |
| else |
| { |
| *inputGateScratch += (0 * nCell * nBatch); |
| *cellScratch += (1 * nCell * nBatch); |
| *forgetGateScratch += (2 * nCell * nBatch); |
| *outputGateScratch += (3 * nCell * nBatch); |
| |
| *inputGateScratchDecoder += (0 * nCell * nBatch); |
| *cellScratchDecoder += (1 * nCell * nBatch); |
| *forgetGateScratchDecoder += (2 * nCell * nBatch); |
| *outputGateScratchDecoder += (3 * nCell * nBatch); |
| } |
| |
| std::unique_ptr<Decoder<float>> inputToInputWeightsTensor; |
| std::unique_ptr<Decoder<float>> inputToForgetWeightsTensor = MakeDecoder<float>( |
| m_InputToForgetWeightsTensor->GetTensorInfo(), m_InputToForgetWeightsTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> inputToCellWeightsTensor = MakeDecoder<float>( |
| m_InputToCellWeightsTensor->GetTensorInfo(), m_InputToCellWeightsTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> inputToOutputWeightsTensor = MakeDecoder<float>( |
| m_InputToOutputWeightsTensor->GetTensorInfo(), m_InputToOutputWeightsTensor->GetConstTensor<void>()); |
| |
| std::unique_ptr<Decoder<float>> recurrentToInputWeightsTensor; |
| std::unique_ptr<Decoder<float>> recurrentToForgetWeightsTensor = MakeDecoder<float>( |
| m_RecurrentToForgetWeightsTensor->GetTensorInfo(), m_RecurrentToForgetWeightsTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> recurrentToCellWeightsTensor = MakeDecoder<float>( |
| m_RecurrentToCellWeightsTensor->GetTensorInfo(), m_RecurrentToCellWeightsTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> recurrentToOutputWeightsTensor = MakeDecoder<float>( |
| m_RecurrentToOutputWeightsTensor->GetTensorInfo(), m_RecurrentToOutputWeightsTensor->GetConstTensor<void>()); |
| |
| std::unique_ptr<Decoder<float>> inputGateBiasTensor; |
| std::unique_ptr<Decoder<float>> forgetGateBiasTensor = MakeDecoder<float>( |
| m_ForgetGateBiasTensor->GetTensorInfo(), m_ForgetGateBiasTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> cellBiasTensor = MakeDecoder<float>( |
| m_CellBiasTensor->GetTensorInfo(), m_CellBiasTensor->GetConstTensor<void>()); |
| std::unique_ptr<Decoder<float>> outputGateBiasTensor = MakeDecoder<float>( |
| m_OutputGateBiasTensor->GetTensorInfo(), m_OutputGateBiasTensor->GetConstTensor<void>()); |
| |
| std::unique_ptr<Decoder<float>> cellToInputWeightsTensor; |
| std::unique_ptr<Decoder<float>> cellToForgetWeightsTensor; |
| std::unique_ptr<Decoder<float>> cellToOutputWeightsTensor; |
| |
| std::unique_ptr<Decoder<float>> projectionWeightsTensor; |
| std::unique_ptr<Decoder<float>> projectionBiasTensor; |
| |
| std::unique_ptr<Decoder<float>> inputLayerNormWeights; |
| std::unique_ptr<Decoder<float>> forgetLayerNormWeights; |
| std::unique_ptr<Decoder<float>> cellLayerNormWeights; |
| std::unique_ptr<Decoder<float>> outputLayerNormWeights; |
| |
| const TensorShape& inputToOutputWeightsShape = m_InputToOutputWeightsTensor->GetShape(); |
| const TensorShape& recurrentToOutputWeightsShape = m_RecurrentToOutputWeightsTensor->GetShape(); |
| |
| if (useLayerNorm) |
| { |
| if (!useCifg) |
| { |
| inputLayerNormWeights = MakeDecoder<float>( |
| m_InputLayerNormWeights->GetTensorInfo(), m_InputLayerNormWeights->GetConstTensor<void>()); |
| } |
| forgetLayerNormWeights = MakeDecoder<float>( |
| m_ForgetLayerNormWeights->GetTensorInfo(), m_ForgetLayerNormWeights->GetConstTensor<void>()); |
| cellLayerNormWeights = MakeDecoder<float>( |
| m_CellLayerNormWeights->GetTensorInfo(), m_CellLayerNormWeights->GetConstTensor<void>()); |
| outputLayerNormWeights = MakeDecoder<float>( |
| m_OutputLayerNormWeights->GetTensorInfo(), m_OutputLayerNormWeights->GetConstTensor<void>()); |
| } |
| |
| if (!useCifg) |
| { |
| inputToInputWeightsTensor = MakeDecoder<float>( |
| m_InputToInputWeightsTensor->GetTensorInfo(), m_InputToInputWeightsTensor->GetConstTensor<void>()); |
| inputGateBiasTensor = MakeDecoder<float>( |
| m_InputGateBiasTensor->GetTensorInfo(), m_InputGateBiasTensor->GetConstTensor<void>()); |
| recurrentToInputWeightsTensor = MakeDecoder<float>( |
| m_RecurrentToInputWeightsTensor->GetTensorInfo(), m_RecurrentToInputWeightsTensor->GetConstTensor<void>()); |
| } |
| |
| if (usePeephole) |
| { |
| cellToForgetWeightsTensor = MakeDecoder<float>( |
| m_CellToForgetWeightsTensor->GetTensorInfo(), m_CellToForgetWeightsTensor->GetConstTensor<void>()); |
| cellToOutputWeightsTensor = MakeDecoder<float>( |
| m_CellToOutputWeightsTensor->GetTensorInfo(), m_CellToOutputWeightsTensor->GetConstTensor<void>()); |
| } |
| |
| if (!useCifg && usePeephole) |
| { |
| cellToInputWeightsTensor = MakeDecoder<float>( |
| m_CellToInputWeightsTensor->GetTensorInfo(), m_CellToInputWeightsTensor->GetConstTensor<void>()); |
| } |
| |
| if (m_Data.m_Parameters.m_ProjectionEnabled) |
| { |
| projectionWeightsTensor = MakeDecoder<float>( |
| m_ProjectionWeightsTensor->GetTensorInfo(), m_ProjectionWeightsTensor->GetConstTensor<void>()); |
| if (m_ProjectionBiasTensor) |
| { |
| projectionBiasTensor = MakeDecoder<float>( |
| m_ProjectionBiasTensor->GetTensorInfo(), m_ProjectionBiasTensor->GetConstTensor<void>()); |
| } |
| } |
| |
| LstmImpl(m_Data.m_Parameters, |
| inputInfo, |
| outputInfo, |
| inputToOutputWeightsShape, |
| recurrentToOutputWeightsShape, |
| inputData, |
| outputStateIn, |
| cellStateIn, |
| outputStateOut, |
| cellStateOut, |
| output, |
| cellStateOutDecoder, |
| outputDecoder, |
| inputToInputWeightsTensor, |
| inputToForgetWeightsTensor, |
| inputToCellWeightsTensor, |
| inputToOutputWeightsTensor, |
| recurrentToInputWeightsTensor, |
| recurrentToForgetWeightsTensor, |
| recurrentToCellWeightsTensor, |
| recurrentToOutputWeightsTensor, |
| cellToInputWeightsTensor, |
| cellToForgetWeightsTensor, |
| cellToOutputWeightsTensor, |
| inputGateBiasTensor, |
| forgetGateBiasTensor, |
| cellBiasTensor, |
| outputGateBiasTensor, |
| projectionWeightsTensor, |
| projectionBiasTensor, |
| inputLayerNormWeights, |
| forgetLayerNormWeights, |
| cellLayerNormWeights, |
| outputLayerNormWeights, |
| inputGateScratch, |
| cellScratch, |
| forgetGateScratch, |
| outputGateScratch, |
| inputGateScratchDecoder, |
| cellScratchDecoder, |
| forgetGateScratchDecoder, |
| outputGateScratchDecoder, |
| m_LayerNormEpsilon); |
| } |
| |
| } //namespace armnn |