blob: ebe320718c28ef80bd8bd4885e567ceb9fdd4187 [file] [log] [blame]
//
// Copyright © 2017,2019-2024 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "QuantizedLstmLayer.hpp"
#include "LayerCloneBase.hpp"
#include <armnn/QuantizedLstmParams.hpp>
#include <armnn/TypesUtils.hpp>
#include <armnn/backends/TensorHandle.hpp>
#include <armnn/backends/WorkloadFactory.hpp>
namespace armnn
{
QuantizedLstmLayer::QuantizedLstmLayer(const char* name)
: Layer(3, 2, LayerType::QuantizedLstm, name)
{
}
std::unique_ptr<IWorkload> QuantizedLstmLayer::CreateWorkload(const IWorkloadFactory& factory) const
{
QuantizedLstmQueueDescriptor descriptor;
// QuantizedLstmLayer parameters - there are no optional params
descriptor.m_InputToInputWeights = m_QuantizedLstmParameters.m_InputToInputWeights.get();
descriptor.m_InputToForgetWeights = m_QuantizedLstmParameters.m_InputToForgetWeights.get();
descriptor.m_InputToCellWeights = m_QuantizedLstmParameters.m_InputToCellWeights.get();
descriptor.m_InputToOutputWeights = m_QuantizedLstmParameters.m_InputToOutputWeights.get();
descriptor.m_RecurrentToInputWeights = m_QuantizedLstmParameters.m_RecurrentToInputWeights.get();
descriptor.m_RecurrentToForgetWeights = m_QuantizedLstmParameters.m_RecurrentToForgetWeights.get();
descriptor.m_RecurrentToCellWeights = m_QuantizedLstmParameters.m_RecurrentToCellWeights.get();
descriptor.m_RecurrentToOutputWeights = m_QuantizedLstmParameters.m_RecurrentToOutputWeights.get();
descriptor.m_InputGateBias = m_QuantizedLstmParameters.m_InputGateBias.get();
descriptor.m_ForgetGateBias = m_QuantizedLstmParameters.m_ForgetGateBias.get();
descriptor.m_CellBias = m_QuantizedLstmParameters.m_CellBias.get();
descriptor.m_OutputGateBias = m_QuantizedLstmParameters.m_OutputGateBias.get();
SetAdditionalInfo(descriptor);
return factory.CreateWorkload(LayerType::QuantizedLstm, descriptor, PrepInfoAndDesc(descriptor));
}
QuantizedLstmLayer* QuantizedLstmLayer::Clone(Graph& graph) const
{
auto layer = CloneBase<QuantizedLstmLayer>(graph, GetName());
layer->m_QuantizedLstmParameters.m_InputToInputWeights = m_QuantizedLstmParameters.m_InputToInputWeights ?
m_QuantizedLstmParameters.m_InputToInputWeights : nullptr;
layer->m_QuantizedLstmParameters.m_InputToForgetWeights = m_QuantizedLstmParameters.m_InputToForgetWeights ?
m_QuantizedLstmParameters.m_InputToForgetWeights : nullptr;
layer->m_QuantizedLstmParameters.m_InputToCellWeights = m_QuantizedLstmParameters.m_InputToCellWeights ?
m_QuantizedLstmParameters.m_InputToCellWeights : nullptr;
layer->m_QuantizedLstmParameters.m_InputToOutputWeights = m_QuantizedLstmParameters.m_InputToOutputWeights ?
m_QuantizedLstmParameters.m_InputToOutputWeights : nullptr;
layer->m_QuantizedLstmParameters.m_RecurrentToInputWeights = m_QuantizedLstmParameters.m_RecurrentToInputWeights ?
m_QuantizedLstmParameters.m_RecurrentToInputWeights : nullptr;
layer->m_QuantizedLstmParameters.m_RecurrentToForgetWeights = m_QuantizedLstmParameters.m_RecurrentToForgetWeights
? m_QuantizedLstmParameters.m_RecurrentToForgetWeights : nullptr;
layer->m_QuantizedLstmParameters.m_RecurrentToCellWeights = m_QuantizedLstmParameters.m_RecurrentToCellWeights ?
m_QuantizedLstmParameters.m_RecurrentToCellWeights : nullptr;
layer->m_QuantizedLstmParameters.m_RecurrentToOutputWeights = m_QuantizedLstmParameters.m_RecurrentToOutputWeights
? m_QuantizedLstmParameters.m_RecurrentToOutputWeights : nullptr;
layer->m_QuantizedLstmParameters.m_InputGateBias = m_QuantizedLstmParameters.m_InputGateBias ?
m_QuantizedLstmParameters.m_InputGateBias : nullptr;
layer->m_QuantizedLstmParameters.m_ForgetGateBias = m_QuantizedLstmParameters.m_ForgetGateBias ?
m_QuantizedLstmParameters.m_ForgetGateBias : nullptr;
layer->m_QuantizedLstmParameters.m_CellBias = m_QuantizedLstmParameters.m_CellBias ?
m_QuantizedLstmParameters.m_CellBias : nullptr;
layer->m_QuantizedLstmParameters.m_OutputGateBias = m_QuantizedLstmParameters.m_OutputGateBias ?
m_QuantizedLstmParameters.m_OutputGateBias : nullptr;
return std::move(layer);
}
std::vector<TensorShape> QuantizedLstmLayer::InferOutputShapes(const std::vector<TensorShape>& inputShapes) const
{
if (inputShapes.size() != 3)
{
throw armnn::Exception("inputShapes' size is \"" + std::to_string(inputShapes.size()) +
"\" - should be \"3\".");
}
// Get input values for validation
unsigned int numBatches = inputShapes[0][0];
unsigned int outputSize = inputShapes[1][1];
std::vector<TensorShape> outShapes;
outShapes.push_back(TensorShape({numBatches, outputSize})); // cellStateOut
outShapes.push_back(TensorShape({numBatches, outputSize})); // output
return outShapes;
}
void QuantizedLstmLayer::ValidateTensorShapesFromInputs()
{
VerifyLayerConnections(3, CHECK_LOCATION());
const TensorShape& outputShape = GetOutputSlot(0).GetTensorInfo().GetShape();
VerifyShapeInferenceType(outputShape, m_ShapeInferenceMethod);
auto inferredShapes = InferOutputShapes(
{
GetInputSlot(0).GetTensorInfo().GetShape(), // input
GetInputSlot(1).GetTensorInfo().GetShape(), // previousCellStateIn
GetInputSlot(2).GetTensorInfo().GetShape() // previousOutputIn
});
if (inferredShapes.size() != 2)
{
throw armnn::LayerValidationException("inferredShapes has "
+ std::to_string(inferredShapes.size()) +
" element(s) - should only have 2.");
}
// Check weights and bias for nullptr
if (!m_QuantizedLstmParameters.m_InputToInputWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_InputToInputWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_InputToForgetWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_InputToForgetWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_InputToCellWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_InputToCellWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_InputToOutputWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_InputToOutputWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_RecurrentToInputWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_RecurrentToInputWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_RecurrentToForgetWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_RecurrentToForgetWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_RecurrentToCellWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_RecurrentToCellWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_RecurrentToOutputWeights)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_RecurrentToOutputWeights "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_InputGateBias)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_InputGateBias "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_ForgetGateBias)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_ForgetGateBias "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_CellBias)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_CellBias "
"should not be null.");
}
if (!m_QuantizedLstmParameters.m_OutputGateBias)
{
throw armnn::LayerValidationException("QuantizedLstmLayer: "
"m_QuantizedLstmParameters.m_OutputGateBias "
"should not be null.");
}
// Check output TensorShape(s) match inferred shape
ValidateAndCopyShape(outputShape, inferredShapes[0], m_ShapeInferenceMethod, "QuantizedLstmLayer");
ValidateAndCopyShape(GetOutputSlot(1).GetTensorInfo().GetShape(),
inferredShapes[1],
m_ShapeInferenceMethod,
"QuantizedLstmLayer",
1);
}
Layer::ImmutableConstantTensors QuantizedLstmLayer::GetConstantTensorsByRef() const
{
// For API stability DO NOT ALTER order and add new members to the end of vector
return
{
m_QuantizedLstmParameters.m_InputToInputWeights,
m_QuantizedLstmParameters.m_InputToForgetWeights,
m_QuantizedLstmParameters.m_InputToCellWeights,
m_QuantizedLstmParameters.m_InputToOutputWeights,
m_QuantizedLstmParameters.m_RecurrentToInputWeights,
m_QuantizedLstmParameters.m_RecurrentToForgetWeights,
m_QuantizedLstmParameters.m_RecurrentToCellWeights,
m_QuantizedLstmParameters.m_RecurrentToOutputWeights,
m_QuantizedLstmParameters.m_InputGateBias,
m_QuantizedLstmParameters.m_ForgetGateBias,
m_QuantizedLstmParameters.m_CellBias,
m_QuantizedLstmParameters.m_OutputGateBias
};
}
void QuantizedLstmLayer::ExecuteStrategy(IStrategy& strategy) const
{
std::vector<ConstTensor> constTensors;
ManagedConstTensorHandle managedInputToInputWeights(m_QuantizedLstmParameters.m_InputToInputWeights);
ManagedConstTensorHandle managedInputToForgetWeights(m_QuantizedLstmParameters.m_InputToForgetWeights);
ManagedConstTensorHandle managedInputToCellWeights(m_QuantizedLstmParameters.m_InputToCellWeights);
ManagedConstTensorHandle managedInputToOutputWeights(m_QuantizedLstmParameters.m_InputToOutputWeights);
ManagedConstTensorHandle managedRecurrentToInputWeights(m_QuantizedLstmParameters.m_RecurrentToInputWeights);
ManagedConstTensorHandle managedRecurrentToForgetWeights(m_QuantizedLstmParameters.m_RecurrentToForgetWeights);
ManagedConstTensorHandle managedRecurrentToCellWeights(m_QuantizedLstmParameters.m_RecurrentToCellWeights);
ManagedConstTensorHandle managedRecurrentToOutputWeights(m_QuantizedLstmParameters.m_RecurrentToOutputWeights);
ManagedConstTensorHandle managedInputGateBias(m_QuantizedLstmParameters.m_InputGateBias);
ManagedConstTensorHandle managedForgetGateBias(m_QuantizedLstmParameters.m_ForgetGateBias);
ManagedConstTensorHandle managedCellBias(m_QuantizedLstmParameters.m_CellBias);
ManagedConstTensorHandle managedOutputGateBias(m_QuantizedLstmParameters.m_OutputGateBias);
// InputToX weight tensors
if (m_QuantizedLstmParameters.m_InputToInputWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(managedInputToInputWeights.GetTensorInfo(),
managedInputToInputWeights.Map()));
}
if (m_QuantizedLstmParameters.m_InputToForgetWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(managedInputToForgetWeights.GetTensorInfo(),
managedInputToForgetWeights.Map()));
}
if (m_QuantizedLstmParameters.m_InputToCellWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(managedInputToCellWeights.GetTensorInfo(),
managedInputToCellWeights.Map()));
}
if (m_QuantizedLstmParameters.m_InputToOutputWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(managedInputToOutputWeights.GetTensorInfo(),
managedInputToOutputWeights.Map()));
}
// RecurrentToX weight tensors
if (m_QuantizedLstmParameters.m_RecurrentToInputWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(
managedRecurrentToInputWeights.GetTensorInfo(),
managedRecurrentToInputWeights.Map()));
}
if (m_QuantizedLstmParameters.m_RecurrentToForgetWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(
managedRecurrentToForgetWeights.GetTensorInfo(),
managedRecurrentToForgetWeights.Map()));
}
if (m_QuantizedLstmParameters.m_RecurrentToCellWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(
managedRecurrentToCellWeights.GetTensorInfo(),
managedRecurrentToCellWeights.Map()));
}
if (m_QuantizedLstmParameters.m_RecurrentToOutputWeights != nullptr)
{
constTensors.emplace_back(ConstTensor(
managedRecurrentToOutputWeights.GetTensorInfo(),
managedRecurrentToOutputWeights.Map()));
}
// Bias tensors
if (m_QuantizedLstmParameters.m_InputGateBias != nullptr)
{
constTensors.emplace_back(ConstTensor(managedInputGateBias.GetTensorInfo(),
managedInputGateBias.Map()));
}
if (m_QuantizedLstmParameters.m_ForgetGateBias != nullptr)
{
constTensors.emplace_back(ConstTensor(managedForgetGateBias.GetTensorInfo(),
managedForgetGateBias.Map()));
}
if (m_QuantizedLstmParameters.m_CellBias != nullptr)
{
constTensors.emplace_back(ConstTensor(managedCellBias.GetTensorInfo(),
managedCellBias.Map()));
}
if (m_QuantizedLstmParameters.m_OutputGateBias != nullptr)
{
constTensors.emplace_back(ConstTensor(managedOutputGateBias.GetTensorInfo(),
managedOutputGateBias.Map()));
}
strategy.ExecuteStrategy(this, BaseDescriptor(), constTensors, GetName());
}
} // namespace armnn