| // |
| // Copyright © 2017 Arm Ltd. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| |
| #include "Deserializer.hpp" |
| |
| #include <armnn/ArmNN.hpp> |
| #include <armnn/Exceptions.hpp> |
| |
| #include <armnnUtils/Permute.hpp> |
| |
| #include <ParserHelper.hpp> |
| #include <VerificationHelpers.hpp> |
| |
| #include <boost/filesystem.hpp> |
| #include <boost/format.hpp> |
| #include <boost/core/ignore_unused.hpp> |
| #include <boost/assert.hpp> |
| #include <boost/format.hpp> |
| #include <boost/format.hpp> |
| #include <boost/numeric/conversion/cast.hpp> |
| #include <boost/polymorphic_cast.hpp> |
| |
| #include <fstream> |
| #include <algorithm> |
| #include <limits> |
| #include <numeric> |
| |
| using armnn::ParseException; |
| using namespace armnn; |
| using namespace armnnSerializer; |
| |
| namespace armnnDeserializer |
| { |
| |
| namespace |
| { |
| |
| const uint32_t VIRTUAL_LAYER_ID = std::numeric_limits<uint32_t>::max(); |
| |
| void CheckGraph(const Deserializer::GraphPtr& graph, |
| unsigned int layersIndex, |
| const CheckLocation& location) |
| { |
| if (graph->layers() == nullptr) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with invalid (null) graph. " |
| "Possible reason is that the graph is not yet loaded and Unpack(ed). " |
| "layers:%2% at %3%") % |
| location.m_Function % |
| layersIndex % |
| location.FileLine())); |
| } |
| else if (layersIndex >= graph->layers()->size()) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with an invalid layers index. " |
| "layers:%2% at %3%") % |
| location.m_Function % |
| layersIndex % |
| location.FileLine())); |
| } |
| } |
| |
| void CheckLayers(const Deserializer::GraphPtr& graph, |
| unsigned int layersIndex, |
| unsigned int layerIndex, |
| const CheckLocation& location) |
| { |
| if (graph->layers() == nullptr) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with invalid (null) graph. " |
| "Possible reason is that the graph is not yet loaded and Unpack(ed). " |
| "layers:%2% at %3%") % |
| location.m_Function % |
| layersIndex % |
| location.FileLine())); |
| } |
| else if (layersIndex >= graph->layers()->size()) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with an invalid layers index. " |
| "layers:%2% at %3%") % |
| location.m_Function % |
| layersIndex % |
| location.FileLine())); |
| } |
| else if (layerIndex >= graph->layers()[layersIndex].size() |
| && layerIndex != VIRTUAL_LAYER_ID) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with an invalid layer index. " |
| "layers:%2% layer:%3% at %4%") % |
| location.m_Function % |
| layersIndex % |
| layerIndex % |
| location.FileLine())); |
| } |
| } |
| |
| void CheckTensorPtr(Deserializer::TensorRawPtr rawPtr, |
| const CheckLocation& location) |
| { |
| if (rawPtr == nullptr) |
| { |
| throw ParseException( |
| boost::str( |
| boost::format("%1% was called with a null tensor pointer. " |
| "at %2%") % |
| location.m_Function % |
| location.FileLine())); |
| |
| } |
| } |
| |
| void CheckConstTensorPtr(Deserializer::ConstTensorRawPtr rawPtr, |
| const CheckLocation& location) |
| { |
| if (rawPtr == nullptr) |
| { |
| throw ParseException(boost::str(boost::format("%1% was called with a null const tensor pointer. at %2%") % |
| location.m_Function % |
| location.FileLine())); |
| } |
| } |
| |
| void CheckConstTensorSize(const unsigned int constTensorSize, |
| const unsigned int tensorSize, |
| const CheckLocation& location) |
| { |
| if (constTensorSize != tensorSize) |
| { |
| throw ParseException(boost::str(boost::format("%1% wrong number of components supplied to tensor. at:%2%") % |
| location.m_Function % |
| location.FileLine())); |
| } |
| } |
| |
| #define CHECK_TENSOR_PTR(TENSOR_PTR) \ |
| CheckTensorPtr(TENSOR_PTR, CHECK_LOCATION()) |
| |
| #define CHECK_CONST_TENSOR_SIZE(CONST_TENSOR_SIZE, TENSOR_SIZE) \ |
| CheckConstTensorSize(CONST_TENSOR_SIZE, TENSOR_SIZE, CHECK_LOCATION()) |
| |
| #define CHECK_CONST_TENSOR_PTR(TENSOR_PTR) \ |
| CheckConstTensorPtr(TENSOR_PTR, CHECK_LOCATION()) |
| |
| #define CHECK_LAYERS(GRAPH, LAYERS_INDEX, LAYER_INDEX) \ |
| CheckLayers(GRAPH, LAYERS_INDEX, LAYER_INDEX, CHECK_LOCATION()) |
| |
| #define CHECK_GRAPH(GRAPH, LAYERS_INDEX) \ |
| CheckGraph(GRAPH, LAYERS_INDEX, CHECK_LOCATION()) |
| } |
| |
| bool CheckShape(const armnn::TensorShape& actual, const std::vector<uint32_t>& expected) |
| { |
| const unsigned int actualSize = actual.GetNumDimensions(); |
| if (actualSize != expected.size()) |
| { |
| return false; |
| } |
| |
| for (unsigned int i = 0u; i < actualSize; i++) |
| { |
| if (actual[i] != static_cast<unsigned int>(expected[i])) |
| { |
| return false; |
| } |
| } |
| |
| return true; |
| } |
| |
| Deserializer::Deserializer() |
| : m_Network(nullptr, nullptr), |
| //May require LayerType_Max to be included |
| m_ParserFunctions(Layer_MAX+1, &Deserializer::ParseUnsupportedLayer) |
| { |
| // register supported layers |
| m_ParserFunctions[Layer_AbsLayer] = &Deserializer::ParseAbs; |
| m_ParserFunctions[Layer_ActivationLayer] = &Deserializer::ParseActivation; |
| m_ParserFunctions[Layer_AdditionLayer] = &Deserializer::ParseAdd; |
| m_ParserFunctions[Layer_ArgMinMaxLayer] = &Deserializer::ParseArgMinMax; |
| m_ParserFunctions[Layer_BatchToSpaceNdLayer] = &Deserializer::ParseBatchToSpaceNd; |
| m_ParserFunctions[Layer_BatchNormalizationLayer] = &Deserializer::ParseBatchNormalization; |
| m_ParserFunctions[Layer_ComparisonLayer] = &Deserializer::ParseComparison; |
| m_ParserFunctions[Layer_ConcatLayer] = &Deserializer::ParseConcat; |
| m_ParserFunctions[Layer_ConstantLayer] = &Deserializer::ParseConstant; |
| m_ParserFunctions[Layer_Convolution2dLayer] = &Deserializer::ParseConvolution2d; |
| m_ParserFunctions[Layer_DepthToSpaceLayer] = &Deserializer::ParseDepthToSpace; |
| m_ParserFunctions[Layer_DepthwiseConvolution2dLayer] = &Deserializer::ParseDepthwiseConvolution2d; |
| m_ParserFunctions[Layer_DequantizeLayer] = &Deserializer::ParseDequantize; |
| m_ParserFunctions[Layer_DetectionPostProcessLayer] = &Deserializer::ParseDetectionPostProcess; |
| m_ParserFunctions[Layer_DivisionLayer] = &Deserializer::ParseDivision; |
| m_ParserFunctions[Layer_EqualLayer] = &Deserializer::ParseEqual; |
| m_ParserFunctions[Layer_FullyConnectedLayer] = &Deserializer::ParseFullyConnected; |
| m_ParserFunctions[Layer_FloorLayer] = &Deserializer::ParseFloor; |
| m_ParserFunctions[Layer_GatherLayer] = &Deserializer::ParseGather; |
| m_ParserFunctions[Layer_GreaterLayer] = &Deserializer::ParseGreater; |
| m_ParserFunctions[Layer_InstanceNormalizationLayer] = &Deserializer::ParseInstanceNormalization; |
| m_ParserFunctions[Layer_L2NormalizationLayer] = &Deserializer::ParseL2Normalization; |
| m_ParserFunctions[Layer_LogSoftmaxLayer] = &Deserializer::ParseLogSoftmax; |
| m_ParserFunctions[Layer_LstmLayer] = &Deserializer::ParseLstm; |
| m_ParserFunctions[Layer_MaximumLayer] = &Deserializer::ParseMaximum; |
| m_ParserFunctions[Layer_MeanLayer] = &Deserializer::ParseMean; |
| m_ParserFunctions[Layer_MinimumLayer] = &Deserializer::ParseMinimum; |
| m_ParserFunctions[Layer_MergeLayer] = &Deserializer::ParseMerge; |
| m_ParserFunctions[Layer_MergerLayer] = &Deserializer::ParseConcat; |
| m_ParserFunctions[Layer_MultiplicationLayer] = &Deserializer::ParseMultiplication; |
| m_ParserFunctions[Layer_NormalizationLayer] = &Deserializer::ParseNormalization; |
| m_ParserFunctions[Layer_PadLayer] = &Deserializer::ParsePad; |
| m_ParserFunctions[Layer_PermuteLayer] = &Deserializer::ParsePermute; |
| m_ParserFunctions[Layer_Pooling2dLayer] = &Deserializer::ParsePooling2d; |
| m_ParserFunctions[Layer_PreluLayer] = &Deserializer::ParsePrelu; |
| m_ParserFunctions[Layer_QuantizeLayer] = &Deserializer::ParseQuantize; |
| m_ParserFunctions[Layer_QuantizedLstmLayer] = &Deserializer::ParseQuantizedLstm; |
| m_ParserFunctions[Layer_ReshapeLayer] = &Deserializer::ParseReshape; |
| m_ParserFunctions[Layer_ResizeBilinearLayer] = &Deserializer::ParseResizeBilinear; |
| m_ParserFunctions[Layer_ResizeLayer] = &Deserializer::ParseResize; |
| m_ParserFunctions[Layer_RsqrtLayer] = &Deserializer::ParseRsqrt; |
| m_ParserFunctions[Layer_SliceLayer] = &Deserializer::ParseSlice; |
| m_ParserFunctions[Layer_SoftmaxLayer] = &Deserializer::ParseSoftmax; |
| m_ParserFunctions[Layer_SpaceToBatchNdLayer] = &Deserializer::ParseSpaceToBatchNd; |
| m_ParserFunctions[Layer_SpaceToDepthLayer] = &Deserializer::ParseSpaceToDepth; |
| m_ParserFunctions[Layer_SplitterLayer] = &Deserializer::ParseSplitter; |
| m_ParserFunctions[Layer_StackLayer] = &Deserializer::ParseStack; |
| m_ParserFunctions[Layer_StandInLayer] = &Deserializer::ParseStandIn; |
| m_ParserFunctions[Layer_StridedSliceLayer] = &Deserializer::ParseStridedSlice; |
| m_ParserFunctions[Layer_SubtractionLayer] = &Deserializer::ParseSubtraction; |
| m_ParserFunctions[Layer_SwitchLayer] = &Deserializer::ParseSwitch; |
| m_ParserFunctions[Layer_TransposeConvolution2dLayer] = &Deserializer::ParseTransposeConvolution2d; |
| } |
| |
| Deserializer::LayerBaseRawPtr Deserializer::GetBaseLayer(const GraphPtr& graphPtr, unsigned int layerIndex) |
| { |
| auto layerType = graphPtr->layers()->Get(layerIndex)->layer_type(); |
| |
| switch(layerType) |
| { |
| case Layer::Layer_AbsLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_AbsLayer()->base(); |
| case Layer::Layer_ActivationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ActivationLayer()->base(); |
| case Layer::Layer_AdditionLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_AdditionLayer()->base(); |
| case Layer::Layer_ArgMinMaxLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ArgMinMaxLayer()->base(); |
| case Layer::Layer_BatchToSpaceNdLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_BatchToSpaceNdLayer()->base(); |
| case Layer::Layer_BatchNormalizationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_BatchNormalizationLayer()->base(); |
| case Layer::Layer_ComparisonLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ComparisonLayer()->base(); |
| case Layer::Layer_ConcatLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ConcatLayer()->base(); |
| case Layer::Layer_ConstantLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ConstantLayer()->base(); |
| case Layer::Layer_Convolution2dLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_Convolution2dLayer()->base(); |
| case Layer::Layer_DepthToSpaceLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_DepthToSpaceLayer()->base(); |
| case Layer::Layer_DepthwiseConvolution2dLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_DepthwiseConvolution2dLayer()->base(); |
| case Layer::Layer_DequantizeLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_DequantizeLayer()->base(); |
| case Layer::Layer_DetectionPostProcessLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_DetectionPostProcessLayer()->base(); |
| case Layer::Layer_DivisionLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_DivisionLayer()->base(); |
| case Layer::Layer_EqualLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_EqualLayer()->base(); |
| case Layer::Layer_FullyConnectedLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_FullyConnectedLayer()->base(); |
| case Layer::Layer_FloorLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_FloorLayer()->base(); |
| case Layer::Layer_GatherLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_GatherLayer()->base(); |
| case Layer::Layer_GreaterLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_GreaterLayer()->base(); |
| case Layer::Layer_InputLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_InputLayer()->base()->base(); |
| case Layer::Layer_InstanceNormalizationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_InstanceNormalizationLayer()->base(); |
| case Layer::Layer_L2NormalizationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_L2NormalizationLayer()->base(); |
| case Layer::Layer_LogSoftmaxLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_LogSoftmaxLayer()->base(); |
| case Layer::Layer_LstmLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_LstmLayer()->base(); |
| case Layer::Layer_MeanLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MeanLayer()->base(); |
| case Layer::Layer_MinimumLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MinimumLayer()->base(); |
| case Layer::Layer_MaximumLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MaximumLayer()->base(); |
| case Layer::Layer_MergeLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MergeLayer()->base(); |
| case Layer::Layer_MergerLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MergerLayer()->base(); |
| case Layer::Layer_MultiplicationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_MultiplicationLayer()->base(); |
| case Layer::Layer_NormalizationLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_NormalizationLayer()->base(); |
| case Layer::Layer_OutputLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_OutputLayer()->base()->base(); |
| case Layer::Layer_PadLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_PadLayer()->base(); |
| case Layer::Layer_PermuteLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_PermuteLayer()->base(); |
| case Layer::Layer_Pooling2dLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_Pooling2dLayer()->base(); |
| case Layer::Layer_PreluLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_PreluLayer()->base(); |
| case Layer::Layer_QuantizeLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_QuantizeLayer()->base(); |
| case Layer::Layer_QuantizedLstmLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_QuantizedLstmLayer()->base(); |
| case Layer::Layer_ReshapeLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ReshapeLayer()->base(); |
| case Layer::Layer_ResizeBilinearLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ResizeBilinearLayer()->base(); |
| case Layer::Layer_ResizeLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_ResizeLayer()->base(); |
| case Layer::Layer_RsqrtLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_RsqrtLayer()->base(); |
| case Layer::Layer_SliceLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SliceLayer()->base(); |
| case Layer::Layer_SoftmaxLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SoftmaxLayer()->base(); |
| case Layer::Layer_SpaceToBatchNdLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SpaceToBatchNdLayer()->base(); |
| case Layer::Layer_SpaceToDepthLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SpaceToDepthLayer()->base(); |
| case Layer::Layer_SplitterLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SplitterLayer()->base(); |
| case Layer::Layer_StackLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_StackLayer()->base(); |
| case Layer::Layer_StandInLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_StandInLayer()->base(); |
| case Layer::Layer_StridedSliceLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_StridedSliceLayer()->base(); |
| case Layer::Layer_SubtractionLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SubtractionLayer()->base(); |
| case Layer::Layer_SwitchLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_SwitchLayer()->base(); |
| case Layer::Layer_TransposeConvolution2dLayer: |
| return graphPtr->layers()->Get(layerIndex)->layer_as_TransposeConvolution2dLayer()->base(); |
| case Layer::Layer_NONE: |
| default: |
| throw ParseException(boost::str( |
| boost::format("Layer type %1% not recognized") % |
| layerType)); |
| } |
| } |
| |
| std::string Deserializer::GetLayerName(const GraphPtr& graph, unsigned int index) |
| { |
| auto layer = GetBaseLayer(graph, index); |
| assert(layer); |
| return layer->layerName()->str(); |
| } |
| |
| int32_t Deserializer::GetBindingLayerInfo(const GraphPtr& graphPtr, unsigned int layerIndex) |
| { |
| auto layerType = graphPtr->layers()->Get(layerIndex)->layer_type(); |
| |
| if (layerType == Layer::Layer_InputLayer) |
| { |
| return graphPtr->layers()->Get(layerIndex)->layer_as_InputLayer()->base()->layerBindingId(); |
| } |
| else if ( layerType == Layer::Layer_OutputLayer ) |
| { |
| return graphPtr->layers()->Get(layerIndex)->layer_as_OutputLayer()->base()->layerBindingId(); |
| } |
| return 0; |
| } |
| |
| armnn::DataLayout ToDataLayout(armnnSerializer::DataLayout dataLayout) |
| { |
| switch (dataLayout) |
| { |
| case armnnSerializer::DataLayout::DataLayout_NHWC: |
| return armnn::DataLayout::NHWC; |
| case armnnSerializer::DataLayout::DataLayout_NCHW: |
| default: |
| return armnn::DataLayout::NCHW; |
| } |
| } |
| |
| armnn::ActivationFunction ToActivationFunction(armnnSerializer::ActivationFunction function) |
| { |
| switch (function) |
| { |
| case armnnSerializer::ActivationFunction_Sigmoid: |
| return armnn::ActivationFunction::Sigmoid; |
| case armnnSerializer::ActivationFunction_TanH: |
| return armnn::ActivationFunction::TanH; |
| case armnnSerializer::ActivationFunction_Linear: |
| return armnn::ActivationFunction::Linear; |
| case armnnSerializer::ActivationFunction_ReLu: |
| return armnn::ActivationFunction::ReLu; |
| case armnnSerializer::ActivationFunction_BoundedReLu: |
| return armnn::ActivationFunction::BoundedReLu; |
| case armnnSerializer::ActivationFunction_LeakyReLu: |
| return armnn::ActivationFunction::LeakyReLu; |
| case armnnSerializer::ActivationFunction_Abs: |
| return armnn::ActivationFunction::Abs; |
| case armnnSerializer::ActivationFunction_Sqrt: |
| return armnn::ActivationFunction::Sqrt; |
| case armnnSerializer::ActivationFunction_Square: |
| return armnn::ActivationFunction::Square; |
| default: |
| return armnn::ActivationFunction::Sigmoid; |
| } |
| } |
| |
| armnn::ArgMinMaxFunction ToArgMinMaxFunction(armnnSerializer::ArgMinMaxFunction function) |
| { |
| switch (function) |
| { |
| case armnnSerializer::ArgMinMaxFunction::ArgMinMaxFunction_Max: |
| return armnn::ArgMinMaxFunction::Max; |
| case armnnSerializer::ArgMinMaxFunction::ArgMinMaxFunction_Min: |
| default: |
| return armnn::ArgMinMaxFunction::Min; |
| } |
| } |
| |
| armnn::ComparisonOperation ToComparisonOperation(armnnSerializer::ComparisonOperation operation) |
| { |
| switch (operation) |
| { |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_Equal: |
| return armnn::ComparisonOperation::Equal; |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_Greater: |
| return armnn::ComparisonOperation::Greater; |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_GreaterOrEqual: |
| return armnn::ComparisonOperation::GreaterOrEqual; |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_Less: |
| return armnn::ComparisonOperation::Less; |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_LessOrEqual: |
| return armnn::ComparisonOperation::LessOrEqual; |
| case armnnSerializer::ComparisonOperation::ComparisonOperation_NotEqual: |
| default: |
| return armnn::ComparisonOperation::NotEqual; |
| } |
| } |
| |
| armnn::ResizeMethod ToResizeMethod(armnnSerializer::ResizeMethod method) |
| { |
| switch (method) |
| { |
| case armnnSerializer::ResizeMethod_NearestNeighbor: |
| return armnn::ResizeMethod::NearestNeighbor; |
| case armnnSerializer::ResizeMethod_Bilinear: |
| return armnn::ResizeMethod::Bilinear; |
| default: |
| return armnn::ResizeMethod::NearestNeighbor; |
| } |
| } |
| |
| armnn::TensorInfo ToTensorInfo(Deserializer::TensorRawPtr tensorPtr) |
| { |
| armnn::DataType type; |
| CHECK_TENSOR_PTR(tensorPtr); |
| |
| switch (tensorPtr->dataType()) |
| { |
| case DataType_QuantisedAsymm8: |
| type = armnn::DataType::QuantisedAsymm8; |
| break; |
| case DataType_QuantisedSymm16: |
| type = armnn::DataType::QuantisedSymm16; |
| break; |
| case DataType_Signed32: |
| type = armnn::DataType::Signed32; |
| break; |
| case DataType_Float32: |
| type = armnn::DataType::Float32; |
| break; |
| case DataType_Float16: |
| type = armnn::DataType::Float16; |
| break; |
| case DataType_Boolean: |
| type = armnn::DataType::Boolean; |
| break; |
| default: |
| { |
| CheckLocation location = CHECK_LOCATION(); |
| throw ParseException( |
| boost::str( |
| boost::format("Unsupported data type %1% = %2%. %3%") % |
| tensorPtr->dataType() % |
| EnumNameDataType(tensorPtr->dataType()) % |
| location.AsString())); |
| } |
| } |
| float quantizationScale = tensorPtr->quantizationScale(); |
| int32_t quantizationOffset = tensorPtr->quantizationOffset(); |
| |
| auto dimensions = tensorPtr->dimensions(); |
| unsigned int size = dimensions->size(); |
| std::vector<unsigned int> outputDims(dimensions->begin(), dimensions->begin() + size); |
| |
| // two statements (on purpose) for easier debugging: |
| armnn::TensorInfo result(size, |
| outputDims.data(), |
| type, |
| quantizationScale, |
| quantizationOffset); |
| return result; |
| } |
| |
| armnn::ConstTensor ToConstTensor(Deserializer::ConstTensorRawPtr constTensorPtr) |
| { |
| CHECK_CONST_TENSOR_PTR(constTensorPtr); |
| armnn::TensorInfo tensorInfo = ToTensorInfo(constTensorPtr->info()); |
| |
| switch (constTensorPtr->data_type()) |
| { |
| case ConstTensorData_ByteData: |
| { |
| auto byteData = constTensorPtr->data_as_ByteData()->data(); |
| CHECK_CONST_TENSOR_SIZE(byteData->size(), tensorInfo.GetNumElements()); |
| return armnn::ConstTensor(tensorInfo, byteData->data()); |
| } |
| case ConstTensorData_ShortData: |
| { |
| auto shortData = constTensorPtr->data_as_ShortData()->data(); |
| CHECK_CONST_TENSOR_SIZE(shortData->size(), tensorInfo.GetNumElements()); |
| return armnn::ConstTensor(tensorInfo, shortData->data()); |
| } |
| case ConstTensorData_IntData: |
| { |
| auto intData = constTensorPtr->data_as_IntData()->data(); |
| CHECK_CONST_TENSOR_SIZE(intData->size(), tensorInfo.GetNumElements()); |
| return armnn::ConstTensor(tensorInfo, intData->data()); |
| } |
| case ConstTensorData_LongData: |
| { |
| auto longData = constTensorPtr->data_as_LongData()->data(); |
| CHECK_CONST_TENSOR_SIZE(longData->size(), tensorInfo.GetNumElements()); |
| return armnn::ConstTensor(tensorInfo, longData->data()); |
| } |
| default: |
| { |
| CheckLocation location = CHECK_LOCATION(); |
| throw ParseException( |
| boost::str(boost::format("Unsupported data type %1% = %2%. %3%") % |
| constTensorPtr->data_type() % |
| EnumNameConstTensorData(constTensorPtr->data_type()) % |
| location.AsString())); |
| } |
| } |
| } |
| |
| Deserializer::TensorRawPtrVector Deserializer::GetInputs(const GraphPtr& graphPtr, |
| unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graphPtr, 0, layerIndex); |
| auto layer = GetBaseLayer(graphPtr, layerIndex); |
| const auto& numInputs = layer->inputSlots()->size(); |
| |
| TensorRawPtrVector result(numInputs); |
| |
| for (unsigned int i=0; i<numInputs; ++i) |
| { |
| auto inputId = CHECKED_NON_NEGATIVE(static_cast<int32_t> |
| (layer->inputSlots()->Get(i)->connection()->sourceLayerIndex())); |
| result[i] = GetBaseLayer(graphPtr, inputId)->outputSlots()->Get(0)->tensorInfo(); |
| } |
| return result; |
| } |
| |
| Deserializer::TensorRawPtrVector Deserializer::GetOutputs(const GraphPtr& graphPtr, |
| unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graphPtr, 0, layerIndex); |
| auto layer = GetBaseLayer(graphPtr, layerIndex); |
| const auto& numOutputs = layer->outputSlots()->size(); |
| |
| TensorRawPtrVector result(numOutputs); |
| |
| for (unsigned int i=0; i<numOutputs; ++i) |
| { |
| result[i] = layer->outputSlots()->Get(i)->tensorInfo(); |
| } |
| return result; |
| } |
| |
| void Deserializer::ParseUnsupportedLayer(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| const auto layerName = GetBaseLayer(graph, layerIndex)->layerName()->c_str(); |
| throw ParseException( |
| boost::str( |
| boost::format("Layer not supported. " |
| "layerIndex: %1% " |
| "layerName: %2% / %3%") % |
| layerIndex % |
| layerName % |
| CHECK_LOCATION().AsString())); |
| } |
| |
| void Deserializer::ResetParser() |
| { |
| m_Network = armnn::INetworkPtr(nullptr, nullptr); |
| m_InputBindings.clear(); |
| m_OutputBindings.clear(); |
| } |
| |
| IDeserializer* IDeserializer::CreateRaw() |
| { |
| return new Deserializer(); |
| } |
| |
| IDeserializerPtr IDeserializer::Create() |
| { |
| return IDeserializerPtr(CreateRaw(), &IDeserializer::Destroy); |
| } |
| |
| void IDeserializer::Destroy(IDeserializer* parser) |
| { |
| delete parser; |
| } |
| |
| INetworkPtr Deserializer::CreateNetworkFromBinary(const std::vector<uint8_t>& binaryContent) |
| { |
| ResetParser(); |
| GraphPtr graph = LoadGraphFromBinary(binaryContent.data(), binaryContent.size()); |
| return CreateNetworkFromGraph(graph); |
| } |
| |
| armnn::INetworkPtr Deserializer::CreateNetworkFromBinary(std::istream& binaryContent) |
| { |
| ResetParser(); |
| std::vector<uint8_t> content((std::istreambuf_iterator<char>(binaryContent)), std::istreambuf_iterator<char>()); |
| GraphPtr graph = LoadGraphFromBinary(content.data(), content.size()); |
| return CreateNetworkFromGraph(graph); |
| } |
| |
| Deserializer::GraphPtr Deserializer::LoadGraphFromBinary(const uint8_t* binaryContent, size_t len) |
| { |
| if (binaryContent == nullptr) |
| { |
| throw InvalidArgumentException(boost::str(boost::format("Invalid (null) binary content %1%") % |
| CHECK_LOCATION().AsString())); |
| } |
| flatbuffers::Verifier verifier(binaryContent, len); |
| if (verifier.VerifyBuffer<SerializedGraph>() == false) |
| { |
| throw ParseException( |
| boost::str(boost::format("Buffer doesn't conform to the expected Armnn " |
| "flatbuffers format. size:%1% %2%") % |
| len % |
| CHECK_LOCATION().AsString())); |
| } |
| return GetSerializedGraph(binaryContent); |
| } |
| |
| INetworkPtr Deserializer::CreateNetworkFromGraph(GraphPtr graph) |
| { |
| m_Network = INetwork::Create(); |
| BOOST_ASSERT(graph != nullptr); |
| unsigned int layerIndex = 0; |
| for (AnyLayer const* layer : *graph->layers()) |
| { |
| if (layer->layer_type() != Layer_InputLayer && |
| layer->layer_type() != Layer_OutputLayer) |
| { |
| // lookup and call the parser function |
| auto& parserFunction = m_ParserFunctions[layer->layer_type()]; |
| (this->*parserFunction)(graph, layerIndex); |
| } |
| ++layerIndex; |
| } |
| |
| SetupInputLayers(graph); |
| SetupOutputLayers(graph); |
| |
| // establish the connections from the layer outputs to the inputs of the subsequent layers |
| for (auto&& graphIt : m_GraphConnections) |
| { |
| Connections& connections = graphIt.second; |
| for (auto&& outputIt : connections.outputSlots) |
| { |
| const unsigned int outputSlotIndex = outputIt.first; |
| IOutputSlot* outputSlot = outputIt.second; |
| if (connections.inputSlots.find(outputSlotIndex) != connections.inputSlots.end()) |
| { |
| for (IInputSlot* inputSlot : connections.inputSlots[outputSlotIndex]) |
| { |
| outputSlot->Connect(*inputSlot); |
| } |
| } |
| } |
| } |
| |
| return std::move(m_Network); |
| } |
| |
| BindingPointInfo Deserializer::GetNetworkInputBindingInfo(unsigned int layerIndex, |
| const std::string& name) const |
| { |
| for (auto inputBinding : m_InputBindings) |
| { |
| if (inputBinding.first == name) |
| { |
| return inputBinding.second; |
| } |
| } |
| throw ParseException( |
| boost::str( |
| boost::format("No input binding found for layer:%1% / %2%") % |
| name % |
| CHECK_LOCATION().AsString())); |
| } |
| |
| BindingPointInfo Deserializer::GetNetworkOutputBindingInfo(unsigned int layerIndex, |
| const std::string& name) const |
| { |
| for (auto outputBinding : m_OutputBindings) |
| { |
| if (outputBinding.first == name) |
| { |
| return outputBinding.second; |
| } |
| } |
| throw ParseException( |
| boost::str( |
| boost::format("No output binding found for layer:%1% / %2%") % |
| name % |
| CHECK_LOCATION().AsString())); |
| } |
| |
| unsigned int Deserializer::GetLayerIndexInVector(GraphPtr graph, unsigned int targetIndex) |
| { |
| for (unsigned int i = 0; i < graph->layers()->size(); i++) |
| { |
| LayerBaseRawPtr layer = GetBaseLayer(graph, i); |
| if (layer->index() == targetIndex) |
| { |
| return i; |
| } |
| } |
| throw ParseException("Layer with given index not found"); |
| } |
| |
| void Deserializer::SetupInputLayers(GraphPtr graph) |
| { |
| CHECK_GRAPH(graph, 0); |
| const unsigned int numInputs = graph->inputIds()->size(); |
| m_InputBindings.clear(); |
| m_InputBindings.reserve(numInputs); |
| |
| for (unsigned int i = 0; i < numInputs; i++) |
| { |
| const unsigned int inputId = graph->inputIds()->Get(i); |
| const unsigned int inputLayerIndex = GetLayerIndexInVector(graph, inputId); |
| LayerBaseRawPtr baseLayer = GetBaseLayer(graph, inputLayerIndex); |
| |
| // GetBindingLayerInfo expect the index to be index in the vector not index property on each layer base |
| LayerBindingId bindingId = GetBindingLayerInfo(graph, inputLayerIndex); |
| BOOST_ASSERT_MSG(baseLayer->layerName()->c_str(), "Input has no name."); |
| |
| IConnectableLayer* inputLayer = |
| m_Network->AddInputLayer(bindingId, baseLayer->layerName()->c_str()); |
| |
| const armnn::TensorInfo& tensorInfo = ToTensorInfo(baseLayer->outputSlots()->Get(0)->tensorInfo()); |
| inputLayer->GetOutputSlot(0).SetTensorInfo(tensorInfo); |
| RegisterOutputSlots(graph, inputLayerIndex, inputLayer); |
| |
| BindingPointInfo bindingInfo = {bindingId, tensorInfo}; |
| m_InputBindings.push_back(std::make_pair(baseLayer->layerName()->c_str(), bindingInfo)); |
| } |
| } |
| |
| void Deserializer::SetupOutputLayers(GraphPtr graph) |
| { |
| CHECK_GRAPH(graph, 0); |
| const unsigned int numOutputs = graph->outputIds()->size(); |
| m_OutputBindings.clear(); |
| m_OutputBindings.reserve(numOutputs); |
| |
| for (unsigned int i = 0; i < numOutputs; i++) |
| { |
| const unsigned int outputId = graph->outputIds()->Get(i); |
| const unsigned int outputLayerIndex = GetLayerIndexInVector(graph, outputId); |
| LayerBaseRawPtr baseLayer = GetBaseLayer(graph, outputLayerIndex); |
| |
| // GetBindingLayerInfo expect the index to be index in the vector not index property on each layer base |
| LayerBindingId bindingId = GetBindingLayerInfo(graph, outputLayerIndex); |
| BOOST_ASSERT_MSG(baseLayer->layerName()->c_str(), "Input has no name."); |
| |
| IConnectableLayer* outputLayer = |
| m_Network->AddOutputLayer(bindingId, baseLayer->layerName()->c_str()); |
| |
| RegisterInputSlots(graph, outputLayerIndex, outputLayer); |
| |
| unsigned int sourceLayerIndex = |
| GetLayerIndexInVector(graph, baseLayer->inputSlots()->Get(0)->connection()->sourceLayerIndex()); |
| LayerBaseRawPtr sourceBaseLayer = GetBaseLayer(graph, sourceLayerIndex); |
| const armnn::TensorInfo& tensorInfo = ToTensorInfo(sourceBaseLayer->outputSlots()->Get(0)->tensorInfo()); |
| |
| BindingPointInfo bindingInfo = {bindingId, tensorInfo}; |
| m_OutputBindings.push_back(std::make_pair(baseLayer->layerName()->c_str(), bindingInfo)); |
| } |
| } |
| |
| void Deserializer::RegisterOutputSlots(GraphPtr graph, |
| uint32_t layerIndex, |
| IConnectableLayer* layer) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| BOOST_ASSERT(layer != nullptr); |
| LayerBaseRawPtr baseLayer = GetBaseLayer(graph, layerIndex); |
| if (baseLayer->outputSlots()->size() != layer->GetNumOutputSlots()) |
| { |
| throw ParseException( |
| boost::str(boost::format("The number of outputslots (%1%) does not match the number expected (%2%)" |
| " for layer index: %3% %4%") % |
| baseLayer->outputSlots()->size() % |
| layer->GetNumOutputSlots() % |
| layerIndex % |
| CHECK_LOCATION().AsString())); |
| } |
| |
| for (unsigned int i = 0; i < layer->GetNumOutputSlots(); ++i) |
| { |
| const unsigned int slotIndex = baseLayer->outputSlots()->Get(i)->index(); |
| armnn::IOutputSlot* outputSlot = &(layer->GetOutputSlot(slotIndex)); |
| // layerIndex is not necessarily the same as baseLayer->index(). The latter is needed here |
| RegisterOutputSlotOfConnection(baseLayer->index(), slotIndex, outputSlot); |
| } |
| } |
| |
| void Deserializer::RegisterInputSlots(GraphPtr graph, |
| uint32_t layerIndex, |
| armnn::IConnectableLayer* layer) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| BOOST_ASSERT(layer != nullptr); |
| LayerBaseRawPtr baseLayer = GetBaseLayer(graph, layerIndex); |
| if (baseLayer->inputSlots()->size() != layer->GetNumInputSlots()) |
| { |
| throw ParseException( |
| boost::str(boost::format("The number of inputslots (%1%) does not match the number expected (%2%)" |
| " for layer index:%3% %4%") % |
| baseLayer->inputSlots()->size() % |
| layer->GetNumInputSlots() % |
| layerIndex % |
| CHECK_LOCATION().AsString())); |
| } |
| |
| for (unsigned int i = 0; i < layer->GetNumInputSlots(); ++i) |
| { |
| auto fbInputSlot = baseLayer->inputSlots()->Get(i); |
| auto fbConnection = fbInputSlot->connection(); |
| armnn::IInputSlot* inputSlot = &(layer->GetInputSlot(fbInputSlot->index())); |
| RegisterInputSlotOfConnection(fbConnection->sourceLayerIndex(), fbConnection->outputSlotIndex(), inputSlot); |
| } |
| } |
| |
| void Deserializer::RegisterInputSlotOfConnection(uint32_t sourceLayerIndex, |
| uint32_t outputSlotIndex, |
| armnn::IInputSlot* inputSlot) |
| { |
| if (m_GraphConnections.find(sourceLayerIndex) == m_GraphConnections.end()) |
| { |
| m_GraphConnections[sourceLayerIndex] = Connections(); |
| } |
| |
| Connections& connections = m_GraphConnections[sourceLayerIndex]; |
| if (connections.inputSlots.find(outputSlotIndex) == connections.inputSlots.end()) |
| { |
| connections.inputSlots[outputSlotIndex] = {inputSlot}; |
| } |
| else |
| { |
| connections.inputSlots[outputSlotIndex].push_back(inputSlot); |
| } |
| } |
| |
| void Deserializer::RegisterOutputSlotOfConnection(uint32_t sourceLayerIndex, |
| uint32_t outputSlotIndex, |
| armnn::IOutputSlot* outputSlot) |
| { |
| if (m_GraphConnections.find(sourceLayerIndex) == m_GraphConnections.end()) |
| { |
| m_GraphConnections[sourceLayerIndex] = Connections(); |
| } |
| |
| Connections& connections = m_GraphConnections[sourceLayerIndex]; |
| if (connections.outputSlots.find(outputSlotIndex) != connections.outputSlots.end()) |
| { |
| throw ParseException("Same output slot index processed twice"); |
| } |
| |
| connections.outputSlots[outputSlotIndex] = outputSlot; |
| } |
| |
| void Deserializer::ParseAbs(armnnDeserializer::Deserializer::GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| IConnectableLayer* layer = m_Network->AddAbsLayer(layerName.c_str()); |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseActivation(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_ActivationLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::ActivationDescriptor descriptor; |
| descriptor.m_Function = ToActivationFunction(serializerDescriptor->activationFunction()); |
| descriptor.m_A = serializerDescriptor->a(); |
| descriptor.m_B = serializerDescriptor->b(); |
| |
| IConnectableLayer* layer = m_Network->AddActivationLayer(descriptor, |
| layerName.c_str()); |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseAdd(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddAdditionLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseArgMinMax(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_ArgMinMaxLayer(); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::ArgMinMaxDescriptor descriptor; |
| descriptor.m_Function = ToArgMinMaxFunction(serializerDescriptor->argMinMaxFunction()); |
| descriptor.m_Axis = serializerDescriptor->axis(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddArgMinMaxLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseBatchToSpaceNd(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_BatchToSpaceNdLayer()->descriptor(); |
| auto flatBufferCrops = flatBufferDescriptor->crops(); |
| auto flatBufferBlockShape = flatBufferDescriptor->blockShape(); |
| |
| if (flatBufferCrops->Length() % 2 != 0) |
| { |
| throw ParseException(boost::str( |
| boost::format("The size of crops must be divisible by 2 %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| std::vector<std::pair<unsigned int, unsigned int>> crops; |
| crops.reserve(flatBufferCrops->Length() / 2); |
| for (unsigned int i = 0; i < flatBufferCrops->Length() - 1; i += 2) |
| { |
| crops.emplace_back(flatBufferCrops->Get(i), flatBufferCrops->Get(i+1)); |
| } |
| |
| armnn::BatchToSpaceNdDescriptor descriptor; |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| descriptor.m_BlockShape = |
| std::vector<unsigned int>(flatBufferBlockShape->begin(), flatBufferBlockShape->end()); |
| descriptor.m_Crops = crops; |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddBatchToSpaceNdLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseBatchNormalization(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_BatchNormalizationLayer(); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::BatchNormalizationDescriptor descriptor; |
| descriptor.m_Eps = serializerDescriptor->eps(); |
| descriptor.m_DataLayout = ToDataLayout(serializerDescriptor->dataLayout()); |
| |
| armnn::ConstTensor mean = ToConstTensor(serializerLayer->mean()); |
| armnn::ConstTensor variance = ToConstTensor(serializerLayer->variance()); |
| armnn::ConstTensor beta = ToConstTensor(serializerLayer->beta()); |
| armnn::ConstTensor gamma = ToConstTensor(serializerLayer->gamma()); |
| |
| IConnectableLayer* layer = m_Network->AddBatchNormalizationLayer(descriptor, |
| mean, |
| variance, |
| beta, |
| gamma, |
| layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseConstant(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| CHECK_LOCATION(); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_ConstantLayer(); |
| auto serializerInput = serializerLayer->input(); |
| |
| armnn::ConstTensor input = ToConstTensor(serializerInput); |
| |
| IConnectableLayer* layer = m_Network->AddConstantLayer(input, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseConvolution2d(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_Convolution2dLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::Convolution2dDescriptor descriptor; |
| descriptor.m_PadLeft = serializerDescriptor->padLeft(); |
| descriptor.m_PadRight = serializerDescriptor->padRight(); |
| descriptor.m_PadTop = serializerDescriptor->padTop(); |
| descriptor.m_PadBottom = serializerDescriptor->padBottom(); |
| descriptor.m_StrideX = serializerDescriptor->strideX(); |
| descriptor.m_StrideY = serializerDescriptor->strideY();; |
| descriptor.m_DilationX = serializerDescriptor->dilationX(); |
| descriptor.m_DilationY = serializerDescriptor->dilationY();; |
| descriptor.m_BiasEnabled = serializerDescriptor->biasEnabled();; |
| descriptor.m_DataLayout = ToDataLayout(serializerDescriptor->dataLayout()); |
| |
| armnn::ConstTensor weights = ToConstTensor(serializerLayer->weights()); |
| armnn::ConstTensor biases; |
| |
| armnn::Optional<armnn::ConstTensor> optionalBiases = armnn::EmptyOptional(); |
| if (descriptor.m_BiasEnabled) |
| { |
| biases = ToConstTensor(serializerLayer->biases()); |
| optionalBiases = armnn::Optional<armnn::ConstTensor>(biases); |
| } |
| IConnectableLayer* layer = m_Network->AddConvolution2dLayer(descriptor, |
| weights, |
| optionalBiases, |
| layerName.c_str()); |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseDepthToSpace(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto fbDescriptor = graph->layers()->Get(layerIndex)->layer_as_DepthToSpaceLayer()->descriptor(); |
| |
| armnn::DepthToSpaceDescriptor descriptor; |
| descriptor.m_BlockSize = fbDescriptor->blockSize(); |
| descriptor.m_DataLayout = ToDataLayout(fbDescriptor->dataLayout()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddDepthToSpaceLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseDepthwiseConvolution2d(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_DepthwiseConvolution2dLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::DepthwiseConvolution2dDescriptor descriptor; |
| descriptor.m_PadLeft = serializerDescriptor->padLeft(); |
| descriptor.m_PadRight = serializerDescriptor->padRight(); |
| descriptor.m_PadTop = serializerDescriptor->padTop(); |
| descriptor.m_PadBottom = serializerDescriptor->padBottom(); |
| descriptor.m_StrideX = serializerDescriptor->strideX(); |
| descriptor.m_StrideY = serializerDescriptor->strideY(); |
| descriptor.m_DilationX = serializerDescriptor->dilationX(); |
| descriptor.m_DilationY = serializerDescriptor->dilationY(); |
| descriptor.m_BiasEnabled = serializerDescriptor->biasEnabled();; |
| descriptor.m_DataLayout = ToDataLayout(serializerDescriptor->dataLayout()); |
| |
| armnn::ConstTensor weights = ToConstTensor(serializerLayer->weights()); |
| armnn::ConstTensor biases; |
| |
| armnn::Optional<armnn::ConstTensor> optionalBiases = armnn::EmptyOptional(); |
| if (descriptor.m_BiasEnabled) |
| { |
| biases = ToConstTensor(serializerLayer->biases()); |
| optionalBiases = armnn::Optional<armnn::ConstTensor>(biases); |
| } |
| IConnectableLayer* layer = m_Network->AddDepthwiseConvolution2dLayer(descriptor, |
| weights, |
| optionalBiases, |
| layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseDetectionPostProcess(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 4); |
| |
| auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_DetectionPostProcessLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto flatBufferDescriptor = flatBufferLayer->descriptor(); |
| |
| armnn::DetectionPostProcessDescriptor descriptor; |
| descriptor.m_MaxDetections = flatBufferDescriptor->maxDetections(); |
| descriptor.m_MaxClassesPerDetection = flatBufferDescriptor->maxClassesPerDetection(); |
| descriptor.m_DetectionsPerClass = flatBufferDescriptor->detectionsPerClass(); |
| descriptor.m_NmsScoreThreshold = flatBufferDescriptor->nmsScoreThreshold(); |
| descriptor.m_NmsIouThreshold = flatBufferDescriptor->nmsIouThreshold(); |
| descriptor.m_NumClasses = flatBufferDescriptor->numClasses(); |
| descriptor.m_UseRegularNms = flatBufferDescriptor->useRegularNms(); |
| descriptor.m_ScaleX = flatBufferDescriptor->scaleX(); |
| descriptor.m_ScaleY = flatBufferDescriptor->scaleY(); |
| descriptor.m_ScaleW = flatBufferDescriptor->scaleW(); |
| descriptor.m_ScaleH = flatBufferDescriptor->scaleH(); |
| |
| armnn::ConstTensor anchors = ToConstTensor(flatBufferLayer->anchors()); |
| |
| IConnectableLayer* layer = m_Network->AddDetectionPostProcessLayer(descriptor, |
| anchors, |
| layerName.c_str()); |
| |
| for (unsigned int i = 0; i < 4; i++) |
| { |
| layer->GetOutputSlot(i).SetTensorInfo(ToTensorInfo(outputs[i])); |
| } |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseDivision(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddDivisionLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseEqual(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| armnn::ComparisonDescriptor descriptor(armnn::ComparisonOperation::Equal); |
| IConnectableLayer* layer = m_Network->AddComparisonLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseGreater(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| armnn::ComparisonDescriptor descriptor(armnn::ComparisonOperation::Greater); |
| IConnectableLayer* layer = m_Network->AddComparisonLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseInstanceNormalization(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto fbLayer = graph->layers()->Get(layerIndex)->layer_as_InstanceNormalizationLayer(); |
| auto fbDescriptor = fbLayer->descriptor(); |
| |
| armnn::InstanceNormalizationDescriptor descriptor; |
| descriptor.m_Gamma = fbDescriptor->gamma(); |
| descriptor.m_Beta = fbDescriptor->beta(); |
| descriptor.m_Eps = fbDescriptor->eps(); |
| descriptor.m_DataLayout = ToDataLayout(fbDescriptor->dataLayout()); |
| |
| const std::string layerName = GetLayerName(graph, layerIndex); |
| const armnn::TensorInfo outputInfo = ToTensorInfo(outputs[0]); |
| |
| IConnectableLayer* layer = m_Network->AddInstanceNormalizationLayer(descriptor, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseL2Normalization(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_L2NormalizationLayer(); |
| auto flatBufferDescriptor = flatBufferLayer->descriptor(); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| armnn::L2NormalizationDescriptor descriptor; |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| descriptor.m_Eps = flatBufferDescriptor->eps(); |
| |
| IConnectableLayer* layer = m_Network->AddL2NormalizationLayer(descriptor, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseLogSoftmax(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| armnn::LogSoftmaxDescriptor descriptor; |
| descriptor.m_Beta = graph->layers()->Get(layerIndex)->layer_as_LogSoftmaxLayer()->descriptor()->beta(); |
| descriptor.m_Axis = graph->layers()->Get(layerIndex)->layer_as_LogSoftmaxLayer()->descriptor()->axis(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| IConnectableLayer* layer = m_Network->AddLogSoftmaxLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseMinimum(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddMinimumLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseMaximum(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddMaximumLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| const armnnSerializer::OriginsDescriptor* GetOriginsDescriptor(const armnnSerializer::SerializedGraph* graph, |
| unsigned int layerIndex) |
| { |
| auto layerType = graph->layers()->Get(layerIndex)->layer_type(); |
| |
| switch (layerType) |
| { |
| case Layer::Layer_ConcatLayer: |
| return graph->layers()->Get(layerIndex)->layer_as_ConcatLayer()->descriptor(); |
| case Layer::Layer_MergerLayer: |
| return graph->layers()->Get(layerIndex)->layer_as_MergerLayer()->descriptor(); |
| default: |
| throw armnn::Exception("unknown layer type, should be concat or merger"); |
| } |
| } |
| |
| void Deserializer::ParseComparison(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| CHECK_LOCATION(); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto fbLayer = graph->layers()->Get(layerIndex)->layer_as_ComparisonLayer(); |
| auto fbDescriptor = fbLayer->descriptor(); |
| |
| armnn::ComparisonDescriptor descriptor; |
| descriptor.m_Operation = ToComparisonOperation(fbDescriptor->operation()); |
| |
| const std::string& layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddComparisonLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseConcat(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| CHECK_LOCATION(); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto originsDescriptor = GetOriginsDescriptor(graph, layerIndex); |
| unsigned int numViews = originsDescriptor->numViews(); |
| unsigned int numDimensions = originsDescriptor->numDimensions(); |
| |
| // can now check the number of inputs == number of views |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), numViews); |
| |
| armnn::OriginsDescriptor descriptor(numViews, numDimensions); |
| auto originsPtr = originsDescriptor->viewOrigins(); |
| for (unsigned int v = 0; v < numViews; ++v) |
| { |
| auto originPtr = originsPtr->Get(v); |
| for (unsigned int d = 0; d < numDimensions; ++d) |
| { |
| uint32_t value = originPtr->data()->Get(d); |
| descriptor.SetViewOriginCoord(v, d, value); |
| } |
| } |
| descriptor.SetConcatAxis(originsDescriptor->concatAxis()); |
| |
| IConnectableLayer* layer = m_Network->AddConcatLayer(descriptor, layerName.c_str()); |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseMultiplication(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddMultiplicationLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseFloor(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| CHECK_LOCATION(); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| armnn::IConnectableLayer* layer; |
| |
| layer = m_Network->AddFloorLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseFullyConnected(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_FullyConnectedLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto flatBufferDescriptor = flatBufferLayer->descriptor(); |
| |
| armnn::FullyConnectedDescriptor fullyConnectedDescriptor; |
| fullyConnectedDescriptor.m_BiasEnabled = flatBufferDescriptor->biasEnabled(); |
| fullyConnectedDescriptor.m_TransposeWeightMatrix = flatBufferDescriptor->transposeWeightsMatrix(); |
| |
| armnn::ConstTensor weightsTensor = ToConstTensor(flatBufferLayer->weights()); |
| |
| armnn::IConnectableLayer* layer; |
| armnn::Optional<armnn::ConstTensor> optionalBiases = armnn::EmptyOptional(); |
| if (flatBufferDescriptor->biasEnabled()) |
| { |
| armnn::ConstTensor biasTensorData = ToConstTensor(flatBufferLayer->biases()); |
| optionalBiases = armnn::Optional<armnn::ConstTensor>(biasTensorData); |
| } |
| layer = m_Network->AddFullyConnectedLayer(fullyConnectedDescriptor, |
| weightsTensor, |
| optionalBiases, |
| layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParsePad(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_PadLayer()->descriptor(); |
| auto flatBufferPadList = flatBufferDescriptor->padList(); |
| float padValue = flatBufferDescriptor->padValue(); |
| |
| if (flatBufferPadList->Length() % 2 != 0) |
| { |
| throw ParseException(boost::str( |
| boost::format("The size of the pad list must be divisible by 2 %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| std::vector<std::pair<unsigned int, unsigned int>> padList; |
| padList.reserve(flatBufferPadList->Length() / 2); |
| for (unsigned int i = 0; i < flatBufferPadList->Length() - 1; i += 2) |
| { |
| padList.emplace_back(flatBufferPadList->Get(i), flatBufferPadList->Get(i+1)); |
| } |
| |
| armnn::PadDescriptor descriptor(padList, padValue); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddPadLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParsePermute(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto dimsMapping = |
| graph->layers()->Get(layerIndex)->layer_as_PermuteLayer()->descriptor()->dimMappings(); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| const armnn::PermuteDescriptor descriptor(armnn::PermutationVector(dimsMapping->data(), dimsMapping->Length())); |
| |
| IConnectableLayer* layer = m_Network->AddPermuteLayer(descriptor, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| armnn::Pooling2dDescriptor Deserializer::GetPoolingDescriptor(Deserializer::PoolingDescriptor pooling2dDesc, |
| unsigned int layerIndex) |
| { |
| armnn::Pooling2dDescriptor desc; |
| |
| switch (pooling2dDesc->poolType()) |
| { |
| case PoolingAlgorithm_Average: |
| { |
| desc.m_PoolType = armnn::PoolingAlgorithm::Average; |
| break; |
| } |
| case PoolingAlgorithm_Max: |
| { |
| desc.m_PoolType = armnn::PoolingAlgorithm::Max; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported pooling algorithm"); |
| } |
| } |
| |
| switch (pooling2dDesc->outputShapeRounding()) |
| { |
| case OutputShapeRounding_Floor: |
| { |
| desc.m_OutputShapeRounding = armnn::OutputShapeRounding::Floor; |
| break; |
| } |
| case OutputShapeRounding_Ceiling: |
| { |
| desc.m_OutputShapeRounding = armnn::OutputShapeRounding::Ceiling; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported output shape rounding"); |
| } |
| } |
| |
| switch (pooling2dDesc->paddingMethod()) |
| { |
| case PaddingMethod_Exclude: |
| { |
| desc.m_PaddingMethod = armnn::PaddingMethod::Exclude; |
| break; |
| } |
| case PaddingMethod_IgnoreValue: |
| { |
| desc.m_PaddingMethod = armnn::PaddingMethod::IgnoreValue; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported padding method"); |
| } |
| } |
| |
| switch (pooling2dDesc->dataLayout()) |
| { |
| case DataLayout_NCHW: |
| { |
| desc.m_DataLayout = armnn::DataLayout::NCHW; |
| break; |
| } |
| case DataLayout_NHWC: |
| { |
| desc.m_DataLayout = armnn::DataLayout::NHWC; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported data layout"); |
| } |
| } |
| |
| desc.m_PadRight = pooling2dDesc->padRight(); |
| desc.m_PadLeft = pooling2dDesc->padLeft(); |
| desc.m_PadBottom = pooling2dDesc->padBottom(); |
| desc.m_PadTop = pooling2dDesc->padTop(); |
| desc.m_StrideX = pooling2dDesc->strideX(); |
| desc.m_StrideY = pooling2dDesc->strideY(); |
| desc.m_PoolWidth = pooling2dDesc->poolWidth(); |
| desc.m_PoolHeight = pooling2dDesc->poolHeight(); |
| |
| return desc; |
| } |
| |
| void Deserializer::ParsePooling2d(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto pooling2dDes = graph->layers()->Get(layerIndex)->layer_as_Pooling2dLayer()->descriptor(); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto pooling2dDescriptor = GetPoolingDescriptor(pooling2dDes, layerIndex); |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddPooling2dLayer(pooling2dDescriptor, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseQuantize(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddQuantizeLayer(layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| armnn::TensorInfo Deserializer::OutputShapeOfReshape(const armnn::TensorInfo& inputTensorInfo, |
| const std::vector<uint32_t>& targetDimsIn) |
| { |
| std::vector<unsigned int> outputDims(targetDimsIn.begin(), targetDimsIn.end()); |
| const auto stretchDim = std::find(targetDimsIn.begin(), targetDimsIn.end(), -1); |
| |
| if (stretchDim != targetDimsIn.end()) |
| { |
| if (std::find(std::next(stretchDim), targetDimsIn.end(), -1) != targetDimsIn.end()) |
| { |
| throw ParseException(boost::str( |
| boost::format("At most one component of shape can be -1 %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| auto targetNumElements = |
| boost::numeric_cast<unsigned int>( |
| std::accumulate(targetDimsIn.begin(), targetDimsIn.end(), -1, std::multiplies<int32_t>())); |
| |
| auto stretchIndex = static_cast<size_t>(std::distance(targetDimsIn.begin(), stretchDim)); |
| outputDims[stretchIndex] = inputTensorInfo.GetNumElements() / targetNumElements; |
| } |
| |
| TensorShape outputShape = TensorShape(static_cast<unsigned int>(outputDims.size()), outputDims.data()); |
| |
| armnn::TensorInfo reshapeInfo = inputTensorInfo; |
| reshapeInfo.SetShape(outputShape); |
| |
| return reshapeInfo; |
| } |
| |
| void Deserializer::ParseReshape(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| armnn::TensorInfo inputTensorInfo = ToTensorInfo(inputs[0]); |
| armnn::TensorInfo actualOutputTensorInfo = ToTensorInfo(outputs[0]); |
| |
| const auto targetDims = graph->layers()->Get(layerIndex)->layer_as_ReshapeLayer()->descriptor()->targetShape(); |
| std::vector<uint32_t> outputDims(targetDims->begin(), targetDims->begin() + targetDims->size()); |
| |
| armnn::TensorInfo reshapeOutputTensorInfo = Deserializer::OutputShapeOfReshape(inputTensorInfo, outputDims); |
| const armnn::TensorShape& reshapeOutputTensorShape = reshapeOutputTensorInfo.GetShape(); |
| |
| const std::vector<uint32_t> expectedDims(outputs[0]->dimensions()->begin(), |
| outputs[0]->dimensions()->begin() + outputs[0]->dimensions()->size()); |
| |
| if (inputs.size() > 1 && !CheckShape(reshapeOutputTensorShape, expectedDims)) |
| { |
| std::stringstream ss; |
| ss << "New shape defined in reshape parameters " |
| << reshapeOutputTensorShape |
| << " does not equal output shape " |
| << actualOutputTensorInfo.GetShape() |
| << ": " |
| << CHECK_LOCATION().AsString(); |
| throw ParseException(ss.str()); |
| } |
| |
| armnn::ReshapeDescriptor reshapeDesc; |
| reshapeDesc.m_TargetShape = reshapeOutputTensorShape; |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddReshapeLayer(reshapeDesc, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(reshapeOutputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseResize(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_ResizeLayer()->descriptor(); |
| |
| armnn::ResizeDescriptor descriptor; |
| descriptor.m_TargetWidth = flatBufferDescriptor->targetWidth(); |
| descriptor.m_TargetHeight = flatBufferDescriptor->targetHeight(); |
| descriptor.m_Method = ToResizeMethod(flatBufferDescriptor->method()); |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddResizeLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseResizeBilinear(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_ResizeBilinearLayer()->descriptor(); |
| |
| armnn::ResizeDescriptor descriptor; |
| descriptor.m_TargetWidth = flatBufferDescriptor->targetWidth(); |
| descriptor.m_TargetHeight = flatBufferDescriptor->targetHeight(); |
| descriptor.m_Method = armnn::ResizeMethod::Bilinear; |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddResizeLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSoftmax(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| armnn::SoftmaxDescriptor descriptor; |
| descriptor.m_Beta = graph->layers()->Get(layerIndex)->layer_as_SoftmaxLayer()->descriptor()->beta(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| IConnectableLayer* layer = m_Network->AddSoftmaxLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSpaceToBatchNd(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_SpaceToBatchNdLayer()->descriptor(); |
| auto flatBufferPadList = flatBufferDescriptor->padList(); |
| auto flatBufferBlockShape = flatBufferDescriptor->blockShape(); |
| |
| if (flatBufferPadList->Length() % 2 != 0) |
| { |
| throw ParseException(boost::str( |
| boost::format("The size of the pad list must be divisible by 2 %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| std::vector<std::pair<unsigned int, unsigned int>> padList; |
| padList.reserve(flatBufferPadList->Length() / 2); |
| for (unsigned int i = 0; i < flatBufferPadList->Length() - 1; i += 2) |
| { |
| padList.emplace_back(flatBufferPadList->Get(i), flatBufferPadList->Get(i+1)); |
| } |
| |
| armnn::SpaceToBatchNdDescriptor descriptor; |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| descriptor.m_BlockShape = |
| std::vector<unsigned int>(flatBufferBlockShape->begin(), flatBufferBlockShape->end()); |
| descriptor.m_PadList = padList; |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSpaceToBatchNdLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSpaceToDepth(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_SpaceToDepthLayer()->descriptor(); |
| |
| armnn::SpaceToDepthDescriptor descriptor; |
| descriptor.m_BlockSize = flatBufferDescriptor->blockSize(); |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSpaceToDepthLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| armnn::NormalizationDescriptor Deserializer::GetNormalizationDescriptor( |
| Deserializer::NormalizationDescriptorPtr normalizationDescriptor, |
| unsigned int layerIndex) |
| { |
| armnn::NormalizationDescriptor desc; |
| |
| switch (normalizationDescriptor->normChannelType()) |
| { |
| case NormalizationAlgorithmChannel_Across: |
| { |
| desc.m_NormChannelType = armnn::NormalizationAlgorithmChannel::Across; |
| break; |
| } |
| case NormalizationAlgorithmChannel_Within: |
| { |
| desc.m_NormChannelType = armnn::NormalizationAlgorithmChannel::Within; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported normalization channel type"); |
| } |
| } |
| |
| switch (normalizationDescriptor->normMethodType()) |
| { |
| case NormalizationAlgorithmMethod_LocalBrightness: |
| { |
| desc.m_NormMethodType = armnn::NormalizationAlgorithmMethod::LocalBrightness; |
| break; |
| } |
| case NormalizationAlgorithmMethod_LocalContrast: |
| { |
| desc.m_NormMethodType = armnn::NormalizationAlgorithmMethod::LocalContrast; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported normalization method type"); |
| } |
| } |
| |
| switch (normalizationDescriptor->dataLayout()) |
| { |
| case DataLayout_NCHW: |
| { |
| desc.m_DataLayout = armnn::DataLayout::NCHW; |
| break; |
| } |
| case DataLayout_NHWC: |
| { |
| desc.m_DataLayout = armnn::DataLayout::NHWC; |
| break; |
| } |
| default: |
| { |
| BOOST_ASSERT_MSG(false, "Unsupported data layout"); |
| } |
| } |
| |
| desc.m_Alpha = normalizationDescriptor->alpha(); |
| desc.m_Beta = normalizationDescriptor->beta(); |
| desc.m_K = normalizationDescriptor->k(); |
| desc.m_NormSize = normalizationDescriptor->normSize(); |
| |
| return desc; |
| } |
| |
| void Deserializer::ParseNormalization(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto normalizationDes = graph->layers()->Get(layerIndex)->layer_as_NormalizationLayer()->descriptor(); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto outputInfo = ToTensorInfo(outputs[0]); |
| |
| auto normalizationDescriptor = GetNormalizationDescriptor(normalizationDes, layerIndex); |
| auto layerName = GetLayerName(graph, layerIndex); |
| |
| IConnectableLayer* layer = m_Network->AddNormalizationLayer(normalizationDescriptor, layerName.c_str()); |
| layer->GetOutputSlot(0).SetTensorInfo(outputInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseRsqrt(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddRsqrtLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSlice(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto fbDescriptor = graph->layers()->Get(layerIndex)->layer_as_SliceLayer()->descriptor(); |
| |
| auto fbBegin = fbDescriptor->begin(); |
| auto fbSize = fbDescriptor->size(); |
| |
| if (fbBegin->Length() != fbSize->Length()) |
| { |
| throw ParseException(boost::str( |
| boost::format("Begin and size descriptors must have the same length %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| armnn::SliceDescriptor descriptor; |
| descriptor.m_Begin.insert(descriptor.m_Begin.end(), fbBegin->begin(), fbBegin->end()); |
| descriptor.m_Size.insert(descriptor.m_Size.end(), fbSize->begin(), fbSize->end()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSliceLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseStridedSlice(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_StridedSliceLayer()->descriptor(); |
| |
| auto flatBufferBegin = flatBufferDescriptor->begin(); |
| auto flatBufferEnd = flatBufferDescriptor->end(); |
| auto flatBufferStride = flatBufferDescriptor->stride(); |
| |
| if (!(flatBufferBegin->Length() == flatBufferEnd->Length() && |
| flatBufferBegin->Length() == flatBufferStride->Length())) |
| { |
| throw ParseException(boost::str( |
| boost::format("The size of the begin, end, and stride must be equal %1%") % CHECK_LOCATION().AsString())); |
| } |
| |
| std::vector<int> begin(flatBufferBegin->begin(), flatBufferBegin->end()); |
| std::vector<int> end(flatBufferEnd->begin(), flatBufferEnd->end()); |
| std::vector<int> stride(flatBufferStride->begin(), flatBufferStride->end()); |
| |
| armnn::StridedSliceDescriptor descriptor(begin, end, stride); |
| descriptor.m_BeginMask = flatBufferDescriptor->beginMask(); |
| descriptor.m_EndMask = flatBufferDescriptor->endMask(); |
| descriptor.m_ShrinkAxisMask = flatBufferDescriptor->shrinkAxisMask(); |
| descriptor.m_EllipsisMask = flatBufferDescriptor->ellipsisMask(); |
| descriptor.m_NewAxisMask = flatBufferDescriptor->newAxisMask(); |
| descriptor.m_DataLayout = ToDataLayout(flatBufferDescriptor->dataLayout()); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddStridedSliceLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSubtraction(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSubtractionLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseGather(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddGatherLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseMean(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_MeanLayer()->descriptor(); |
| auto flatBufferAxis = flatBufferDescriptor->axis(); |
| auto flatBufferKeepDims = flatBufferDescriptor->keepDims(); |
| |
| armnn::MeanDescriptor descriptor; |
| descriptor.m_Axis = std::vector<unsigned int>(flatBufferAxis->begin(), flatBufferAxis->end()); |
| descriptor.m_KeepDims = flatBufferKeepDims; |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddMeanLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSplitter(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| |
| auto flatBufferViewsDescriptor = graph->layers()->Get(layerIndex)->layer_as_SplitterLayer()->descriptor(); |
| auto flatBufferViewSizes = flatBufferViewsDescriptor->viewSizes(); |
| auto flatBufferOriginsDescriptor = flatBufferViewsDescriptor->origins(); |
| auto flatBufferViewOrigins = flatBufferOriginsDescriptor->viewOrigins(); |
| uint32_t numViews = flatBufferOriginsDescriptor->numViews(); |
| uint32_t numDimensions = flatBufferOriginsDescriptor->numDimensions(); |
| |
| // Check numViews and numDimensions corresponds to the ones already serialized ... |
| // numViews == flatBufferViewSizes.size(); |
| // foreach: numDimensions == flatBufferViewSizes[x].size(); |
| |
| armnn::ViewsDescriptor viewsDescriptor(numViews, numDimensions); |
| for(unsigned int vIdx = 0; vIdx < numViews; ++vIdx) |
| { |
| for (unsigned int dIdx = 0; dIdx < numDimensions; ++dIdx) |
| { |
| viewsDescriptor.SetViewSize(vIdx, dIdx, flatBufferViewSizes->Get(vIdx)->data()->Get(dIdx)); |
| viewsDescriptor.SetViewOriginCoord(vIdx, dIdx, flatBufferViewOrigins->Get(vIdx)->data()->Get(dIdx)); |
| } |
| } |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSplitterLayer(viewsDescriptor, layerName.c_str()); |
| |
| // I could have as many outputs as views ... |
| for(unsigned int vIdx = 0; vIdx < numViews; ++vIdx) |
| { |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[vIdx]); |
| layer->GetOutputSlot(vIdx).SetTensorInfo(outputTensorInfo); |
| } |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| armnn::LstmDescriptor Deserializer::GetLstmDescriptor(Deserializer::LstmDescriptorPtr lstmDescriptor) |
| { |
| armnn::LstmDescriptor desc; |
| |
| desc.m_ActivationFunc = lstmDescriptor->activationFunc(); |
| desc.m_ClippingThresCell = lstmDescriptor->clippingThresCell(); |
| desc.m_ClippingThresProj = lstmDescriptor->clippingThresProj(); |
| desc.m_CifgEnabled = lstmDescriptor->cifgEnabled(); |
| desc.m_PeepholeEnabled = lstmDescriptor->peepholeEnabled(); |
| desc.m_ProjectionEnabled = lstmDescriptor->projectionEnabled(); |
| desc.m_LayerNormEnabled = lstmDescriptor->layerNormEnabled(); |
| |
| return desc; |
| } |
| |
| void Deserializer::ParseLstm(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 3); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 4); |
| |
| auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_LstmLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto flatBufferDescriptor = flatBufferLayer->descriptor(); |
| auto flatBufferInputParams = flatBufferLayer->inputParams(); |
| |
| auto lstmDescriptor = GetLstmDescriptor(flatBufferDescriptor); |
| |
| armnn::LstmInputParams lstmInputParams; |
| |
| armnn::ConstTensor inputToForgetWeights = ToConstTensor(flatBufferInputParams->inputToForgetWeights()); |
| armnn::ConstTensor inputToCellWeights = ToConstTensor(flatBufferInputParams->inputToCellWeights()); |
| armnn::ConstTensor inputToOutputWeights = ToConstTensor(flatBufferInputParams->inputToOutputWeights()); |
| armnn::ConstTensor recurrentToForgetWeights = ToConstTensor(flatBufferInputParams->recurrentToForgetWeights()); |
| armnn::ConstTensor recurrentToCellWeights = ToConstTensor(flatBufferInputParams->recurrentToCellWeights()); |
| armnn::ConstTensor recurrentToOutputWeights = ToConstTensor(flatBufferInputParams->recurrentToOutputWeights()); |
| armnn::ConstTensor forgetGateBias = ToConstTensor(flatBufferInputParams->forgetGateBias()); |
| armnn::ConstTensor cellBias = ToConstTensor(flatBufferInputParams->cellBias()); |
| armnn::ConstTensor outputGateBias = ToConstTensor(flatBufferInputParams->outputGateBias()); |
| |
| lstmInputParams.m_InputToForgetWeights = &inputToForgetWeights; |
| lstmInputParams.m_InputToCellWeights = &inputToCellWeights; |
| lstmInputParams.m_InputToOutputWeights = &inputToOutputWeights; |
| lstmInputParams.m_RecurrentToForgetWeights = &recurrentToForgetWeights; |
| lstmInputParams.m_RecurrentToCellWeights = &recurrentToCellWeights; |
| lstmInputParams.m_RecurrentToOutputWeights = &recurrentToOutputWeights; |
| lstmInputParams.m_ForgetGateBias = &forgetGateBias; |
| lstmInputParams.m_CellBias = &cellBias; |
| lstmInputParams.m_OutputGateBias = &outputGateBias; |
| |
| armnn::ConstTensor inputToInputWeights; |
| armnn::ConstTensor recurrentToInputWeights; |
| armnn::ConstTensor cellToInputWeights; |
| armnn::ConstTensor inputGateBias; |
| if (!lstmDescriptor.m_CifgEnabled) |
| { |
| inputToInputWeights = ToConstTensor(flatBufferInputParams->inputToInputWeights()); |
| recurrentToInputWeights = ToConstTensor(flatBufferInputParams->recurrentToInputWeights()); |
| cellToInputWeights = ToConstTensor(flatBufferInputParams->cellToInputWeights()); |
| inputGateBias = ToConstTensor(flatBufferInputParams->inputGateBias()); |
| |
| lstmInputParams.m_InputToInputWeights = &inputToInputWeights; |
| lstmInputParams.m_RecurrentToInputWeights = &recurrentToInputWeights; |
| lstmInputParams.m_CellToInputWeights = &cellToInputWeights; |
| lstmInputParams.m_InputGateBias = &inputGateBias; |
| } |
| |
| armnn::ConstTensor projectionWeights; |
| armnn::ConstTensor projectionBias; |
| if (lstmDescriptor.m_ProjectionEnabled) |
| { |
| projectionWeights = ToConstTensor(flatBufferInputParams->projectionWeights()); |
| projectionBias = ToConstTensor(flatBufferInputParams->projectionBias()); |
| |
| lstmInputParams.m_ProjectionWeights = &projectionWeights; |
| lstmInputParams.m_ProjectionBias = &projectionBias; |
| } |
| |
| armnn::ConstTensor cellToForgetWeights; |
| armnn::ConstTensor cellToOutputWeights; |
| if (lstmDescriptor.m_PeepholeEnabled) |
| { |
| cellToForgetWeights = ToConstTensor(flatBufferInputParams->cellToForgetWeights()); |
| cellToOutputWeights = ToConstTensor(flatBufferInputParams->cellToOutputWeights()); |
| |
| lstmInputParams.m_CellToForgetWeights = &cellToForgetWeights; |
| lstmInputParams.m_CellToOutputWeights = &cellToOutputWeights; |
| } |
| |
| armnn::ConstTensor inputLayerNormWeights; |
| armnn::ConstTensor forgetLayerNormWeights; |
| armnn::ConstTensor cellLayerNormWeights; |
| armnn::ConstTensor outputLayerNormWeights; |
| if (lstmDescriptor.m_LayerNormEnabled) |
| { |
| if (!lstmDescriptor.m_CifgEnabled) |
| { |
| inputLayerNormWeights = ToConstTensor(flatBufferInputParams->inputLayerNormWeights()); |
| lstmInputParams.m_InputLayerNormWeights = &inputLayerNormWeights; |
| } |
| forgetLayerNormWeights = ToConstTensor(flatBufferInputParams->forgetLayerNormWeights()); |
| cellLayerNormWeights = ToConstTensor(flatBufferInputParams->cellLayerNormWeights()); |
| outputLayerNormWeights = ToConstTensor(flatBufferInputParams->outputLayerNormWeights()); |
| |
| lstmInputParams.m_ForgetLayerNormWeights = &forgetLayerNormWeights; |
| lstmInputParams.m_CellLayerNormWeights = &cellLayerNormWeights; |
| lstmInputParams.m_OutputLayerNormWeights = &outputLayerNormWeights; |
| } |
| |
| IConnectableLayer* layer = m_Network->AddLstmLayer(lstmDescriptor, lstmInputParams, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo1 = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo1); |
| |
| armnn::TensorInfo outputTensorInfo2 = ToTensorInfo(outputs[1]); |
| layer->GetOutputSlot(1).SetTensorInfo(outputTensorInfo2); |
| |
| armnn::TensorInfo outputTensorInfo3 = ToTensorInfo(outputs[2]); |
| layer->GetOutputSlot(2).SetTensorInfo(outputTensorInfo3); |
| |
| armnn::TensorInfo outputTensorInfo4 = ToTensorInfo(outputs[3]); |
| layer->GetOutputSlot(3).SetTensorInfo(outputTensorInfo4); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseQuantizedLstm(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 3); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 2); |
| |
| auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_QuantizedLstmLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto flatBufferInputParams = flatBufferLayer->inputParams(); |
| |
| armnn::QuantizedLstmInputParams lstmInputParams; |
| |
| armnn::ConstTensor inputToInputWeights = ToConstTensor(flatBufferInputParams->inputToInputWeights()); |
| armnn::ConstTensor inputToForgetWeights = ToConstTensor(flatBufferInputParams->inputToForgetWeights()); |
| armnn::ConstTensor inputToCellWeights = ToConstTensor(flatBufferInputParams->inputToCellWeights()); |
| armnn::ConstTensor inputToOutputWeights = ToConstTensor(flatBufferInputParams->inputToOutputWeights()); |
| armnn::ConstTensor recurrentToInputWeights = ToConstTensor(flatBufferInputParams->recurrentToInputWeights()); |
| armnn::ConstTensor recurrentToForgetWeights = ToConstTensor(flatBufferInputParams->recurrentToForgetWeights()); |
| armnn::ConstTensor recurrentToCellWeights = ToConstTensor(flatBufferInputParams->recurrentToCellWeights()); |
| armnn::ConstTensor recurrentToOutputWeights = ToConstTensor(flatBufferInputParams->recurrentToOutputWeights()); |
| armnn::ConstTensor inputGateBias = ToConstTensor(flatBufferInputParams->inputGateBias()); |
| armnn::ConstTensor forgetGateBias = ToConstTensor(flatBufferInputParams->forgetGateBias()); |
| armnn::ConstTensor cellBias = ToConstTensor(flatBufferInputParams->cellBias()); |
| armnn::ConstTensor outputGateBias = ToConstTensor(flatBufferInputParams->outputGateBias()); |
| |
| lstmInputParams.m_InputToInputWeights = &inputToInputWeights; |
| lstmInputParams.m_InputToForgetWeights = &inputToForgetWeights; |
| lstmInputParams.m_InputToCellWeights = &inputToCellWeights; |
| lstmInputParams.m_InputToOutputWeights = &inputToOutputWeights; |
| lstmInputParams.m_RecurrentToInputWeights = &recurrentToInputWeights; |
| lstmInputParams.m_RecurrentToForgetWeights = &recurrentToForgetWeights; |
| lstmInputParams.m_RecurrentToCellWeights = &recurrentToCellWeights; |
| lstmInputParams.m_RecurrentToOutputWeights = &recurrentToOutputWeights; |
| lstmInputParams.m_InputGateBias = &inputGateBias; |
| lstmInputParams.m_ForgetGateBias = &forgetGateBias; |
| lstmInputParams.m_CellBias = &cellBias; |
| lstmInputParams.m_OutputGateBias = &outputGateBias; |
| |
| IConnectableLayer* layer = m_Network->AddQuantizedLstmLayer(lstmInputParams, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo1 = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo1); |
| |
| armnn::TensorInfo outputTensorInfo2 = ToTensorInfo(outputs[1]); |
| layer->GetOutputSlot(1).SetTensorInfo(outputTensorInfo2); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseDequantize(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| const std::string layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddDequantizeLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseMerge(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| Deserializer::TensorRawPtrVector inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| Deserializer::TensorRawPtrVector outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| const std::string layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddMergeLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseSwitch(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 2); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddSwitchLayer(layerName.c_str()); |
| |
| armnn::TensorInfo output0TensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(output0TensorInfo); |
| |
| armnn::TensorInfo output1TensorInfo = ToTensorInfo(outputs[1]); |
| layer->GetOutputSlot(1).SetTensorInfo(output1TensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParsePrelu(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_LOCATION(); |
| CHECK_VALID_SIZE(inputs.size(), 2); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddPreluLayer(layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseTransposeConvolution2d(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| CHECK_VALID_SIZE(inputs.size(), 1); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_TransposeConvolution2dLayer(); |
| auto layerName = GetLayerName(graph, layerIndex); |
| auto serializerDescriptor = serializerLayer->descriptor(); |
| |
| armnn::TransposeConvolution2dDescriptor descriptor; |
| descriptor.m_PadLeft = serializerDescriptor->padLeft(); |
| descriptor.m_PadRight = serializerDescriptor->padRight(); |
| descriptor.m_PadTop = serializerDescriptor->padTop(); |
| descriptor.m_PadBottom = serializerDescriptor->padBottom(); |
| descriptor.m_StrideX = serializerDescriptor->strideX(); |
| descriptor.m_StrideY = serializerDescriptor->strideY();; |
| descriptor.m_BiasEnabled = serializerDescriptor->biasEnabled();; |
| descriptor.m_DataLayout = ToDataLayout(serializerDescriptor->dataLayout()); |
| |
| // weights & biases |
| armnn::ConstTensor weights = ToConstTensor(serializerLayer->weights()); |
| armnn::Optional<armnn::ConstTensor> optionalBiases; |
| if (descriptor.m_BiasEnabled) |
| { |
| armnn::ConstTensor biases = ToConstTensor(serializerLayer->biases()); |
| optionalBiases = armnn::MakeOptional<armnn::ConstTensor>(biases); |
| } |
| |
| IConnectableLayer* layer = m_Network->AddTransposeConvolution2dLayer(descriptor, |
| weights, |
| optionalBiases, |
| layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseStack(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| auto inputs = GetInputs(graph, layerIndex); |
| |
| auto outputs = GetOutputs(graph, layerIndex); |
| CHECK_VALID_SIZE(outputs.size(), 1); |
| |
| auto flatBufferDescriptor = graph->layers()->Get(layerIndex)->layer_as_StackLayer()->descriptor(); |
| unsigned int axis = flatBufferDescriptor->axis(); |
| unsigned int numInputs = flatBufferDescriptor->numInputs(); |
| CHECK_VALID_SIZE(inputs.size(), numInputs); |
| |
| auto flatBufferInputShape = flatBufferDescriptor->inputShape(); |
| std::vector<uint32_t> vectorInputShape(flatBufferInputShape->begin(), |
| flatBufferInputShape->begin() + flatBufferInputShape->size()); |
| |
| TensorShape inputShape(static_cast<unsigned int>(vectorInputShape.size()), vectorInputShape.data()); |
| armnn::StackDescriptor descriptor(axis, numInputs, inputShape); |
| |
| for (unsigned int i=0; i<inputs.size(); ++i) |
| { |
| armnn::TensorShape inputShape = ToTensorInfo(inputs[i]).GetShape(); |
| if (descriptor.m_InputShape != inputShape) |
| { |
| std::stringstream ss; |
| ss << "Shape of input " |
| << i |
| << " " |
| << inputShape |
| << " does not equal defined input shape " |
| << descriptor.m_InputShape |
| << ": " |
| << CHECK_LOCATION().AsString(); |
| throw ParseException(ss.str()); |
| } |
| } |
| |
| auto layerName = GetLayerName(graph, layerIndex); |
| IConnectableLayer* layer = m_Network->AddStackLayer(descriptor, layerName.c_str()); |
| |
| armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); |
| layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| void Deserializer::ParseStandIn(GraphPtr graph, unsigned int layerIndex) |
| { |
| CHECK_LAYERS(graph, 0, layerIndex); |
| |
| auto inputs = GetInputs(graph, layerIndex); |
| auto outputs = GetOutputs(graph, layerIndex); |
| |
| auto fbLayer = graph->layers()->Get(layerIndex)->layer_as_StandInLayer(); |
| auto fbDescriptor = fbLayer->descriptor(); |
| |
| armnn::StandInDescriptor descriptor; |
| descriptor.m_NumInputs = fbDescriptor->numInputs(); |
| descriptor.m_NumOutputs = fbDescriptor->numOutputs(); |
| |
| CHECK_VALID_SIZE(inputs.size(), descriptor.m_NumInputs); |
| CHECK_VALID_SIZE(outputs.size(), descriptor.m_NumOutputs); |
| |
| const std::string layerName = GetLayerName(graph, layerIndex); |
| armnn::IConnectableLayer* layer = m_Network->AddStandInLayer(descriptor, layerName.c_str()); |
| |
| for (unsigned int i = 0u; i < descriptor.m_NumOutputs; ++i) |
| { |
| armnn::TensorInfo outputInfo = ToTensorInfo(outputs[i]); |
| layer->GetOutputSlot(i).SetTensorInfo(outputInfo); |
| } |
| |
| RegisterInputSlots(graph, layerIndex, layer); |
| RegisterOutputSlots(graph, layerIndex, layer); |
| } |
| |
| } // namespace armnnDeserializer |