| // |
| // Copyright © 2020 Arm Ltd and Contributors. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| |
| #pragma once |
| |
| #include <armnn/backends/OptimizationViews.hpp> |
| |
| namespace armnn |
| { |
| |
| namespace |
| { |
| |
| // |
| // this helper only works if all layers where the inputs connect to are not selected |
| // |
| SubgraphView::InputSlots CreateInputsFrom(const std::vector<Layer*>& layers) |
| { |
| SubgraphView::InputSlots result; |
| for (auto&& layer : layers) |
| { |
| for (auto&& it = layer->BeginInputSlots(); it != layer->EndInputSlots(); ++it) |
| { |
| result.push_back(&(*it)); |
| } |
| } |
| return result; |
| } |
| |
| // |
| // this helper only works if all layers where the outputs connect to are not selected |
| // |
| SubgraphView::OutputSlots CreateOutputsFrom(const std::vector<Layer*>& layers) |
| { |
| SubgraphView::OutputSlots result; |
| for (auto&& layer : layers) |
| { |
| for (auto&& it = layer->BeginOutputSlots(); it != layer->EndOutputSlots(); ++it) |
| { |
| result.push_back(&(*it)); |
| } |
| } |
| return result; |
| } |
| |
| bool checkDataTypeInputandOutput(const Layer& layer) |
| { |
| auto inputInfo = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); |
| auto outputInfo = layer.GetOutputSlot(0).GetTensorInfo(); |
| bool sameDataType = (inputInfo.GetDataType() == outputInfo.GetDataType()); |
| |
| // Check is same quantization info (same scale and offset) |
| if (sameDataType) |
| { |
| if (IsQuantizedType(inputInfo.GetDataType())) |
| { |
| bool sameScale = (inputInfo.GetQuantizationScale() == outputInfo.GetQuantizationScale()); |
| bool sameOffset = (inputInfo.GetQuantizationOffset() == outputInfo.GetQuantizationOffset()); |
| |
| return (sameScale && sameOffset); |
| } |
| else |
| { |
| return true; |
| } |
| } |
| else |
| { |
| return false; |
| } |
| } |
| |
| } // namespace |
| |
| inline void ReportUntouchedLayers(OptimizationViews& optimizationViews, std::map<LayerGuid, Layer*> untouched) |
| { |
| std::vector<Layer*> untouchedVector; |
| for (const auto& pair : untouched) |
| { |
| Layer* layer = pair.second; |
| SubgraphView subgraphView(CreateInputsFrom({layer}), |
| CreateOutputsFrom({layer}), |
| {layer}); |
| optimizationViews.AddUntouchedSubgraph(std::move(subgraphView)); |
| } |
| } |
| |
| template<typename LayerType> |
| LayerType* FuseLayerWithoutParameters(OptimizationViews& optimizationViews, |
| LayerType* baseLayer, |
| ActivationLayer* activationLayer, |
| ActivationDescriptor& activationDesc, |
| std::string name) |
| { |
| LayerType* replacementLayer = optimizationViews.GetGraph().AddLayer<LayerType>(name.c_str()); |
| |
| replacementLayer->SetAdditionalInfoForObject(std::make_shared<ActivationDescriptor>(activationDesc)); |
| |
| SubgraphView substitutionSubgraph(CreateInputsFrom({baseLayer}), |
| CreateOutputsFrom({activationLayer}), |
| {baseLayer, activationLayer}); |
| SubgraphView replacementSubgraph(replacementLayer); |
| |
| optimizationViews.AddSubstitution({substitutionSubgraph, replacementSubgraph}); |
| return replacementLayer; |
| } |
| |
| template<typename LayerType> |
| LayerType* FuseLayerWithParameters(OptimizationViews& optimizationViews, |
| LayerType* baseLayer, |
| ActivationLayer* activationLayer, |
| ActivationDescriptor& activationDesc, |
| std::string name) |
| { |
| LayerType* replacementLayer = optimizationViews.GetGraph().AddLayer<LayerType>(baseLayer->GetParameters(), |
| name.c_str()); |
| |
| replacementLayer->SetAdditionalInfoForObject(std::make_shared<ActivationDescriptor>(activationDesc)); |
| |
| SubgraphView substitutionSubgraph(CreateInputsFrom({baseLayer}), |
| CreateOutputsFrom({activationLayer}), |
| {baseLayer, activationLayer}); |
| SubgraphView replacementSubgraph(replacementLayer); |
| |
| optimizationViews.AddSubstitution({substitutionSubgraph, replacementSubgraph}); |
| return replacementLayer; |
| } |
| |
| template<typename LayerType> |
| LayerType* FuseLayerWithWeightsAndBiases(OptimizationViews& optimizationViews, |
| LayerType* baseLayer, |
| ActivationLayer* activationLayer, |
| ActivationDescriptor& activationDesc, |
| std::string name) |
| { |
| LayerType* replacementLayer = FuseLayerWithParameters(optimizationViews, |
| baseLayer, |
| activationLayer, |
| activationDesc, |
| name); |
| |
| replacementLayer->m_Weight = std::move(baseLayer->m_Weight); |
| replacementLayer->m_Bias = std::move(baseLayer->m_Bias); |
| |
| return replacementLayer; |
| } |
| |
| } // namespace armnn |