blob: a0fca4633077668bf840a6b2c6e634665fc975b0 [file] [log] [blame]
//
// Copyright © 2020 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#pragma once
#include <armnn/backends/OptimizationViews.hpp>
namespace armnn
{
namespace
{
//
// this helper only works if all layers where the inputs connect to are not selected
//
SubgraphView::InputSlots CreateInputsFrom(const std::vector<Layer*>& layers)
{
SubgraphView::InputSlots result;
for (auto&& layer : layers)
{
for (auto&& it = layer->BeginInputSlots(); it != layer->EndInputSlots(); ++it)
{
result.push_back(&(*it));
}
}
return result;
}
//
// this helper only works if all layers where the outputs connect to are not selected
//
SubgraphView::OutputSlots CreateOutputsFrom(const std::vector<Layer*>& layers)
{
SubgraphView::OutputSlots result;
for (auto&& layer : layers)
{
for (auto&& it = layer->BeginOutputSlots(); it != layer->EndOutputSlots(); ++it)
{
result.push_back(&(*it));
}
}
return result;
}
bool checkDataTypeInputandOutput(const Layer& layer)
{
auto inputInfo = layer.GetInputSlot(0).GetConnection()->GetTensorInfo();
auto outputInfo = layer.GetOutputSlot(0).GetTensorInfo();
bool sameDataType = (inputInfo.GetDataType() == outputInfo.GetDataType());
// Check is same quantization info (same scale and offset)
if (sameDataType)
{
if (IsQuantizedType(inputInfo.GetDataType()))
{
bool sameScale = (inputInfo.GetQuantizationScale() == outputInfo.GetQuantizationScale());
bool sameOffset = (inputInfo.GetQuantizationOffset() == outputInfo.GetQuantizationOffset());
return (sameScale && sameOffset);
}
else
{
return true;
}
}
else
{
return false;
}
}
} // namespace
inline void ReportUntouchedLayers(OptimizationViews& optimizationViews, std::map<LayerGuid, Layer*> untouched)
{
std::vector<Layer*> untouchedVector;
for (const auto& pair : untouched)
{
Layer* layer = pair.second;
SubgraphView subgraphView(CreateInputsFrom({layer}),
CreateOutputsFrom({layer}),
{layer});
optimizationViews.AddUntouchedSubgraph(std::move(subgraphView));
}
}
template<typename LayerType>
LayerType* FuseLayerWithoutParameters(OptimizationViews& optimizationViews,
LayerType* baseLayer,
ActivationLayer* activationLayer,
ActivationDescriptor& activationDesc,
std::string name)
{
LayerType* replacementLayer = optimizationViews.GetGraph().AddLayer<LayerType>(name.c_str());
replacementLayer->SetAdditionalInfoForObject(std::make_shared<ActivationDescriptor>(activationDesc));
SubgraphView substitutionSubgraph(CreateInputsFrom({baseLayer}),
CreateOutputsFrom({activationLayer}),
{baseLayer, activationLayer});
SubgraphView replacementSubgraph(replacementLayer);
optimizationViews.AddSubstitution({substitutionSubgraph, replacementSubgraph});
return replacementLayer;
}
template<typename LayerType>
LayerType* FuseLayerWithParameters(OptimizationViews& optimizationViews,
LayerType* baseLayer,
ActivationLayer* activationLayer,
ActivationDescriptor& activationDesc,
std::string name)
{
LayerType* replacementLayer = optimizationViews.GetGraph().AddLayer<LayerType>(baseLayer->GetParameters(),
name.c_str());
replacementLayer->SetAdditionalInfoForObject(std::make_shared<ActivationDescriptor>(activationDesc));
SubgraphView substitutionSubgraph(CreateInputsFrom({baseLayer}),
CreateOutputsFrom({activationLayer}),
{baseLayer, activationLayer});
SubgraphView replacementSubgraph(replacementLayer);
optimizationViews.AddSubstitution({substitutionSubgraph, replacementSubgraph});
return replacementLayer;
}
template<typename LayerType>
LayerType* FuseLayerWithWeightsAndBiases(OptimizationViews& optimizationViews,
LayerType* baseLayer,
ActivationLayer* activationLayer,
ActivationDescriptor& activationDesc,
std::string name)
{
LayerType* replacementLayer = FuseLayerWithParameters(optimizationViews,
baseLayer,
activationLayer,
activationDesc,
name);
replacementLayer->m_Weight = std::move(baseLayer->m_Weight);
replacementLayer->m_Bias = std::move(baseLayer->m_Bias);
return replacementLayer;
}
} // namespace armnn