blob: 53136b521e0d4619c6f1c746a09492f9313c910d [file] [log] [blame]
//
// Copyright © 2021-2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "SharedFunctions.hpp"
#include <ClassicDelegateUtils.hpp>
#include <tensorflow/lite/builtin_ops.h>
#include <tensorflow/lite/c/builtin_op_data.h>
#include <tensorflow/lite/c/common.h>
#include <tensorflow/lite/minimal_logging.h>
namespace armnnDelegate
{
TfLiteStatus ValidateFloorOperator(DelegateData& delegateData,
TfLiteContext* tfLiteContext,
const armnn::TensorInfo& inputTensorInfo,
const armnn::TensorInfo& outputTensorInfo)
{
bool isSupported = false;
auto validateFunc = [&](const armnn::TensorInfo& outInfo, bool& isSupported)
{
FORWARD_LAYER_SUPPORT_FUNC("FLOOR",
tfLiteContext,
IsFloorSupported,
delegateData.m_Backends,
isSupported,
armnn::BackendId(),
inputTensorInfo,
outInfo);
};
validateFunc(outputTensorInfo, isSupported);
return isSupported ? kTfLiteOk : kTfLiteError;
}
TfLiteStatus ValidateFusedActivationOperator(DelegateData& delegateData,
TfLiteContext* tfLiteContext,
const armnn::TensorInfo& inputInfo,
const armnn::TensorInfo& outputInfo,
TfLiteFusedActivation activationType)
{
armnn::ActivationDescriptor activationDesc;
switch (activationType)
{
case kTfLiteActNone:
{
// No Activation
return kTfLiteOk;
}
case kTfLiteActRelu:
{
activationDesc.m_Function = armnn::ActivationFunction::ReLu;
break;
}
// The name of kTfLiteActRelu1 changed after TF Lite v2.3
#if defined(ARMNN_POST_TFLITE_2_3)
case kTfLiteActReluN1To1:
#else
case kTfLiteActRelu1:
#endif
{
activationDesc.m_Function = armnn::ActivationFunction::BoundedReLu;
activationDesc.m_A = 1.0f;
activationDesc.m_B = -1.0f;
break;
}
case kTfLiteActRelu6:
{
activationDesc.m_Function = armnn::ActivationFunction::BoundedReLu;
activationDesc.m_A = 6.0f;
activationDesc.m_B = 0.0f;
break;
}
case kTfLiteActSigmoid:
{
activationDesc.m_Function = armnn::ActivationFunction::Sigmoid;
break;
}
case kTfLiteActTanh:
{
activationDesc.m_Function = armnn::ActivationFunction::TanH;
activationDesc.m_A = 1.0f;
activationDesc.m_B = 1.0f;
break;
}
default:
return kTfLiteError;
}
bool isSupported = false;
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
FORWARD_LAYER_SUPPORT_FUNC("ACTIVATION",
tfLiteContext,
IsActivationSupported,
delegateData.m_Backends,
isSupported,
armnn::BackendId(),
inputInfo,
outputInfo,
activationDesc);
};
validateFunc(outputInfo, isSupported);
return isSupported ? kTfLiteOk : kTfLiteError;
}
TfLiteNode* GetNodeConnectedToInput(TfLiteContext* tfLiteContext,
int32_t& connectedIndex,
int32_t inputIdx)
{
TfLiteIntArray* executionPlan = nullptr;
if (tfLiteContext->GetExecutionPlan(tfLiteContext, &executionPlan) != kTfLiteOk)
{
TF_LITE_KERNEL_LOG(tfLiteContext, "TfLiteArmnnDelegate: Unable to get graph execution plan.");
return nullptr;
}
for (int i = 0; i < executionPlan->size; ++i)
{
connectedIndex = executionPlan->data[i];
// If TfLite nodes can be delegated to ArmNN
TfLiteNode* connectedNode = nullptr;
TfLiteRegistration* tfLiteRegistration = nullptr;
if (tfLiteContext->GetNodeAndRegistration(
tfLiteContext, connectedIndex, &connectedNode, &tfLiteRegistration) != kTfLiteOk)
{
TF_LITE_KERNEL_LOG(tfLiteContext,
"TfLiteArmnnDelegate: Unable to get node and registration for node %d.",
connectedIndex);
continue;
}
for (int j= 0; j < connectedNode->outputs->size; ++j)
{
if (connectedNode->outputs->data[j] == inputIdx)
{
return connectedNode;
}
}
}
// No node found so set connectedIndex to -1
connectedIndex = -1;
return nullptr;
}
bool WillInputBeOptimizedToConst(TfLiteContext* tfLiteContext, int32_t inputIdx)
{
int32_t connectedIndex;
TfLiteNode* connectedNode = GetNodeConnectedToInput(tfLiteContext, connectedIndex, inputIdx);
if (connectedNode)
{
TfLiteRegistration* tfLiteRegistration = nullptr;
if (tfLiteContext->GetNodeAndRegistration(tfLiteContext, connectedIndex, &connectedNode, &tfLiteRegistration)
== kTfLiteOk)
{
switch (tfLiteRegistration->builtin_code)
{
case kTfLiteBuiltinDequantize:
{
if (connectedNode->inputs->size >= 1)
{
const TfLiteTensor* tfLiteTensors = tfLiteContext->tensors;
const TfLiteTensor& tfLiteInputTensor = tfLiteTensors[connectedNode->inputs->data[0]];
// If the input to the Dequantize is a Constant then both that Constant layer and the Dequantize
// layer will be replaced by a single Constant layer containing the dequantized values.
if (tflite::IsConstantTensor(&tfLiteInputTensor))
{
return true;
}
}
break;
}
default:
{
}
}
}
}
return false;
}
} // namespace armnnDelegate