IVGCVSW-1951 Update NeonWorkloadUtils
Change-Id: I147dbf6811f84ec4588264d636a36efc8ec56f72
diff --git a/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp b/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
index 8cebb4f..51fd7af 100644
--- a/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
+++ b/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
@@ -69,22 +69,22 @@
// Allocate
if (m_Data.m_Weight->GetTensorInfo().GetDataType() == DataType::QuantisedAsymm8)
{
- InitialiseArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight->GetConstTensor<uint8_t>());
+ InitializeArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight);
}
else
{
- InitializeArmComputeTensorDataForFloatTypes(*m_WeightsTensor, m_Data.m_Weight);
+ InitializeArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight);
}
if (m_BiasesTensor)
{
if (m_Data.m_Bias->GetTensorInfo().GetDataType() == DataType::Signed32)
{
- InitialiseArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias->GetConstTensor<int32_t>());
+ InitializeArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias);
}
else
{
- InitializeArmComputeTensorDataForFloatTypes(*m_BiasesTensor, m_Data.m_Bias);
+ InitializeArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias);
}
}
@@ -107,4 +107,3 @@
}
} //namespace armnn
-