IVGCVSW-1951 Update NeonWorkloadUtils

Change-Id: I147dbf6811f84ec4588264d636a36efc8ec56f72
diff --git a/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp b/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
index 8cebb4f..51fd7af 100644
--- a/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
+++ b/src/backends/neon/workloads/NeonFullyConnectedWorkload.cpp
@@ -69,22 +69,22 @@
     // Allocate
     if (m_Data.m_Weight->GetTensorInfo().GetDataType() == DataType::QuantisedAsymm8)
     {
-        InitialiseArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight->GetConstTensor<uint8_t>());
+        InitializeArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight);
     }
     else
     {
-        InitializeArmComputeTensorDataForFloatTypes(*m_WeightsTensor, m_Data.m_Weight);
+        InitializeArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight);
     }
 
     if (m_BiasesTensor)
     {
         if (m_Data.m_Bias->GetTensorInfo().GetDataType() == DataType::Signed32)
         {
-            InitialiseArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias->GetConstTensor<int32_t>());
+            InitializeArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias);
         }
         else
         {
-            InitializeArmComputeTensorDataForFloatTypes(*m_BiasesTensor, m_Data.m_Bias);
+            InitializeArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias);
         }
     }
 
@@ -107,4 +107,3 @@
 }
 
 } //namespace armnn
-