IVGCVSW-4386 Add ArmNN reference support for QAsymmS8

 * Added Quantization Scheme for QAsymmS8
 * Added Unit Tests for QAsymmS8
 * Renamed QAsymm8 calls to QAsymmU8

Signed-off-by: Ryan OShea <Ryan.OShea2@arm.com>
Change-Id: I897b4e018ba1d808cc3f8c113f2be2dbad49c8db
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp
index 0ca19bb..c60348e 100644
--- a/src/backends/reference/RefLayerSupport.cpp
+++ b/src/backends/reference/RefLayerSupport.cpp
@@ -610,7 +610,8 @@
 {
    bool supported = true;
 
-    std::array<DataType,3> supportedInputTypes = {
+    std::array<DataType,4> supportedInputTypes = {
+        DataType::QAsymmS8,
         DataType::QAsymmU8,
         DataType::QSymmS8,
         DataType::QSymmS16
@@ -1439,9 +1440,10 @@
    bool supported = true;
 
     // Define supported input types.
-    std::array<DataType,5> supportedInputTypes = {
+    std::array<DataType,6> supportedInputTypes = {
         DataType::Float32,
         DataType::Float16,
+        DataType::QAsymmS8,
         DataType::QAsymmU8,
         DataType::QSymmS8,
         DataType::QSymmS16
@@ -1451,8 +1453,9 @@
                                   "Reference quantize: input type not supported.");
 
     // Define supported output types.
-    std::array<DataType,3> supportedOutputTypes = {
+    std::array<DataType,4> supportedOutputTypes = {
         DataType::QAsymmU8,
+        DataType::QAsymmS8,
         DataType::QSymmS8,
         DataType::QSymmS16
     };
diff --git a/src/backends/reference/test/RefLayerTests.cpp b/src/backends/reference/test/RefLayerTests.cpp
index b0d8db8..99468e0 100644
--- a/src/backends/reference/test/RefLayerTests.cpp
+++ b/src/backends/reference/test/RefLayerTests.cpp
@@ -1419,6 +1419,8 @@
 // Dequantize
 ARMNN_AUTO_TEST_CASE(DequantizeSimpleUint8, DequantizeSimpleUint8Test)
 ARMNN_AUTO_TEST_CASE(DequantizeOffsetUint8, DequantizeOffsetUint8Test)
+ARMNN_AUTO_TEST_CASE(DequantizeSimpleAsymmInt8, DequantizeSimpleAsymmInt8Test)
+ARMNN_AUTO_TEST_CASE(DequantizeOffsetAsymmInt8, DequantizeOffsetAsymmInt8Test)
 ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt8, DequantizeSimpleInt8Test)
 ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt16, DequantizeSimpleInt16Test)
 ARMNN_AUTO_TEST_CASE(DequantizeSimpleUint8ToFp16, DequantizeSimpleUint8ToFp16Test)
@@ -1428,6 +1430,7 @@
 // Quantize
 ARMNN_AUTO_TEST_CASE(QuantizeSimpleUint8, QuantizeSimpleUint8Test)
 ARMNN_AUTO_TEST_CASE(QuantizeClampUint8, QuantizeClampUint8Test)
+ARMNN_AUTO_TEST_CASE(QuantizeClampAsymmInt8, QuantizeClampAsymmInt8Test)
 ARMNN_AUTO_TEST_CASE(QuantizeClampInt8, QuantizeClampInt8Test)
 ARMNN_AUTO_TEST_CASE(QuantizeClampInt16, QuantizeClampInt16Test)
 
diff --git a/src/backends/reference/workloads/BaseIterator.hpp b/src/backends/reference/workloads/BaseIterator.hpp
index 581aabf..c482018 100644
--- a/src/backends/reference/workloads/BaseIterator.hpp
+++ b/src/backends/reference/workloads/BaseIterator.hpp
@@ -137,6 +137,25 @@
     const int32_t m_Offset;
 };
 
+class QASymmS8Decoder : public TypedIterator<const int8_t, Decoder<float>>
+{
+public:
+    QASymmS8Decoder(const int8_t* data, const float scale, const int32_t offset)
+        : TypedIterator(data), m_Scale(scale), m_Offset(offset) {}
+
+    QASymmS8Decoder(const float scale, const int32_t offset)
+        : QASymmS8Decoder(nullptr, scale, offset) {}
+
+    float Get() const override
+    {
+        return armnn::Dequantize(*m_Iterator, m_Scale, m_Offset);
+    }
+
+private:
+    const float m_Scale;
+    const int32_t m_Offset;
+};
+
 class QSymmS8Decoder : public TypedIterator<const int8_t, Decoder<float>>
 {
 public:
@@ -264,6 +283,30 @@
     const int32_t m_Offset;
 };
 
+class QASymmS8Encoder : public TypedIterator<int8_t, Encoder<float>>
+{
+public:
+    QASymmS8Encoder(int8_t* data, const float scale, const int32_t offset)
+        : TypedIterator(data), m_Scale(scale), m_Offset(offset) {}
+
+    QASymmS8Encoder(const float scale, const int32_t offset)
+        : QASymmS8Encoder(nullptr, scale, offset) {}
+
+    void Set(float right) override
+    {
+        *m_Iterator = armnn::Quantize<int8_t>(right, m_Scale, m_Offset);
+    }
+
+    float Get() const override
+    {
+        return armnn::Dequantize(*m_Iterator, m_Scale, m_Offset);
+    }
+
+private:
+    const float m_Scale;
+    const int32_t m_Offset;
+};
+
 class QSymmS8Encoder : public TypedIterator<int8_t, Encoder<float>>
 {
 public:
diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp
index 6f30978..6a8c756 100644
--- a/src/backends/reference/workloads/Decoders.hpp
+++ b/src/backends/reference/workloads/Decoders.hpp
@@ -81,6 +81,13 @@
                 params.first);
         }
         ARMNN_NO_DEPRECATE_WARN_END
+        case DataType::QAsymmS8:
+        {
+            return std::make_unique<QASymmS8Decoder>(
+                static_cast<const int8_t*>(data),
+                info.GetQuantizationScale(),
+                info.GetQuantizationOffset());
+        }
         case DataType::QAsymmU8:
         {
             return std::make_unique<QASymm8Decoder>(
diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp
index 8ddd559..f522976 100644
--- a/src/backends/reference/workloads/Encoders.hpp
+++ b/src/backends/reference/workloads/Encoders.hpp
@@ -32,6 +32,13 @@
                 params.first);
         }
         ARMNN_NO_DEPRECATE_WARN_END
+        case armnn::DataType::QAsymmS8:
+        {
+            return std::make_unique<QASymmS8Encoder>(
+                static_cast<int8_t*>(data),
+                info.GetQuantizationScale(),
+                info.GetQuantizationOffset());
+        }
         case armnn::DataType::QAsymmU8:
         {
             return std::make_unique<QASymm8Encoder>(