Rename quantized data types
!armnn:2571
Signed-off-by: Derek Lamberti <derek.lamberti@arm.com>
Change-Id: I06977553a097479a2a996e76a106249673d31ed7
diff --git a/Utils.cpp b/Utils.cpp
index 0211e92..cdebfae 100644
--- a/Utils.cpp
+++ b/Utils.cpp
@@ -43,7 +43,7 @@
{
case armnn::DataType::Float16:
case armnn::DataType::Float32:
- case armnn::DataType::QuantisedAsymm8:
+ case armnn::DataType::QAsymmU8:
case armnn::DataType::QuantizedSymm8PerAxis:
SwizzleAndroidNn4dTensorToArmNn(tensor.GetShape(), input, output, armnn::GetDataTypeSize(dataType), mappings);
break;
@@ -83,7 +83,7 @@
type = armnn::DataType::Float32;
break;
case V1_0::OperandType::TENSOR_QUANT8_ASYMM:
- type = armnn::DataType::QuantisedAsymm8;
+ type = armnn::DataType::QAsymmU8;
break;
case V1_0::OperandType::TENSOR_INT32:
type = armnn::DataType::Signed32;
@@ -119,13 +119,13 @@
type = armnn::DataType::QuantizedSymm8PerAxis;
break;
case V1_2::OperandType::TENSOR_QUANT8_ASYMM:
- type = armnn::DataType::QuantisedAsymm8;
+ type = armnn::DataType::QAsymmU8;
break;
case V1_2::OperandType::TENSOR_QUANT8_SYMM:
type = armnn::DataType::QSymmS8;
break;
case V1_2::OperandType::TENSOR_QUANT16_SYMM:
- type = armnn::DataType::QuantisedSymm16;
+ type = armnn::DataType::QSymmS16;
break;
case V1_2::OperandType::TENSOR_INT32:
type = armnn::DataType::Signed32;
@@ -228,7 +228,7 @@
dumpElementFunction = &DumpTensorElement<float>;
break;
}
- case armnn::DataType::QuantisedAsymm8:
+ case armnn::DataType::QAsymmU8:
{
dumpElementFunction = &DumpTensorElement<uint8_t, uint32_t>;
break;