IVGCVSW-2657: Model update to fix failing timings test: TfLiteVGG16Quantized
Signed-off-by: David Monahan <david.monahan@arm.com>
Change-Id: I49bd93766f8f3cf41a9471c87c1c61ed1ec6034c
diff --git a/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
index 84d5292..0ba1e5d 100644
--- a/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
+++ b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
@@ -34,10 +34,10 @@
retVal = armnn::test::ClassifierInferenceTestMain<DatabaseType,
ParserType>(
argc, argv,
- "vgg_16_u8_batch1.tflite", // model name
+ "vgg_16_int8.tflite", // model name
true, // model is binary
- "content_vgg/concat", // input tensor name
- "content_vgg/prob", // output tensor name
+ "input", // input tensor name
+ "vgg_16/fc8/squeezed", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
[&imageSet](const char* dataDir, const ModelType & model) {
// we need to get the input quantization parameters from