IVGCVSW-2657: Model update to fix failing timings test: TfLiteVGG16Quantized
Signed-off-by: David Monahan <david.monahan@arm.com>

Change-Id: I49bd93766f8f3cf41a9471c87c1c61ed1ec6034c
diff --git a/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
index 84d5292..0ba1e5d 100644
--- a/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
+++ b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp
@@ -34,10 +34,10 @@
         retVal = armnn::test::ClassifierInferenceTestMain<DatabaseType,
                                                           ParserType>(
                      argc, argv,
-                     "vgg_16_u8_batch1.tflite", // model name
+                     "vgg_16_int8.tflite",      // model name
                      true,                      // model is binary
-                     "content_vgg/concat",      // input tensor name
-                     "content_vgg/prob",        // output tensor name
+                     "input",                   // input tensor name
+                     "vgg_16/fc8/squeezed",     // output tensor name
                      { 0, 1, 2 },               // test images to test with as above
                      [&imageSet](const char* dataDir, const ModelType & model) {
                          // we need to get the input quantization parameters from