Add INT64 support to inference_parser

Change-Id: I2462632f3b87df193c69b277b0944eb4d5d0efac
Signed-off-by: Jonny Svärd <jonny.svaerd@arm.com>
diff --git a/applications/inference_process/include/inference_parser.hpp b/applications/inference_process/include/inference_parser.hpp
index 1251648..a4168f0 100644
--- a/applications/inference_process/include/inference_parser.hpp
+++ b/applications/inference_process/include/inference_parser.hpp
@@ -1,6 +1,5 @@
 /*
- * SPDX-FileCopyrightText: Copyright 2022-2023 Arm Limited and/or its affiliates <open-source-office@arm.com>
- *
+ * SPDX-FileCopyrightText: Copyright 2022-2024 Arm Limited and/or its affiliates <open-source-office@arm.com>
  * SPDX-License-Identifier: Apache-2.0
  *
  * Licensed under the Apache License, Version 2.0 (the License); you may
@@ -137,6 +136,9 @@
         case tflite::TensorType::TensorType_FLOAT32:
             size = 4;
             break;
+        case tflite::TensorType::TensorType_INT64:
+            size = 8;
+            break;
         default:
             printf("Warning: Unsupported tensor type\n");
             return true;