IVGCVSW-6289 Separate tensor shape inference and validation calls
* Pass m_shapeInferenceMethod to OptimizerOptions in ExecuteNetwork
Signed-off-by: Finn Williams <Finn.Williams@arm.com>
Change-Id: I90280fb7629092d3b66e8a3968ca9e35a0df854a
diff --git a/src/armnn/Network.cpp b/src/armnn/Network.cpp
index 365f1bd..42d7ae3 100644
--- a/src/armnn/Network.cpp
+++ b/src/armnn/Network.cpp
@@ -1594,13 +1594,22 @@
// Get the optimized graph
Graph& optGraph = optNetObjPtr->pOptimizedNetworkImpl->GetGraph();
- // Infer the tensor infos for all output slots. Throws an exception on failure
- optGraph.InferTensorInfos();
+ if(options.m_shapeInferenceMethod == ShapeInferenceMethod::InferAndValidate)
+ {
+ // Infer the tensor infos for all output slots. Throws an exception on failure
+ optGraph.InferTensorInfos();
+ }
// Perform AddBroadcastReshapeLayer optimisation
using namespace optimizations;
Optimizer::Pass(optGraph, MakeOptimizations(AddBroadcastReshapeLayer()));
+ if(options.m_shapeInferenceMethod == ShapeInferenceMethod::ValidateOnly)
+ {
+ // Validate the tensor infos for all output slots. Throws an exception on failure
+ optGraph.InferTensorInfos();
+ }
+
// Perform optimisation passes
Optimizer::Pass(optGraph, MakeOptimizations(SquashEqualPermuteSiblings(),
SquashEqualTransposeSiblings(),