MLBEDSW-7501: Vela unnecessary adds reshaped weights tensors

- Weights are internally cloned and reshaped/transposed when
running on the NPU. This happens already in the reader. If
the op is passed through to the CPU there are code that writes
backs these clones but with another round of reshape/transpose.
This adds extra tensors in the optimized file compared to the
original file if the original tensors are subgraph inputs.

- If the op is passed trough to the CPU the clones should not
be written to the file. Solved this by setting the src_tensor
when making the clone.

Change-Id: I9f55d542c099882882920bffe8e15b43b2ca2c8d
Signed-off-by: Johan Alfven <johan.alfven@arm.com>
diff --git a/ethosu/vela/tensor.py b/ethosu/vela/tensor.py
index 9ba6ab7..6ba331c 100644
--- a/ethosu/vela/tensor.py
+++ b/ethosu/vela/tensor.py
@@ -506,6 +506,7 @@
         res.name = res.name + suffix
         res.ops = []
         res.consumer_list = []
+        res.src_tensor = self
 
         return res