MLBEDSW-2935: LUT fusing with preceding operator

Allows fusing of LUT with a preceding operator regardless of
input/output scale.

Change-Id: Ia378adbb3fe61d71299feb085f7313377e0efa39
Signed-off-by: Louis Verhaard <louis.verhaard@arm.com>
diff --git a/ethosu/vela/weight_compressor.py b/ethosu/vela/weight_compressor.py
index 175646b..2374cd4 100644
--- a/ethosu/vela/weight_compressor.py
+++ b/ethosu/vela/weight_compressor.py
@@ -416,13 +416,13 @@
     first_consumer_op = tens.consumer_list[0]
     ifm_dtype = first_consumer_op.inputs[0].dtype
     ifm_scale = first_consumer_op.inputs[0].quantization.scale_f32
-    ofm_scale = first_consumer_op.outputs[0].quantization.scale_f32
+    ofm_scale = first_consumer_op.get_output_quantization().scale_f32
     weight_scales = first_consumer_op.inputs[1].quantization.scale_f32
 
     # biases can have multiple consumers for rnn cells. if so, then check that they are all the same
     for op in tens.consumer_list[1:]:
         assert ifm_scale == op.inputs[0].quantization.scale_f32
-        assert ofm_scale == op.outputs[0].quantization.scale_f32
+        assert ofm_scale == op.get_output_quantization().scale_f32
         assert weight_scales == op.inputs[1].quantization.scale_f32
 
     if not hasattr(weight_scales, "__iter__"):