IVGCVSW-7333 Add missing TFLite Delegate operator support to documentation

Signed-off-by: Matthew Sloyan <matthew.sloyan@arm.com>
Change-Id: Id7747e85119eb70254ad7795168ce31b137f2f49
diff --git a/docs/05_03_delegate.dox b/docs/05_03_delegate.dox
index 9d92627..ffd583d 100644
--- a/docs/05_03_delegate.dox
+++ b/docs/05_03_delegate.dox
@@ -37,13 +37,13 @@
 
 - ADD
 
-- ARGMAX
+- ARG_MAX
 
-- ARGMIN
+- ARG_MIN
 
-- AVERAGE_POOL_2D, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- AVERAGE_POOL_2D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
-- AVERAGE_POOL_3D
+- AVERAGE_POOL_3D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, SIGN_BIT, TANH, NONE
 
 - BATCH_MATMUL
 
@@ -51,15 +51,15 @@
 
 - CAST
 
-- CONCATENATION, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- CONCATENATION, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
-- CONV_2D, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- CONV_2D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
-- CONV_3D, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- CONV_3D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
 - DEPTH_TO_SPACE
 
-- DEPTHWISE_CONV_2D, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- DEPTHWISE_CONV_2D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
 - DEQUANTIZE
 
@@ -71,23 +71,33 @@
 
 - EXP
 
-- FULLY_CONNECTED, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- EXPAND_DIMS
+
+- FILL
 
 - FLOOR
 
+- FLOOR_DIV
+
+- FULLY_CONNECTED, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
+
 - GATHER
 
 - GATHER_ND
 
 - GREATER
 
-- GREATER_OR_EQUAL
+- GREATER_EQUAL
 
 - HARD_SWISH
 
+- L2_NORMALIZATION
+
+- L2_POOL_2D
+
 - LESS
 
-- LESS_OR_EQUAL
+- LESS_EQUAL
 
 - LOCAL_RESPONSE_NORMALIZATION
 
@@ -105,15 +115,11 @@
 
 - LSTM
 
-- L2_NORMALIZATION
-
-- L2_POOL_2D
-
 - MAXIMUM
 
-- MAX_POOL_2D, Supported Fused Activation: RELU , RELU6 , TANH, NONE
+- MAX_POOL_2D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, TANH, NONE
 
-- MAX_POOL_3D
+- MAX_POOL_3D, Supported Fused Activation: RELU, RELU6, RELU_N1_TO_1, SIGMOID, SIGN_BIT, TANH, NONE
 
 - MEAN
 
@@ -131,6 +137,8 @@
 
 - PAD
 
+- PADV2
+
 - PRELU
 
 - QUANTIZE
@@ -141,16 +149,20 @@
 
 - REDUCE_MIN
 
+- REDUCE_PROD
+
+- RELU
+
+- RELU6
+
+- RELU_N1_TO_1
+
 - RESHAPE
 
 - RESIZE_BILINEAR
 
 - RESIZE_NEAREST_NEIGHBOR
 
-- RELU
-
-- RELU6
-
 - RSQRT
 
 - SHAPE
@@ -169,6 +181,8 @@
 
 - SQRT
 
+- SQUEEZE
+
 - STRIDED_SLICE
 
 - SUB