COMPMID-1710: Remove Soft Relu activation from large runs.
Soft relu results in calculating infinities when running in large
shapes, thus disabling for now.
Change-Id: I61eb4d3198c69d2762001299228836e445802509
Signed-off-by: Georgios Pinitas <georgios.pinitas@arm.com>
Reviewed-on: https://review.mlplatform.org/c/784
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Reviewed-by: Michalis Spyrou <michalis.spyrou@arm.com>
diff --git a/tests/validation/CL/Winograd.cpp b/tests/validation/CL/Winograd.cpp
index 3f38cfd..e744473 100644
--- a/tests/validation/CL/Winograd.cpp
+++ b/tests/validation/CL/Winograd.cpp
@@ -147,14 +147,14 @@
ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::RELU),
ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::BOUNDED_RELU),
ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LU_BOUNDED_RELU),
- ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU),
- ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::SOFT_RELU)
+ ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU)
});
const auto ActivationFunctionsSmallDataset = framework::dataset::make("ActivationInfo",
{
ActivationLayerInfo(),
ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LU_BOUNDED_RELU),
- ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU)
+ ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU),
+ ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::SOFT_RELU)
});
} // namespace