blob: 42adefeef6f1a84c68a456edeeaeeb0d99417790 [file] [log] [blame]
SiCong Li1e5c1572017-07-27 17:58:52 +01001/*
Anthony Barbier72856ab2018-01-11 10:45:24 +00002 * Copyright (c) 2017-2018 ARM Limited.
SiCong Li1e5c1572017-07-27 17:58:52 +01003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
24#ifndef ARM_COMPUTE_TEST_BATCHNORMALIZATIONLAYERFIXTURE
25#define ARM_COMPUTE_TEST_BATCHNORMALIZATIONLAYERFIXTURE
26
27#include "arm_compute/core/TensorShape.h"
28#include "arm_compute/core/Types.h"
SiCong Li1e5c1572017-07-27 17:58:52 +010029#include "tests/Globals.h"
30#include "tests/Utils.h"
Moritz Pflanzera09de0c2017-09-01 20:41:12 +010031#include "tests/framework/Fixture.h"
SiCong Li1e5c1572017-07-27 17:58:52 +010032
33namespace arm_compute
34{
35namespace test
36{
Michalis Spyrou724079d2017-12-15 11:37:37 +000037namespace benchmark
38{
SiCong Li1e5c1572017-07-27 17:58:52 +010039/** Fixture that can be used for NEON and CL */
40template <typename TensorType, typename Function, typename Accessor>
41class BatchNormalizationLayerFixture : public framework::Fixture
42{
43public:
44 template <typename...>
Michele Di Giorgio0cbb9272018-03-01 16:56:48 +000045 void setup(TensorShape tensor_shape, TensorShape param_shape, float epsilon, bool use_gamma, bool use_beta, ActivationLayerInfo act_info, DataType data_type, DataLayout data_layout, int batches)
SiCong Li1e5c1572017-07-27 17:58:52 +010046 {
47 // Set batched in source and destination shapes
Vidhya Sudhan Loganathan014333d2018-07-02 09:13:49 +010048
SiCong Li1e5c1572017-07-27 17:58:52 +010049 tensor_shape.set(tensor_shape.num_dimensions(), batches);
Michele Di Giorgio0cbb9272018-03-01 16:56:48 +000050 if(data_layout == DataLayout::NHWC)
51 {
52 permute(tensor_shape, PermutationVector(2U, 0U, 1U));
53 }
SiCong Li1e5c1572017-07-27 17:58:52 +010054
55 // Create tensors
Vidhya Sudhan Loganathan014333d2018-07-02 09:13:49 +010056 src = create_tensor<TensorType>(tensor_shape, data_type, 1, QuantizationInfo(), data_layout);
57 dst = create_tensor<TensorType>(tensor_shape, data_type, 1, QuantizationInfo(), data_layout);
58 mean = create_tensor<TensorType>(param_shape, data_type, 1);
59 variance = create_tensor<TensorType>(param_shape, data_type, 1);
60 beta = create_tensor<TensorType>(param_shape, data_type, 1);
61 gamma = create_tensor<TensorType>(param_shape, data_type, 1);
SiCong Li1e5c1572017-07-27 17:58:52 +010062
63 // Create and configure function
Michele Di Giorgio4d336302018-03-02 09:43:54 +000064 TensorType *beta_ptr = use_beta ? &beta : nullptr;
65 TensorType *gamma_ptr = use_gamma ? &gamma : nullptr;
66 batch_norm_layer.configure(&src, &dst, &mean, &variance, beta_ptr, gamma_ptr, epsilon, act_info);
SiCong Li1e5c1572017-07-27 17:58:52 +010067
68 // Allocate tensors
69 src.allocator()->allocate();
70 dst.allocator()->allocate();
71 mean.allocator()->allocate();
72 variance.allocator()->allocate();
73 beta.allocator()->allocate();
74 gamma.allocator()->allocate();
SiCong Li1e5c1572017-07-27 17:58:52 +010075 }
76
77 void run()
78 {
79 batch_norm_layer.run();
Joel Liang1c5ffd62017-12-28 10:09:51 +080080 }
81
82 void sync()
83 {
84 sync_if_necessary<TensorType>();
85 sync_tensor_if_necessary<TensorType>(dst);
SiCong Li1e5c1572017-07-27 17:58:52 +010086 }
87
88 void teardown()
89 {
90 src.allocator()->free();
91 dst.allocator()->free();
92 mean.allocator()->free();
93 variance.allocator()->free();
94 beta.allocator()->free();
95 gamma.allocator()->free();
96 }
97
98private:
99 TensorType src{};
100 TensorType dst{};
101 TensorType mean{};
102 TensorType variance{};
103 TensorType beta{};
104 TensorType gamma{};
105 Function batch_norm_layer{};
106};
Michalis Spyrou724079d2017-12-15 11:37:37 +0000107} // namespace benchmark
SiCong Li1e5c1572017-07-27 17:58:52 +0100108} // namespace test
109} // namespace arm_compute
110#endif /* ARM_COMPUTE_TEST_BATCHNORMALIZATIONLAYERFIXTURE */