blob: 232b9f59b6569b44c9967ae20881de519541715a [file] [log] [blame]
giuros014a8ec802019-03-18 13:25:05 +00001/*
Michele Di Giorgiod9eaf612020-07-08 11:12:57 +01002 * Copyright (c) 2019-2020 Arm Limited.
giuros014a8ec802019-03-18 13:25:05 +00003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
Michalis Spyrouf4643372019-11-29 16:17:13 +000024#ifndef ARM_COMPUTE_CLDIRECTDECONVOLUTIONLAYER_H
25#define ARM_COMPUTE_CLDIRECTDECONVOLUTIONLAYER_H
giuros014a8ec802019-03-18 13:25:05 +000026
27#include "arm_compute/runtime/CL/functions/CLConvolutionLayer.h"
28#include "arm_compute/runtime/CL/functions/CLDeconvolutionLayerUpsample.h"
Georgios Pinitasdbfc2dc2019-04-02 12:51:21 +010029#include "arm_compute/runtime/CL/functions/CLReverse.h"
giuros014a8ec802019-03-18 13:25:05 +000030#include "arm_compute/runtime/CL/functions/CLTranspose.h"
31
giuros014a8ec802019-03-18 13:25:05 +000032#include "arm_compute/runtime/CL/CLTensor.h"
33#include "arm_compute/runtime/IFunction.h"
34#include "arm_compute/runtime/IMemoryManager.h"
Georgios Pinitas26014cf2019-09-09 19:00:57 +010035#include "arm_compute/runtime/MemoryGroup.h"
giuros014a8ec802019-03-18 13:25:05 +000036
37#include <memory>
38
39namespace arm_compute
40{
41class ICLTensor;
42/** Function to run the deconvolution layer.
43 *
44 * Deconvolution Layer is the backward pass of Convolution Layer. First we transform the input depending on the stride and pad info and then perform a 1x1
45 * convolution pass. Input stride defines how many zeroes we should put between each element of the input and pad is the amount of padding.
46 *
47 * The relation between input to output is as follows:
48 * \f[
49 * width\_output = (width\_input - 1) \cdot stride\_x - 2 \cdot padding\_x + kernel\_x
50 * \f]
51 * \f[
52 * height\_output = (height\_input - 1) \cdot stride\_y - 2 \cdot padding\_y + kernel\_y
53 * \f]
54 *
55 * where:
56 * width_input is the size of the first input dimension.
57 * height_input is the size of the second input dimension.
58 * width_output is the size of the first output dimension.
59 * height_output is the size of the second output dimension.
60 * kernel_x and kernel_y are the convolution sizes in x and y.
61 * stride_x and stride_y is the input stride of the first and second dimension.
62 *
63 * The weights used by Deconvolution are supposed to be the same as the ones used for Convolution. Therefore, it will be necessary to use the weights in the
Luca Foschianifedefc32020-02-17 17:02:49 +000064 * reverse order to perform an actual convolution. This is achieved by using @ref CLReverse.
giuros014a8ec802019-03-18 13:25:05 +000065 *
66 * This function calls the following OpenCL kernels/functions:
67 *
68 * -# @ref CLDeconvolutionLayerUpsample
69 * -# @ref CLConvolutionLayer
70 *
71 * And the following CPP kernels:
Georgios Pinitasdbfc2dc2019-04-02 12:51:21 +010072 * -# @ref CLReverse
giuros014a8ec802019-03-18 13:25:05 +000073 *
74 */
75class CLDirectDeconvolutionLayer : public IFunction
76{
77public:
78 /** Constructor */
79 CLDirectDeconvolutionLayer(std::shared_ptr<IMemoryManager> memory_manager = nullptr);
80 /** Prevent instances of this class from being copied (As this class contains pointers) */
81 CLDirectDeconvolutionLayer(const CLDirectDeconvolutionLayer &) = delete;
82 /** Default move constructor */
83 CLDirectDeconvolutionLayer(CLDirectDeconvolutionLayer &&) = default;
84 /** Prevent instances of this class from being copied (As this class contains pointers) */
85 CLDirectDeconvolutionLayer &operator=(const CLDirectDeconvolutionLayer &) = delete;
86 /** Default move assignment operator */
87 CLDirectDeconvolutionLayer &operator=(CLDirectDeconvolutionLayer &&) = default;
88 /** Set the input, weights, biases and output tensors.
89 *
Sheri Zhang0ef60322020-02-20 17:37:12 +000090 * @param[in,out] input Input tensor. 3 lower dimensions represent a single input, and an optional 4th dimension for batch of inputs.
91 * Data types supported: QASYMM8_SIGNED/QASYMM8/F16/F32.
giuros014a8ec802019-03-18 13:25:05 +000092 * @param[in] weights The 4d weights with dimensions [width, height, IFM, OFM]. Data type supported: Same as @p input.
Sheri Zhang0ef60322020-02-20 17:37:12 +000093 * @param[in] bias (Optional) The biases have one dimension.
94 * Data type supported: Should match @p input data type, except for input of QASYMM8 and QASYMM8_SIGNED type where biases should be of S32 type
giuros014a8ec802019-03-18 13:25:05 +000095 * @param[out] output Output tensor. The output has the same number of dimensions as the @p input.
96 * @param[in] info Contains padding and policies to be used in the deconvolution, this is decribed in @ref PadStrideInfo.
97 * @param[in] weights_info (Optional) Weights information needed for @ref CLConvolutionLayer, specifies if the weights tensor has been reshaped with @ref CLWeightsReshapeKernel.
98 *
99 */
100 void configure(ICLTensor *input, ICLTensor *weights, const ICLTensor *bias, ICLTensor *output, const PadStrideInfo &info, const WeightsInfo &weights_info = WeightsInfo());
Manuel Bottini2b84be52020-04-08 10:15:51 +0100101 /** Set the input, weights, biases and output tensors.
102 *
103 * @param[in] compile_context The compile context to be used.
104 * @param[in,out] input Input tensor. 3 lower dimensions represent a single input, and an optional 4th dimension for batch of inputs.
105 * Data types supported: QASYMM8_SIGNED/QASYMM8/F16/F32.
106 * @param[in] weights The 4d weights with dimensions [width, height, IFM, OFM]. Data type supported: Same as @p input.
107 * @param[in] bias (Optional) The biases have one dimension.
108 * Data type supported: Should match @p input data type, except for input of QASYMM8 and QASYMM8_SIGNED type where biases should be of S32 type
109 * @param[out] output Output tensor. The output has the same number of dimensions as the @p input.
110 * @param[in] info Contains padding and policies to be used in the deconvolution, this is decribed in @ref PadStrideInfo.
111 * @param[in] weights_info (Optional) Weights information needed for @ref CLConvolutionLayer, specifies if the weights tensor has been reshaped with @ref CLWeightsReshapeKernel.
112 *
113 */
114 void configure(const CLCompileContext &compile_context, ICLTensor *input, ICLTensor *weights, const ICLTensor *bias, ICLTensor *output, const PadStrideInfo &info,
115 const WeightsInfo &weights_info = WeightsInfo());
giuros014a8ec802019-03-18 13:25:05 +0000116 /** Static function to check if given info will lead to a valid configuration of @ref CLDirectDeconvolutionLayer
117 *
Sheri Zhang0ef60322020-02-20 17:37:12 +0000118 * @param[in] input Input tensor info. 3 lower dimensions represent a single input, and an optional 4th dimension for batch of inputs.
119 * Data types supported: QASYMM8_SIGNED/QASYMM8/F16/F32.
giuros014a8ec802019-03-18 13:25:05 +0000120 * @param[in] weights The 4d weights info with dimensions [width, height, IFM, OFM]. Data type supported: Same as @p input.
Sheri Zhang0ef60322020-02-20 17:37:12 +0000121 * @param[in] bias (Optional) The biases have one dimension.
122 * Data type supported: Should match @p input data type, except for input of QASYMM8 and QASYMM8_SIGNED type where biases should be of S32 type
giuros014a8ec802019-03-18 13:25:05 +0000123 * @param[in] output Output tensor info. The output has the same number of dimensions as the @p input.
124 * @param[in] info Contains padding and policies to be used in the deconvolution, this is decribed in @ref PadStrideInfo.
125 * @param[in] weights_info (Optional) Weights information needed for @ref CLConvolutionLayer, specifies if the weights tensor has been reshaped with @ref CLWeightsReshapeKernel.
126 *
127 * @return a status
128 */
129 static Status validate(const ITensorInfo *input, const ITensorInfo *weights, const ITensorInfo *bias, ITensorInfo *output, const PadStrideInfo &info,
130 const WeightsInfo &weights_info = WeightsInfo());
131
132 // Inherited methods overridden:
133 void run() override;
134 void prepare() override;
135
136private:
Georgios Pinitas26014cf2019-09-09 19:00:57 +0100137 MemoryGroup _memory_group;
giuros014a8ec802019-03-18 13:25:05 +0000138 CLDeconvolutionLayerUpsample _scale_f;
139 CLConvolutionLayer _conv_f;
Georgios Pinitasdbfc2dc2019-04-02 12:51:21 +0100140 CLReverse _flip_weights;
giuros014a8ec802019-03-18 13:25:05 +0000141
142 CLTensor _scaled_output;
143 ICLTensor *_original_weights;
144 CLTensor _weights_flipped;
Georgios Pinitasdbfc2dc2019-04-02 12:51:21 +0100145 CLTensor _flip_axis;
giuros014a8ec802019-03-18 13:25:05 +0000146
147 bool _is_prepared;
148};
149} // namespace arm_compute
Michalis Spyrouf4643372019-11-29 16:17:13 +0000150#endif /* ARM_COMPUTE_CLDECONVOLUTIONLAYER_H */