blob: ab10a64de47ce34b020b16ab50cadaa6fd38a41e [file] [log] [blame]
Anthony Barbier6ff3b192017-09-04 18:44:23 +01001/*
Michele Di Giorgiod9eaf612020-07-08 11:12:57 +01002 * Copyright (c) 2017-2020 Arm Limited.
Anthony Barbier6ff3b192017-09-04 18:44:23 +01003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
Michalis Spyrouf4643372019-11-29 16:17:13 +000024#ifndef ARM_COMPUTE_CLSOFTMAXLAYER_H
25#define ARM_COMPUTE_CLSOFTMAXLAYER_H
Anthony Barbier6ff3b192017-09-04 18:44:23 +010026
Anthony Barbier6ff3b192017-09-04 18:44:23 +010027#include "arm_compute/runtime/CL/CLTensor.h"
SiCong Li96209c72020-08-21 12:28:30 +010028#include "arm_compute/runtime/CL/functions/CLPermute.h"
Anthony Barbier6ff3b192017-09-04 18:44:23 +010029#include "arm_compute/runtime/IFunction.h"
Georgios Pinitasbaf174e2017-09-08 19:47:30 +010030#include "arm_compute/runtime/IMemoryManager.h"
Georgios Pinitas26014cf2019-09-09 19:00:57 +010031#include "arm_compute/runtime/MemoryGroup.h"
Georgios Pinitasbaf174e2017-09-08 19:47:30 +010032
33#include <memory>
Anthony Barbier6ff3b192017-09-04 18:44:23 +010034
35namespace arm_compute
36{
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +010037class CLCompileContext;
38class CLLogits1DMaxShiftExpSumKernel;
39class CLLogits1DNormKernel;
Anthony Barbier6ff3b192017-09-04 18:44:23 +010040class ICLTensor;
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +010041class ITensorInfo;
Anthony Barbier6ff3b192017-09-04 18:44:23 +010042
43/** Basic function to compute a SoftmaxLayer.
44 *
45 * Softmax is calculated by :
Georgios Pinitas388d3ec2017-11-02 12:17:56 +000046 * @f[ out = exp((x - max(x)) * beta) / sum(exp((x - max(x)) * beta)) @f]
Anthony Barbier6ff3b192017-09-04 18:44:23 +010047 *
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000048 * Log Softmax is calculated by :
Sang-Hoon Parka0205b92020-07-07 09:36:09 +010049 * @f[ out = (x - max(x) * beta) - log(\sum{e^{x - max(x) * beta}}) @f]
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000050 *
Anthony Barbier6ff3b192017-09-04 18:44:23 +010051 * This function runs the following kernels:
SiCong Li96209c72020-08-21 12:28:30 +010052 * -# If axis is not 0:
53 * -# @ref CLPermute
Anthony Barbier6ff3b192017-09-04 18:44:23 +010054 * -# @ref CLLogits1DNormKernel
SiCong Li96209c72020-08-21 12:28:30 +010055 * -# @ref CLLogits1DMaxShiftExpSumKernel
Anthony Barbier6ff3b192017-09-04 18:44:23 +010056 */
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000057template <bool IS_LOG = false>
58class CLSoftmaxLayerGeneric : public IFunction
Anthony Barbier6ff3b192017-09-04 18:44:23 +010059{
60public:
61 /** Constructor */
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000062 CLSoftmaxLayerGeneric(std::shared_ptr<IMemoryManager> memory_manager = nullptr);
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +010063 /** Prevent instances of this class from being copied */
64 CLSoftmaxLayerGeneric(const CLSoftmaxLayerGeneric &) = delete;
65 /** Prevent instances of this class from being copied */
66 CLSoftmaxLayerGeneric &operator=(const CLSoftmaxLayerGeneric &) = delete;
67 /** Prevent instances of this class to be moved */
68 CLSoftmaxLayerGeneric(CLSoftmaxLayerGeneric &&) = delete;
69 /** Prevent instances of this class to be moved */
70 CLSoftmaxLayerGeneric &operator=(CLSoftmaxLayerGeneric &&) = delete;
71 /** Default destructor */
72 ~CLSoftmaxLayerGeneric();
Anthony Barbier6ff3b192017-09-04 18:44:23 +010073 /** Set the input and output tensors.
74 *
morgolock9c7fed82020-08-05 12:30:56 +010075 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
76 * @param[out] output Destination tensor. Data types supported: same as @p input
77 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010078 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
79 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
Anthony Barbier6ff3b192017-09-04 18:44:23 +010080 */
SiCong Li96209c72020-08-21 12:28:30 +010081 void configure(const ICLTensor *input, ICLTensor *output, float beta = 1.0f, int32_t axis = 0);
Manuel Bottini2b84be52020-04-08 10:15:51 +010082 /** Set the input and output tensors.
83 *
84 * @param[in] compile_context The compile context to be used.
Michele Di Giorgiof6f78762020-07-06 11:27:21 +010085 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
Manuel Bottini2b84be52020-04-08 10:15:51 +010086 * @param[out] output Destination tensor. Data types supported: same as @p input
87 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010088 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
89 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
Manuel Bottini2b84be52020-04-08 10:15:51 +010090 */
SiCong Li96209c72020-08-21 12:28:30 +010091 void configure(const CLCompileContext &compile_context, const ICLTensor *input, ICLTensor *output, float beta = 1.0f, int32_t axis = 0);
Georgios Pinitas30902ed2017-11-14 15:32:57 +000092 /** Static function to check if given info will lead to a valid configuration of @ref CLSoftmaxLayer
93 *
morgolock9c7fed82020-08-05 12:30:56 +010094 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
95 * @param[in] output Destination tensor. Data types supported: same as @p input
96 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010097 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
98 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
morgolock9c7fed82020-08-05 12:30:56 +010099 *
Georgios Pinitas631c41a2017-12-06 11:53:03 +0000100 * @return a status
Georgios Pinitas30902ed2017-11-14 15:32:57 +0000101 */
SiCong Li96209c72020-08-21 12:28:30 +0100102 static Status validate(const ITensorInfo *input, const ITensorInfo *output, float beta = 1.0f, int32_t axis = 0);
Anthony Barbier6ff3b192017-09-04 18:44:23 +0100103
104 // Inherited methods overridden:
105 void run() override;
106
107private:
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +0100108 MemoryGroup _memory_group;
109 CLPermute _permute_input;
110 CLPermute _permute_output;
111 std::unique_ptr<CLLogits1DMaxShiftExpSumKernel> _max_shift_exp_sum_kernel;
112 std::unique_ptr<CLLogits1DNormKernel> _norm_kernel;
113 CLTensor _max;
114 CLTensor _sum;
115 CLTensor _tmp;
116 CLTensor _input_permuted;
117 CLTensor _output_permuted;
118 bool _needs_permute;
Anthony Barbier6ff3b192017-09-04 18:44:23 +0100119};
Sang-Hoon Park62eeb532019-10-29 13:13:19 +0000120
121using CLSoftmaxLayer = CLSoftmaxLayerGeneric<false>;
122using CLLogSoftmaxLayer = CLSoftmaxLayerGeneric<true>;
123} // namespace arm_compute
Michalis Spyrouf4643372019-11-29 16:17:13 +0000124#endif /* ARM_COMPUTE_CLSOFTMAXLAYER_H */