blob: 721a47144e99efd1ac36ca603167392cda1a6623 [file] [log] [blame]
Anthony Barbier6ff3b192017-09-04 18:44:23 +01001/*
Sang-Hoon Park201e0fe2021-01-27 13:14:56 +00002 * Copyright (c) 2017-2021 Arm Limited.
Anthony Barbier6ff3b192017-09-04 18:44:23 +01003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
Michalis Spyrouf4643372019-11-29 16:17:13 +000024#ifndef ARM_COMPUTE_CLSOFTMAXLAYER_H
25#define ARM_COMPUTE_CLSOFTMAXLAYER_H
Anthony Barbier6ff3b192017-09-04 18:44:23 +010026
Anthony Barbier6ff3b192017-09-04 18:44:23 +010027#include "arm_compute/runtime/IFunction.h"
Georgios Pinitasbaf174e2017-09-08 19:47:30 +010028#include "arm_compute/runtime/IMemoryManager.h"
Georgios Pinitas26014cf2019-09-09 19:00:57 +010029#include "arm_compute/runtime/MemoryGroup.h"
Georgios Pinitasbaf174e2017-09-08 19:47:30 +010030
31#include <memory>
Anthony Barbier6ff3b192017-09-04 18:44:23 +010032
33namespace arm_compute
34{
35class ICLTensor;
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +010036class ITensorInfo;
Sang-Hoon Park201e0fe2021-01-27 13:14:56 +000037class CLCompileContext;
Anthony Barbier6ff3b192017-09-04 18:44:23 +010038
39/** Basic function to compute a SoftmaxLayer.
40 *
41 * Softmax is calculated by :
Georgios Pinitas388d3ec2017-11-02 12:17:56 +000042 * @f[ out = exp((x - max(x)) * beta) / sum(exp((x - max(x)) * beta)) @f]
Anthony Barbier6ff3b192017-09-04 18:44:23 +010043 *
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000044 * Log Softmax is calculated by :
Sang-Hoon Parka0205b92020-07-07 09:36:09 +010045 * @f[ out = (x - max(x) * beta) - log(\sum{e^{x - max(x) * beta}}) @f]
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000046 *
Sang-Hoon Park201e0fe2021-01-27 13:14:56 +000047 * This function runs the following operators/kernels:
SiCong Li96209c72020-08-21 12:28:30 +010048 * -# If axis is not 0:
Sang-Hoon Park201e0fe2021-01-27 13:14:56 +000049 * -# @ref opencl::ClPermute
50 * -# @ref opencl::kernels::ClLogits1DNormKernel
51 * -# @ref opencl::kernels::ClLogits1DMaxShiftExpSumKernel
Anthony Barbier6ff3b192017-09-04 18:44:23 +010052 */
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000053template <bool IS_LOG = false>
54class CLSoftmaxLayerGeneric : public IFunction
Anthony Barbier6ff3b192017-09-04 18:44:23 +010055{
56public:
57 /** Constructor */
Sang-Hoon Park62eeb532019-10-29 13:13:19 +000058 CLSoftmaxLayerGeneric(std::shared_ptr<IMemoryManager> memory_manager = nullptr);
Sang-Hoon Parkbef7fa22020-10-21 15:58:54 +010059 /** Default destructor */
60 ~CLSoftmaxLayerGeneric();
Anthony Barbier6ff3b192017-09-04 18:44:23 +010061 /** Set the input and output tensors.
62 *
Sheri Zhang6124ce62021-05-04 14:03:13 +010063 * Valid data layouts:
64 * - All
65 *
66 * Valid data type configurations:
67 * |src |dst |
68 * |:--------------|:--------------|
69 * |QASYMM8 |QASYMM8 |
70 * |QASYMM8_SIGNED |QASYMM8_SIGNED |
71 * |F16 |F16 |
72 * |F32 |F32 |
73 *
morgolock9c7fed82020-08-05 12:30:56 +010074 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
75 * @param[out] output Destination tensor. Data types supported: same as @p input
76 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010077 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
78 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
Anthony Barbier6ff3b192017-09-04 18:44:23 +010079 */
SiCong Li96209c72020-08-21 12:28:30 +010080 void configure(const ICLTensor *input, ICLTensor *output, float beta = 1.0f, int32_t axis = 0);
Manuel Bottini2b84be52020-04-08 10:15:51 +010081 /** Set the input and output tensors.
82 *
83 * @param[in] compile_context The compile context to be used.
Michele Di Giorgiof6f78762020-07-06 11:27:21 +010084 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
Manuel Bottini2b84be52020-04-08 10:15:51 +010085 * @param[out] output Destination tensor. Data types supported: same as @p input
86 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010087 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
88 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
Manuel Bottini2b84be52020-04-08 10:15:51 +010089 */
SiCong Li96209c72020-08-21 12:28:30 +010090 void configure(const CLCompileContext &compile_context, const ICLTensor *input, ICLTensor *output, float beta = 1.0f, int32_t axis = 0);
Georgios Pinitas30902ed2017-11-14 15:32:57 +000091 /** Static function to check if given info will lead to a valid configuration of @ref CLSoftmaxLayer
92 *
morgolock9c7fed82020-08-05 12:30:56 +010093 * @param[in] input Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32 for Softmax and F16/F32 for Log Softmax
94 * @param[in] output Destination tensor. Data types supported: same as @p input
95 * @param[in] beta (Optional) A scaling factor for the exponent. Defaults to 1.f
SiCong Li96209c72020-08-21 12:28:30 +010096 * @param[in] axis (Optional) The dimension in which to apply the function. E.g. for input of shape 4x5x6 and
97 * axis=1, softmax will be applied to 4x6=24 vectors of size 5. Defaults to 0
morgolock9c7fed82020-08-05 12:30:56 +010098 *
Georgios Pinitas631c41a2017-12-06 11:53:03 +000099 * @return a status
Georgios Pinitas30902ed2017-11-14 15:32:57 +0000100 */
SiCong Li96209c72020-08-21 12:28:30 +0100101 static Status validate(const ITensorInfo *input, const ITensorInfo *output, float beta = 1.0f, int32_t axis = 0);
Anthony Barbier6ff3b192017-09-04 18:44:23 +0100102
103 // Inherited methods overridden:
104 void run() override;
105
106private:
Sang-Hoon Park201e0fe2021-01-27 13:14:56 +0000107 struct Impl;
108 std::unique_ptr<Impl> _impl;
109
110 /** Allocate workspace required by the operator */
111 void allocate_workspace();
Anthony Barbier6ff3b192017-09-04 18:44:23 +0100112};
Sang-Hoon Park62eeb532019-10-29 13:13:19 +0000113
114using CLSoftmaxLayer = CLSoftmaxLayerGeneric<false>;
115using CLLogSoftmaxLayer = CLSoftmaxLayerGeneric<true>;
116} // namespace arm_compute
Michalis Spyrouf4643372019-11-29 16:17:13 +0000117#endif /* ARM_COMPUTE_CLSOFTMAXLAYER_H */