blob: 42d2ffce58bcc022e09e2eaed1ab6d0d4ca891a0 [file] [log] [blame]
Gian Marcoe75a02b2017-11-08 12:24:09 +00001/*
Michele Di Giorgio9c700372020-01-08 11:33:44 +00002 * Copyright (c) 2017-2020 ARM Limited.
Gian Marcoe75a02b2017-11-08 12:24:09 +00003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
24#include "arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h"
25
26#include "arm_compute/core/ITensor.h"
Gian Marco Iodicebc415af2019-06-13 15:58:32 +010027#include "arm_compute/core/NEON/kernels/NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel.h"
Georgios Pinitas448a81f2019-11-21 14:10:25 +000028#include "arm_compute/core/NEON/kernels/NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel.h"
Gian Marco58c57942017-11-28 09:10:03 +000029#include "arm_compute/core/NEON/kernels/NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel.h"
Gian Marcoe75a02b2017-11-08 12:24:09 +000030#include "arm_compute/core/NEON/kernels/NEGEMMLowpQuantizeDownInt32ToUint8ScaleKernel.h"
Michele Di Giorgio9c700372020-01-08 11:33:44 +000031#include "arm_compute/core/Validate.h"
Matthew Bentham758b5ba2020-03-05 23:37:48 +000032#include "support/MemorySupport.h"
Gian Marcoe75a02b2017-11-08 12:24:09 +000033
Gian Marco Iodicebc415af2019-06-13 15:58:32 +010034namespace arm_compute
35{
Gian Marco6b77e912017-11-17 09:27:57 +000036void NEGEMMLowpQuantizeDownInt32ToUint8Scale::configure(const ITensor *input, const ITensor *bias, ITensor *output, int result_offset, int result_mult_int, int result_shift, int min, int max)
Gian Marcoe75a02b2017-11-08 12:24:09 +000037{
38 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToUint8ScaleKernel>();
Gian Marco6b77e912017-11-17 09:27:57 +000039 k->configure(input, bias, output, result_offset, result_mult_int, result_shift, min, max);
Gian Marcoe75a02b2017-11-08 12:24:09 +000040 _kernel = std::move(k);
Georgios Pinitasa3b1b462017-11-16 19:24:39 +000041}
42
Georgios Pinitas631c41a2017-12-06 11:53:03 +000043Status NEGEMMLowpQuantizeDownInt32ToUint8Scale::validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min, int max)
Georgios Pinitasa3b1b462017-11-16 19:24:39 +000044{
45 return NEGEMMLowpQuantizeDownInt32ToUint8ScaleKernel::validate(input, bias, output, min, max);
Gian Marco58c57942017-11-28 09:10:03 +000046}
47
48void NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint::configure(const ITensor *input, const ITensor *bias, ITensor *output, int result_fixedpoint_multiplier, int result_shift,
Georgios Pinitasbb081ca2018-11-08 10:22:01 +000049 int result_offset_after_shift, int min, int max)
Gian Marco58c57942017-11-28 09:10:03 +000050{
51 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel>();
Georgios Pinitasbb081ca2018-11-08 10:22:01 +000052 k->configure(input, bias, output, result_fixedpoint_multiplier, result_shift, result_offset_after_shift, min, max);
Gian Marco58c57942017-11-28 09:10:03 +000053 _kernel = std::move(k);
54}
55
Georgios Pinitasbb081ca2018-11-08 10:22:01 +000056Status NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint::validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min, int max)
Gian Marco58c57942017-11-28 09:10:03 +000057{
Georgios Pinitasbb081ca2018-11-08 10:22:01 +000058 return NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel::validate(input, bias, output, min, max);
Gian Marco Iodicebc415af2019-06-13 15:58:32 +010059}
60
Georgios Pinitas448a81f2019-11-21 14:10:25 +000061void NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint::configure(const ITensor *input, const ITensor *bias, ITensor *output, int result_fixedpoint_multiplier, int result_shift,
62 int result_offset_after_shift, int min, int max)
63{
64 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel>();
65 k->configure(input, bias, output, result_fixedpoint_multiplier, result_shift, result_offset_after_shift, min, max);
66 _kernel = std::move(k);
67}
68
69Status NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint::validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min, int max)
70{
71 return NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel::validate(input, bias, output, min, max);
72}
73
Gian Marco Iodicebc415af2019-06-13 15:58:32 +010074void NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint::configure(const ITensor *input, const ITensor *bias, ITensor *output, int result_fixedpoint_multiplier, int result_shift, int min, int max)
75{
76 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel>();
77 k->configure(input, bias, output, result_fixedpoint_multiplier, result_shift, min, max);
78 _kernel = std::move(k);
79}
80
81Status NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint::validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min, int max)
82{
83 return NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel::validate(input, bias, output, min, max);
84}
Michele Di Giorgio9c700372020-01-08 11:33:44 +000085
86void NEGEMMLowpOutputStage::configure(const ITensor *input, const ITensor *bias, ITensor *output, const GEMMLowpOutputStageInfo &info)
87{
88 // Perform validate step
89 ARM_COMPUTE_ERROR_ON_NULLPTR(input, output);
90 ARM_COMPUTE_ERROR_THROW_ON(NEGEMMLowpOutputStage::validate(input->info(), bias != nullptr ? bias->info() : nullptr, output->info(), info));
91
92 if(info.type == GEMMLowpOutputStageType::QUANTIZE_DOWN)
93 {
94 switch(output->info()->data_type())
95 {
96 case DataType::QASYMM8:
97 {
98 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToUint8ScaleKernel>();
99 k->configure(input, bias, output, info.gemmlowp_multiplier, info.gemmlowp_shift, info.gemmlowp_offset, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
100 _kernel = std::move(k);
101 break;
102 }
103 default:
104 ARM_COMPUTE_ERROR("Unsupported output data type.");
105 }
106 }
107 else if(info.type == GEMMLowpOutputStageType::QUANTIZE_DOWN_FIXEDPOINT)
108 {
109 switch(output->info()->data_type())
110 {
111 case DataType::QASYMM8:
112 {
113 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel>();
114 k->configure(input, bias, output, info.gemmlowp_multiplier, info.gemmlowp_shift, info.gemmlowp_offset, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
115 _kernel = std::move(k);
116 break;
117 }
118 case DataType::QASYMM8_SIGNED:
119 {
120 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel>();
121 k->configure(input, bias, output, info.gemmlowp_multiplier, info.gemmlowp_shift, info.gemmlowp_offset, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
122 _kernel = std::move(k);
123 break;
124 }
125 case DataType::QSYMM16:
126 {
127 auto k = arm_compute::support::cpp14::make_unique<NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel>();
128 k->configure(input, bias, output, info.gemmlowp_multiplier, info.gemmlowp_shift, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
129 _kernel = std::move(k);
130 break;
131 }
132 default:
133 ARM_COMPUTE_ERROR("Unsupported output data type.");
134 }
135 }
136 else
137 {
138 ARM_COMPUTE_ERROR("Unsupported output stage quantization type.");
139 }
140}
141
142Status NEGEMMLowpOutputStage::validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, const GEMMLowpOutputStageInfo &info)
143{
144 ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(output);
145 ARM_COMPUTE_RETURN_ERROR_ON_MSG(output->data_type() == DataType::UNKNOWN, "NEGEMMLowpQuantizeDownScaleByFixedPoint cannot be used with UNKNOWN output data type.");
146 ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output, 1, DataType::QASYMM8, DataType::QASYMM8_SIGNED, DataType::QSYMM16);
147
148 ARM_COMPUTE_RETURN_ERROR_ON((info.type != GEMMLowpOutputStageType::QUANTIZE_DOWN) && (info.type != GEMMLowpOutputStageType::QUANTIZE_DOWN_FIXEDPOINT));
149
150 if(info.type == GEMMLowpOutputStageType::QUANTIZE_DOWN)
151 {
152 switch(output->data_type())
153 {
154 case DataType::QASYMM8:
155 return NEGEMMLowpQuantizeDownInt32ToUint8ScaleKernel::validate(input, bias, output, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
156 default:
157 return ARM_COMPUTE_CREATE_ERROR(ErrorCode::RUNTIME_ERROR, "Unsupported output data type.");
158 }
159 }
160 else
161 {
162 switch(output->data_type())
163 {
164 case DataType::QASYMM8:
165 return NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel::validate(input, bias, output, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
166 case DataType::QASYMM8_SIGNED:
167 return NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel::validate(input, bias, output, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
168 case DataType::QSYMM16:
169 return NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel::validate(input, bias, output, info.gemmlowp_min_bound, info.gemmlowp_max_bound);
170 default:
171 return ARM_COMPUTE_CREATE_ERROR(ErrorCode::RUNTIME_ERROR, "Unsupported output data type.");
172 }
173 }
174}
Gian Marco Iodicebc415af2019-06-13 15:58:32 +0100175} // namespace arm_compute