blob: 36275e68bf612b3ec9f732d5fb34a12ee1f1f6e2 [file] [log] [blame]
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001///
Renato Arantes36a75da2024-01-26 17:31:18 +00002/// Copyright (c) 2021-2023,2024 Arm Limited.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01003///
4/// SPDX-License-Identifier: MIT
5///
6/// Permission is hereby granted, free of charge, to any person obtaining a copy
7/// of this software and associated documentation files (the "Software"), to
8/// deal in the Software without restriction, including without limitation the
9/// rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10/// sell copies of the Software, and to permit persons to whom the Software is
11/// furnished to do so, subject to the following conditions:
12///
13/// The above copyright notice and this permission notice shall be included in all
14/// copies or substantial portions of the Software.
15///
16/// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17/// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18/// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19/// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20/// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21/// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22/// SOFTWARE.
23///
24namespace arm_compute
25{
26/**
27@page operators_list Supported Operators
28
29@tableofcontents
30
31@section S9_1_operators_list Supported Operators
32
33Compute Library supports operators that are listed in below table.
34
35Compute Library supports a wide list of data-types, information can been directly found in the documentation of each kernel/function.
36The main data-types that the Machine Learning functions support are the following:
37 <ul>
38 <li>BFLOAT16: 16-bit non-standard brain floating point
39 <li>QASYMM8: 8-bit unsigned asymmetric quantized
40 <li>QASYMM8_SIGNED: 8-bit signed asymmetric quantized
41 <li>QSYMM8_PER_CHANNEL: 8-bit signed symmetric quantized (Used for the weights)
42 <li>QSYMM8: 8-bit unsigned symmetric quantized
43 <li>QSYMM16: 16-bit unsigned symmetric quantized
44 <li>F32: 32-bit single precision floating point
45 <li>F16: 16-bit half precision floating point
46 <li>S32: 32-bit signed integer
47 <li>U8: 8-bit unsigned char
Jakub Sujakee301b32021-06-04 09:46:08 +010048 <li>All: Agnostic to any specific data type
Sheri Zhanga47dcc22021-04-22 14:41:12 +010049 </ul>
50
51Compute Library supports the following data layouts (fast changing dimension from right to left):
52 <ul>
53 <li>NHWC: The native layout of Compute Library that delivers the best performance where channels are in the fastest changing dimension
54 <li>NCHW: Legacy layout where width is in the fastest changing dimension
Sheri Zhang5dda2172021-10-15 19:54:17 +010055 <li>NDHWC: New data layout for supporting 3D operators
Jakub Sujakee301b32021-06-04 09:46:08 +010056 <li>All: Agnostic to any specific data layout
Sheri Zhanga47dcc22021-04-22 14:41:12 +010057 </ul>
Sheri Zhang5dda2172021-10-15 19:54:17 +010058where N = batches, C = channels, H = height, W = width, D = depth
Sheri Zhanga47dcc22021-04-22 14:41:12 +010059
60<table>
61<caption id="multi_row"></caption>
62<tr>
63 <th>Function
64 <th>Description
65 <th>Equivalent Android NNAPI Op
66 <th>Backends
67 <th>Data Layouts
68 <th>Data Types
69<tr>
70 <td rowspan="2">ActivationLayer
71 <td rowspan="2" style="width:200px;"> Function to simulate an activation layer with the specified activation function.
72 <td rowspan="2">
73 <ul>
74 <li>ANEURALNETWORKS_ELU
75 <li>ANEURALNETWORKS_HARD_SWISH
76 <li>ANEURALNETWORKS_LOGISTIC
77 <li>ANEURALNETWORKS_RELU
78 <li>ANEURALNETWORKS_RELU1
79 <li>ANEURALNETWORKS_RELU6
80 <li>ANEURALNETWORKS_TANH
81 </ul>
82 <td>NEActivationLayer
83 <td>
84 <ul>
85 <li>All
86 </ul>
87 <td>
88 <table>
89 <tr><th>src<th>dst
90 <tr><td>QASYMM8<td>QASYMM8
91 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
92 <tr><td>QSYMM16<td>QSYMM16
93 <tr><td>F16<td>F16
94 <tr><td>F32<td>F32
95 </table>
96<tr>
97 <td>CLActivationLayer
98 <td>
99 <ul>
100 <li>All
101 </ul>
102 <td>
103 <table>
104 <tr><th>src<th>dst
105 <tr><td>QASYMM8<td>QASYMM8
106 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
107 <tr><td>QSYMM16<td>QSYMM16
108 <tr><td>F16<td>F16
109 <tr><td>F32<td>F32
110 </table>
111<tr>
Jakub Sujak667e82f2023-11-07 22:39:30 +0000112 <td rowspan="1">AddMulAdd
113 <td rowspan="1" style="width:200px;"> Performs a fused Add + Mul + Add [+ Relu-based-Activation] operation.
114 <td rowspan="1">
115 <ul>
116 <li>n/a
117 </ul>
118 <td>NEAddMulAdd
119 <td>
120 <ul>
121 <li>Any
122 </ul>
123 <td>
124 <table>
125 <tr><th>input1<th>input2<th>bn_mul<th>bn_add<th>add_output<th>final_output
126 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8<td>QASYMM8<td>QASYMM8<td>QASYMM8
127 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
128 <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16
129 <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32
130 </table>
131<tr>
Teresa Charlin62687422021-04-28 10:58:49 +0100132 <td rowspan="2">ArgMinMaxLayer
133 <td rowspan="2" style="width:200px;"> Function to calculate the index of the minimum or maximum values in a tensor based on an axis.
134 <td rowspan="2">
135 <ul>
136 <li>ANEURALNETWORKS_ARGMAX
137 <li>ANEURALNETWORKS_ARGMIN
138 </ul>
139 <td>NEArgMinMaxLayer
140 <td>
141 <ul>
142 <li>All
143 </ul>
144 <td>
145 <table>
146 <tr><th>src<th>dst
147 <tr><td>QASYMM8<td>U32, S32
148 <tr><td>QASYMM8_SIGNED<td>U32, S32
Pablo Marquez Tello29e27b02023-08-03 14:47:31 +0100149 <tr><td>S32<td>U32, S32, S64
Teresa Charlin62687422021-04-28 10:58:49 +0100150 <tr><td>F16<td>U32, S32
151 <tr><td>F32<td>U32, S32
152 </table>
153<tr>
154 <td>CLArgMinMaxLayer
155 <td>
156 <ul>
157 <li>All
158 </ul>
159 <td>
160 <table>
161 <tr><th>src<th>dst
162 <tr><td>QASYMM8<td>U32, S32
163 <tr><td>QASYMM8_SIGNED<td>U32, S32
164 <tr><td>S32<td>U32, S32
165 <tr><td>F16<td>U32, S32
166 <tr><td>F32<td>U32, S32
167 </table>
168<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100169 <td rowspan="1">ArithmeticAddition
170 <td rowspan="1" style="width:200px;"> Function to add 2 tensors.
171 <td rowspan="1">
172 <ul>
173 <li>ANEURALNETWORKS_ADD
174 </ul>
175 <td>NEArithmeticAddition
176 <td>
177 <ul>
178 <li>All
179 </ul>
180 <td>
181 <table>
182 <tr><th>src0<th>src1<th>dst
183 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
184 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
185 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
186 <tr><td>QSYMM16<td>QSYMM16<td>S32
187 <tr><td>U8<td>U8<td>U8
Sheri Zhang6124ce62021-05-04 14:03:13 +0100188 <tr><td>S16<td>S16<td>S16
189 <tr><td>S32<td>S32<td>S32
190 <tr><td>F16<td>F16<td>F16
191 <tr><td>F32<td>F32<td>F32
192 </table>
193<tr>
194 <td rowspan="1">ArithmeticSubtraction
195 <td rowspan="1" style="width:200px;"> Function to substract 2 tensors.
196 <td rowspan="1">
197 <ul>
198 <li>ANEURALNETWORKS_SUB
199 </ul>
200 <td>NEArithmeticSubtraction
201 <td>
202 <ul>
203 <li>All
204 </ul>
205 <td>
206 <table>
207 <tr><th>src0<th>src1<th>dst
208 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
209 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
210 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
211 <tr><td>QSYMM16<td>QSYMM16<td>S32
212 <tr><td>U8<td>U8<td>U8
Sheri Zhang6124ce62021-05-04 14:03:13 +0100213 <tr><td>S16<td>S16<td>S16
214 <tr><td>S32<td>S32<td>S32
215 <tr><td>F16<td>F16<td>F16
216 <tr><td>F32<td>F32<td>F32
217 </table>
218<tr>
Teresa Charlin62687422021-04-28 10:58:49 +0100219 <td rowspan="2">BatchNormalizationLayer
220 <td rowspan="2" style="width:200px;"> Function to perform batch normalization.
221 <td rowspan="2">
222 <ul>
223 <li>n/a
224 </ul>
225 <td>NEBatchNormalizationLayer
226 <td>
227 <ul>
228 <li>NHWC
229 <li>NCHW
230 </ul>
231 <td>
232 <table>
233 <tr><th>src<th>dst
234 <tr><td>F32<td>F32
235 <tr><td>F16<td>F16
236 </table>
237<tr>
238 <td>CLBatchNormalizationLayer
239 <td>
240 <ul>
241 <li>NHWC
242 <li>NCHW
243 </ul>
244 <td>
245 <table>
246 <tr><th>src<th>dst
247 <tr><td>F32<td>F32
248 <tr><td>F16<td>F16
249 </table>
250<tr>
251 <td rowspan="2">BatchToSpaceLayer
252 <td rowspan="2" style="width:200px;"> Batch to space transformation.
253 <td rowspan="2">
254 <ul>
255 <li>ANEURALNETWORKS_BATCH_TO_SPACE_ND
256 </ul>
257 <td>NEBatchToSpaceLayer
258 <td>
259 <ul>
260 <li>NHWC
261 <li>NCHW
262 </ul>
263 <td>
264 <table>
265 <tr><th>src0<th>src1<th>dst
266 <tr><td>All<td>s32<td>All
267 </table>
268<tr>
269 <td>CLBatchToSpaceLayer
270 <td>
271 <ul>
272 <li>NHWC
273 <li>NCHW
274 </ul>
275 <td>
276 <table>
277 <tr><th>src0<th>src1<th>dst
278 <tr><td>All<td>s32<td>All
279 </table>
280<tr>
281 <td rowspan="2">BitwiseAnd
Jakub Sujakee301b32021-06-04 09:46:08 +0100282 <td rowspan="2" style="width:200px;"> Function to perform bitwise AND between 2 tensors.
Teresa Charlin62687422021-04-28 10:58:49 +0100283 <td rowspan="2">
284 <ul>
285 <li>ANEURALNETWORKS_LOGICAL_AND
286 </ul>
287 <td>NEBitwiseAnd
288 <td>
289 <ul>
290 <li>All
291 </ul>
292 <td>
293 <table>
294 <tr><th>src<th>dst
295 <tr><td>U8<td>U8
296 </table>
297<tr>
298 <td>CLBitwiseAnd
299 <td>
300 <ul>
301 <li>All
302 </ul>
303 <td>
304 <table>
305 <tr><th>src<th>dst
306 <tr><td>U8<td>U8
307 </table>
308<tr>
309 <td rowspan="2">BitwiseNot
Jakub Sujakee301b32021-06-04 09:46:08 +0100310 <td rowspan="2" style="width:200px;"> Function to perform bitwise NOT.
Teresa Charlin62687422021-04-28 10:58:49 +0100311 <td rowspan="2">
312 <ul>
313 <li>ANEURALNETWORKS_LOGICAL_NOT
314 </ul>
315 <td>NEBitwiseNot
316 <td>
317 <ul>
318 <li>All
319 </ul>
320 <td>
321 <table>
322 <tr><th>src<th>dst
323 <tr><td>U8<td>U8
324 </table>
325<tr>
326 <td>CLBitwiseNot
327 <td>
328 <ul>
329 <li>All
330 </ul>
331 <td>
332 <table>
333 <tr><th>src<th>dst
334 <tr><td>U8<td>U8
335 </table>
336<tr>
337 <td rowspan="2">BitwiseOr
Jakub Sujakee301b32021-06-04 09:46:08 +0100338 <td rowspan="2" style="width:200px;"> Function to perform bitwise OR between 2 tensors.
Teresa Charlin62687422021-04-28 10:58:49 +0100339 <td rowspan="2">
340 <ul>
341 <li>ANEURALNETWORKS_LOGICAL_OR
342 </ul>
343 <td>NEBitwiseOr
344 <td>
345 <ul>
346 <li>All
347 </ul>
348 <td>
349 <table>
350 <tr><th>src<th>dst
351 <tr><td>U8<td>U8
352 </table>
353<tr>
354 <td>CLBitwiseOr
355 <td>
356 <ul>
357 <li>All
358 </ul>
359 <td>
360 <table>
361 <tr><th>src<th>dst
362 <tr><td>U8<td>U8
363 </table>
364<tr>
365 <td rowspan="2">BitwiseXor
Jakub Sujakee301b32021-06-04 09:46:08 +0100366 <td rowspan="2" style="width:200px;"> Function to perform bitwise XOR between 2 tensors.
Teresa Charlin62687422021-04-28 10:58:49 +0100367 <td rowspan="2">
368 <ul>
369 <li>n/a
370 </ul>
371 <td>NEBitwiseXor
372 <td>
373 <ul>
374 <li>All
375 </ul>
376 <td>
377 <table>
378 <tr><th>src<th>dst
379 <tr><td>U8<td>U8
380 </table>
381<tr>
382 <td>CLBitwiseXor
383 <td>
384 <ul>
385 <li>All
386 </ul>
387 <td>
388 <table>
389 <tr><th>src<th>dst
390 <tr><td>U8<td>U8
391 </table>
392<tr>
393 <td rowspan="2">BoundingBoxTransform
394 <td rowspan="2" style="width:200px;"> Transform proposal bounding boxes to target bounding box using bounding box deltas.
395 <td rowspan="2">
396 <ul>
397 <li>n/a
398 </ul>
399 <td>NEBoundingBoxTransform
400 <td>
401 <ul>
402 <li>NHWC
403 <li>NCHW
404 </ul>
405 <td>
406 <table>
407 <tr><th>src0<th>src1<th>dst
408 <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16
409 <tr><td>F16<td>F16<td>F16
410 <tr><td>F32<td>F32<td>F32
411 </table>
412<tr>
413 <td>CLBoundingBoxTransform
414 <td>
415 <ul>
416 <li>NHWC
417 <li>NCHW
418 </ul>
419 <td>
420 <table>
421 <tr><th>src0<th>src1<th>dst
422 <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16
423 <tr><td>F16<td>F16<td>F16
424 <tr><td>F32<td>F32<td>F32
425 </table>
426<tr>
427 <td rowspan="2">Cast
428 <td rowspan="2" style="width:200px;"> Function to cast a tensor.
429 <td rowspan="2">
430 <ul>
431 <li>ANEURALNETWORKS_CAST
432 </ul>
433 <td>NECast
434 <td>
435 <ul>
436 <li>All
437 </ul>
438 <td>
439 <table>
440 <tr><th>src<th>dst
441 <tr><td>QASYMM8_SIGNED<td>S16, S32, F32, F16
442 <tr><td>QASYMM8<td>U16, S16, S32, F32, F16
443 <tr><td>U8<td>U16, S16, S32, F32, F16
444 <tr><td>U16<td>U8, U32
445 <tr><td>S16<td>QASYMM8_SIGNED, U8, S32
446 <tr><td>F16<td>QASYMM8_SIGNED, QASYMM8, F32, S32, U8
447 <tr><td>S32<td>QASYMM8_SIGNED, QASYMM8, F16, F32, U8
448 <tr><td>F32<td>QASYMM8_SIGNED, QASYMM8, BFLOAT16, F16, S32, U8
449 </table>
450<tr>
451 <td>CLCast
452 <td>
453 <ul>
454 <li>All
455 </ul>
456 <td>
457 <table>
458 <tr><th>src<th>dst
459 <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32
Pablo Marquez Tello205ba242023-07-12 14:29:58 +0100460 <tr><td>S8<td>U8, U16, S16, U32, S32, F16, F32
Teresa Charlin62687422021-04-28 10:58:49 +0100461 <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32
462 <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32
463 <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32
464 <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32
Pablo Marquez Tello205ba242023-07-12 14:29:58 +0100465 <tr><td>U64<td>U8, S8, U16, S16, U32, S32, F16, F32
466 <tr><td>S64<td>U8, S8, U16, S16, U32, S32, F16, F32
467 <tr><td>F16<td>U8, S8, U16, S16, S32, U32, F32
468 <tr><td>F32<td>U8, S8, U16, S16, S32, U32, F16
Teresa Charlin62687422021-04-28 10:58:49 +0100469 </table>
470<tr>
471 <td rowspan="2">ChannelShuffleLayer
472 <td rowspan="2" style="width:200px;"> Function to shuffle the channels of the input tensor.
473 <td rowspan="2">
474 <ul>
475 <li>ANEURALNETWORKS_CHANNEL_SHUFFLE
476 </ul>
477 <td>NEChannelShuffleLayer
478 <td>
479 <ul>
480 <li>NCHW
Michele Di Giorgiob8025b32021-09-03 10:29:49 +0100481 <li>NHWC
Teresa Charlin62687422021-04-28 10:58:49 +0100482 </ul>
483 <td>
484 <table>
485 <tr><th>src<th>dst
486 <tr><td>All<td>All
487 </table>
488<tr>
489 <td>CLChannelShuffleLayer
490 <td>
491 <ul>
492 <li>NCHW
Michele Di Giorgiob8025b32021-09-03 10:29:49 +0100493 <li>NHWC
Teresa Charlin62687422021-04-28 10:58:49 +0100494 </ul>
495 <td>
496 <table>
497 <tr><th>src<th>dst
498 <tr><td>All<td>All
499 </table>
500<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100501 <td rowspan="1">Comparison
502 <td rowspan="1" style="width:200px;"> Function to compare 2 tensors.
503 <td rowspan="1">
504 <ul>
505 <li>ANEURALNETWORKS_EQUAL
506 <li>ANEURALNETWORKS_GREATER
507 <li>ANEURALNETWORKS_GREATER_EQUAL
508 <li>ANEURALNETWORKS_LESS
509 <li>ANEURALNETWORKS_LESS_EQUAL
510 <li>ANEURALNETWORKS_NOT_EQUAL
511 </ul>
512 <td>CLComparison
513 <td>
514 <ul>
515 <li>All
516 </ul>
517 <td>
518 <table>
519 <tr><th>src0<th>src1<th>dst
520 <tr><td>All<td>All<td>U8
521 </table>
522<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100523 <td rowspan="2">ConcatenateLayer
524 <td rowspan="2" style="width:200px;"> Function to concatenate tensors along a given axis.
525 <td rowspan="2">
526 <ul>
527 <li>ANEURALNETWORKS_CONCATENATION
528 </ul>
529 <td>NEConcatenateLayer
530 <td>
531 <ul>
532 <li>All
533 </ul>
534 <td>
535 <table>
536 <tr><th>src<th>dst
537 <tr><td>QASYMM8<td>QASYMM8
538 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
539 <tr><td>F16<td>F16
540 <tr><td>F32<td>F32
541 </table>
542<tr>
543 <td>CLConcatenateLayer
544 <td>
545 <ul>
546 <li>All
547 </ul>
548 <td>
549 <table>
550 <tr><th>src<th>dst
551 <tr><td>QASYMM8<td>QASYMM8
552 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
553 <tr><td>F16<td>F16
554 <tr><td>F32<td>F32
555 </table>
556<tr>
557 <td rowspan="2">ConvertFullyConnectedWeights
Jakub Sujakee301b32021-06-04 09:46:08 +0100558 <td rowspan="2" style="width:200px;"> Function to transpose the weights for the fully connected layer.
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100559 <td rowspan="2">
560 <ul>
Teresa Charlin62687422021-04-28 10:58:49 +0100561 <li>n/a
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100562 </ul>
563 <td>NEConvertFullyConnectedWeights
564 <td>
565 <ul>
566 <li>NHWC
567 <li>NCHW
568 </ul>
569 <td>
570 <table>
571 <tr><th>src<th>dst
572 <tr><td>All<td>All
573 </table>
574<tr>
575 <td>CLConvertFullyConnectedWeights
576 <td>
577 <ul>
578 <li>NHWC
579 <li>NCHW
580 </ul>
581 <td>
582 <table>
583 <tr><th>src<th>dst
584 <tr><td>All<td>All
585 </table>
586<tr>
Teresa Charlin62687422021-04-28 10:58:49 +0100587 <td rowspan="2">ConvolutionLayer
588 <td rowspan="2" style="width:200px;"> Function to compute a convolution layer.
589 <td rowspan="2">
590 <ul>
591 <li>ANEURALNETWORKS_CONV_2D
592 </ul>
593 <td>NEConvolutionLayer
594 <td>
595 <ul>
596 <li>NHWC
597 <li>NCHW
598 </ul>
599 <td>
600 <table>
601 <tr><th>src0<th>src1<th>src2<th>dst
602 <tr><td>F16<td>F16<td>F16<td>F16
603 <tr><td>F32<td>F32<td>F32<td>F32
604 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
605 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
606 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
607 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
608 </table>
609<tr>
610 <td>CLConvolutionLayer
611 <td>
612 <ul>
613 <li>NHWC
614 <li>NCHW
615 </ul>
616 <td>
617 <table>
618 <tr><th>src0<th>src1<th>src2<th>dst
619 <tr><td>F16<td>F16<td>F16<td>F16
620 <tr><td>F32<td>F32<td>F32<td>F32
621 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
622 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
623 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
624 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
625 </table>
626<tr>
Sheri Zhang6d9c9822021-09-24 16:02:57 +0100627 <td rowspan="2">Conv3D
628 <td rowspan="2" style="width:200px;"> Function to compute a 3d convolution layer.
629 <td rowspan="2">
630 <ul>
631 <li>ANEURALNETWORKS_CONV_3D
632 </ul>
633 <td>NEConv3D
634 <td>
635 <ul>
636 <li>NDHWC
637 </ul>
638 <td>
639 <table>
640 <tr><th>src0<th>src1<th>src2<th>dst
641 <tr><td>F16<td>F16<td>F16<td>F16
642 <tr><td>F32<td>F32<td>F32<td>F32
Freddie Liardetf727ef42021-10-18 13:28:57 +0100643 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
644 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
Sheri Zhang6d9c9822021-09-24 16:02:57 +0100645 </table>
646<tr>
647 <td>CLConv3D
648 <td>
649 <ul>
650 <li>NDHWC
651 </ul>
652 <td>
653 <table>
654 <tr><th>src0<th>src1<th>src2<th>dst
655 <tr><td>F16<td>F16<td>F16<td>F16
656 <tr><td>F32<td>F32<td>F32<td>F32
Giorgio Arena51847d52021-10-19 15:45:57 +0100657 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
658 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
Sheri Zhang6d9c9822021-09-24 16:02:57 +0100659 </table>
660<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100661 <td rowspan="2">Copy
662 <td rowspan="2" style="width:200px;"> Function to copy a tensor.
663 <td rowspan="2">
664 <ul>
Teresa Charlin62687422021-04-28 10:58:49 +0100665 <li>n/a
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100666 </ul>
667 <td>NECopy
668 <td>
669 <ul>
670 <li>All
671 </ul>
672 <td>
673 <table>
674 <tr><th>src<th>dst
675 <tr><td>All<td>All
676 </table>
677<tr>
678 <td>CLCopy
679 <td>
680 <ul>
681 <li>All
682 </ul>
683 <td>
684 <table>
685 <tr><th>src<th>dst
686 <tr><td>All<td>All
687 </table>
688<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100689 <td rowspan="1">Crop
690 <td rowspan="1" style="width:200px;"> Performs a copy of input tensor to the output tensor.
691 <td rowspan="1">
692 <ul>
693 <li>n/a
694 </ul>
695 <td>CLCrop
696 <td>
697 <ul>
698 <li>NHWC
699 </ul>
700 <td>
701 <table>
702 <tr><th>src<th>dst
703 <tr><td>All<td>F32
704 </table>
705<tr>
Teresa Charlin62687422021-04-28 10:58:49 +0100706 <td rowspan="2">CropResize
707 <td rowspan="2" style="width:200px;"> Function to perform cropping and resizing.
708 <td rowspan="2">
709 <ul>
710 <li>n/a
711 </ul>
712 <td>NECropResize
713 <td>
714 <ul>
715 <li>NHWC
716 </ul>
717 <td>
718 <table>
719 <tr><th>src0<th>src1<th>src2<th>dst
720 <tr><td>All<td>F32<td>F32<td>F32
721 </table>
722<tr>
723 <td>CLCropResize
724 <td>
725 <ul>
726 <li>NHWC
727 </ul>
728 <td>
729 <table>
730 <tr><th>src0<th>src1<th>src2<th>dst
731 <tr><td>All<td>F32<td>F32<td>F32
732 </table>
733<tr>
734 <td rowspan="2">DeconvolutionLayer
Jakub Sujakee301b32021-06-04 09:46:08 +0100735 <td rowspan="2" style="width:200px;"> Function to compute a deconvolution or transpose convolution.
Teresa Charlin62687422021-04-28 10:58:49 +0100736 <td rowspan="2">
737 <ul>
738 <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
739 </ul>
740 <td>NEDeconvolutionLayer
741 <td>
742 <ul>
743 <li>NHWC
744 <li>NCHW
745 </ul>
746 <td>
747 <table>
748 <tr><th>src0<th>src1<th>src2<th>dst
749 <tr><td>F16<td>F16<td>F16<td>F16
750 <tr><td>F32<td>F32<td>F32<td>F32
751 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
752 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
753 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
754 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
755 </table>
756<tr>
757 <td>CLDeconvolutionLayer
758 <td>
759 <ul>
760 <li>NHWC
761 <li>NCHW
762 </ul>
763 <td>
764 <table>
765 <tr><th>src0<th>src1<th>src2<th>dst
766 <tr><td>F16<td>F16<td>F16<td>F16
767 <tr><td>F32<td>F32<td>F32<td>F32
768 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
769 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
770 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
771 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
772 </table>
773<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100774 <td rowspan="1">DeconvolutionLayerUpsample
775 <td rowspan="1" style="width:200px;"> Function to execute deconvolution upsample on OpenCL.
776 <td rowspan="1">
777 <ul>
778 <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
779 </ul>
780 <td>CLDeconvolutionLayerUpsample
781 <td>
782 <ul>
783 <li>NHWC
784 <li>NCHW
785 </ul>
786 <td>
787 <table>
788 <tr><th>src<th>dst
789 <tr><td>All<td>All
790 </table>
791<tr>
Teresa Charlin62687422021-04-28 10:58:49 +0100792 <td rowspan="2">DepthConvertLayer
793 <td rowspan="2" style="width:200px;"> Performs a down-scaling depth conversion.
794 <td rowspan="2">
795 <ul>
796 <li>n/a
797 </ul>
798 <td>NEDepthConvertLayer
799 <td>
800 <ul>
801 <li>All
802 </ul>
803 <td>
804 <table>
805 <tr><th>src<th>dst
806 <tr><td>QASYMM8<td>F16, F32
807 <tr><td>U8<td>U16, S16, S32
808 <tr><td>U16<td>U8, U32
809 <tr><td>S16<td>U8, S32
810 <tr><td>BFLOAT16<td>F32
811 <tr><td>F16<td>QASYMM8, F32
812 <tr><td>F32<td>QASYMM8, F16, BFLOAT16
813 </table>
814<tr>
815 <td>CLDepthConvertLayer
816 <td>
817 <ul>
818 <li>All
819 </ul>
820 <td>
821 <table>
822 <tr><th>src<th>dst
823 <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32
824 <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32
825 <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32
826 <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32
827 <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32
828 <tr><td>F16<td>U8, S8, U16, S16, U32, F32
829 <tr><td>F32<td>U8, S8, U16, S16, U32, F16
830 </table>
831<tr>
832 <td rowspan="2">DepthToSpaceLayer
833 <td rowspan="2" style="width:200px;"> Depth to Space transformation.
834 <td rowspan="2">
835 <ul>
836 <li>ANEURALNETWORKS_DEPTH_TO_SPACE
837 </ul>
838 <td>NEDepthToSpaceLayer
839 <td>
840 <ul>
841 <li>NHWC
842 <li>NCHW
843 </ul>
844 <td>
845 <table>
846 <tr><th>src<th>dst
847 <tr><td>All<td>All
848 </table>
849<tr>
850 <td>CLDepthToSpaceLayer
851 <td>
852 <ul>
853 <li>NHWC
854 <li>NCHW
855 </ul>
856 <td>
857 <table>
858 <tr><th>src<th>dst
859 <tr><td>All<td>All
860 </table>
861<tr>
862 <td rowspan="2">DepthwiseConvolutionLayer
863 <td rowspan="2" style="width:200px;"> Function to perform depthwise separable convolution.
864 <td rowspan="2">
865 <ul>
866 <li>ANEURALNETWORKS_DEPTHWISE_CONV_2D
867 </ul>
868 <td>NEDepthwiseConvolutionLayer
869 <td>
870 <ul>
871 <li>NHWC
872 <li>NCHW
873 </ul>
874 <td>
875 <table>
876 <tr><th>src0<th>src1<th>src2<th>dst
877 <tr><td>F16<td>F16<td>F16<td>F16
878 <tr><td>F32<td>F32<td>F32<td>F32
879 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
880 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
881 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
882 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
883 </table>
884<tr>
885 <td>CLDepthwiseConvolutionLayer
886 <td>
887 <ul>
888 <li>NHWC
889 <li>NCHW
890 </ul>
891 <td>
892 <table>
893 <tr><th>src0<th>src1<th>src2<th>dst
894 <tr><td>F16<td>F16<td>F16<td>F16
895 <tr><td>F32<td>F32<td>F32<td>F32
896 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
897 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
898 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
899 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
900 </table>
901<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100902 <td rowspan="2">DequantizationLayer
Teresa Charlin62687422021-04-28 10:58:49 +0100903 <td rowspan="2" style="width:200px;"> Function to dequantize the values in a tensor.
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100904 <td rowspan="2">
905 <ul>
906 <li>ANEURALNETWORKS_DEQUANTIZE
907 </ul>
908 <td>NEDequantizationLayer
909 <td>
910 <ul>
911 <li>All
912 </ul>
913 <td>
914 <table>
915 <tr><th>src<th>dst
Teresa Charlin62687422021-04-28 10:58:49 +0100916 <tr><td>QASYMM8<td>F16, F32
917 <tr><td>QASYMM8_SIGNED<td>F16, F32
918 <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32
919 <tr><td>QSYMM8<td>F16, F32
920 <tr><td>QSYMM16<td>F16, F32
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100921 </table>
922<tr>
923 <td>CLDequantizationLayer
924 <td>
925 <ul>
926 <li>All
927 </ul>
928 <td>
929 <table>
930 <tr><th>src<th>dst
Teresa Charlin62687422021-04-28 10:58:49 +0100931 <tr><td>QASYMM8<td>F16, F32
932 <tr><td>QASYMM8_SIGNED<td>F16, F32
933 <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32
934 <tr><td>QSYMM8<td>F16, F32
935 <tr><td>QSYMM16<td>F16, F32
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100936 </table>
937<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100938 <td rowspan="1">DetectionPostProcessLayer
939 <td rowspan="1" style="width:200px;"> Function to generate the detection output based on center size encoded boxes, class prediction and anchors by doing non maximum suppression (NMS).
940 <td rowspan="1">
941 <ul>
942 <li>ANEURALNETWORKS_DETECTION_POSTPROCESSING
943 </ul>
944 <td>NEDetectionPostProcessLayer
945 <td>
946 <ul>
947 <li>All
948 </ul>
949 <td>
950 <table>
951 <tr><th>src0 - src2<th>dst0 - dst3
952 <tr><td>QASYMM8<td>F32
953 <tr><td>QASYMM8_SIGNED<td>F32
954 <tr><td>F32<td>F32
955 </table>
956<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100957 <td rowspan="2">DirectConvolutionLayer
Teresa Charlin62687422021-04-28 10:58:49 +0100958 <td rowspan="2" style="width:200px;"> Function to compute direct convolution.
Sheri Zhanga47dcc22021-04-22 14:41:12 +0100959 <td rowspan="2">
960 <ul>
961 <li>ANEURALNETWORKS_CONV_2D
962 </ul>
963 <td>NEDirectConvolutionLayer
964 <td>
965 <ul>
966 <li>NHWC
967 <li>NCHW
968 </ul>
969 <td>
970 <table>
971 <tr><th>src0<th>src1<th>src2<th>dst
972 <tr><td>F16<td>F16<td>F16<td>F16
973 <tr><td>F32<td>F32<td>F32<td>F32
974 </table>
975<tr>
976 <td>CLDirectConvolutionLayer
977 <td>
978 <ul>
979 <li>NHWC
980 <li>NCHW
981 </ul>
982 <td>
983 <table>
984 <tr><th>src0<th>src1<th>src2<th>dst
985 <tr><td>F16<td>F16<td>F16<td>F16
986 <tr><td>F32<td>F32<td>F32<td>F32
987 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
988 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
989 </table>
990<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +0100991 <td rowspan="1">DirectDeconvolutionLayer
992 <td rowspan="1" style="width:200px;"> Function to run the deconvolution layer.
993 <td rowspan="1">
994 <ul>
995 <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
996 </ul>
997 <td>CLDirectDeconvolutionLayer
998 <td>
999 <ul>
1000 <li>NHWC
1001 <li>NCHW
1002 </ul>
1003 <td>
1004 <table>
1005 <tr><th>src0<th>src1<th>src2<th>dst
1006 <tr><td>F16<td>F16<td>F16<td>F16
1007 <tr><td>F32<td>F32<td>F32<td>F32
1008 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1009 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1010 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
1011 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
1012 </table>
1013<tr>
Jakub Sujakee301b32021-06-04 09:46:08 +01001014 <td rowspan="13">ElementwiseOperations
Sheri Zhang6124ce62021-05-04 14:03:13 +01001015 <td rowspan="13" style="width:200px;"> Function to perform in Cpu: - Div - Max - Min - Pow - SquaredDiff - Comparisons (Equal, greater, greater_equal, less, less_equal, not_equal) Function to perform in CL: - Add - Sub - Div - Max - Min - Pow - SquaredDiff
1016 <td rowspan="13">
1017 <ul>
1018 <li>ANEURALNETWORKS_MAXIMUM
1019 <li>ANEURALNETWORKS_MINIMUM
1020 <li>ANEURALNETWORKS_POW
1021 <li>ANEURALNETWORKS_DIV
1022 <li>ANEURALNETWORKS_ADD
1023 <li>ANEURALNETWORKS_SUB
1024 <li>ANEURALNETWORKS_EQUAL
1025 <li>ANEURALNETWORKS_GREATER
1026 <li>ANEURALNETWORKS_GREATER_EQUAL
1027 <li>ANEURALNETWORKS_LESS
1028 <li>ANEURALNETWORKS_LESS_EQUAL
1029 <li>ANEURALNETWORKS_NOT_EQUAL
1030 </ul>
1031 <td>NEElementwiseMax
1032 <td>
1033 <ul>
1034 <li>All
1035 </ul>
1036 <td>
1037 <table>
1038 <tr><th>src0<th>src1<th>dst
1039 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1040 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1041 <tr><td>S32<td>S32<td>S32
1042 <tr><td>S16<td>S16<td>S16
1043 <tr><td>F16<td>F16<td>F16
1044 <tr><td>F32<td>F32<td>F32
1045 </table>
1046<tr>
1047 <td>NEElementwiseMin
1048 <td>
1049 <ul>
1050 <li>All
1051 </ul>
1052 <td>
1053 <table>
1054 <tr><th>src0<th>src1<th>dst
1055 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1056 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1057 <tr><td>S32<td>S32<td>S32
1058 <tr><td>S16<td>S16<td>S16
1059 <tr><td>F16<td>F16<td>F16
1060 <tr><td>F32<td>F32<td>F32
1061 </table>
1062<tr>
1063 <td>NEElementwiseSquaredDiff
1064 <td>
1065 <ul>
1066 <li>All
1067 </ul>
1068 <td>
1069 <table>
1070 <tr><th>src0<th>src1<th>dst
1071 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1072 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1073 <tr><td>S32<td>S32<td>S32
1074 <tr><td>S16<td>S16<td>S16
1075 <tr><td>F16<td>F16<td>F16
1076 <tr><td>F32<td>F32<td>F32
1077 </table>
1078<tr>
1079 <td>NEElementwiseDivision
1080 <td>
1081 <ul>
1082 <li>All
1083 </ul>
1084 <td>
1085 <table>
1086 <tr><th>src0<th>src1<th>dst
1087 <tr><td>F16<td>F16<td>F16
1088 <tr><td>F32<td>F32<td>F32
1089 </table>
1090<tr>
1091 <td>NEElementwisePower
1092 <td>
1093 <ul>
1094 <li>All
1095 </ul>
1096 <td>
1097 <table>
1098 <tr><th>src0<th>src1<th>dst
1099 <tr><td>F16<td>F16<td>F16
1100 <tr><td>F32<td>F32<td>F32
1101 </table>
1102<tr>
1103 <td>NEElementwiseComparison
1104 <td>
1105 <ul>
1106 <li>All
1107 </ul>
1108 <td>
1109 <table>
1110 <tr><th>src0<th>src1<th>dst
1111 <tr><td>QASYMM8<td>QASYMM8<td>U8
1112 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>U8
1113 <tr><td>S32<td>S32<td>U8
1114 <tr><td>U8<td>U8<td>U8
1115 <tr><td>S16<td>S16<td>U8
1116 <tr><td>F16<td>F16<td>U8
1117 <tr><td>F32<td>F32<td>U8
1118 </table>
1119<tr>
1120 <td>CLArithmeticAddition
1121 <td>
1122 <ul>
1123 <li>All
1124 </ul>
1125 <td>
1126 <table>
1127 <tr><th>src0<th>src1<th>dst
1128 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1129 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1130 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
1131 <tr><td>U8<td>U8<td>U8
1132 <tr><td>U8<td>U8<td>S16
1133 <tr><td>U8<td>S16<td>S16
1134 <tr><td>S16<td>U8<td>S16
1135 <tr><td>S16<td>S16<td>S16
1136 <tr><td>S32<td>S32<td>S32
1137 <tr><td>F16<td>F16<td>F16
1138 <tr><td>F32<td>F32<td>F32
1139 </table>
1140<tr>
1141 <td>CLArithmeticSubtraction
1142 <td>
1143 <ul>
1144 <li>All
1145 </ul>
1146 <td>
1147 <table>
1148 <tr><th>src0<th>src1<th>dst
1149 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1150 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1151 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
1152 <tr><td>U8<td>U8<td>U8
1153 <tr><td>U8<td>U8<td>S16
1154 <tr><td>U8<td>S16<td>S16
1155 <tr><td>S16<td>U8<td>S16
1156 <tr><td>S16<td>S16<td>S16
1157 <tr><td>S32<td>S32<td>S32
1158 <tr><td>F16<td>F16<td>F16
1159 <tr><td>F32<td>F32<td>F32
1160 </table>
1161<tr>
1162 <td>CLArithmeticDivision
1163 <td>
1164 <ul>
1165 <li>All
1166 </ul>
1167 <td>
1168 <table>
1169 <tr><th>src0<th>src1<th>dst
1170 <tr><td>F16<td>F16<td>F16
1171 <tr><td>F32<td>F32<td>F32
1172 </table>
1173<tr>
1174 <td>CLElementwiseMax
1175 <td>
1176 <ul>
1177 <li>All
1178 </ul>
1179 <td>
1180 <table>
1181 <tr><th>src0<th>src1<th>dst
1182 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1183 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1184 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
1185 <tr><td>U8<td>U8<td>U8
1186 <tr><td>S16<td>S16<td>S16
1187 <tr><td>S32<td>S32<td>S32
1188 <tr><td>U32<td>U32<td>U32
1189 <tr><td>F16<td>F16<td>F16
1190 <tr><td>F32<td>F32<td>F32
1191 </table>
1192<tr>
1193 <td>CLElementwiseMin
1194 <td>
1195 <ul>
1196 <li>All
1197 </ul>
1198 <td>
1199 <table>
1200 <tr><th>src0<th>src1<th>dst
1201 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1202 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1203 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
1204 <tr><td>U8<td>U8<td>U8
1205 <tr><td>S16<td>S16<td>S16
1206 <tr><td>S32<td>S32<td>S32
1207 <tr><td>U32<td>U32<td>U32
1208 <tr><td>F16<td>F16<td>F16
1209 <tr><td>F32<td>F32<td>F32
1210 </table>
1211<tr>
1212 <td>CLElementwiseSquaredDiff
1213 <td>
1214 <ul>
1215 <li>All
1216 </ul>
1217 <td>
1218 <table>
1219 <tr><th>src0<th>src1<th>dst
1220 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
1221 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
1222 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
1223 <tr><td>U8<td>U8<td>U8
1224 <tr><td>S16<td>S16<td>S16
1225 <tr><td>F16<td>F16<td>F16
1226 <tr><td>F32<td>F32<td>F32
1227 </table>
1228<tr>
1229 <td>CLElementwisePower
1230 <td>
1231 <ul>
1232 <li>All
1233 </ul>
1234 <td>
1235 <table>
1236 <tr><th>src0<th>src1<th>dst
1237 <tr><td>F16<td>F16<td>F16
1238 <tr><td>F32<td>F32<td>F32
1239 </table>
1240<tr>
1241 <td rowspan="8">ElementwiseUnaryLayer
1242 <td rowspan="8" style="width:200px;"> Function to perform: - Rsqrt - Exp - Neg - Log - Abs - Round - Sin
1243 <td rowspan="8">
1244 <ul>
1245 <li>ANEURALNETWORKS_ABS
1246 <li>ANEURALNETWORKS_EXP
1247 <li>ANEURALNETWORKS_LOG
1248 <li>ANEURALNETWORKS_NEG
1249 <li>ANEURALNETWORKS_RSQRT
1250 <li>ANEURALNETWORKS_SIN
1251 </ul>
1252 <td>NEElementwiseUnaryLayer
1253 <td>
1254 <ul>
1255 <li>All
1256 </ul>
1257 <td>
1258 <table>
1259 <tr><th>src<th>dst
1260 <tr><td>F16<td>F16
1261 <tr><td>F32<td>F32
1262 <tr><td>S32<td>S32
1263 </table>
1264<tr>
1265 <td>CLRsqrtLayer
1266 <td>
1267 <ul>
1268 <li>All
1269 </ul>
1270 <td>
1271 <table>
1272 <tr><th>src<th>dst
1273 <tr><td>F16<td>F16
1274 <tr><td>F32<td>F32
1275 </table>
1276<tr>
1277 <td>CLExpLayer
1278 <td>
1279 <ul>
1280 <li>All
1281 </ul>
1282 <td>
1283 <table>
1284 <tr><th>src<th>dst
1285 <tr><td>F16<td>F16
1286 <tr><td>F32<td>F32
1287 </table>
1288<tr>
1289 <td>CLNegLayer
1290 <td>
1291 <ul>
1292 <li>All
1293 </ul>
1294 <td>
1295 <table>
1296 <tr><th>src<th>dst
1297 <tr><td>F16<td>F16
1298 <tr><td>F32<td>F32
Jakub Sujakee301b32021-06-04 09:46:08 +01001299 <tr><td>S32<td>S32
Sheri Zhang6124ce62021-05-04 14:03:13 +01001300 </table>
1301<tr>
1302 <td>CLSinLayer
1303 <td>
1304 <ul>
1305 <li>All
1306 </ul>
1307 <td>
1308 <table>
1309 <tr><th>src<th>dst
1310 <tr><td>F16<td>F16
1311 <tr><td>F32<td>F32
1312 </table>
1313<tr>
1314 <td>CLLogLayer
1315 <td>
1316 <ul>
1317 <li>All
1318 </ul>
1319 <td>
1320 <table>
1321 <tr><th>src<th>dst
1322 <tr><td>F16<td>F16
1323 <tr><td>F32<td>F32
1324 </table>
1325<tr>
1326 <td>CLAbsLayer
1327 <td>
1328 <ul>
1329 <li>All
1330 </ul>
1331 <td>
1332 <table>
1333 <tr><th>src<th>dst
1334 <tr><td>F16<td>F16
1335 <tr><td>F32<td>F32
1336 </table>
1337<tr>
1338 <td>CLRoundLayer
1339 <td>
1340 <ul>
1341 <li>All
1342 </ul>
1343 <td>
1344 <table>
1345 <tr><th>src<th>dst
1346 <tr><td>F16<td>F16
1347 <tr><td>F32<td>F32
1348 </table>
1349<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001350 <td rowspan="2">FFT1D
Teresa Charlin62687422021-04-28 10:58:49 +01001351 <td rowspan="2" style="width:200px;"> Fast Fourier Transform 1D.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001352 <td rowspan="2">
1353 <ul>
Teresa Charlin62687422021-04-28 10:58:49 +01001354 <li>n/a
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001355 </ul>
1356 <td>NEFFT1D
1357 <td>
1358 <ul>
1359 <li>All
1360 </ul>
1361 <td>
1362 <table>
1363 <tr><th>src<th>dst
1364 <tr><td>F32<td>F32
1365 </table>
1366<tr>
1367 <td>CLFFT1D
1368 <td>
1369 <ul>
1370 <li>All
1371 </ul>
1372 <td>
1373 <table>
1374 <tr><th>src<th>dst
1375 <tr><td>F32<td>F32
1376 <tr><td>F16<td>F16
1377 </table>
1378<tr>
1379 <td rowspan="2">FFT2D
Teresa Charlin62687422021-04-28 10:58:49 +01001380 <td rowspan="2" style="width:200px;"> Fast Fourier Transform 2D.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001381 <td rowspan="2">
1382 <ul>
Teresa Charlin62687422021-04-28 10:58:49 +01001383 <li>n/a
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001384 </ul>
1385 <td>NEFFT2D
1386 <td>
1387 <ul>
1388 <li>All
1389 </ul>
1390 <td>
1391 <table>
1392 <tr><th>src<th>dst
1393 <tr><td>F32<td>F32
1394 </table>
1395<tr>
1396 <td>CLFFT2D
1397 <td>
1398 <ul>
1399 <li>All
1400 </ul>
1401 <td>
1402 <table>
1403 <tr><th>src<th>dst
1404 <tr><td>F32<td>F32
1405 <tr><td>F16<td>F16
1406 </table>
1407<tr>
1408 <td rowspan="2">FFTConvolutionLayer
Teresa Charlin62687422021-04-28 10:58:49 +01001409 <td rowspan="2" style="width:200px;"> Fast Fourier Transform Convolution.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001410 <td rowspan="2">
1411 <ul>
1412 <li>ANEURALNETWORKS_CONV_2D
1413 </ul>
1414 <td>NEFFTConvolutionLayer
1415 <td>
1416 <ul>
1417 <li>All
1418 </ul>
1419 <td>
1420 <table>
1421 <tr><th>src<th>dst
1422 <tr><td>F32<td>F32
1423 </table>
1424<tr>
1425 <td>CLFFTConvolutionLayer
1426 <td>
1427 <ul>
1428 <li>All
1429 </ul>
1430 <td>
1431 <table>
1432 <tr><th>src<th>dst
1433 <tr><td>F32<td>F32
1434 <tr><td>F16<td>F16
1435 </table>
1436<tr>
1437 <td rowspan="2">Fill
Teresa Charlin62687422021-04-28 10:58:49 +01001438 <td rowspan="2" style="width:200px;"> Set the values of a tensor with a given value.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001439 <td rowspan="2">
1440 <ul>
1441 <li>ANEURALNETWORKS_FILL
1442 </ul>
1443 <td>NEFill
1444 <td>
1445 <ul>
1446 <li>All
1447 </ul>
1448 <td>
1449 <table>
1450 <tr><th>src<th>dst
1451 <tr><td>All<td>All
1452 </table>
1453<tr>
1454 <td>CLFill
1455 <td>
1456 <ul>
1457 <li>All
1458 </ul>
1459 <td>
1460 <table>
1461 <tr><th>src<th>dst
1462 <tr><td>All<td>All
1463 </table>
1464<tr>
Georgios Pinitasb6af4822021-09-14 12:33:34 +01001465 <td rowspan="1">FillBorder
1466 <td rowspan="1" style="width:200px;"> Function to fill the borders within the XY-planes.
1467 <td rowspan="1">
Teresa Charlin62687422021-04-28 10:58:49 +01001468 <ul>
1469 <li>n/a
1470 </ul>
1471 <td>NEFillBorder
1472 <td>
1473 <ul>
1474 <li>All
1475 </ul>
1476 <td>
1477 <table>
1478 <tr><th>src<th>dst
1479 <tr><td>All<td>All
1480 </table>
1481<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01001482 <td rowspan="2">FlattenLayer
1483 <td rowspan="2" style="width:200px;"> Reshape a tensor to be 1D
1484 <td rowspan="2">
1485 <ul>
1486 <li>ANEURALNETWORKS_RESHAPE
1487 </ul>
1488 <td>NEFlattenLayer
1489 <td>
1490 <ul>
1491 <li>All
1492 </ul>
1493 <td>
1494 <table>
1495 <tr><th>src<th>dst
1496 <tr><td>All<td>All
1497 </table>
1498<tr>
1499 <td>CLFlattenLayer
1500 <td>
1501 <ul>
1502 <li>All
1503 </ul>
1504 <td>
1505 <table>
1506 <tr><th>src<th>dst
1507 <tr><td>All<td>All
1508 </table>
1509<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001510 <td rowspan="2">Floor
Teresa Charlin62687422021-04-28 10:58:49 +01001511 <td rowspan="2" style="width:200px;"> Round the value to the lowest number.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01001512 <td rowspan="2">
1513 <ul>
1514 <li>ANEURALNETWORKS_FLOOR
1515 </ul>
1516 <td>NEFloor
1517 <td>
1518 <ul>
1519 <li>All
1520 </ul>
1521 <td>
1522 <table>
1523 <tr><th>src<th>dst
1524 <tr><td>F32<td>F32
1525 <tr><td>F16<td>F16
1526 </table>
1527<tr>
1528 <td>CLFloor
1529 <td>
1530 <ul>
1531 <li>All
1532 </ul>
1533 <td>
1534 <table>
1535 <tr><th>src<th>dst
1536 <tr><td>F32<td>F32
1537 <tr><td>F16<td>F16
1538 </table>
1539<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01001540 <td rowspan="2">FullyConnectedLayer
1541 <td rowspan="2" style="width:200px;"> Function to perform a fully connected / dense layer.
1542 <td rowspan="2">
1543 <ul>
1544 <li>ANEURALNETWORKS_FULLY_CONNECTED
1545 </ul>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001546 <td>NEFullyConnectedLayer
Teresa Charlin62687422021-04-28 10:58:49 +01001547 <td>
1548 <ul>
1549 <li>NHWC
1550 <li>NCHW
1551 </ul>
1552 <td>
1553 <table>
1554 <tr><th>src0<th>src1<th>src2<th>dst
1555 <tr><td>F16<td>F16<td>F16<td>F16
1556 <tr><td>F32<td>F32<td>F32<td>F32
1557 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1558 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1559 </table>
1560<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001561 <td>CLFullyConnectedLayer
Teresa Charlin62687422021-04-28 10:58:49 +01001562 <td>
1563 <ul>
1564 <li>NHWC
1565 <li>NCHW
1566 </ul>
1567 <td>
1568 <table>
1569 <tr><th>src0<th>src1<th>src2<th>dst
1570 <tr><td>F16<td>F16<td>F16<td>F16
1571 <tr><td>F32<td>F32<td>F32<td>F32
1572 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1573 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1574 </table>
1575<tr>
1576 <td rowspan="2">FuseBatchNormalization
1577 <td rowspan="2" style="width:200px;"> Function to fuse the batch normalization node to a preceding convolution node.
1578 <td rowspan="2">
1579 <ul>
1580 <li>n/a
1581 </ul>
1582 <td>NEFuseBatchNormalization
1583 <td>
1584 <ul>
1585 <li>NHWC
1586 <li>NCHW
1587 </ul>
1588 <td>
1589 <table>
1590 <tr><th>src<th>dst
1591 <tr><td>F32<td>F32
1592 <tr><td>F16<td>F16
1593 </table>
1594<tr>
1595 <td>CLFuseBatchNormalization
1596 <td>
1597 <ul>
1598 <li>NHWC
1599 <li>NCHW
1600 </ul>
1601 <td>
1602 <table>
1603 <tr><th>src<th>dst
1604 <tr><td>F32<td>F32
1605 <tr><td>F16<td>F16
1606 </table>
1607<tr>
1608 <td rowspan="2">Gather
1609 <td rowspan="2" style="width:200px;"> Performs the Gather operation along the chosen axis.
1610 <td rowspan="2">
1611 <ul>
1612 <li>ANEURALNETWORKS_GATHER
1613 </ul>
1614 <td>NEGather
1615 <td>
1616 <ul>
1617 <li>All
1618 </ul>
1619 <td>
1620 <table>
1621 <tr><th>src<th>dst
1622 <tr><td>All<td>All
1623 </table>
1624<tr>
1625 <td>CLGather
1626 <td>
1627 <ul>
1628 <li>All
1629 </ul>
1630 <td>
1631 <table>
1632 <tr><th>src<th>dst
1633 <tr><td>All<td>All
1634 </table>
1635<tr>
1636 <td rowspan="2">GEMM
1637 <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
1638 <td rowspan="2">
1639 <ul>
1640 <li>n/a
1641 </ul>
1642 <td>NEGEMM
1643 <td>
1644 <ul>
1645 <li>All
1646 </ul>
1647 <td>
1648 <table>
1649 <tr><th>src0<th>src1<th>src2<th>dst
1650 <tr><td>F32<td>F32<td>F32<td>F32
1651 <tr><td>F16<td>F16<td>F16<td>F16
1652 <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
1653 </table>
1654<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001655 <td>CLGEMM
Teresa Charlin62687422021-04-28 10:58:49 +01001656 <td>
1657 <ul>
1658 <li>All
1659 </ul>
1660 <td>
1661 <table>
1662 <tr><th>src0<th>src1<th>src2<th>dst
1663 <tr><td>F32<td>F32<td>F32<td>F32
1664 <tr><td>F16<td>F16<td>F16<td>F16
1665 </table>
1666<tr>
Jakub Sujakee301b32021-06-04 09:46:08 +01001667 <td rowspan="1">GEMMConv2d
Sheri Zhang6124ce62021-05-04 14:03:13 +01001668 <td rowspan="1" style="width:200px;"> General Matrix Multiplication.
1669 <td rowspan="1">
1670 <ul>
1671 <li>ANEURALNETWORKS_CONV_2D
1672 </ul>
1673 <td>NEGEMMConv2d
1674 <td>
1675 <ul>
1676 <li>All
1677 </ul>
1678 <td>
1679 <table>
1680 <tr><th>src0<th>src1<th>src2<th>dst
1681 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1682 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1683 <tr><td>F16<td>F16<td>F16<td>F16
1684 <tr><td>F32<td>F32<td>F32<td>F32
1685 <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
1686 </table>
1687<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01001688 <td rowspan="2">GEMMConvolutionLayer
1689 <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
1690 <td rowspan="2">
1691 <ul>
1692 <li>ANEURALNETWORKS_CONV_2D
1693 </ul>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001694 <td>NEGEMMConvolutionLayer
Teresa Charlin62687422021-04-28 10:58:49 +01001695 <td>
1696 <ul>
1697 <li>NHWC
1698 <li>NCHW
1699 </ul>
1700 <td>
1701 <table>
1702 <tr><th>src0<th>src1<th>src2<th>dst
1703 <tr><td>F16<td>F16<td>F16<td>F16
1704 <tr><td>F32<td>F32<td>F32<td>F32
1705 <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
1706 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1707 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
1708 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1709 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
1710 </table>
1711<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001712 <td>CLGEMMConvolutionLayer
Teresa Charlin62687422021-04-28 10:58:49 +01001713 <td>
1714 <ul>
1715 <li>NHWC
1716 <li>NCHW
1717 </ul>
1718 <td>
1719 <table>
1720 <tr><th>src0<th>src1<th>src2<th>dst
1721 <tr><td>F16<td>F16<td>F16<td>F16
1722 <tr><td>F32<td>F32<td>F32<td>F32
1723 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1724 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
1725 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1726 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
1727 </table>
1728<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001729 <td rowspan="1">GEMMDeconvolutionLayer
1730 <td rowspan="1" style="width:200px;"> General Matrix Multiplication.
1731 <td rowspan="1">
1732 <ul>
1733 <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
1734 </ul>
1735 <td>CLGEMMDeconvolutionLayer
1736 <td>
1737 <ul>
1738 <li>NHWC
1739 </ul>
1740 <td>
1741 <table>
1742 <tr><th>src0<th>src1<th>src2<th>dst
1743 <tr><td>F16<td>F16<td>F16<td>F16
1744 <tr><td>F32<td>F32<td>F32<td>F32
1745 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1746 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1747 </table>
1748<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01001749 <td rowspan="2">GEMMLowpMatrixMultiplyCore
1750 <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
1751 <td rowspan="2">
1752 <ul>
1753 <li>n/a
1754 </ul>
1755 <td>NEGEMMLowpMatrixMultiplyCore
1756 <td>
1757 <ul>
1758 <li>NHWC
1759 <li>NCHW
1760 </ul>
1761 <td>
1762 <table>
1763 <tr><th>src0<th>src1<th>src2<th>dst
1764 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1765 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
1766 <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8
1767 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32
1768 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
1769 <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32
1770 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1771 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
1772 <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED
1773 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32
1774 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
1775 <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32
1776 </table>
1777<tr>
1778 <td>CLGEMMLowpMatrixMultiplyCore
1779 <td>
1780 <ul>
1781 <li>NHWC
1782 <li>NCHW
1783 </ul>
1784 <td>
1785 <table>
1786 <tr><th>src0<th>src1<th>src2<th>dst
1787 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
1788 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
1789 <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8
1790 <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32
1791 <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
1792 <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32
1793 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
1794 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
1795 <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED
1796 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32
1797 <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
1798 <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32
1799 </table>
1800<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001801 <td rowspan="2">GEMMLowpOutputStage
1802 <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
1803 <td rowspan="2">
1804 <ul>
1805 <li>n/a
1806 </ul>
1807 <td>NEGEMMLowpOutputStage
1808 <td>
1809 <ul>
1810 <li>All
1811 </ul>
1812 <td>
1813 <table>
1814 <tr><th>src0<th>src1<th>dst
1815 <tr><td>S32<td>S32<td>QASYMM8
1816 <tr><td>S32<td>S32<td>QASYMM8_SIGNED
1817 <tr><td>S32<td>S32<td>QSYMM16
1818 </table>
1819<tr>
1820 <td>CLGEMMLowpOutputStage
1821 <td>
1822 <ul>
1823 <li>All
1824 </ul>
1825 <td>
1826 <table>
1827 <tr><th>src0<th>src1<th>dst
1828 <tr><td>S32<td>S32<td>QASYMM8
1829 <tr><td>S32<td>S32<td>QASYMM8_SIGNED
1830 <tr><td>S32<td>S32<td>QSYMM16
1831 </table>
1832<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01001833 <td rowspan="2">GenerateProposalsLayer
1834 <td rowspan="2" style="width:200px;"> Function to generate proposals for a RPN (Region Proposal Network).
1835 <td rowspan="2">
1836 <ul>
1837 <li>ANEURALNETWORKS_GENERATE_PROPOSALS
1838 </ul>
1839 <td>NEGenerateProposalsLayer
1840 <td>
1841 <ul>
1842 <li>All
1843 </ul>
1844 <td>
1845 <table>
1846 <tr><th>src0<th>src1<th>src2<th>dst
1847 <tr><td>F16<td>F16<td>F16<td>F16
1848 <tr><td>F32<td>F32<td>F32<td>F32
1849 <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8
1850 </table>
1851<tr>
1852 <td>CLGenerateProposalsLayer
1853 <td>
1854 <ul>
1855 <li>All
1856 </ul>
1857 <td>
1858 <table>
1859 <tr><th>src0<th>src1<th>src2<th>dst
1860 <tr><td>F16<td>F16<td>F16<td>F16
1861 <tr><td>F32<td>F32<td>F32<td>F32
1862 <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8
1863 </table>
1864<tr>
1865 <td rowspan="2">InstanceNormalizationLayer
1866 <td rowspan="2" style="width:200px;"> Function to perform a Instance normalization on a given axis.
1867 <td rowspan="2">
1868 <ul>
1869 <li>ANEURALNETWORKS_INSTANCE_NORMALIZATION
1870 </ul>
1871 <td>NEInstanceNormalizationLayer
1872 <td>
1873 <ul>
1874 <li>NHWC
1875 <li>NCHW
1876 </ul>
1877 <td>
1878 <table>
1879 <tr><th>src<th>dst
1880 <tr><td>F16<td>F16
1881 <tr><td>F32<td>F32
1882 </table>
1883<tr>
1884 <td>CLInstanceNormalizationLayer
1885 <td>
1886 <ul>
1887 <li>NHWC
1888 <li>NCHW
1889 </ul>
1890 <td>
1891 <table>
1892 <tr><th>src<th>dst
1893 <tr><td>F16<td>F16
1894 <tr><td>F32<td>F32
1895 </table>
1896<tr>
1897 <td rowspan="2">L2NormalizeLayer
1898 <td rowspan="2" style="width:200px;"> Function to perform a L2 normalization on a given axis.
1899 <td rowspan="2">
1900 <ul>
1901 <li>ANEURALNETWORKS_L2_NORMALIZATION
1902 </ul>
1903 <td>NEL2NormalizeLayer
1904 <td>
1905 <ul>
1906 <li>NHWC
1907 <li>NCHW
1908 </ul>
1909 <td>
1910 <table>
1911 <tr><th>src<th>dst
1912 <tr><td>F16<td>F16
1913 <tr><td>F32<td>F32
1914 </table>
1915<tr>
1916 <td>CLL2NormalizeLayer
1917 <td>
1918 <ul>
1919 <li>NHWC
1920 <li>NCHW
1921 </ul>
1922 <td>
1923 <table>
1924 <tr><th>src<th>dst
1925 <tr><td>F16<td>F16
1926 <tr><td>F32<td>F32
1927 </table>
1928<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01001929 <td rowspan="3">Logical
1930 <td rowspan="3" style="width:200px;"> Function to perform: - Logical AND - Logical OR - Logical NOT
1931 <td rowspan="3">
1932 <ul>
1933 <li>n/a
1934 </ul>
1935 <td>NELogicalAnd
1936 <td>
1937 <ul>
1938 <li>All
1939 </ul>
1940 <td>
1941 <table>
1942 <tr><th>src0<th>src1<th>dst
1943 <tr><td>U8<td>U8<td>U8
1944 </table>
1945<tr>
1946 <td>NELogicalOr
1947 <td>
1948 <ul>
1949 <li>All
1950 </ul>
1951 <td>
1952 <table>
1953 <tr><th>src0<th>src1<th>dst
1954 <tr><td>U8<td>U8<td>U8
1955 </table>
1956<tr>
1957 <td>NELogicalNot
1958 <td>
1959 <ul>
1960 <li>All
1961 </ul>
1962 <td>
1963 <table>
1964 <tr><th>src<th>dst
1965 <tr><td>U8<td>U8
1966 </table>
1967<tr>
1968 <td rowspan="1">LogicalAnd
1969 <td rowspan="1" style="width:200px;"> Function to perform Logical AND.
1970 <td rowspan="1">
1971 <ul>
1972 <li>n/a
1973 </ul>
1974 <td>CLLogicalAnd
1975 <td>
1976 <ul>
1977 <li>All
1978 </ul>
1979 <td>
1980 <table>
1981 <tr><th>src0<th>src1<th>dst
1982 <tr><td>U8<td>U8<td>U8
1983 </table>
1984<tr>
1985 <td rowspan="1">LogicalOr
1986 <td rowspan="1" style="width:200px;"> Function to perform Logical OR.
1987 <td rowspan="1">
1988 <ul>
1989 <li>n/a
1990 </ul>
1991 <td>CLLogicalOr
1992 <td>
1993 <ul>
1994 <li>All
1995 </ul>
1996 <td>
1997 <table>
1998 <tr><th>src0<th>src1<th>dst
1999 <tr><td>U8<td>U8<td>U8
2000 </table>
2001<tr>
2002 <td rowspan="1">LogicalNot
2003 <td rowspan="1" style="width:200px;"> Function to perform Logical NOT.
2004 <td rowspan="1">
2005 <ul>
2006 <li>n/a
2007 </ul>
2008 <td>CLLogicalNot
2009 <td>
2010 <ul>
2011 <li>All
2012 </ul>
2013 <td>
2014 <table>
2015 <tr><th>src<th>dst
2016 <tr><td>U8<td>U8
2017 </table>
2018<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002019 <td rowspan="2">LSTMLayer
2020 <td rowspan="2" style="width:200px;"> Function to perform a single time step in a Long Short-Term Memory (LSTM) layer.
2021 <td rowspan="2">
2022 <ul>
2023 <li>ANEURALNETWORKS_LSTM
2024 </ul>
2025 <td>NELSTMLayer
2026 <td>
2027 <ul>
2028 <li>All
2029 </ul>
2030 <td>
2031 <table>
2032 <tr><th>src0 - src13<th>dst0 - dst3
2033 <tr><td>F16<td>F16
2034 <tr><td>F32<td>F32
2035 </table>
2036<tr>
2037 <td>CLLSTMLayer
2038 <td>
2039 <ul>
2040 <li>All
2041 </ul>
2042 <td>
2043 <table>
2044 <tr><th>src0 - src13<th>dst0 - dst3
2045 <tr><td>F16<td>F16
2046 <tr><td>F32<td>F32
2047 </table>
2048<tr>
2049 <td rowspan="2">LSTMLayerQuantized
2050 <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory)
2051 <td rowspan="2">
2052 <ul>
2053 <li>ANEURALNETWORKS_QUANTIZED_LSTM
2054 <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
2055 </ul>
2056 <td>NELSTMLayerQuantized
2057 <td>
2058 <ul>
2059 <li>All
2060 </ul>
2061 <td>
2062 <table>
2063 <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1
2064 <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8
2065 </table>
2066<tr>
2067 <td>CLLSTMLayerQuantized
2068 <td>
2069 <ul>
2070 <li>All
2071 </ul>
2072 <td>
2073 <table>
2074 <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1
2075 <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8
2076 </table>
2077<tr>
Jakub Sujak667e82f2023-11-07 22:39:30 +00002078 <td rowspan="2">MatMul
2079 <td rowspan="2" style="width:200px;"> Computes a matrix multiplication in batches.
2080 <td rowspan="2">
2081 <ul>
2082 <li>ANEURALNETWORKS_BATCH_MATMUL
2083 </ul>
2084 <td>NEMatMul
2085 <td>
2086 <ul>
2087 <li>Any
2088 </ul>
2089 <td>
2090 <table>
2091 <tr><th>lhs<th>rhs<th>dst
2092 <tr><td>F32<td>F32<td>F32
2093 <tr><td>F16<td>F16<td>F16
Renato Arantes36a75da2024-01-26 17:31:18 +00002094 <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
Jakub Sujak667e82f2023-11-07 22:39:30 +00002095 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2096 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
2097 </table>
2098<tr>
2099 <td>CLMatMul
2100 <td>
2101 <ul>
2102 <li>All
2103 </ul>
2104 <td>
2105 <table>
2106 <tr><th>lhs<th>rhs<th>dst
2107 <tr><td>F32<td>F32<td>F32
2108 <tr><td>F16<td>F16<td>F16
2109 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2110 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
2111 </table>
2112<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002113 <td rowspan="2">MaxUnpoolingLayer
2114 <td rowspan="2" style="width:200px;"> Function to perform MaxUnpooling.
2115 <td rowspan="2">
2116 <ul>
2117 <li>n/a
2118 </ul>
2119 <td>NEMaxUnpoolingLayer
2120 <td>
2121 <ul>
2122 <li>NHWC
2123 <li>NCHW
2124 </ul>
2125 <td>
2126 <table>
2127 <tr><th>src<th>dst
2128 <tr><td>QASYMM8<td>QASYMM8
2129 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2130 <tr><td>F16<td>F16
2131 <tr><td>F32<td>F32
2132 </table>
2133<tr>
2134 <td>CLMaxUnpoolingLayer
2135 <td>
2136 <ul>
2137 <li>NHWC
2138 <li>NCHW
2139 </ul>
2140 <td>
2141 <table>
2142 <tr><th>src<th>dst
2143 <tr><td>QASYMM8<td>QASYMM8
2144 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2145 <tr><td>F16<td>F16
2146 <tr><td>F32<td>F32
2147 </table>
2148<tr>
2149 <td rowspan="2">MeanStdDevNormalizationLayer
2150 <td rowspan="2" style="width:200px;"> Function to execute mean and standard deviation normalization.
2151 <td rowspan="2">
2152 <ul>
2153 <li>n/a
2154 </ul>
2155 <td>NEMeanStdDevNormalizationLayer
2156 <td>
2157 <ul>
2158 <li>NHWC
2159 <li>NCHW
2160 </ul>
2161 <td>
2162 <table>
2163 <tr><th>src<th>dst
2164 <tr><td>F32<td>F32
2165 <tr><td>F16<td>F16
2166 </table>
2167<tr>
2168 <td>CLMeanStdDevNormalizationLayer
2169 <td>
2170 <ul>
2171 <li>NHWC
2172 <li>NCHW
2173 </ul>
2174 <td>
2175 <table>
2176 <tr><th>src<th>dst
2177 <tr><td>F32<td>F32
2178 <tr><td>F16<td>F16
2179 </table>
2180<tr>
2181 <td rowspan="2">NormalizationLayer
2182 <td rowspan="2" style="width:200px;"> Function to compute normalization layer.
2183 <td rowspan="2">
2184 <ul>
2185 <li>ANEURALNETWORKS_LOCAL_RESPONSE_NORMALIZATION
2186 </ul>
2187 <td>NENormalizationLayer
2188 <td>
2189 <ul>
2190 <li>NHWC
2191 <li>NCHW
2192 </ul>
2193 <td>
2194 <table>
2195 <tr><th>src<th>dst
2196 <tr><td>F32<td>F32
2197 <tr><td>F16<td>F16
2198 </table>
2199<tr>
2200 <td>CLNormalizationLayer
2201 <td>
2202 <ul>
2203 <li>NHWC
2204 <li>NCHW
2205 </ul>
2206 <td>
2207 <table>
2208 <tr><th>src<th>dst
2209 <tr><td>F32<td>F32
2210 <tr><td>F16<td>F16
2211 </table>
2212<tr>
Jakub Sujak667e82f2023-11-07 22:39:30 +00002213 <td rowspan="1">NormalizePlanarYUVLayer
2214 <td rowspan="1" style="width:200px;"> Function to compute normalization planar YUV layer.
2215 <td rowspan="1">
2216 <ul>
2217 <li>n/a
2218 </ul>
2219 <td>CLNormalizePlanarYUVLayer
2220 <td>
2221 <ul>
2222 <li>NHWC
2223 <li>NCHW
2224 </ul>
2225 <td>
2226 <table>
2227 <tr><th>src<th>dst
2228 <tr><td>F32<td>F32
2229 <tr><td>F16<td>F16
2230 <tr><td>QASYMM8<td>QASYMM8
2231 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2232 </table>
2233<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002234 <td rowspan="2">PadLayer
2235 <td rowspan="2" style="width:200px;"> Function to pad a tensor.
2236 <td rowspan="2">
2237 <ul>
2238 <li>ANEURALNETWORKS_PAD
2239 <li>ANEURALNETWORKS_PAD_V2
2240 </ul>
2241 <td>NEPadLayer
2242 <td>
2243 <ul>
2244 <li>NHWC
2245 <li>NCHW
2246 </ul>
2247 <td>
2248 <table>
2249 <tr><th>src<th>dst
2250 <tr><td>All<td>All
2251 </table>
2252<tr>
2253 <td>CLPadLayer
2254 <td>
2255 <ul>
2256 <li>NHWC
2257 <li>NCHW
2258 </ul>
2259 <td>
2260 <table>
2261 <tr><th>src<th>dst
2262 <tr><td>All<td>All
2263 </table>
2264<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002265 <td rowspan="2">Permute
2266 <td rowspan="2" style="width:200px;"> Function to transpose an ND tensor.
2267 <td rowspan="2">
2268 <ul>
2269 <li>ANEURALNETWORKS_TRANSPOSE
2270 </ul>
2271 <td>NEPermute
2272 <td>
2273 <ul>
2274 <li>NHWC
2275 <li>NCHW
2276 </ul>
2277 <td>
2278 <table>
2279 <tr><th>src<th>dst
2280 <tr><td>All<td>All
2281 </table>
2282<tr>
2283 <td>CLPermute
2284 <td>
2285 <ul>
2286 <li>NHWC
2287 <li>NCHW
2288 </ul>
2289 <td>
2290 <table>
2291 <tr><th>src<th>dst
2292 <tr><td>All<td>All
2293 </table>
2294<tr>
2295 <td rowspan="2">PixelWiseMultiplication
Jakub Sujakee301b32021-06-04 09:46:08 +01002296 <td rowspan="2" style="width:200px;"> Function to perform a multiplication.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002297 <td rowspan="2">
2298 <ul>
2299 <li>ANEURALNETWORKS_MUL
2300 </ul>
2301 <td>NEPixelWiseMultiplication
2302 <td>
2303 <ul>
2304 <li>All
2305 </ul>
2306 <td>
2307 <table>
2308 <tr><th>src0<th>src1<th>dst
2309 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
2310 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2311 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
2312 <tr><td>QSYMM16<td>QSYMM16<td>S32
2313 <tr><td>U8<td>U8<td>U8
2314 <tr><td>U8<td>U8<td>S16
2315 <tr><td>U8<td>S16<td>S16
2316 <tr><td>S16<td>U8<td>S16
2317 <tr><td>S16<td>S16<td>S16
2318 <tr><td>F16<td>F16<td>F16
2319 <tr><td>F32<td>S32<td>F32
2320 </table>
2321<tr>
2322 <td>CLPixelWiseMultiplication
2323 <td>
2324 <ul>
2325 <li>All
2326 </ul>
2327 <td>
2328 <table>
2329 <tr><th>src0<th>src1<th>dst
2330 <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
2331 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2332 <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
2333 <tr><td>QSYMM16<td>QSYMM16<td>S32
2334 <tr><td>U8<td>U8<td>U8
2335 <tr><td>U8<td>U8<td>S16
2336 <tr><td>U8<td>S16<td>S16
2337 <tr><td>S16<td>U8<td>S16
2338 <tr><td>S16<td>S16<td>S16
2339 <tr><td>F16<td>F16<td>F16
Jakub Sujakee301b32021-06-04 09:46:08 +01002340 <tr><td>F32<td>F32<td>F32
2341 <tr><td>S32<td>S32<td>S32
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002342 </table>
2343<tr>
2344 <td rowspan="2">PoolingLayer
Jakub Sujakee301b32021-06-04 09:46:08 +01002345 <td rowspan="2" style="width:200px;"> Function to perform pooling with the specified pooling operation.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002346 <td rowspan="2">
2347 <ul>
2348 <li>ANEURALNETWORKS_AVERAGE_POOL_2D
2349 <li>ANEURALNETWORKS_L2_POOL_2D
2350 <li>ANEURALNETWORKS_MAX_POOL_2D
2351 </ul>
2352 <td>NEPoolingLayer
2353 <td>
2354 <ul>
2355 <li>NHWC
2356 <li>NCHW
2357 </ul>
2358 <td>
2359 <table>
2360 <tr><th>src<th>dst
2361 <tr><td>QASYMM8<td>QASYMM8
2362 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2363 <tr><td>F16<td>F16
2364 <tr><td>F32<td>F32
2365 </table>
2366<tr>
2367 <td>CLPoolingLayer
2368 <td>
2369 <ul>
2370 <li>NHWC
2371 <li>NCHW
2372 </ul>
2373 <td>
2374 <table>
2375 <tr><th>src<th>dst
2376 <tr><td>QASYMM8<td>QASYMM8
2377 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2378 <tr><td>F16<td>F16
2379 <tr><td>F32<td>F32
2380 </table>
2381<tr>
Adnan AlSinan171fc3d2022-03-15 18:46:42 +00002382 <td rowspan="2">Pooling3dLayer
2383 <td rowspan="2" style="width:200px;"> Function to perform pooling 3D with the specified pooling operation.
2384 <td rowspan="2">
2385 <ul>
2386 <li>N/A
2387 </ul>
2388 <td>NEPooling3dLayer
2389 <td>
2390 <ul>
2391 <li>NDHWC
2392 </ul>
2393 <td>
2394 <table>
2395 <tr><th>src<th>dst
2396 <tr><td>F16<td>F16
2397 <tr><td>F32<td>F32
Adnan AlSinan9104cd52022-04-06 16:19:31 +01002398 <tr><td>QASYMM8<td>QASYMM8
2399 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
Adnan AlSinan171fc3d2022-03-15 18:46:42 +00002400 </table>
2401<tr>
2402 <td>CLPooling3dLayer
2403 <td>
2404 <ul>
2405 <li>NDHWC
2406 </ul>
2407 <td>
2408 <table>
2409 <tr><th>src<th>dst
2410 <tr><td>F16<td>F16
2411 <tr><td>F32<td>F32
Mohammed Suhail Munshi5e549fa2022-03-16 11:14:06 +00002412 <tr><td>QASYMM8<td>QASYMM8
2413 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
Adnan AlSinan171fc3d2022-03-15 18:46:42 +00002414 </table>
2415<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002416 <td rowspan="2">PReluLayer
2417 <td rowspan="2" style="width:200px;"> Function to compute the activation layer with the PRELU activation function.
2418 <td rowspan="2">
2419 <ul>
2420 <li>ANEURALNETWORKS_PRELU
2421 </ul>
2422 <td>NEPReluLayer
2423 <td>
2424 <ul>
2425 <li>All
2426 </ul>
2427 <td>
2428 <table>
2429 <tr><th>src<th>dst
2430 <tr><td>QASYMM8<td>QASYMM8
2431 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2432 <tr><td>F16<td>F16
2433 <tr><td>F32<td>F32
2434 </table>
2435<tr>
2436 <td>CLPReluLayer
2437 <td>
2438 <ul>
2439 <li>All
2440 </ul>
2441 <td>
2442 <table>
2443 <tr><th>src<th>dst
2444 <tr><td>QASYMM8<td>QASYMM8
2445 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2446 <tr><td>F16<td>F16
2447 <tr><td>F32<td>F32
2448 </table>
2449<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002450 <td rowspan="2">PriorBoxLayer
Sheri Zhang6124ce62021-05-04 14:03:13 +01002451 <td rowspan="2" style="width:200px;"> Function to compute prior boxes and clip.
Teresa Charlin62687422021-04-28 10:58:49 +01002452 <td rowspan="2">
2453 <ul>
2454 <li>n/a
2455 </ul>
2456 <td>NEPriorBoxLayer
2457 <td>
2458 <ul>
2459 <li>NHWC
2460 <li>NCHW
2461 </ul>
2462 <td>
2463 <table>
2464 <tr><th>src0<th>src1<th>dst
2465 <tr><td>F32<td>F32<td>F32
2466 </table>
2467<tr>
2468 <td>CLPriorBoxLayer
2469 <td>
2470 <ul>
2471 <li>NHWC
2472 <li>NCHW
2473 </ul>
2474 <td>
2475 <table>
2476 <tr><th>src0<th>src1<th>dst
2477 <tr><td>F32<td>F32<td>F32
2478 </table>
2479<tr>
2480 <td rowspan="2">QLSTMLayer
2481 <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory).
2482 <td rowspan="2">
2483 <ul>
2484 <li>ANEURALNETWORKS_QUANTIZED_LSTM
2485 <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
2486 </ul>
2487 <td>NEQLSTMLayer
2488 <td>
2489 <ul>
2490 <li>All
2491 </ul>
2492 <td>
2493 <table>
2494 <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2
2495 <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED
2496 </table>
2497<tr>
2498 <td>CLQLSTMLayer
2499 <td>
2500 <ul>
2501 <li>All
2502 </ul>
2503 <td>
2504 <table>
2505 <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2
2506 <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED
2507 </table>
2508<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002509 <td rowspan="2">QuantizationLayer
2510 <td rowspan="2" style="width:200px;"> Function to perform quantization layer
2511 <td rowspan="2">
2512 <ul>
2513 <li>ANEURALNETWORKS_QUANTIZE
2514 </ul>
2515 <td>NEQuantizationLayer
2516 <td>
2517 <ul>
2518 <li>All
2519 </ul>
2520 <td>
2521 <table>
2522 <tr><th>src<th>dst
Teresa Charlin62687422021-04-28 10:58:49 +01002523 <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2524 <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2525 <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2526 <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002527 </table>
2528<tr>
2529 <td>CLQuantizationLayer
2530 <td>
2531 <ul>
2532 <li>All
2533 </ul>
2534 <td>
2535 <table>
2536 <tr><th>src<th>dst
Teresa Charlin62687422021-04-28 10:58:49 +01002537 <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2538 <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2539 <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2540 <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
2541 </table>
2542<tr>
2543 <td rowspan="2">Range
2544 <td rowspan="2" style="width:200px;"> Function to generates a sequence of numbers starting from START and extends by increments of 'STEP' up to but not including 'END'.
2545 <td rowspan="2">
2546 <ul>
2547 <li>n/a
2548 </ul>
2549 <td>NERange
2550 <td>
2551 <ul>
2552 <li>All
2553 </ul>
2554 <td>
2555 <table>
2556 <tr><th>dst
2557 <tr><td>U8
2558 <tr><td>S8
2559 <tr><td>U16
2560 <tr><td>S16
2561 <tr><td>U32
2562 <tr><td>S32
2563 <tr><td>F16
2564 <tr><td>F32
2565 </table>
2566<tr>
2567 <td>CLRange
2568 <td>
2569 <ul>
2570 <li>All
2571 </ul>
2572 <td>
2573 <table>
2574 <tr><th>dst
2575 <tr><td>U8
2576 <tr><td>S8
2577 <tr><td>QASYMM8
2578 <tr><td>U16
2579 <tr><td>S16
2580 <tr><td>U32
2581 <tr><td>S32
2582 <tr><td>F16
2583 <tr><td>F32
2584 </table>
2585<tr>
2586 <td rowspan="2">ReduceMean
Jakub Sujakee301b32021-06-04 09:46:08 +01002587 <td rowspan="2" style="width:200px;"> Function to perform reduce mean operation.
Teresa Charlin62687422021-04-28 10:58:49 +01002588 <td rowspan="2">
2589 <ul>
2590 <li>ANEURALNETWORKS_MEAN
2591 </ul>
2592 <td>NEReduceMean
2593 <td>
2594 <ul>
2595 <li>All
2596 </ul>
2597 <td>
2598 <table>
2599 <tr><th>src<th>dst
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002600 <tr><td>QASYMM8<td>QASYMM8
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002601 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
Teresa Charlin62687422021-04-28 10:58:49 +01002602 <tr><td>F16<td>F16
2603 <tr><td>F32<td>F32
2604 </table>
2605<tr>
2606 <td>CLReduceMean
2607 <td>
2608 <ul>
2609 <li>All
2610 </ul>
2611 <td>
2612 <table>
2613 <tr><th>src<th>dst
2614 <tr><td>QASYMM8<td>QASYMM8
2615 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2616 <tr><td>F16<td>F16
2617 <tr><td>F32<td>F32
2618 </table>
2619<tr>
2620 <td rowspan="2">ReductionOperation
Jakub Sujakee301b32021-06-04 09:46:08 +01002621 <td rowspan="2" style="width:200px;"> Function to perform reduce with the following operations - ARG_IDX_MAX: Index of the max value - ARG_IDX_MIN: Index of the min value - MEAN_SUM: Mean of sum - PROD: Product - SUM_SQUARE: Sum of squares - SUM: Sum - MIN: Min - MAX: Max
Teresa Charlin62687422021-04-28 10:58:49 +01002622 <td rowspan="2">
2623 <ul>
2624 <li>ANEURALNETWORKS_REDUCE_ALL
2625 <li>ANEURALNETWORKS_REDUCE_ANY
2626 <li>ANEURALNETWORKS_REDUCE_MAX
2627 <li>ANEURALNETWORKS_REDUCE_MIN
2628 <li>ANEURALNETWORKS_REDUCE_PROD
2629 <li>ANEURALNETWORKS_REDUCE_SUM
2630 </ul>
2631 <td>NEReductionOperation
2632 <td>
2633 <ul>
2634 <li>All
2635 </ul>
2636 <td>
2637 <table>
2638 <tr><th>src<th>dst
2639 <tr><td>QASYMM8<td>QASYMM8
2640 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2641 <tr><td>F16<td>F16
2642 <tr><td>F32<td>F32
2643 <tr><td>S32<td>S32
2644 </table>
2645<tr>
2646 <td>CLReductionOperation
2647 <td>
2648 <ul>
2649 <li>All
2650 </ul>
2651 <td>
2652 <table>
2653 <tr><th>src<th>dst
2654 <tr><td>QASYMM8<td>QASYMM8
2655 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2656 <tr><td>F16<td>F16
2657 <tr><td>F32<td>F32
2658 <tr><td>S32<td>S32
2659 </table>
2660<tr>
Jakub Sujak667e82f2023-11-07 22:39:30 +00002661 <td rowspan="1">ReorderLayer
2662 <td rowspan="1" style="width:200px;"> Reorders a tensor to a different weights format.
2663 <td rowspan="1">
2664 <ul>
2665 <li>n/a
2666 </ul>
2667 <td>NEReorderLayer
2668 <td>
2669 <ul>
2670 <li>NCHW
2671 </ul>
2672 <td>
2673 <table>
2674 <tr><th>src<th>dst
2675 <tr><td>F32<td>F32
2676 </table>
2677<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002678 <td rowspan="2">ReorgLayer
2679 <td rowspan="2" style="width:200px;"> Performs a reorganization layer of input tensor to the output tensor.
2680 <td rowspan="2">
2681 <ul>
2682 <li>n/a
2683 </ul>
2684 <td>NEReorgLayer
2685 <td>
2686 <ul>
2687 <li>NHWC
2688 <li>NCHW
2689 </ul>
2690 <td>
2691 <table>
2692 <tr><th>src<th>dst
2693 <tr><td>All<td>All
2694 </table>
2695<tr>
2696 <td>CLReorgLayer
2697 <td>
2698 <ul>
2699 <li>NHWC
2700 <li>NCHW
2701 </ul>
2702 <td>
2703 <table>
2704 <tr><th>src<th>dst
2705 <tr><td>All<td>All
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002706 </table>
2707<tr>
2708 <td rowspan="2">ReshapeLayer
Teresa Charlin62687422021-04-28 10:58:49 +01002709 <td rowspan="2" style="width:200px;"> Function to reshape a tensor.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002710 <td rowspan="2">
2711 <ul>
2712 <li>ANEURALNETWORKS_RESHAPE
2713 <li>ANEURALNETWORKS_SQUEEZE
2714 </ul>
2715 <td>NEReshapeLayer
2716 <td>
2717 <ul>
2718 <li>All
2719 </ul>
2720 <td>
2721 <table>
2722 <tr><th>src<th>dst
2723 <tr><td>All<td>All
2724 </table>
2725<tr>
2726 <td>CLReshapeLayer
2727 <td>
2728 <ul>
2729 <li>All
2730 </ul>
2731 <td>
2732 <table>
2733 <tr><th>src<th>dst
2734 <tr><td>All<td>All
2735 </table>
2736<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002737 <td rowspan="2">Reverse
2738 <td rowspan="2" style="width:200px;"> Function to reverse tensor according to axis.
2739 <td rowspan="2">
2740 <ul>
2741 <li>n/a
2742 </ul>
2743 <td>NEReverse
2744 <td>
2745 <ul>
2746 <li>All
2747 </ul>
2748 <td>
2749 <table>
2750 <tr><th>src0<th>src1<th>dst
Adnan AlSinanbdcb4c12023-09-18 14:49:45 +01002751 <tr><td>All<td>U32, S32<td>All
Teresa Charlin62687422021-04-28 10:58:49 +01002752 </table>
2753<tr>
2754 <td>CLReverse
2755 <td>
2756 <ul>
2757 <li>All
2758 </ul>
2759 <td>
2760 <table>
2761 <tr><th>src0<th>src1<th>dst
Adnan AlSinan704c22f2023-10-24 11:05:56 +01002762 <tr><td>All<td>U32, S32<td>All
Teresa Charlin62687422021-04-28 10:58:49 +01002763 </table>
2764<tr>
2765 <td rowspan="2">RNNLayer
2766 <td rowspan="2" style="width:200px;"> Function to perform recurrent neural network layer.
2767 <td rowspan="2">
2768 <ul>
2769 <li>ANEURALNETWORKS_RNN
2770 </ul>
2771 <td>NERNNLayer
2772 <td>
2773 <ul>
2774 <li>NHWC
2775 <li>NCHW
2776 </ul>
2777 <td>
2778 <table>
2779 <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1
2780 <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16
2781 <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32
2782 </table>
2783<tr>
2784 <td>CLRNNLayer
2785 <td>
2786 <ul>
2787 <li>NHWC
2788 <li>NCHW
2789 </ul>
2790 <td>
2791 <table>
2792 <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1
2793 <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16
2794 <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32
2795 </table>
2796<tr>
2797 <td rowspan="2">ROIAlignLayer
2798 <td rowspan="2" style="width:200px;"> Function to perform ROI alignment.
2799 <td rowspan="2">
2800 <ul>
2801 <li>ANEURALNETWORKS_ROI_ALIGN
2802 </ul>
2803 <td>NEROIAlignLayer
2804 <td>
2805 <ul>
2806 <li>All
2807 </ul>
2808 <td>
2809 <table>
2810 <tr><th>src0<th>src1<th>dst
2811 <tr><td>F16<td>F16<td>F16
2812 <tr><td>F32<td>F32<td>F32
2813 <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8
2814 <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED
2815 </table>
2816<tr>
2817 <td>CLROIAlignLayer
2818 <td>
2819 <ul>
2820 <li>All
2821 </ul>
2822 <td>
2823 <table>
2824 <tr><th>src0<th>src1<th>dst
2825 <tr><td>F16<td>F16<td>F16
2826 <tr><td>F32<td>F32<td>F32
2827 <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8
2828 <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED
2829 </table>
2830<tr>
2831 <td rowspan="2">ROIPoolingLayer
2832 <td rowspan="2" style="width:200px;"> Function to perform ROI pooling.
2833 <td rowspan="2">
2834 <ul>
2835 <li>ANEURALNETWORKS_ROI_POOLING
2836 </ul>
2837 <td>NEROIPoolingLayer
2838 <td>
2839 <ul>
2840 <li>All
2841 </ul>
2842 <td>
2843 <table>
2844 <tr><th>src0<th>src1<th>dst
2845 <tr><td>F32<td>U16<td>F32
2846 <tr><td>QASYMM8<td>U16<td>QASYMM8
2847 </table>
2848<tr>
2849 <td>CLROIPoolingLayer
2850 <td>
2851 <ul>
2852 <li>All
2853 </ul>
2854 <td>
2855 <table>
2856 <tr><th>src0<th>src1<th>dst
2857 <tr><td>F16<td>U16<td>F16
2858 <tr><td>F32<td>U16<td>F32
2859 <tr><td>QASYMM8<td>U16<td>QASYMM8
2860 </table>
2861<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002862 <td rowspan="2">Scale
Teresa Charlin62687422021-04-28 10:58:49 +01002863 <td rowspan="2" style="width:200px;"> Function to perform resize a tensor using to interpolate: - Bilinear - Nearest neighbor
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002864 <td rowspan="2">
2865 <ul>
2866 <li>ANEURALNETWORKS_RESIZE_BILINEAR
2867 <li>ANEURALNETWORKS_RESIZE_NEAREST_NEIGHBOR
2868 </ul>
2869 <td>NEScale
2870 <td>
2871 <ul>
2872 <li>NHWC
2873 <li>NCHW
2874 </ul>
2875 <td>
2876 <table>
2877 <tr><th>src<th>dst
2878 <tr><td>QASYMM8<td>QASYMM8
2879 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2880 <tr><td>F16<td>F16
2881 <tr><td>F32<td>F32
2882 <tr><td>U8<td>U8
Gunes Bayirc4f27432022-09-11 15:59:19 +01002883 <tr><td>S8<td>S8
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002884 <tr><td>S16<td>S16
2885 </table>
2886<tr>
2887 <td>CLScale
2888 <td>
2889 <ul>
2890 <li>NHWC
2891 <li>NCHW
2892 </ul>
2893 <td>
2894 <table>
2895 <tr><th>src<th>dst
2896 <tr><td>QASYMM8<td>QASYMM8
2897 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2898 <tr><td>F16<td>F16
2899 <tr><td>F32<td>F32
2900 <tr><td>U8<td>U8
2901 <tr><td>S16<td>S16
2902 </table>
2903<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002904 <td rowspan="2">Select
2905 <td rowspan="2" style="width:200px;"> Function to select values from 2 tensors depending on an input tensor of booleans.
2906 <td rowspan="2">
2907 <ul>
2908 <li>ANEURALNETWORKS_SELECT
2909 </ul>
2910 <td>NESelect
2911 <td>
2912 <ul>
2913 <li>All
2914 </ul>
2915 <td>
2916 <table>
2917 <tr><th>src0<th>src1<th>src2<th>dst
2918 <tr><td>U8<td>All<td>All<td>All
2919 </table>
2920<tr>
2921 <td>CLSelect
2922 <td>
2923 <ul>
2924 <li>All
2925 </ul>
2926 <td>
2927 <table>
2928 <tr><th>src0<th>src1<th>src2<th>dst
2929 <tr><td>U8<td>All<td>All<td>All
2930 </table>
2931<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01002932 <td rowspan="2">Slice
2933 <td rowspan="2" style="width:200px;"> Function to perform tensor slicing.
2934 <td rowspan="2">
2935 <ul>
2936 <li>ANEURALNETWORKS_SLICE
2937 </ul>
2938 <td>NESlice
2939 <td>
2940 <ul>
2941 <li>All
2942 </ul>
2943 <td>
2944 <table>
2945 <tr><th>src<th>dst
2946 <tr><td>All<td>All
2947 </table>
2948<tr>
2949 <td>CLSlice
2950 <td>
2951 <ul>
2952 <li>All
2953 </ul>
2954 <td>
2955 <table>
2956 <tr><th>src<th>dst
2957 <tr><td>All<td>All
2958 </table>
2959<tr>
Sheri Zhang6124ce62021-05-04 14:03:13 +01002960 <td rowspan="2">SoftmaxLayer
2961 <td rowspan="2" style="width:200px;"> Function to compute a SoftmaxLayer and a Log SoftmaxLayer.
2962 <td rowspan="2">
2963 <ul>
2964 <li>ANEURALNETWORKS_LOG_SOFTMAX
2965 <li>ANEURALNETWORKS_SOFTMAX
2966 </ul>
2967 <td>NESoftmaxLayerGeneric
2968 <td>
2969 <ul>
2970 <li>All
2971 </ul>
2972 <td>
2973 <table>
2974 <tr><th>src<th>dst
2975 <tr><td>QASYMM8<td>QASYMM8
2976 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2977 <tr><td>F16<td>F16
2978 <tr><td>F32<td>F32
2979 </table>
2980<tr>
2981 <td>CLSoftmaxLayerGeneric
2982 <td>
2983 <ul>
2984 <li>All
2985 </ul>
2986 <td>
2987 <table>
2988 <tr><th>src<th>dst
2989 <tr><td>QASYMM8<td>QASYMM8
2990 <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
2991 <tr><td>F16<td>F16
2992 <tr><td>F32<td>F32
2993 </table>
2994<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01002995 <td rowspan="2">SpaceToBatchLayer
2996 <td rowspan="2" style="width:200px;"> Function to divide a tensor spatially.
2997 <td rowspan="2">
2998 <ul>
2999 <li>ANEURALNETWORKS_SPACE_TO_BATCH_ND
3000 </ul>
3001 <td>NESpaceToBatchLayer
3002 <td>
3003 <ul>
3004 <li>NHWC
3005 <li>NCHW
3006 </ul>
3007 <td>
3008 <table>
3009 <tr><th>src0<th>src1<th>src2<th>dst
3010 <tr><td>All<td>S32<td>S32<td>All
3011 </table>
3012<tr>
3013 <td>CLSpaceToBatchLayer
3014 <td>
3015 <ul>
3016 <li>NHWC
3017 <li>NCHW
3018 </ul>
3019 <td>
3020 <table>
3021 <tr><th>src0<th>src1<th>src2<th>dst
3022 <tr><td>All<td>S32<td>S32<td>All
3023 </table>
3024<tr>
3025 <td rowspan="2">SpaceToDepthLayer
3026 <td rowspan="2" style="width:200px;"> Function to rearrange blocks of spatial data into depth.
3027 <td rowspan="2">
3028 <ul>
3029 <li>ANEURALNETWORKS_SPACE_TO_DEPTH
3030 </ul>
3031 <td>NESpaceToDepthLayer
3032 <td>
3033 <ul>
3034 <li>NHWC
3035 <li>NCHW
3036 </ul>
3037 <td>
3038 <table>
3039 <tr><th>src<th>dst
3040 <tr><td>All<td>All
3041 </table>
3042<tr>
3043 <td>CLSpaceToDepthLayer
3044 <td>
3045 <ul>
3046 <li>NHWC
3047 <li>NCHW
3048 </ul>
3049 <td>
3050 <table>
3051 <tr><th>src<th>dst
3052 <tr><td>All<td>All
3053 </table>
3054<tr>
3055 <td rowspan="2">Split
3056 <td rowspan="2" style="width:200px;"> Function to split a tensor along a given axis.
3057 <td rowspan="2">
3058 <ul>
3059 <li>ANEURALNETWORKS_SPLIT
3060 </ul>
3061 <td>NESplit
3062 <td>
3063 <ul>
3064 <li>All
3065 </ul>
3066 <td>
3067 <table>
3068 <tr><th>src<th>dst
3069 <tr><td>All<td>All
3070 </table>
3071<tr>
3072 <td>CLSplit
3073 <td>
3074 <ul>
3075 <li>All
3076 </ul>
3077 <td>
3078 <table>
3079 <tr><th>src<th>dst
3080 <tr><td>All<td>All
3081 </table>
3082<tr>
3083 <td rowspan="2">StackLayer
3084 <td rowspan="2" style="width:200px;"> Function to stack tensors along an axis.
3085 <td rowspan="2">
3086 <ul>
3087 <li>n/a
3088 </ul>
3089 <td>NEStackLayer
3090 <td>
3091 <ul>
3092 <li>All
3093 </ul>
3094 <td>
3095 <table>
3096 <tr><th>src<th>dst
3097 <tr><td>All<td>All
3098 </table>
3099<tr>
3100 <td>CLStackLayer
3101 <td>
3102 <ul>
3103 <li>All
3104 </ul>
3105 <td>
3106 <table>
3107 <tr><th>src<th>dst
3108 <tr><td>All<td>All
3109 </table>
3110<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01003111 <td rowspan="2">StridedSlice
3112 <td rowspan="2" style="width:200px;"> Function to extract a strided slice of a tensor.
3113 <td rowspan="2">
3114 <ul>
3115 <li>ANEURALNETWORKS_STRIDED_SLICE
3116 </ul>
3117 <td>NEStridedSlice
3118 <td>
3119 <ul>
3120 <li>All
3121 </ul>
3122 <td>
3123 <table>
3124 <tr><th>src<th>dst
3125 <tr><td>All<td>All
3126 </table>
3127<tr>
3128 <td>CLStridedSlice
3129 <td>
3130 <ul>
3131 <li>All
3132 </ul>
3133 <td>
3134 <table>
3135 <tr><th>src<th>dst
3136 <tr><td>All<td>All
3137 </table>
3138<tr>
Teresa Charlin62687422021-04-28 10:58:49 +01003139 <td rowspan="2">Tile
3140 <td rowspan="2" style="width:200px;"> Function to construct a tensor by tiling a given tensor.
3141 <td rowspan="2">
3142 <ul>
3143 <li>ANEURALNETWORKS_TILE
3144 </ul>
3145 <td>NETile
3146 <td>
3147 <ul>
3148 <li>All
3149 </ul>
3150 <td>
3151 <table>
3152 <tr><th>src<th>dst
3153 <tr><td>All<td>All
3154 </table>
3155<tr>
3156 <td>CLTile
3157 <td>
3158 <ul>
3159 <li>All
3160 </ul>
3161 <td>
3162 <table>
3163 <tr><th>src<th>dst
3164 <tr><td>All<td>All
3165 </table>
3166<tr>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01003167 <td rowspan="2">Transpose
Teresa Charlin62687422021-04-28 10:58:49 +01003168 <td rowspan="2" style="width:200px;"> Function to transpose a 2D tensor.
Sheri Zhanga47dcc22021-04-22 14:41:12 +01003169 <td rowspan="2">
3170 <ul>
3171 <li>ANEURALNETWORKS_TRANSPOSE
3172 </ul>
3173 <td>NETranspose
3174 <td>
3175 <ul>
3176 <li>All
3177 </ul>
3178 <td>
3179 <table>
3180 <tr><th>src<th>dst
3181 <tr><td>All<td>All
3182 </table>
3183<tr>
3184 <td>CLTranspose
3185 <td>
3186 <ul>
3187 <li>All
3188 </ul>
3189 <td>
3190 <table>
3191 <tr><th>src<th>dst
3192 <tr><td>All<td>All
3193 </table>
Teresa Charlin62687422021-04-28 10:58:49 +01003194<tr>
3195 <td rowspan="2">Unstack
3196 <td rowspan="2" style="width:200px;"> Function to unpack a rank-R tensor into rank-(R-1) tensors.
3197 <td rowspan="2">
3198 <ul>
3199 <li>n/a
3200 </ul>
3201 <td>NEUnstack
3202 <td>
3203 <ul>
3204 <li>All
3205 </ul>
3206 <td>
3207 <table>
3208 <tr><th>src<th>dst
3209 <tr><td>All<td>All
3210 </table>
3211<tr>
3212 <td>CLUnstack
3213 <td>
3214 <ul>
3215 <li>All
3216 </ul>
3217 <td>
3218 <table>
3219 <tr><th>src<th>dst
3220 <tr><td>All<td>All
3221 </table>
3222<tr>
3223 <td rowspan="2">WinogradConvolutionLayer
3224 <td rowspan="2" style="width:200px;"> Function to do Winograd Convolution.
3225 <td rowspan="2">
3226 <ul>
3227 <li>ANEURALNETWORKS_CONV_2D
3228 </ul>
3229 <td>NEWinogradConvolutionLayer
3230 <td>
3231 <ul>
3232 <li>NHWC
3233 <li>NCHW
3234 </ul>
3235 <td>
3236 <table>
3237 <tr><th>src0<th>src1<th>src2<th>dst
3238 <tr><td>F16<td>F16<td>F16<td>F16
3239 <tr><td>F32<td>F32<td>F32<td>F32
3240 </table>
3241<tr>
3242 <td>CLWinogradConvolutionLayer
3243 <td>
3244 <ul>
3245 <li>NHWC
3246 <li>NCHW
3247 </ul>
3248 <td>
3249 <table>
3250 <tr><th>src0<th>src1<th>src2<th>dst
3251 <tr><td>F16<td>F16<td>F16<td>F16
3252 <tr><td>F32<td>F32<td>F32<td>F32
3253 </table>
Sheri Zhanga47dcc22021-04-22 14:41:12 +01003254</table>
3255
3256*/
Mohammed Suhail Munshi5e549fa2022-03-16 11:14:06 +00003257} // namespace