blob: cbad7e84cca1629dfb7566517a1f6cd3be62424e [file] [log] [blame]
Sadik Armagan1a9c9f62021-08-05 09:25:15 +01001/// Copyright (c) 2021 ARM Limited and Contributors. All rights reserved.
2///
3/// SPDX-License-Identifier: MIT
4///
5
6namespace armnn
7{
8/**
9@page operator_list Arm NN Operators
10
11@tableofcontents
12
13@section S5_1_operator_list Arm NN Operators
14
15Arm NN supports operators that are listed in below table.
16
17Arm NN supports a wide list of data-types.
18The main data-types that the Machine Learning functions support are the following:
19 <ul>
20 <li><b>BFLOAT16:</b> 16-bit non-standard brain floating point
21 <li><b>QASYMMU8:</b> 8-bit unsigned asymmetric quantized
22 <li><b>QASYMMS8:</b> 8-bit signed asymmetric quantized
23 <li><b>QUANTIZEDSYMM8PERAXIS:</b> 8-bit signed symmetric quantized
24 <li><b>QSYMMS8:</b> 8-bit unsigned symmetric quantized
25 <li><b>QSYMMS16:</b> 16-bit unsigned symmetric quantized
26 <li><b>FLOAT32:</b> 32-bit single precision floating point
27 <li><b>FLOAT16:</b> 16-bit half precision floating point
28 <li><b>SIGNED32:</b> 32-bit signed integer
29 <li><b>BOOLEAN:</b> 8-bit unsigned char
30 <li><b>All:</b> Agnostic to any specific data type
31 </ul>
32
33Arm NN supports the following data layouts (fast changing dimension from right to left):
34 <ul>
35 <li><b>NHWC:</b> Layout where channels are in the fastest changing dimension
36 <li><b>NCHW:</b> Layout where width is in the fastest changing dimension
37 <li><b>All:</b> Agnostic to any specific data layout
38 </ul>
39where N = batches, C = channels, H = height, W = width
40
41<table>
42<caption id="multi_row"></caption>
43<tr>
44 <th>Operator
45 <th>Description
46 <th>Equivalent Android NNAPI Operator
47 <th>Backends
48 <th>Data Layouts
49 <th>Data Types
50<tr>
51 <td rowspan="3">AbsLayer
52 <td rowspan="3"> Layer to perform absolute operation.
53 <td rowspan="3">
54 <ul>
55 <li>ANEURALNETWORKS_ABS
56 </ul>
57 <td>CpuRef
58 <td>
59 <ul>
60 <li>All
61 </ul>
62 <td>
63 <table>
64 <tr><th>
65 <tr><td>BFLOAT16
66 <tr><td>FLOAT16
67 <tr><td>FLOAT32
68 <tr><td>QASYMMS8
69 <tr><td>QASYMMU8
70 <tr><td>QSYMMS16
71 <tr><td>SIGNED32
72 </table>
73<tr>
74 <td>CpuAcc
75 <td>
76 <ul>
77 <li>All
78 </ul>
79 <td>
80 <table>
81 <tr><th>
82 <tr><td>FLOAT16
83 <tr><td>FLOAT32
84 <tr><td>SIGNED32
85 </table>
86<tr>
87 <td>GpuAcc
88 <td>
89 <ul>
90 <li>All
91 </ul>
92 <td>
93 <table>
94 <tr><th>
95 <tr><td>FLOAT16
96 <tr><td>FLOAT32
97 </table>
98<tr>
99 <td rowspan="3">ActivationLayer
100 <td rowspan="3" style="width:200px;"> Layer to simulate an activation layer with the specified activation function.
101 <td rowspan="3">
102 <ul>
103 <li>ANEURALNETWORKS_ABS
104 <li>ANEURALNETWORKS_ELU
105 <li>ANEURALNETWORKS_HARD_SWISH
106 <li>ANEURALNETWORKS_LOGISTIC
107 <li>ANEURALNETWORKS_PRELU
108 <li>ANEURALNETWORKS_RELU
109 <li>ANEURALNETWORKS_RELU1
110 <li>ANEURALNETWORKS_RELU6
111 <li>ANEURALNETWORKS_SQRT
112 <li>ANEURALNETWORKS_TANH
113 </ul>
114 <td>CpuRef
115 <td>
116 <ul>
117 <li>All
118 </ul>
119 <td>
120 <table>
121 <tr><th>
122 <tr><td>BFLOAT16
123 <tr><td>FLOAT16
124 <tr><td>FLOAT32
125 <tr><td>QASYMMS8
126 <tr><td>QASYMMU8
127 <tr><td>QSYMMS16
128 </table>
129<tr>
130 <td>CpuAcc
131 <td>
132 <ul>
133 <li>All
134 </ul>
135 <td>
136 <table>
137 <tr><th>
138 <tr><td>QASYMMU8
139 <tr><td>QASYMMS8
140 <tr><td>QSYMMS16
141 <tr><td>FLOAT16
142 <tr><td>FLOAT32
143 </table>
144<tr>
145 <td>GpuAcc
146 <td>
147 <ul>
148 <li>All
149 </ul>
150 <td>
151 <table>
152 <tr><th>
153 <tr><td>QASYMMU8
154 <tr><td>QASYMMS8
155 <tr><td>QSYMMS16
156 <tr><td>FLOAT16
157 <tr><td>FLOAT32
158 </table>
159<tr>
160 <td rowspan="3">AdditionLayer
161 <td rowspan="3" style="width:200px;"> Layer to add 2 tensors.
162 <td rowspan="3">
163 <ul>
164 <li>ANEURALNETWORKS_ADD
165 </ul>
166 <td>CpuRef
167 <td>
168 <ul>
169 <li>All
170 </ul>
171 <td>
172 <table>
173 <tr><th>
174 <tr><td>BFLOAT16
175 <tr><td>FLOAT16
176 <tr><td>FLOAT32
177 <tr><td>QASYMMS8
178 <tr><td>QASYMMU8
179 <tr><td>QSYMMS16
180 <tr><td>SIGNED32
181 </table>
182<tr>
183 <td>CpuAcc
184 <td>
185 <ul>
186 <li>All
187 </ul>
188 <td>
189 <table>
190 <tr><th>
191 <tr><td>QASYMMU8
192 <tr><td>QASYMMS8
193 <tr><td>QSYMMS16
194 <tr><td>SIGNED32
195 <tr><td>FLOAT16
196 <tr><td>FLOAT32
197 </table>
198<tr>
199 <td>GpuAcc
200 <td>
201 <ul>
202 <li>All
203 </ul>
204 <td>
205 <table>
206 <tr><th>
207 <tr><td>QASYMMU8
208 <tr><td>QASYMMS8
209 <tr><td>QSYMMS16
210 <tr><td>SIGNED32
211 <tr><td>FLOAT16
212 <tr><td>FLOAT32
213 </table>
214<tr>
215 <td rowspan="3">ArgMinMaxLayer
216 <td rowspan="3" style="width:200px;"> Layer to calculate the index of the minimum or maximum values in a tensor
217 based on an axis.
218 <td rowspan="3">
219 <ul>
220 <li>ANEURALNETWORKS_ARGMAX
221 <li>ANEURALNETWORKS_ARGMIN
222 </ul>
223 <td>CpuRef
224 <td>
225 <ul>
226 <li>All
227 </ul>
228 <td>
229 <table>
230 <tr><th>
231 <tr><td>BFLOAT16
232 <tr><td>FLOAT16
233 <tr><td>FLOAT32
234 <tr><td>QASYMMS8
235 <tr><td>QASYMMU8
236 <tr><td>QSYMMS16
237 <tr><td>SIGNED32
238 <tr><td>SIGNED64
239 </table>
240<tr>
241 <td>CpuAcc
242 <td>
243 <ul>
244 <li>All
245 </ul>
246 <td>
247 <table>
248 <tr><th>
249 <tr><td>QASYMMU8
250 <tr><td>QASYMMS8
251 <tr><td>SIGNED32
252 <tr><td>FLOAT16
253 <tr><td>FLOAT32
254 </table>
255<tr>
256 <td>GpuAcc
257 <td>
258 <ul>
259 <li>All
260 </ul>
261 <td>
262 <table>
263 <tr><th>
264 <tr><td>QASYMMU8
265 <tr><td>QASYMMS8
266 <tr><td>SIGNED32
267 <tr><td>FLOAT16
268 <tr><td>FLOAT32
269 </table>
270<tr>
271 <td rowspan="3">BatchNormalizationLayer
272 <td rowspan="3" style="width:200px;"> Layer to perform batch normalization.
273 <td rowspan="3">
274 <ul>
275 <li>N/A
276 </ul>
277 <td>CpuRef
278 <td>
279 <ul>
280 <li>All
281 </ul>
282 <td>
283 <table>
284 <tr><th>
285 <tr><td>BFLOAT16
286 <tr><td>FLOAT16
287 <tr><td>FLOAT32
288 <tr><td>QASYMMS8
289 <tr><td>QASYMMU8
290 <tr><td>QSYMMS16
291 </table>
292<tr>
293 <td>CpuAcc
294 <td>
295 <ul>
296 <li>NHWC
297 <li>NCHW
298 </ul>
299 <td>
300 <table>
301 <tr><th>
302 <tr><td>FLOAT32
303 <tr><td>FLOAT16
304 </table>
305<tr>
306 <td>GpuAcc
307 <td>
308 <ul>
309 <li>NHWC
310 <li>NCHW
311 </ul>
312 <td>
313 <table>
314 <tr><th>
315 <tr><td>FLOAT32
316 <tr><td>FLOAT16
317 </table>
318<tr>
319 <td rowspan="3">BatchToSpaceNdLayer
320 <td rowspan="3" style="width:200px;"> Layer to perform a batch to space transformation.
321 <td rowspan="3">
322 <ul>
323 <li>ANEURALNETWORKS_BATCH_TO_SPACE_ND
324 </ul>
325 <td>CpuRef
326 <td>
327 <ul>
328 <li>All
329 </ul>
330 <td>
331 <table>
332 <tr><th>
333 <tr><td>BFLOAT16
334 <tr><td>FLOAT16
335 <tr><td>FLOAT32
336 <tr><td>QASYMMS8
337 <tr><td>QASYMMU8
338 <tr><td>QSYMMS16
339 </table>
340<tr>
341 <td>CpuAcc
342 <td>
343 <ul>
344 <li>NHWC
345 <li>NCHW
346 </ul>
347 <td>
348 <table>
349 <tr><th>
350 <tr><td>All
351 </table>
352<tr>
353 <td>GpuAcc
354 <td>
355 <ul>
356 <li>NHWC
357 <li>NCHW
358 </ul>
359 <td>
360 <table>
361 <tr><th>
362 <tr><td>All
363 </table>
364<tr>
365 <td rowspan="3">CastLayer
366 <td rowspan="3" style="width:200px;"> Layer to cast a tensor to a type.
367 <td rowspan="3">
368 <ul>
369 <li>ANEURALNETWORKS_CAST
370 </ul>
371 <td>CpuRef
372 <td>
373 <ul>
374 <li>All
375 </ul>
376 <td>
377 <table>
378 <tr><th>
379 <tr><td>BFLOAT16
380 <tr><td>FLOAT16
381 <tr><td>FLOAT32
382 <tr><td>QSYMMS8
383 <tr><td>QASYMMS8
384 <tr><td>QASYMMU8
385 <tr><td>QSYMMS16
386 <tr><td>SIGNED32
387 </table>
388<tr>
389 <td>CpuAcc
390 <td>
391 <ul>
392 <li>All
393 </ul>
394 <td>
395 <table>
396 <tr><th>
397 <tr><td>QASYMMS8
398 <tr><td>QASYMMU8
399 <tr><td>FLOAT16
400 <tr><td>SIGNED32
401 <tr><td>FLOAT32
402 </table>
403<tr>
404 <td>GpuAcc
405 <td>
406 <ul>
407 <li>All
408 </ul>
409 <td>
410 <table>
411 <tr><th>
412 <tr><td>QASYMMS8
413 <tr><td>QASYMMU8
414 <tr><td>SIGNED32
415 <tr><td>FLOAT16
416 <tr><td>FLOAT32
417 </table>
418<tr>
419 <td rowspan="3">ComparisonLayer
420 <td rowspan="3" style="width:200px;"> Layer to compare 2 tensors.
421 <td rowspan="3">
422 <ul>
423 <li>ANEURALNETWORKS_EQUAL
424 <li>ANEURALNETWORKS_GREATER
425 <li>ANEURALNETWORKS_GREATER_EQUAL
426 <li>ANEURALNETWORKS_LESS
427 <li>ANEURALNETWORKS_LESS_EQUAL
428 <li>ANEURALNETWORKS_NOT_EQUAL
429 </ul>
430 <td>CpuRef
431 <td>
432 <ul>
433 <li>All
434 </ul>
435 <td>
436 <table>
437 <tr><th>
438 <tr><td>BFLOAT16
439 <tr><td>FLOAT16
440 <tr><td>FLOAT32
441 <tr><td>BOOLEAN
442 <tr><td>QASYMMS8
443 <tr><td>QASYMMU8
444 <tr><td>QSYMMS16
445 <tr><td>SIGNED32
446 </table>
447<tr>
448 <td>CpuAcc
449 <td>
450 <ul>
451 <li>All
452 </ul>
453 <td>
454 <table>
455 <tr><th>
456 <tr><td>All
457 </table>
458<tr>
459 <td>GpuAcc
460 <td>
461 <ul>
462 <li>All
463 </ul>
464 <td>
465 <table>
466 <tr><th>
467 <tr><td>All
468 </table>
469<tr>
470 <td rowspan="3">ConcatLayer
471 <td rowspan="3" style="width:200px;"> Layer to concatenate tensors along a given axis.
472 <td rowspan="3">
473 <ul>
474 <li>ANEURALNETWORKS_CONCATENATION
475 </ul>
476 <td>CpuRef
477 <td>
478 <ul>
479 <li>All
480 </ul>
481 <td>
482 <table>
483 <tr><th>
484 <tr><td>BFLOAT16
485 <tr><td>FLOAT16
486 <tr><td>FLOAT32
487 <tr><td>QASYMMS8
488 <tr><td>QASYMMU8
489 <tr><td>QSYMMS16
490 </table>
491<tr>
492 <td>CpuAcc
493 <td>
494 <ul>
495 <li>All
496 </ul>
497 <td>
498 <table>
499 <tr><th>
500 <tr><td>QASYMMU8
501 <tr><td>QASYMMS8
502 <tr><td>FLOAT16
503 <tr><td>FLOAT32
504 </table>
505<tr>
506 <td>GpuAcc
507 <td>
508 <ul>
509 <li>All
510 </ul>
511 <td>
512 <table>
513 <tr><th>
514 <tr><td>QASYMMU8
515 <tr><td>QASYMMS8
516 <tr><td>FLOAT16
517 <tr><td>FLOAT32
518 </table>
519<tr>
520 <td rowspan="3">ConstantLayer
521 <td rowspan="3" style="width:200px;"> Layer to provide a constant tensor.
522 <td rowspan="3">
523 <ul>
524 <li>N/A
525 </ul>
526 <td>CpuRef
527 <td>
528 <ul>
529 <li>All
530 </ul>
531 <td>
532 <table>
533 <tr><th>
534 <tr><td>BFLOAT16
535 <tr><td>FLOAT16
536 <tr><td>FLOAT32
537 <tr><td>QASYMMS8
538 <tr><td>QASYMMU8
539 <tr><td>QSYMMS8
540 <tr><td>QSYMMS16
541 <tr><td>SIGNED32
542 </table>
543<tr>
544 <td>CpuAcc
545 <td>
546 <ul>
547 <li>All
548 </ul>
549 <td>
550 <table>
551 <tr><th>
552 <tr><td>All
553 </table>
554<tr>
555 <td>GpuAcc
556 <td>
557 <ul>
558 <li>All
559 </ul>
560 <td>
561 <table>
562 <tr><th>
563 <tr><td>All
564 </table>
565<tr>
566 <td rowspan="3">ConvertBf16ToFp32Layer
567 <td rowspan="3" style="width:200px;"> Layer to convert BFloat16 tensor to Float32 tensor.
568 <td rowspan="3">
569 <ul>
570 <li>N/A
571 </ul>
572 <td>CpuRef
573 <td>
574 <ul>
575 <li>All
576 </ul>
577 <td>
578 <table>
579 <tr><th>
580 <tr><td>BFLOAT16
581 <tr><td>FLOAT32
582 </table>
583<tr>
584 <td>CpuAcc
585 <td>
586 <ul>
587 <li>All
588 </ul>
589 <td>
590 <table>
591 <tr><th>
592 <tr><td>BFLOAT16
593 <tr><td>FLOAT32
594 </table>
595<tr>
596 <td>GpuAcc
597 <td>
598 <ul>
599 <li>All
600 </ul>
601 <td>
602 <table>
603 <tr><th>
604 <tr><td>BFLOAT16
605 <tr><td>FLOAT32
606 </table>
607<tr>
608 <td rowspan="3">ConvertFp16ToFp32Layer
609 <td rowspan="3" style="width:200px;"> Layer to convert Float16 tensor to Float32 tensor.
610 <td rowspan="3">
611 <ul>
612 <li>N/A
613 </ul>
614 <td>CpuRef
615 <td>
616 <ul>
617 <li>All
618 </ul>
619 <td>
620 <table>
621 <tr><th>
622 <tr><td>FLOAT16
623 <tr><td>FLOAT32
624 </table>
625<tr>
626 <td>CpuAcc
627 <td>
628 <ul>
629 <li>All
630 </ul>
631 <td>
632 <table>
633 <tr><th>
634 <tr><td>FLOAT16
635 <tr><td>FLOAT32
636 </table>
637<tr>
638 <td>GpuAcc
639 <td>
640 <ul>
641 <li>All
642 </ul>
643 <td>
644 <table>
645 <tr><th>
646 <tr><td>FLOAT16
647 <tr><td>FLOAT32
648 </table>
649<tr>
650 <td rowspan="3">ConvertFp32ToBf16Layer
651 <td rowspan="3" style="width:200px;"> Layer to convert Float32 tensor to BFloat16 tensor.
652 <td rowspan="3">
653 <ul>
654 <li>N/A
655 </ul>
656 <td>CpuRef
657 <td>
658 <ul>
659 <li>All
660 </ul>
661 <td>
662 <table>
663 <tr><th>
664 <tr><td>BFLOAT16
665 <tr><td>FLOAT32
666 </table>
667<tr>
668 <td>CpuAcc
669 <td>
670 <ul>
671 <li>All
672 </ul>
673 <td>
674 <table>
675 <tr><th>
676 <tr><td>BFLOAT16
677 <tr><td>FLOAT32
678 </table>
679<tr>
680 <td>GpuAcc
681 <td>
682 <ul>
683 <li>All
684 </ul>
685 <td>
686 <table>
687 <tr><th>
688 <tr><td>BFLOAT16
689 <tr><td>FLOAT32
690 </table>
691<tr>
692 <td rowspan="3">ConvertFp32ToFp16Layer
693 <td rowspan="3" style="width:200px;"> Layer to convert Float32 tensor to Float16 tensor.
694 <td rowspan="3">
695 <ul>
696 <li>N/A
697 </ul>
698 <td>CpuRef
699 <td>
700 <ul>
701 <li>All
702 </ul>
703 <td>
704 <table>
705 <tr><th>
706 <tr><td>FLOAT16
707 <tr><td>FLOAT32
708 </table>
709<tr>
710 <td>CpuAcc
711 <td>
712 <ul>
713 <li>All
714 </ul>
715 <td>
716 <table>
717 <tr><th>
718 <tr><td>FLOAT16
719 <tr><td>FLOAT32
720 </table>
721<tr>
722 <td>GpuAcc
723 <td>
724 <ul>
725 <li>All
726 </ul>
727 <td>
728 <table>
729 <tr><th>
730 <tr><td>FLOAT16
731 <tr><td>FLOAT32
732 </table>
733<tr>
734 <td rowspan="3">Convolution2dLayer
735 <td rowspan="3" style="width:200px;"> Layer to compute a convolution operation.
736 <td rowspan="3">
737 <ul>
738 <li>ANEURALNETWORKS_CONV_2D
739 <li>ANEURALNETWORKS_GROUPED_CONV_2D
740 </ul>
741 <td>CpuRef
742 <td>
743 <ul>
744 <li>All
745 </ul>
746 <td>
747 <table>
748 <tr><th>
749 <tr><td>BFLOAT16
750 <tr><td>FLOAT16
751 <tr><td>FLOAT32
752 <tr><td>QASYMMS8
753 <tr><td>QASYMMU8
754 <tr><td>QSYMMS16
755 </table>
756<tr>
757 <td>CpuAcc
758 <td>
759 <ul>
760 <li>NHWC
761 <li>NCHW
762 </ul>
763 <td>
764 <table>
765 <tr><th>
766 <tr><td>SIGNED32
767 <tr><td>FLOAT16
768 <tr><td>FLOAT32
769 <tr><td>QASYMMU8
770 <tr><td>QASYMMS8
771 <tr><td>QUANTIZEDSYMM8PERAXIS
772 </table>
773<tr>
774 <td>GpuAcc
775 <td>
776 <ul>
777 <li>NHWC
778 <li>NCHW
779 </ul>
780 <td>
781 <table>
782 <tr><th>
783 <tr><td>SIGNED32
784 <tr><td>FLOAT16
785 <tr><td>FLOAT32
786 <tr><td>QASYMMU8
787 <tr><td>QASYMMS8
788 <tr><td>QUANTIZEDSYMM8PERAXIS
789 </table>
790<tr>
791 <td rowspan="1">DebugLayer
792 <td rowspan="1" style="width:200px;"> Layer to print out inter layer tensor information.
793 <td rowspan="1">
794 <ul>
795 <li>N/A
796 </ul>
797 <td>CpuRef
798 <td>
799 <ul>
800 <li>All
801 </ul>
802 <td>
803 <table>
804 <tr><th>
805 <tr><td>BFLOAT16
806 <tr><td>FLOAT16
807 <tr><td>FLOAT32
808 <tr><td>QASYMMS8
809 <tr><td>QASYMMU8
810 <tr><td>QSYMMS8
811 <tr><td>QSYMMS16
812 <tr><td>SIGNED32
813 </table>
814<tr>
815 <td rowspan="3">DepthToSpaceLayer
816 <td rowspan="3" style="width:200px;"> Layer to perform Depth to Space transformation.
817 <td rowspan="3">
818 <ul>
819 <li>ANEURALNETWORKS_DEPTH_TO_SPACE
820 </ul>
821 <td>CpuRef
822 <td>
823 <ul>
824 <li>All
825 </ul>
826 <td>
827 <table>
828 <tr><th>
829 <tr><td>BFLOAT16
830 <tr><td>FLOAT16
831 <tr><td>FLOAT32
832 <tr><td>QASYMMS8
833 <tr><td>QASYMMU8
834 <tr><td>QSYMMS16
835 </table>
836<tr>
837 <td>CpuAcc
838 <td>
839 <ul>
840 <li>NHWC
841 <li>NCHW
842 </ul>
843 <td>
844 <table>
845 <tr><th>
846 <tr><td>All
847 </table>
848<tr>
849 <td>GpuAcc
850 <td>
851 <ul>
852 <li>NHWC
853 <li>NCHW
854 </ul>
855 <td>
856 <table>
857 <tr><th>
858 <tr><td>All
859 </table>
860<tr>
861 <td rowspan="3">DepthwiseConvolution2dLayer
862 <td rowspan="3" style="width:200px;"> Layer to compute a deconvolution or transpose convolution.
863 <td rowspan="3">
864 <ul>
865 <li>ANEURALNETWORKS_DEPTHWISE_CONV_2D
866 </ul>
867 <td>CpuRef
868 <td>
869 <ul>
870 <li>All
871 </ul>
872 <td>
873 <table>
874 <tr><th>
875 <tr><td>BFLOAT16
876 <tr><td>FLOAT16
877 <tr><td>FLOAT32
878 <tr><td>QASYMMS8
879 <tr><td>QASYMMU8
880 <tr><td>QSYMMS8
881 <tr><td>QSYMMS16
882 </table>
883<tr>
884 <td>CpuAcc
885 <td>
886 <ul>
887 <li>NHWC
888 <li>NCHW
889 </ul>
890 <td>
891 <table>
892 <tr><th>
893 <tr><td>FLOAT16
894 <tr><td>FLOAT32
895 <tr><td>SIGNED32
896 <tr><td>QASYMMU8
897 <tr><td>QASYMMS8
898 <tr><td>QUANTIZEDSYMM8PERAXIS
899 </table>
900<tr>
901 <td>GpuAcc
902 <td>
903 <ul>
904 <li>NHWC
905 <li>NCHW
906 </ul>
907 <td>
908 <table>
909 <tr><th>
910 <tr><td>FLOAT16
911 <tr><td>FLOAT32
912 <tr><td>SIGNED32
913 <tr><td>QASYMMU8
914 <tr><td>QASYMMS8
915 <tr><td>QUANTIZEDSYMM8PERAXIS
916 </table>
917<tr>
918 <td rowspan="3">DequantizeLayer
919 <td rowspan="3" style="width:200px;"> Layer to dequantize the values in a tensor.
920 <td rowspan="3">
921 <ul>
922 <li>ANEURALNETWORKS_DEQUANTIZE
923 </ul>
924 <td>CpuRef
925 <td>
926 <ul>
927 <li>All
928 </ul>
929 <td>
930 <table>
931 <tr><th>
932 <tr><td>QASYMMS8
933 <tr><td>QASYMMU8
934 <tr><td>QSYMMS8
935 <tr><td>QSYMMS16
936 </table>
937<tr>
938 <td>CpuAcc
939 <td>
940 <ul>
941 <li>All
942 </ul>
943 <td>
944 <table>
945 <tr><th>
946 <tr><td>FLOAT16
947 <tr><td>FLOAT32
948 <tr><td>QASYMMU8
949 <tr><td>QASYMMS8
950 <tr><td>QUANTIZEDSYMM8PERAXIS
951 <tr><td>QSYMMS8
952 <tr><td>QSYMMS16
953 </table>
954<tr>
955 <td>GpuAcc
956 <td>
957 <ul>
958 <li>All
959 </ul>
960 <td>
961 <table>
962 <tr><th>
963 <tr><td>FLOAT16
964 <tr><td>FLOAT32
965 <tr><td>QASYMMU8
966 <tr><td>QASYMMS8
967 <tr><td>QUANTIZEDSYMM8PERAXIS
968 <tr><td>QSYMMS8
969 <tr><td>QSYMMS16
970 </table>
971<tr>
972 <td rowspan="2">DetectionPostProcessLayer
973 <td rowspan="2" style="width:200px;"> Layer to generate the detection output based on center size encoded boxes, class prediction and anchors by doing non maximum suppression (NMS).
974 <td rowspan="2">
975 <ul>
976 <li>ANEURALNETWORKS_DETECTION_POSTPROCESSING
977 </ul>
978 <td>CpuRef
979 <td>
980 <ul>
981 <li>All
982 </ul>
983 <td>
984 <table>
985 <tr><th>
986 <tr><td>BFLOAT16
987 <tr><td>FLOAT16
988 <tr><td>FLOAT32
989 <tr><td>QASYMMS8
990 <tr><td>QASYMMU8
991 <tr><td>QSYMMS16
992 </table>
993<tr>
994 <td>CpuAcc
995 <td>
996 <ul>
997 <li>All
998 </ul>
999 <td>
1000 <table>
1001 <tr><th>
1002 <tr><td>QASYMMU8
1003 <tr><td>QASYMMS8
1004 <tr><td>FLOAT32
1005 </table>
1006<tr>
1007 <td rowspan="3">DivisionLayer
1008 <td rowspan="3" style="width:200px;"> Layer to divide 2 tensors.
1009 <td rowspan="3">
1010 <ul>
1011 <li>ANEURALNETWORKS_DIV
1012 </ul>
1013 <td>CpuRef
1014 <td>
1015 <ul>
1016 <li>All
1017 </ul>
1018 <td>
1019 <table>
1020 <tr><th>
1021 <tr><td>BFLOAT16
1022 <tr><td>FLOAT16
1023 <tr><td>FLOAT32
1024 <tr><td>QASYMMS8
1025 <tr><td>QASYMMU8
1026 <tr><td>QSYMMS16
1027 <tr><td>SIGNED32
1028 </table>
1029<tr>
1030 <td>CpuAcc
1031 <td>
1032 <ul>
1033 <li>All
1034 </ul>
1035 <td>
1036 <table>
1037 <tr><th>
1038 <tr><td>FLOAT16
1039 <tr><td>FLOAT32
1040 </table>
1041<tr>
1042 <td>GpuAcc
1043 <td>
1044 <ul>
1045 <li>All
1046 </ul>
1047 <td>
1048 <table>
1049 <tr><th>
1050 <tr><td>FLOAT16
1051 <tr><td>FLOAT32
1052 </table>
1053<tr>
1054 <td rowspan="3">ElementwiseBaseLayer
1055 <td rowspan="3" style="width:200px;"> Layer to perform Add - Div - Max - Min - Mul operations.
1056 <td rowspan="3">
1057 <ul>
1058 <li>ANEURALNETWORKS_ADD
1059 <li>ANEURALNETWORKS_DIV
1060 <li>ANEURALNETWORKS_MAXIMUM
1061 <li>ANEURALNETWORKS_MINIMUM
1062 <li>ANEURALNETWORKS_MUL
1063 </ul>
1064 <td>CpuRef
1065 <td>
1066 <ul>
1067 <li>All
1068 </ul>
1069 <td>
1070 <table>
1071 <tr><th>
1072 <tr><td>BFLOAT16
1073 <tr><td>FLOAT16
1074 <tr><td>FLOAT32
1075 <tr><td>QASYMMS8
1076 <tr><td>QASYMMU8
1077 <tr><td>QSYMMS16
1078 <tr><td>SIGNED32
1079 </table>
1080<tr>
1081 <td>CpuAcc
1082 <td>
1083 <ul>
1084 <li>All
1085 </ul>
1086 <td>
1087 <table>
1088 <tr><th>
1089 <tr><td>QASYMMU8
1090 <tr><td>QASYMMS8
1091 <tr><td>QSYMMS16
1092 <tr><td>SIGNED32
1093 <tr><td>FLOAT16
1094 <tr><td>FLOAT32
1095 </table>
1096<tr>
1097 <td>GpuAcc
1098 <td>
1099 <ul>
1100 <li>All
1101 </ul>
1102 <td>
1103 <table>
1104 <tr><th>
1105 <tr><td>QASYMMU8
1106 <tr><td>QASYMMS8
1107 <tr><td>QSYMMS16
1108 <tr><td>SIGNED32
1109 <tr><td>FLOAT16
1110 <tr><td>FLOAT32
1111 </table>
1112<tr>
1113 <td rowspan="3">ElementwiseUnaryLayer
1114 <td rowspan="3" style="width:200px;"> Layer to perform Rsqrt - Exp - Neg - Log - Abs - Sin - Sqrt operations.
1115 <td rowspan="3">
1116 <ul>
1117 <li>ANEURALNETWORKS_ABS
1118 <li>ANEURALNETWORKS_EXP
1119 <li>ANEURALNETWORKS_LOG
1120 <li>ANEURALNETWORKS_NEG
1121 <li>ANEURALNETWORKS_RSQRT
1122 <li>ANEURALNETWORKS_SIN
1123 <li>ANEURALNETWORKS_SQRT
1124 </ul>
1125 <td>CpuRef
1126 <td>
1127 <ul>
1128 <li>All
1129 </ul>
1130 <td>
1131 <table>
1132 <tr><th>
1133 <tr><td>BFLOAT16
1134 <tr><td>FLOAT16
1135 <tr><td>FLOAT32
1136 <tr><td>QASYMMS8
1137 <tr><td>QASYMMU8
1138 <tr><td>QSYMMS16
1139 </table>
1140<tr>
1141 <td>CpuAcc
1142 <td>
1143 <ul>
1144 <li>All
1145 </ul>
1146 <td>
1147 <table>
1148 <tr><th>
1149 <tr><td>FLOAT16
1150 <tr><td>FLOAT32
1151 <tr><td>SIGNED32
1152 </table>
1153<tr>
1154 <td>GpuAcc
1155 <td>
1156 <ul>
1157 <li>All
1158 </ul>
1159 <td>
1160 <table>
1161 <tr><th>
1162 <tr><td>FLOAT16
1163 <tr><td>FLOAT32
1164 </table>
1165<tr>
1166 <td rowspan="1">FakeQuantizationLayer
1167 <td rowspan="1" style="width:200px;"> Layer to quantize float values and dequantize afterwards. The current implementation does not dequantize the values.
1168 <td rowspan="1">
1169 <ul>
1170 <li>N/A
1171 </ul>
1172 <td>CpuRef
1173 <td>
1174 <ul>
1175 <li>All
1176 </ul>
1177 <td>
1178 <table>
1179 <tr><th>
1180 <tr><td>FLOAT32
1181 </table>
1182<tr>
1183 <td rowspan="3">FillLayer
1184 <td rowspan="3" style="width:200px;"> Layer to set the values of a tensor with a given value.
1185 <td rowspan="3">
1186 <ul>
1187 <li>ANEURALNETWORKS_FILL
1188 </ul>
1189 <td>CpuRef
1190 <td>
1191 <ul>
1192 <li>All
1193 </ul>
1194 <td>
1195 <table>
1196 <tr><th>
1197 <tr><td>FLOAT16
1198 <tr><td>FLOAT32
1199 <tr><td>SIGNED32
1200 </table>
1201<tr>
1202 <td>CpuAcc
1203 <td>
1204 <ul>
1205 <li>All
1206 </ul>
1207 <td>
1208 <table>
1209 <tr><th>
1210 <tr><td>All
1211 </table>
1212<tr>
1213 <td>GpuAcc
1214 <td>
1215 <ul>
1216 <li>All
1217 </ul>
1218 <td>
1219 <table>
1220 <tr><th>
1221 <tr><td>All
1222 </table>
1223<tr>
1224 <td rowspan="3">FloorLayer
1225 <td rowspan="3" style="width:200px;"> Layer to round the value to the lowest whole number.
1226 <td rowspan="3">
1227 <ul>
1228 <li>ANEURALNETWORKS_FLOOR
1229 </ul>
1230 <td>CpuRef
1231 <td>
1232 <ul>
1233 <li>All
1234 </ul>
1235 <td>
1236 <table>
1237 <tr><th>
1238 <tr><td>BFLOAT16
1239 <tr><td>FLOAT16
1240 <tr><td>FLOAT32
1241 </table>
1242<tr>
1243 <td>CpuAcc
1244 <td>
1245 <ul>
1246 <li>All
1247 </ul>
1248 <td>
1249 <table>
1250 <tr><th>
1251 <tr><td>FLOAT32
1252 <tr><td>FLOAT16
1253 </table>
1254<tr>
1255 <td>GpuAcc
1256 <td>
1257 <ul>
1258 <li>All
1259 </ul>
1260 <td>
1261 <table>
1262 <tr><th>
1263 <tr><td>FLOAT32
1264 <tr><td>FLOAT16
1265 </table>
1266<tr>
1267 <td rowspan="3">FullyConnectedLayer
1268 <td rowspan="3" style="width:200px;"> Layer to perform a fully connected / dense operation.
1269 <td rowspan="3">
1270 <ul>
1271 <li>ANEURALNETWORKS_FULLY_CONNECTED
1272 </ul>
1273 <td>CpuRef
1274 <td>
1275 <ul>
1276 <li>All
1277 </ul>
1278 <td>
1279 <table>
1280 <tr><th>
1281 <tr><td>BFLOAT16
1282 <tr><td>FLOAT16
1283 <tr><td>FLOAT32
1284 <tr><td>QASYMMS8
1285 <tr><td>QASYMMU8
1286 <tr><td>QSYMMS16
1287 </table>
1288<tr>
1289 <td>CpuAcc
1290 <td>
1291 <ul>
1292 <li>NHWC
1293 <li>NCHW
1294 </ul>
1295 <td>
1296 <table>
1297 <tr><th>
1298 <tr><td>SIGNED32
1299 <tr><td>FLOAT16
1300 <tr><td>FLOAT32
1301 <tr><td>QASYMMU8
1302 <tr><td>QASYMMS8
1303 </table>
1304<tr>
1305 <td>GpuAcc
1306 <td>
1307 <ul>
1308 <li>NHWC
1309 <li>NCHW
1310 </ul>
1311 <td>
1312 <table>
1313 <tr><th>
1314 <tr><td>SIGNED32
1315 <tr><td>FLOAT16
1316 <tr><td>FLOAT32
1317 <tr><td>QASYMMU8
1318 <tr><td>QASYMMS8
1319 </table>
1320<tr>
1321 <td rowspan="3">GatherLayer
1322 <td rowspan="3" style="width:200px;"> Layer to perform the gather operation along the chosen axis.
1323 <td rowspan="3">
1324 <ul>
1325 <li>ANEURALNETWORKS_GATHER
1326 </ul>
1327 <td>CpuRef
1328 <td>
1329 <ul>
1330 <li>All
1331 </ul>
1332 <td>
1333 <table>
1334 <tr><th>
1335 <tr><td>BFLOAT16
1336 <tr><td>FLOAT16
1337 <tr><td>FLOAT32
1338 <tr><td>QASYMMS8
1339 <tr><td>QASYMMU8
1340 <tr><td>QSYMMS16
1341 <tr><td>SIGNED32
1342 </table>
1343<tr>
1344 <td>CpuAcc
1345 <td>
1346 <ul>
1347 <li>All
1348 </ul>
1349 <td>
1350 <table>
1351 <tr><th>
1352 <tr><td>All
1353 </table>
1354<tr>
1355 <td>GpuAcc
1356 <td>
1357 <ul>
1358 <li>All
1359 </ul>
1360 <td>
1361 <table>
1362 <tr><th>
1363 <tr><td>All
1364 </table>
1365<tr>
1366 <td rowspan="1">InputLayer
1367 <td rowspan="1" style="width:200px;"> Special layer used to provide input data to the computational network.
1368 <td rowspan="1">
1369 <ul>
1370 <li>N/A
1371 </ul>
1372 <td>All
1373 <td>
1374 <ul>
1375 <li>All
1376 </ul>
1377 <td>
1378 <table>
1379 <tr><th>
1380 <tr><td>All
1381 </table>
1382<tr>
1383 <td rowspan="3">InstanceNormalizationLayer
1384 <td rowspan="3" style="width:200px;"> Layer to perform an instance normalization on a given axis.
1385 <td rowspan="3">
1386 <ul>
1387 <li>ANEURALNETWORKS_INSTANCE_NORMALIZATION
1388 </ul>
1389 <td>CpuRef
1390 <td>
1391 <ul>
1392 <li>All
1393 </ul>
1394 <td>
1395 <table>
1396 <tr><th>
1397 <tr><td>BFLOAT16
1398 <tr><td>FLOAT16
1399 <tr><td>FLOAT32
1400 </table>
1401<tr>
1402 <td>CpuAcc
1403 <td>
1404 <ul>
1405 <li>NHWC
1406 <li>NCHW
1407 </ul>
1408 <td>
1409 <table>
1410 <tr><th>
1411 <tr><td>FLOAT16
1412 <tr><td>FLOAT32
1413 </table>
1414<tr>
1415 <td>GpuAcc
1416 <td>
1417 <ul>
1418 <li>NHWC
1419 <li>NCHW
1420 </ul>
1421 <td>
1422 <table>
1423 <tr><th>
1424 <tr><td>FLOAT16
1425 <tr><td>FLOAT32
1426 </table>
1427<tr>
1428 <td rowspan="3">L2NormalizationLayer
1429 <td rowspan="3" style="width:200px;"> Layer to perform an L2 normalization on a given axis.
1430 <td rowspan="3">
1431 <ul>
1432 <li>ANEURALNETWORKS_L2_NORMALIZATION
1433 </ul>
1434 <td>CpuRef
1435 <td>
1436 <ul>
1437 <li>All
1438 </ul>
1439 <td>
1440 <table>
1441 <tr><th>
1442 <tr><td>BFLOAT16
1443 <tr><td>FLOAT16
1444 <tr><td>FLOAT32
1445 <tr><td>QASYMMS8
1446 <tr><td>QASYMMU8
1447 <tr><td>QSYMMS16
1448 </table>
1449<tr>
1450 <td>CpuAcc
1451 <td>
1452 <ul>
1453 <li>NHWC
1454 <li>NCHW
1455 </ul>
1456 <td>
1457 <table>
1458 <tr><th>
1459 <tr><td>FLOAT16
1460 <tr><td>FLOAT32
1461 </table>
1462<tr>
1463 <td>GpuAcc
1464 <td>
1465 <ul>
1466 <li>NHWC
1467 <li>NCHW
1468 </ul>
1469 <td>
1470 <table>
1471 <tr><th>
1472 <tr><td>FLOAT16
1473 <tr><td>FLOAT32
1474 </table>
1475<tr>
1476 <td rowspan="3">LogSoftmaxLayer
1477 <td rowspan="3" style="width:200px;"> Layer to perform the log softmax activations given logits.
1478 <td rowspan="3">
1479 <ul>
1480 <li>N/A
1481 </ul>
1482 <td>CpuRef
1483 <td>
1484 <ul>
1485 <li>All
1486 </ul>
1487 <td>
1488 <table>
1489 <tr><th>
1490 <tr><td>BFLOAT16
1491 <tr><td>FLOAT16
1492 <tr><td>FLOAT32
1493 </table>
1494<tr>
1495 <td>CpuAcc
1496 <td>
1497 <ul>
1498 <li>All
1499 </ul>
1500 <td>
1501 <table>
1502 <tr><th>
1503 <tr><td>QASYMMU8
1504 <tr><td>QASYMMS8
1505 <tr><td>FLOAT16
1506 <tr><td>FLOAT32
1507 </table>
1508<tr>
1509 <td>GpuAcc
1510 <td>
1511 <ul>
1512 <li>All
1513 </ul>
1514 <td>
1515 <table>
1516 <tr><th>
1517 <tr><td>QASYMMU8
1518 <tr><td>QASYMMS8
1519 <tr><td>FLOAT16
1520 <tr><td>FLOAT32
1521 </table>
1522<tr>
1523 <td rowspan="3">LogicalBinaryLayer
1524 <td rowspan="3" style="width:200px;"> Layer to perform Logical AND - Logical NOT - Logical OR operations.
1525 <td rowspan="3">
1526 <ul>
1527 <li>ANEURALNETWORKS_LOGICAL_AND
1528 <li>ANEURALNETWORKS_LOGICAL_NOT
1529 <li>ANEURALNETWORKS_LOGICAL_OR
1530 </ul>
1531 <td>CpuRef
1532 <td>
1533 <ul>
1534 <li>All
1535 </ul>
1536 <td>
1537 <table>
1538 <tr><th>
1539 <tr><td>BOOLEAN
1540 </table>
1541<tr>
1542 <td>CpuAcc
1543 <td>
1544 <ul>
1545 <li>All
1546 </ul>
1547 <td>
1548 <table>
1549 <tr><th>
1550 <tr><td>BOOLEAN
1551 </table>
1552<tr>
1553 <td>GpuAcc
1554 <td>
1555 <ul>
1556 <li>All
1557 </ul>
1558 <td>
1559 <table>
1560 <tr><th>
1561 <tr><td>BOOLEAN
1562 </table>
1563<tr>
1564 <td rowspan="3">LstmLayer
1565 <td rowspan="3" style="width:200px;"> Layer to perform a single time step in a Long Short-Term Memory (LSTM) operation.
1566 <td rowspan="3">
1567 <ul>
1568 <li>ANEURALNETWORKS_LSTM
1569 </ul>
1570 <td>CpuRef
1571 <td>
1572 <ul>
1573 <li>All
1574 </ul>
1575 <td>
1576 <table>
1577 <tr><th>
1578 <tr><td>BFLOAT16
1579 <tr><td>FLOAT16
1580 <tr><td>QSYMMS16
1581 </table>
1582<tr>
1583 <td>CpuAcc
1584 <td>
1585 <ul>
1586 <li>All
1587 </ul>
1588 <td>
1589 <table>
1590 <tr><th>
1591 <tr><td>FLOAT16
1592 <tr><td>FLOAT32
1593 </table>
1594<tr>
1595 <td>GpuAcc
1596 <td>
1597 <ul>
1598 <li>All
1599 </ul>
1600 <td>
1601 <table>
1602 <tr><th>
1603 <tr><td>FLOAT16
1604 <tr><td>FLOAT32
1605 </table>
1606<tr>
1607 <td rowspan="3">MapLayer
1608 <td rowspan="3" style="width:200px;"> Layer to perform map operation on tensor.
1609 <td rowspan="3">
1610 <ul>
1611 <li>N/A
1612 </ul>
1613 <td>CpuRef
1614 <td>
1615 <ul>
1616 <li>All
1617 </ul>
1618 <td>
1619 <table>
1620 <tr><th>
1621 <tr><td>All
1622 </table>
1623<tr>
1624 <td>CpuAcc
1625 <td>
1626 <ul>
1627 <li>All
1628 </ul>
1629 <td>
1630 <table>
1631 <tr><th>
1632 <tr><td>All
1633 </table>
1634<tr>
1635 <td>GpuAcc
1636 <td>
1637 <ul>
1638 <li>All
1639 </ul>
1640 <td>
1641 <table>
1642 <tr><th>
1643 <tr><td>All
1644 </table>
1645<tr>
1646 <td rowspan="3">MaximumLayer
1647 <td rowspan="3" style="width:200px;"> Layer to perform an elementwise maximum of two tensors.
1648 <td rowspan="3">
1649 <ul>
1650 <li>N/A
1651 </ul>
1652 <td>CpuRef
1653 <td>
1654 <ul>
1655 <li>All
1656 </ul>
1657 <td>
1658 <table>
1659 <tr><th>
1660 <tr><td>BFLOAT16
1661 <tr><td>FLOAT16
1662 <tr><td>FLOAT32
1663 <tr><td>QASYMMS8
1664 <tr><td>QASYMMU8
1665 <tr><td>QSYMMS16
1666 <tr><td>SIGNED32
1667 </table>
1668<tr>
1669 <td>CpuAcc
1670 <td>
1671 <ul>
1672 <li>All
1673 </ul>
1674 <td>
1675 <table>
1676 <tr><th>
1677 <tr><td>QASYMMU8
1678 <tr><td>QASYMMS8
1679 <tr><td>FLOAT16
1680 <tr><td>FLOAT32
1681 <tr><td>SIGNED32
1682 </table>
1683<tr>
1684 <td>GpuAcc
1685 <td>
1686 <ul>
1687 <li>All
1688 </ul>
1689 <td>
1690 <table>
1691 <tr><th>
1692 <tr><td>QASYMMU8
1693 <tr><td>QASYMMS8
1694 <tr><td>QSYMMS16
1695 <tr><td>FLOAT16
1696 <tr><td>FLOAT32
1697 <tr><td>SIGNED32
1698 </table>
1699<tr>
1700 <td rowspan="3">MeanLayer
1701 <td rowspan="3" style="width:200px;"> Layer to perform reduce mean operation.
1702 <td rowspan="3">
1703 <ul>
1704 <li>ANEURALNETWORKS_MEAN
1705 </ul>
1706 <td>CpuRef
1707 <td>
1708 <ul>
1709 <li>All
1710 </ul>
1711 <td>
1712 <table>
1713 <tr><th>
1714 <tr><td>BFLOAT16
1715 <tr><td>FLOAT16
1716 <tr><td>FLOAT32
1717 <tr><td>QASYMMS8
1718 <tr><td>QASYMMU8
1719 <tr><td>QSYMMS16
1720 </table>
1721<tr>
1722 <td>CpuAcc
1723 <td>
1724 <ul>
1725 <li>All
1726 </ul>
1727 <td>
1728 <table>
1729 <tr><th>
1730 <tr><td>QASYMMU8
1731 <tr><td>QASYMMS8
1732 <tr><td>FLOAT16
1733 <tr><td>FLOAT32
1734 </table>
1735<tr>
1736 <td>GpuAcc
1737 <td>
1738 <ul>
1739 <li>All
1740 </ul>
1741 <td>
1742 <table>
1743 <tr><th>
1744 <tr><td>QASYMMU8
1745 <tr><td>QASYMMS8
1746 <tr><td>FLOAT16
1747 <tr><td>FLOAT32
1748 </table>
1749<tr>
1750 <td rowspan="3">MemCopyLayer
1751 <td rowspan="3" style="width:200px;"> Layer to perform memory copy operation.
1752 <td rowspan="3">
1753 <ul>
1754 <li>N/A
1755 </ul>
1756 <td>CpuRef
1757 <td>
1758 <ul>
1759 <li>All
1760 </ul>
1761 <td>
1762 <table>
1763 <tr><th>
1764 <tr><td>BFLOAT16
1765 <tr><td>FLOAT16
1766 <tr><td>FLOAT32
1767 <tr><td>QASYMMS8
1768 <tr><td>QASYMMU8
1769 <tr><td>QSYMMS16
1770 <tr><td>BOOLEAN
1771 </table>
1772<tr>
1773 <td>CpuAcc
1774 <td>
1775 <ul>
1776 <li>All
1777 </ul>
1778 <td>
1779 <table>
1780 <tr><th>
1781 <tr><td>All
1782 </table>
1783<tr>
1784 <td>GpuAcc
1785 <td>
1786 <ul>
1787 <li>All
1788 </ul>
1789 <td>
1790 <table>
1791 <tr><th>
1792 <tr><td>All
1793 </table>
1794<tr>
1795 <td rowspan="3">MemImportLayer
1796 <td rowspan="3" style="width:200px;"> Layer to perform memory import operation.
1797 <td rowspan="3">
1798 <ul>
1799 <li>N/A
1800 </ul>
1801 <td>CpuRef
1802 <td>
1803 <ul>
1804 <li>All
1805 </ul>
1806 <td>
1807 <table>
1808 <tr><th>
1809 <tr><td>All
1810 </table>
1811<tr>
1812 <td>CpuAcc
1813 <td>
1814 <ul>
1815 <li>All
1816 </ul>
1817 <td>
1818 <table>
1819 <tr><th>
1820 <tr><td>All
1821 </table>
1822<tr>
1823 <td>GpuAcc
1824 <td>
1825 <ul>
1826 <li>All
1827 </ul>
1828 <td>
1829 <table>
1830 <tr><th>
1831 <tr><td>All
1832 </table>
1833<tr>
1834 <td rowspan="3">MergeLayer
1835 <td rowspan="3" style="width:200px;"> Layer to concatenate tensors along a given axis.
1836 <td rowspan="3">
1837 <ul>
1838 <li>ANEURALNETWORKS_CONCATENATION
1839 </ul>
1840 <td>CpuRef
1841 <td>
1842 <ul>
1843 <li>All
1844 </ul>
1845 <td>
1846 <table>
1847 <tr><th>
1848 <tr><td>BFLOAT16
1849 <tr><td>FLOAT16
1850 <tr><td>FLOAT32
1851 <tr><td>QASYMMS8
1852 <tr><td>QASYMMU8
1853 <tr><td>QSYMMS16
1854 </table>
1855<tr>
1856 <td>CpuAcc
1857 <td>
1858 <ul>
1859 <li>All
1860 </ul>
1861 <td>
1862 <table>
1863 <tr><th>
1864 <tr><td>QASYMMU8
1865 <tr><td>QASYMMS8
1866 <tr><td>FLOAT16
1867 <tr><td>FLOAT32
1868 </table>
1869<tr>
1870 <td>GpuAcc
1871 <td>
1872 <ul>
1873 <li>All
1874 </ul>
1875 <td>
1876 <table>
1877 <tr><th>
1878 <tr><td>QASYMMU8
1879 <tr><td>QASYMMS8
1880 <tr><td>FLOAT16
1881 <tr><td>FLOAT32
1882 </table>
1883<tr>
1884 <td rowspan="3">MinimumLayer
1885 <td rowspan="3" style="width:200px;"> Layer to perform an elementwise minimum of two tensors.
1886 <td rowspan="3">
1887 <ul>
1888 <li>ANEURALNETWORKS_MINIMUM
1889 </ul>
1890 <td>CpuRef
1891 <td>
1892 <ul>
1893 <li>All
1894 </ul>
1895 <td>
1896 <table>
1897 <tr><th>
1898 <tr><td>BFLOAT16
1899 <tr><td>FLOAT16
1900 <tr><td>FLOAT32
1901 <tr><td>QASYMMS8
1902 <tr><td>QASYMMU8
1903 <tr><td>QSYMMS16
1904 <tr><td>SIGNED32
1905 </table>
1906<tr>
1907 <td>CpuAcc
1908 <td>
1909 <ul>
1910 <li>All
1911 </ul>
1912 <td>
1913 <table>
1914 <tr><th>
1915 <tr><td>QASYMMU8
1916 <tr><td>QASYMMS8
1917 <tr><td>QSYMMS16
1918 <tr><td>FLOAT16
1919 <tr><td>FLOAT32
1920 </table>
1921<tr>
1922 <td>GpuAcc
1923 <td>
1924 <ul>
1925 <li>All
1926 </ul>
1927 <td>
1928 <table>
1929 <tr><th>
1930 <tr><td>QASYMMU8
1931 <tr><td>QASYMMS8
1932 <tr><td>QSYMMS16
1933 <tr><td>FLOAT16
1934 <tr><td>FLOAT32
1935 <tr><td>SIGNED32
1936 </table>
1937<tr>
1938 <td rowspan="3">MultiplicationLayer
1939 <td rowspan="3" style="width:200px;"> Layer to perform an elementwise multiplication of two tensors.
1940 <td rowspan="3">
1941 <ul>
1942 <li>ANEURALNETWORKS_MUL
1943 </ul>
1944 <td>CpuRef
1945 <td>
1946 <ul>
1947 <li>All
1948 </ul>
1949 <td>
1950 <table>
1951 <tr><th>
1952 <tr><td>BFLOAT16
1953 <tr><td>FLOAT16
1954 <tr><td>FLOAT32
1955 <tr><td>QASYMMS8
1956 <tr><td>QASYMMU8
1957 <tr><td>QSYMMS16
1958 <tr><td>SIGNED32
1959 </table>
1960<tr>
1961 <td>CpuAcc
1962 <td>
1963 <ul>
1964 <li>All
1965 </ul>
1966 <td>
1967 <table>
1968 <tr><th>
1969 <tr><td>QASYMMU8
1970 <tr><td>QASYMMS8
1971 <tr><td>QSYMMS16
1972 <tr><td>SIGNED32
1973 <tr><td>FLOAT16
1974 <tr><td>FLOAT32
1975 </table>
1976<tr>
1977 <td>GpuAcc
1978 <td>
1979 <ul>
1980 <li>All
1981 </ul>
1982 <td>
1983 <table>
1984 <tr><th>
1985 <tr><td>QASYMMU8
1986 <tr><td>QASYMMS8
1987 <tr><td>QSYMMS16
1988 <tr><td>SIGNED32
1989 <tr><td>FLOAT16
1990 <tr><td>FLOAT32
1991 <tr><td>SIGNED32
1992 </table>
1993<tr>
1994 <td rowspan="3">NormalizationLayer
1995 <td rowspan="3" style="width:200px;"> Layer to compute normalization operation.
1996 <td rowspan="3">
1997 <ul>
1998 <li>ANEURALNETWORKS_LOCAL_RESPONSE_NORMALIZATION
1999 </ul>
2000 <td>CpuRef
2001 <td>
2002 <ul>
2003 <li>All
2004 </ul>
2005 <td>
2006 <table>
2007 <tr><th>
2008 <tr><td>BFLOAT16
2009 <tr><td>FLOAT16
2010 <tr><td>FLOAT32
2011 <tr><td>QASYMMS8
2012 <tr><td>QASYMMU8
2013 <tr><td>QSYMMS16
2014 </table>
2015<tr>
2016 <td>CpuAcc
2017 <td>
2018 <ul>
2019 <li>NHWC
2020 <li>NCHW
2021 </ul>
2022 <td>
2023 <table>
2024 <tr><th>
2025 <tr><td>FLOAT32
2026 <tr><td>FLOAT16
2027 </table>
2028<tr>
2029 <td>GpuAcc
2030 <td>
2031 <ul>
2032 <li>NHWC
2033 <li>NCHW
2034 </ul>
2035 <td>
2036 <table>
2037 <tr><th>
2038 <tr><td>FLOAT32
2039 <tr><td>FLOAT16
2040 </table>
2041<tr>
2042 <td rowspan="1">OutputLayer
2043 <td rowspan="1" style="width:200px;"> A special layer providing access to a user supplied buffer into which the output of a network can be written.
2044 <td rowspan="1">
2045 <ul>
2046 <li>N/A
2047 </ul>
2048 <td>All
2049 <td>
2050 <ul>
2051 <li>All
2052 </ul>
2053 <td>
2054 <table>
2055 <tr><th>
2056 <tr><td>All
2057 </table>
2058<tr>
2059 <td rowspan="3">PadLayer
2060 <td rowspan="3" style="width:200px;"> Layer to pad a tensor.
2061 <td rowspan="3">
2062 <ul>
2063 <li>ANEURALNETWORKS_PAD
2064 <li>ANEURALNETWORKS_PAD_V2
2065 </ul>
2066 <td>CpuRef
2067 <td>
2068 <ul>
2069 <li>All
2070 </ul>
2071 <td>
2072 <table>
2073 <tr><th>
2074 <tr><td>BFLOAT16
2075 <tr><td>FLOAT16
2076 <tr><td>FLOAT32
2077 <tr><td>QASYMMS8
2078 <tr><td>QASYMMU8
2079 <tr><td>QSYMMS16
2080 </table>
2081<tr>
2082 <td>CpuAcc
2083 <td>
2084 <ul>
2085 <li>NHWC
2086 <li>NCHW
2087 </ul>
2088 <td>
2089 <table>
2090 <tr><th>
2091 <tr><td>All
2092 </table>
2093<tr>
2094 <td>GpuAcc
2095 <td>
2096 <ul>
2097 <li>NHWC
2098 <li>NCHW
2099 </ul>
2100 <td>
2101 <table>
2102 <tr><th>
2103 <tr><td>All
2104 </table>
2105<tr>
2106 <td rowspan="3">PermuteLayer
2107 <td rowspan="3" style="width:200px;"> Layer to transpose an ND tensor.
2108 <td rowspan="3">
2109 <ul>
2110 <li>ANEURALNETWORKS_TRANSPOSE
2111 </ul>
2112 <td>CpuRef
2113 <td>
2114 <ul>
2115 <li>All
2116 </ul>
2117 <td>
2118 <table>
2119 <tr><th>
2120 <tr><td>BFLOAT16
2121 <tr><td>FLOAT16
2122 <tr><td>FLOAT32
2123 <tr><td>QASYMMS8
2124 <tr><td>QASYMMU8
2125 <tr><td>QSYMMS16
2126 </table>
2127<tr>
2128 <td>CpuAcc
2129 <td>
2130 <ul>
2131 <li>NHWC
2132 <li>NCHW
2133 </ul>
2134 <td>
2135 <table>
2136 <tr><th>
2137 <tr><td>All
2138 </table>
2139<tr>
2140 <td>GpuAcc
2141 <td>
2142 <ul>
2143 <li>NHWC
2144 <li>NCHW
2145 </ul>
2146 <td>
2147 <table>
2148 <tr><th>
2149 <tr><td>All
2150 </table>
2151<tr>
2152 <td rowspan="3">Pooling2dLayer
2153 <td rowspan="3" style="width:200px;"> Layer to perform pooling with the specified pooling operation.
2154 <td rowspan="3">
2155 <ul>
2156 <li>ANEURALNETWORKS_AVERAGE_POOL_2D
2157 <li>ANEURALNETWORKS_L2_POOL_2D
2158 <li>ANEURALNETWORKS_MAX_POOL_2D
2159 </ul>
2160 <td>CpuRef
2161 <td>
2162 <ul>
2163 <li>All
2164 </ul>
2165 <td>
2166 <table>
2167 <tr><th>
2168 <tr><td>BFLOAT16
2169 <tr><td>FLOAT16
2170 <tr><td>FLOAT32
2171 <tr><td>QASYMMS8
2172 <tr><td>QASYMMU8
2173 <tr><td>QSYMMS16
2174 </table>
2175<tr>
2176 <td>CpuAcc
2177 <td>
2178 <ul>
2179 <li>NHWC
2180 <li>NCHW
2181 </ul>
2182 <td>
2183 <table>
2184 <tr><th>
2185 <tr><td>QASYMMU8
2186 <tr><td>QASYMMS8
2187 <tr><td>FLOAT16
2188 <tr><td>FLOAT32
2189 </table>
2190<tr>
2191 <td>GpuAcc
2192 <td>
2193 <ul>
2194 <li>NHWC
2195 <li>NCHW
2196 </ul>
2197 <td>
2198 <table>
2199 <tr><th>
2200 <tr><td>QASYMMU8
2201 <tr><td>QASYMMS8
2202 <tr><td>FLOAT16
2203 <tr><td>FLOAT32
2204 </table>
2205<tr>
2206 <td rowspan="1">PreCompiledLayer
2207 <td rowspan="1" style="width:200px;"> Opaque layer provided by a backend which provides an executable representation of a subgraph from the original network.
2208 <td rowspan="1">
2209 <ul>
2210 <li>N/A
2211 </ul>
2212 <td>N/A
2213 <td>N/A
2214 <td>N/A
2215<tr>
2216 <td rowspan="3">PreluLayer
2217 <td rowspan="3" style="width:200px;"> Layer to compute the activation layer with the PRELU activation function.
2218 <td rowspan="3">
2219 <ul>
2220 <li>ANEURALNETWORKS_PRELU
2221 </ul>
2222 <td>CpuRef
2223 <td>
2224 <ul>
2225 <li>All
2226 </ul>
2227 <td>
2228 <table>
2229 <tr><th>
2230 <tr><td>BFLOAT16
2231 <tr><td>FLOAT16
2232 <tr><td>FLOAT32
2233 <tr><td>QASYMMS8
2234 <tr><td>QASYMMU8
2235 <tr><td>QSYMMS16
2236 </table>
2237<tr>
2238 <td>CpuAcc
2239 <td>
2240 <ul>
2241 <li>All
2242 </ul>
2243 <td>
2244 <table>
2245 <tr><th>
2246 <tr><td>QASYMMU8
2247 <tr><td>QASYMMS8
2248 <tr><td>FLOAT16
2249 <tr><td>FLOAT32
2250 </table>
2251<tr>
2252 <td>GpuAcc
2253 <td>
2254 <ul>
2255 <li>All
2256 </ul>
2257 <td>
2258 <table>
2259 <tr><th>
2260 <tr><td>QASYMMU8
2261 <tr><td>QASYMMS8
2262 <tr><td>FLOAT16
2263 <tr><td>FLOAT32
2264 </table>
2265<tr>
2266 <td rowspan="3">QLstmLayer
2267 <td rowspan="3" style="width:200px;"> Layer to perform quantized LSTM (Long Short-Term Memory) operation.
2268 <td rowspan="3">
2269 <ul>
2270 <li>ANEURALNETWORKS_QUANTIZED_LSTM
2271 <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
2272 </ul>
2273 <td>CpuRef
2274 <td>
2275 <ul>
2276 <li>All
2277 </ul>
2278 <td>
2279 <table>
2280 <tr><th>
2281 <tr><td>All
2282 </table>
2283<tr>
2284 <td>CpuAcc
2285 <td>
2286 <ul>
2287 <li>All
2288 </ul>
2289 <td>
2290 <table>
2291 <tr><th>
2292 <tr><td>QASYMMS8
2293 <tr><td>QASYMMU8
2294 <tr><td>SIGNED32
2295 <tr><td>QSYMMS16
2296 </table>
2297<tr>
2298 <td>GpuAcc
2299 <td>
2300 <ul>
2301 <li>All
2302 </ul>
2303 <td>
2304 <table>
2305 <tr><th>
2306 <tr><td>QASYMMS8
2307 <tr><td>QASYMMU8
2308 <tr><td>SIGNED32
2309 <tr><td>QSYMMS16
2310 </table>
2311<tr>
2312 <td rowspan="3">QuantizeLayer
2313 <td rowspan="3" style="width:200px;"> Layer to perform quantization operation.
2314 <td rowspan="3">
2315 <ul>
2316 <li>ANEURALNETWORKS_QUANTIZE
2317 </ul>
2318 <td>CpuRef
2319 <td>
2320 <ul>
2321 <li>All
2322 </ul>
2323 <td>
2324 <table>
2325 <tr><th>
2326 <tr><td>BFLOAT16
2327 <tr><td>FLOAT16
2328 <tr><td>FLOAT32
2329 <tr><td>QASYMMS8
2330 <tr><td>QASYMMU8
2331 <tr><td>QSYMMS8
2332 <tr><td>QSYMMS16
2333 </table>
2334<tr>
2335 <td>CpuAcc
2336 <td>
2337 <ul>
2338 <li>All
2339 </ul>
2340 <td>
2341 <table>
2342 <tr><th>
2343 <tr><td>QASYMMU8
2344 <tr><td>QASYMMS8
2345 <tr><td>QASYMM16
2346 <tr><td>FLOAT16
2347 <tr><td>FLOAT32
2348 </table>
2349<tr>
2350 <td>GpuAcc
2351 <td>
2352 <ul>
2353 <li>All
2354 </ul>
2355 <td>
2356 <table>
2357 <tr><th>
2358 <tr><td>QASYMMU8
2359 <tr><td>QASYMMS8
2360 <tr><td>QASYMM16
2361 <tr><td>FLOAT16
2362 <tr><td>FLOAT32
2363 </table>
2364<tr>
2365 <td rowspan="3">QuantizedLstmLayer
2366 <td rowspan="3" style="width:200px;"> Layer to perform quantized LSTM (Long Short-Term Memory) operation.
2367 <td rowspan="3">
2368 <ul>
2369 <li>ANEURALNETWORKS_QUANTIZED_LSTM
2370 <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
2371 </ul>
2372 <td>CpuRef
2373 <td>
2374 <ul>
2375 <li>All
2376 </ul>
2377 <td>
2378 <table>
2379 <tr><th>
2380 <tr><td>All
2381 </table>
2382<tr>
2383 <td>CpuAcc
2384 <td>
2385 <ul>
2386 <li>All
2387 </ul>
2388 <td>
2389 <table>
2390 <tr><th>
2391 <tr><td>SIGNED32
2392 <tr><td>QASYMMU8
2393 <tr><td>QSYMMS16
2394 </table>
2395<tr>
2396 <td>GpuAcc
2397 <td>
2398 <ul>
2399 <li>All
2400 </ul>
2401 <td>
2402 <table>
2403 <tr><th>
2404 <tr><td>SIGNED32
2405 <tr><td>QASYMMU8
2406 <tr><td>QSYMMS16
2407 </table>
2408<tr>
2409 <td rowspan="3">RankLayer
2410 <td rowspan="3" style="width:200px;"> Layer to perform a rank operation.
2411 <td rowspan="3">
2412 <ul>
2413 <li>ANEURALNETWORKS_RANK
2414 </ul>
2415 <td>CpuRef
2416 <td>
2417 <ul>
2418 <li>All
2419 </ul>
2420 <td>
2421 <table>
2422 <tr><th>
2423 <tr><td>All
2424 </table>
2425<tr>
2426 <td>CpuAcc
2427 <td>
2428 <ul>
2429 <li>All
2430 </ul>
2431 <td>
2432 <table>
2433 <tr><th>
2434 <tr><td>All
2435 </table>
2436<tr>
2437 <td>GpuAcc
2438 <td>
2439 <ul>
2440 <li>All
2441 </ul>
2442 <td>
2443 <table>
2444 <tr><th>
2445 <tr><td>All
2446 </table>
2447<tr>
2448 <td rowspan="3">ReduceLayer
2449 <td rowspan="3" style="width:200px;"> Layer to perform reduce with the following operations - ARG_IDX_MAX: Index of the max value - ARG_IDX_MIN: Index of the min value - MEAN_SUM: Mean of sum - PROD: Product - SUM_SQUARE: Sum of squares - SUM: Sum - MIN: Min - MAX: Max
2450 <td rowspan="3">
2451 <ul>
2452 <li>ANEURALNETWORKS_REDUCE_MAX
2453 <li>ANEURALNETWORKS_REDUCE_MIN
2454 <li>ANEURALNETWORKS_REDUCE_SUM
Teresa Charlin32b78702021-09-03 11:25:54 +01002455 <li>ANEURALNETWORKS_REDUCE_PROD
Sadik Armagan1a9c9f62021-08-05 09:25:15 +01002456 </ul>
2457 <td>CpuRef
2458 <td>
2459 <ul>
2460 <li>All
2461 </ul>
2462 <td>
2463 <table>
2464 <tr><th>
2465 <tr><td>BFLOAT16
2466 <tr><td>FLOAT16
2467 <tr><td>FLOAT32
2468 <tr><td>QASYMMS8
2469 <tr><td>QASYMMU8
2470 <tr><td>QSYMMS16
2471 <tr><td>SIGNED32
2472 </table>
2473<tr>
2474 <td>CpuAcc
2475 <td>
2476 <ul>
2477 <li>All
2478 </ul>
2479 <td>
2480 <table>
2481 <tr><th>
2482 <tr><td>QASYMMU8
2483 <tr><td>QASYMMS8
2484 <tr><td>FLOAT16
2485 <tr><td>FLOAT32
2486 <tr><td>SIGNED32
2487 </table>
2488<tr>
2489 <td>GpuAcc
2490 <td>
2491 <ul>
2492 <li>All
2493 </ul>
2494 <td>
2495 <table>
2496 <tr><th>
2497 <tr><td>QASYMMU8
2498 <tr><td>QASYMMS8
2499 <tr><td>FLOAT16
2500 <tr><td>FLOAT32
2501 <tr><td>SIGNED32
2502 </table>
2503<tr>
2504 <td rowspan="3">ReshapeLayer
2505 <td rowspan="3" style="width:200px;"> Layer to reshape a tensor.
2506 <td rowspan="3">
2507 <ul>
2508 <li>ANEURALNETWORKS_RESHAPE
2509 <li>ANEURALNETWORKS_SQUEEZE
2510 <li>ANEURALNETWORKS_EXPAND_DIMS
2511 </ul>
2512 <td>CpuRef
2513 <td>
2514 <ul>
2515 <li>All
2516 </ul>
2517 <td>
2518 <table>
2519 <tr><th>
2520 <tr><td>BFLOAT16
2521 <tr><td>FLOAT16
2522 <tr><td>FLOAT32
2523 <tr><td>QASYMMS8
2524 <tr><td>QASYMMU8
2525 <tr><td>QSYMMS16
2526 <tr><td>SIGNED32
2527 <tr><td>BOOLEAN
2528 </table>
2529<tr>
2530 <td>CpuAcc
2531 <td>
2532 <ul>
2533 <li>All
2534 </ul>
2535 <td>
2536 <table>
2537 <tr><th>
2538 <tr><td>All
2539 </table>
2540<tr>
2541 <td>GpuAcc
2542 <td>
2543 <ul>
2544 <li>All
2545 </ul>
2546 <td>
2547 <table>
2548 <tr><th>
2549 <tr><td>All
2550 </table>
2551<tr>
2552 <td rowspan="3">ResizeLayer
2553 <td rowspan="3" style="width:200px;"> Layer to perform resize of a tensor using one of the interpolation methods: - Bilinear - Nearest Neighbor.
2554 <td rowspan="3">
2555 <ul>
2556 <li>ANEURALNETWORKS_RESIZE_BILINEAR
2557 <li>ANEURALNETWORKS_RESIZE_NEAREST_NEIGHBOR
2558 </ul>
2559 <td>CpuRef
2560 <td>
2561 <ul>
2562 <li>All
2563 </ul>
2564 <td>
2565 <table>
2566 <tr><th>
2567 <tr><td>BFLOAT16
2568 <tr><td>FLOAT16
2569 <tr><td>FLOAT32
2570 <tr><td>QASYMMS8
2571 <tr><td>QASYMMU8
2572 <tr><td>QSYMMS16
2573 </table>
2574<tr>
2575 <td>CpuAcc
2576 <td>
2577 <ul>
2578 <li>NHWC
2579 <li>NCHW
2580 </ul>
2581 <td>
2582 <table>
2583 <tr><th>
2584 <tr><td>QASYMMU8
2585 <tr><td>QASYMMS8
2586 <tr><td>FLOAT16
2587 <tr><td>FLOAT32
2588 </table>
2589<tr>
2590 <td>GpuAcc
2591 <td>
2592 <ul>
2593 <li>NHWC
2594 <li>NCHW
2595 </ul>
2596 <td>
2597 <table>
2598 <tr><th>
2599 <tr><td>QASYMMU8
2600 <tr><td>QASYMMS8
2601 <tr><td>FLOAT16
2602 <tr><td>FLOAT32
2603 </table>
2604<tr>
2605 <td rowspan="3">RsqrtLayer
2606 <td rowspan="3" style="width:200px;"> Layer to perform Rsqrt operation.
2607 <td rowspan="3">
2608 <ul>
2609 <li>ANEURALNETWORKS_RSQRT
2610 </ul>
2611 <td>CpuRef
2612 <td>
2613 <ul>
2614 <li>All
2615 </ul>
2616 <td>
2617 <table>
2618 <tr><th>
2619 <tr><td>BFLOAT16
2620 <tr><td>FLOAT16
2621 <tr><td>FLOAT32
2622 <tr><td>QASYMMS8
2623 <tr><td>QASYMMU8
2624 <tr><td>QSYMMS16
2625 <tr><td>SIGNED32
2626 </table>
2627<tr>
2628 <td>CpuAcc
2629 <td>
2630 <ul>
2631 <li>All
2632 </ul>
2633 <td>
2634 <table>
2635 <tr><th>
2636 <tr><td>FLOAT16
2637 <tr><td>FLOAT32
2638 <tr><td>SIGNED32
2639 </table>
2640<tr>
2641 <td>GpuAcc
2642 <td>
2643 <ul>
2644 <li>All
2645 </ul>
2646 <td>
2647 <table>
2648 <tr><th>
2649 <tr><td>FLOAT16
2650 <tr><td>FLOAT32
2651 </table>
2652<tr>
2653 <td rowspan="3">ShapeLayer
2654 <td rowspan="3" style="width:200px;"> Layer to return the shape of the input tensor.
2655 <td rowspan="3">
2656 <ul>
2657 <li>N/A
2658 </ul>
2659 <td>CpuRef
2660 <td>
2661 <ul>
2662 <li>All
2663 </ul>
2664 <td>
2665 <table>
2666 <tr><th>
2667 <tr><td>All
2668 </table>
2669<tr>
2670 <td>CpuAcc
2671 <td>
2672 <ul>
2673 <li>All
2674 </ul>
2675 <td>
2676 <table>
2677 <tr><th>
2678 <tr><td>All
2679 </table>
2680<tr>
2681 <td>GpuAcc
2682 <td>
2683 <ul>
2684 <li>All
2685 </ul>
2686 <td>
2687 <table>
2688 <tr><th>
2689 <tr><td>All
2690 </table>
2691<tr>
2692 <td rowspan="3">SliceLayer
2693 <td rowspan="3" style="width:200px;"> Layer to perform tensor slicing.
2694 <td rowspan="3">
2695 <ul>
2696 <li>ANEURALNETWORKS_SLICE
2697 </ul>
2698 <td>CpuRef
2699 <td>
2700 <ul>
2701 <li>All
2702 </ul>
2703 <td>
2704 <table>
2705 <tr><th>
2706 <tr><td>BFLOAT16
2707 <tr><td>FLOAT32
2708 <tr><td>QASYMMS8
2709 <tr><td>QASYMMU8
2710 <tr><td>QSYMMS16
2711 </table>
2712<tr>
2713 <td>CpuAcc
2714 <td>
2715 <ul>
2716 <li>All
2717 </ul>
2718 <td>
2719 <table>
2720 <tr><th>
2721 <tr><td>All
2722 </table>
2723<tr>
2724 <td>GpuAcc
2725 <td>
2726 <ul>
2727 <li>All
2728 </ul>
2729 <td>
2730 <table>
2731 <tr><th>
2732 <tr><td>All
2733 </table>
2734<tr>
2735 <td rowspan="3">SoftmaxLayer
2736 <td rowspan="3" style="width:200px;"> Layer to perform softmax, log-softmax operation over the specified axis.
2737 <td rowspan="3">
2738 <ul>
2739 <li>ANEURALNETWORKS_LOG_SOFTMAX
2740 <li>ANEURALNETWORKS_SOFTMAX
2741 </ul>
2742 <td>CpuRef
2743 <td>
2744 <ul>
2745 <li>All
2746 </ul>
2747 <td>
2748 <table>
2749 <tr><th>
2750 <tr><td>BFLOAT16
2751 <tr><td>FLOAT16
2752 <tr><td>FLOAT32
2753 <tr><td>QASYMMS8
2754 <tr><td>QASYMMU8
2755 <tr><td>QSYMMS8
2756 <tr><td>QSYMMS16
2757 </table>
2758<tr>
2759 <td>CpuAcc
2760 <td>
2761 <ul>
2762 <li>All
2763 </ul>
2764 <td>
2765 <table>
2766 <tr><th>
2767 <tr><td>QASYMMU8
2768 <tr><td>QASYMMS8
2769 <tr><td>FLOAT16
2770 <tr><td>FLOAT32
2771 </table>
2772<tr>
2773 <td>GpuAcc
2774 <td>
2775 <ul>
2776 <li>All
2777 </ul>
2778 <td>
2779 <table>
2780 <tr><th>
2781 <tr><td>QASYMMU8
2782 <tr><td>QASYMMS8
2783 <tr><td>FLOAT16
2784 <tr><td>FLOAT32
2785 </table>
2786<tr>
2787 <td rowspan="3">SpaceToBatchNdLayer
2788 <td rowspan="3" style="width:200px;"> Layer to divide spatial dimensions of the tensor into a grid of blocks and interleaves these blocks with the batch dimension.
2789 <td rowspan="3">
2790 <ul>
2791 <li>ANEURALNETWORKS_SPACE_TO_BATCH_ND
2792 </ul>
2793 <td>CpuRef
2794 <td>
2795 <ul>
2796 <li>All
2797 </ul>
2798 <td>
2799 <table>
2800 <tr><th>
2801 <tr><td>BFLOAT16
2802 <tr><td>FLOAT16
2803 <tr><td>FLOAT32
2804 <tr><td>QASYMMS8
2805 <tr><td>QASYMMU8
2806 <tr><td>QSYMMS16
2807 </table>
2808<tr>
2809 <td>CpuAcc
2810 <td>
2811 <ul>
2812 <li>NHWC
2813 <li>NCHW
2814 </ul>
2815 <td>
2816 <table>
2817 <tr><th>
2818 <tr><td>All
2819 </table>
2820<tr>
2821 <td>GpuAcc
2822 <td>
2823 <ul>
2824 <li>NHWC
2825 <li>NCHW
2826 </ul>
2827 <td>
2828 <table>
2829 <tr><th>
2830 <tr><td>All
2831 </table>
2832<tr>
2833 <td rowspan="3">SpaceToDepthLayer
2834 <td rowspan="3" style="width:200px;"> Layer to rearrange blocks of spatial data into depth.
2835 <td rowspan="3">
2836 <ul>
2837 <li>ANEURALNETWORKS_SPACE_TO_DEPTH
2838 </ul>
2839 <td>CpuRef
2840 <td>
2841 <ul>
2842 <li>All
2843 </ul>
2844 <td>
2845 <table>
2846 <tr><th>
2847 <tr><td>BFLOAT16
2848 <tr><td>FLOAT16
2849 <tr><td>FLOAT32
2850 <tr><td>QASYMMS8
2851 <tr><td>QASYMMU8
2852 <tr><td>QSYMMS16
2853 </table>
2854<tr>
2855 <td>CpuAcc
2856 <td>
2857 <ul>
2858 <li>NHWC
2859 <li>NCHW
2860 </ul>
2861 <td>
2862 <table>
2863 <tr><th>
2864 <tr><td>All
2865 </table>
2866<tr>
2867 <td>GpuAcc
2868 <td>
2869 <ul>
2870 <li>NHWC
2871 <li>NCHW
2872 </ul>
2873 <td>
2874 <table>
2875 <tr><th>
2876 <tr><td>All
2877 </table>
2878<tr>
2879 <td rowspan="3">SplitterLayer
2880 <td rowspan="3" style="width:200px;"> Layer to split a tensor along a given axis.
2881 <td rowspan="3">
2882 <ul>
2883 <li>ANEURALNETWORKS_SPLIT
2884 </ul>
2885 <td>CpuRef
2886 <td>
2887 <ul>
2888 <li>All
2889 </ul>
2890 <td>
2891 <table>
2892 <tr><th>
2893 <tr><td>BFLOAT16
2894 <tr><td>FLOAT16
2895 <tr><td>FLOAT32
2896 <tr><td>QASYMMS8
2897 <tr><td>QASYMMU8
2898 <tr><td>QSYMMS16
2899 </table>
2900<tr>
2901 <td>CpuAcc
2902 <td>
2903 <ul>
2904 <li>All
2905 </ul>
2906 <td>
2907 <table>
2908 <tr><th>
2909 <tr><td>All
2910 </table>
2911<tr>
2912 <td>GpuAcc
2913 <td>
2914 <ul>
2915 <li>All
2916 </ul>
2917 <td>
2918 <table>
2919 <tr><th>
2920 <tr><td>All
2921 </table>
2922<tr>
2923 <td rowspan="3">StackLayer
2924 <td rowspan="3" style="width:200px;"> Layer to stack tensors along an axis.
2925 <td rowspan="3">
2926 <ul>
2927 <li>N/A
2928 </ul>
2929 <td>CpuRef
2930 <td>
2931 <ul>
2932 <li>All
2933 </ul>
2934 <td>
2935 <table>
2936 <tr><th>
2937 <tr><td>BFLOAT16
2938 <tr><td>FLOAT16
2939 <tr><td>FLOAT32
2940 <tr><td>QASYMMS8
2941 <tr><td>QASYMMU8
2942 <tr><td>QSYMMS16
2943 </table>
2944<tr>
2945 <td>CpuAcc
2946 <td>
2947 <ul>
2948 <li>All
2949 </ul>
2950 <td>
2951 <table>
2952 <tr><th>
2953 <tr><td>All
2954 </table>
2955<tr>
2956 <td>GpuAcc
2957 <td>
2958 <ul>
2959 <li>All
2960 </ul>
2961 <td>
2962 <table>
2963 <tr><th>
2964 <tr><td>All
2965 </table>
2966<tr>
2967 <td rowspan="1">StandInLayer
2968 <td rowspan="1" style="width:200px;"> A layer to represent "unknown" or "unsupported" operations in the input graph. It has a configurable number of input and output slots and an optional name.
2969 <td rowspan="1">
2970 <ul>
2971 <li>N/A
2972 </ul>
2973 <td>N/A
2974 <td>N/A
2975 <td>N/A
2976<tr>
2977 <td rowspan="3">StridedSliceLayer
2978 <td rowspan="3" style="width:200px;"> Layer to extract a strided slice of a tensor.
2979 <td rowspan="3">
2980 <ul>
2981 <li>ANEURALNETWORKS_STRIDED_SLICE
2982 </ul>
2983 <td>CpuRef
2984 <td>
2985 <ul>
2986 <li>All
2987 </ul>
2988 <td>
2989 <table>
2990 <tr><th>
2991 <tr><td>BFLOAT16
2992 <tr><td>FLOAT32
2993 <tr><td>QASYMMS8
2994 <tr><td>QASYMMU8
2995 <tr><td>QSYMMS16
2996 </table>
2997<tr>
2998 <td>CpuAcc
2999 <td>
3000 <ul>
3001 <li>All
3002 </ul>
3003 <td>
3004 <table>
3005 <tr><th>
3006 <tr><td>All
3007 </table>
3008<tr>
3009 <td>GpuAcc
3010 <td>
3011 <ul>
3012 <li>All
3013 </ul>
3014 <td>
3015 <table>
3016 <tr><th>
3017 <tr><td>All
3018 </table>
3019<tr>
3020 <td rowspan="3">SubtractionLayer
3021 <td rowspan="3" style="width:200px;"> Layer to perform an elementwise subtract of 2 tensors.
3022 <td rowspan="3">
3023 <ul>
3024 <li>ANEURALNETWORKS_SUB
3025 </ul>
3026 <td>CpuRef
3027 <td>
3028 <ul>
3029 <li>All
3030 </ul>
3031 <td>
3032 <table>
3033 <tr><th>
3034 <tr><td>BFLOAT16
3035 <tr><td>FLOAT16
3036 <tr><td>FLOAT32
3037 <tr><td>QASYMMS8
3038 <tr><td>QASYMMU8
3039 <tr><td>QSYMMS16
3040 <tr><td>SIGNED32
3041 </table>
3042<tr>
3043 <td>CpuAcc
3044 <td>
3045 <ul>
3046 <li>All
3047 </ul>
3048 <td>
3049 <table>
3050 <tr><th>
3051 <tr><td>QASYMMU8
3052 <tr><td>QASYMMS8
3053 <tr><td>QSYMMS16
3054 <tr><td>SIGNED32
3055 <tr><td>FLOAT16
3056 <tr><td>FLOAT32
3057 </table>
3058<tr>
3059 <td>GpuAcc
3060 <td>
3061 <ul>
3062 <li>All
3063 </ul>
3064 <td>
3065 <table>
3066 <tr><th>
3067 <tr><td>QASYMMU8
3068 <tr><td>QASYMMS8
3069 <tr><td>QSYMMS16
3070 <tr><td>SIGNED32
3071 <tr><td>FLOAT16
3072 <tr><td>FLOAT32
3073 </table>
3074<tr>
3075 <td rowspan="3">TransposeConvolution2dLayer
3076 <td rowspan="3" style="width:200px;"> Layer to perform 2D transpose convolution (deconvolution) operation.
3077 <td rowspan="3">
3078 <ul>
3079 <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
3080 </ul>
3081 <td>CpuRef
3082 <td>
3083 <ul>
3084 <li>All
3085 </ul>
3086 <td>
3087 <table>
3088 <tr><th>
3089 <tr><td>BFLOAT16
3090 <tr><td>FLOAT16
3091 <tr><td>FLOAT32
3092 <tr><td>QASYMMS8
3093 <tr><td>QASYMMU8
3094 <tr><td>QSYMMS8
3095 <tr><td>QSYMMS16
3096 </table>
3097<tr>
3098 <td>CpuAcc
3099 <td>
3100 <ul>
3101 <li>NHWC
3102 <li>NCHW
3103 </ul>
3104 <td>
3105 <table>
3106 <tr><th>
3107 <tr><td>SIGNED32
3108 <tr><td>FLOAT16
3109 <tr><td>FLOAT32
3110 <tr><td>QASYMMU8
3111 <tr><td>QASYMMS8
3112 <tr><td>QUANTIZEDSYMM8PERAXIS
3113 </table>
3114<tr>
3115 <td>GpuAcc
3116 <td>
3117 <ul>
3118 <li>NHWC
3119 <li>NCHW
3120 </ul>
3121 <td>
3122 <table>
3123 <tr><th>
3124 <tr><td>SIGNED32
3125 <tr><td>FLOAT16
3126 <tr><td>FLOAT32
3127 <tr><td>QASYMMU8
3128 <tr><td>QASYMMS8
3129 <tr><td>QUANTIZEDSYMM8PERAXIS
3130 </table>
3131<tr>
3132 <td rowspan="3">TransposeLayer
3133 <td rowspan="3" style="width:200px;"> Layer to transpose a tensor.
3134 <td rowspan="3">
3135 <ul>
3136 <li>ANEURALNETWORKS_TRANSPOSE
3137 </ul>
3138 <td>CpuRef
3139 <td>
3140 <ul>
3141 <li>All
3142 </ul>
3143 <td>
3144 <table>
3145 <tr><th>
3146 <tr><td>BFLOAT16
3147 <tr><td>FLOAT16
3148 <tr><td>FLOAT32
3149 <tr><td>QASYMMS8
3150 <tr><td>QASYMMU8
3151 <tr><td>QSYMMS16
3152 </table>
3153<tr>
3154 <td>CpuAcc
3155 <td>
3156 <ul>
3157 <li>All
3158 </ul>
3159 <td>
3160 <table>
3161 <tr><th>
3162 <tr><td>All
3163 </table>
3164<tr>
3165 <td>GpuAcc
3166 <td>
3167 <ul>
3168 <li>All
3169 </ul>
3170 <td>
3171 <table>
3172 <tr><th>
3173 <tr><td>All
3174 </table>
3175<tr>
3176 <td rowspan="3">UnidirectionalSquenceLstmLayer
Narumol Prangnawaratbd575b22021-08-31 16:53:54 +01003177 <td rowspan="3" style="width:200px;"> Layer to perform unidirectional sequence LSTM operation.
Sadik Armagan1a9c9f62021-08-05 09:25:15 +01003178 <td rowspan="3">
3179 <ul>
3180 <li>ANEURALNETWORKS_UNIDIRECTIONAL_SEQUENCE_LSTM
3181 </ul>
3182 <td>CpuRef
3183 <td>
3184 <ul>
3185 <li>All
3186 </ul>
3187 <td>
3188 <table>
Narumol Prangnawaratbd575b22021-08-31 16:53:54 +01003189 <tr><th>Input Types
3190 <tr><td>FLOAT32
Sadik Armagan1a9c9f62021-08-05 09:25:15 +01003191 </table>
Narumol Prangnawaratbd575b22021-08-31 16:53:54 +01003192 <table>
3193 <tr><th>Weight Types
3194 <tr><td>FLOAT32
3195 <tr><td>QASYMMS8
3196 </table>
Sadik Armagan1a9c9f62021-08-05 09:25:15 +01003197<tr>
3198 <td rowspan="3">UnmapLayer
3199 <td rowspan="3" style="width:200px;"> Layer to perform unmap operation on tensor.
3200 <td rowspan="3">
3201 <ul>
3202 <li>N/A
3203 </ul>
3204 <td>CpuRef
3205 <td>
3206 <ul>
3207 <li>All
3208 </ul>
3209 <td>
3210 <table>
3211 <tr><th>
3212 <tr><td>All
3213 </table>
3214<tr>
3215 <td>CpuAcc
3216 <td>
3217 <ul>
3218 <li>NHWC
3219 <li>NCHW
3220 </ul>
3221 <td>
3222 <table>
3223 <tr><th>
3224 <tr><td>All
3225 </table>
3226<tr>
3227 <td>GpuAcc
3228 <td>
3229 <ul>
3230 <li>NHWC
3231 <li>NCHW
3232 </ul>
3233 <td>
3234 <table>
3235 <tr><th>
3236 <tr><td>All
3237 </table>
3238</table>
3239
3240*/
3241} // namespace