Gitiles
Code Review
Sign In
review.mlplatform.org
/
ml
/
ComputeLibrary
/
eff8d95991205e874091576e2d225f63246dd0bb
/
src
/
core
/
CL
/
cl_kernels
« Previous
f6402dd
COMPMID-834 Fix arm_compute_nightly_validation getting killed
by Michalis Spyrou
· 7 years ago
36a0a46
COMPMID-748 - Integrating optimized SGEMM for bifrost
by Gian Marco
· 7 years ago
7b4d547
COMPMID-816 - Optimizing CLGEMMLowpMatrixMultiplyCore - Part1
by Gian Marco
· 7 years ago
a1f7e33
COMPMID-841: Add CL QASYMM8 RELU Activation
by Michele Di Giorgio
· 7 years ago
5237e01
COMPMID-838 Implement CLPermute
by Michalis Spyrou
· 7 years ago
652bde5
COMPMID-674 - Create Google InceptionV3 example
by Georgios Pinitas
· 7 years ago
a0d1183
COMPMID-751 QASYMM8 ActivationLayer optimisation: don't requantize if not necessary
by Giorgio Arena
· 7 years ago
944d3f7
COMPMID-751 Processing 8 elements makes computation up to 80us faster on MobileNet QASYMM8 dwc layers
by Giorgio Arena
· 7 years ago
780db4e
COMPMID-471 Implement Deconvolution on OpenCL
by Michalis Spyrou
· 7 years ago
25f2368
COMPMID-589: Port HOGDescriptor to new validation
by John Richardson
· 7 years ago
1d08a31
COMPMID-765: Collapse execution window in CL kernels.
by Georgios Pinitas
· 7 years ago
5124be5
COMPMID-661: Convolution quantized (#32)
by Chunosov
· 7 years ago
fcd52fb
COMPMID-661: Vectorize im2col and add lws heuristics for convolution kernels #46
by Anthony Barbier
· 7 years ago
58c5794
COMPMID-706 - Add GEMMLowp output stage for scaling by a fixed point number
by Gian Marco
· 7 years ago
0162436
COMPMID-684: 2D In-Map normalization support for CL
by Georgios Pinitas
· 7 years ago
45bcc3a
COMPMID-661: QASYMM8 support for fully connected layer.
by Georgios Pinitas
· 7 years ago
6fdfaa8
COMPMID-713: Address failures in OCLGrind for CLDirectConvolution
by Georgios Pinitas
· 7 years ago
47b5603
COMPMID-712: OCLGrind CLSoftmaxLayer quantized failures
by Georgios Pinitas
· 7 years ago
c000fb8
COMPMID-714: Resolve failures in OCLGrind for CLPhase
by Georgios Pinitas
· 7 years ago
05288a2
COMPMID-697 - Rework GEMMLowp interface on OpenCL
by Gian Marco
· 7 years ago
f202e50
COMPMID-556 Improved indentation and error handling in format_doxygen.py
by Anthony Barbier
· 7 years ago
c809712
COMPMID-556 Add saturation to 8-bit activation. This prevents undefined overflow
by Rob Hughes
· 7 years ago
02bf80d
COMPMID-661: Fix scale border issue (#38)
by Daniil Efremov
· 7 years ago
3e80c7f
COMPMID-661: Optimize FC layer with 2 new Bifrost kernels and LWS tuning (#33)
by Anton Lokhmotov
· 7 years ago
d7295b7
COMPMID-661: Add QASYMM8 support (and basic tests) to CLDepthwiseConvolution3x3 kernel (#28)
by Dmitry Savenko
· 7 years ago
540d008
COMPMID-556: Fixes bias in CLDirectConvolutionLayer to be int32.
by Georgios Pinitas
· 7 years ago
7a49c79
COMPMID-661: issue# 23 Scale border fix (#26)
by Daniil Efremov
· 7 years ago
f1f3ebd
APPBROWSER-298, APPBROWSER-306: Reimplement the common code of compute shader
by Joel Liang
· 7 years ago
624b778
COMPMID-556: Fix CLNormalization issues.
by Georgios Pinitas
· 7 years ago
2f18579
COMPMID-661: Fix rounding in average pooling for uint8.
by Georgios Pinitas
· 7 years ago
f450caa
COMPMID-661: softmax-uint8 implementation (#16)
by Chunosov
· 7 years ago
7068f99
COMPMID-631: Merge branches/gles_compute branch
by Anthony Barbier
· 7 years ago
4df76c9
COMPMID-661: Fix beta in softmax_layer kernel
by Georgios Pinitas
· 7 years ago
af6204c
COMPMID-661: Add avgpool-uint8 support. Optimize avgpool-fp32 for Bifrost. (#13)
by Anton Lokhmotov
· 7 years ago
d6afedc
COMPMID-661: softmax-fp32 optimisation (#14)
by Chunosov
· 7 years ago
d621bca
COMPMID-661: directconv-uint8 (#20)
by Chunosov
· 7 years ago
6f31f8c
Allow running without cl_khr_fp16
by Matthew Bentham
· 7 years ago
adaae7e
COMPMID-647: Exclude padding pixels from averaging factor.
by Georgios Pinitas
· 7 years ago
0063380
IVGCVSW-619: Support for Cl u8 bounded Relu
by Michel Iwaniec
· 7 years ago
00f4d00
COMPMID-648: Fix initial value for MAX pooling for QS types in CL.
by Georgios Pinitas
· 7 years ago
81a26ad
COMPMID-643: Add bias to CLDepthwiseConvolution.
by Georgios Pinitas
· 7 years ago
96880cf
COMPMID-640: FullyConnectedLayer failures on both NEON/CL
by Georgios Pinitas
· 7 years ago
13fc22c
COMPMID-556: Fix CLPoolingLayer checks
by Georgios Pinitas
· 7 years ago
a1ed41f
IVGCVSW-601: support for asymetric padding in cl conv and depthwise conv
by Jaroslaw Rzepecki
· 7 years ago
48a60f9
IVGCVSW-632 CL support for Softmax beta parameter
by Pablo Palmier
· 7 years ago
744b5ed
COMPMID-606 - Fix for S8 failures
by Gian Marco Iodice
· 7 years ago
040bffe
COMPMID417 - Fix illegal scalar access in color_convert.cl
by Gian Marco Iodice
· 7 years ago
f01f9de
COMPMID-545 add CL printf support
by steniu01
· 7 years ago
349feef
COMPMID-417 - Added validation for FP16 CLBatchNormalizationLayer
by Gian Marco Iodice
· 7 years ago
83be745
COMPMID-424 Implemented reference implementation and tests for WarpAffine
by Isabella Gottardi
· 7 years ago
54f366a
COMPMID-417: Fix CL compiler warnings
by Moritz Pflanzer
· 7 years ago
4726fdf
COMPMID-541: Fix padding in CLMinMaxLocationKernel
by Moritz Pflanzer
· 7 years ago
cdf5145
COMPMID-515: L2 Pooling for FP32/FP16 in CL.
by Georgios Pinitas
· 7 years ago
f81652d
COMPMID-516 Increase tolerance rate of Scale, Conv, fully connected and GEMM
by steniu01
· 7 years ago
9fe4144
COMPMID-452 CL Generic Depthwise Convolution implementation.
by Giorgio Arena
· 7 years ago
bf17955
COMPMID-522 - Added support for GlobalPooling in CLPoolingLayer and CLFlattening for 3D tensor
by Gian Marco Iodice
· 7 years ago
52f8b39
COMPMID-417: Fix CLNonLinearFilter
by Georgios Pinitas
· 7 years ago
5ee66ea
COMPMID-462: Implement TensorReshape for NEON and CL.
by Georgios Pinitas
· 7 years ago
56dd726
COMPMID-448: Implement CL Quantization/Dequantization Layer.
by Michele Di Giorgio
· 7 years ago
1c8409d
COMPMID-477 - Optimized CLDirectConvolution1x1 for Bifrost
by Gian Marco Iodice
· 7 years ago
cfb6553
COMPMID-417 Fix ROIPooling
by SiCong Li
· 7 years ago
64ebe5b
COMPMID-519: Add support for Lower and Upper Bounded RELU for CL/NEON
by Georgios Pinitas
· 7 years ago
1fab09f
COMPMID-424 Implemented reference implementation, new output valid region and validation tests (NEON and CL) for Scale
by Isabella Gottardi
· 7 years ago
04f089c
COMPMID-476 L2 Normalization for CL
by Michalis Spyrou
· 7 years ago
3e36369
COMPMID-358 Implement OpenCL ROI Pooling
by SiCong Li
· 7 years ago
edfa9f4
COMPMID-477 - Optimized batched case in CLConvolutionLayer
by Gian Marco Iodice
· 7 years ago
93a690e
COMPMID-452 CL Depthwise Separable Convolution Layer kernel implementation, validation and benchmarking for 3x3xC depthwise filter and DataType::F32.
by Giorgio Arena
· 7 years ago
d60a6b9
COMPMID-477 - Optimized CLNormalizationLayer
by Gian Marco Iodice
· 7 years ago
0c7614f
COMPMID-431 Port OpenCL pooling layer to use fixed point
by steniu01
· 7 years ago
cb29283
COMPMID-477 - Optimizing Pooling 3x3 with stride_x <= 3 on OpenCL
by Gian Marco Iodice
· 7 years ago
1246b63
COMPMID-477 - Optimized Direct Convolution 3x3 and 5x5 (f32) for Bifrost.
by Gian Marco Iodice
· 7 years ago
409ee0a
COMPMID-417: Add in-place support for batch-normalization.
by Georgios Pinitas
· 7 years ago
1e5c157
COMPMID-450 Add YOLOV2 benchmark tests
by SiCong Li
· 7 years ago
db00668
COMPMID-478 Implemnt CL direct convolution 5x5
by steniu01
· 7 years ago
def665a
COMPMID-474 - Add support for QS8/QS16 DirectConvolution CL
by Michalis Spyrou
· 7 years ago
2eac5bd
COMPMID-417 - Fixed bug in CLCol2ImKernek related to the stride passed during the configuration
by Gian Marco Iodice
· 7 years ago
868e541
COMPMID-459 Collapse CL Im2col's higher dimensions
by steniu01
· 7 years ago
5cb4d6a
COMPMID-477 - Optimizing CLDirectConvolution 3x3 on OpenCL and added the auto configuration
by Gian Marco Iodice
· 7 years ago
d8e765b
COMPMID-472 : Implement Floor for CL and NEON.
by Georgios Pinitas
· 7 years ago
cfc6fe8
COMPMID-443 collapse higher dimension for CL col2im kernel
by steniu01
· 7 years ago
c51b72f
COMPMID-355 Implement CL DirectConvolution1x1
by SiCong Li
· 7 years ago
ef4b4ae
COMPMID-438: Add support for floating point Min-Max Location layer.
by Michele Di Giorgio
· 7 years ago
6c92834
COMPMID-413: Add support for QS8 and QS16 CLNormalizationLayer.
by Michele Di Giorgio
· 7 years ago
f9bae2e
COMPMID-417 - Bug Fix WarpPerspective kernel
by Isabella Gottardi
· 7 years ago
02dfb2c
COMPMID-457 Fix F16 NormalizationLayer CL kernel
by SiCong Li
· 7 years ago
a36ccf1
COMPMID-417: Fix CL F16 ActivationLayer
by Moritz Pflanzer
· 7 years ago
3a62324
COMPMID-455 - Optimizing CLIm2ColKernel
by Gian Marco Iodice
· 7 years ago
e49e266
COMPMID-415: Use half_float library for F16
by Moritz Pflanzer
· 7 years ago
27b386c
COMPMID-355 Implement 3x3 CL direct convolution
by steniu01
· 7 years ago
7281834
COMPMID-446: Add support for QS8/QS16 CL Arithmetic Add/Sub
by Michele Di Giorgio
· 7 years ago
172e570
COMPMID-425 Port CLBatchnormalization to support QS8/QS16
by Michalis Spyrou
· 7 years ago
579c049
COMPMID-417: Add Leaky RELU support for both NEON/CL.
by Georgios Pinitas
· 7 years ago
0d523cc
COMPMID-443 Change CLSoftMaxLayerKernel to use 3D tensor and collapse the higer dimension
by steniu01
· 7 years ago
00394ae
COMPMID-406: Port CLActivationLayer to use QS8/QS16.
by Georgios Pinitas
· 7 years ago
ac4e873
COMPMID-417: Port DepthConcatenate to QS8/QS16 for NEON/CL.
by Georgios Pinitas
· 7 years ago
9a7182e
COMPMID-443 Use 3D tensor for pixel multiply (Needed for Normalization Layer)
by Anthony Barbier
· 7 years ago
7ff47a3
COMPMID-443: Use 3D tensors for fill_border_image
by Anthony Barbier
· 7 years ago
da37e2f
COMPMID-431 Port CLDepthConvert to use 8-bit and 16-bit fixed point
by steniu01
· 7 years ago
0979675
COMPMID-429: Port CLSoftmaxLayer to QS16.
by Georgios Pinitas
· 7 years ago
7d323a6
COMPMID-440, COMPMID-441 - Port CLConvolutionLayer and CLFullyConnectedLayer to support 16 bit fixed point
by Gian Marco Iodice
· 7 years ago
Next »