1. b75d624 Graph Fusion With Post Ops Fix by ramelg01 · 2 years, 7 months ago
  2. 1d4a320 Fix node fusion mutator ordering by SiCongLi · 2 years, 8 months ago
  3. c65023e Refine code for graph level fusion by Sheri Zhang · 2 years, 8 months ago
  4. fb22803 Add graph level convolution fusion with post operator by Sheri Zhang · 2 years, 8 months ago
  5. cc171f9 Provide tests for fusing pad layer in graph API by Gunes Bayir · 2 years, 10 months ago
  6. 814bddf Fuse pad layer with subsequent convolution layer by Gunes Bayir · 2 years, 10 months ago
  7. a387e27 Add in-place calculation support for CL elementwise arithmetic kernels by Sheri Zhang · 3 years ago
  8. bc4e311 Implement FP GPU depthwise convolution 1x1 kernel for in-place computation by SiCongLi · 3 years ago
  9. 1d35927 Add in-place computation for elementwise operations by Sheri Zhang · 3 years, 1 month ago
  10. bc91297 Fix node fusion mutator by SiCongLi · 3 years, 1 month ago
  11. f466d75 Add QASYMM8_SIGNED support to graph examples via graph mutator by SiCongLi · 3 years, 4 months ago
  12. 40f51a6 Update default C++ standard to C++14 by Georgios Pinitas · 3 years, 7 months ago
  13. 68dd25f COMPMID-3637: Move utility headers from arm_compute to src by Sang-Hoon Park · 3 years, 9 months ago
  14. 047c6fc COMPMID-3791: Add support for all activation types in NodeFusionMutator by Gian Marco Iodice · 3 years, 9 months ago
  15. d9eaf61 COMPMID-3324: Adjusting capitalization of Arm copyright claim to reflect Arm preferred presentation by Michele Di Giorgio · 4 years ago
  16. 294f6ff COMPMID-3324: Allow in-place mutation on nodes with separate multiple inputs by Michele Di Giorgio · 4 years ago
  17. 80feed5 COMPMID-3479: Perform in-place computations in NEElementwiseUnaryKernel by Manuel Bottini · 4 years, 1 month ago
  18. 3be0b8c COMPMID-3480: Enable in-place elementwise operations in graph API by Michele Di Giorgio · 4 years ago
  19. 1902383 Revert "COMPMID-3480: Perform in-place computations in NEArithmeticAdditionKernel" by Michele Di Giorgio · 4 years ago
  20. 16dddd2 COMPMID-3381: Implement graph example for YoLo v3 output detector by Sheri Zhang · 4 years, 1 month ago
  21. 4a61653 COMPMID-3480: Perform in-place computations in NEArithmeticAdditionKernel by Michele Di Giorgio · 4 years, 1 month ago
  22. 8b2a7d3 COMPMID-3101 Fuse activation with floating point elementwise operation layers in CL by Giorgio Arena · 4 years, 5 months ago
  23. 2e5fd63 COMPMID-3098 Fuse Relu and Bounded Relu with FullyConnected NEON by SiCongLi · 4 years, 4 months ago
  24. 758b5ba COMPMID-3069: Improve build time by splitting up ToolchainSupport.h by Matthew Bentham · 4 years, 4 months ago
  25. 1856ff7 COMPMID-3097 Fuse activation with fully connected layer CL by Giorgio Arena · 4 years, 5 months ago
  26. 6e9d0e0 COMPMID-2856 Add PrintLayer at graph level by Giorgio Arena · 4 years, 6 months ago
  27. f4261ad COMPMID-2779: Add support for generating synthetic int8 graphs. by Georgios Pinitas · 4 years, 7 months ago
  28. 5dea19e COMPMID-2579: Fuse batch normalization with convolution and depthwise convolution at graph level on NEON by Gian Marco Iodice · 4 years, 8 months ago
  29. 04ea4e8 COMPMID-2581: Fuse batch normalization with convolution and depthwise convolution at graph level for OpenCL - FP16 by giuros01 · 4 years, 9 months ago
  30. 351bd13 compmid-2573: Investigate FP16 Winograd reference implementations by giuros01 · 4 years, 10 months ago
  31. f948b10 COMPMID-2582: Disable fuse batch normalization on OpenCL for FP16 by Gian Marco Iodice · 4 years, 11 months ago
  32. 169cda3 COMPMID-2055: Fusion of ConvolutionLayer with BatchNormalization at graph only for CL by Manuel Bottini · 5 years ago
  33. 2ea3761 COMPMID-2336: Fix InPlaceMutator condition and add SaveNumpyAccessor by Isabella Gottardi · 5 years ago
  34. bffb41e COMPMID-2273: Fuse Batch Normalization with Depthwise Convolution layer at graph level (only for CL) by Manuel Bottini · 5 years ago
  35. 299fdd3 COMPMID-2177 Fix clang warnings by Michalis Spyrou · 5 years ago
  36. 9e4824c COMPMID-2111: ConcatenateLayer API should accept an index instead of an enum by Georgios Pinitas · 5 years ago
  37. cadb368 COMPMID-1995: Fixed graph fusion mutator for float types. by Georgios Pinitas · 5 years ago
  38. 0ae5de9 COMPMID-1995: Prepare Graph to support different input/output quantization info by Isabella Gottardi · 5 years ago
  39. 749021a COMPMID-1995: Revert fusing convolution to batch norm due to performance regressions by giuros01 · 5 years ago
  40. acce504 COMPMID-1740: Fuse batch normalization with Convolution Layer at graph level by giuros01 · 5 years ago
  41. 1c32bf396 COMPMID-1451: Perform fusion before GroupConvolution unrolling by Georgios Pinitas · 6 years ago
  42. 60e9825 COMPMID-1451: Fuse activation in DepthwiseConvolution. by Georgios Pinitas · 6 years ago
  43. 08346e9 COMPMID-1451:Fuse RELU,LU_BOUNDED_RELU with requantization in NEGEMMConvolutionLayer. by Georgios Pinitas · 6 years ago
  44. 890ad1b COMPMID-1246: Fix bug in handling backends that can't be loaded in the Graph API by Anthony Barbier · 6 years ago
  45. 2a2db59 COMPMID-1505: Add native grouping support at graph level by Georgios Pinitas · 6 years ago
  46. e222055 COMPMID-1367: Enable NHWC in graph examples by Georgios Pinitas · 6 years ago
  47. 6f109bd COMPMID-1409: Disable BN + Act fusion in the graph if activation is not supported by Georgios Pinitas · 6 years ago
  48. d3a78ab COMPMID-1283: (GitHub issue) after convolution output data is zero by Georgios Pinitas · 6 years ago
  49. cac13b1 COMPMID-1097: Port mobilenet to NHWC by Georgios Pinitas · 6 years ago
  50. d9eb275 COMPMID-797: Switch to new graph. by Georgios Pinitas · 6 years ago