1. 814bddf Fuse pad layer with subsequent convolution layer by Gunes Bayir · 2 years, 10 months ago
  2. bc91297 Fix node fusion mutator by SiCongLi · 3 years, 1 month ago
  3. 68dd25f COMPMID-3637: Move utility headers from arm_compute to src by Sang-Hoon Park · 3 years, 9 months ago
  4. 047c6fc COMPMID-3791: Add support for all activation types in NodeFusionMutator by Gian Marco Iodice · 3 years, 9 months ago
  5. d9eaf61 COMPMID-3324: Adjusting capitalization of Arm copyright claim to reflect Arm preferred presentation by Michele Di Giorgio · 4 years ago
  6. 16dddd2 COMPMID-3381: Implement graph example for YoLo v3 output detector by Sheri Zhang · 4 years, 1 month ago
  7. 8b2a7d3 COMPMID-3101 Fuse activation with floating point elementwise operation layers in CL by Giorgio Arena · 4 years, 5 months ago
  8. 2e5fd63 COMPMID-3098 Fuse Relu and Bounded Relu with FullyConnected NEON by SiCongLi · 4 years, 4 months ago
  9. 1856ff7 COMPMID-3097 Fuse activation with fully connected layer CL by Giorgio Arena · 4 years, 5 months ago
  10. f4261ad COMPMID-2779: Add support for generating synthetic int8 graphs. by Georgios Pinitas · 4 years, 7 months ago
  11. 5dea19e COMPMID-2579: Fuse batch normalization with convolution and depthwise convolution at graph level on NEON by Gian Marco Iodice · 4 years, 8 months ago
  12. 04ea4e8 COMPMID-2581: Fuse batch normalization with convolution and depthwise convolution at graph level for OpenCL - FP16 by giuros01 · 4 years, 9 months ago
  13. 351bd13 compmid-2573: Investigate FP16 Winograd reference implementations by giuros01 · 4 years, 10 months ago
  14. f948b10 COMPMID-2582: Disable fuse batch normalization on OpenCL for FP16 by Gian Marco Iodice · 4 years, 11 months ago
  15. 169cda3 COMPMID-2055: Fusion of ConvolutionLayer with BatchNormalization at graph only for CL by Manuel Bottini · 5 years ago
  16. bffb41e COMPMID-2273: Fuse Batch Normalization with Depthwise Convolution layer at graph level (only for CL) by Manuel Bottini · 5 years ago
  17. 299fdd3 COMPMID-2177 Fix clang warnings by Michalis Spyrou · 5 years ago
  18. cadb368 COMPMID-1995: Fixed graph fusion mutator for float types. by Georgios Pinitas · 5 years ago
  19. 0ae5de9 COMPMID-1995: Prepare Graph to support different input/output quantization info by Isabella Gottardi · 5 years ago
  20. 749021a COMPMID-1995: Revert fusing convolution to batch norm due to performance regressions by giuros01 · 5 years ago
  21. acce504 COMPMID-1740: Fuse batch normalization with Convolution Layer at graph level by giuros01 · 5 years ago
  22. 1c32bf396 COMPMID-1451: Perform fusion before GroupConvolution unrolling by Georgios Pinitas · 6 years ago
  23. 60e9825 COMPMID-1451: Fuse activation in DepthwiseConvolution. by Georgios Pinitas · 6 years ago
  24. 08346e9 COMPMID-1451:Fuse RELU,LU_BOUNDED_RELU with requantization in NEGEMMConvolutionLayer. by Georgios Pinitas · 6 years ago
  25. 2a2db59 COMPMID-1505: Add native grouping support at graph level by Georgios Pinitas · 6 years ago
  26. 6f109bd COMPMID-1409: Disable BN + Act fusion in the graph if activation is not supported by Georgios Pinitas · 6 years ago
  27. d3a78ab COMPMID-1283: (GitHub issue) after convolution output data is zero by Georgios Pinitas · 6 years ago
  28. d9eb275 COMPMID-797: Switch to new graph. by Georgios Pinitas · 6 years ago[Renamed (94%) from src/graph2/mutators/NodeFusionMutator.cpp]
  29. d8734b5 COMPMID-793 : Add graph intermediate representation by Georgios Pinitas · 7 years ago