Tengine is a lite, high performance, modular inference engine for embedded device
Dynamic graph segmentation
C++ API (experiment)
Python API (experiment)
support ARM-Mali GPU with ACL
support others GPU with Vulkan (experiment)
support fp16 inference with armv8.2 (experiment)
uint8 reference op (experiment)
mish activation op
Initial Tengine Lite release v0.1
Separate cpu operator implementation and the framework into two so. Add serializer for TFLite, and reference implementation on TFLite op. Add RNN/GRU/LSTM reference implementation
With the new API 2.0 and a few new features and bug fixes.
Android build to run ACL MSSD can use GPU to accelerate Android build with c++_shared instead of gnustl_shared
Support GPU fp16. Only works with ACL 18.05 More tensorflow model and onnx model support
This is a first version which implements many basic features for an inference engine