libonnx

Inference engine

A lightweight onnx inference engine for embedded devices with hardware acceleration support

A lightweight, portable pure C99 onnx inference engine for embedded devices with hardware acceleration support.

GitHub

583 stars
28 watching
107 forks
Language: C
last commit: almost 2 years ago
Linked from 1 awesome list

aibaremetalcdedeep-neural-networksdeep-learningembeddedembedded-systemshardware-accelerationinferencelibrarylightweightmachine-learningneural-networkonnxportable

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
xilinx/finn Fast and scalable neural network inference framework for FPGAs. 747
emlearn/emlearn A machine learning inference engine designed to be portable and efficient for embedded systems with minimal dependencies. 511
alrevuelta/connxr An embedded device-friendly C ONNX runtime with zero dependencies 193
tpoisonooo/llama.onnx A project providing onnx models and tools for inference with LLaMa transformer model on various devices 352
microsoft/onnxruntime-inference-examples Repository providing examples for using ONNX Runtime (ORT) to perform machine learning inferencing. 1,212
jatinchowdhury18/rtneural Provides a lightweight neural network inferencing engine for real-time systems 601
microsoft/deepspeed-mii A Python library designed to accelerate model inference with high-throughput and low latency capabilities 1,898
kraiskil/onnx2c Generates C code from ONNX files for efficient neural network inference on microcontrollers 223
emergentorder/onnx-scala An API and backend for running ONNX models in Scala 3 using typeful, functional deep learning and classical machine learning. 138
xilinx/logicnets Designs and deploys neural networks integrated with Xilinx FPGAs for high-throughput applications 83
utensor/utensor A lightweight machine learning inference framework built on Tensorflow optimized for Arm targets. 1,729
xmartlabs/bender An abstraction layer for building and running neural networks on iOS using MetalPerformanceShaders and pre-trained models. 1,795
mit-han-lab/proxylessnas Direct neural architecture search on target task and hardware for efficient model deployment 1,425
xternalz/sdpoint A deep learning method for optimizing convolutional neural networks by reducing computational cost while improving regularization and inference efficiency. 18
albermax/innvestigate A toolbox to help understand neural networks' predictions by providing different analysis methods and a common interface. 1,265