libonnx
Inference engine
A lightweight onnx inference engine for embedded devices with hardware acceleration support
A lightweight, portable pure C99 onnx inference engine for embedded devices with hardware acceleration support.
583 stars
28 watching
107 forks
Language: C
last commit: almost 2 years ago
Linked from 1 awesome list
aibaremetalcdedeep-neural-networksdeep-learningembeddedembedded-systemshardware-accelerationinferencelibrarylightweightmachine-learningneural-networkonnxportable
Related projects:
Repository | Description | Stars |
---|---|---|
xilinx/finn | Fast and scalable neural network inference framework for FPGAs. | 747 |
emlearn/emlearn | A machine learning inference engine designed to be portable and efficient for embedded systems with minimal dependencies. | 511 |
alrevuelta/connxr | An embedded device-friendly C ONNX runtime with zero dependencies | 193 |
tpoisonooo/llama.onnx | A project providing onnx models and tools for inference with LLaMa transformer model on various devices | 352 |
microsoft/onnxruntime-inference-examples | Repository providing examples for using ONNX Runtime (ORT) to perform machine learning inferencing. | 1,212 |
jatinchowdhury18/rtneural | Provides a lightweight neural network inferencing engine for real-time systems | 601 |
microsoft/deepspeed-mii | A Python library designed to accelerate model inference with high-throughput and low latency capabilities | 1,898 |
kraiskil/onnx2c | Generates C code from ONNX files for efficient neural network inference on microcontrollers | 223 |
emergentorder/onnx-scala | An API and backend for running ONNX models in Scala 3 using typeful, functional deep learning and classical machine learning. | 138 |
xilinx/logicnets | Designs and deploys neural networks integrated with Xilinx FPGAs for high-throughput applications | 83 |
utensor/utensor | A lightweight machine learning inference framework built on Tensorflow optimized for Arm targets. | 1,729 |
xmartlabs/bender | An abstraction layer for building and running neural networks on iOS using MetalPerformanceShaders and pre-trained models. | 1,795 |
mit-han-lab/proxylessnas | Direct neural architecture search on target task and hardware for efficient model deployment | 1,425 |
xternalz/sdpoint | A deep learning method for optimizing convolutional neural networks by reducing computational cost while improving regularization and inference efficiency. | 18 |
albermax/innvestigate | A toolbox to help understand neural networks' predictions by providing different analysis methods and a common interface. | 1,265 |