PowerInfer
LLM inference engine
An efficient Large Language Model inference engine leveraging consumer-grade GPUs on PCs
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
8k stars
78 watching
415 forks
Language: C++
last commit: 6 months ago
Linked from 2 awesome lists
bamboo-7bfalconlarge-language-modelsllamallmllm-inferencelocal-inference
Related projects:
Repository | Description | Stars |
---|---|---|
| Generates large language model outputs in high-throughput mode on single GPUs | 9,236 |
| A toolkit for optimizing and serving large language models | 4,854 |
| A deep learning optimization library that simplifies distributed training and inference on modern computing hardware. | 35,863 |
| Optimizes large language model inference on limited GPU resources | 5,446 |
| An open-source implementation of a large bilingual language model pre-trained on vast amounts of text data. | 7,672 |
| A fast serving framework for large language models and vision language models. | 6,551 |
| An inference and serving engine for large language models | 31,982 |
| A toolkit for training and deploying large AI models in parallel on distributed computing infrastructure | 38,907 |
| A framework for deep learning inference on mobile devices | 4,949 |
| A suite of libraries implementing machine learning algorithms and mathematical primitives on NVIDIA GPUs | 4,292 |
| An open-source software project that enables efficient and accurate low-bit weight quantization for large language models. | 2,593 |
| A toolkit for deploying and serving Large Language Models (LLMs) for high-performance text generation | 9,456 |
| An open machine learning framework for building classical, deep, or hybrid models on various hardware platforms. | 5,555 |
| A PyTorch model definition and inference/sampling code repository for a powerful diffusion transformer with fine-grained Chinese understanding | 3,678 |
| A framework for efficient and fault-tolerant distributed training of large neural networks on multiple GPUs. | 3,299 |