llama2.c
LLama inference engine
A minimalistic C implementation of the Llama 2 language model inference engine.
Inference Llama 2 in one file of pure C
17k stars
192 watching
2k forks
Language: C
last commit: 4 months ago
Linked from 1 awesome list
Related projects:
Repository | Description | Stars |
---|---|---|
ggerganov/llama.cpp | Enables efficient inference of large language models using optimized C/C++ implementations and various backend frameworks | 67,866 |
meta-llama/llama | A collection of tools and utilities for deploying, fine-tuning, and utilizing large language models. | 56,437 |
meta-llama/codellama | Provides inference code and tools for fine-tuning large language models, specifically designed for code generation tasks | 16,039 |
meta-llama/llama-recipes | Provides tools and examples for fine-tuning the Meta Llama model and building applications with it | 15,126 |
opengvlab/llama-adapter | An implementation of a method for fine-tuning language models to follow instructions with high efficiency and accuracy | 5,754 |
lyogavin/airllm | A Python library that optimizes inference memory usage for large language models on limited GPU resources. | 5,259 |
modeltc/lightllm | An LLM inference and serving framework providing a lightweight design, scalability, and high-speed performance for large language models. | 2,609 |
meta-llama/llama3 | Provides pre-trained and instruction-tuned Llama 3 language models and tools for loading and running inference | 27,138 |
lightning-ai/lit-llama | An implementation of a large language model using the nanoGPT architecture | 5,993 |
meta-llama/llama-stack | Provides a set of standardized APIs and tools to build generative AI applications | 4,591 |
rasbt/llms-from-scratch | Developing and pretraining a GPT-like Large Language Model from scratch | 32,908 |
alpha-vllm/llama2-accessory | An open-source toolkit for pretraining and fine-tuning large language models | 2,720 |
optimalscale/lmflow | A toolkit for finetuning large language models and providing efficient inference capabilities | 8,273 |
ericlbuehler/mistral.rs | A fast and flexible LLM inference platform supporting various models and devices | 4,466 |
scisharp/llamasharp | A C#/.NET library to efficiently run Large Language Models (LLMs) on local devices | 2,673 |