alpaca-lora
Language model tuning
Tuning a large language model on consumer hardware using low-rank adaptation
Instruct-tune LLaMA on consumer hardware
19k stars
154 watching
2k forks
Language: Jupyter Notebook
last commit: 4 months ago
Linked from 2 awesome lists
Related projects:
Repository | Description | Stars |
---|---|---|
phoebussi/alpaca-cot | Provides a unified interface for fine-tuning large language models with parameter-efficient methods and instruction collection data | 2,619 |
antimatter15/alpaca.cpp | A locally run Instruction-Tuned Chat-Style LLM project that combines foundation models and fine-tuning to create a chat interface. | 10,250 |
tatsu-lab/stanford_alpaca | Develops an instruction-following LLaMA model for research use only, with the goal of fine-tuning and releasing it under specific licenses and restrictions. | 29,554 |
facico/chinese-vicuna | An instruction-following Chinese LLaMA-based model project aimed at training and fine-tuning models on specific hardware configurations for efficient deployment. | 4,142 |
hiyouga/llama-factory | A unified platform for fine-tuning multiple large language models with various training approaches and methods | 34,436 |
lightning-ai/lit-llama | An implementation of a large language model using the nanoGPT architecture | 5,993 |
microsoft/lora | A method to adapt large language models by reducing their parameter count using low-rank adaptation matrices | 10,752 |
haotian-liu/llava | A system that uses large language and vision models to generate and process visual instructions | 20,359 |
meta-llama/llama-recipes | Provides tools and examples for fine-tuning the Meta Llama model and building applications with it | 15,288 |
opengvlab/llama-adapter | An implementation of a method for fine-tuning language models to follow instructions with high efficiency and accuracy | 5,754 |
liguodongiot/llm-action | A comprehensive resource sharing project focused on large language model (LLM) engineering and applications, covering various aspects from training to inference and deployment. | 10,677 |
ggerganov/llama.cpp | Enables efficient inference of large language models using optimized C/C++ implementations and various backend frameworks | 68,190 |
lyogavin/airllm | A Python library that optimizes inference memory usage for large language models on limited GPU resources. | 5,259 |
lc1332/chinese-alpaca-lora | Develops and maintains a Chinese language model finetuned on LLaMA, used for text generation and summarization tasks. | 711 |
alpha-vllm/llama2-accessory | An open-source toolkit for pretraining and fine-tuning large language models | 2,722 |