Alpaca-CoT
LLM fine-tuner
Provides a unified interface for fine-tuning large language models with parameter-efficient methods and instruction collection data
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
3k stars
36 watching
246 forks
Language: Jupyter Notebook
last commit: 12 months ago
Linked from 1 awesome list
alpacachatglmchatgptcotinstruction-tuningllamallmloramossp-tuningparameter-efficientpytorchtabultabular-datatabular-model
Related projects:
Repository | Description | Stars |
---|---|---|
tloen/alpaca-lora | Tuning a large language model on consumer hardware using low-rank adaptation | 18,651 |
tatsu-lab/stanford_alpaca | Develops an instruction-following LLaMA model for research use only, with the goal of fine-tuning and releasing it under specific licenses and restrictions. | 29,554 |
zjh-819/llmdatahub | A curated collection of high-quality datasets for training large language models. | 2,635 |
instruction-tuning-with-gpt-4/gpt-4-llm | This project generates instruction-following data using GPT-4 to fine-tune large language models for real-world tasks. | 4,210 |
lc1332/chinese-alpaca-lora | Develops and maintains a Chinese language model finetuned on LLaMA, used for text generation and summarization tasks. | 711 |
mooler0410/llmspracticalguide | A curated list of resources to help developers navigate the landscape of large language models and their applications in NLP | 9,489 |
hiyouga/llama-factory | A unified platform for fine-tuning multiple large language models with various training approaches and methods | 34,436 |
antimatter15/alpaca.cpp | A locally run Instruction-Tuned Chat-Style LLM project that combines foundation models and fine-tuning to create a chat interface. | 10,250 |
mlabonne/llm-course | A comprehensive course and resource package on building and deploying Large Language Models (LLMs) | 39,120 |
microsoft/lora | A method to adapt large language models by reducing their parameter count using low-rank adaptation matrices | 10,752 |
opengvlab/llama-adapter | An implementation of a method for fine-tuning language models to follow instructions with high efficiency and accuracy | 5,754 |
haotian-liu/llava | A system that uses large language and vision models to generate and process visual instructions | 20,232 |
facico/chinese-vicuna | An instruction-following Chinese LLaMA-based model project aimed at training and fine-tuning models on specific hardware configurations for efficient deployment. | 4,142 |
alpha-vllm/llama2-accessory | An open-source toolkit for pretraining and fine-tuning large language models | 2,720 |
lightning-ai/lit-llama | An implementation of a large language model using the nanoGPT architecture | 5,993 |