LLaVA-Plus-Codebase
Model trainer
A platform for training and deploying large language and vision models that can use tools to perform tasks
LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills
717 stars
12 watching
53 forks
Language: Python
last commit: about 1 year ago agentlarge-language-modelslarge-multimodal-modelsmultimodal-large-language-modelstool-use
Related projects:
Repository | Description | Stars |
---|---|---|
| A system designed to enable large multimodal models to understand arbitrary visual prompts | 302 |
| An all-in-one demo for interactive image processing and generation | 353 |
| This project presents an optimization technique for large-scale image models to reduce computational requirements while maintaining performance. | 106 |
| Transfers visual prompt generators across large language models to reduce training costs and enable customization of multimodal LLMs | 270 |
| A framework for training and fine-tuning multimodal language models on various data types | 601 |
| A tool for training and fine-tuning large language models using advanced techniques | 387 |
| A live training platform for large-scale deep learning models, allowing community participation and collaboration in model development and deployment. | 511 |
| Provides pre-trained language models and tools for fine-tuning and evaluation | 439 |
| Develops training methods to improve the politeness and natural flow of multi-modal Large Language Models | 63 |
| Debiasing techniques to minimize hallucinations in large visual language models | 75 |
| A guide to using pre-trained large language models in source code analysis and generation | 1,789 |
| A multimodal LLM designed to handle text-rich visual questions | 270 |
| This is an open-source project that proposes a novel method to train large-scale vision-language models with minimal resources and no fine-tuning required. | 94 |
| A PyTorch-based framework for training large language models in parallel on multiple devices | 679 |
| A research project aimed at building large language and vision models for biomedical applications with capabilities comparable to GPT-4. | 1,622 |