metaseq
Transformer platform
A codebase for working with Open Pre-trained Transformers, enabling deployment and fine-tuning of transformer models on various platforms.
Repo for external large-scale work
Archived
7k stars
112 watching
728 forks
Language: Python
last commit: 9 months ago Related projects:
Repository | Description | Stars |
---|---|---|
huggingface/transformers | A collection of pre-trained machine learning models for various natural language and computer vision tasks, enabling developers to fine-tune and deploy these models on their own projects. | 136,357 |
google-research/vision_transformer | Provides pre-trained models and code for training vision transformers and mixers using JAX/Flax | 10,620 |
eleutherai/gpt-neox | Provides a framework for training large-scale language models on GPUs with advanced features and optimizations. | 6,997 |
nvidia/megatron-lm | A framework for training large language models using scalable and optimized GPU techniques | 10,804 |
google-research/big_vision | Supports large-scale vision model training on GPU machines or Google Cloud TPUs using scalable input pipelines. | 2,439 |
facebookresearch/fairseq | A toolkit for training custom sequence-to-sequence models for various NLP tasks | 30,675 |
huggingface/optimum | A toolkit for optimizing and accelerating the training and inference of machine learning models on various hardware platforms. | 2,618 |
huggingface/peft | An efficient method for fine-tuning large pre-trained models by adapting only a small fraction of their parameters | 16,699 |
huggingface/transformers.js | An open-source JavaScript library for running machine learning models in the browser without a server. | 12,363 |
opennmt/ctranslate2 | A high-performance inference engine for transformer models | 3,467 |
optimalscale/lmflow | A toolkit for fine-tuning and inferring large machine learning models | 8,312 |
facebookresearch/xformers | A library of optimized building blocks for deep learning models using the Transformer architecture. | 8,780 |
microsoft/megatron-deepspeed | Research tool for training large transformer language models at scale | 1,926 |
nvidia/fastertransformer | A high-performance transformer-based NLP component optimized for GPU acceleration and integration into various frameworks. | 5,937 |
google/big-bench | A benchmark designed to probe large language models and extrapolate their future capabilities through a diverse set of tasks. | 2,899 |