gpt-2
Language model repository
A repository providing code and models for research into language modeling and multitask learning
Code for the paper "Language Models are Unsupervised Multitask Learners"
23k stars
631 watching
6k forks
Language: Python
last commit: 8 months ago paper
Related projects:
Repository | Description | Stars |
---|---|---|
| A tool for retraining and fine-tuning the OpenAI GPT-2 text generation model on new datasets. | 3,398 |
| A collection of GPT-2 model outputs for research on detection and biases. | 1,948 |
| Guides software developers on how to effectively use and build systems around Large Language Models like GPT-4. | 8,487 |
| A minimal PyTorch implementation of a transformer-based language model | 20,474 |
| Provides a framework for training large-scale language models on GPUs with advanced features and optimizations. | 6,997 |
| Compiles and organizes key papers on pre-trained language models, providing a resource for developers and researchers. | 3,331 |
| A PyTorch-based framework for training and sampling consistency models in image generation | 6,199 |
| An explanation of key concepts and advancements in the field of Machine Learning | 7,352 |
| This project generates instruction-following data using GPT-4 to fine-tune large language models for real-world tasks. | 4,244 |
| Fine-tuned language models trained on mixed-quality data | 5,273 |
| An open-source toolkit for training and deploying large-scale AI models on various downstream tasks with multi-modality | 3,840 |
| This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,167 |
| A curated list of papers on prompt-based tuning for pre-trained language models, providing insights and advancements in the field. | 4,112 |
| An interactive code generation and execution tool using AI models | 3,567 |
| Provides pre-trained language models and tools for fine-tuning and evaluation | 439 |