minGPT
Transformer model
A minimal PyTorch implementation of a transformer-based language model
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
20k stars
259 watching
3k forks
Language: Python
last commit: 6 months ago Related projects:
Repository | Description | Stars |
---|---|---|
| A tool for retraining and fine-tuning the OpenAI GPT-2 text generation model on new datasets. | 3,398 |
| Provides a framework for training large-scale language models on GPUs with advanced features and optimizations. | 6,997 |
| A repository providing code and models for research into language modeling and multitask learning | 22,644 |
| A library for creating scalable and flexible Gaussian process models with ease | 3,605 |
| A high-performance inference engine for transformer models | 3,467 |
| A Rust implementation of a minimal Generative Pretrained Transformer architecture. | 845 |
| A tool for automating the process of generating and ranking effective prompts for AI models like GPT-4, GPT-3.5-Turbo, or Claude 3 Opus. | 9,411 |
| Provides pre-trained models and code for training vision transformers and mixers using JAX/Flax | 10,620 |
| A command-line tool using AI-powered language models to generate shell commands and code snippets | 9,933 |
| Research tool for training large transformer language models at scale | 1,926 |
| A collection of tools and scripts for training large transformer language models at scale | 1,342 |
| Analyzing knowledge development and evolution in large language models during training | 2,309 |
| This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,167 |
| Developing and pretraining a GPT-like Large Language Model from scratch | 35,405 |
| An AI-powered platform to experiment with software engineering tasks using natural language input. | 52,634 |