PLMpapers
Language model research paper collection
Compiles and organizes key papers on pre-trained language models, providing a resource for developers and researchers.
Must-read Papers on pre-trained language models.
3k stars
148 watching
436 forks
last commit: over 2 years ago
Linked from 2 awesome lists
Related projects:
Repository | Description | Stars |
---|---|---|
| A curated list of resources to help developers navigate the landscape of large language models and their applications in NLP | 9,551 |
| A curated list of papers on prompt-based tuning for pre-trained language models, providing insights and advancements in the field. | 4,112 |
| A comprehensive course and resource package on building and deploying Large Language Models (LLMs) | 40,053 |
| A collection of pre-trained language models and optimization techniques for efficient natural language processing | 3,039 |
| A collection of pre-trained natural language processing models | 170 |
| An NLP project offering various text classification models and techniques for deep learning exploration | 7,881 |
| A general-purpose language model pre-trained with an autoregressive blank-filling objective and designed for various natural language understanding and generation tasks. | 3,207 |
| A curated collection of high-quality datasets for training large language models. | 2,708 |
| Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. | 806 |
| Guides software developers on how to effectively use and build systems around Large Language Models like GPT-4. | 8,487 |
| An explanation of key concepts and advancements in the field of Machine Learning | 7,352 |
| A repository of pre-trained language models for natural language processing tasks in Chinese | 977 |
| This project generates instruction-following data using GPT-4 to fine-tune large language models for real-world tasks. | 4,244 |
| A framework for training and serving large language models using JAX/Flax | 2,428 |
| Developing and pretraining a GPT-like Large Language Model from scratch | 35,405 |