GAU-alpha
Transformer model
An implementation of a transformer-based NLP model utilizing gated attention units
基于Gated Attention Unit的Transformer模型(尝鲜版)
98 stars
4 watching
9 forks
Language: Python
last commit: almost 2 years ago Related projects:
Repository | Description | Stars |
---|---|---|
| An enhanced transformer model with improved relative position embeddings for natural language processing tasks | 837 |
| This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,167 |
| A collection of pre-trained language models for natural language processing tasks | 989 |
| Trains German transformer models to improve language understanding | 23 |
| A series of large language models trained from scratch to excel in multiple NLP tasks | 7,743 |
| An implementation of transformer models in PyTorch for natural language processing tasks | 1,257 |
| An improved version of a transformer-based language model with enhanced speed and accuracy through structural simplification and pre-training | 148 |
| A pre-trained transformer model for natural language understanding and generation tasks in Chinese | 482 |
| Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture | 1,652 |
| A Word-based Chinese BERT model trained on large-scale text data using pre-trained models as a foundation | 460 |
| Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing | 591 |
| Implementation of a transformer-based translation model in PyTorch | 240 |
| Implementing OpenAI's transformer language model in PyTorch with pre-trained weights and fine-tuning capabilities | 1,511 |
| This project provides pre-trained models and tools for natural language understanding (NLU) and generation (NLG) tasks in Chinese. | 439 |
| A pre-trained Chinese language model based on the Transformer-XL architecture. | 218 |