GAU-alpha
Transformer model
An implementation of a Gated Attention Unit-based Transformer model for natural language processing tasks
基于Gated Attention Unit的Transformer模型(尝鲜版)
96 stars
4 watching
9 forks
Language: Python
last commit: over 1 year ago Related projects:
Repository | Description | Stars |
---|---|---|
zhuiyitechnology/roformer | An enhanced transformer model with improved relative position embeddings for natural language processing tasks | 819 |
openai/finetune-transformer-lm | This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,160 |
zhuiyitechnology/pretrained-models | A collection of pre-trained language models for natural language processing tasks | 987 |
german-nlp-group/german-transformer-training | Trains German transformer models to improve language understanding | 23 |
01-ai/yi | A series of large language models trained from scratch to excel in multiple NLP tasks | 7,719 |
tongjilibo/bert4torch | An implementation of transformer models in PyTorch for natural language processing tasks | 1,241 |
zhuiyitechnology/roformer-v2 | A faster and more effective text processing model based on the RoFormer architecture | 149 |
fastnlp/cpt | A pre-trained transformer model for natural language understanding and generation tasks in Chinese | 481 |
ymcui/chinese-xlnet | Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture | 1,653 |
zhuiyitechnology/wobert | A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text | 458 |
shawn-ieitsystems/yuan-1.0 | Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing | 591 |
leviswind/pytorch-transformer | Implementation of a transformer-based translation model in PyTorch | 239 |
huggingface/pytorch-openai-transformer-lm | Implementing OpenAI's transformer language model in PyTorch with pre-trained weights and fine-tuning capabilities | 1,511 |
yunwentechnology/unilm | This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. | 438 |
thudm/chinese-transformer-xl | A pre-trained Chinese language model based on the Transformer-XL architecture. | 218 |