GAU-alpha

Transformer model

An implementation of a Gated Attention Unit-based Transformer model for natural language processing tasks

基于Gated Attention Unit的Transformer模型(尝鲜版)

GitHub

96 stars
4 watching
9 forks
Language: Python
last commit: over 1 year ago

Related projects:

Repository Description Stars
zhuiyitechnology/roformer An enhanced transformer model with improved relative position embeddings for natural language processing tasks 819
openai/finetune-transformer-lm This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. 2,160
zhuiyitechnology/pretrained-models A collection of pre-trained language models for natural language processing tasks 987
german-nlp-group/german-transformer-training Trains German transformer models to improve language understanding 23
01-ai/yi A series of large language models trained from scratch to excel in multiple NLP tasks 7,719
tongjilibo/bert4torch An implementation of transformer models in PyTorch for natural language processing tasks 1,241
zhuiyitechnology/roformer-v2 A faster and more effective text processing model based on the RoFormer architecture 149
fastnlp/cpt A pre-trained transformer model for natural language understanding and generation tasks in Chinese 481
ymcui/chinese-xlnet Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture 1,653
zhuiyitechnology/wobert A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text 458
shawn-ieitsystems/yuan-1.0 Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing 591
leviswind/pytorch-transformer Implementation of a transformer-based translation model in PyTorch 239
huggingface/pytorch-openai-transformer-lm Implementing OpenAI's transformer language model in PyTorch with pre-trained weights and fine-tuning capabilities 1,511
yunwentechnology/unilm This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. 438
thudm/chinese-transformer-xl A pre-trained Chinese language model based on the Transformer-XL architecture. 218