roformer

Transformer model

An enhanced transformer model with improved relative position embeddings for natural language processing tasks

Rotary Transformer

GitHub

837 stars
8 watching
51 forks
Language: Python
last commit: almost 3 years ago

Related projects:

Repository Description Stars
zhuiyitechnology/roformer-v2 An improved version of a transformer-based language model with enhanced speed and accuracy through structural simplification and pre-training 148
zhuiyitechnology/roformer-sim An upgraded version of SimBERT with integrated retrieval and generation capabilities 441
zhuiyitechnology/gau-alpha An implementation of a transformer-based NLP model utilizing gated attention units 98
tongjilibo/bert4torch An implementation of transformer models in PyTorch for natural language processing tasks 1,257
thudm/chinese-transformer-xl A pre-trained Chinese language model based on the Transformer-XL architecture. 218
zhuiyitechnology/wobert A Word-based Chinese BERT model trained on large-scale text data using pre-trained models as a foundation 460
lucidrains/reformer-pytorch An implementation of Reformer, an efficient Transformer model for natural language processing tasks. 2,132
fastnlp/cpt A pre-trained transformer model for natural language understanding and generation tasks in Chinese 482
zhuiyitechnology/t5-pegasus Pretrained Chinese text generation model trained on large-scale data 558
leviswind/pytorch-transformer Implementation of a transformer-based translation model in PyTorch 240
yangjianxin1/ofa-chinese Transforms the OFA-Chinese model to work with the Hugging Face Transformers framework 123
chrislemke/sk-transformers Provides a collection of reusable data transformation tools 10
german-nlp-group/german-transformer-training Trains German transformer models to improve language understanding 23
robostack/jupyter-ros Provides ROS support for Jupyter notebooks to enable robotics developers to create interactive and dynamic visualizations of robot behavior. 592
rosinality/glow-pytorch A PyTorch implementation of Glow, a generative flow model using invertible 1x1 convolutions. 522