attention-is-all-you-need-pytorch
Sequence translator
An implementation of the Transformer model in PyTorch, a deep learning framework for sequence-to-sequence tasks like language translation.
A PyTorch implementation of the Transformer model in "Attention is All You Need".
9k stars
97 watching
2k forks
Language: Python
last commit: 7 months ago attentionattention-is-all-you-needdeep-learningnatural-language-processingnlppytorch
Related projects:
Repository | Description | Stars |
---|---|---|
leviswind/pytorch-transformer | Implementation of a transformer-based translation model in PyTorch | 239 |
diego999/pygat | An implementation of the Graph Attention Network model using PyTorch. | 2,914 |
kimiyoung/transformer-xl | Implementations of a neural network architecture for language modeling | 3,611 |
codertimo/bert-pytorch | An implementation of Google's 2018 BERT model in PyTorch, allowing pre-training and fine-tuning for natural language processing tasks | 6,222 |
javeywang/pyramid-attention-networks-pytorch | An implementation of a deep learning model using PyTorch for semantic segmentation tasks. | 235 |
pytorch/pytorch | A Python library providing tensors and dynamic neural networks with strong GPU acceleration | 83,959 |
lucidrains/reformer-pytorch | An implementation of Reformer, an efficient Transformer model for natural language processing tasks. | 2,120 |
dair-ai/ml-papers-explained | An explanation of key concepts and advancements in the field of Machine Learning | 7,315 |
maximumentropy/seq2seq-pytorch | An implementation of Sequence to Sequence models in PyTorch with various attention mechanisms and extensions for machine translation tasks. | 736 |
huggingface/transformers | A collection of pre-trained machine learning models for various natural language and computer vision tasks, enabling developers to fine-tune and deploy these models on their own projects. | 135,022 |
awslabs/sockeye | An open-source sequence-to-sequence framework for neural machine translation built on PyTorch. | 1,212 |
ukplab/sentence-transformers | Provides dense vector representations for text using transformer networks | 15,329 |
amdegroot/ssd.pytorch | An implementation of a deep learning-based object detection system in PyTorch. | 5,146 |
rdspring1/pytorch_gbw_lm | Trains a large-scale PyTorch language model on the 1-Billion Word dataset | 123 |
asyml/texar-pytorch | A toolkit providing easy-to-use machine learning modules and functionalities for natural language processing and text generation tasks | 745 |