attention-is-all-you-need-pytorch
Sequence translator
An implementation of the Transformer model in PyTorch, a deep learning framework for sequence-to-sequence tasks like language translation.
A PyTorch implementation of the Transformer model in "Attention is All You Need".
9k stars
97 watching
2k forks
Language: Python
last commit: 10 months ago attentionattention-is-all-you-needdeep-learningnatural-language-processingnlppytorch
Related projects:
Repository | Description | Stars |
---|---|---|
| Implementation of a transformer-based translation model in PyTorch | 240 |
| An implementation of the Graph Attention Network model using PyTorch. | 2,935 |
| Implementations of a neural network architecture for language modeling | 3,619 |
| An implementation of Google's 2018 BERT model in PyTorch, allowing pre-training and fine-tuning for natural language processing tasks | 6,251 |
| An implementation of a deep learning model using PyTorch for semantic segmentation tasks. | 237 |
| A Python library providing tensors and dynamic neural networks with strong GPU acceleration | 84,978 |
| An implementation of Reformer, an efficient Transformer model for natural language processing tasks. | 2,132 |
| An explanation of key concepts and advancements in the field of Machine Learning | 7,352 |
| An implementation of Sequence to Sequence models in PyTorch with various attention mechanisms and extensions for machine translation tasks. | 738 |
| A collection of pre-trained machine learning models for various natural language and computer vision tasks, enabling developers to fine-tune and deploy these models on their own projects. | 136,357 |
| An open-source sequence-to-sequence framework for neural machine translation built on PyTorch. | 1,212 |
| Provides dense vector representations for text using transformer networks | 15,556 |
| An implementation of a deep learning-based object detection system in PyTorch. | 5,160 |
| Trains a large-scale PyTorch language model on the 1-Billion Word dataset | 123 |
| A toolkit providing easy-to-use machine learning modules and functionalities for natural language processing and text generation tasks | 745 |