Skip-gram
Word embedding model
A Python implementation of a neural network model for learning word embeddings from text data
6 stars
1 watching
4 forks
Language: Python
last commit: over 4 years ago Related projects:
Repository | Description | Stars |
---|---|---|
| A PyTorch implementation of the skip-gram model for learning word embeddings. | 188 |
| A toolkit for learning high-quality word and text representations from ngram co-occurrence statistics | 848 |
| A project implementing a method to incorporate morphological information into word embeddings using a neural network model | 52 |
| A deep learning model that generates word embeddings by predicting words based on their dependency context | 291 |
| Provides training and testing code for a CNN-based sentence embedding model | 2 |
| Provides fast and efficient word embeddings for natural language processing. | 223 |
| An implementation of a word embedding model that uses character n-grams and achieves state-of-the-art results in multiple NLP tasks | 803 |
| A tool for training and using character n-gram based word and sentence embeddings in natural language processing. | 125 |
| Reproduces a SGNN's word projections preprocessing pipeline using word n-grams instead of skip-grams | 23 |
| Develops unified sentence embedding models for NLP tasks | 840 |
| This is a repository for a subword ELMo model pre-trained on a large corpus of text. | 12 |
| A codebase for training and using models of sentence embeddings. | 33 |
| A foundation model for single-cell biology tasks by leveraging large language model embeddings of gene descriptions. | 211 |
| Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
| A platform for designing and evaluating Graph Neural Networks (GNN) models | 1,738 |