hashembedding

Hash embedding library

Software component providing efficient word representation using hash embeddings

Hash Embedding code for the paper "Hash Embeddings for Efficient Word Representations"

GitHub

42 stars
3 watching
17 forks
Language: Python
last commit: almost 7 years ago

Related projects:

Repository Description Stars
hslcy/vcwe This project provides code and corpora for creating word embeddings by considering the visual characteristics of words. 15
zhezhaoa/ngram2vec A toolkit for learning high-quality word and text representations from ngram co-occurrence statistics 846
uhh-lt/sensegram Tools and techniques for analyzing word meanings from word embeddings 212
vzhong/embeddings Provides fast and efficient word embeddings for natural language processing. 223
mitchellvanw/hashids Provides a simple way to encode and decode unique identifiers in web applications. 106
hassygo/charngram2vec A repository providing a re-implementation of character n-gram embeddings for pre-training in natural language processing tasks 23
danieldk/go2vec A package for reading and analyzing word embeddings from the word2vec format in Go. 56
galuhsahid/indonesian-word-embedding Demonstrates word embedding in Indonesian language using pre-trained Word2vec models 20
tca19/dict2vec A framework to learn word embeddings using lexical dictionaries 115
atgreen/cl-embeddings A Common Lisp library for generating word embeddings using neural network models. 8
dccuchile/spanish-word-embeddings A collection of precomputed word embeddings for the Spanish language, derived from different corpora and computational methods. 356
malllabiisc/wordgcn A deep learning model that generates word embeddings by predicting words based on their dependency context 290
greg7mdp/sparsepp A high-performance hash map implementation for C++ with low memory usage and efficient resizing 1,253
jwieting/paragram-word Trains word embeddings from a paraphrase database to represent semantic relationships between words. 30
bheinzerling/bpemb A collection of pre-trained subword embeddings in 275 languages, useful for natural language processing tasks. 1,184