protein-sequence-embedding-iclr2019
Protein embedding model
Developing models to learn and represent protein sequences based on their structure
Source code for "Learning protein sequence embeddings using information from structure" - ICLR 2019
259 stars
11 watching
75 forks
Language: Python
last commit: over 3 years ago
Linked from 1 awesome list
deep-learninglanguage-modelprotein-embeddingprotein-modelingprotein-representation-learningprotein-sequenceprotein-structurepytorchrecurrent-neural-networks
Related projects:
Repository | Description | Stars |
---|---|---|
| Code for training universal paraphrastic sentence embeddings and models on semantic similarity tasks | 193 |
| A software framework for evaluating protein embeddings and benchmarking semi-supervised learning tasks in protein biology | 118 |
| Provides pre-trained protein embeddings and benchmarking tools for semi-supervised learning tasks in protein biology | 671 |
| A codebase for training and using models of sentence embeddings. | 33 |
| Provides training and testing code for a CNN-based sentence embedding model | 2 |
| This repository presents a novel neural network architecture and its applications in sequence modeling tasks such as language modeling and classification. | 472 |
| An implementation of a protein language model that uses prompts to learn from multi-level structural information in proteins. | 32 |
| Unsupervised word embeddings capture latent knowledge from materials science literature | 624 |
| Provides code and tools for learning joint text-video embeddings using the HowTo100M dataset | 254 |
| A PyTorch implementation of a representation learning framework for dynamic temporal networks | 362 |
| Pre-trained word and sentence embeddings for biomedical text analysis | 578 |
| Provides tools and frameworks for training sequence-to-sequence models using PyTorch | 523 |
| A sparsity-aware implementation of a deep learning algorithm for graph embedding with text information. | 74 |
| A Python framework for building deep learning models with optimized encoding layers and batch normalization. | 2,044 |
| Trains word embeddings from a paraphrase database to represent semantic relationships between words. | 30 |