lstm
LSTM model trainer
A package to train a neural network for language modeling using a variant of the Long Short Term Memory algorithm
664 stars
48 watching
283 forks
Language: Lua
last commit: over 6 years ago
Linked from 2 awesome lists
Related projects:
Repository | Description | Stars |
---|---|---|
wojzaremba/algorithm-learning | A framework to learn simple algorithms from examples by generating and visualizing intermediate solutions | 180 |
bobazooba/xllm | A tool for training and fine-tuning large language models using advanced techniques | 381 |
wojciechz/learning_to_execute | This software trains an LSTM-based neural network to predict Python code snippets | 480 |
huashiyiqike/lstm-matlab | An implementation of the Long Short-term Memory (LSTM) algorithm in MATLAB | 252 |
cwxcode/lstm-matlab | A Matlab implementation of a Long Short-Term Memory neural network architecture | 50 |
csuhan/onellm | A framework for training and fine-tuning multimodal language models on various data types | 588 |
yoonkim/lstm-char-cnn | An LSTM language model with a convolutional neural network over characters as input to predict words. | 826 |
openai/finetune-transformer-lm | This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,160 |
volcengine/vescale | A PyTorch-based framework for training large language models in parallel on multiple devices | 663 |
robromijnders/lstm_tsc | An implementation of a Long Short-term memory model for time series classification using Python and TensorFlow. | 408 |
dasguptar/treelstm.pytorch | An implementation of Tree-LSTM networks in PyTorch for semantic similarity tasks using tree structures | 550 |
gt-vision-lab/vqa_lstm_cnn | A Visual Question Answering model using a deeper LSTM and normalized CNN architecture. | 376 |
luogen1996/lavin | An open-source implementation of a vision-language instructed large language model | 508 |
taketwo/llm-ollama | A plugin providing access to local Ollama models using an HTTP API | 129 |
ermlab/politbert | Trains a language model using a RoBERTa architecture on high-quality Polish text data | 33 |