TransformerLens

Model decipherer

A library for reverse engineering the algorithms learned by large language models from their weights

A library for mechanistic interpretability of GPT-style language models

GitHub

2k stars
16 watching
315 forks
Language: Python
last commit: about 1 month ago

Related projects:

Repository Description Stars
matlab-deep-learning/transformer-models An implementation of deep learning transformer models in MATLAB 209
lucidrains/reformer-pytorch An implementation of Reformer, an efficient Transformer model for natural language processing tasks. 2,132
jlevy/repren A command-line tool for mass-replacing text patterns in files and renaming directories recursively. 355
openai/finetune-transformer-lm This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. 2,167
chrislemke/sk-transformers Provides a collection of reusable data transformation tools 10
leviswind/pytorch-transformer Implementation of a transformer-based translation model in PyTorch 240
feature-engine/feature_engine A Python library with multiple transformers to engineer and select features for use in machine learning models. 1,956
fastnlp/cpt A pre-trained transformer model for natural language understanding and generation tasks in Chinese 482
marella/ctransformers Provides a unified interface to various transformer models implemented in C/C++ using GGML library 1,823
nlgranger/seqtools A Python library to manipulate and transform indexable data 49
pylons/colander A library for serializing and deserializing data structures into strings, mappings, and lists while performing validation. 451
jbloomaus/decisiontransformerinterpretability An open-source project that provides tools and utilities to understand how transformers are used in reinforcement learning tasks. 75
microsoft/megatron-deepspeed Research tool for training large transformer language models at scale 1,926
bigscience-workshop/megatron-deepspeed A collection of tools and scripts for training large transformer language models at scale 1,342
neukg/techgpt A generative transformer model designed to process and generate text in various vertical domains, including computer science, finance, and more. 217