transformer-explainer
Model explainer
An interactive visualization tool to help users understand how large language models like GPT work
Transformer Explained Visually: Learn How LLM Transformer Models Work with Interactive Visualization
4k stars
35 watching
319 forks
Language: JavaScript
last commit: 3 months ago
Linked from 1 awesome list
deep-learninggenerative-aigptlangauge-modelllmvisualization
Related projects:
Repository | Description | Stars |
---|---|---|
| Provides pre-trained models and code for training vision transformers and mixers using JAX/Flax | 10,620 |
| An interactive visualization system to help non-experts learn about Convolutional Neural Networks (CNNs) by visualizing the learning process. | 8,204 |
| Implementations of a neural network architecture for language modeling | 3,619 |
| A collection of pre-trained machine learning models for various natural language and computer vision tasks, enabling developers to fine-tune and deploy these models on their own projects. | 136,357 |
| An open-source project that provides tools and utilities to understand how transformers are used in reinforcement learning tasks. | 75 |
| Provides tools and libraries for training and fine-tuning large language models using transformer architectures | 6,215 |
| This project provides code and model for improving language understanding through generative pre-training using a transformer-based architecture. | 2,167 |
| An explanation of key concepts and advancements in the field of Machine Learning | 7,352 |
| An implementation of deep learning transformer models in MATLAB | 209 |
| Converts popular transformer models to run on Android devices for efficient inference and generation tasks. | 396 |
| A pre-trained transformer model for natural language understanding and generation tasks in Chinese | 482 |
| An open-source tool that helps investigate specific behaviors of small language models by combining automated interpretability techniques with sparse autoencoders. | 4,047 |
| An implementation of a transformer-based NLP model utilizing gated attention units | 98 |
| A library designed to train transformer language models with reinforcement learning using various optimization techniques and fine-tuning methods. | 10,308 |
| Provides a unified interface to various transformer models implemented in C/C++ using GGML library | 1,823 |