DecisionTransformerInterpretability
Transformer Explainability Tool
An open-source project that provides tools and utilities to understand how transformers are used in reinforcement learning tasks.
Interpreting how transformers simulate agents performing RL tasks
75 stars
4 watching
17 forks
Language: Jupyter Notebook
last commit: over 1 year ago mechanistic-interpretabilityreinforcement-learning
Related projects:
Repository | Description | Stars |
---|---|---|
| An R package for providing explanations of predictions made by black box classifiers. | 486 |
| An R package and workshop materials for explaining machine learning models using explainable AI techniques | 52 |
| A library for reverse engineering the algorithms learned by large language models from their weights | 1,653 |
| Provides a method to generate explanations for predictions made by any black box classifier. | 798 |
| A comprehensive guide to using the Transformers library for natural language processing tasks | 1,220 |
| Provides a unified interface to various transformer models implemented in C/C++ using GGML library | 1,823 |
| An implementation of transformer models in PyTorch for natural language processing tasks | 1,257 |
| An interpretability method that provides explanations for neural network predictions by highlighting high-level concepts relevant to classification tasks. | 633 |
| An eXplainable AI toolkit for evaluating and interpreting neural network explanations in various deep learning frameworks. | 567 |
| An implementation of deep neural network architectures, including Transformers, in Python. | 214 |
| An implementation of deep learning transformer models in MATLAB | 209 |
| Converts popular transformer models to run on Android devices for efficient inference and generation tasks. | 396 |
| A command-line tool for mass-replacing text patterns in files and renaming directories recursively. | 355 |
| A command-line utility for transforming strings using various encoding and decoding algorithms. | 966 |
| An implementation of Reformer, an efficient Transformer model for natural language processing tasks. | 2,132 |