TDB

Creator
Creator
Seonglae ChoSeonglae Cho
Created
Created
2024 Mar 9 11:37
Editor
Edited
Edited
2024 Oct 24 11:41

Transformer Debugger

Component can be an attention head or neuron, or autoencoder latent
Each node only exists in one forward/backward pass. If you modify the prompt and rerun, that would create different nodes
Residual stream stands for
Circuit is a set of nodes that work together to perform some behavior/reasoning.
TDB Usages
 
 
 

Terms

github.com
1. TDB Intro
In this video, I will walk you through Transformer Debugger, a tool developed to perform exploratory analyses on activations of Transformer language models. Similar to a Python debugger, Transformer Debugger allows you to step through language model outputs, trace important activations, and analyze upstream activations. I will explain how to use prompts, view model outputs, and interpret the node table.
1. TDB Intro
2. TDB neuron-viewer pages
In this video, I explain how the Transformer Debugger tool provides a prompt-centric view of important activations and model components in a Transformer model. I demonstrate how to navigate the tool and interpret the visualizations, such as the color map indicating attention strength. I also discuss the significance of specific attention heads and MLP neurons, highlighting their role in capturing patterns and making predictions.
2. TDB neuron-viewer pages
 
 
 

Recommendations