New in This Fork:
- About This Repository - Fork lineage and feature history
- Installation
- Dedicated Feature Crosscoders (DFCs) - ⭐ NEW: Feature partitioning for model diffing
From Previous Forks & Original:
- Using Trained Dictionaries
- Training Your Own Dictionaries
- Downloading Pre-trained Dictionaries
- Dictionary Statistics
This repository is a fork of science-of-finetuning/crosscoder_learning, which is itself a fork of the original saprmarks/dictionary_learning.
Original: saprmarks/dictionary_learning
- Foundation for dictionary learning via sparse autoencoders (SAEs) on neural network activations
- Multiple SAE architectures: Standard, Gated, Top-K
- Training infrastructure with neuron resampling and LR warmup
- Integration with
nnsightfor activation extraction
Fork 1: science-of-finetuning/crosscoder_learning
pipinstallable packageCrossCoderclass for comparing model variants (Anthropic paper)BatchTopKCrossCoderwith batch-wise top-k sparsity (paper)- Activation caching system (
cache.py) for efficient training - Training scripts for CrossCoders with pre-computed activations (
scripts/train_crosscoder.py) - HuggingFace Hub integration (
push_to_hub(),from_pretrained(..., from_hub=True))
Fork 2: This Repository (superkaiba/dfc_learning) - NEW
- Dedicated Feature Crosscoders (DFCs) with feature partitioning for improved model diffing
- Model-exclusive and shared feature subspaces
- Gradient masking to prevent auxiliary loss leakage
- Comprehensive test suite (19 tests) validating partition integrity
- Drop-in compatible with existing
BatchTopKCrossCoderTrainer
pip install git+https://github.com/superkaiba/dfc_learningfrom dictionary_learning import CrossCoder, DedicatedFeatureBatchTopKCrossCoder
from nnsight import LanguageModel
import torch as th
# Load pretrained crosscoder
crosscoder = CrossCoder.from_pretrained("Butanium/gemma-2-2b-crosscoder-l13-mu4.1e-02-lr1e-04", from_hub=True)
# Load models
gemma_2 = LanguageModel("google/gemma-2-2b", device_map="cuda:0")
gemma_2_it = LanguageModel("google/gemma-2-2b-it", device_map="cuda:1")
prompt = "quick fox brown"
# Extract activations
with gemma_2.trace(prompt):
l13_act_base = gemma_2.model.layers[13].output[0][:, -1].save() # (1, 2304)
gemma_2.model.layers[13].output.stop()
with gemma_2_it.trace(prompt):
l13_act_it = gemma_2_it.model.layers[13].output[0][:, -1].save() # (1, 2304)
gemma_2_it.model.layers[13].output.stop()
# Run crosscoder
crosscoder_input = th.cat([l13_act_base, l13_act_it], dim=0).unsqueeze(0).cpu() # (batch, 2, 2304)
reconstruction, features = crosscoder(crosscoder_input, output_features=True)
# Print metrics
print(f"MSE loss: {th.nn.functional.mse_loss(reconstruction, crosscoder_input).item():.2f}")
print(f"L1 sparsity: {features.abs().sum():.1f}")
print(f"L0 sparsity: {(features > 1e-4).sum()}")⭐ NEW in this fork - DFCs extend the CrossCoder architecture with feature partitioning for improved model diffing. The feature dictionary is divided into three disjoint sets:
- Model A-exclusive features: Only Model A can encode/decode from these
- Model B-exclusive features: Only Model B can encode/decode from these
- Shared features: Both models can encode/decode from these
This architectural constraint creates a prior favoring model-exclusive feature discovery, which is valuable for identifying behavioral differences between model variants.
- Gradient Masking: Prevents auxiliary loss from leaking gradients to the wrong model's exclusive features
- Partition Integrity: Forbidden weights (e.g., Model A's decoder for B-exclusive features) remain zero throughout training
- Automatic Training: Works with existing
BatchTopKCrossCoderTrainerwithout modification - Comprehensive Testing: 19 tests validate gradient masking and auxiliary loss isolation
from dictionary_learning import DedicatedFeatureBatchTopKCrossCoder
from dictionary_learning.trainers.crosscoder import BatchTopKCrossCoderTrainer
# Create DFC with 5% features dedicated to each model
dfc = DedicatedFeatureBatchTopKCrossCoder(
activation_dim=2304,
dict_size=131072,
num_layers=2,
k=200,
model_a_exclusive_pct=0.05, # 5% = 6,554 features for Model A
model_b_exclusive_pct=0.05, # 5% = 6,554 features for Model B
# Remaining 90% = 117,964 shared features
)
# Train using standard CrossCoder trainer
trainer = BatchTopKCrossCoderTrainer(
steps=100000,
activation_dim=2304,
dict_size=131072,
k=200,
layer=13,
lm_name="model-a_vs_model-b",
lr=1e-4,
dict_class=DedicatedFeatureBatchTopKCrossCoder,
dict_class_kwargs={
"model_a_exclusive_pct": 0.05,
"model_b_exclusive_pct": 0.05,
},
)
# Verify partition integrity during/after training
integrity = dfc.verify_partition_integrity()
print(f"Max encoder violation: {integrity['encoder']['max_violation']:.2e}")
print(f"Max decoder violation: {integrity['decoder']['max_violation']:.2e}")
# Values should stay near zero (< 1e-6) throughout training# Comprehensive test suite (19 tests including gradient masking validation)
pytest tests/test_dfc.py -v- Smaller partitions (1-2%): Better shared feature quality, may miss fine-grained differences
- Larger partitions (5-10%): Captures more model-specific features, trades off shared capacity
- Recommended start: 5% for each model (default)
For more details, see dictionary_learning/feature_partition.py and dictionary_learning/dictionary.py.
This is a repository for doing dictionary learning via sparse autoencoders on neural network activations. It was developed by Samuel Marks and Aaron Mueller.
For accessing, saving, and intervening on NN activations, we use the nnsight package; as of March 2024, nnsight is under active development and may undergo breaking changes. That said, nnsight is easy to use and quick to learn; if you plan to modify this repo, then we recommend going through the main nnsight demo here.
Some dictionaries trained using this repository (and asociated training checkpoints) can be accessed at https://baulab.us/u/smarks/autoencoders/. See below for more information about these dictionaries.
Navigate to the to the location where you would like to clone this repo, clone and enter the repo, and install the requirements.
git clone https://github.com/saprmarks/dictionary_learning
cd dictionary_learning
pip install -r requirements.txtTo use dictionary_learning, include it as a subdirectory in some project's directory and import it; see the examples below.
You can load and used a pretrained dictionary as follows
from dictionary_learning import AutoEncoder
# load autoencoder
ae = AutoEncoder.from_pretrained("path/to/dictionary/weights")
# get NN activations using your preferred method: hooks, transformer_lens, nnsight, etc. ...
# for now we'll just use random activations
activations = torch.randn(64, activation_dim)
features = ae.encode(activations) # get features from activations
reconstructed_activations = ae.decode(features)
# you can also just get the reconstruction ...
reconstructed_activations = ae(activations)
# ... or get the features and reconstruction at the same time
reconstructed_activations, features = ae(activations, output_features=True)Dictionaries have encode, decode, and forward methods -- see dictionary.py.
We have limited support for automatically converting SAEs from sae_lens; currently this is only supported for JumpReLU SAEs, but we may expand support if users are interested.
from dictionary_learning import JumpReluAutoEncoder
ae = JumpReluAutoEncoder.from_pretrained(
load_from_sae_lens=True,
release="your_release_name",
sae_id="your_sae_id"
)The arguments should should match those used in the SAE.from_pretrained call you would use to load an SAE in sae_lens. For this to work, sae_lens should be installed in your environment.
To train your own dictionaries, you'll need to understand a bit about our infrastructure. (See below for downloading our dictionaries.)
This repository supports different sparse autoencoder architectures, including standard AutoEncoder (Bricken et al., 2023), GatedAutoEncoder (Rajamanoharan et al., 2024), and AutoEncoderTopK (Gao et al., 2024).
Each sparse autoencoder architecture is implemented with a corresponding trainer that implements the training protocol described by the authors.
This allows us to implement different training protocols (e.g. p-annealing) for different architectures without a lot of overhead.
Specifically, this repository supports the following trainers:
StandardTrainer: Implements a training scheme similar to that of Bricken et al., 2023.GatedSAETrainer: Implements the training scheme for Gated SAEs described in Rajamanoharan et al., 2024.AutoEncoderTopK: Implemented the training scheme for Top-K SAEs described in Gao et al., 2024.PAnnealTrainer: Extends theStandardTrainerby providing the option to anneal the sparsity parameter p.GatedAnnealTrainer: Extends theGatedSAETrainerby providing the option for p-annealing, similar toPAnnealTrainer.
Another key object is the ActivationBuffer, defined in buffer.py. Following Neel Nanda's appraoch, ActivationBuffers maintain a buffer of NN activations, which it outputs in batches.
An ActivationBuffer is initialized from an nnsight LanguageModel object, a submodule (e.g. an MLP), and a generator which yields strings (the text data). It processes a large number of strings, up to some capacity, and saves the submodule's activations. You sample batches from it, and when it is half-depleted, it refreshes itself with new text data.
Here's an example for training a dictionary; in it we load a language model as an nnsight LanguageModel (this will work for any Huggingface model), specify a submodule, create an ActivationBuffer, and then train an autoencoder with trainSAE.
from nnsight import LanguageModel
from dictionary_learning import ActivationBuffer, AutoEncoder
from dictionary_learning.trainers import StandardTrainer
from dictionary_learning.training import trainSAE
device = "cuda:0"
model_name = "EleutherAI/pythia-70m-deduped" # can be any Huggingface model
model = LanguageModel(
model_name,
device_map=device,
)
submodule = model.gpt_neox.layers[1].mlp # layer 1 MLP
activation_dim = 512 # output dimension of the MLP
dictionary_size = 16 * activation_dim
# data must be an iterator that outputs strings
data = iter(
[
"This is some example data",
"In real life, for training a dictionary",
"you would need much more data than this",
]
)
buffer = ActivationBuffer(
data=data,
model=model,
submodule=submodule,
d_submodule=activation_dim, # output dimension of the model component
n_ctxs=3e4, # you can set this higher or lower dependong on your available memory
device=device,
) # buffer will yield batches of tensors of dimension = submodule's output dimension
trainer_cfg = {
"trainer": StandardTrainer,
"dict_class": AutoEncoder,
"activation_dim": activation_dim,
"dict_size": dictionary_size,
"lr": 1e-3,
"device": device,
}
# train the sparse autoencoder (SAE)
ae = trainSAE(
data=buffer, # you could also use another (i.e. pytorch dataloader) here instead of buffer
trainer_configs=[trainer_cfg],
)Some technical notes our training infrastructure and supported features:
- Training uses the
ConstrainedAdamoptimizer defined intraining.py. This is a variant of Adam which supports constraining theAutoEncoder's decoder weights to be norm 1. - Neuron resampling: if a
resample_stepsargument is passed totrainSAE, then dead neurons will periodically be resampled according to the procedure specified here. - Learning rate warmup: if a
warmup_stepsargument is passed totrainSAE, then a linear LR warmup is used at the start of training and, if doing neuron resampling, also after every time neurons are resampled.
If submodule is a model component where the activations are tuples (e.g. this is common when working with residual stream activations), then the buffer yields the first coordinate of the tuple.
To download our pretrained dictionaries automatically, run:
./pretrained_dictionary_downloader.shThis will download dictionaries of all submodules (~2.5 GB) hosted on huggingface. Currently, we provide dictionaries from the 10_32768 training run. This set has dictionaries for MLP outputs, attention outputs, and residual streams (including embeddings) in all layers of EleutherAI's Pythia-70m-deduped model. These dictionaries were trained on 2B tokens from The Pile.
Let's explain the directory structure by example. After using the script above, you'll have a dictionaries/pythia-70m-deduped/mlp_out_layer1/10_32768 directory corresponding to the layer 1 MLP dictionary from the 10_32768 set. This directory contains:
ae.pt: thestate_dictof the fully trained dictionaryconfig.json: a json file which specifies the hyperparameters used to train the dictionarycheckpoints/: a directory containing training checkpoints of the formae_step.pt(only if you used the--checkpointsflag)
We've also previously released other dictionaries which can be found and downloaded here.
We'll report the following statistics for our 10_32768 dictionaries. These were measured using the code in evaluation.py.
- MSE loss: average squared L2 distance between an activation and the autoencoder's reconstruction of it
- L1 loss: a measure of the autoencoder's sparsity
- L0: average number of features active above a random token
- Percentage of neurons alive: fraction of the dictionary features which are active on at least one token out of 8192 random tokens
- CE diff: difference between the usual cross-entropy loss of the model for next token prediction and the cross entropy when replacing activations with our dictionary's reconstruction
- Percentage of CE loss recovered: when replacing the activation with the dictionary's reconstruction, the percentage of the model's cross-entropy loss on next token prediction that is recovered (relative to the baseline of zero ablating the activation)
| Layer | Variance Explained (%) | L1 | L0 | % Alive | CE Diff | % CE Recovered |
|---|---|---|---|---|---|---|
| 0 | 92 | 8 | 128 | 17 | 0.02 | 99 |
| 1 | 87 | 9 | 127 | 17 | 0.03 | 94 |
| 2 | 90 | 19 | 215 | 12 | 0.05 | 93 |
| 3 | 89 | 12 | 169 | 13 | 0.03 | 93 |
| 4 | 83 | 8 | 132 | 14 | 0.01 | 95 |
| 5 | 89 | 11 | 144 | 20 | 0.02 | 93 |
| Layer | Variance Explained (%) | L1 | L0 | % Alive | CE Diff | % CE Recovered |
|---|---|---|---|---|---|---|
| 0 | 97 | 5 | 5 | 40 | 0.10 | 99 |
| 1 | 85 | 8 | 69 | 44 | 0.06 | 95 |
| 2 | 99 | 12 | 88 | 31 | 0.11 | 88 |
| 3 | 88 | 20 | 160 | 25 | 0.12 | 94 |
| 4 | 92 | 20 | 100 | 29 | 0.14 | 90 |
| 5 | 96 | 31 | 102 | 35 | 0.15 | 97 |
NOTE: these are indexed so that the resid_i dictionary is the output of the ith layer. Thus embeddings go first, then layer 0, etc.
| Layer | Variance Explained (%) | L1 | L0 | % Alive | CE Diff | % CE Recovered |
|---|---|---|---|---|---|---|
| embed | 96 | 1 | 3 | 36 | 0.17 | 98 |
| 0 | 92 | 11 | 59 | 41 | 0.24 | 97 |
| 1 | 85 | 13 | 54 | 38 | 0.45 | 95 |
| 2 | 96 | 24 | 108 | 27 | 0.55 | 94 |
| 3 | 96 | 23 | 68 | 22 | 0.58 | 95 |
| 4 | 88 | 23 | 61 | 27 | 0.48 | 95 |
| 5 | 90 | 35 | 72 | 45 | 0.55 | 92 |
Note: these features are likely to be depricated in future releases.
We've included support for some experimental features. We briefly investigated them as an alternative approaches to training dictionaries.
-
MLP stretchers. Based on the perspective that one may be able to identify features with "neurons in a sufficiently large model," we experimented with training "autoencoders" to, given as input an MLP input activation
$x$ , output not$x$ but$MLP(x)$ (the same output as the MLP). For instance, given an MLP which maps a 512-dimensional input$x$ to a 1024-dimensional hidden state$h$ and then a 512-dimensional output$y$ , we train a dictionary$A$ with hidden dimension 16384 = 16 x 1024 so that$A(x)$ is close to$y$ (and, as usual, so that the hidden state of the dictionary is sparse).- The resulting dictionaries seemed decent, but we decided not to pursue the idea further.
- To use this functionality, set the
ioparameter of an activaiton buffer to'in_to_out'(default is'out'). - h/t to Max Li for this suggestion.
- Replacing L1 loss with entropy. Based on the ideas in this post, we experimented with using entropy to regularize a dictionary's hidden state instead of L1 loss. This seemed to cause the features to split into dead features (which never fired) and very high-frequency features which fired on nearly every input, which was not the desired behavior. But plausibly there is a way to make this work better.
- Ghost grads, as described here.