Sparsity is the enemy of feature extraction (ft. absorption)

Sparse Autoencoders (and other related feature extraction tools) often optimize for sparsity to extract human-interpretable latent representations from a model’s activation space. We show analytically that sparsity naturally leads to feature absorption in a simplified untied SAE, and discuss how this makes SAEs less trustworthy to use for AI safety with some ongoing efforts to fix this. This might be obvious to people working in the field—but we ended up writing a proof sketch so we’re putting it out here. Produced as part of the ML Alignment & Theory Scholars Program—Winter 2024-25 Cohort.

The dataset (a distribution with feature hierarchy)

In this proof, we consider a dataset with points sampled to exhibit features from a set of feature . Particularly, we will consider two features that follow the hierarchy (think f_2 = elephant and f_1 = animal for instance), where existence of implies existence of .

Hierarchy in human-interpretable features is prevalent (and hard to study in LLMs). While other unrelated features still exist, for and , we can partition the probability of this dataset for four combinations:

Features

So these are the individual probabilities of a datapoint eliciting these combinations of features:

  • (both features present, think elephant, which implies animal),

  • (only present, think cat or dog)

  • (only present, which should be zero, because of hierarchy)

  • (neither feature present, maybe talking about volcanoes)

Each feature is a vector with unit norm, and we assume that all features are mutually orthogonal, so . Each activation in the model’s residual stream is a sum of all active features.

Training a Sparse Autoencoder (SAE)

Given a language model with residual stream activations , the sparse autoencoder learns a mapping such that reconstructs . The total loss consists of a reconstruction term , which minimizes the squared error, and a sparsity term , enforcing sparsity via an -norm penalty on . The model parameters are optimized via gradient descent to minimize . The complete loss:

where (the reconstruction loss) and (the sparsity loss).

The SAE constructs latent activations as , where .

We consider encoders and decoders without biases for simplicity, and we look at the latents of the SAE that are related to the two features and , and show that absorption doesn’t affect reconstruction and increases sparsity, and that optimizing for sparsity pushes for higher absorption.

Evaluating the SAE loss under -absorption

We use to denote the hidden activation for latent , to denote the SAE encoder for latent , and to denote the SAE decoder for latent . We assume the decoder is linear, that is, the reconstruction is a linear combination of latents . We assume that the first latents ( and ) track and without loss of generality.

We define -absorption as part-way between no absorption () and full absorption (. Thus, , , , and , and we look at the reconstruction and sparsity losses under varying amounts of absorption.

Reconstruction under -absorption

Case 1: (parent feature only)

In this case, the parent feature fires on its own, so the SAE just needs to reconstruct .

The first latent fires with magnitude 1, since but since .

The second latent fires with magnitude 0 since ..

The decoder output is thus just , perfectly reconstructing the input.

Case 2: (parent and child together)

In this case, the parent and child are firing together, so the SAE needs to reconstruct the sum of the parent and child features .

The first latent fires with magnitude since and , but since . We have .

The second latent fires with magnitude 1 since is present in .

The decoder output sums to , again perfectly reconstructing the input.

Case 3: (nothing fires)

Here, and .

When neither feature is present, neither latents fire:

And thus the decoder output is still 0, achieving perfect reconstruction.

We do not need to consider the case where fires alone, as this is not allowed by our absorption setup. As we see above, in all cases, any level of absorption still achieves perfect reconstruction.

Intuitively, absorption does not hinder reconstruction because knowing the elephant feature is active is enough to infer animal even if that feature gets absorbed.

Sparsity under -absorption

We calculate the sparsity loss as follows:

where represents the two features that are related to and , and is the norm (to disambiguate with ). Since the representation and sparsity of the unrelated features stays the same, we look at this value of sparsity and how it changes with the amount of absorption in our data distribution.

Case 1: (no absorption)

With no absorption (which means the SAE learned the true features and in both the encoder and decoder), the sparsity comes out to be (with -norm, where we assume for simplicity):

Case 2: (absolute absorption)

With absolute absorption of a feature (all datapoints exhibiting it), with the decoder learning and respectively for the encoder features (exclusion due to absorption) and . In this case, we get the following sparsity loss:

Case 3: General Case (some amount of absorption)

Finally, with an arbitrary amount of absorption of into , we get:

It is evident that absorption leads to higher sparsity in encoder activations for hierarichical features. Now, we show that minimizing sparsity via differentiating naturally leads to increasing absorption.

Differentiating the Sparsity Loss

Assume that the sparsity penalty is given by an norm on the encoder activation:

We can see that the derivative of the sparsity loss with respect to promotes absorption:

This derivative is always negative as long as , i.e., datapoints with both features exist in our distribution, so increasing sparsity (decreasing the loss) always increases absorption. Next, we show that absorption makes SAEs (and other related feature extractions) much less trustworthy for safety.

Absorption makes SAEs less trustworthy for safety

The hope with SAEs (and other related research) is that we extract model-internal latents representing human-interpretable features that we can detect and control. For this to be usable, we want this to work successfully for complex, harmful behaviours such as lying, strategic deception, power-seeking, sycophancy, backdoors, etc.

Absorption, and sparsity in feature-extraction which promotes absorption, means that even if we find linear latents for these features (for which we don’t really have good progress in the first place [TODO: cite], we can’t trust them.

A feature for deception, with absorption, might actually have learned deception-except-deception-in-2027, or a feature with power seeking might have learned power-seeking-except-instrumental-convergence. Unless we fix our feature extractors to not optimize to maximise just sparsity in latent activations (some other ideas and starting points include KL-divergence and parameter-space related directions), we can’t trust our SAEs, and their features can potentially be more misleading and harmful than helpful.

An example of feature absorption in real-world models

In the original feature absorption paper, absorption is shown to occur in real SAEs trained on LLMs. SAE latents that seem to track starting letter information (e.g. a feature that fires on tokens that start with ) fails to fire on seemingly arbitrary tokens that start with (like the token "_short"). The paper shows that this is due to more specific latents “absorbing” the “starts with ” direction.

In feature absorption, we find gerrymandered SAE latents which appear to track an interpretable concept but have holes in their recall. Here, we see the dashboard for Gemma Scope layer 3 16k, latent 6510 which appears to track “starts with S”, but mysteriously doesn’t fire on “_short”.

The original paper hypothesizes that feature absorption is a logical consequence of the sparsity penalty in SAE training, but we now have a proof that naively optimizing an SAE for sparsity will indeed lead to feature absorption.

How much of this applies to other SAE variants and feature-extraction methods?

We should expect feature absorption in any SAE architecture that incentivizes sparsity, which at present is all common architectures. JumpReLU, TopK, and standard L1 SAEs all extract feature representations by making use of sparsity and we should thus expect them all to engage in feature absorption.

Cross-layer SAE variants such as crosscoders[1] and transcoders[2] also rely on sparsity to extract features and we should thus also expect that these new architectures will also suffer from feature absorption. Indeed, recent work on transcoders finds they do experience feature absorption[3].

Matryoshka SAEs to fix Absorption

Matryoshka SAEs are a promising approach to fixing absorption in SAEs. These SAEs encode a notion of hierarchy by forcing earlier latents to reconstruct the full output on their own, making it more difficult for parent latents to have holes in their recall for child features.

  1. ^

    Sparse Crosscoders for Cross-Layer Features and Model Diffing [link], Jack Lindsey, Adly Templeton, Jonathan Marcus, Thomas Conerly, Joshua Batson, Christopher Olah, Transformer Circuits Thread 2024

  2. ^

    Transcoders Find Interpretable LLM Feature Circuits [link], Jacob Dunefsky, Philippe Chlenski, Neel Nanda, arXiv:2406.11944, 2024

  3. ^

    Transcoders Beat Sparse Autoencoders for Interpretability [link], Gonçalo Paulo, Stepan Shabalin, Nora Belrose, arXiv:2501.18823, 2025

No comments.