Apollo Research (London).

My main research interests are mechanistic interpretability and inner alignment.

# Lee Sharkey(Lee Sharkey)

# Decomposing the QK circuit with Bilinear Sparse Dictionary Learning

I’ll add a strong plus one to this and a note for emphasis:

Representational geometry is already a long-standing theme in computational neuroscience (e.g. Kriegeskorte et al., 2013)

Overall I think mech interp practitioners would do well to pay more attention to ideas and methods in computational neuroscience. I think mech interp as a field has a habit of overlooking some hard-won lessons learned by that community.

# Apollo Research 1-year update

# Identifying Functionally Important Features with End-to-End Sparse Dictionary Learning

I’m pretty sure that there’s at least one other MATS group (unrelated to us) currently working on this, although I’m not certain about any of the details. Hopefully they release their research soon!

There’s recent work published on this here by Chris Mathwin, Dennis Akar, and me. The gated attention block is a kind of transcoder adapted for attention blocks.

Nice work by the way! I think this is a promising direction.

Note also the similar, but substantially different, use of the term transcoder here, whose problems were pointed out to me by Lucius. Addressing those problems helped to motivate our interest in the kind of transcoders that you’ve trained in your work!

Trying to summarize my current understanding of what you’re saying:

Yes all four sound right to me.

To avoid any confusion, I’d just add an emphasis that the descriptions are*mathematical*, as opposed semantic.I’d guess you have intuitions that the “short description length” framing is philosophically the right one, and I probably don’t quite share those and feel more confused how to best think about “short descriptions” if we don’t just allow arbitrary Turing machines (basically because deciding what allowable “parts” or mathematical objects are seems to be doing a lot of work). Not sure how feasible converging on this is in this format (though I’m happy to keep trying a bit more in case you’re excited to explain).

I too am keen to converge on a format in terms of Turing machines or Kolmogorov complexity or something else more formal. But I don’t feel very well placed to do that, unfortunately, since thinking in those terms isn’t very natural to me yet.

Hm I think of the (network, dataset) as scaling multiplicatively with size of network and size of dataset. In the thread with Erik above, I touched a little bit on why:

“*SAEs (or decompiled networks that use SAEs as the building block) are supposed to approximate the original network behaviour. So SAEs are mathematical descriptions of the network, but not of the (network, dataset). What’s a mathematical description of the (network, dataset), then? It’s just what you get when you pass the dataset through the network; this datum interacts with this weight to produce this activation, that datum interacts with this weight to produce that activation, and so on. A mathematical description of the (network, dataset) in terms of SAEs are: this datum activates dictionary features xyz (where xyz is just indices and has no semantic info), that datum activates dictionary features abc, and so on.*”And spiritually, we only need to understand behavior on the training dataset to understand everything that SGD has taught the model.

Yes, I roughly agree with the spirit of this.

Is there some formal-ish definition of “explanation of (network, dataset)” and “mathematical description length of an explanation” such that you think SAEs are especially short explanations? I still don’t think I have whatever intuition you’re describing, and I feel like the issue is that I don’t know how you’re measuring description length and what class of “explanations” you’re considering.

**I’ll register that I prefer using ‘description’ instead of ‘explanation’ in most places. The reason is that ‘explanation’ invokes a notion of understanding, which requires both a mathematical description and a semantic description. So I regret using the word explanation in the comment above (although not completely wrong to use it—but it did risk confusion). I’ll edit to replace it with ‘description’ and strikethrough ‘explanation’.***“explanation of (network, dataset)”*: I’m afraid I don’t have a great formalish definition beyond just pointing at the intuitive notion. But formalizing what an explanation is seems like a high bar. If it’s helpful, a mathematical description is just a statement of what the network is in terms of particular kinds of mathematical objects.*“mathematical description length of an explanation”*: (Note: Mathematical descriptions are of networks, not of explanations.) It’s just the set of objects used to describe the network. Maybe helpful to think in terms of maps between different descriptions: E.g. there is a many-to-one map between a description of a neural network in terms of polytopes and in terms of neurons. There are ~exponentially many more polytopes. Hence the mathematical description of the network in terms of individual polytopes is much larger.

Focusing instead on what an “explanation” is: would you say the network itself is an “explanation of (network, dataset)” and just has high description length?

I would not. So:

If not, then the thing I don’t understand is more about what an explanation is and why SAEs are one, rather than how you measure description length.

I think that the confusion might again be from using ‘explanation’ rather than description.

SAEs (or decompiled networks that use SAEs as the building block) are supposed to approximate the original network behaviour. So SAEs are mathematical descriptions of the network, but not of the (network, dataset). What’s a mathematical description of the (network, dataset), then? It’s just what you get when you pass the dataset through the network;*this*datum interacts with this weight to produce this activation,*that*datum interacts with this weight to produce that activation, and so on. A mathematical description of the (network, dataset) in terms of SAEs are:*this*datum activates dictionary features xyz (where xyz is just indices and has no semantic info),*that*datum activates dictionary features abc, and so on.

Lmk if that’s any clearer.

# Gated Attention Blocks: Preliminary Progress toward Removing Attention Head Superposition

Thanks Aidan!

I’m not sure I follow this bit:

In my mind, the reconstruction loss is more of a non-degeneracy control to encourage almost-orthogonality between features.

I don’t currently see why reconstruction would encourage features to be different directions from each other in any way unless paired with an L_{0<p<1}. And I specifically don’t mean L1, because in toy data settings with recon+L1, you can end up with features pointing in exactly the same direction.

Thanks Erik :) And I’m glad you raised this.

One of the things that many researchers I’ve talked to don’t appreciate is that, if we accept networks can do computation in superposition, then we also have to accept that

**we can’t just understand the network alone**. We want to understand the network’s behaviour*on a dataset,*where the dataset contains potentially lots of features. And depending on the features that are active in a given datum, the network can do different computations in superposition (unlike in a linear network that can’t do superposition). The combined object ‘(network, dataset)’ is much larger than the network itself.~~Explanations~~Descriptions of the (network, dataset) object can actually be compressions despite potentially being larger than the network.So,

One might say that SAEs lead to something like a shorter “description length of what happens on any individual input” (in the sense that fewer features are active). But I don’t think there’s a formalization of this claim that captures what we want. In the limit of very many SAE features, we can just have one feature active at a time, but clearly that’s not helpful.

You can have one feature active for each datapoint, but now we’ve got an

~~explanation~~description of the (network, dataset) that scales linearly in the size of the dataset, which sucks! Instead, if we look for regularities (opportunities for compression) in how the network treats data, then we have a better chance at~~explanations~~descriptions that scale better with dataset size. Suppose a datum consists of a novel combination of previously~~explained~~described circuits. Then our~~explanation~~description of the (network, dataset) is much smaller than if we~~explained~~described every datapoint anew.In light of that, you can understand my disagreement with “

*in that case, I could also reduce the description length by training a smaller model*.” No! Assuming the network is smaller yet as performant (therefore presumably doing more computation in superposition), then the~~explanation~~description of the (network, dataset) is basically unchanged.

So, for models that are 10 terabytes in size, you should perhaps be expecting a “model manual” which is around 10 terabytes in size.

Yep, that seems reasonable.

I’m guessing you’re not satisfied with the retort that we should expect AIs to do the heavy lifting here?

Or perhaps you don’t think you need something which is close in accuracy to a full explanation of the network’s behavior.

I think the accuracy you need will depend on your use case. I don’t think of it as a globally applicable quantity for all of interp.

For instance, maybe to ‘audit for deception’ you really only need identify and detect when the deception circuits are active, which will involve explaining only 0.0001% of the network.

But maybe to make robust-to-training interpretability methods you need to understand 99.99...99%.

It seem likely to me that we can unlock more and more interpretability use cases by understanding more and more of the network.

Thanks for this feedback! I agree that the task & demo you suggested should be of interest to those working on the agenda.

It makes me a bit worried that this post seems to implicitly assume that SAEs work well at their stated purpose.

There were a few purposes proposed, and at multiple levels of abstraction, e.g.

The purpose of being the main building block of a mathematical description used in an ambitious mech interp solution

The purpose of being the main building block of decompiled networks

The purpose of taking features out of superposition

I’m going to assume you meant the first one (and maybe the second). Lmk if not.

Fwiw I’m not totally convinced that SAEs are the ultimate solution for the purposes in the first two bullet points. But I do think they’re currently SOTA for ambitious mech interp purposes, and there is usually scientific benefit of using imperfect but SOTA methods to push the frontier of what we know about network internals. Indeed, I view this as beneficial in the same way that historical applications of (e.g.) causal scrubbing for circuit discovery were beneficial, despite the imperfections of both methods.

I’ll also add a persnickety note that I do explicitly say in the agenda that we should be looking for better methods than SAEs: “*It would be nice to have a formal justification for why we should expect sparsification to yield short semantic descriptions. Currently, the justification is simply that it appears to work and a vague assumption about the data distribution containing sparse features. I would support work that critically examines this assumption (though I don’t currently intend to work on it directly), since it may yield a better criterion to optimize than simply ‘sparsity’ or may yield even better interpretability methods than SAEs.*”

However, to concede to your overall point, the rest of the article does kinda suggest that we can make progress in interp with SAEs. But as argued above, I’m comfortable that some people in the field proceed with inquiries that use probably imperfect methods.Precisely, I would bet against “mild tweaks on SAEs will allow for interpretability researchers to produce succinct and human understandable explanations that allow for recovering >75% of the training compute of model components”.

I’m curious if you believe that, even if SAEs aren’t the right solution, there realistically exists a potential solution that would allow researchers to produce succinct, human understandable explanation that allow for recovering >75% of the training compute of model components?

I’m wondering if the issue you’re pointing at is the goal rather than the method.

# Sparsify: A mechanistic interpretability research agenda

This is a good idea and is something we’re (Apollo + MATS stream) working on atm. We’re planning on releasing our agenda related to this and, of course, results whenever they’re ready to share.

# Addressing Feature Suppression in SAEs

Makes sense! Thanks!

Great! I’m curious, what was it about the sparsity penalty that you changed your mind about?

Hey thanks for your review! Though I’m not sure that either this article or Cunningham et al. can reasonably be described as a reproduction of Anthropic’s results (by which I assume you’re talking about Bricken et al.), given their relative timings and contents.

I’ve no great resources in mind for this. Off the top of my head, a few examples of ideas that are common in comp neuro that might have bought mech interp some time if more people in mech interp were familiar with them when they were needed:

- Polysemanticity/distributed representations/mixed selectivity

- Sparse coding

- Representational geometry

I am not sure about what future mech interp needs will be, so it’s hard to predict which ideas or methods that are common in neuro will be useful (topological data analysis? Dynamical systems?). But it just seems pretty likely that a field that tackles such a similar problem will continue to be a useful source of intuitions and methods. I’d love if someone were to write a review or post on the interplay between the parallel fields of comp neuro and mech interp. It might help flag places where there ought to be more interplay.