Is there a list of projects to get started with Interpretability?

Big parts of Alignment outreach is done by trying to draw CS students into the filed, which gets even more present in the light of the EAGx’s that happen around the world right now. There are endless courses that teach basic skills in the field, loads of research agendas of the field, lists of specific open questions and posts giving a high-level overview.

Now if you have gotten some general overview, it naturally gets interesting to learn about something very specific by actively doing a specific technical project, however I feel like there’s a lack of formulated problems such as this one from Buck Shlegris, one get just get started with.

More specifically: We are a group of 3 CS undergrads (3rd and 4th year), who have done some of the big picture general AI alignment stuff, probably could go much deeper & broader in the general alignment resources by just passively reading but for now want to dive deeper into Interpretability research, not by keeping reading, but by actively implementing something and getting started hands-on.

Does anyone have a list of projects we could look at like the one of Buck (maybe a bit longer or more focused on interpretability), mainly focused on providing a good learning experience? Or can point me to some training projects? I think this kind of list might be helpful for loads of student interested in alignment research and the fact that effective thesis does not feature something like this is unfortunate