RSS

Towards_Keeperhood

Karma: 1,002

Simon Skade

I did (mostly non-prosaic) alignment research between Feb 2022 and Aug 2025. (Won $10k in the ELK contest, participated in MLAB and SERI MATS 3.0 & 3.1, then independent research. I mostly worked on an ambitious attempt to better understand minds to figure out how to create more understandable and pointable AIs. I started with agent foundations but then developed a more sciency agenda where I also studied concrete observations from language/​linguistics, pychology, (neuroscience—though didn’t study much here yet), and from tracking my thoughts on problems I solved (aka a good kind of introspection).)

I’m now exploring advocacy for making it more likely that we get sth like the MIRI treaty (ideally with a good exit plan like human intelligence augmentation, or possibly an alignment project with actually competent leadership).

Currently based in Germany.

[Ad­vanced In­tro to AI Align­ment] 2. What Values May an AI Learn? — 4 Key Problems

Towards_Keeperhood2 Jan 2026 14:51 UTC
33 points
10 comments19 min readLW link

[Ad­vanced In­tro to AI Align­ment] 1. Goal-Directed Rea­son­ing and Why It Matters

Towards_Keeperhood30 Dec 2025 15:48 UTC
12 points
0 comments10 min readLW link

[Ad­vanced In­tro to AI Align­ment] 0. Overview and Foundations

Towards_Keeperhood22 Dec 2025 21:20 UTC
15 points
0 comments5 min readLW link