Full time independent deconfusion researcher (https://www.alignmentforum.org/posts/5Nz4PJgvLCpJd6YTA/looking-deeper-at-deconfusion) in AI Alignment. (Also PhD in the theory of distributed computing).
If you’re interested by some research ideas that you see in my posts, know that I keep private docs with the most compressed version of my deconfusion ideas in the process of getting feedback. I can give you access if you PM me!
A list of topics I’m currently doing deconfusion on:
Goal-directedness for discussing AI Risk
Myopic Decision Theories for dealing with deception (with Evan Hubinger)
Universality for many alignment ideas of Paul Christiano
Deconfusion itself to get better at it
Models of Languages Models to clarify the alignment issues surrounding them.