RSS

Joseph Bloom

Karma: 262

I’m an independently funded AI Alignment Research Engineer focussing on mechanistic interpretability in reinforcement learning. I’m particularly interested in comparing circuits in decision transformers to those generated by other techniques.

Joseph Bloom on choos­ing AI Align­ment over bio, what many as­piring re­searchers get wrong, and more (in­ter­view)

17 Sep 2023 18:45 UTC
25 points
2 comments8 min readLW link

A Mechanis­tic In­ter­pretabil­ity Anal­y­sis of a GridWorld Agent-Si­mu­la­tor (Part 1 of N)

Joseph Bloom16 May 2023 22:59 UTC
36 points
2 comments16 min readLW link