RSS

Evan R. Murphy

Karma: 1,195

Send me feedback anonymously

I’m doing research and other work focused on AI safety/​security, governance and risk reduction. Currently my top projects are (last updated Feb 26, 2025):

General areas of interest for me are AI safety strategy, comparative AI alignment research, prioritizing technical alignment work, analyzing the published alignment plans of major AI labs, interpretability, deconfusion research and other AI safety-related topics.

Research that I’ve authored or co-authored:

Before getting into AI safety, I was a software engineer for 11 years at Google and various startups. You can find details about my previous work on my LinkedIn.

While I’m not always great at responding, I’m happy to connect with other researchers or people interested in AI alignment and effective altruism. Feel free to send me a private message!

AI Risk: Can We Thread the Nee­dle? [Recorded Talk from EA Sum­mit Van­cou­ver ’25]

Evan R. MurphyOct 2, 2025, 7:08 PM
6 points

1 vote

Overall karma indicates overall quality.

0 comments2 min readLW link

[Question] Does the Univer­sal Geom­e­try of Embed­dings pa­per have big im­pli­ca­tions for in­ter­pretabil­ity?

Evan R. MurphyMay 26, 2025, 6:20 PM
43 points

13 votes

Overall karma indicates overall quality.

6 comments1 min readLW link

Evan R. Mur­phy’s Shortform

Evan R. MurphyFeb 28, 2025, 12:56 AM
6 points

1 vote

Overall karma indicates overall quality.

6 comments1 min readLW link