Working on making AI systems reason safely about decision theory and acausal interactions and eliciting their conceptual reasoning abilities, collaborating with Caspar Oesterheld and Emery Cooper.
I try to check the LessWrong infrequently. If you wanna get in touch, you can use my admonymous link and leave your email address there, so I can reply to you! (If you don’t include some contact details in your message, I can’t reply)
You can also just send me thoughts and questions anonymously!
How others can help me
Be interested in working on/implementing ideas from research on acausal cooperations! Or connect me with people who might be.
How I can help others
Ask me about acausal stuff!
Or any of my background: Before doing independent research, I worked for the Center on Long-Term Risk on s-risk reduction projects (hiring, community building, and grantmaking.) Previously, I was a guest manager at the EA Infrastructure Fund (2021), did some research for 1 Day Sooner on Human Challenge Trials for Covid vaccines (2020), did the summer research fellowship at FHI writing about IDA (2019), worked a few hours a week for CEA on local groups mentoring for a few months (2018), and helped a little bit with organizing EA Oxford (2018/19). I studied PPE at Oxford (2018-2021) and psychology in Freiburg (2015-2018.)
I also have things to say about mental health and advice for taking a break from work.
Hello! Just chiming in here to add a few things.
As one of the “academics” (scare quotes because I’m unfortunately quite academically uncredentialed), I just wanna note that our entire team actually cares a lot about this past an intellectual exercise in case people will read this as implying the opposite! I think of the entire point of my career as making decision theory and acausal dynamics go well in AIs (while also hopefully helping with automating alignment as a massive positive side effect) and I think others on the team feel similarly.
On a different note, I think the difference between FDT and empirically updateless EDT often feels overblown on LessWrong. I don’t actually know of cases where the two very clearly come apart, so I’m often confused when people seem confident in FDT’s superiority. There are cases where I’m not super sure what FDT would do, so they might come apart in those but if so, I think these cases haven’t been clearly articulated.