Daniel Dewey of Open Phil gives a very well-articulated summary (IMO) of MIRI’s work on trying to develop High Reliable Agents, which is in their Agent Foundations agenda and covers things like reasoning under uncertainty / decision theory / epistemology.
He examines the arguments behind why MIRI thinks it’s worth pursuing, and he evaluates them against other possible approaches to AI safety.
Short summary:
Daniel Dewey of Open Phil gives a very well-articulated summary (IMO) of MIRI’s work on trying to develop High Reliable Agents, which is in their Agent Foundations agenda and covers things like reasoning under uncertainty / decision theory / epistemology.
He examines the arguments behind why MIRI thinks it’s worth pursuing, and he evaluates them against other possible approaches to AI safety.