Alignment & Agency

An Ortho­dox Case Against Utility Functions

The Poin­t­ers Prob­lem: Hu­man Values Are A Func­tion Of Hu­mans’ La­tent Variables

Align­ment By Default

An overview of 11 pro­pos­als for build­ing safe ad­vanced AI

The ground of optimization

Search ver­sus design

In­ner Align­ment: Ex­plain like I’m 12 Edition

Inac­cessible information

AGI safety from first prin­ci­ples: Introduction

Is Suc­cess the Enemy of Free­dom? (Full)