Alternate Alignment Ideas

These are ‘brainstorming’ posts, around the theme of what it means for a system to be helpful to a human.

Stable Poin­t­ers to Value: An Agent Embed­ded in Its Own Utility Function

Stable Poin­t­ers to Value II: En­vi­ron­men­tal Goals

Stable Poin­t­ers to Value III: Re­cur­sive Quantilization

Policy Alignment

Non-Con­se­quen­tial­ist Co­op­er­a­tion?