RSS

Michael Soareverix

Karma: 78

Could an AI Align­ment Sand­box be use­ful?

Michael Soareverix2 Jul 2022 5:06 UTC
2 points
1 comment1 min readLW link

Three Min­i­mum Pivotal Acts Pos­si­ble by Nar­row AI

Michael Soareverix12 Jul 2022 9:51 UTC
0 points
4 comments2 min readLW link

Mus­ings on the Hu­man Ob­jec­tive Function

Michael Soareverix15 Jul 2022 7:13 UTC
3 points
0 comments3 min readLW link

Our Ex­ist­ing Solu­tions to AGI Align­ment (semi-safe)

Michael Soareverix21 Jul 2022 19:00 UTC
12 points
1 comment3 min readLW link

A rough idea for solv­ing ELK: An ap­proach for train­ing gen­er­al­ist agents like GATO to make plans and de­scribe them to hu­mans clearly and hon­estly.

Michael Soareverix8 Sep 2022 15:20 UTC
2 points
2 comments2 min readLW link

A Good Fu­ture (rough draft)

Michael Soareverix24 Oct 2022 20:45 UTC
10 points
5 comments3 min readLW link