RSS

Stuart_Armstrong

Karma: 21,149 (LW), 336 (AF)
Page 1

Re­search Agenda in re­verse: what *would* a solu­tion look like?

Stuart_Armstrong
25 Jun 2019 13:52 UTC
21 points
1 comment1 min readLW link

Re­search Agenda v0.9: Syn­the­sis­ing a hu­man’s prefer­ences into a util­ity function

Stuart_Armstrong
17 Jun 2019 17:46 UTC
49 points
10 comments32 min readLW link

Prefer­ence con­di­tional on cir­cum­stances and past prefer­ence satisfaction

Stuart_Armstrong
17 Jun 2019 15:30 UTC
11 points
1 comment1 min readLW link

For the past, in some ways only, we are moral degenerates

Stuart_Armstrong
7 Jun 2019 15:57 UTC
29 points
14 comments2 min readLW link

To first or­der, moral re­al­ism and moral anti-re­al­ism are the same thing

Stuart_Armstrong
3 Jun 2019 15:04 UTC
17 points
5 comments3 min readLW link

Con­di­tional meta-preferences

Stuart_Armstrong
3 Jun 2019 14:09 UTC
6 points
0 comments1 min readLW link

Uncer­tainty ver­sus fuzzi­ness ver­sus ex­trap­o­la­tion desiderata

Stuart_Armstrong
30 May 2019 13:52 UTC
20 points
8 comments4 min readLW link

And the AI would have got away with it too, if...

Stuart_Armstrong
22 May 2019 21:35 UTC
71 points
4 comments1 min readLW link

By de­fault, avoid am­bigu­ous dis­tant situations

Stuart_Armstrong
21 May 2019 14:48 UTC
31 points
15 comments3 min readLW link

Or­a­cles, se­quence pre­dic­tors, and self-con­firm­ing predictions

Stuart_Armstrong
3 May 2019 14:09 UTC
21 points
0 comments3 min readLW link

Self-con­firm­ing pre­dic­tions can be ar­bi­trar­ily bad

Stuart_Armstrong
3 May 2019 11:34 UTC
43 points
11 comments5 min readLW link

Nash equil­ibriums can be ar­bi­trar­ily bad

Stuart_Armstrong
1 May 2019 14:58 UTC
35 points
24 comments2 min readLW link

Defeat­ing Good­hart and the “clos­est un­blocked strat­egy” problem

Stuart_Armstrong
3 Apr 2019 14:46 UTC
39 points
15 comments6 min readLW link

Learn­ing “known” in­for­ma­tion when the in­for­ma­tion is not ac­tu­ally known

Stuart_Armstrong
1 Apr 2019 17:56 UTC
13 points
0 comments1 min readLW link

Rel­a­tive ex­change rate be­tween preferences

Stuart_Armstrong
29 Mar 2019 11:46 UTC
12 points
1 comment1 min readLW link

Be­ing wrong in ethics

Stuart_Armstrong
29 Mar 2019 11:28 UTC
22 points
0 comments3 min readLW link

Models of prefer­ences in dis­tant situations

Stuart_Armstrong
29 Mar 2019 10:42 UTC
11 points
0 comments2 min readLW link

The low cost of hu­man prefer­ence incoherence

Stuart_Armstrong
27 Mar 2019 11:58 UTC
19 points
5 comments2 min readLW link

“Mo­ral” as a prefer­ence label

Stuart_Armstrong
26 Mar 2019 10:30 UTC
14 points
1 comment1 min readLW link

Par­tial prefer­ences and models

Stuart_Armstrong
19 Mar 2019 16:29 UTC
13 points
6 comments2 min readLW link