RSS

capybaralet

Karma: 215 (LW), 0 (AF)
Page 1

Disam­biguat­ing “al­ign­ment” and re­lated no­tions

capybaralet
5 Jun 2018 15:35 UTC
43 points
21 comments2 min readLW link

Con­cep­tual Anal­y­sis for AI Align­ment

capybaralet
30 Dec 2018 0:46 UTC
26 points
2 comments2 min readLW link

Thoughts on Ben Garfinkel’s “How sure are we about this AI stuff?”

capybaralet
6 Feb 2019 19:09 UTC
25 points
17 comments1 min readLW link

The role of epistemic vs. aleatory un­cer­tainty in quan­tify­ing AI-Xrisk

capybaralet
31 Jan 2019 6:13 UTC
14 points
6 comments2 min readLW link

Ineffi­cient Games

capybaralet
23 Aug 2016 17:47 UTC
14 points
13 comments1 min readLW link

My use of the phrase “Su­per-Hu­man Feed­back”

capybaralet
6 Feb 2019 19:11 UTC
12 points
0 comments1 min readLW link

X-risks are a tragedies of the commons

capybaralet
7 Feb 2019 2:48 UTC
9 points
19 comments1 min readLW link

Imi­ta­tion learn­ing con­sid­ered un­safe?

capybaralet
6 Jan 2019 15:48 UTC
9 points
11 comments1 min readLW link