RSS

David Scott Krueger (formerly: capybaralet)(David Krueger)

Karma: 1,840

I’m more active on Twitter than LW/​AF these days: https://​​twitter.com/​​DavidSKrueger

Bio from https://​​www.davidscottkrueger.com/​​:
I am an Assistant Professor at the University of Cambridge and a member of Cambridge’s Computational and Biological Learning lab (CBL). My research group focuses on Deep Learning, AI Alignment, and AI safety. I’m broadly interested in work (including in areas outside of Machine Learning, e.g. AI governance) that could reduce the risk of human extinction (“x-risk”) resulting from out-of-control AI systems. Particular interests include:

A Some­what Vague Pro­posal for Ground­ing Ethics in Physics

David Scott Krueger (formerly: capybaralet)27 Jan 2015 5:45 UTC
−3 points
15 comments1 min readLW link

A Ba­sic Prob­lem of Ethics: Panpsy­chism?

David Scott Krueger (formerly: capybaralet)27 Jan 2015 6:27 UTC
−5 points
16 comments1 min readLW link

Should we en­able pub­lic bind­ing pre­com­mit­ments?

David Scott Krueger (formerly: capybaralet)31 Jul 2016 19:47 UTC
1 point
19 comments1 min readLW link

Ineffi­cient Games

David Scott Krueger (formerly: capybaralet)23 Aug 2016 17:47 UTC
26 points
13 comments1 min readLW link

Risks from Ap­prox­i­mate Value Learning

David Scott Krueger (formerly: capybaralet)27 Aug 2016 19:34 UTC
7 points
10 comments1 min readLW link

Prob­lems with learn­ing val­ues from observation

David Scott Krueger (formerly: capybaralet)21 Sep 2016 0:40 UTC
2 points
4 comments1 min readLW link

Disam­biguat­ing “al­ign­ment” and re­lated no­tions

David Scott Krueger (formerly: capybaralet)5 Jun 2018 15:35 UTC
22 points
21 comments2 min readLW link

Con­cep­tual Anal­y­sis for AI Align­ment

David Scott Krueger (formerly: capybaralet)30 Dec 2018 0:46 UTC
26 points
3 comments2 min readLW link

Imi­ta­tion learn­ing con­sid­ered un­safe?

David Scott Krueger (formerly: capybaralet)6 Jan 2019 15:48 UTC
20 points
11 comments1 min readLW link

The role of epistemic vs. aleatory un­cer­tainty in quan­tify­ing AI-Xrisk

David Scott Krueger (formerly: capybaralet)31 Jan 2019 6:13 UTC
15 points
6 comments2 min readLW link

Thoughts on Ben Garfinkel’s “How sure are we about this AI stuff?”

David Scott Krueger (formerly: capybaralet)6 Feb 2019 19:09 UTC
25 points
17 comments1 min readLW link

My use of the phrase “Su­per-Hu­man Feed­back”

David Scott Krueger (formerly: capybaralet)6 Feb 2019 19:11 UTC
13 points
0 comments1 min readLW link

X-risks are a tragedies of the commons

David Scott Krueger (formerly: capybaralet)7 Feb 2019 2:48 UTC
9 points
19 comments1 min readLW link

Let’s talk about “Con­ver­gent Ra­tion­al­ity”

David Scott Krueger (formerly: capybaralet)12 Jun 2019 21:53 UTC
44 points
33 comments6 min readLW link

False as­sump­tions and leaky ab­strac­tions in ma­chine learn­ing and AI safety

David Scott Krueger (formerly: capybaralet)28 Jun 2019 4:54 UTC
21 points
3 comments1 min readLW link

Pro­ject Pro­posal: Con­sid­er­a­tions for trad­ing off ca­pa­bil­ities and safety im­pacts of AI research

David Scott Krueger (formerly: capybaralet)6 Aug 2019 22:22 UTC
25 points
11 comments2 min readLW link

[Question] What are the rea­sons to *not* con­sider re­duc­ing AI-Xrisk the high­est pri­or­ity cause?

David Scott Krueger (formerly: capybaralet)20 Aug 2019 21:45 UTC
29 points
27 comments1 min readLW link

[Question] Can in­differ­ence meth­ods re­deem per­son-af­fect­ing views?

David Scott Krueger (formerly: capybaralet)12 Nov 2019 4:23 UTC
10 points
3 comments1 min readLW link

A fun cal­ibra­tion game: “0-hit Google phrases”

David Scott Krueger (formerly: capybaralet)21 Nov 2019 1:13 UTC
6 points
1 comment1 min readLW link

What I talk about when I talk about AI x-risk: 3 core claims I want ma­chine learn­ing re­searchers to ad­dress.

David Scott Krueger (formerly: capybaralet)2 Dec 2019 18:20 UTC
29 points
13 comments3 min readLW link