RSS

AI Risk

TagLast edit: 16 Jul 2020 10:29 UTC by Ben Pace

AI Risk is analysis of the risks associated with building powerful AI systems.

What failure looks like

paulfchristiano17 Mar 2019 20:18 UTC
253 points
48 comments8 min readLW link2 nominations2 reviews

Su­per­in­tel­li­gence FAQ

Scott Alexander20 Sep 2016 19:00 UTC
58 points
11 comments27 min readLW link

Speci­fi­ca­tion gam­ing ex­am­ples in AI

Vika3 Apr 2018 12:30 UTC
39 points
9 comments1 min readLW link

In­tu­itions about goal-di­rected behavior

rohinmshah1 Dec 2018 4:25 UTC
46 points
15 comments6 min readLW link

Episte­molog­i­cal Fram­ing for AI Align­ment Research

adamShimi8 Mar 2021 22:05 UTC
50 points
6 comments9 min readLW link

What can the prin­ci­pal-agent liter­a­ture tell us about AI risk?

Alexis Carlier8 Feb 2020 21:28 UTC
99 points
31 comments16 min readLW link

Devel­op­men­tal Stages of GPTs

orthonormal26 Jul 2020 22:03 UTC
131 points
73 comments7 min readLW link

[Question] Will OpenAI’s work un­in­ten­tion­ally in­crease ex­is­ten­tial risks re­lated to AI?

adamShimi11 Aug 2020 18:16 UTC
49 points
54 comments1 min readLW link

Another (outer) al­ign­ment failure story

paulfchristiano7 Apr 2021 20:12 UTC
183 points
37 comments12 min readLW link

How good is hu­man­ity at co­or­di­na­tion?

Buck21 Jul 2020 20:01 UTC
73 points
43 comments3 min readLW link

A Gym Grid­world En­vi­ron­ment for the Treach­er­ous Turn

Michaël Trazzi28 Jul 2018 21:27 UTC
68 points
9 comments3 min readLW link
(github.com)

Are min­i­mal cir­cuits de­cep­tive?

evhub7 Sep 2019 18:11 UTC
55 points
8 comments8 min readLW link

Soft take­off can still lead to de­ci­sive strate­gic advantage

Daniel Kokotajlo23 Aug 2019 16:39 UTC
113 points
46 comments8 min readLW link2 nominations4 reviews

Should we post­pone AGI un­til we reach safety?

otto.barten18 Nov 2020 15:43 UTC
24 points
36 comments3 min readLW link

Cri­tiquing “What failure looks like”

Grue_Slinky27 Dec 2019 23:59 UTC
35 points
6 comments3 min readLW link

The Main Sources of AI Risk?

21 Mar 2019 18:28 UTC
75 points
22 comments2 min readLW link

Clar­ify­ing some key hy­pothe­ses in AI alignment

15 Aug 2019 21:29 UTC
77 points
11 comments9 min readLW link

“Tak­ing AI Risk Se­ri­ously” (thoughts by Critch)

Raemon29 Jan 2018 9:27 UTC
109 points
68 comments13 min readLW link

Some con­cep­tual high­lights from “Disjunc­tive Sce­nar­ios of Catas­trophic AI Risk”

Kaj_Sotala12 Feb 2018 12:30 UTC
29 points
4 comments6 min readLW link
(kajsotala.fi)

Non-Ad­ver­sar­ial Good­hart and AI Risks

Davidmanheim27 Mar 2018 1:39 UTC
22 points
9 comments6 min readLW link

Six AI Risk/​Strat­egy Ideas

Wei_Dai27 Aug 2019 0:40 UTC
63 points
18 comments4 min readLW link2 nominations1 review

[Question] Did AI pi­o­neers not worry much about AI risks?

lisperati9 Feb 2020 19:58 UTC
42 points
9 comments1 min readLW link

Some dis­junc­tive rea­sons for ur­gency on AI risk

Wei_Dai15 Feb 2019 20:43 UTC
36 points
24 comments1 min readLW link

Drexler on AI Risk

PeterMcCluskey1 Feb 2019 5:11 UTC
34 points
10 comments9 min readLW link
(www.bayesianinvestor.com)

A shift in ar­gu­ments for AI risk

Richard_Ngo28 May 2019 13:47 UTC
32 points
7 comments1 min readLW link
(fragile-credences.github.io)

Disen­tan­gling ar­gu­ments for the im­por­tance of AI safety

Richard_Ngo21 Jan 2019 12:41 UTC
123 points
23 comments8 min readLW link

AI Safety “Suc­cess Sto­ries”

Wei_Dai7 Sep 2019 2:54 UTC
105 points
27 comments4 min readLW link2 nominations1 review

De­bate on In­stru­men­tal Con­ver­gence be­tween LeCun, Rus­sell, Ben­gio, Zador, and More

Ben Pace4 Oct 2019 4:08 UTC
177 points
57 comments15 min readLW link2 nominations2 reviews

[AN #80]: Why AI risk might be solved with­out ad­di­tional in­ter­ven­tion from longtermists

rohinmshah2 Jan 2020 18:20 UTC
35 points
93 comments10 min readLW link
(mailchi.mp)

The strat­egy-steal­ing assumption

paulfchristiano16 Sep 2019 15:23 UTC
68 points
46 comments12 min readLW link2 nominations3 reviews

Think­ing soberly about the con­text and con­se­quences of Friendly AI

Mitchell_Porter16 Oct 2012 4:33 UTC
20 points
39 comments1 min readLW link

An­nounce­ment: AI al­ign­ment prize win­ners and next round

cousin_it15 Jan 2018 14:33 UTC
80 points
68 comments2 min readLW link

What Failure Looks Like: Distill­ing the Discussion

Ben Pace29 Jul 2020 21:49 UTC
74 points
12 comments7 min readLW link

Uber Self-Driv­ing Crash

jefftk7 Nov 2019 15:00 UTC
110 points
1 comment2 min readLW link
(www.jefftk.com)

Re­ply to Holden on ‘Tool AI’

Eliezer Yudkowsky12 Jun 2012 18:00 UTC
146 points
357 comments17 min readLW link

Stan­ford En­cy­clo­pe­dia of Philos­o­phy on AI ethics and superintelligence

Kaj_Sotala2 May 2020 7:35 UTC
43 points
19 comments7 min readLW link
(plato.stanford.edu)

AGI Safety Liter­a­ture Re­view (Ever­itt, Lea & Hut­ter 2018)

Kaj_Sotala4 May 2018 8:56 UTC
13 points
1 comment1 min readLW link
(arxiv.org)

Re­sponse to Oren Etz­ioni’s “How to know if ar­tifi­cial in­tel­li­gence is about to de­stroy civ­i­liza­tion”

Daniel Kokotajlo27 Feb 2020 18:10 UTC
27 points
5 comments8 min readLW link

Why don’t sin­gu­lar­i­tar­i­ans bet on the cre­ation of AGI by buy­ing stocks?

John_Maxwell11 Mar 2020 16:27 UTC
41 points
19 comments4 min readLW link

The prob­lem/​solu­tion ma­trix: Calcu­lat­ing the prob­a­bil­ity of AI safety “on the back of an en­velope”

John_Maxwell20 Oct 2019 8:03 UTC
22 points
4 comments2 min readLW link

Three Sto­ries for How AGI Comes Be­fore FAI

John_Maxwell17 Sep 2019 23:26 UTC
27 points
8 comments6 min readLW link

Brain­storm­ing ad­di­tional AI risk re­duc­tion ideas

John_Maxwell14 Jun 2012 7:55 UTC
19 points
37 comments1 min readLW link

AI Align­ment 2018-19 Review

rohinmshah28 Jan 2020 2:19 UTC
121 points
6 comments35 min readLW link

The Fu­sion Power Gen­er­a­tor Scenario

johnswentworth8 Aug 2020 18:31 UTC
109 points
25 comments3 min readLW link

A guide to Iter­ated Am­plifi­ca­tion & Debate

Rafael Harth15 Nov 2020 17:14 UTC
59 points
8 comments15 min readLW link

Work on Se­cu­rity In­stead of Friendli­ness?

Wei_Dai21 Jul 2012 18:28 UTC
49 points
107 comments2 min readLW link

An un­al­igned benchmark

paulfchristiano17 Nov 2018 15:51 UTC
29 points
0 comments9 min readLW link

Ar­tifi­cial In­tel­li­gence: A Modern Ap­proach (4th edi­tion) on the Align­ment Problem

Zack_M_Davis17 Sep 2020 2:23 UTC
72 points
12 comments5 min readLW link
(aima.cs.berkeley.edu)

Clar­ify­ing “What failure looks like” (part 1)

Sam Clarke20 Sep 2020 20:40 UTC
82 points
13 comments17 min readLW link

Re­laxed ad­ver­sar­ial train­ing for in­ner alignment

evhub10 Sep 2019 23:03 UTC
54 points
10 comments27 min readLW link

An overview of 11 pro­pos­als for build­ing safe ad­vanced AI

evhub29 May 2020 20:38 UTC
158 points
30 comments38 min readLW link

Risks from Learned Op­ti­miza­tion: Introduction

31 May 2019 23:44 UTC
142 points
40 comments12 min readLW link3 nominations3 reviews

Risks from Learned Op­ti­miza­tion: Con­clu­sion and Re­lated Work

7 Jun 2019 19:53 UTC
70 points
4 comments6 min readLW link

De­cep­tive Alignment

5 Jun 2019 20:16 UTC
69 points
11 comments17 min readLW link

The In­ner Align­ment Problem

4 Jun 2019 1:20 UTC
76 points
17 comments13 min readLW link

Con­di­tions for Mesa-Optimization

1 Jun 2019 20:52 UTC
62 points
47 comments12 min readLW link

AI risk hub in Sin­ga­pore?

Daniel Kokotajlo29 Oct 2020 11:45 UTC
50 points
18 comments4 min readLW link

Thoughts on Robin Han­son’s AI Im­pacts interview

Steven Byrnes24 Nov 2019 1:40 UTC
25 points
3 comments7 min readLW link

The AI Safety Game (UPDATED)

Daniel Kokotajlo5 Dec 2020 10:27 UTC
40 points
5 comments3 min readLW link

[Question] Sugges­tions of posts on the AF to review

adamShimi16 Feb 2021 12:40 UTC
51 points
20 comments1 min readLW link

Google’s Eth­i­cal AI team and AI Safety

magfrump20 Feb 2021 9:42 UTC
12 points
15 comments7 min readLW link

Be­hav­ioral Suffi­cient Statis­tics for Goal-Directedness

adamShimi11 Mar 2021 15:01 UTC
21 points
12 comments9 min readLW link

Re­view of “Fun with +12 OOMs of Com­pute”

28 Mar 2021 14:55 UTC
55 points
18 comments8 min readLW link

April drafts

AI Impacts1 Apr 2021 18:10 UTC
49 points
2 comments1 min readLW link
(aiimpacts.org)

25 Min Talk on Me­taEth­i­cal.AI with Ques­tions from Stu­art Armstrong

June Ku29 Apr 2021 15:38 UTC
17 points
5 comments1 min readLW link

Less Real­is­tic Tales of Doom

Mark Xu6 May 2021 23:01 UTC
94 points
13 comments4 min readLW link

Rogue AGI Em­bod­ies Valuable In­tel­lec­tual Property

3 Jun 2021 20:37 UTC
69 points
9 comments3 min readLW link

En­vi­ron­men­tal Struc­ture Can Cause In­stru­men­tal Convergence

TurnTrout22 Jun 2021 22:26 UTC
69 points
30 comments16 min readLW link
(arxiv.org)

Alex Turner’s Re­search, Com­pre­hen­sive In­for­ma­tion Gathering

adamShimi23 Jun 2021 9:44 UTC
15 points
3 comments3 min readLW link

Sam Alt­man and Ezra Klein on the AI Revolution

Zack_M_Davis27 Jun 2021 4:53 UTC
37 points
16 comments1 min readLW link
(www.nytimes.com)

Ac­cess to AI: a hu­man right?

dmtea25 Jul 2020 9:38 UTC
5 points
3 comments2 min readLW link

Agen­tic Lan­guage Model Memes

FactorialCode1 Aug 2020 18:03 UTC
16 points
1 comment2 min readLW link

Con­ver­sa­tion with Paul Christiano

abergal11 Sep 2019 23:20 UTC
44 points
6 comments30 min readLW link
(aiimpacts.org)

Tran­scrip­tion of Eliezer’s Jan­uary 2010 video Q&A

curiousepic14 Nov 2011 17:02 UTC
109 points
9 comments56 min readLW link

Re­sponses to Catas­trophic AGI Risk: A Survey

lukeprog8 Jul 2013 14:33 UTC
17 points
8 comments1 min readLW link

How can I re­duce ex­is­ten­tial risk from AI?

lukeprog13 Nov 2012 21:56 UTC
60 points
92 comments8 min readLW link

Thoughts on Ben Garfinkel’s “How sure are we about this AI stuff?”

capybaralet6 Feb 2019 19:09 UTC
25 points
17 comments1 min readLW link

Refram­ing mis­al­igned AGI’s: well-in­ten­tioned non-neu­rotyp­i­cal assistants

zhukeepa1 Apr 2018 1:22 UTC
46 points
14 comments2 min readLW link

When is un­al­igned AI morally valuable?

paulfchristiano25 May 2018 1:57 UTC
57 points
52 comments10 min readLW link

In­tro­duc­ing the AI Align­ment Fo­rum (FAQ)

29 Oct 2018 21:07 UTC
86 points
8 comments6 min readLW link

Swim­ming Up­stream: A Case Study in In­stru­men­tal Rationality

TurnTrout3 Jun 2018 3:16 UTC
64 points
7 comments8 min readLW link

Cur­rent AI Safety Roles for Soft­ware Engineers

ozziegooen9 Nov 2018 20:57 UTC
69 points
9 comments4 min readLW link

[Question] Why is so much dis­cus­sion hap­pen­ing in pri­vate Google Docs?

Wei_Dai12 Jan 2019 2:19 UTC
93 points
21 comments1 min readLW link

Prob­lems in AI Align­ment that philoso­phers could po­ten­tially con­tribute to

Wei_Dai17 Aug 2019 17:38 UTC
71 points
14 comments2 min readLW link

Two Ne­glected Prob­lems in Hu­man-AI Safety

Wei_Dai16 Dec 2018 22:13 UTC
73 points
23 comments2 min readLW link

An­nounce­ment: AI al­ign­ment prize round 4 winners

cousin_it20 Jan 2019 14:46 UTC
74 points
41 comments1 min readLW link

Soon: a weekly AI Safety pre­req­ui­sites mod­ule on LessWrong

toonalfrink30 Apr 2018 13:23 UTC
35 points
10 comments1 min readLW link

And the AI would have got away with it too, if...

Stuart_Armstrong22 May 2019 21:35 UTC
75 points
7 comments1 min readLW link

2017 AI Safety Liter­a­ture Re­view and Char­ity Com­par­i­son

Larks24 Dec 2017 18:52 UTC
41 points
5 comments23 min readLW link

Should ethi­cists be in­side or out­side a pro­fes­sion?

Eliezer Yudkowsky12 Dec 2018 1:40 UTC
79 points
6 comments9 min readLW link

I Vouch For MIRI

Zvi17 Dec 2017 17:50 UTC
34 points
9 comments5 min readLW link
(thezvi.wordpress.com)

Be­ware of black boxes in AI al­ign­ment research

cousin_it18 Jan 2018 15:07 UTC
39 points
10 comments1 min readLW link

AI Align­ment Prize: Round 2 due March 31, 2018

Zvi12 Mar 2018 12:10 UTC
28 points
2 comments3 min readLW link
(thezvi.wordpress.com)

Three AI Safety Re­lated Ideas

Wei_Dai13 Dec 2018 21:32 UTC
62 points
38 comments2 min readLW link

A rant against robots

Lê Nguyên Hoang14 Jan 2020 22:03 UTC
60 points
7 comments5 min readLW link

Op­por­tu­ni­ties for in­di­vi­d­ual donors in AI safety

alexflint31 Mar 2018 18:37 UTC
30 points
3 comments11 min readLW link

But ex­actly how com­plex and frag­ile?

KatjaGrace3 Nov 2019 18:20 UTC
65 points
32 comments3 min readLW link2 nominations1 review
(meteuphoric.com)

Course recom­men­da­tions for Friendli­ness researchers

Louie9 Jan 2013 14:33 UTC
92 points
112 comments10 min readLW link

AI Safety Re­search Camp—Pro­ject Proposal

David_Kristoffersson2 Feb 2018 4:25 UTC
29 points
11 comments8 min readLW link

AI Sum­mer Fel­lows Program

colm21 Mar 2018 15:32 UTC
21 points
0 comments1 min readLW link

The ge­nie knows, but doesn’t care

Rob Bensinger6 Sep 2013 6:42 UTC
88 points
519 comments8 min readLW link

Align­ment Newslet­ter #13: 07/​02/​18

rohinmshah2 Jul 2018 16:10 UTC
70 points
12 comments8 min readLW link
(mailchi.mp)

An In­creas­ingly Ma­nipu­la­tive Newsfeed

Michaël Trazzi1 Jul 2019 15:26 UTC
58 points
14 comments5 min readLW link

The sim­ple pic­ture on AI safety

alexflint27 May 2018 19:43 UTC
25 points
10 comments2 min readLW link

Elon Musk donates $10M to the Fu­ture of Life In­sti­tute to keep AI benefi­cial

Paul Crowley15 Jan 2015 16:33 UTC
78 points
52 comments1 min readLW link

Strate­gic im­pli­ca­tions of AIs’ abil­ity to co­or­di­nate at low cost, for ex­am­ple by merging

Wei_Dai25 Apr 2019 5:08 UTC
60 points
45 comments2 min readLW link2 nominations1 review

Model­ing AGI Safety Frame­works with Causal In­fluence Diagrams

Ramana Kumar21 Jun 2019 12:50 UTC
43 points
6 comments1 min readLW link
(arxiv.org)

Henry Kiss­inger: AI Could Mean the End of Hu­man History

ESRogs15 May 2018 20:11 UTC
17 points
12 comments1 min readLW link
(www.theatlantic.com)

Toy model of the AI con­trol prob­lem: an­i­mated version

Stuart_Armstrong10 Oct 2017 11:06 UTC
25 points
8 comments1 min readLW link

A Vi­su­al­iza­tion of Nick Bostrom’s Superintelligence

[deleted]23 Jul 2014 0:24 UTC
62 points
28 comments3 min readLW link

AI Align­ment Re­search Overview (by Ja­cob Stein­hardt)

Ben Pace6 Nov 2019 19:24 UTC
43 points
0 comments7 min readLW link
(docs.google.com)

A gen­eral model of safety-ori­ented AI development

Wei_Dai11 Jun 2018 21:00 UTC
65 points
8 comments1 min readLW link

Coun­ter­fac­tual Or­a­cles = on­line su­per­vised learn­ing with ran­dom se­lec­tion of train­ing episodes

Wei_Dai10 Sep 2019 8:29 UTC
44 points
26 comments3 min readLW link

Siren wor­lds and the per­ils of over-op­ti­mised search

Stuart_Armstrong7 Apr 2014 11:00 UTC
69 points
417 comments7 min readLW link

Top 9+2 myths about AI risk

Stuart_Armstrong29 Jun 2015 20:41 UTC
64 points
46 comments2 min readLW link

Ro­hin Shah on rea­sons for AI optimism

abergal31 Oct 2019 12:10 UTC
40 points
58 comments1 min readLW link
(aiimpacts.org)

Plau­si­bly, al­most ev­ery pow­er­ful al­gorithm would be manipulative

Stuart_Armstrong6 Feb 2020 11:50 UTC
38 points
25 comments3 min readLW link

The Mag­ni­tude of His Own Folly

Eliezer Yudkowsky30 Sep 2008 11:31 UTC
52 points
128 comments6 min readLW link

AI al­ign­ment landscape

paulfchristiano13 Oct 2019 2:10 UTC
40 points
3 comments1 min readLW link
(ai-alignment.com)

Launched: Friend­ship is Optimal

iceman15 Nov 2012 4:57 UTC
63 points
31 comments1 min readLW link

Friend­ship is Op­ti­mal: A My Lit­tle Pony fan­fic about an op­ti­miza­tion process

iceman8 Sep 2012 6:16 UTC
98 points
152 comments1 min readLW link

Do Earths with slower eco­nomic growth have a bet­ter chance at FAI?

Eliezer Yudkowsky12 Jun 2013 19:54 UTC
54 points
176 comments4 min readLW link

Idea: Open Ac­cess AI Safety Journal

G Gordon Worley III23 Mar 2018 18:27 UTC
28 points
11 comments1 min readLW link

G.K. Ch­ester­ton On AI Risk

Scott Alexander1 Apr 2017 19:00 UTC
12 points
0 comments7 min readLW link

The Hid­den Com­plex­ity of Wishes

Eliezer Yudkowsky24 Nov 2007 0:12 UTC
112 points
135 comments7 min readLW link

The Friendly AI Game

bentarm15 Mar 2011 16:45 UTC
50 points
178 comments1 min readLW link

Q&A with Jür­gen Sch­mid­hu­ber on risks from AI

XiXiDu15 Jun 2011 15:51 UTC
54 points
45 comments4 min readLW link

[Question] What should an Ein­stein-like figure in Ma­chine Learn­ing do?

Razied5 Aug 2020 23:52 UTC
3 points
3 comments1 min readLW link

Take­aways from safety by de­fault interviews

3 Apr 2020 17:20 UTC
23 points
2 comments13 min readLW link
(aiimpacts.org)

Field-Build­ing and Deep Models

Ben Pace13 Jan 2018 21:16 UTC
21 points
12 comments4 min readLW link

Cri­tique my Model: The EV of AGI to Selfish Individuals

ozziegooen8 Apr 2018 20:04 UTC
19 points
9 comments4 min readLW link

‘Dumb’ AI ob­serves and ma­nipu­lates controllers

Stuart_Armstrong13 Jan 2015 13:35 UTC
52 points
19 comments2 min readLW link

2019 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks19 Dec 2019 3:00 UTC
130 points
18 comments62 min readLW link

Book re­view: Ar­chi­tects of In­tel­li­gence by Martin Ford (2018)

ofer11 Aug 2020 17:30 UTC
15 points
0 comments2 min readLW link

Qual­i­ta­tive Strate­gies of Friendliness

Eliezer Yudkowsky30 Aug 2008 2:12 UTC
13 points
56 comments12 min readLW link

Dreams of Friendliness

Eliezer Yudkowsky31 Aug 2008 1:20 UTC
24 points
80 comments9 min readLW link

Con­cep­tual is­sues in AI safety: the paradig­matic gap

vedevazz24 Jun 2018 15:09 UTC
33 points
0 comments1 min readLW link
(www.foldl.me)

On un­fix­ably un­safe AGI architectures

Steven Byrnes19 Feb 2020 21:16 UTC
30 points
8 comments5 min readLW link

A toy model of the treach­er­ous turn

Stuart_Armstrong8 Jan 2016 12:58 UTC
35 points
13 comments6 min readLW link

Alle­gory On AI Risk, Game The­ory, and Mithril

James_Miller13 Feb 2017 20:41 UTC
41 points
57 comments3 min readLW link

1hr talk: In­tro to AGI safety

Steven Byrnes18 Jun 2019 21:41 UTC
33 points
4 comments24 min readLW link

The Evil AI Over­lord List

Stuart_Armstrong20 Nov 2012 17:02 UTC
44 points
80 comments1 min readLW link

What I would like the SIAI to publish

XiXiDu1 Nov 2010 14:07 UTC
36 points
225 comments3 min readLW link

Eval­u­at­ing the fea­si­bil­ity of SI’s plan

JoshuaFox10 Jan 2013 8:17 UTC
38 points
188 comments4 min readLW link

Q&A with ex­perts on risks from AI #1

XiXiDu8 Jan 2012 11:46 UTC
45 points
67 comments9 min readLW link

Algo trad­ing is a cen­tral ex­am­ple of AI risk

Vanessa Kosoy28 Jul 2018 20:31 UTC
24 points
5 comments1 min readLW link

Will the world’s elites nav­i­gate the cre­ation of AI just fine?

lukeprog31 May 2013 18:49 UTC
36 points
266 comments2 min readLW link

Let’s talk about “Con­ver­gent Ra­tion­al­ity”

capybaralet12 Jun 2019 21:53 UTC
35 points
33 comments6 min readLW link

Break­ing Or­a­cles: su­per­ra­tional­ity and acausal trade

Stuart_Armstrong25 Nov 2019 10:40 UTC
25 points
15 comments1 min readLW link

Q&A with Stan Fran­klin on risks from AI

XiXiDu11 Jun 2011 15:22 UTC
36 points
10 comments2 min readLW link

Muehlhauser-Go­ertzel Dialogue, Part 1

lukeprog16 Mar 2012 17:12 UTC
42 points
161 comments33 min readLW link

[LINK] NYT Ar­ti­cle about Ex­is­ten­tial Risk from AI

[deleted]28 Jan 2013 10:37 UTC
38 points
23 comments1 min readLW link

Refram­ing the Prob­lem of AI Progress

Wei_Dai12 Apr 2012 19:31 UTC
32 points
47 comments1 min readLW link

New AI risks re­search in­sti­tute at Oxford University

lukeprog16 Nov 2011 18:52 UTC
36 points
10 comments1 min readLW link

Thoughts on the Fea­si­bil­ity of Pro­saic AGI Align­ment?

iamthouthouarti21 Aug 2020 23:25 UTC
8 points
10 comments1 min readLW link

Memes and Ra­tional Decisions

inferential9 Jan 2015 6:42 UTC
35 points
17 comments10 min readLW link

Levels of AI Self-Im­prove­ment

avturchin29 Apr 2018 11:45 UTC
9 points
0 comments39 min readLW link

Op­ti­mis­ing So­ciety to Con­strain Risk of War from an Ar­tifi­cial Su­per­in­tel­li­gence

JohnCDraper30 Apr 2020 10:47 UTC
3 points
0 comments51 min readLW link

Some Thoughts on Sin­gu­lar­ity Strategies

Wei_Dai13 Jul 2011 2:41 UTC
37 points
29 comments3 min readLW link

A trick for Safer GPT-N

Razied23 Aug 2020 0:39 UTC
7 points
1 comment2 min readLW link

against “AI risk”

Wei_Dai11 Apr 2012 22:46 UTC
35 points
91 comments1 min readLW link

“Smarter than us” is out!

Stuart_Armstrong25 Feb 2014 15:50 UTC
41 points
57 comments1 min readLW link

Analysing: Danger­ous mes­sages from fu­ture UFAI via Oracles

Stuart_Armstrong22 Nov 2019 14:17 UTC
22 points
16 comments4 min readLW link

Q&A with Abram Dem­ski on risks from AI

XiXiDu17 Jan 2012 9:43 UTC
33 points
71 comments9 min readLW link

Q&A with ex­perts on risks from AI #2

XiXiDu9 Jan 2012 19:40 UTC
22 points
29 comments7 min readLW link

AI Safety Dis­cus­sion Day

Linda Linsefors15 Sep 2020 14:40 UTC
20 points
0 comments1 min readLW link

A long re­ply to Ben Garfinkel on Scru­ti­niz­ing Clas­sic AI Risk Arguments

Søren Elverlin27 Sep 2020 17:51 UTC
16 points
6 comments1 min readLW link

On­line AI Safety Dis­cus­sion Day

Linda Linsefors8 Oct 2020 12:11 UTC
5 points
0 comments1 min readLW link

Mili­tary AI as a Con­ver­gent Goal of Self-Im­prov­ing AI

avturchin13 Nov 2017 12:17 UTC
5 points
3 comments1 min readLW link

Neu­ral pro­gram syn­the­sis is a dan­ger­ous technology

syllogism12 Jan 2018 16:19 UTC
9 points
6 comments2 min readLW link

New, Brief Pop­u­lar-Level In­tro­duc­tion to AI Risks and Superintelligence

LyleN23 Jan 2015 15:43 UTC
33 points
3 comments1 min readLW link

FAI Re­search Con­straints and AGI Side Effects

JustinShovelain3 Jun 2015 19:25 UTC
26 points
59 comments7 min readLW link

Euro­pean Master’s Pro­grams in Ma­chine Learn­ing, Ar­tifi­cial In­tel­li­gence, and re­lated fields

Master Programs ML/AI14 Nov 2020 15:51 UTC
25 points
8 comments1 min readLW link

The mind-killer

Paul Crowley2 May 2009 16:49 UTC
29 points
160 comments2 min readLW link

[Question] Should I do it?

MrLight19 Nov 2020 1:08 UTC
−3 points
16 comments2 min readLW link

Ra­tion­al­is­ing hu­mans: an­other mug­ging, but not Pas­cal’s

Stuart_Armstrong14 Nov 2017 15:46 UTC
7 points
1 comment3 min readLW link

Ma­chine learn­ing could be fun­da­men­tally unexplainable

George16 Dec 2020 13:32 UTC
25 points
15 comments15 min readLW link
(cerebralab.com)

[Question] What do you make of AGI:un­al­igned::space­ships:not enough food?

Ronny22 Feb 2020 14:14 UTC
4 points
3 comments1 min readLW link

Risk Map of AI Systems

15 Dec 2020 9:16 UTC
25 points
3 comments8 min readLW link

Edge of the Cliff

akaTrickster5 Jan 2021 17:21 UTC
1 point
0 comments5 min readLW link

[Question] Does it be­come eas­ier, or harder, for the world to co­or­di­nate around not build­ing AGI as time goes on?

Eli Tyre29 Jul 2019 22:59 UTC
86 points
31 comments3 min readLW link2 nominations2 reviews

Grey Goo Re­quires AI

harsimony15 Jan 2021 4:45 UTC
8 points
11 comments4 min readLW link
(harsimony.wordpress.com)

AISU 2021

Linda Linsefors30 Jan 2021 17:40 UTC
28 points
2 comments1 min readLW link

Non­per­son Predicates

Eliezer Yudkowsky27 Dec 2008 1:47 UTC
44 points
176 comments6 min readLW link

En­gag­ing First In­tro­duc­tions to AI Risk

Rob Bensinger19 Aug 2013 6:26 UTC
31 points
21 comments3 min readLW link

For­mal Solu­tion to the In­ner Align­ment Problem

michaelcohen18 Feb 2021 14:51 UTC
46 points
123 comments2 min readLW link

[Question] What are the biggest cur­rent im­pacts of AI?

Sam Clarke7 Mar 2021 21:44 UTC
15 points
4 comments1 min readLW link

[Question] Is a Self-Iter­at­ing AGI Vuln­er­a­ble to Thomp­son-style Tro­jans?

sxae25 Mar 2021 14:46 UTC
15 points
7 comments3 min readLW link

AI or­a­cles on blockchain

Caravaggio6 Apr 2021 20:13 UTC
4 points
0 comments3 min readLW link

What if AGI is near?

Wulky Wilkinsen14 Apr 2021 0:05 UTC
11 points
5 comments1 min readLW link

[Question] Is there any­thing that can stop AGI de­vel­op­ment in the near term?

Wulky Wilkinsen22 Apr 2021 20:37 UTC
4 points
5 comments1 min readLW link

[Question] [time­boxed ex­er­cise] write me your model of AI hu­man-ex­is­ten­tial safety and the al­ign­ment prob­lems in 15 minutes

Quinn4 May 2021 19:10 UTC
6 points
2 comments1 min readLW link

AI Safety Re­search Pro­ject Ideas

Owain_Evans21 May 2021 13:39 UTC
56 points
1 comment3 min readLW link

Sur­vey on AI ex­is­ten­tial risk scenarios

8 Jun 2021 17:12 UTC
54 points
10 comments7 min readLW link

[Question] What are some claims or opinions about multi-multi del­e­ga­tion you’ve seen in the meme­plex that you think de­serve scrutiny?

Quinn27 Jun 2021 17:44 UTC
16 points
6 comments2 min readLW link

Mauhn Re­leases AI Safety Documentation

Berg Severens3 Jul 2021 21:23 UTC
4 points
0 comments1 min readLW link
No comments.