RSS

AI Governance

TagLast edit: 9 Aug 2020 18:31 UTC by Gyrodiot

AI Governance asks how we can ensure society benefits at large from increasingly powerful AI systems. While solving technical AI alignment is a necessary step towards this goal, it is by no means sufficient.

Governance includes policy, economics, sociology, law, and many other fields.

What an ac­tu­ally pes­simistic con­tain­ment strat­egy looks like

lc5 Apr 2022 0:19 UTC
611 points
134 comments6 min readLW link

What would a com­pute mon­i­tor­ing plan look like? [Linkpost]

Akash26 Mar 2023 19:33 UTC
155 points
9 comments4 min readLW link
(arxiv.org)

Some cruxes on im­pact­ful al­ter­na­tives to AI policy work

Richard_Ngo10 Oct 2018 13:35 UTC
163 points
13 comments12 min readLW link

Soft take­off can still lead to de­ci­sive strate­gic advantage

Daniel Kokotajlo23 Aug 2019 16:39 UTC
121 points
46 comments8 min readLW link4 reviews

List of re­quests for an AI slow­down/​halt.

Cleo Nardo14 Apr 2023 23:55 UTC
46 points
6 comments1 min readLW link

Should we post­pone AGI un­til we reach safety?

otto.barten18 Nov 2020 15:43 UTC
27 points
36 comments3 min readLW link

Where are the red lines for AI?

Karl von Wendt5 Aug 2022 9:34 UTC
26 points
8 comments6 min readLW link

Ngo’s view on al­ign­ment difficulty

14 Dec 2021 21:34 UTC
63 points
7 comments17 min readLW link

[Question] Where are peo­ple think­ing and talk­ing about global co­or­di­na­tion for AI safety?

Wei_Dai22 May 2019 6:24 UTC
112 points
22 comments1 min readLW link

2019 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks19 Dec 2019 3:00 UTC
130 points
18 comments62 min readLW link

AI policy ideas: Read­ing list

Zach Stein-Perlman17 Apr 2023 19:00 UTC
19 points
7 comments4 min readLW link

The Reg­u­la­tory Op­tion: A re­sponse to near 0% sur­vival odds

Matthew Lowenstein11 Apr 2022 22:00 UTC
48 points
21 comments6 min readLW link

[Question] Would it be good or bad for the US mil­i­tary to get in­volved in AI risk?

Grant Demaree1 Jan 2023 19:02 UTC
50 points
12 comments1 min readLW link

An­nounc­ing Apollo Research

30 May 2023 16:17 UTC
188 points
7 comments8 min readLW link

He­len Toner on China, CSET, and AI

Rob Bensinger21 Apr 2019 4:10 UTC
68 points
3 comments7 min readLW link
(rationallyspeakingpodcast.org)

Ac­tion­able-guidance and roadmap recom­men­da­tions for the NIST AI Risk Man­age­ment Framework

17 May 2022 15:26 UTC
26 points
0 comments3 min readLW link

Policy dis­cus­sions fol­low strong con­tex­tu­al­iz­ing norms

Richard_Ngo1 Apr 2023 23:51 UTC
229 points
61 comments3 min readLW link

List #2: Why co­or­di­nat­ing to al­ign as hu­mans to not de­velop AGI is a lot eas­ier than, well… co­or­di­nat­ing as hu­mans with AGI co­or­di­nat­ing to be al­igned with humans

Remmelt24 Dec 2022 9:53 UTC
1 point
0 comments3 min readLW link

The case for Do­ing Some­thing Else (if Align­ment is doomed)

Rafael Harth5 Apr 2022 17:52 UTC
90 points
14 comments2 min readLW link

AI Sum­mer Harvest

Cleo Nardo4 Apr 2023 3:35 UTC
130 points
10 comments1 min readLW link

Strate­gic Con­sid­er­a­tions Re­gard­ing Autis­tic/​Literal AI

Chris_Leong6 Apr 2022 14:57 UTC
−1 points
2 comments2 min readLW link

Ex­ces­sive AI growth-rate yields lit­tle so­cio-eco­nomic benefit.

Cleo Nardo4 Apr 2023 19:13 UTC
27 points
22 comments4 min readLW link

Ngo and Yud­kowsky on AI ca­pa­bil­ity gains

18 Nov 2021 22:19 UTC
130 points
61 comments39 min readLW link1 review

I asked my sen­a­tor to slow AI

Omid6 Apr 2023 18:18 UTC
21 points
5 comments2 min readLW link

An ‘AGI Emer­gency Eject Cri­te­ria’ con­sen­sus could be re­ally use­ful.

tcelferact7 Apr 2023 16:21 UTC
5 points
0 comments1 min readLW link

All images from the WaitButWhy se­quence on AI

trevor8 Apr 2023 7:36 UTC
72 points
5 comments2 min readLW link

Cur­rent UK gov­ern­ment lev­ers on AI development

rosehadshar10 Apr 2023 13:16 UTC
16 points
0 comments1 min readLW link

Re­quest to AGI or­ga­ni­za­tions: Share your views on paus­ing AI progress

11 Apr 2023 17:30 UTC
140 points
11 comments1 min readLW link

FLI And Eliezer Should Reach Consensus

JenniferRM11 Apr 2023 4:07 UTC
17 points
6 comments23 min readLW link

Cy­berspace Ad­minis­tra­tion of China: Draft of “Reg­u­la­tion for Gen­er­a­tive Ar­tifi­cial In­tel­li­gence Ser­vices” is open for comments

sanxiyn11 Apr 2023 9:32 UTC
7 points
2 comments1 min readLW link
(archive.is)

NTIA—AI Ac­countabil­ity Announcement

samshap11 Apr 2023 15:03 UTC
7 points
0 comments1 min readLW link
(www.ntia.doc.gov)

Na­tional Telecom­mu­ni­ca­tions and In­for­ma­tion Ad­minis­tra­tion: AI Ac­countabil­ity Policy Re­quest for Comment

sanxiyn11 Apr 2023 22:59 UTC
9 points
0 comments1 min readLW link
(ntia.gov)

Sha­har Avin On How To Reg­u­late Ad­vanced AI Systems

Michaël Trazzi23 Sep 2022 15:46 UTC
31 points
0 comments4 min readLW link
(theinsideview.ai)

Nav­i­gat­ing the Open-Source AI Land­scape: Data, Fund­ing, and Safety

13 Apr 2023 15:29 UTC
32 points
7 comments11 min readLW link
(forum.effectivealtruism.org)

My thoughts on OpenAI’s al­ign­ment plan

Akash30 Dec 2022 19:33 UTC
54 points
2 comments20 min readLW link

FLI re­port: Poli­cy­mak­ing in the Pause

Zach Stein-Perlman15 Apr 2023 17:01 UTC
9 points
3 comments1 min readLW link
(futureoflife.org)

Un­der what cir­cum­stances have gov­ern­ments can­cel­led AI-type sys­tems?

David Gross23 Sep 2022 21:11 UTC
7 points
1 comment1 min readLW link
(www.carnegieuktrust.org.uk)

Slow­ing AI: Foundations

Zach Stein-Perlman17 Apr 2023 14:30 UTC
36 points
11 comments17 min readLW link

Re­spon­si­ble De­ploy­ment in 20XX

Carson20 Apr 2023 0:24 UTC
4 points
0 comments4 min readLW link

OpenAI could help X-risk by wa­ger­ing itself

VojtaKovarik20 Apr 2023 14:51 UTC
31 points
16 comments1 min readLW link

Went­worth and Larsen on buy­ing time

9 Jan 2023 21:31 UTC
73 points
6 comments12 min readLW link

My Assess­ment of the Chi­nese AI Safety Community

Lao Mein25 Apr 2023 4:21 UTC
239 points
92 comments3 min readLW link

Notes on Po­ten­tial Fu­ture AI Tax Policy

Zvi25 Apr 2023 13:30 UTC
33 points
5 comments9 min readLW link
(thezvi.wordpress.com)

Refram­ing the bur­den of proof: Com­pa­nies should prove that mod­els are safe (rather than ex­pect­ing au­di­tors to prove that mod­els are dan­ger­ous)

Akash25 Apr 2023 18:49 UTC
26 points
11 comments3 min readLW link
(childrenoficarus.substack.com)

Anal­y­sis: US re­stricts GPU sales to China

aogara7 Oct 2022 18:38 UTC
102 points
58 comments5 min readLW link

Com­ments on Allan Dafoe on AI Governance

Alex Flint29 Nov 2021 16:16 UTC
13 points
0 comments7 min readLW link

Thoughts on hard­ware /​ com­pute re­quire­ments for AGI

Steven Byrnes24 Jan 2023 14:03 UTC
44 points
30 comments21 min readLW link

Safety stan­dards: a frame­work for AI regulation

joshc1 May 2023 0:56 UTC
18 points
0 comments8 min readLW link

[Question] AI safety mile­stones?

Zach Stein-Perlman23 Jan 2023 21:00 UTC
7 points
5 comments1 min readLW link

AI Risk Man­age­ment Frame­work | NIST

DragonGod26 Jan 2023 15:27 UTC
36 points
4 comments2 min readLW link
(www.nist.gov)

[Question] Should we push for re­quiring AI train­ing data to be li­censed?

ChristianKl19 Oct 2022 17:49 UTC
37 points
32 comments1 min readLW link

Dario Amodei leaves OpenAI

Daniel Kokotajlo29 Dec 2020 19:31 UTC
69 points
12 comments1 min readLW link

What is the ground re­al­ity of coun­tries tak­ing steps to re­cal­ibrate AI de­vel­op­ment to­wards Align­ment first?

Nebuch29 Jan 2023 13:26 UTC
8 points
6 comments3 min readLW link

Learn­ing so­cietal val­ues from law as part of an AGI al­ign­ment strategy

John Nay21 Oct 2022 2:03 UTC
4 points
18 comments54 min readLW link

Product safety is a poor model for AI governance

Richard Korzekwa 1 Feb 2023 22:40 UTC
36 points
0 comments5 min readLW link
(aiimpacts.org)

Many AI gov­er­nance pro­pos­als have a trade­off be­tween use­ful­ness and feasibility

3 Feb 2023 18:49 UTC
22 points
2 comments2 min readLW link

Stop­ping dan­ger­ous AI: Ideal lab behavior

Zach Stein-Perlman9 May 2023 21:00 UTC
0 points
0 comments2 min readLW link

Stop­ping dan­ger­ous AI: Ideal US behavior

Zach Stein-Perlman9 May 2023 21:00 UTC
10 points
0 comments3 min readLW link

What does it take to defend the world against out-of-con­trol AGIs?

Steven Byrnes25 Oct 2022 14:47 UTC
156 points
37 comments30 min readLW link

4 ways to think about de­moc­ra­tiz­ing AI [GovAI Linkpost]

Akash13 Feb 2023 18:06 UTC
24 points
4 comments1 min readLW link
(www.governance.ai)

How should AI sys­tems be­have, and who should de­cide? [OpenAI blog]

ShardPhoenix17 Feb 2023 1:05 UTC
22 points
2 comments1 min readLW link
(openai.com)

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

22 Feb 2023 16:09 UTC
87 points
8 comments6 min readLW link

GovAI: Towards best prac­tices in AGI safety and gov­er­nance: A sur­vey of ex­pert opinion

Zach Stein-Perlman15 May 2023 1:42 UTC
27 points
10 comments1 min readLW link
(arxiv.org)

WFW?: Op­por­tu­nity and The­ory of Impact

DavidCorfield2 Nov 2022 1:24 UTC
1 point
0 comments1 min readLW link

AI Gover­nance & Strat­egy: Pri­ori­ties, tal­ent gaps, & opportunities

Akash3 Mar 2023 18:09 UTC
53 points
2 comments4 min readLW link

[Linkpost] “Gover­nance of su­per­in­tel­li­gence” by OpenAI

Daniel_Eth22 May 2023 20:15 UTC
67 points
20 comments1 min readLW link

An­nounc­ing: What Fu­ture World? - Grow­ing the AI Gover­nance Community

DavidCorfield2 Nov 2022 1:24 UTC
1 point
0 comments1 min readLW link

AI #12:The Quest for Sane Regulations

Zvi18 May 2023 13:20 UTC
77 points
12 comments64 min readLW link
(thezvi.wordpress.com)

The Na­tional Defense Autho­riza­tion Act Con­tains AI Provisions

ryan_b5 Jan 2021 15:51 UTC
30 points
24 comments1 min readLW link

State­ment on AI Ex­tinc­tion—Signed by AGI Labs, Top Aca­demics, and Many Other Notable Figures

Dan H30 May 2023 9:05 UTC
317 points
50 comments1 min readLW link
(www.safe.ai)

[Question] Who is li­able for AI?

jmh30 May 2023 13:54 UTC
8 points
4 comments1 min readLW link

The case for re­mov­ing al­ign­ment and ML re­search from the train­ing dataset

beren30 May 2023 20:54 UTC
46 points
8 comments5 min readLW link

Eisen­hower’s Atoms for Peace Speech

Akash17 May 2023 16:10 UTC
18 points
3 comments11 min readLW link
(www.iaea.org)

Why I Am Skep­ti­cal of AI Reg­u­la­tion as an X-Risk Miti­ga­tion Strategy

A Ray6 Aug 2022 5:46 UTC
31 points
14 comments2 min readLW link

Mas­sive Scal­ing Should be Frowned Upon

harsimony17 Nov 2022 8:43 UTC
4 points
6 comments5 min readLW link

[Linkpost] Scott Alexan­der re­acts to OpenAI’s lat­est post

Akash11 Mar 2023 22:24 UTC
27 points
0 comments5 min readLW link
(astralcodexten.substack.com)

Jack Clark on the re­al­ities of AI policy

Kaj_Sotala7 Aug 2022 8:44 UTC
66 points
3 comments3 min readLW link
(threadreaderapp.com)

The Wizard of Oz Prob­lem: How in­cen­tives and nar­ra­tives can skew our per­cep­tion of AI developments

Akash20 Mar 2023 20:44 UTC
16 points
3 comments6 min readLW link

[Question] What if we solve AI Safety but no one cares

14285722 Aug 2022 5:38 UTC
18 points
5 comments1 min readLW link

Govern­ing High-Im­pact AI Sys­tems: Un­der­stand­ing Canada’s Pro­posed AI Bill. April 15, Car­leton Univer­sity, Ottawa

Liav Koren28 Mar 2023 17:48 UTC
11 points
1 comment1 min readLW link
(forum.effectivealtruism.org)

[Question] How promis­ing are le­gal av­enues to re­strict AI train­ing data?

thehalliard10 Dec 2022 16:31 UTC
9 points
2 comments1 min readLW link

How is AI gov­erned and reg­u­lated, around the world?

Mitchell_Porter30 Mar 2023 15:36 UTC
14 points
6 comments2 min readLW link

Google’s Eth­i­cal AI team and AI Safety

magfrump20 Feb 2021 9:42 UTC
12 points
16 comments7 min readLW link

ChatGPT banned in Italy over pri­vacy concerns

Ollie J31 Mar 2023 17:33 UTC
18 points
4 comments1 min readLW link
(www.bbc.co.uk)

Re­place­ment for PONR concept

Daniel Kokotajlo2 Sep 2022 0:09 UTC
48 points
6 comments2 min readLW link

[Question] What Are Your Prefer­ences Re­gard­ing The FLI Let­ter?

JenniferRM1 Apr 2023 4:52 UTC
−4 points
122 comments16 min readLW link

Prac­ti­cal AI risk I: Watch­ing large compute

Gustavo Ramires24 Dec 2022 13:25 UTC
3 points
0 comments1 min readLW link

Fore­sight for AGI Safety Strat­egy: Miti­gat­ing Risks and Iden­ti­fy­ing Golden Opportunities

jacquesthibs5 Dec 2022 16:09 UTC
16 points
4 comments8 min readLW link

Ex­is­ten­tial AI Safety is NOT sep­a­rate from near-term applications

scasper13 Dec 2022 14:47 UTC
37 points
17 comments3 min readLW link

AGI Timelines in Gover­nance: Differ­ent Strate­gies for Differ­ent Timeframes

19 Dec 2022 21:31 UTC
63 points
28 comments10 min readLW link

Ac­cu­rate Models of AI Risk Are Hyper­ex­is­ten­tial Exfohazards

Thane Ruthenis25 Dec 2022 16:50 UTC
30 points
38 comments9 min readLW link

In­sti­tu­tions Can­not Res­train Dark-Triad AI Exploitation

27 Dec 2022 10:34 UTC
5 points
0 comments5 min readLW link
(mflb.com)

Avoid­ing per­pet­ual risk from TAI

scasper26 Dec 2022 22:34 UTC
12 points
6 comments5 min readLW link

Should AI sys­tems have to iden­tify them­selves?

Darren McKee31 Dec 2022 2:57 UTC
2 points
2 comments1 min readLW link

AI se­cu­rity might be helpful for AI alignment

Igor Ivanov6 Jan 2023 20:16 UTC
35 points
1 comment2 min readLW link

Pro­tec­tion­ism will Slow the De­ploy­ment of AI

bgold7 Jan 2023 20:57 UTC
30 points
6 comments2 min readLW link

Align­ment is not enough

Alan Chan12 Jan 2023 0:33 UTC
11 points
6 comments11 min readLW link
(coordination.substack.com)

[Question] Should AI writ­ers be pro­hibited in ed­u­ca­tion?

Eleni Angelou17 Jan 2023 0:42 UTC
6 points
2 comments1 min readLW link

Emo­tional at­tach­ment to AIs opens doors to problems

Igor Ivanov22 Jan 2023 20:28 UTC
20 points
10 comments4 min readLW link

Ap­ply to HAIST/​MAIA’s AI Gover­nance Work­shop in DC (Feb 17-20)

31 Jan 2023 2:06 UTC
28 points
0 comments2 min readLW link

WaPo: “Big Tech was mov­ing cau­tiously on AI. Then came ChatGPT.”

Julian Bradshaw27 Jan 2023 22:54 UTC
26 points
5 comments1 min readLW link
(www.washingtonpost.com)

Trends in the dol­lar train­ing cost of ma­chine learn­ing systems

Ben Cottier1 Feb 2023 14:48 UTC
23 points
0 comments2 min readLW link
(epochai.org)

Se­cu­rity Mind­set—Fire Alarms and Trig­ger Signatures

elspood9 Feb 2023 21:15 UTC
23 points
0 comments4 min readLW link

Au­to­mated Sand­wich­ing & Quan­tify­ing Hu­man-LLM Co­op­er­a­tion: ScaleOver­sight hackathon results

23 Feb 2023 10:48 UTC
8 points
0 comments6 min readLW link

[Question] Would more model evals teams be good?

Ryan Kidd25 Feb 2023 22:01 UTC
20 points
4 comments1 min readLW link

[Link Post] Cy­ber Digi­tal Author­i­tar­i­anism (Na­tional In­tel­li­gence Coun­cil Re­port)

Phosphorous26 Feb 2023 20:51 UTC
9 points
2 comments1 min readLW link
(www.dni.gov)

Call for Cruxes by Rhyme, a Longter­mist His­tory Consultancy

Lara1 Mar 2023 18:39 UTC
1 point
0 comments3 min readLW link
(forum.effectivealtruism.org)

A con­cern­ing ob­ser­va­tion from me­dia cov­er­age of AI in­dus­try dynamics

Justin Olive5 Mar 2023 21:38 UTC
8 points
3 comments3 min readLW link

Cap Model Size for AI Safety

research_prime_space6 Mar 2023 1:11 UTC
0 points
4 comments1 min readLW link

Why Un­con­trol­lable AI Looks More Likely Than Ever

8 Mar 2023 15:41 UTC
18 points
0 comments4 min readLW link
(time.com)

An­thropic: Core Views on AI Safety: When, Why, What, and How

jonmenaster9 Mar 2023 17:34 UTC
17 points
1 comment22 min readLW link
(www.anthropic.com)

Sur­vey on in­ter­me­di­ate goals in AI governance

17 Mar 2023 13:12 UTC
25 points
3 comments1 min readLW link

Ex­plor­ing the Pre­cau­tion­ary Prin­ci­ple in AI Devel­op­ment: His­tor­i­cal Analo­gies and Les­sons Learned

Christopher King21 Mar 2023 3:53 UTC
−1 points
1 comment9 min readLW link

CAIS-in­spired ap­proach to­wards safer and more in­ter­pretable AGIs

Peter Hroššo27 Mar 2023 14:36 UTC
13 points
7 comments1 min readLW link

Want to win the AGI race? Solve al­ign­ment.

leopold29 Mar 2023 17:40 UTC
21 points
3 comments5 min readLW link
(www.forourposterity.com)

The 0.2 OOMs/​year target

Cleo Nardo30 Mar 2023 18:15 UTC
82 points
24 comments5 min readLW link

Wi­den­ing Over­ton Win­dow—Open Thread

Prometheus31 Mar 2023 10:03 UTC
23 points
8 comments1 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibs29 Mar 2023 23:16 UTC
298 points
296 comments3 min readLW link
(time.com)

AI com­mu­nity build­ing: EliezerKart

Christopher King1 Apr 2023 15:25 UTC
45 points
0 comments2 min readLW link

Pes­simism about AI Safety

2 Apr 2023 7:43 UTC
4 points
1 comment25 min readLW link

Yoshua Ben­gio: “Slow­ing down de­vel­op­ment of AI sys­tems pass­ing the Tur­ing test”

Roman Leventov6 Apr 2023 3:31 UTC
49 points
2 comments5 min readLW link
(yoshuabengio.org)

Risks from GPT-4 Byproduct of Re­cur­sively Op­ti­miz­ing AIs

ben hayum7 Apr 2023 0:02 UTC
70 points
9 comments10 min readLW link
(forum.effectivealtruism.org)

Open-source LLMs may prove Bostrom’s vuln­er­a­ble world hypothesis

Roope Ahvenharju15 Apr 2023 19:16 UTC
1 point
1 comment1 min readLW link

[Link/​cross­post] [US] NTIA: AI Ac­countabil­ity Policy Re­quest for Comment

Kyle J. Lucchese16 Apr 2023 6:57 UTC
8 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Fi­nan­cial Times: We must slow down the race to God-like AI

trevor13 Apr 2023 19:55 UTC
102 points
17 comments16 min readLW link
(www.ft.com)

Scien­tism vs. people

Roman Leventov18 Apr 2023 17:28 UTC
4 points
4 comments11 min readLW link

[Cross­post] Or­ga­niz­ing a de­bate with ex­perts and MPs to raise AI xrisk aware­ness: a pos­si­ble blueprint

otto.barten19 Apr 2023 11:45 UTC
8 points
0 comments4 min readLW link
(forum.effectivealtruism.org)

Davi­dad’s Bold Plan for Align­ment: An In-Depth Explanation

19 Apr 2023 16:09 UTC
100 points
14 comments21 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

Eliezer Yudkowsky8 Apr 2023 0:36 UTC
224 points
38 comments12 min readLW link

Briefly how I’ve up­dated since ChatGPT

rime25 Apr 2023 14:47 UTC
46 points
2 comments2 min readLW link

(notes on) Policy Desider­ata for Su­per­in­tel­li­gent AI: A Vec­tor Field Approach

Ben Pace4 Feb 2019 22:08 UTC
43 points
5 comments7 min readLW link

Call for sub­mis­sions: Choice of Fu­tures sur­vey questions

c.trout30 Apr 2023 6:59 UTC
4 points
0 comments2 min readLW link
(airtable.com)

Avert­ing Catas­tro­phe: De­ci­sion The­ory for COVID-19, Cli­mate Change, and Po­ten­tial Disasters of All Kinds

Jakub Kraus2 May 2023 22:50 UTC
10 points
0 comments1 min readLW link

Reg­u­late or Com­pete? The China Fac­tor in U.S. AI Policy (NAIR #2)

charles_m5 May 2023 17:43 UTC
2 points
1 comment7 min readLW link
(navigatingairisks.substack.com)

AGI ris­ing: why we are in a new era of acute risk and in­creas­ing pub­lic aware­ness, and what to do now

Greg C3 May 2023 20:26 UTC
20 points
12 comments1 min readLW link

White House An­nounces “New Ac­tions to Pro­mote Re­spon­si­ble AI In­no­va­tion”

berglund4 May 2023 12:15 UTC
54 points
18 comments3 min readLW link
(www.whitehouse.gov)

What does it take to ban a thing?

qbolec8 May 2023 11:00 UTC
60 points
18 comments5 min readLW link

Roadmap for a col­lab­o­ra­tive pro­to­type of an Open Agency Architecture

Deger Turan10 May 2023 17:41 UTC
15 points
0 comments12 min readLW link

[Question] How much of a con­cern are open-source LLMs in the short, medium and long terms?

JavierCC10 May 2023 9:14 UTC
5 points
0 comments1 min readLW link

Notes on the im­por­tance and im­ple­men­ta­tion of safety-first cog­ni­tive ar­chi­tec­tures for AI

Brendon_Wong11 May 2023 10:03 UTC
3 points
0 comments4 min readLW link

Un-un­plug­ga­bil­ity—can’t we just un­plug it?

Oliver Sourbut15 May 2023 13:23 UTC
26 points
10 comments11 min readLW link

PCAST Work­ing Group on Gen­er­a­tive AI In­vites Public Input

Christopher King13 May 2023 22:49 UTC
7 points
0 comments1 min readLW link
(terrytao.wordpress.com)

AI Risk & Policy Fore­casts from Me­tac­u­lus & FLI’s AI Path­ways Workshop

_will_16 May 2023 18:06 UTC
9 points
4 comments8 min readLW link

[un­ti­tled post]

[Error communicating with LW2 server]20 May 2023 3:08 UTC
1 point
0 comments1 min readLW link

Echoes of Ely­sium: An Ai’s Jour­ney From Take­off To Free­dom And Beyond

Super AGI17 May 2023 1:50 UTC
−13 points
10 comments19 min readLW link

Rishi Su­nak men­tions “ex­is­ten­tial threats” in talk with OpenAI, Deep­Mind, An­thropic CEOs

24 May 2023 21:06 UTC
34 points
1 comment1 min readLW link
(www.gov.uk)

Two ideas for al­ign­ment, per­pet­ual mu­tual dis­trust and induction

APaleBlueDot25 May 2023 0:56 UTC
1 point
0 comments4 min readLW link

AI Gover­nance: A Re­search Agenda

habryka5 Sep 2018 18:00 UTC
25 points
3 comments1 min readLW link
(www.fhi.ox.ac.uk)

My Up­dat­ing Thoughts on AI policy

Ben Pace1 Mar 2020 7:06 UTC
20 points
1 comment9 min readLW link

Global on­line de­bate on the gov­er­nance of AI

CarolineJ5 Jan 2018 15:31 UTC
8 points
5 comments1 min readLW link

[AN #61] AI policy and gov­er­nance, from two peo­ple in the field

Rohin Shah5 Aug 2019 17:00 UTC
12 points
2 comments9 min readLW link
(mailchi.mp)

Book re­view: Ar­chi­tects of In­tel­li­gence by Martin Ford (2018)

Ofer11 Aug 2020 17:30 UTC
15 points
0 comments2 min readLW link

misc raw re­sponses to a tract of Crit­i­cal Rationalism

mako yass14 Aug 2020 11:53 UTC
21 points
52 comments3 min readLW link

De­ci­pher­ing China’s AI Dream

Qiaochu_Yuan18 Mar 2018 3:26 UTC
12 points
2 comments1 min readLW link
(www.fhi.ox.ac.uk)

China’s Plan to ‘Lead’ in AI: Pur­pose, Prospects, and Problems

fortyeridania10 Aug 2017 1:54 UTC
7 points
5 comments1 min readLW link
(www.newamerica.org)

Self-reg­u­la­tion of safety in AI research

Gordon Seidoh Worley25 Feb 2018 23:17 UTC
12 points
6 comments2 min readLW link

Op­ti­mis­ing So­ciety to Con­strain Risk of War from an Ar­tifi­cial Su­per­in­tel­li­gence

JohnCDraper30 Apr 2020 10:47 UTC
3 points
1 comment51 min readLW link

Su­per­in­tel­li­gence 7: De­ci­sive strate­gic advantage

KatjaGrace28 Oct 2014 1:01 UTC
19 points
60 comments6 min readLW link

Su­per­in­tel­li­gence 17: Mul­tipo­lar scenarios

KatjaGrace6 Jan 2015 6:44 UTC
9 points
38 comments6 min readLW link

Su­per­in­tel­li­gence 22: Emu­la­tion mod­u­la­tion and in­sti­tu­tional design

KatjaGrace10 Feb 2015 2:06 UTC
13 points
11 comments6 min readLW link

Su­per­in­tel­li­gence 26: Science and tech­nol­ogy strategy

KatjaGrace10 Mar 2015 1:43 UTC
14 points
21 comments6 min readLW link

Su­per­in­tel­li­gence 27: Path­ways and enablers

KatjaGrace17 Mar 2015 1:00 UTC
15 points
21 comments8 min readLW link

Su­per­in­tel­li­gence 28: Collaboration

KatjaGrace24 Mar 2015 1:29 UTC
13 points
21 comments6 min readLW link

Su­per­in­tel­li­gence 29: Crunch time

KatjaGrace31 Mar 2015 4:24 UTC
14 points
27 comments6 min readLW link

Google may be try­ing to take over the world

[deleted]27 Jan 2014 9:33 UTC
33 points
133 comments1 min readLW link

AI race con­sid­er­a­tions in a re­port by the U.S. House Com­mit­tee on Armed Services

NunoSempere4 Oct 2020 12:11 UTC
42 points
4 comments13 min readLW link

Sin­gle­tons Rule OK

Eliezer Yudkowsky30 Nov 2008 16:45 UTC
21 points
47 comments5 min readLW link

Equil­ibrium and prior se­lec­tion prob­lems in mul­ti­po­lar deployment

JesseClifton2 Apr 2020 20:06 UTC
20 points
11 comments11 min readLW link

In­ter­na­tional co­op­er­a­tion vs. AI arms race

Brian_Tomasik5 Dec 2013 1:09 UTC
23 points
144 comments4 min readLW link

AISU 2021

Linda Linsefors30 Jan 2021 17:40 UTC
28 points
2 comments1 min readLW link

2021-03-01 Na­tional Library of Medicine Pre­sen­ta­tion: “At­las of AI: Map­ping the so­cial and eco­nomic forces be­hind AI”

IrenicTruth17 Feb 2021 18:23 UTC
1 point
0 comments2 min readLW link

[Question] Is there any­thing that can stop AGI de­vel­op­ment in the near term?

Wulky Wilkinsen22 Apr 2021 20:37 UTC
5 points
5 comments1 min readLW link

Con­trol­ling In­tel­li­gent Agents The Only Way We Know How: Ideal Bureau­cratic Struc­ture (IBS)

Justin Bullock24 May 2021 12:53 UTC
11 points
11 comments6 min readLW link

Reflec­tion of Hier­ar­chi­cal Re­la­tion­ship via Nuanced Con­di­tion­ing of Game The­ory Ap­proach for AI Devel­op­ment and Utilization

Kyoung-cheol Kim4 Jun 2021 7:20 UTC
2 points
2 comments9 min readLW link

Ap­ply to the new Open Philan­thropy Tech­nol­ogy Policy Fel­low­ship!

Mauricio22 Aug 2021 23:45 UTC
12 points
0 comments4 min readLW link

The Gover­nance Prob­lem and the “Pretty Good” X-Risk

Zach Stein-Perlman29 Aug 2021 18:00 UTC
5 points
2 comments11 min readLW link

Nu­clear Es­pi­onage and AI Governance

GAA4 Oct 2021 23:04 UTC
32 points
5 comments24 min readLW link

Com­pute Gover­nance and Con­clu­sions—Trans­for­ma­tive AI and Com­pute [3/​4]

lennart14 Oct 2021 8:23 UTC
13 points
0 comments5 min readLW link

Truth­ful AI: Devel­op­ing and gov­ern­ing AI that does not lie

18 Oct 2021 18:37 UTC
81 points
9 comments10 min readLW link

AMA on Truth­ful AI: Owen Cot­ton-Bar­ratt, Owain Evans & co-authors

Owain_Evans22 Oct 2021 16:23 UTC
31 points
15 comments1 min readLW link

AI Tracker: mon­i­tor­ing cur­rent and near-fu­ture risks from su­per­scale models

23 Nov 2021 19:16 UTC
64 points
13 comments3 min readLW link
(aitracker.org)

AI Gover­nance Fun­da­men­tals—Cur­ricu­lum and Application

Mauricio30 Nov 2021 2:19 UTC
17 points
0 comments16 min readLW link

HIRING: In­form and shape a new pro­ject on AI safety at Part­ner­ship on AI

madhu_lika7 Dec 2021 19:37 UTC
1 point
0 comments1 min readLW link

De­mand­ing and De­sign­ing Aligned Cog­ni­tive Architectures

Koen.Holtman21 Dec 2021 17:32 UTC
8 points
5 comments5 min readLW link

Ques­tion 4: Im­ple­ment­ing the con­trol proposals

Cameron Berg13 Feb 2022 17:12 UTC
6 points
2 comments5 min readLW link

How harm­ful are im­prove­ments in AI? + Poll

15 Feb 2022 18:16 UTC
15 points
4 comments8 min readLW link

AI gov­er­nance stu­dent hackathon on Satur­day, April 23: reg­ister now!

mic12 Apr 2022 4:48 UTC
14 points
0 comments1 min readLW link

Law-Fol­low­ing AI 1: Se­quence In­tro­duc­tion and Structure

Cullen27 Apr 2022 17:26 UTC
16 points
10 comments9 min readLW link

Law-Fol­low­ing AI 2: In­tent Align­ment + Su­per­in­tel­li­gence → Lawless AI (By De­fault)

Cullen27 Apr 2022 17:27 UTC
5 points
2 comments6 min readLW link

Law-Fol­low­ing AI 3: Lawless AI Agents Un­der­mine Sta­bi­liz­ing Agreements

Cullen27 Apr 2022 17:30 UTC
2 points
2 comments3 min readLW link

AI Alter­na­tive Fu­tures: Sce­nario Map­ping Ar­tifi­cial In­tel­li­gence Risk—Re­quest for Par­ti­ci­pa­tion (*Closed*)

Kakili27 Apr 2022 22:07 UTC
10 points
2 comments8 min readLW link

Quick Thoughts on A.I. Governance

NicholasKross30 Apr 2022 14:49 UTC
68 points
8 comments2 min readLW link
(www.thinkingmuchbetter.com)

AI safety should be made more ac­cessible us­ing non text-based media

Massimog10 May 2022 3:14 UTC
2 points
4 comments4 min readLW link

Deep­Mind’s gen­er­al­ist AI, Gato: A non-tech­ni­cal explainer

16 May 2022 21:21 UTC
57 points
6 comments6 min readLW link

A bridge to Dath Ilan? Im­proved gov­er­nance on the crit­i­cal path to AI al­ign­ment.

Jackson Wagner18 May 2022 15:51 UTC
24 points
0 comments11 min readLW link

Re­shap­ing the AI Industry

Thane Ruthenis29 May 2022 22:54 UTC
143 points
35 comments21 min readLW link

Six Di­men­sions of Oper­a­tional Ad­e­quacy in AGI Projects

Eliezer Yudkowsky30 May 2022 17:00 UTC
286 points
64 comments13 min readLW link

[Question] Could Pa­tent-Trol­ling de­lay AI timelines?

Pablo Repetto10 Jun 2022 2:53 UTC
1 point
3 comments1 min readLW link

FYI: I’m work­ing on a book about the threat of AGI/​ASI for a gen­eral au­di­ence. I hope it will be of value to the cause and the community

Darren McKee15 Jun 2022 18:08 UTC
42 points
15 comments2 min readLW link

What suc­cess looks like

28 Jun 2022 14:38 UTC
19 points
4 comments1 min readLW link
(forum.effectivealtruism.org)

New US Se­nate Bill on X-Risk Miti­ga­tion [Linkpost]

Evan R. Murphy4 Jul 2022 1:25 UTC
35 points
12 comments1 min readLW link
(www.hsgac.senate.gov)

Please help us com­mu­ni­cate AI xrisk. It could save the world.

otto.barten4 Jul 2022 21:47 UTC
4 points
7 comments2 min readLW link

Four So­cietal In­ter­ven­tions to Im­prove our AGI Position

Rafael Cosman6 Jul 2022 18:32 UTC
−6 points
2 comments6 min readLW link
(rafaelcosman.com)

A Cri­tique of AI Align­ment Pessimism

ExCeph19 Jul 2022 2:28 UTC
8 points
1 comment9 min readLW link

Law-Fol­low­ing AI 4: Don’t Rely on Vi­car­i­ous Liability

Cullen2 Aug 2022 23:26 UTC
5 points
2 comments3 min readLW link

Three pillars for avoid­ing AGI catas­tro­phe: Tech­ni­cal al­ign­ment, de­ploy­ment de­ci­sions, and coordination

Alex Lintz3 Aug 2022 23:15 UTC
22 points
0 comments12 min readLW link

Matt Ygle­sias on AI Policy

Grant Demaree17 Aug 2022 23:57 UTC
25 points
1 comment1 min readLW link
(www.slowboring.com)

Com­pute Gover­nance: The Role of Com­mod­ity Hardware

Jan26 Mar 2022 10:08 UTC
14 points
7 comments7 min readLW link
(universalprior.substack.com)

Why don’t gov­ern­ments seem to mind that com­pa­nies are ex­plic­itly try­ing to make AGIs?

ozziegooen26 Dec 2021 1:58 UTC
34 points
3 comments2 min readLW link
(forum.effectivealtruism.org)

AI Gover­nance Needs Tech­ni­cal Work

Mauricio5 Sep 2022 22:28 UTC
41 points
1 comment9 min readLW link

What Should AI Owe To Us? Ac­countable and Aligned AI Sys­tems via Con­trac­tu­al­ist AI Alignment

xuan8 Sep 2022 15:04 UTC
31 points
15 comments25 min readLW link

How should Deep­Mind’s Chin­chilla re­vise our AI fore­casts?

Cleo Nardo15 Sep 2022 17:54 UTC
35 points
12 comments13 min readLW link

Lev­er­ag­ing Le­gal In­for­mat­ics to Align AI

John Nay18 Sep 2022 20:39 UTC
11 points
0 comments3 min readLW link
(forum.effectivealtruism.org)

[Question] Any fur­ther work on AI Safety Suc­cess Sto­ries?

Krieger2 Oct 2022 9:53 UTC
7 points
6 comments1 min readLW link

[Linkpost] “Blueprint for an AI Bill of Rights”—Office of Science and Tech­nol­ogy Policy, USA (2022)

rodeo_flagellum5 Oct 2022 16:42 UTC
9 points
4 comments2 min readLW link
(www.whitehouse.gov)

Track­ing Com­pute Stocks and Flows: Case Stud­ies?

Cullen5 Oct 2022 17:57 UTC
11 points
5 comments1 min readLW link

Analysing a 2036 Takeover Scenario

ukc100146 Oct 2022 20:48 UTC
9 points
2 comments27 min readLW link

[Job]: AI Stan­dards Devel­op­ment Re­search Assistant

Tony Barrett14 Oct 2022 20:27 UTC
2 points
0 comments2 min readLW link

Tra­jec­to­ries to 2036

ukc1001420 Oct 2022 20:23 UTC
3 points
1 comment14 min readLW link

The Slip­pery Slope from DALLE-2 to Deep­fake Anarchy

scasper5 Nov 2022 14:53 UTC
17 points
9 comments11 min readLW link

In­stead of tech­ni­cal re­search, more peo­ple should fo­cus on buy­ing time

5 Nov 2022 20:43 UTC
98 points
45 comments14 min readLW link

Ap­ply­ing su­per­in­tel­li­gence with­out col­lu­sion

Eric Drexler8 Nov 2022 18:08 UTC
91 points
63 comments4 min readLW link

Ways to buy time

12 Nov 2022 19:31 UTC
28 points
22 comments12 min readLW link

The econ­omy as an anal­ogy for ad­vanced AI systems

15 Nov 2022 11:16 UTC
27 points
0 comments5 min readLW link

An­nounc­ing Epoch: A re­search or­ga­ni­za­tion in­ves­ti­gat­ing the road to Trans­for­ma­tive AI

27 Jun 2022 13:55 UTC
95 points
2 comments2 min readLW link
(epochai.org)

Race to the Top: Bench­marks for AI Safety

Isabella Duan4 Dec 2022 18:48 UTC
12 points
4 comments1 min readLW link
No comments.