A brief col­lec­tion of Hin­ton’s re­cent com­ments on AGI risk

Kaj_Sotala4 May 2023 23:31 UTC
143 points
9 comments11 min readLW link

Robin Han­son and I talk about AI risk

KatjaGrace4 May 2023 22:20 UTC
39 points
8 comments1 min readLW link
(worldspiritsockpuppet.com)

Who reg­u­lates the reg­u­la­tors? We need to go be­yond the re­view-and-ap­proval paradigm

jasoncrawford4 May 2023 22:11 UTC
122 points
29 comments13 min readLW link
(rootsofprogress.org)

Re­cur­sive Mid­dle Man­ager Hell: AI Edition

VojtaKovarik4 May 2023 20:08 UTC
30 points
11 comments2 min readLW link

AI risk/​re­ward: A sim­ple model

Nathan Young4 May 2023 19:25 UTC
3 points
0 comments1 min readLW link

Google “We Have No Moat, And Nei­ther Does OpenAI”

Chris_Leong4 May 2023 18:23 UTC
61 points
28 comments1 min readLW link
(www.semianalysis.com)

Try­ing to mea­sure AI de­cep­tion ca­pa­bil­ities us­ing tem­po­rary simu­la­tion fine-tuning

alenoach4 May 2023 17:59 UTC
4 points
0 comments7 min readLW link

[Linkpost]Trans­former-Based LM Sur­prisal Pre­dicts Hu­man Read­ing Times Best with About Two Billion Train­ing Tokens

Curtis Huebner4 May 2023 17:16 UTC
10 points
1 comment1 min readLW link
(arxiv.org)

Clar­ify­ing and pre­dict­ing AGI

Richard_Ngo4 May 2023 15:55 UTC
129 points
42 comments4 min readLW link

[Cross­post] AI X-risk in the News: How Effec­tive are Re­cent Me­dia Items and How is Aware­ness Chang­ing? Our New Sur­vey Re­sults.

otto.barten4 May 2023 14:09 UTC
5 points
0 comments9 min readLW link
(forum.effectivealtruism.org)

AI #10: Code In­ter­preter and Ge­off Hinton

Zvi4 May 2023 14:00 UTC
80 points
7 comments78 min readLW link
(thezvi.wordpress.com)

Ad­vice for in­ter­act­ing with busy people

Severin T. Seehrich4 May 2023 13:31 UTC
66 points
4 comments4 min readLW link

We don’t need AGI for an amaz­ing future

Karl von Wendt4 May 2023 12:10 UTC
18 points
32 comments5 min readLW link

Has the Sym­bol Ground­ing Prob­lem just gone away?

RussellThor4 May 2023 7:46 UTC
12 points
3 comments1 min readLW link

Opinion merg­ing for AI control

David Johnston4 May 2023 2:43 UTC
6 points
0 comments11 min readLW link

Un­der­stand­ing why illu­sion­ism does not deny the ex­is­tence of qualia

Mergimio H. Doefevmil4 May 2023 2:13 UTC
0 points
17 comments1 min readLW link

[New] Re­jected Con­tent Section

4 May 2023 1:43 UTC
65 points
21 comments5 min readLW link

How MATS ad­dresses “mass move­ment build­ing” concerns

Ryan Kidd4 May 2023 0:55 UTC
62 points
9 comments3 min readLW link

Mov­ing VPS Again

jefftk4 May 2023 0:30 UTC
9 points
2 comments1 min readLW link
(www.jefftk.com)

Prizes for ma­trix com­ple­tion problems

paulfchristiano3 May 2023 23:30 UTC
163 points
51 comments1 min readLW link
(www.alignment.org)

Align­ment Re­search @ EleutherAI

Curtis Huebner3 May 2023 22:45 UTC
40 points
1 comment3 min readLW link
(blog.eleuther.ai)

«Boundaries/​Mem­branes» and AI safety compilation

Chipmonk3 May 2023 21:41 UTC
53 points
17 comments8 min readLW link

[Question] What con­straints does deep learn­ing place on al­ign­ment plans?

Garrett Baker3 May 2023 20:40 UTC
9 points
0 comments1 min readLW link

AGI ris­ing: why we are in a new era of acute risk and in­creas­ing pub­lic aware­ness, and what to do now

Greg C3 May 2023 20:26 UTC
23 points
12 comments1 min readLW link

For­mal­iz­ing the “AI x-risk is un­likely be­cause it is ridicu­lous” argument

Christopher King3 May 2023 18:56 UTC
47 points
17 comments3 min readLW link

[Question] List of no­table peo­ple who be­lieve in AI X-risk?

vlad.proex3 May 2023 18:46 UTC
14 points
4 comments1 min readLW link

[Question] LessWrong ex­port­ing?

axiomAdministrator3 May 2023 18:34 UTC
0 points
3 comments1 min readLW link

Progress links and tweets, 2023-05-03

jasoncrawford3 May 2023 16:23 UTC
13 points
0 comments2 min readLW link
(rootsofprogress.org)

Per­son­hood is a Reli­gious Belief

jan Sijan3 May 2023 16:16 UTC
−42 points
28 comments6 min readLW link

Slow­ing AI: Crunch time

Zach Stein-Perlman3 May 2023 15:00 UTC
11 points
1 comment2 min readLW link

Find­ing Neu­rons in a Haystack: Case Stud­ies with Sparse Probing

3 May 2023 13:30 UTC
33 points
5 comments2 min readLW link
(arxiv.org)

Monthly Roundup #6: May 2023

Zvi3 May 2023 12:50 UTC
31 points
12 comments24 min readLW link
(thezvi.wordpress.com)

[Question] How much do per­sonal bi­ases in risk as­sess­ment af­fect as­sess­ment of AI risks?

Gordon Seidoh Worley3 May 2023 6:12 UTC
10 points
8 comments1 min readLW link

Com­mu­ni­ca­tion strate­gies for autism, with examples

stonefly3 May 2023 5:25 UTC
15 points
2 comments7 min readLW link

Un­der­stand how other peo­ple think: a the­ory of wor­ld­views.

spencerg3 May 2023 3:57 UTC
2 points
8 comments1 min readLW link

“Copi­lot” type AI in­te­gra­tion could lead to train­ing data needed for AGI

anithite3 May 2023 0:57 UTC
6 points
0 comments2 min readLW link

Avert­ing Catas­tro­phe: De­ci­sion The­ory for COVID-19, Cli­mate Change, and Po­ten­tial Disasters of All Kinds

JakubK2 May 2023 22:50 UTC
10 points
0 comments1 min readLW link

A Case for the Least For­giv­ing Take On Alignment

Thane Ruthenis2 May 2023 21:34 UTC
99 points
82 comments22 min readLW link

Are Emer­gent Abil­ities of Large Lan­guage Models a Mirage? [linkpost]

Matthew Barnett2 May 2023 21:01 UTC
52 points
19 comments1 min readLW link
(arxiv.org)

Does descal­ing a ket­tle help? The­ory and practice

philh2 May 2023 20:20 UTC
35 points
25 comments8 min readLW link
(reasonableapproximation.net)

Avoid­ing xrisk from AI doesn’t mean fo­cus­ing on AI xrisk

Stuart_Armstrong2 May 2023 19:27 UTC
64 points
7 comments3 min readLW link

AI Safety Newslet­ter #4: AI and Cy­ber­se­cu­rity, Per­sua­sive AIs, Weaponiza­tion, and Ge­offrey Hin­ton talks AI risks

2 May 2023 18:41 UTC
32 points
0 comments5 min readLW link
(newsletter.safe.ai)

My best sys­tem yet: text-based pro­ject management

jt2 May 2023 17:44 UTC
6 points
6 comments5 min readLW link

[Question] What’s the state of AI safety in Ja­pan?

ChristianKl2 May 2023 17:06 UTC
5 points
1 comment1 min readLW link

Five Wor­lds of AI (by Scott Aaron­son and Boaz Barak)

mishka2 May 2023 13:23 UTC
21 points
5 comments1 min readLW link
(scottaaronson.blog)

Sys­tems that can­not be un­safe can­not be safe

Davidmanheim2 May 2023 8:53 UTC
62 points
27 comments2 min readLW link

AGI safety ca­reer advice

Richard_Ngo2 May 2023 7:36 UTC
131 points
24 comments13 min readLW link

An Im­pos­si­bil­ity Proof Rele­vant to the Shut­down Prob­lem and Corrigibility

Audere2 May 2023 6:52 UTC
65 points
13 comments9 min readLW link

Some Thoughts on Virtue Ethics for AIs

peligrietzer2 May 2023 5:46 UTC
74 points
7 comments4 min readLW link

Tech­nolog­i­cal un­em­ploy­ment as an­other test for ra­tio­nal­ist winning

RomanHauksson2 May 2023 4:16 UTC
14 points
5 comments1 min readLW link