Ben­gio’s FAQ on Catas­trophic AI Risks

VaniverJun 29, 2023, 11:04 PM
39 points
0 comments1 min readLW link
(yoshuabengio.org)

AGI & War

CalecuteJun 29, 2023, 10:20 PM
9 points
1 comment1 min readLW link

Biosafety Reg­u­la­tions (BMBL) and their rele­vance for AI

Štěpán LosJun 29, 2023, 7:22 PM
4 points
0 comments4 min readLW link

Na­ture Re­leases A Stupid Edi­to­rial On AI Risk

Bentham's BulldogJun 29, 2023, 7:00 PM
2 points
1 comment3 min readLW link

AI Safety with­out Align­ment: How hu­mans can WIN against AI

vicchainJun 29, 2023, 5:53 PM
1 point
1 comment2 min readLW link

Challenge pro­posal: small­est pos­si­ble self-hard­en­ing back­door for RLHF

Christopher KingJun 29, 2023, 4:56 PM
7 points
0 comments2 min readLW link

AI #18: The Great De­bate Debate

ZviJun 29, 2023, 4:20 PM
47 points
9 comments52 min readLW link
(thezvi.wordpress.com)

Bruce Ster­ling on the AI ma­nia of 2023

Mitchell_PorterJun 29, 2023, 5:00 AM
25 points
1 comment1 min readLW link
(www.newsweek.com)

Cheat sheet of AI X-risk

momom2Jun 29, 2023, 4:28 AM
19 points
1 comment7 min readLW link

An­throp­i­cally Blind: the an­thropic shadow is re­flec­tively inconsistent

Christopher KingJun 29, 2023, 2:36 AM
43 points
40 comments10 min readLW link

One path to co­her­ence: con­di­tion­al­iza­tion

porbyJun 29, 2023, 1:08 AM
28 points
4 comments4 min readLW link

AXRP an­nounce­ment: Sur­vey, Store Clos­ing, Patreon

DanielFilanJun 28, 2023, 11:40 PM
14 points
0 comments1 min readLW link

Me­taphors for AI, and why I don’t like them

boazbarakJun 28, 2023, 10:47 PM
38 points
18 comments12 min readLW link

Trans­form­ing Democ­racy: A Unique Fund­ing Op­por­tu­nity for US Fed­eral Ap­proval Voting

Aaron HamlinJun 28, 2023, 10:07 PM
25 points
6 comments2 min readLW link

AGI x An­i­mal Welfare: A High-EV Outreach Op­por­tu­nity?

simeon_cJun 28, 2023, 8:44 PM
29 points
0 commentsLW link

A “weak” AGI may at­tempt an un­likely-to-suc­ceed takeover

RobertMJun 28, 2023, 8:31 PM
56 points
17 comments3 min readLW link

Progress links and tweets, 2023-06-28: “We can do big things again in Penn­syl­va­nia”

jasoncrawfordJun 28, 2023, 8:23 PM
14 points
1 comment1 min readLW link
(rootsofprogress.org)

[Question] What money-pumps ex­ist, if any, for de­on­tol­o­gists?

Daniel KokotajloJun 28, 2023, 7:08 PM
39 points
35 comments1 min readLW link

[Question] What is your fi­nan­cial port­fo­lio?

AlgonJun 28, 2023, 6:39 PM
11 points
11 comments1 min readLW link

Levels of safety for AI and other technologies

jasoncrawfordJun 28, 2023, 6:35 PM
16 points
0 comments2 min readLW link
(rootsofprogress.org)

LeCun says mak­ing a util­ity func­tion is intractable

IknownothingJun 28, 2023, 6:02 PM
2 points
3 comments1 min readLW link

My re­search agenda in agent foundations

Alex_AltairJun 28, 2023, 6:00 PM
75 points
9 comments11 min readLW link

AI In­ci­dent Shar­ing—Best prac­tices from other fields and a com­pre­hen­sive list of ex­ist­ing platforms

Štěpán LosJun 28, 2023, 5:21 PM
20 points
0 comments4 min readLW link

The Case for Over­con­fi­dence is Overstated

Kevin DorstJun 28, 2023, 5:21 PM
50 points
13 comments8 min readLW link
(kevindorst.substack.com)

When do “brains beat brawn” in Chess? An experiment

titotalJun 28, 2023, 1:33 PM
318 points
106 comments7 min readLW link2 reviews
(titotal.substack.com)

Giv­ing an evolu­tion­ary ex­pla­na­tion for Kah­ne­man and Tver­sky’s in­sights on sub­jec­tive satisfaction

LionelJun 28, 2023, 12:17 PM
−7 points
1 comment1 min readLW link
(lionelpage.substack.com)

Na­ture: “Stop talk­ing about to­mor­row’s AI dooms­day when AI poses risks to­day”

Ben SmithJun 28, 2023, 5:59 AM
40 points
8 comments2 min readLW link
(www.nature.com)

Re­quest: Put Carl Shul­man’s re­cent pod­cast into an or­ga­nized writ­ten format

Aryeh EnglanderJun 28, 2023, 2:58 AM
19 points
4 comments1 min readLW link

Pre­dic­tion Mar­ket: Will I Pull “The One Ring To Rule Them All?”

Connor TabarrokJun 28, 2023, 2:41 AM
1 point
0 comments1 min readLW link
(manifold.markets)

Carl Shul­man on The Lu­nar So­ciety (7 hour, two-part pod­cast)

ESRogsJun 28, 2023, 1:23 AM
79 points
17 comments1 min readLW link
(www.dwarkeshpatel.com)

Brief sum­mary of ai-plans.com

IknownothingJun 28, 2023, 12:33 AM
9 points
4 comments2 min readLW link
(ai-plans.com)

Catas­trophic Risks from AI #6: Dis­cus­sion and FAQ

Jun 27, 2023, 11:23 PM
24 points
1 comment13 min readLW link
(arxiv.org)

Catas­trophic Risks from AI #5: Rogue AIs

Jun 27, 2023, 10:06 PM
15 points
0 comments22 min readLW link
(arxiv.org)

AISN #12: Policy Pro­pos­als from NTIA’s Re­quest for Com­ment and Re­con­sid­er­ing In­stru­men­tal Convergence

Dan HJun 27, 2023, 5:20 PM
6 points
0 commentsLW link

The Weight of the Fu­ture (Why The Apoca­lypse Can Be A Relief)

SableJun 27, 2023, 5:18 PM
18 points
14 comments3 min readLW link
(affablyevil.substack.com)

Align­ing AI by op­ti­miz­ing for “wis­dom”

Jun 27, 2023, 3:20 PM
28 points
8 comments12 min readLW link

Free­dom un­der Nat­u­ral­is­tic Dualism

Arturo MaciasJun 27, 2023, 2:34 PM
1 point
36 comments1 min readLW link
(www.jneurophilosophy.com)

Munk AI de­bate: con­fu­sions and pos­si­ble cruxes

Steven ByrnesJun 27, 2023, 2:18 PM
244 points
21 comments8 min readLW link

Ate­liers: Motivation

Stephen FowlerJun 27, 2023, 1:07 PM
7 points
0 comments2 min readLW link

Self-Blinded Caf­feine RCT

niplavJun 27, 2023, 12:38 PM
45 points
9 comments8 min readLW link

An overview of the points system

IknownothingJun 27, 2023, 9:09 AM
3 points
4 comments1 min readLW link
(ai-plans.com)

AISC team re­port: Soft-op­ti­miza­tion, Bayes and Goodhart

Jun 27, 2023, 6:05 AM
38 points
2 comments15 min readLW link

Epistemic spot check­ing one claim in The Precipice

Isaac KingJun 27, 2023, 1:03 AM
33 points
3 comments1 min readLW link

nu­clear costs are inflation

bhauthJun 26, 2023, 10:30 PM
8 points
42 comments5 min readLW link
(www.bhauth.com)

Man in the Arena

Richard_NgoJun 26, 2023, 9:57 PM
66 points
6 comments8 min readLW link

Catas­trophic Risks from AI #4: Or­ga­ni­za­tional Risks

Jun 26, 2023, 7:36 PM
23 points
0 comments21 min readLW link
(arxiv.org)

The fraught voy­age of al­igned novelty

TsviBTJun 26, 2023, 7:10 PM
13 points
0 comments17 min readLW link

[Question] De­cep­tive AI vs. shift­ing in­stru­men­tal incentives

Aryeh EnglanderJun 26, 2023, 6:09 PM
7 points
2 comments3 min readLW link

On the Cost of Thriv­ing Index

ZviJun 26, 2023, 3:30 PM
33 points
6 comments9 min readLW link
(thezvi.wordpress.com)

“Safety Cul­ture for AI” is im­por­tant, but isn’t go­ing to be easy

DavidmanheimJun 26, 2023, 12:52 PM
47 points
2 comments2 min readLW link
(forum.effectivealtruism.org)