Luck based medicine: my re­sent­ful story of be­com­ing a med­i­cal miracle

Elizabeth16 Oct 2022 17:40 UTC
480 points
119 comments12 min readLW link3 reviews
(acesounderglass.com)

Coun­ter­ar­gu­ments to the ba­sic AI x-risk case

KatjaGrace14 Oct 2022 13:00 UTC
369 points
124 comments34 min readLW link1 review
(aiimpacts.org)

So, geez there’s a lot of AI con­tent these days

Raemon6 Oct 2022 21:32 UTC
255 points
140 comments6 min readLW link

In­tro­duc­tion to ab­stract entropy

Alex_Altair20 Oct 2022 21:03 UTC
229 points
78 comments18 min readLW link1 review

Les­sons learned from talk­ing to >100 aca­demics about AI safety

Marius Hobbhahn10 Oct 2022 13:16 UTC
214 points
17 comments12 min readLW link1 review

What does it take to defend the world against out-of-con­trol AGIs?

Steven Byrnes25 Oct 2022 14:47 UTC
194 points
47 comments30 min readLW link1 review

De­ci­sion the­ory does not im­ply that we get to have nice things

So8res18 Oct 2022 3:04 UTC
168 points
58 comments26 min readLW link2 reviews

The So­cial Re­ces­sion: By the Numbers

antonomon29 Oct 2022 18:45 UTC
165 points
29 comments8 min readLW link
(novum.substack.com)

Why I think there’s a one-in-six chance of an im­mi­nent global nu­clear war

Max Tegmark8 Oct 2022 6:26 UTC
163 points
169 comments4 min readLW link

Six (and a half) in­tu­itions for KL divergence

CallumMcDougall12 Oct 2022 21:07 UTC
154 points
25 comments10 min readLW link1 review
(www.perfectlynormal.co.uk)

Age changes what you care about

Dentin16 Oct 2022 15:36 UTC
140 points
36 comments2 min readLW link

AI Timelines via Cu­mu­la­tive Op­ti­miza­tion Power: Less Long, More Short

jacob_cannell6 Oct 2022 0:21 UTC
139 points
33 comments6 min readLW link

Ap­ply to the Red­wood Re­search Mechanis­tic In­ter­pretabil­ity Ex­per­i­ment (REMIX), a re­search pro­gram in Berkeley

27 Oct 2022 1:32 UTC
135 points
14 comments12 min readLW link

Warn­ing Shots Prob­a­bly Wouldn’t Change The Pic­ture Much

So8res6 Oct 2022 5:15 UTC
122 points
42 comments2 min readLW link

Nice­ness is unnatural

So8res13 Oct 2022 1:30 UTC
121 points
20 comments8 min readLW link1 review

Mnestics

Jarred Filmer23 Oct 2022 0:30 UTC
117 points
5 comments4 min readLW link

Why Weren’t Hot Air Bal­loons In­vented Sooner?

Lost Futures18 Oct 2022 0:41 UTC
115 points
52 comments6 min readLW link
(lostfutures.substack.com)

Am I se­cretly ex­cited for AI get­ting weird?

porby29 Oct 2022 22:16 UTC
115 points
4 comments4 min readLW link

Don’t leave your finger­prints on the future

So8res8 Oct 2022 0:35 UTC
114 points
46 comments5 min readLW link

Ac­tu­ally, All Nu­clear Famine Papers are Bunk

Lao Mein12 Oct 2022 5:58 UTC
113 points
37 comments2 min readLW link1 review

That one apoc­a­lyp­tic nu­clear famine pa­per is bunk

Lao Mein12 Oct 2022 3:33 UTC
110 points
10 comments1 min readLW link

Con­sider your ap­petite for disagreements

Adam Zerner8 Oct 2022 23:25 UTC
107 points
18 comments6 min readLW link1 review

Plans Are Pre­dic­tions, Not Op­ti­miza­tion Targets

johnswentworth20 Oct 2022 21:17 UTC
105 points
20 comments4 min readLW link1 review

Align­ment 201 curriculum

Richard_Ngo12 Oct 2022 18:03 UTC
102 points
3 comments1 min readLW link
(www.agisafetyfundamentals.com)

Anal­y­sis: US re­stricts GPU sales to China

aogara7 Oct 2022 18:38 UTC
102 points
58 comments5 min readLW link

Scal­ing Laws for Re­ward Model Overoptimization

20 Oct 2022 0:20 UTC
102 points
13 comments1 min readLW link
(arxiv.org)

Con­tra shard the­ory, in the con­text of the di­a­mond max­i­mizer problem

So8res13 Oct 2022 23:51 UTC
101 points
19 comments2 min readLW link1 review

Some Les­sons Learned from Study­ing Indi­rect Ob­ject Iden­ti­fi­ca­tion in GPT-2 small

28 Oct 2022 23:55 UTC
99 points
9 comments9 min readLW link2 reviews
(arxiv.org)

How To Make Pre­dic­tion Mar­kets Use­ful For Align­ment Work

johnswentworth18 Oct 2022 19:01 UTC
97 points
18 comments2 min readLW link

Trans­for­ma­tive VR Is Likely Com­ing Soon

jimrandomh13 Oct 2022 6:25 UTC
92 points
46 comments2 min readLW link

A shot at the di­a­mond-al­ign­ment problem

TurnTrout6 Oct 2022 18:29 UTC
92 points
58 comments15 min readLW link

A blog post is a very long and com­plex search query to find fas­ci­nat­ing peo­ple and make them route in­ter­est­ing stuff to your inbox

Henrik Karlsson5 Oct 2022 19:07 UTC
89 points
12 comments11 min readLW link
(escapingflatland.substack.com)

Poly­se­man­tic­ity and Ca­pac­ity in Neu­ral Networks

7 Oct 2022 17:51 UTC
87 points
14 comments3 min readLW link

Why Balsa Re­search is Worthwhile

Zvi10 Oct 2022 13:50 UTC
87 points
12 comments8 min readLW link
(thezvi.wordpress.com)

«Boundaries», Part 3a: Defin­ing bound­aries as di­rected Markov blankets

Andrew_Critch30 Oct 2022 6:31 UTC
86 points
20 comments15 min readLW link

Paper: Dis­cov­er­ing novel al­gorithms with AlphaTen­sor [Deep­mind]

LawrenceC5 Oct 2022 16:20 UTC
82 points
18 comments1 min readLW link
(www.deepmind.com)

The Teacup Test

lsusr8 Oct 2022 4:25 UTC
82 points
31 comments2 min readLW link

More Re­cent Progress in the The­ory of Neu­ral Networks

jylin046 Oct 2022 16:57 UTC
81 points
6 comments4 min readLW link

“Nor­mal” is the equil­ibrium state of past op­ti­miza­tion processes

Alex_Altair30 Oct 2022 19:03 UTC
81 points
5 comments5 min readLW link

Vot­ing The­ory Introduction

Scott Garrabrant17 Oct 2022 8:48 UTC
81 points
7 comments6 min readLW link

The her­i­ta­bil­ity of hu­man val­ues: A be­hav­ior ge­netic cri­tique of Shard Theory

geoffreymiller20 Oct 2022 15:51 UTC
80 points
59 comments21 min readLW link

I learn bet­ter when I frame learn­ing as Vengeance for losses in­curred through ig­no­rance, and you might too

chaosmage15 Oct 2022 12:41 UTC
79 points
9 comments3 min readLW link1 review

Un­tapped Po­ten­tial at 13-18

belkarx18 Oct 2022 18:09 UTC
78 points
53 comments1 min readLW link

Re­sources that (I think) new al­ign­ment re­searchers should know about

Akash28 Oct 2022 22:13 UTC
77 points
9 comments4 min readLW link

Re­sponse to Katja Grace’s AI x-risk counterarguments

19 Oct 2022 1:17 UTC
76 points
18 comments15 min readLW link

The “you-can-just” alarm

Emrik8 Oct 2022 10:43 UTC
76 points
3 comments1 min readLW link

Neu­ral Tan­gent Ker­nel Distillation

5 Oct 2022 18:11 UTC
75 points
20 comments8 min readLW link

Max­i­mal Lotteries

Scott Garrabrant17 Oct 2022 8:54 UTC
74 points
11 comments7 min readLW link

What does it mean for an AGI to be ‘safe’?

So8res7 Oct 2022 4:13 UTC
74 points
29 comments3 min readLW link

Open Prob­lem in Vot­ing Theory

Scott Garrabrant17 Oct 2022 20:42 UTC
74 points
16 comments6 min readLW link