Hospi­tal­iza­tion: A Review

Logan Riggs9 Oct 2025 14:36 UTC
363 points
21 comments9 min readLW link

Towards a Ty­pol­ogy of Strange LLM Chains-of-Thought

1a3orn9 Oct 2025 22:02 UTC
301 points
29 comments9 min readLW link

Con­sider donat­ing to Alex Bores, au­thor of the RAISE Act

Eric Neyman20 Oct 2025 14:50 UTC
259 points
20 comments18 min readLW link
(ericneyman.wordpress.com)

I take an­tide­pres­sants. You’re welcome

Elizabeth9 Oct 2025 19:30 UTC
258 points
11 comments3 min readLW link
(acesounderglass.com)

On Flesh­ling Safety: A De­bate by Klurl and Tra­pau­cius.

Eliezer Yudkowsky26 Oct 2025 23:44 UTC
253 points
52 comments79 min readLW link

EU ex­plained in 10 minutes

Martin Sustrik21 Oct 2025 4:40 UTC
244 points
49 comments8 min readLW link
(www.250bpm.com)

The Most Com­mon Bad Ar­gu­ment In Th­ese Parts

J Bostock11 Oct 2025 16:29 UTC
243 points
60 comments4 min readLW link

Ome­las Is Perfectly Misread

Tobias H2 Oct 2025 23:11 UTC
221 points
59 comments5 min readLW link

The Memet­ics of AI Successionism

Jan_Kulveit28 Oct 2025 15:04 UTC
212 points
54 comments9 min readLW link

If Any­one Builds It Every­one Dies, a semi-out­sider re­view

dvd13 Oct 2025 22:10 UTC
212 points
67 comments15 min readLW link

Do One New Thing A Day To Solve Your Problems

Algon3 Oct 2025 17:08 UTC
208 points
28 comments2 min readLW link

The Doomers Were Right

Algon22 Oct 2025 22:18 UTC
204 points
26 comments3 min readLW link

The Origami Men

Tomás B.6 Oct 2025 15:25 UTC
189 points
14 comments16 min readLW link

The “Length” of “Hori­zons”

Adam Scholl14 Oct 2025 14:48 UTC
183 points
27 comments7 min readLW link

An Opinionated Guide to Pri­vacy De­spite Authoritarianism

TurnTrout29 Oct 2025 20:32 UTC
179 points
27 comments4 min readLW link
(turntrout.com)

That Mad Olympiad

Tomás B.15 Oct 2025 13:45 UTC
178 points
14 comments14 min readLW link

Hu­man­ity Learned Al­most Noth­ing From COVID-19

niplav19 Oct 2025 21:24 UTC
163 points
38 comments4 min readLW link

Don’t Mock Yourself

Algon12 Oct 2025 22:40 UTC
163 points
18 comments2 min readLW link

Inoc­u­la­tion prompt­ing: In­struct­ing mod­els to mis­be­have at train-time can im­prove run-time behavior

8 Oct 2025 22:02 UTC
156 points
37 comments2 min readLW link

Nice-ish, smooth take­off (with im­perfect safe­guards) prob­a­bly kills most “clas­sic hu­mans” in a few decades.

Raemon2 Oct 2025 21:03 UTC
155 points
19 comments12 min readLW link

Med­i­ta­tion is dangerous

Algon17 Oct 2025 22:52 UTC
155 points
40 comments4 min readLW link

AIs should also re­fuse to work on ca­pa­bil­ities research

Davidmanheim27 Oct 2025 8:42 UTC
150 points
20 comments3 min readLW link

Son­net 4.5′s eval gam­ing se­ri­ously un­der­mines al­ign­ment evals, and this seems caused by train­ing on al­ign­ment evals

30 Oct 2025 15:34 UTC
143 points
21 comments14 min readLW link

Real­is­tic Re­ward Hack­ing In­duces Differ­ent and Deeper Misalignment

Jozdien9 Oct 2025 18:45 UTC
143 points
2 comments23 min readLW link

Which side of the AI safety com­mu­nity are you in?

Max Tegmark22 Oct 2025 21:17 UTC
141 points
88 comments2 min readLW link

Cheap Labour Every­where

Morpheus16 Oct 2025 13:15 UTC
136 points
34 comments2 min readLW link

Con­sider donat­ing to AI safety cham­pion Scott Wiener

Eric Neyman22 Oct 2025 18:40 UTC
133 points
9 comments18 min readLW link
(ericneyman.wordpress.com)

How Well Does RL Scale?

Toby_Ord22 Oct 2025 13:16 UTC
131 points
22 comments7 min readLW link

Plans A, B, C, and D for mis­al­ign­ment risk

ryan_greenblatt8 Oct 2025 17:18 UTC
131 points
75 comments6 min readLW link

Emer­gent In­tro­spec­tive Aware­ness in Large Lan­guage Models

Drake Thomas30 Oct 2025 4:42 UTC
129 points
19 comments1 min readLW link
(transformer-circuits.pub)

Check­ing in on AI-2027

Baybar2 Oct 2025 18:46 UTC
128 points
22 comments4 min readLW link

Cancer has a sur­pris­ing amount of detail

Abhishaike Mahajan26 Oct 2025 20:33 UTC
127 points
18 comments11 min readLW link
(www.owlposting.com)

Re­con­tex­tu­al­iza­tion Miti­gates Speci­fi­ca­tion Gam­ing Without Mod­ify­ing the Specification

14 Oct 2025 0:53 UTC
125 points
15 comments9 min readLW link

The main way I’ve seen peo­ple turn ide­olog­i­cally crazy [Linkpost]

Noosphere8923 Oct 2025 20:09 UTC
123 points
22 comments8 min readLW link
(andymasley.substack.com)

Grad­ual Disem­pow­er­ment Monthly Roundup

Raymond Douglas6 Oct 2025 15:36 UTC
119 points
9 comments6 min readLW link

Give Me Your Data: The Ra­tion­al­ist Mind Meld

Taylor G. Lunt19 Oct 2025 2:25 UTC
114 points
14 comments4 min readLW link

LLM robots can’t pass but­ter (and they are hav­ing an ex­is­ten­tial crisis about it)

Lukas Petersson28 Oct 2025 14:14 UTC
105 points
7 comments4 min readLW link

OpenAI #15: More on OpenAI’s Para­noid Law­fare Against Ad­vo­cates of SB 53

Zvi13 Oct 2025 15:00 UTC
104 points
2 comments23 min readLW link
(thezvi.wordpress.com)

Mus­ings on Re­ported Cost of Com­pute (Oct 2025)

Vladimir_Nesov24 Oct 2025 20:42 UTC
103 points
11 comments2 min readLW link

Where does Son­net 4.5′s de­sire to “not get too com­fortable” come from?

Kaj_Sotala4 Oct 2025 10:19 UTC
103 points
23 comments64 min readLW link

Con­sid­er­a­tions around ca­reer costs of poli­ti­cal donations

GradientDissenter20 Oct 2025 12:51 UTC
97 points
17 comments15 min readLW link

You Should Get a Reusable Mask

jefftk8 Oct 2025 2:40 UTC
96 points
28 comments1 min readLW link
(www.jefftk.com)

The Think­ing Machines Tinker API is good news for AI con­trol and security

Buck9 Oct 2025 15:22 UTC
91 points
10 comments6 min readLW link

Bend­ing The Curve

Zvi7 Oct 2025 20:00 UTC
91 points
12 comments21 min readLW link
(thezvi.wordpress.com)

Is 90% of code at An­thropic be­ing writ­ten by AIs?

ryan_greenblatt22 Oct 2025 14:50 UTC
90 points
14 comments5 min readLW link

Learn­ing to In­ter­pret Weight Differ­ences in Lan­guage Models

avichal23 Oct 2025 3:55 UTC
89 points
2 comments5 min readLW link
(arxiv.org)

Mak­ing Your Pain Worse can Get You What You Want

Logan Riggs5 Oct 2025 0:19 UTC
87 points
5 comments3 min readLW link

Rea­sons to sign a state­ment to ban su­per­in­tel­li­gence (+ FAQ for those on the fence)

13 Oct 2025 19:00 UTC
83 points
4 comments13 min readLW link

The Bio­chem­i­cal Beauty of Re­ta­tru­tide: How GLP-1s Ac­tu­ally Work

Elizabeth14 Oct 2025 16:00 UTC
82 points
3 comments7 min readLW link
(acesounderglass.com)

New State­ment Calls For Not Build­ing Su­per­in­tel­li­gence For Now

Zvi24 Oct 2025 17:40 UTC
80 points
3 comments7 min readLW link
(thezvi.wordpress.com)