Anti-memes: x-risk edition

WillPetillo10 Apr 2025 23:35 UTC
15 points
0 comments7 min readLW link

Fore­cast­ing time to au­to­mated su­per­hu­man coders [AI 2027 Timelines Fore­cast]

10 Apr 2025 23:10 UTC
35 points
0 comments18 min readLW link
(ai-2027.com)

AI could cause a drop in GDP, even if mar­kets are com­pet­i­tive and efficient

Casey Barkan10 Apr 2025 22:35 UTC
29 points
0 comments5 min readLW link

Not The End of All Value

Ben Ihrig10 Apr 2025 20:53 UTC
−13 points
0 comments3 min readLW link

EA Reflec­tions on my Mili­tary Career

TomGardiner10 Apr 2025 19:01 UTC
7 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Text First, Ev­i­dence Later? Manag­ing Qual­ity and Trust in an Era of AI-Aug­mented Research

Thehumanproject.ai10 Apr 2025 18:52 UTC
1 point
1 comment5 min readLW link

Nuanced Models for the In­fluence of Information

ozziegooen10 Apr 2025 18:28 UTC
8 points
0 comments1 min readLW link

Play­ing in the Creek

Hastings10 Apr 2025 17:39 UTC
396 points
13 comments2 min readLW link
(hgreer.com)

The Three Boxes: A Sim­ple Model for Spread­ing Ideas

JohnGreer10 Apr 2025 17:15 UTC
6 points
0 comments5 min readLW link

Re­ac­tions to METR task length pa­per are insane

Cole Wyeth10 Apr 2025 17:13 UTC
59 points
43 comments4 min readLW link

Ex­ist­ing Safety Frame­works Im­ply Un­rea­son­able Confidence

10 Apr 2025 16:31 UTC
46 points
3 comments15 min readLW link
(intelligence.org)

Ar­gu­ments for and against grad­ual change

Gustavo Ramires10 Apr 2025 14:43 UTC
3 points
0 comments6 min readLW link

Disem­pow­er­ment spirals as a likely mechanism for ex­is­ten­tial catastrophe

10 Apr 2025 14:37 UTC
74 points
7 comments5 min readLW link

AI #111: Giv­ing Us Pause

Zvi10 Apr 2025 14:00 UTC
26 points
4 comments34 min readLW link
(thezvi.wordpress.com)

Forg­ing A New AGI So­cial Contract

Deric Cheng10 Apr 2025 13:41 UTC
23 points
3 comments7 min readLW link
(agisocialcontract.substack.com)

Why Ex­pe­rienced Pro­fes­sion­als Fail to Land High-Im­pact Roles (FBB #5)

gergogaspar10 Apr 2025 12:46 UTC
12 points
4 comments9 min readLW link

Linkpost to a Sum­mary of “Imag­in­ing and build­ing wise ma­chines: The cen­tral­ity of AI metacog­ni­tion” by John­son, Karimi, Ben­gio, et al.

Chris_Leong10 Apr 2025 11:54 UTC
8 points
0 comments2 min readLW link

Grounded Ghosts in the Ma­chine—Fris­ton Blan­kets, Mir­ror Neu­rons, and the Quest for Co­op­er­a­tive AI

Davidmanheim10 Apr 2025 10:15 UTC
9 points
0 comments9 min readLW link
(davidmanheim.com)

New Paper: In­fra-Bayesian De­ci­sion-Es­ti­ma­tion Theory

10 Apr 2025 9:17 UTC
77 points
4 comments1 min readLW link
(arxiv.org)

Elec­tric Lunchbox

jefftk10 Apr 2025 2:40 UTC
15 points
0 comments1 min readLW link
(www.jefftk.com)

Scop­ing LLMs

10 Apr 2025 0:32 UTC
4 points
0 comments22 min readLW link

[Question] How fa­mil­iar is the Less­wrong com­mu­nity as a whole with the con­cept of Re­ward-mod­el­ling?

Oxidize9 Apr 2025 23:33 UTC
1 point
8 comments1 min readLW link

What can we learn from ex­pert AGI fore­casts?

Benjamin_Todd9 Apr 2025 21:34 UTC
5 points
0 comments5 min readLW link
(80000hours.org)

Thoughts on AI 2027

Max Harms9 Apr 2025 21:26 UTC
222 points
61 comments21 min readLW link
(intelligence.org)

The case for AGI by 2030

Benjamin_Todd9 Apr 2025 20:35 UTC
40 points
6 comments42 min readLW link
(80000hours.org)

Anti-au­toma­tion policy as a bot­tle­neck to eco­nomic growth

mhampton9 Apr 2025 20:12 UTC
4 points
0 comments4 min readLW link

Rea­son­ing mod­els don’t always say what they think

9 Apr 2025 19:48 UTC
28 points
4 comments1 min readLW link
(www.anthropic.com)

Re­v­erse en­g­ineer­ing the mem­ory lay­out of GPU inference

Paul Bricman9 Apr 2025 15:40 UTC
5 points
0 comments6 min readLW link
(noemaresearch.com)

Llama Does Not Look Good 4 Anything

Zvi9 Apr 2025 13:20 UTC
31 points
1 comment16 min readLW link
(thezvi.wordpress.com)

Learned pain as a lead­ing cause of chronic pain

SoerenMind9 Apr 2025 11:57 UTC
210 points
38 comments9 min readLW link

Tax­on­omy of possibility

dkl99 Apr 2025 4:24 UTC
13 points
1 comment5 min readLW link
(dkl9.net)

Short Timelines Don’t De­value Long Hori­zon Research

Vladimir_Nesov9 Apr 2025 0:42 UTC
170 points
24 comments1 min readLW link

A Plat­form for Falsifi­able Con­jec­tures and Public Re­fu­ta­tion — Would This Be Use­ful?

PetrusNonius8 Apr 2025 21:09 UTC
1 point
1 comment1 min readLW link

Quan­tify­ing SAE Qual­ity with Fea­ture Steer­abil­ity Metrics

phenomanon8 Apr 2025 20:55 UTC
2 points
0 comments4 min readLW link

MATS is hiring!

8 Apr 2025 20:45 UTC
8 points
0 comments6 min readLW link

birds and mam­mals in­de­pen­dently evolved intelligence

bhauth8 Apr 2025 20:00 UTC
73 points
23 comments1 min readLW link
(www.quantamagazine.org)

Align­ment Fak­ing Re­vis­ited: Im­proved Clas­sifiers and Open Source Extensions

8 Apr 2025 17:32 UTC
146 points
20 comments12 min readLW link

Lon­don Work­ing Group for Short/​Medium Term AI Risks

scronkfinkle8 Apr 2025 17:32 UTC
5 points
0 comments2 min readLW link

Think­ing Machines

Knight Lee8 Apr 2025 17:27 UTC
3 points
0 comments6 min readLW link

Digi­tal Er­ror Cor­rec­tion and Lock-In

alamerton8 Apr 2025 15:46 UTC
1 point
0 comments5 min readLW link
(alfielamerton.substack.com)

[Question] What faith­ful­ness met­rics should gen­eral claims about CoT faith­ful­ness be based upon?

Rauno Arike8 Apr 2025 15:27 UTC
24 points
0 comments4 min readLW link

AI 2027: Responses

Zvi8 Apr 2025 12:50 UTC
111 points
3 comments30 min readLW link
(thezvi.wordpress.com)

The first AI war will be in your com­puter

Viliam8 Apr 2025 9:28 UTC
43 points
10 comments3 min readLW link

Who wants to bet me $25k at 1:7 odds that there won’t be an AI mar­ket crash in the next year?

Remmelt8 Apr 2025 8:31 UTC
25 points
19 comments1 min readLW link

A Path­way to Fully Au­tonomous Therapists

Declan Molony8 Apr 2025 4:10 UTC
6 points
2 comments6 min readLW link

Re­think­ing Fric­tion: Equity and Mo­ti­va­tion Across Domains

eltimbalino8 Apr 2025 3:58 UTC
−1 points
0 comments2 min readLW link
(www.lesswrong.com)

On differ­ent dis­cus­sion traditions

Eugene Shcherbinin7 Apr 2025 23:00 UTC
1 point
0 comments2 min readLW link

Mis­in­for­ma­tion is the de­fault, and in­for­ma­tion is the gov­ern­ment tel­ling you your tap wa­ter is safe to drink

d_el_ez7 Apr 2025 22:28 UTC
10 points
2 comments9 min readLW link

Log-lin­ear Scal­ing is Worth the Cost due to Gains in Long-Hori­zon Tasks

shash427 Apr 2025 21:50 UTC
16 points
2 comments1 min readLW link

AI Safety at the Fron­tier: Paper High­lights, March ’25

gasteigerjo7 Apr 2025 20:17 UTC
9 points
0 comments9 min readLW link
(aisafetyfrontier.substack.com)