Ti­maeus in 2024

Feb 20, 2025, 11:54 PM
99 points
1 comment8 min readLW link

Biolog­i­cal hu­mans col­lec­tively ex­ert at most 400 gi­gabits/​s of con­trol over the world.

benwrFeb 20, 2025, 11:44 PM
15 points
3 comments1 min readLW link

The first RCT for GLP-1 drugs and al­co­holism isn’t what we hoped

dynomightFeb 20, 2025, 10:30 PM
60 points
4 comments6 min readLW link
(dynomight.net)

Pub­lished re­port: Path­ways to short TAI timelines

Zershaaneh QureshiFeb 20, 2025, 10:10 PM
22 points
0 commentsLW link
(www.convergenceanalysis.org)

Neu­ral Scal­ing Laws Rooted in the Data Distribution

aribrillFeb 20, 2025, 9:22 PM
7 points
0 comments1 min readLW link
(arxiv.org)

De­mon­strat­ing speci­fi­ca­tion gam­ing in rea­son­ing models

Matrice JacobineFeb 20, 2025, 7:26 PM
4 points
0 comments1 min readLW link
(arxiv.org)

What makes a the­ory of in­tel­li­gence use­ful?

Cole WyethFeb 20, 2025, 7:22 PM
16 points
0 comments11 min readLW link

AI #104: Amer­i­can State Ca­pac­ity on the Brink

ZviFeb 20, 2025, 2:50 PM
37 points
9 comments44 min readLW link
(thezvi.wordpress.com)

US AI Safety In­sti­tute will be ‘gut­ted,’ Ax­ios reports

Matrice JacobineFeb 20, 2025, 2:40 PM
11 points
1 commentLW link
(www.zdnet.com)

Hu­man-AI Re­la­tion­al­ity is Already Here

bridgebotFeb 20, 2025, 7:08 AM
17 points
0 comments15 min readLW link

Safe Distil­la­tion With a Pow­er­ful Un­trusted AI

Alek WestoverFeb 20, 2025, 3:14 AM
5 points
1 comment5 min readLW link

Mo­du­lar­ity and as­sem­bly: AI safety via think­ing smaller

D WongFeb 20, 2025, 12:58 AM
2 points
0 comments11 min readLW link
(criticalreason.substack.com)

Eliezer’s Lost Align­ment Ar­ti­cles /​ The Ar­bital Sequence

Feb 20, 2025, 12:48 AM
207 points
10 comments5 min readLW link

Ar­bital has been im­ported to LessWrong

Feb 20, 2025, 12:47 AM
281 points
30 comments5 min readLW link

The Dilemma’s Dilemma

James Stephen BrownFeb 19, 2025, 11:50 PM
6 points
11 comments7 min readLW link
(nonzerosum.games)

[Question] Why do we have the NATO logo?

KvmanThinkingFeb 19, 2025, 10:59 PM
1 point
4 comments1 min readLW link

Me­tac­u­lus Q4 AI Bench­mark­ing: Bots Are Clos­ing The Gap

Feb 19, 2025, 10:42 PM
13 points
0 comments13 min readLW link
(www.metaculus.com)

Sev­eral Ar­gu­ments Against the Math­e­mat­i­cal Uni­verse Hy­poth­e­sis

Vittu PerkeleFeb 19, 2025, 10:13 PM
−4 points
6 comments3 min readLW link
(open.substack.com)

Liter­a­ture Re­view of Text AutoEncoders

NickyPFeb 19, 2025, 9:54 PM
20 points
5 comments8 min readLW link

Deep­Seek Made it Even Harder for US AI Com­pa­nies to Ever Reach Profitability

garrisonFeb 19, 2025, 9:02 PM
10 points
1 commentLW link
(garrisonlovely.substack.com)

Won’t vs. Can’t: Sand­bag­ging-like Be­hav­ior from Claude Models

Feb 19, 2025, 8:47 PM
15 points
1 comment1 min readLW link
(alignment.anthropic.com)

AI Align­ment and the Fi­nan­cial War Against Nar­cis­sis­tic Manipulation

henophiliaFeb 19, 2025, 8:42 PM
−17 points
2 comments3 min readLW link

How to Make Superbabies

Feb 19, 2025, 8:39 PM
608 points
351 comments31 min readLW link

The New­bie’s Guide to Nav­i­gat­ing AI Futures

keithjmenezesFeb 19, 2025, 8:37 PM
−1 points
0 comments40 min readLW link

Against Un­limited Ge­nius for Baby-Killers

gggggFeb 19, 2025, 8:33 PM
−7 points
1 comment3 min readLW link
(ggggggggggggggggggggggg.substack.com)

New LLM Scal­ing Law

wrmedfordFeb 19, 2025, 8:21 PM
2 points
0 comments1 min readLW link
(github.com)

Go Grok Yourself

ZviFeb 19, 2025, 8:20 PM
57 points
2 comments17 min readLW link
(thezvi.wordpress.com)

[Question] Take over my pro­ject: do com­putable agents plan against the uni­ver­sal dis­tri­bu­tion pes­simisti­cally?

Cole WyethFeb 19, 2025, 8:17 PM
25 points
3 comments3 min readLW link

When should we worry about AI power-seek­ing?

Joe CarlsmithFeb 19, 2025, 7:44 PM
20 points
0 comments18 min readLW link
(joecarlsmith.substack.com)

Su­perBa­bies pod­cast with Gene Smith

EneaszFeb 19, 2025, 7:36 PM
35 points
1 comment1 min readLW link
(thebayesianconspiracy.substack.com)

Un­de­sir­able Con­clu­sions and Ori­gin Adjustment

JerdleFeb 19, 2025, 6:35 PM
3 points
0 comments5 min readLW link

How might we safely pass the buck to AI?

joshcFeb 19, 2025, 5:48 PM
83 points
58 comments31 min readLW link

Us­ing Prompt Eval­u­a­tion to Com­bat Bio-Weapon Research

Feb 19, 2025, 12:39 PM
11 points
2 comments3 min readLW link

In­tel­li­gence Is Jagged

Adam TrainFeb 19, 2025, 7:08 AM
6 points
1 comment3 min readLW link

Closed-ended ques­tions aren’t as hard as you think

electroswingFeb 19, 2025, 3:53 AM
6 points
0 comments3 min readLW link

Un­der­grad AI Safety Conference

JoNeedsSleepFeb 19, 2025, 3:43 AM
18 points
0 comments1 min readLW link

Per­ma­nent prop­er­ties of things are a self-fulfilling prophecy

YanLyutnevFeb 19, 2025, 12:08 AM
4 points
0 comments9 min readLW link

Places of Lov­ing Grace [Story]

ankFeb 18, 2025, 11:49 PM
−1 points
0 comments4 min readLW link

Are SAE fea­tures from the Base Model still mean­ingful to LLaVA?

Shan23ChenFeb 18, 2025, 10:16 PM
8 points
2 comments10 min readLW link
(www.lesswrong.com)

Sparse Au­toen­coder Fea­tures for Clas­sifi­ca­tions and Transferability

Shan23ChenFeb 18, 2025, 10:14 PM
5 points
0 comments1 min readLW link
(arxiv.org)

A fable on AI x-risk

bgaesopFeb 18, 2025, 8:15 PM
8 points
4 comments1 min readLW link

The Un­earned Priv­ilege We Rarely Dis­cuss: Cog­ni­tive Capability

DiegoRojasFeb 18, 2025, 8:06 PM
−21 points
7 comments3 min readLW link

Call for Ap­pli­ca­tions: XLab Sum­mer Re­search Fel­low­ship

JoNeedsSleepFeb 18, 2025, 7:19 PM
9 points
0 comments1 min readLW link

AISN #48: Utility Eng­ineer­ing and EnigmaEval

Feb 18, 2025, 7:15 PM
4 points
0 comments4 min readLW link
(newsletter.safe.ai)

Ab­stract Math­e­mat­i­cal Con­cepts vs. Ab­strac­tions Over Real-World Systems

Thane RuthenisFeb 18, 2025, 6:04 PM
32 points
10 comments4 min readLW link

How ac­cu­rate was my “Altered Traits” book re­view?

lsusrFeb 18, 2025, 5:00 PM
41 points
3 comments3 min readLW link

Med­i­cal Roundup #4

ZviFeb 18, 2025, 1:40 PM
24 points
3 comments10 min readLW link
(thezvi.wordpress.com)

Dear AGI,

Nathan YoungFeb 18, 2025, 10:48 AM
88 points
11 comments3 min readLW link

There are a lot of up­com­ing re­treats/​con­fer­ences be­tween March and July (2025)

Feb 18, 2025, 9:30 AM
6 points
0 comments1 min readLW link

Sea Change

Charlie SandersFeb 18, 2025, 6:03 AM
−2 points
2 comments5 min readLW link
(www.dailymicrofiction.com)