In­tent al­ign­ment seems incoherent

Joe Rogero7 Oct 2025 23:01 UTC
20 points
1 comment6 min readLW link

Petri: An open-source au­dit­ing tool to ac­cel­er­ate AI safety research

Sam Marks7 Oct 2025 20:39 UTC
60 points
0 comments1 min readLW link
(alignment.anthropic.com)

Bend­ing The Curve

Zvi7 Oct 2025 20:00 UTC
73 points
6 comments21 min readLW link
(thezvi.wordpress.com)

Kairos is hiring: Found­ing Gen­er­al­ist & SPAR Contractor

agucova7 Oct 2025 18:43 UTC
8 points
0 comments4 min readLW link

Messy on Pur­pose: Part 2 of A Con­ser­va­tive Vi­sion for the Future

7 Oct 2025 17:00 UTC
9 points
0 comments12 min readLW link

Go­ing Phoneless

robotelvis7 Oct 2025 16:40 UTC
17 points
0 comments5 min readLW link
(messyprogress.substack.com)

The Tower of Ba­bel in Reverse

Nostradamus_27 Oct 2025 16:27 UTC
6 points
0 comments7 min readLW link
(terminalvel0city.substack.com)

The Align­ment Para­dox: Why Trans­parency Can Breed Deception

Joseph Banks7 Oct 2025 13:28 UTC
4 points
0 comments7 min readLW link

Notes on “Ho­mol­ogy, Genes and Evolu­tion­ary In­no­va­tion”

Morpheus7 Oct 2025 12:45 UTC
8 points
1 comment2 min readLW link

Re­search Robots: When AIs Ex­per­i­ment on Us

Shoshannah Tekofsky7 Oct 2025 12:10 UTC
18 points
0 comments7 min readLW link
(theaidigest.org)

Top Warn­ing Signs Your Friends are Be­ing Oneshot­ted By AI

Charlie Edwards7 Oct 2025 11:56 UTC
−18 points
1 comment6 min readLW link

LLMs as a limiter of so­cial intercourse

Adam Zerner7 Oct 2025 6:38 UTC
17 points
4 comments2 min readLW link

[Question] Gen­er­al­iza­tion and the Mul­ti­ple Stage Fal­lacy?

Zack_M_Davis7 Oct 2025 6:20 UTC
34 points
6 comments3 min readLW link

Tel­ling the Differ­ence Between Me­mories & Log­i­cal Guesses

Logan Riggs7 Oct 2025 5:46 UTC
25 points
3 comments4 min readLW link

Notes from Euro­pean Progress Conference

Martin Sustrik7 Oct 2025 3:50 UTC
9 points
0 comments4 min readLW link
(www.250bpm.com)

“In­tel­li­gence” → “Re­lentless, Creative Re­source­ful­ness”

Raemon7 Oct 2025 0:28 UTC
63 points
28 comments17 min readLW link

Chaos Alone is No Bar to Superintelligence

Algon6 Oct 2025 22:45 UTC
11 points
0 comments2 min readLW link
(aisafety.info)

We won’t get AIs smart enough to solve al­ign­ment but too dumb to rebel

Joe Rogero6 Oct 2025 21:49 UTC
28 points
16 comments5 min readLW link

Notes on the need to lose

Algon6 Oct 2025 21:27 UTC
2 points
6 comments2 min readLW link

Ex­cerpts from my neu­ro­science to-do list

Steven Byrnes6 Oct 2025 21:05 UTC
26 points
1 comment4 min readLW link

Ex­pe­rience Re­port—ML4Good Boot­camp Sin­ga­pore, Sep′25

NurAlam6 Oct 2025 18:49 UTC
2 points
0 comments4 min readLW link

Which differ­ences be­tween sand­bag­ging eval­u­a­tions and sand­bag­ging safety re­search are im­por­tant for con­trol?

lennie6 Oct 2025 18:20 UTC
1 point
0 comments11 min readLW link

Grad­ual Disem­pow­er­ment Monthly Roundup

Raymond Douglas6 Oct 2025 15:36 UTC
93 points
7 comments6 min readLW link

Sublimi­nal Learn­ing, the Lot­tery-Ticket Hy­poth­e­sis, and Mode Connectivity

David Africa6 Oct 2025 15:26 UTC
16 points
3 comments7 min readLW link

The Origami Men

Tomás B.6 Oct 2025 15:25 UTC
138 points
9 comments16 min readLW link

Med­i­cal Roundup #5

Zvi6 Oct 2025 15:10 UTC
26 points
2 comments26 min readLW link
(thezvi.wordpress.com)

Sand­bag­ging: dis­t­in­guish­ing de­tec­tion of un­der­perfor­mance from in­crim­i­na­tion, and the im­pli­ca­tions for down­stream in­ter­ven­tions.

lennie6 Oct 2025 14:00 UTC
1 point
0 comments8 min readLW link

Why I think ECL shouldn’t make you up­date your cause prio

Jim Buhler6 Oct 2025 13:01 UTC
2 points
0 comments11 min readLW link

[Question] Did Tyler Robin­son carry his rifle as claimed by the gov­ern­ment?

ChristianKl6 Oct 2025 12:46 UTC
4 points
9 comments1 min readLW link

AI Science Com­pa­nies: Ev­i­dence AGI Is Near

Josh Snider6 Oct 2025 10:13 UTC
5 points
3 comments1 min readLW link
(www.joshuasnider.com)

LLMs one-box when in a “hos­tile telepath” ver­sion of New­comb’s Para­dox, ex­cept for the one that beat the predictor

Kaj_Sotala6 Oct 2025 8:44 UTC
47 points
6 comments17 min readLW link

Align­ment Fak­ing Demo for Con­gres­sional Staffers

Alice Blair6 Oct 2025 1:44 UTC
19 points
0 comments3 min readLW link

Do Things for as Many Rea­sons as Possible

Philipreal6 Oct 2025 0:28 UTC
35 points
1 comment2 min readLW link

One Does Not Sim­ply Walk Away from Omelas

Taylor G. Lunt6 Oct 2025 0:04 UTC
4 points
5 comments7 min readLW link

The quo­ta­tion mark

Maxwell Peterson5 Oct 2025 23:23 UTC
19 points
8 comments13 min readLW link

The Sadism Spec­trum and How to Ac­cess It

Dawn Drescher5 Oct 2025 23:09 UTC
13 points
2 comments20 min readLW link
(impartial-priorities.org)

Maybe so­cial me­dia al­gorithms don’t suck

Algon5 Oct 2025 18:47 UTC
64 points
18 comments3 min readLW link

Base64Bench: How good are LLMs at base64, and why care about it?

richbc5 Oct 2025 18:07 UTC
31 points
6 comments11 min readLW link

[Question] What can Cana­di­ans do to help end the AI arms race?

Tom9385 Oct 2025 18:03 UTC
8 points
7 comments2 min readLW link

17 years old, self-taught state con­trol—look­ing for peo­ple who ac­tu­ally get this

Cornelius Caspian5 Oct 2025 18:02 UTC
−3 points
3 comments1 min readLW link

Be­hav­ior Best-of-N achieves Near Hu­man Perfor­mance on Com­puter Tasks

Baybar5 Oct 2025 16:53 UTC
6 points
0 comments3 min readLW link

Ac­cel­er­at­ing AI Safety Progress via Tech­ni­cal Meth­ods- Cal­ling Re­searchers, Founders, and Funders

Martin Leitgab5 Oct 2025 16:40 UTC
1 point
0 comments1 min readLW link

Mini-Sym­po­sium on Ac­cel­er­at­ing AI Safety Progress via Tech­ni­cal Meth­ods—Hy­brid In-Per­son and Virtual

Martin Leitgab5 Oct 2025 16:05 UTC
1 point
0 comments1 min readLW link

[Question] How likely are “s-risks” (large-scale suffer­ing out­comes) from un­al­igned AI com­pared to ex­tinc­tion risks?

CanYouFeelTheBenefits5 Oct 2025 14:38 UTC
14 points
1 comment1 min readLW link

LLMs are badly misaligned

Joe Rogero5 Oct 2025 14:00 UTC
27 points
25 comments3 min readLW link

The Coun­ter­fac­tual Quiet AGI Timeline

Davidmanheim5 Oct 2025 9:09 UTC
64 points
5 comments9 min readLW link

AISafety.com Read­ing Group ses­sion 328

Søren Elverlin5 Oct 2025 7:51 UTC
5 points
0 comments1 min readLW link

How the NanoGPT Speedrun WR dropped by 20% in 3 months

larry-dial5 Oct 2025 1:05 UTC
26 points
9 comments9 min readLW link

a quick thought about AI alignment

foodforthought5 Oct 2025 0:51 UTC
10 points
4 comments1 min readLW link

Mak­ing Your Pain Worse can Get You What You Want

Logan Riggs5 Oct 2025 0:19 UTC
76 points
4 comments3 min readLW link