Why’s equal­ity in logic less flex­ible than in cat­e­gory the­ory?

Algon1 Oct 2025 22:03 UTC
17 points
24 comments3 min readLW link

[Linkpost] A Field Guide to Writ­ing Styles

Linch1 Oct 2025 21:49 UTC
17 points
0 comments17 min readLW link
(linch.substack.com)

</​rant> </​un­char­i­ta­ble> </​psy­chol­o­giz­ing>

Raemon1 Oct 2025 21:20 UTC
53 points
11 comments2 min readLW link

How I think about al­ign­ment and ethics as a co­op­er­a­tion pro­to­col soft­ware

Burny1 Oct 2025 21:09 UTC
3 points
0 comments1 min readLW link

In­tro­duc­ing the Mox Guest Program

1 Oct 2025 18:35 UTC
11 points
0 comments2 min readLW link
(moxsf.com)

The Prob­lem of the Con­cen­tra­tion of Power

hazem1 Oct 2025 18:13 UTC
−5 points
2 comments2 min readLW link

Claude Son­net 4.5 Is A Very Good Model

Zvi1 Oct 2025 18:00 UTC
40 points
2 comments24 min readLW link
(thezvi.wordpress.com)

My Brush with Su­per­hu­man Persuasion

Ben S.1 Oct 2025 17:50 UTC
18 points
13 comments9 min readLW link
(thebsdetector.substack.com)

AI and Cheap Weapons

Felix C.1 Oct 2025 17:31 UTC
31 points
3 comments23 min readLW link

But what kind of stuff can you just do?

Bastiaan1 Oct 2025 16:58 UTC
25 points
5 comments1 min readLW link

AI Safety at the Fron­tier: Paper High­lights, Septem­ber ’25

gasteigerjo1 Oct 2025 16:24 UTC
5 points
0 comments6 min readLW link
(aisafetyfrontier.substack.com)

Uncer­tain Up­dates: Septem­ber 2025

Gordon Seidoh Worley1 Oct 2025 14:50 UTC
11 points
0 comments1 min readLW link
(uncertainupdates.substack.com)

[CS2881r] Op­ti­miz­ing Prompts with Re­in­force­ment Learning

1 Oct 2025 14:02 UTC
2 points
0 comments5 min readLW link

“Pes­simiza­tion” is Just Or­di­nary Failure

J Bostock1 Oct 2025 13:48 UTC
56 points
2 comments6 min readLW link

Beyond the Zom­bie Argument

James Diacoumis1 Oct 2025 13:16 UTC
7 points
23 comments2 min readLW link
(jamesdiacoumis.substack.com)

Against the Inevita­bil­ity of Ha­bit­u­a­tion to Con­tin­u­ous Bliss

CanYouFeelTheBenefits1 Oct 2025 12:12 UTC
8 points
0 comments1 min readLW link

Lec­tures on statis­ti­cal learn­ing the­ory for al­ign­ment researchers

Vanessa Kosoy1 Oct 2025 8:36 UTC
41 points
1 comment1 min readLW link
(www.youtube.com)

Claude Son­net 4.5: Sys­tem Card and Alignment

Zvi30 Sep 2025 20:50 UTC
72 points
4 comments27 min readLW link
(thezvi.wordpress.com)

Halfhaven vir­tual blog­ger camp

Viliam30 Sep 2025 20:22 UTC
87 points
6 comments2 min readLW link

Masks: On the benefits and draw­backs of a so­ciety where ev­ery­one cov­er­ing their face is the norm

3Nora30 Sep 2025 18:43 UTC
−3 points
1 comment3 min readLW link

How reimag­in­ing the na­ture of con­scious­ness en­tirely changes the AI game

Jáchym Fibír30 Sep 2025 18:30 UTC
−9 points
0 comments14 min readLW link
(www.phiand.ai)

The Ba­sic Case For Doom

Bentham's Bulldog30 Sep 2025 16:04 UTC
26 points
4 comments5 min readLW link

AI Safety Re­search Futarchy: Us­ing Pre­dic­tion Mar­kets to Choose Re­search Pro­jects for MARS

JasonBrown30 Sep 2025 15:37 UTC
32 points
8 comments4 min readLW link

ARENA 7.0 - Call for Applicants

30 Sep 2025 14:54 UTC
22 points
0 comments6 min readLW link

The fa­mous sur­vivor­ship bias image is a “loose re­con­struc­tion” of meth­ods used on a hy­po­thet­i­cal dataset

Lao Mein30 Sep 2025 13:13 UTC
47 points
0 comments1 min readLW link

[GDP­val] Models Could Au­to­mate the U.S. Econ­omy by 2027

bira30 Sep 2025 11:53 UTC
14 points
0 comments1 min readLW link

Eth­i­cal De­sign Patterns

AnnaSalamon30 Sep 2025 11:52 UTC
210 points
39 comments20 min readLW link

What is the Base Model Si­mu­la­tion of Hu­man AI-As­sis­tant Con­ver­sa­tion?:

bodry30 Sep 2025 7:08 UTC
5 points
0 comments21 min readLW link

First­post: First impressions

Shell30 Sep 2025 2:23 UTC
14 points
1 comment1 min readLW link

Ex­plo­ra­tion of Coun­ter­fac­tual Im­por­tance and At­ten­tion Heads

Realmbird30 Sep 2025 1:17 UTC
12 points
0 comments6 min readLW link

Why Cor­rigi­bil­ity is Hard and Im­por­tant (i.e. “Whence the high MIRI con­fi­dence in al­ign­ment difficulty?”)

30 Sep 2025 0:12 UTC
80 points
52 comments17 min readLW link

What SB 53, Cal­ifor­nia’s new AI law, does

tlevin29 Sep 2025 23:29 UTC
104 points
12 comments4 min readLW link

Why Most Efforts Towards “Demo­cratic AI” Fall Short

jacobhaimes29 Sep 2025 20:52 UTC
2 points
0 comments6 min readLW link
(www.odysseaninstitute.org)

You’re prob­a­bly over­es­ti­mat­ing how well you un­der­stand Dun­ning-Kruger

abstractapplic29 Sep 2025 19:27 UTC
216 points
24 comments4 min readLW link

On Dwarkesh Pa­tel’s Pod­cast With Richard Sutton

Zvi29 Sep 2025 19:20 UTC
54 points
10 comments23 min readLW link
(thezvi.wordpress.com)

Con­trol­ling the op­tions AIs can pursue

Joe Carlsmith29 Sep 2025 17:23 UTC
15 points
0 comments35 min readLW link

Ex­po­nen­tial in­crease is the de­fault (as­sum­ing it in­creases at all) [Linkpost]

Noosphere8929 Sep 2025 16:13 UTC
13 points
0 comments2 min readLW link
(x.com)

[Question] How does the cur­rent AI paradigm give rise to the “su­per­a­gency” that IABIED is con­cerned with?

jchan29 Sep 2025 15:23 UTC
3 points
4 comments1 min readLW link

AI com­pa­nies’ policy ad­vo­cacy (Sep 2025)

Zach Stein-Perlman29 Sep 2025 15:00 UTC
43 points
0 comments3 min readLW link

KYC for ChatGPT? Prevent­ing AI Harms for Youth Should Not Mean Vio­lat­ing Every­one Else’s Pri­vacy Rights

Noah Weinberger29 Sep 2025 14:18 UTC
7 points
0 comments7 min readLW link

Sys­tem Level Safety Evaluations

29 Sep 2025 13:57 UTC
14 points
0 comments9 min readLW link
(equilibria1.substack.com)

I have de­cided to stop ly­ing to Amer­i­cans about 9/​11

Lao Mein29 Sep 2025 13:55 UTC
86 points
24 comments1 min readLW link

[Re­tracted] Guess I Was Wrong About AIxBio Risks

J Bostock29 Sep 2025 11:44 UTC
62 points
7 comments5 min readLW link

If Drexler Is Wrong, He May as Well Be Right

Tomás B.29 Sep 2025 7:00 UTC
51 points
8 comments2 min readLW link

Ap­plied Mur­phyjitsu Meditation

Alice Blair29 Sep 2025 6:31 UTC
20 points
0 comments3 min readLW link

The per­sonal in­tel­li­gence I want

Rebecca Dai29 Sep 2025 4:09 UTC
20 points
9 comments8 min readLW link
(rebeccadai.substack.com)

Why ASI Align­ment Is Hard (an overview)

Yotam29 Sep 2025 4:05 UTC
16 points
1 comment25 min readLW link

When the AI Dam Breaks: From Surveillance to Game The­ory in AI Alignment

pataphor29 Sep 2025 4:01 UTC
5 points
7 comments5 min readLW link

Yet Another IABIED Review

PeterMcCluskey28 Sep 2025 21:36 UTC
15 points
0 comments7 min readLW link
(bayesianinvestor.com)

A non-re­view of “If Any­one Builds It, Every­one Dies”

boazbarak28 Sep 2025 17:34 UTC
125 points
50 comments4 min readLW link