GiveCalc: Open-source tool to calcu­late the true cost of char­i­ta­ble giving

Max Ghenis2 Dec 2025 23:56 UTC
5 points
1 comment2 min readLW link

Effec­tive Pizzaism

Screwtape2 Dec 2025 23:50 UTC
45 points
1 comment8 min readLW link

TastyBench: Toward Mea­sur­ing Re­search Taste in LLM

2 Dec 2025 23:26 UTC
27 points
2 comments6 min readLW link

AI Safety at the Fron­tier: Paper High­lights of Novem­ber 2025

gasteigerjo2 Dec 2025 21:11 UTC
6 points
0 comments8 min readLW link
(aisafetyfrontier.substack.com)

Open Thread Win­ter 2025/​26

kave2 Dec 2025 19:27 UTC
21 points
59 comments1 min readLW link

Prac­ti­cal AI risk II: Train­ing transparency

Gustavo Ramires2 Dec 2025 19:26 UTC
1 point
0 comments1 min readLW link

Five ways AI can tell you’re test­ing it

sjadler2 Dec 2025 17:25 UTC
16 points
0 comments15 min readLW link
(stevenadler.substack.com)

Why Moloch is ac­tu­ally the God of Evolu­tion­ary Pri­soner’s Dilemmas

Jonah Wilberg2 Dec 2025 16:54 UTC
32 points
2 comments11 min readLW link

Re­ward Mis­matches in RL Cause Emer­gent Misalignment

Zvi2 Dec 2025 16:31 UTC
70 points
1 comment7 min readLW link
(thezvi.wordpress.com)

Sci.STEPS in­vites mentee applications

Valentin20262 Dec 2025 13:33 UTC
7 points
0 comments1 min readLW link

How Claude Opus 4.5 de­scribes its ex­pe­rience of var­i­ous concepts

Kaj_Sotala2 Dec 2025 13:05 UTC
16 points
1 comment65 min readLW link

Safety Cases Ex­plained: How to Ar­gue an AI is Safe

JanWehner2 Dec 2025 11:03 UTC
16 points
2 comments9 min readLW link

The Hid­den Asym­me­try in Per­sonal Pre­pared­ness: Early Costs, Late Losses

Ulrik Horn2 Dec 2025 10:33 UTC
6 points
5 comments15 min readLW link

Halfhaven Digest 6 + Retrospective

Taylor G. Lunt2 Dec 2025 5:27 UTC
20 points
2 comments3 min readLW link

Met­ric-haven (quick stats on how Inkhaven im­pacted LessWrong)

Ruby2 Dec 2025 3:31 UTC
26 points
3 comments1 min readLW link

MIRI’s 2025 Fundraiser

alexvermeer2 Dec 2025 1:53 UTC
176 points
7 comments8 min readLW link

Every­one Can Be High Sta­tus In Utopia

Algon1 Dec 2025 23:43 UTC
12 points
5 comments2 min readLW link

GRPO is terrible

RobinHa1 Dec 2025 22:54 UTC
4 points
2 comments5 min readLW link
(robinhaselhorst.com)

How to Write Fast, Weird, and Well

Linch1 Dec 2025 21:30 UTC
44 points
1 comment18 min readLW link
(inchpin.substack.com)

The 2024 LessWrong Review

RobertM1 Dec 2025 21:06 UTC
63 points
10 comments7 min readLW link

Fu­ture Proofing Solstice

Raemon1 Dec 2025 20:57 UTC
51 points
7 comments1 min readLW link

Why ra­tio­nal­ists get depressed

Pjain1 Dec 2025 20:07 UTC
28 points
0 comments17 min readLW link

An­nounc­ing: OpenAI’s Align­ment Re­search Blog

Naomi Bashkansky1 Dec 2025 19:52 UTC
120 points
11 comments1 min readLW link

AI Men­tal Health Chat­bots for Low-Re­source Set­tings: A Pri­ori­ti­za­tion Framework

Dawn Drescher1 Dec 2025 17:41 UTC
6 points
0 comments16 min readLW link

Which planet is clos­est to the Earth, and why is it Mer­cury?

Menotim1 Dec 2025 17:16 UTC
27 points
5 comments4 min readLW link

How mid­dle pow­ers may pre­vent the de­vel­op­ment of ar­tifi­cial superintelligence

1 Dec 2025 16:48 UTC
127 points
12 comments3 min readLW link
(asi-prevention.com)

Be­com­ing a Chi­nese Room

Raelifin1 Dec 2025 16:34 UTC
39 points
3 comments6 min readLW link
(raelifin.substack.com)

Well, Sea­sons Great­ings Every­one! [Short Fic­tion]

Shiva's Right Foot1 Dec 2025 16:29 UTC
15 points
0 comments3 min readLW link

23 thoughts on Ar­tifi­cial In­tel­li­gence (2025)

Annapurna1 Dec 2025 16:01 UTC
1 point
0 comments5 min readLW link

Lorxus Does Halfhaven: 11/​22~11/​28

Lorxus1 Dec 2025 14:47 UTC
5 points
0 comments2 min readLW link
(tiled-with-pentagons.blogspot.com)

Would ASI de­vel­op­ment in non-party states un­der­mine a non­pro­lifer­a­tion agree­ment?

Robi Rahman1 Dec 2025 14:22 UTC
13 points
0 comments9 min readLW link

Claude Opus 4.5 Is The Best Model Available

Zvi1 Dec 2025 14:01 UTC
74 points
2 comments21 min readLW link
(thezvi.wordpress.com)

How Can In­ter­pretabil­ity Re­searchers Help AGI Go Well?

1 Dec 2025 13:05 UTC
66 points
1 comment14 min readLW link

A Prag­matic Vi­sion for Interpretability

1 Dec 2025 13:05 UTC
131 points
39 comments27 min readLW link

HERMES: Towards Effi­cient and Ver­ifi­able Math­e­mat­i­cal Rea­son­ing in LLMs

Gunnar_Zarncke1 Dec 2025 10:07 UTC
8 points
0 comments1 min readLW link
(arxiv.org)

Align­ment as an Eval­u­a­tion Problem

wolverdude1 Dec 2025 10:04 UTC
16 points
0 comments6 min readLW link

In­ter­view: What it’s like to be a bat

Saul Munn1 Dec 2025 9:35 UTC
77 points
6 comments4 min readLW link
(www.brasstacks.blog)

Slack Observability

Alice Blair1 Dec 2025 7:52 UTC
32 points
0 comments2 min readLW link

A Statis­ti­cal Anal­y­sis of Inkhaven

Ben Pace1 Dec 2025 7:47 UTC
34 points
0 comments8 min readLW link

Is the ev­i­dence in “Lan­guage Models Learn to Mislead Hu­mans via RLHF” valid?

1 Dec 2025 6:50 UTC
35 points
0 comments19 min readLW link

How To De­ploy a (Tiny) AI

michaelwaves1 Dec 2025 6:22 UTC
1 point
0 comments2 min readLW link

10 Apho­risms from 𝘛𝘩𝘦 𝘉𝘦𝘥 𝘰𝘧 𝘗𝘳𝘰𝘤𝘳𝘶𝘴𝘵𝘦𝘴

lsusr1 Dec 2025 5:55 UTC
10 points
7 comments1 min readLW link

Inkhaven: 30 Days, 30 Memories

Collisteru1 Dec 2025 5:24 UTC
42 points
2 comments3 min readLW link

Novem­ber Retrospective

johnswentworth1 Dec 2025 4:20 UTC
53 points
3 comments2 min readLW link

Models of Mind: Are You Mono- or Multi-Threaded?

Shoshannah Tekofsky1 Dec 2025 3:35 UTC
25 points
5 comments3 min readLW link
(shoshanigans.substack.com)

Video qual­ity is mainly not de­ter­mined by resolution

Awaclus1 Dec 2025 2:52 UTC
8 points
7 comments3 min readLW link

In­vert­ing qualia with group theory

jessicata1 Dec 2025 1:44 UTC
39 points
23 comments8 min readLW link
(unstableontology.com)

What makes a good egg?

Metacelsus1 Dec 2025 0:31 UTC
34 points
0 comments4 min readLW link
(denovo.substack.com)

Soul-Whore

Eneasz1 Dec 2025 0:06 UTC
47 points
0 comments5 min readLW link
(deathisbad.substack.com)

Choose Your Failure Modes

Algon30 Nov 2025 23:56 UTC
9 points
0 comments2 min readLW link