Proac­tive ‘If-Then’ Safety Cases

Nathan Helm-Burger18 Nov 2024 21:16 UTC
10 points
0 comments4 min readLW link

[Question] Will Orion/​Gem­ini 2/​Llama-4 out­perform o1

LuigiPagani18 Nov 2024 21:15 UTC
2 points
3 comments1 min readLW link

How to use bright light to im­prove your life.

Nat Martin18 Nov 2024 19:32 UTC
40 points
10 comments10 min readLW link

How likely is brain preser­va­tion to work?

Andy_McKenzie18 Nov 2024 16:58 UTC
26 points
3 comments6 min readLW link

Why im­perfect ad­ver­sar­ial ro­bust­ness doesn’t doom AI control

18 Nov 2024 16:05 UTC
62 points
25 comments2 min readLW link

Eth­i­cal Im­pli­ca­tions of the Quan­tum Multiverse

Jonah Wilberg18 Nov 2024 16:00 UTC
7 points
22 comments6 min readLW link

Re­duc­ing x-risk might be ac­tively harmful

MountainPath18 Nov 2024 14:25 UTC
5 points
5 comments1 min readLW link

Monthly Roundup #24: Novem­ber 2024

Zvi18 Nov 2024 13:20 UTC
44 points
14 comments50 min readLW link
(thezvi.wordpress.com)

A Straight­for­ward Ex­pla­na­tion of the Good Reg­u­la­tor Theorem

Alfred Harwood18 Nov 2024 12:45 UTC
82 points
29 comments14 min readLW link

The Choice Transition

18 Nov 2024 12:30 UTC
54 points
4 comments15 min readLW link
(strangecities.substack.com)

Chat Bankman-Fried: an Ex­plo­ra­tion of LLM Align­ment in Finance

claudia.biancotti18 Nov 2024 9:38 UTC
26 points
4 comments1 min readLW link

Pro­posal to in­crease fer­til­ity: Univer­sity par­ent clubs

Fluffnutt18 Nov 2024 4:21 UTC
17 points
3 comments1 min readLW link

A small im­prove­ment to Wikipe­dia page on Pareto Efficiency

Edwin Evans18 Nov 2024 2:13 UTC
8 points
0 comments1 min readLW link

[Question] Why is Gem­ini tel­ling the user to die?

Burny18 Nov 2024 1:44 UTC
13 points
1 comment1 min readLW link

“It’s a 10% chance which I did 10 times, so it should be 100%”

egor.timatkov18 Nov 2024 1:14 UTC
159 points
59 comments2 min readLW link

The Catas­tro­phe of Shiny Objects

mindprison18 Nov 2024 0:24 UTC
−11 points
0 comments3 min readLW link

Do Deep Neu­ral Net­works Have Brain-like Rep­re­sen­ta­tions?: A Sum­mary of Disagreements

Joseph Emerson18 Nov 2024 0:07 UTC
9 points
0 comments26 min readLW link

Truth Ter­mi­nal: A re­con­struc­tion of events

17 Nov 2024 23:51 UTC
5 points
1 comment7 min readLW link

Which AI Safety Bench­mark Do We Need Most in 2025?

17 Nov 2024 23:50 UTC
2 points
2 comments8 min readLW link

“The Solomonoff Prior is Mal­ign” is a spe­cial case of a sim­pler argument

David Matolcsi17 Nov 2024 21:32 UTC
131 points
46 comments12 min readLW link

Chess As The Model Game

criticalpoints17 Nov 2024 19:45 UTC
19 points
0 comments8 min readLW link
(eregis.github.io)

The grass is always greener in the en­vi­ron­ment that shaped your values

Karl Faulks17 Nov 2024 18:00 UTC
8 points
0 comments3 min readLW link

An­nounc­ing turn­trout.com, my new digi­tal home

TurnTrout17 Nov 2024 17:42 UTC
108 points
33 comments1 min readLW link
(turntrout.com)

Sec­u­lar Sols­tice Song­book Update

jefftk17 Nov 2024 17:30 UTC
14 points
2 comments1 min readLW link
(www.jefftk.com)

Ger­many-wide ACX Meetup

Fernand017 Nov 2024 10:08 UTC
4 points
0 comments1 min readLW link

Pro­ject Ad­e­quate: Seek­ing Cofounders/​Funders

Lorec17 Nov 2024 3:12 UTC
5 points
7 comments8 min readLW link

Try­ing Bluesky

jefftk17 Nov 2024 2:50 UTC
26 points
16 comments1 min readLW link
(www.jefftk.com)

AXRP Epi­sode 38.1 - Alan Chan on Agent Infrastructure

DanielFilan16 Nov 2024 23:30 UTC
12 points
0 comments14 min readLW link

Cross-con­text ab­duc­tion: LLMs make in­fer­ences about pro­ce­du­ral train­ing data lev­er­ag­ing declar­a­tive facts in ear­lier train­ing data

Sohaib Imran16 Nov 2024 23:22 UTC
36 points
11 comments14 min readLW link

Why We Wouldn’t Build Aligned AI Even If We Could

Snowyiu16 Nov 2024 20:19 UTC
10 points
7 comments10 min readLW link

Which evals re­sources would be good?

Marius Hobbhahn16 Nov 2024 14:24 UTC
51 points
4 comments5 min readLW link

OpenAI Email Archives (from Musk v. Alt­man and OpenAI blog)

habryka16 Nov 2024 6:38 UTC
533 points
81 comments51 min readLW link

Us­ing Danger­ous AI, But Safely?

habryka16 Nov 2024 4:29 UTC
17 points
2 comments43 min readLW link

Ayn Rand’s model of “liv­ing money”; and an up­side of burnout

AnnaSalamon16 Nov 2024 2:59 UTC
237 points
59 comments5 min readLW link

Fun­da­men­tal Uncer­tainty: Epilogue

Gordon Seidoh Worley16 Nov 2024 0:57 UTC
10 points
0 comments1 min readLW link

Mak­ing a con­ser­va­tive case for alignment

15 Nov 2024 18:55 UTC
208 points
67 comments7 min readLW link

The Case For Giv­ing To The Shrimp Welfare Project

Bentham's Bulldog15 Nov 2024 16:03 UTC
−4 points
14 comments7 min readLW link

Win/​con­tinue/​lose sce­nar­ios and ex­e­cute/​re­place/​au­dit protocols

Buck15 Nov 2024 15:47 UTC
64 points
2 comments7 min readLW link

Antonym Heads Pre­dict Se­man­tic Op­po­sites in Lan­guage Models

Jake Ward15 Nov 2024 15:32 UTC
3 points
0 comments5 min readLW link

Propos­ing the Con­di­tional AI Safety Treaty (linkpost TIME)

otto.barten15 Nov 2024 13:59 UTC
11 points
9 comments3 min readLW link
(time.com)

A The­ory of Equil­ibrium in the Offense-Defense Balance

Maxwell Tabarrok15 Nov 2024 13:51 UTC
25 points
6 comments2 min readLW link
(www.maximum-progress.com)

Bos­ton Sec­u­lar Sols­tice 2024: Call for Singers and Musicans

jefftk15 Nov 2024 13:50 UTC
22 points
0 comments1 min readLW link
(www.jefftk.com)

An Un­canny Moat

Adam Newgas15 Nov 2024 11:39 UTC
13 points
0 comments4 min readLW link
(www.boristhebrave.com)

If I care about mea­sure, choices have ad­di­tional bur­den (+AI gen­er­ated LW-com­ments)

avturchin15 Nov 2024 10:27 UTC
5 points
11 comments2 min readLW link

What are Emo­tions?

Myles H15 Nov 2024 4:20 UTC
5 points
13 comments8 min readLW link

The Third Fun­da­men­tal Question

Screwtape15 Nov 2024 4:01 UTC
66 points
7 comments6 min readLW link

Dance Differentiation

jefftk15 Nov 2024 2:30 UTC
14 points
0 comments1 min readLW link
(www.jefftk.com)

Break­ing be­liefs about sav­ing the world

Oxidize15 Nov 2024 0:46 UTC
−1 points
3 comments9 min readLW link

Col­lege tech­ni­cal AI safety hackathon ret­ro­spec­tive—Ge­or­gia Tech

yix15 Nov 2024 0:22 UTC
44 points
2 comments5 min readLW link
(open.substack.com)

Gw­ern Bran­wen in­ter­view on Dwarkesh Pa­tel’s pod­cast: “How an Anony­mous Re­searcher Pre­dicted AI’s Tra­jec­tory”

Said Achmiz14 Nov 2024 23:53 UTC
87 points
0 comments1 min readLW link
(www.dwarkeshpatel.com)