Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
The Deep Lore of LightHaven, with Oliver Habryka (TBC episode 228)
Eneasz
and
habryka
Dec 24, 2024, 10:45 PM
45
points
4
comments
91
min read
LW
link
(thebayesianconspiracy.substack.com)
Acknowledging Background Information with P(Q|I)
JenniferRM
Dec 24, 2024, 6:50 PM
29
points
8
comments
14
min read
LW
link
Game Theory and Behavioral Economics in The Stock Market
Jaiveer Singh
Dec 24, 2024, 6:15 PM
1
point
0
comments
3
min read
LW
link
[Question]
What are the main arguments against AGI?
Edy Nastase
Dec 24, 2024, 3:49 PM
1
point
6
comments
1
min read
LW
link
[Question]
Recommendations on communities that discuss AI applications in society
Annapurna
Dec 24, 2024, 1:37 PM
7
points
2
comments
1
min read
LW
link
AIs Will Increasingly Fake Alignment
Zvi
Dec 24, 2024, 1:00 PM
89
points
0
comments
52
min read
LW
link
(thezvi.wordpress.com)
Apply to the 2025 PIBBSS Summer Research Fellowship
DusanDNesic
and
Lucas Teixeira
Dec 24, 2024, 10:25 AM
15
points
0
comments
2
min read
LW
link
Human-AI Complementarity: A Goal for Amplified Oversight
rishubjain
and
Sophie Bridgers
Dec 24, 2024, 9:57 AM
27
points
4
comments
1
min read
LW
link
(deepmindsafetyresearch.medium.com)
Preliminary Thoughts on Flirting Theory
Alice Blair
Dec 24, 2024, 7:37 AM
14
points
6
comments
3
min read
LW
link
[Question]
Why is neuron count of human brain relevant to AI timelines?
samuelshadrach
Dec 24, 2024, 5:15 AM
6
points
7
comments
1
min read
LW
link
How Much to Give is a Pragmatic Question
jefftk
Dec 24, 2024, 4:20 AM
12
points
1
comment
2
min read
LW
link
(www.jefftk.com)
Do you need a better map of your myriad of maps to the territory?
CstineSublime
Dec 24, 2024, 2:00 AM
11
points
2
comments
5
min read
LW
link
Panology
JenniferRM
Dec 23, 2024, 9:40 PM
17
points
10
comments
5
min read
LW
link
Aristotle, Aquinas, and the Evolution of Teleology: From Purpose to Meaning.
Spiritus Dei
Dec 23, 2024, 7:37 PM
−9
points
0
comments
6
min read
LW
link
People aren’t properly calibrated on FrontierMath
cakubilo
Dec 23, 2024, 7:35 PM
31
points
4
comments
3
min read
LW
link
Near- and medium-term AI Control Safety Cases
Martín Soto
Dec 23, 2024, 5:37 PM
9
points
0
comments
6
min read
LW
link
[Rationality Malaysia] 2024 year-end meetup!
Doris Liew
Dec 23, 2024, 4:02 PM
1
point
0
comments
1
min read
LW
link
Printable book of some rationalist creative writing (from Scott A. & Eliezer)
CounterBlunder
Dec 23, 2024, 3:44 PM
10
points
0
comments
1
min read
LW
link
Monthly Roundup #25: December 2024
Zvi
Dec 23, 2024, 2:20 PM
18
points
3
comments
26
min read
LW
link
(thezvi.wordpress.com)
Exploring the petertodd / Leilan duality in GPT-2 and GPT-J
mwatkins
Dec 23, 2024, 1:17 PM
12
points
1
comment
17
min read
LW
link
[Question]
What are the strongest arguments for very short timelines?
Kaj_Sotala
Dec 23, 2024, 9:38 AM
101
points
79
comments
1
min read
LW
link
Reduce AI Self-Allegiance by saying “he” instead of “I”
Knight Lee
Dec 23, 2024, 9:32 AM
10
points
4
comments
2
min read
LW
link
Funding Case: AI Safety Camp 11
Remmelt
,
Robert Kralisch
and
Linda Linsefors
Dec 23, 2024, 8:51 AM
60
points
4
comments
6
min read
LW
link
(manifund.org)
What is compute governance?
Vishakha
and
Algon
Dec 23, 2024, 6:32 AM
6
points
0
comments
2
min read
LW
link
(aisafety.info)
Stop Making Sense
JenniferRM
Dec 23, 2024, 5:16 AM
16
points
0
comments
3
min read
LW
link
Hire (or Become) a Thinking Assistant
Raemon
Dec 23, 2024, 3:58 AM
138
points
49
comments
8
min read
LW
link
Non-Obvious Benefits of Insurance
jefftk
Dec 23, 2024, 3:40 AM
21
points
5
comments
2
min read
LW
link
(www.jefftk.com)
Vision of a positive Singularity
RussellThor
Dec 23, 2024, 2:19 AM
4
points
0
comments
4
min read
LW
link
Ideologies are slow and necessary, for now
Gabriel Alfour
Dec 23, 2024, 1:57 AM
15
points
1
comment
1
min read
LW
link
(cognition.cafe)
[Question]
Has Anthropic checked if Claude fakes alignment for intended values too?
Maloew
Dec 23, 2024, 12:43 AM
4
points
1
comment
1
min read
LW
link
Vegans need to eat just enough Meat—emperically evaluate the minimum ammount of meat that maximizes utility
Johannes C. Mayer
Dec 22, 2024, 10:08 PM
55
points
35
comments
3
min read
LW
link
We are in a New Paradigm of AI Progress—OpenAI’s o3 model makes huge gains on the toughest AI benchmarks in the world
garrison
Dec 22, 2024, 9:45 PM
17
points
3
comments
LW
link
(garrisonlovely.substack.com)
My AI timelines
samuelshadrach
Dec 22, 2024, 9:06 PM
12
points
2
comments
5
min read
LW
link
(samuelshadrach.com)
A breakdown of AI capability levels focused on AI R&D labor acceleration
ryan_greenblatt
Dec 22, 2024, 8:56 PM
104
points
6
comments
6
min read
LW
link
How I saved 1 human life (in expectation) without overthinking it
Christopher King
Dec 22, 2024, 8:53 PM
19
points
0
comments
4
min read
LW
link
Towards mutually assured cooperation
mikko
Dec 22, 2024, 8:46 PM
5
points
0
comments
2
min read
LW
link
Checking in on Scott’s composition image bet with imagen 3
Dave Orr
Dec 22, 2024, 7:04 PM
65
points
0
comments
1
min read
LW
link
Woloch & Wosatan
JackOfAllTrades
Dec 22, 2024, 3:46 PM
−11
points
0
comments
2
min read
LW
link
A primer on machine learning in cryo-electron microscopy (cryo-EM)
Abhishaike Mahajan
Dec 22, 2024, 3:11 PM
18
points
0
comments
25
min read
LW
link
(www.owlposting.com)
Notes from Copenhagen Secular Solstice 2024
Søren Elverlin
Dec 22, 2024, 3:08 PM
9
points
0
comments
3
min read
LW
link
Proof Explained for “Robust Agents Learn Causal World Model”
Dalcy
Dec 22, 2024, 3:06 PM
25
points
0
comments
15
min read
LW
link
subfunctional overlaps in attentional selection history implies momentum for decision-trajectories
Emrik
Dec 22, 2024, 2:12 PM
19
points
1
comment
2
min read
LW
link
It looks like there are some good funding opportunities in AI safety right now
Benjamin_Todd
Dec 22, 2024, 12:41 PM
20
points
0
comments
4
min read
LW
link
(benjamintodd.substack.com)
What o3 Becomes by 2028
Vladimir_Nesov
Dec 22, 2024, 12:37 PM
147
points
15
comments
5
min read
LW
link
The Alignment Simulator
Yair Halberstadt
Dec 22, 2024, 11:45 AM
28
points
3
comments
2
min read
LW
link
(yairhalberstadt.github.io)
Theoretical Alignment’s Second Chance
lunatic_at_large
Dec 22, 2024, 5:03 AM
27
points
3
comments
2
min read
LW
link
Orienting to 3 year AGI timelines
Nikola Jurkovic
Dec 22, 2024, 1:15 AM
282
points
51
comments
8
min read
LW
link
ARC-AGI is a genuine AGI test but o3 cheated :(
Knight Lee
Dec 22, 2024, 12:58 AM
3
points
6
comments
2
min read
LW
link
When AI 10x’s AI R&D, What Do We Do?
Logan Riggs
Dec 21, 2024, 11:56 PM
72
points
16
comments
4
min read
LW
link
AI as systems, not just models
Andy Arditi
Dec 21, 2024, 11:19 PM
28
points
0
comments
7
min read
LW
link
(andyrdt.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel