Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
[Question]
If alignment problem was unsolvable, would that avoid doom?
Kinrany
May 7, 2023, 10:13 PM
3
points
3
comments
1
min read
LW
link
An artificially structured argument for expecting AGI ruin
Rob Bensinger
May 7, 2023, 9:52 PM
91
points
26
comments
19
min read
LW
link
Where “the Sequences” Are Wrong
Thoth Hermes
May 7, 2023, 8:21 PM
−15
points
5
comments
14
min read
LW
link
(thothhermes.substack.com)
What’s wrong with being dumb?
Adam Zerner
May 7, 2023, 6:31 PM
14
points
17
comments
2
min read
LW
link
Categories of Arguing Style : Why being good among rationalists isn’t enough to argue with everyone
Camille Berger
May 7, 2023, 5:45 PM
16
points
0
comments
23
min read
LW
link
Self-Administered Gell-Mann Amnesia
krs
May 7, 2023, 5:44 PM
1
point
1
comment
1
min read
LW
link
Understanding mesa-optimization using toy models
tilmanr
,
rusheb
,
Guillaume Corlouer
,
Dan Valentine
,
afspies
,
mivanitskiy
and
Can
May 7, 2023, 5:00 PM
45
points
2
comments
10
min read
LW
link
How to have Polygenically Screened Children
GeneSmith
May 7, 2023, 4:01 PM
367
points
128
comments
27
min read
LW
link
1
review
Statistical models & the irrelevance of rare exceptions
patrissimo
May 7, 2023, 3:59 PM
36
points
6
comments
2
min read
LW
link
Let’s look for coherence theorems
Valdes
May 7, 2023, 2:45 PM
25
points
18
comments
6
min read
LW
link
Graphical Representations of Paul Christiano’s Doom Model
Nathan Young
May 7, 2023, 1:03 PM
7
points
0
comments
LW
link
An anthropomorphic AI dilemma
TsviBT
May 7, 2023, 12:44 PM
26
points
0
comments
7
min read
LW
link
Violin Supports
jefftk
May 7, 2023, 12:10 PM
12
points
1
comment
1
min read
LW
link
(www.jefftk.com)
Properties of Good Textbooks
niplav
May 7, 2023, 8:38 AM
50
points
11
comments
1
min read
LW
link
Against sacrificing AI transparency for generality gains
Ape in the coat
May 7, 2023, 6:52 AM
4
points
0
comments
2
min read
LW
link
TED talk by Eliezer Yudkowsky: Unleashing the Power of Artificial Intelligence
bayesed
May 7, 2023, 5:45 AM
49
points
36
comments
1
min read
LW
link
(www.youtube.com)
Thinking of Convenience as an Economic Term
ozziegooen
May 7, 2023, 1:21 AM
6
points
0
comments
12
min read
LW
link
(forum.effectivealtruism.org)
Corrigibility, Much more detail than anyone wants to Read
Logan Zoellner
May 7, 2023, 1:02 AM
26
points
2
comments
7
min read
LW
link
Residual stream norms grow exponentially over the forward pass
StefanHex
and
TurnTrout
May 7, 2023, 12:46 AM
77
points
24
comments
11
min read
LW
link
On the Loebner Silver Prize (a Turing test)
hold_my_fish
May 7, 2023, 12:39 AM
18
points
2
comments
2
min read
LW
link
Time and Energy Costs to Erase a Bit
DaemonicSigil
May 6, 2023, 11:29 PM
24
points
32
comments
7
min read
LW
link
How much do you believe your results?
Eric Neyman
May 6, 2023, 8:31 PM
507
points
18
comments
15
min read
LW
link
4
reviews
(ericneyman.wordpress.com)
Long Covid Risks: 2023 Update
Elizabeth
May 6, 2023, 6:20 PM
69
points
11
comments
4
min read
LW
link
(acesounderglass.com)
Is “red” for GPT-4 the same as “red” for you?
Yusuke Hayashi
May 6, 2023, 5:55 PM
9
points
6
comments
2
min read
LW
link
The Broader Fossil Fuel Community
Jeffrey Heninger
May 6, 2023, 2:49 PM
16
points
1
comment
3
min read
LW
link
Estimating Norovirus Prevalence
jefftk
May 6, 2023, 11:40 AM
16
points
0
comments
2
min read
LW
link
(www.jefftk.com)
Alignment as Function Fitting
A.H.
May 6, 2023, 11:38 AM
7
points
0
comments
12
min read
LW
link
My preferred framings for reward misspecification and goal misgeneralisation
Yi-Yang
May 6, 2023, 4:48 AM
27
points
1
comment
8
min read
LW
link
You don’t need to be a genius to be in AI safety research
Claire Short
May 6, 2023, 2:32 AM
14
points
1
comment
6
min read
LW
link
Naturalist Collection
LoganStrohl
May 6, 2023, 12:37 AM
71
points
7
comments
15
min read
LW
link
Do you work at an AI lab? Please quit
Nik Samoylov
May 5, 2023, 11:41 PM
−29
points
9
comments
1
min read
LW
link
Explaining “Hell is Game Theory Folk Theorems”
electroswing
May 5, 2023, 11:33 PM
57
points
21
comments
5
min read
LW
link
Sleeping Beauty – the Death Hypothesis
Guillaume Charrier
May 5, 2023, 11:32 PM
7
points
8
comments
5
min read
LW
link
Orthogonal’s Formal-Goal Alignment theory of change
Tamsin Leake
May 5, 2023, 10:36 PM
69
points
13
comments
4
min read
LW
link
(carado.moe)
A smart enough LLM might be deadly simply if you run it for long enough
Mikhail Samin
May 5, 2023, 8:49 PM
19
points
16
comments
8
min read
LW
link
What Jason has been reading, May 2023: “Protopia,” complex systems, Daedalus vs. Icarus, and more
jasoncrawford
May 5, 2023, 7:54 PM
25
points
2
comments
11
min read
LW
link
(rootsofprogress.org)
CHAT Diplomacy: LLMs and National Security
SebastianG
May 5, 2023, 7:45 PM
25
points
6
comments
7
min read
LW
link
Linkpost for Accursed Farms Discussion / debate with AI expert Eliezer Yudkowsky
gilch
May 5, 2023, 6:20 PM
14
points
2
comments
1
min read
LW
link
(www.youtube.com)
Regulate or Compete? The China Factor in U.S. AI Policy (NAIR #2)
charles_m
May 5, 2023, 5:43 PM
2
points
1
comment
7
min read
LW
link
(navigatingairisks.substack.com)
Kingfisher Live CD Process
jefftk
May 5, 2023, 5:00 PM
13
points
0
comments
3
min read
LW
link
(www.jefftk.com)
What can we learn from Bayes about reasoning?
jasoncrawford
May 5, 2023, 3:52 PM
22
points
11
comments
1
min read
LW
link
[Question]
Why not use active SETI to prevent AI Doom?
RomanS
May 5, 2023, 2:41 PM
13
points
13
comments
1
min read
LW
link
Investigating Emergent Goal-Like Behavior in Large Language Models using Experimental Economics
phelps-sg
5 May 2023 11:15 UTC
6
points
1
comment
4
min read
LW
link
Monthly Shorts 4/23
Celer
5 May 2023 7:20 UTC
8
points
1
comment
3
min read
LW
link
(keller.substack.com)
[Question]
What is it like to be a compatibilist?
tslarm
5 May 2023 2:56 UTC
8
points
72
comments
1
min read
LW
link
Transcript of a presentation on catastrophic risks from AI
RobertM
5 May 2023 1:38 UTC
6
points
0
comments
8
min read
LW
link
How to get good at programming
Ulisse Mini
5 May 2023 1:14 UTC
39
points
3
comments
2
min read
LW
link
An Update On The Campaign For AI Safety Dot Org
yanni kyriacos
5 May 2023 0:21 UTC
−13
points
2
comments
1
min read
LW
link
A brief collection of Hinton’s recent comments on AGI risk
Kaj_Sotala
4 May 2023 23:31 UTC
143
points
9
comments
11
min read
LW
link
Robin Hanson and I talk about AI risk
KatjaGrace
4 May 2023 22:20 UTC
39
points
8
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel