Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Latacora might be of interest to some AI Safety organizations
NunoSempere
Nov 25, 2021, 11:57 PM
14
points
10
comments
1
min read
LW
link
(www.latacora.com)
Christiano, Cotra, and Yudkowsky on AI progress
Eliezer Yudkowsky
and
Ajeya Cotra
Nov 25, 2021, 4:45 PM
119
points
95
comments
66
min read
LW
link
Covid 11/25: Another Thanksgiving
Zvi
Nov 25, 2021, 1:40 PM
73
points
9
comments
21
min read
LW
link
(thezvi.wordpress.com)
Coordinating the Unequal Treaties
lsusr
Nov 25, 2021, 10:47 AM
34
points
4
comments
2
min read
LW
link
First Strike and Second Strike
lsusr
Nov 25, 2021, 9:23 AM
28
points
5
comments
1
min read
LW
link
You are way more fallible than you think
Shmi
Nov 25, 2021, 5:52 AM
4
points
14
comments
2
min read
LW
link
[Linkpost] Danger of motivatiogenesis in interdisciplinary work
particlemania
Nov 25, 2021, 12:13 AM
9
points
0
comments
1
min read
LW
link
Meetup for The Roots of Progress in San Diego, Dec 1
jasoncrawford
Nov 24, 2021, 10:50 PM
7
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
Base Rates and Reference Classes
jsteinhardt
Nov 24, 2021, 10:30 PM
20
points
7
comments
5
min read
LW
link
(bounded-regret.ghost.io)
Why do you need the story?
George3d6
Nov 24, 2021, 8:26 PM
52
points
11
comments
5
min read
LW
link
(cerebralab.com)
[AN #169]: Collaborating with humans without human data
Rohin Shah
Nov 24, 2021, 6:30 PM
33
points
0
comments
8
min read
LW
link
(mailchi.mp)
Paxlovid Remains Illegal: 11/24 Update
Zvi
Nov 24, 2021, 1:40 PM
54
points
21
comments
7
min read
LW
link
(thezvi.wordpress.com)
HIRING: Inform and shape a new project on AI safety at Partnership on AI
Madhulika Srikumar
Nov 24, 2021, 8:27 AM
6
points
0
comments
1
min read
LW
link
[Question]
How much Bayesian evidence from rapid antigen and PCR tests?
mingyuan
Nov 24, 2021, 6:54 AM
8
points
4
comments
1
min read
LW
link
French long COVID study: Belief vs Infection
Bucky
Nov 23, 2021, 11:14 PM
40
points
11
comments
5
min read
LW
link
[Question]
Cornell Meetup
Lionel Levine
Nov 23, 2021, 9:28 PM
6
points
4
comments
1
min read
LW
link
AI Tracker: monitoring current and near-future risks from superscale models
Edouard Harris
and
Jeremie Harris
Nov 23, 2021, 7:16 PM
67
points
13
comments
3
min read
LW
link
(aitracker.org)
Laplace’s rule of succession
Ege Erdil
Nov 23, 2021, 3:48 PM
52
points
2
comments
7
min read
LW
link
AI Safety Needs Great Engineers
Andy Jones
Nov 23, 2021, 3:40 PM
90
points
43
comments
4
min read
LW
link
Slightly advanced decision theory 102: Four reasons not to be a (naive) utility maximizer
Jan
Nov 23, 2021, 11:02 AM
10
points
1
comment
15
min read
LW
link
(universalprior.substack.com)
Use Tools For What They’re For
DirectedEvolution
Nov 23, 2021, 8:26 AM
28
points
14
comments
8
min read
LW
link
[linkpost] Acquisition of Chess Knowledge in AlphaZero
Quintin Pope
Nov 23, 2021, 7:55 AM
8
points
1
comment
1
min read
LW
link
[linkpost] Why Going to the Doctor Sucks (WaitButWhy)
mike_hawke
Nov 23, 2021, 3:02 AM
5
points
11
comments
1
min read
LW
link
(waitbutwhy.com)
Integrating Three Models of (Human) Cognition
jbkjr
Nov 23, 2021, 1:06 AM
40
points
4
comments
32
min read
LW
link
Potential Alignment mental tool: Keeping track of the types
Donald Hobson
Nov 22, 2021, 8:05 PM
29
points
1
comment
2
min read
LW
link
Yudkowsky and Christiano discuss “Takeoff Speeds”
Eliezer Yudkowsky
Nov 22, 2021, 7:35 PM
210
points
176
comments
60
min read
LW
link
1
review
Morally underdefined situations can be deadly
Stuart_Armstrong
Nov 22, 2021, 2:48 PM
17
points
8
comments
2
min read
LW
link
A Bayesian Aggregation Paradox
Jsevillamol
Nov 22, 2021, 10:39 AM
87
points
23
comments
7
min read
LW
link
[Question]
Do factored sets elucidate anything about how to update everyday beliefs?
TekhneMakre
Nov 22, 2021, 6:51 AM
5
points
1
comment
1
min read
LW
link
Even if you’re right, you’re wrong
DanielFilan
Nov 22, 2021, 5:40 AM
17
points
5
comments
1
min read
LW
link
(danielfilan.com)
The Meta-Puzzle
DanielFilan
Nov 22, 2021, 5:30 AM
23
points
27
comments
3
min read
LW
link
(danielfilan.com)
Some real examples of gradient hacking
Oliver Sourbut
Nov 22, 2021, 12:11 AM
15
points
8
comments
2
min read
LW
link
“The Wisdom of the Lazy Teacher”
Richard_Kennaway
Nov 21, 2021, 9:11 PM
16
points
5
comments
1
min read
LW
link
Vitalik: Cryptoeconomics and X-Risk Researchers Should Listen to Each Other More
Emerson Spartz
Nov 21, 2021, 6:53 PM
47
points
9
comments
5
min read
LW
link
Giving Up On T-Mobile
jefftk
Nov 21, 2021, 4:50 PM
13
points
6
comments
2
min read
LW
link
(www.jefftk.com)
From language to ethics by automated reasoning
Michele Campolo
Nov 21, 2021, 3:16 PM
4
points
4
comments
6
min read
LW
link
Split and Commit
Duncan Sabien (Inactive)
Nov 21, 2021, 6:27 AM
191
points
34
comments
7
min read
LW
link
1
review
What’s the weirdest way to win this game?
Adam Scherlis
Nov 21, 2021, 5:18 AM
9
points
5
comments
1
min read
LW
link
(adam.scherlis.com)
Eat the cute animals instead
Andrew Vlahos
Nov 21, 2021, 1:06 AM
−4
points
2
comments
1
min read
LW
link
Chris Voss negotiation MasterClass: review
VipulNaik
Nov 20, 2021, 10:39 PM
70
points
15
comments
33
min read
LW
link
ACX Montreal Meetup Dec 4 2021
E
Nov 20, 2021, 5:49 PM
8
points
0
comments
1
min read
LW
link
The Maker of MIND
Tomás B.
Nov 20, 2021, 4:28 PM
112
points
19
comments
11
min read
LW
link
South Bay ACX/LW Meetup—CHANGED LOCATION
IS
Nov 20, 2021, 2:42 PM
11
points
0
comments
1
min read
LW
link
The Emperor’s New Clothes: a story of motivated stupidity
David Hugh-Jones
20 Nov 2021 13:24 UTC
10
points
5
comments
3
min read
LW
link
(wyclif.substack.com)
[Book Review] “Sorceror’s Apprentice” by Tahir Shah
lsusr
20 Nov 2021 11:29 UTC
92
points
11
comments
7
min read
LW
link
Competence/Confidence
Duncan Sabien (Inactive)
20 Nov 2021 8:59 UTC
60
points
19
comments
1
min read
LW
link
Awesome-github Post-Scarcity List
lorepieri
20 Nov 2021 8:47 UTC
3
points
6
comments
1
min read
LW
link
A Certain Formalization of Corrigibility Is VNM-Incoherent
TurnTrout
20 Nov 2021 0:30 UTC
68
points
24
comments
8
min read
LW
link
More detailed proposal for measuring alignment of current models
Beth Barnes
20 Nov 2021 0:03 UTC
31
points
0
comments
8
min read
LW
link
Ambitious Altruistic Software Engineering Efforts: Opportunities and Benefits
ozziegooen
19 Nov 2021 17:55 UTC
42
points
1
comment
9
min read
LW
link
(forum.effectivealtruism.org)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel