Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Verification and Transparency
DanielFilan
Aug 8, 2019, 1:50 AM
35
points
6
comments
2
min read
LW
link
(danielfilan.com)
AI Alignment Open Thread August 2019
habryka
Aug 4, 2019, 10:09 PM
35
points
96
comments
1
min read
LW
link
AI Forecasting Resolution Council (Forecasting infrastructure, part 2)
Bird Concept
and
Ben Goldhaber
Aug 29, 2019, 5:35 PM
35
points
2
comments
3
min read
LW
link
[Question]
What authors consistently give accurate pictures of complex topics they discuss?
seez
Aug 21, 2019, 12:09 AM
34
points
3
comments
1
min read
LW
link
[Site Update] Weekly/Monthly/Yearly on All Posts
Raemon
Aug 2, 2019, 12:39 AM
33
points
7
comments
1
min read
LW
link
“Can We Survive Technology” by von Neumann
Ben Pace
Aug 18, 2019, 6:58 PM
33
points
2
comments
1
min read
LW
link
(geosci.uchicago.edu)
[Question]
What experiments would demonstrate “upper limits of augmented working memory?”
Raemon
Aug 15, 2019, 10:09 PM
33
points
6
comments
2
min read
LW
link
AI Alignment Writing Day Roundup #1
Ben Pace
Aug 30, 2019, 1:26 AM
32
points
12
comments
1
min read
LW
link
Calibrating With Cards
lifelonglearner
Aug 8, 2019, 6:44 AM
32
points
3
comments
3
min read
LW
link
Distance Functions are Hard
Grue_Slinky
Aug 13, 2019, 5:33 PM
31
points
19
comments
6
min read
LW
link
Don’t Pull a Broken Chain
johnswentworth
Aug 28, 2019, 1:21 AM
31
points
6
comments
5
min read
LW
link
[Question]
What explanatory power does Kahneman’s System 2 possess?
Richard_Ngo
Aug 12, 2019, 3:23 PM
31
points
2
comments
1
min read
LW
link
When do utility functions constrain?
Hoagy
Aug 23, 2019, 5:19 PM
30
points
8
comments
7
min read
LW
link
Self-Supervised Learning and AGI Safety
Steven Byrnes
Aug 7, 2019, 2:21 PM
30
points
9
comments
12
min read
LW
link
Help forecast study replication in this social science prediction market
rosiecam
Aug 7, 2019, 6:18 PM
29
points
3
comments
1
min read
LW
link
A Survey of Early Impact Measures
Matthew Barnett
Aug 6, 2019, 1:22 AM
29
points
0
comments
8
min read
LW
link
[Question]
Could we solve this email mess if we all moved to paid emails?
Bird Concept
Aug 11, 2019, 4:31 PM
29
points
50
comments
4
min read
LW
link
Inspection Paradox as a Driver of Group Separation
Shmi
Aug 17, 2019, 9:47 PM
29
points
0
comments
1
min read
LW
link
[Question]
What are the reasons to *not* consider reducing AI-Xrisk the highest priority cause?
David Scott Krueger (formerly: capybaralet)
Aug 20, 2019, 9:45 PM
29
points
27
comments
1
min read
LW
link
Predicted AI alignment event/meeting calendar
rmoehn
Aug 14, 2019, 7:14 AM
29
points
14
comments
1
min read
LW
link
Announcement: Writing Day Today (Thursday)
Ben Pace
Aug 22, 2019, 4:48 AM
29
points
5
comments
1
min read
LW
link
GPT-2: 6-Month Follow-Up
lifelonglearner
Aug 21, 2019, 5:06 AM
28
points
1
comment
1
min read
LW
link
“Designing agent incentives to avoid reward tampering”, DeepMind
gwern
Aug 14, 2019, 4:57 PM
28
points
15
comments
LW
link
(medium.com)
[AN #62] Are adversarial examples caused by real but imperceptible features?
Rohin Shah
Aug 22, 2019, 5:10 PM
28
points
10
comments
9
min read
LW
link
(mailchi.mp)
Algorithmic Similarity
LukasM
Aug 23, 2019, 4:39 PM
28
points
10
comments
11
min read
LW
link
[Question]
What is the state of the ego depletion field?
Eli Tyre
Aug 9, 2019, 8:30 PM
27
points
10
comments
1
min read
LW
link
[Question]
Why are the people who could be doing safety research, but aren’t, doing something else?
Adam Scholl
Aug 29, 2019, 8:51 AM
27
points
19
comments
1
min read
LW
link
Raph Koster on Virtual Worlds vs Games (notes)
Raemon
Aug 18, 2019, 7:01 PM
26
points
8
comments
2
min read
LW
link
Reversible changes: consider a bucket of water
Stuart_Armstrong
Aug 26, 2019, 10:55 PM
25
points
18
comments
2
min read
LW
link
Project Proposal: Considerations for trading off capabilities and safety impacts of AI research
David Scott Krueger (formerly: capybaralet)
Aug 6, 2019, 10:22 PM
25
points
11
comments
2
min read
LW
link
Inversion of theorems into definitions when generalizing
riceissa
Aug 4, 2019, 5:44 PM
25
points
3
comments
5
min read
LW
link
Goodhart’s Curse and Limitations on AI Alignment
Gordon Seidoh Worley
Aug 19, 2019, 7:57 AM
25
points
18
comments
10
min read
LW
link
Why Gradients Vanish and Explode
Matthew Barnett
Aug 9, 2019, 2:54 AM
25
points
9
comments
3
min read
LW
link
Which of these five AI alignment research projects ideas are no good?
rmoehn
Aug 8, 2019, 7:17 AM
25
points
13
comments
1
min read
LW
link
[Question]
Why do humans not have built-in neural i/o channels?
Richard_Ngo
Aug 8, 2019, 1:09 PM
25
points
23
comments
1
min read
LW
link
Negative “eeny meeny miny moe”
jefftk
Aug 20, 2019, 2:48 AM
25
points
6
comments
1
min read
LW
link
A Primer on Matrix Calculus, Part 1: Basic review
Matthew Barnett
Aug 12, 2019, 11:44 PM
25
points
4
comments
7
min read
LW
link
Emotions are not beliefs
Chris_Leong
Aug 7, 2019, 6:27 AM
25
points
2
comments
2
min read
LW
link
Implications of Quantum Computing for Artificial Intelligence Alignment Research
Jsevillamol
and
PabloAMC
Aug 22, 2019, 10:33 AM
24
points
3
comments
13
min read
LW
link
Understanding understanding
mthq
Aug 23, 2019, 6:10 PM
24
points
1
comment
2
min read
LW
link
July 2019 gwern.net newsletter
gwern
Aug 1, 2019, 4:19 PM
23
points
0
comments
LW
link
(www.gwern.net)
[Site Update] Behind the scenes data-layer and caching improvements
habryka
Aug 7, 2019, 12:49 AM
23
points
3
comments
1
min read
LW
link
Cartographic Processes
johnswentworth
Aug 27, 2019, 8:02 PM
23
points
3
comments
4
min read
LW
link
[Question]
Do you do weekly or daily reviews? What are they like?
benwr
Aug 5, 2019, 1:23 AM
23
points
8
comments
1
min read
LW
link
Practical consequences of impossibility of value learning
Stuart_Armstrong
2 Aug 2019 23:06 UTC
23
points
13
comments
3
min read
LW
link
A Primer on Matrix Calculus, Part 2: Jacobians and other fun
Matthew Barnett
15 Aug 2019 1:13 UTC
22
points
7
comments
7
min read
LW
link
In defense of Oracle (“Tool”) AI research
Steven Byrnes
7 Aug 2019 19:14 UTC
22
points
11
comments
4
min read
LW
link
Four Ways An Impact Measure Could Help Alignment
Matthew Barnett
8 Aug 2019 0:10 UTC
21
points
1
comment
9
min read
LW
link
[Question]
Is LW making progress?
zulupineapple
24 Aug 2019 0:32 UTC
21
points
11
comments
1
min read
LW
link
Problems with AI debate
Stuart_Armstrong
26 Aug 2019 19:21 UTC
21
points
3
comments
5
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel