Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
Page
1
South Bay Meetup
DavidFriedman
Feb 10, 2020, 10:36 PM
4
points
0
comments
LW
link
Simulation of technological progress (work in progress)
Daniel Kokotajlo
Feb 10, 2020, 8:39 PM
21
points
9
comments
5
min read
LW
link
[Question]
Why do we refuse to take action claiming our impact would be too small?
hookdump
Feb 10, 2020, 7:33 PM
5
points
31
comments
1
min read
LW
link
Gricean communication and meta-preferences
Charlie Steiner
Feb 10, 2020, 5:05 AM
24
points
0
comments
3
min read
LW
link
Attainable Utility Landscape: How The World Is Changed
TurnTrout
Feb 10, 2020, 12:58 AM
52
points
7
comments
6
min read
LW
link
A Simple Introduction to Neural Networks
Rafael Harth
Feb 9, 2020, 10:02 PM
34
points
13
comments
18
min read
LW
link
[Question]
Did AI pioneers not worry much about AI risks?
lisperati
Feb 9, 2020, 7:58 PM
42
points
9
comments
1
min read
LW
link
[Question]
Source of Karma
jmh
Feb 9, 2020, 2:13 PM
4
points
14
comments
1
min read
LW
link
State Space of X-Risk Trajectories
David_Kristoffersson
Feb 9, 2020, 1:56 PM
11
points
0
comments
9
min read
LW
link
[Question]
Does there exist an AGI-level parameter setting for modern DRL architectures?
TurnTrout
Feb 9, 2020, 5:09 AM
15
points
3
comments
1
min read
LW
link
[Question]
Who… (or what) designed this site and where did they come from?
thedayismine
Feb 9, 2020, 4:04 AM
12
points
3
comments
1
min read
LW
link
How to Frame Negative Feedback as Forward-Facing Guidance
Liron
Feb 9, 2020, 2:47 AM
46
points
7
comments
3
min read
LW
link
Relationship Outcomes Are Not Particularly Sensitive to Small Variations in Verbal Ability
Zack_M_Davis
Feb 9, 2020, 12:34 AM
14
points
2
comments
1
min read
LW
link
(zackmdavis.net)
What can the principal-agent literature tell us about AI risk?
apc
Feb 8, 2020, 9:28 PM
104
points
29
comments
16
min read
LW
link
A Cautionary Note on Unlocking the Emotional Brain
eapache
Feb 8, 2020, 5:21 PM
55
points
20
comments
2
min read
LW
link
[Question]
What is this review feature?
Long try
Feb 8, 2020, 3:30 PM
1
point
1
comment
1
min read
LW
link
Halifax SSC Meetup—FEB 8
interstice
Feb 8, 2020, 12:45 AM
4
points
0
comments
1
min read
LW
link
On the falsifiability of hypercomputation
jessicata
Feb 7, 2020, 8:16 AM
24
points
4
comments
4
min read
LW
link
(unstableontology.com)
More writeups!
jefftk
Feb 7, 2020, 3:10 AM
40
points
5
comments
1
min read
LW
link
(www.jefftk.com)
Book Review: Decisive by Chip and Dan Heath
Ian David Moss
Feb 6, 2020, 8:15 PM
4
points
0
comments
2
min read
LW
link
(medium.com)
Bayes-Up: An App for Sharing Bayesian-MCQ
Louis Faucon
Feb 6, 2020, 7:01 PM
53
points
9
comments
1
min read
LW
link
Mazes Sequence Roundup: Final Thoughts and Paths Forward
Zvi
Feb 6, 2020, 4:10 PM
88
points
28
comments
14
min read
LW
link
1
review
(thezvi.wordpress.com)
Plausibly, almost every powerful algorithm would be manipulative
Stuart_Armstrong
Feb 6, 2020, 11:50 AM
38
points
25
comments
3
min read
LW
link
Some quick notes on hand hygiene
willbradshaw
Feb 6, 2020, 2:47 AM
68
points
52
comments
3
min read
LW
link
Potential Research Topic: Vingean Reflection, Value Alignment and Aspiration
Vaughn Papenhausen
Feb 6, 2020, 1:09 AM
15
points
4
comments
4
min read
LW
link
Synthesizing amplification and debate
evhub
Feb 5, 2020, 10:53 PM
33
points
10
comments
4
min read
LW
link
Writeup: Progress on AI Safety via Debate
Beth Barnes
and
paulfchristiano
Feb 5, 2020, 9:04 PM
103
points
18
comments
33
min read
LW
link
[AN #85]: The normative questions we should be asking for AI alignment, and a surprisingly good chatbot
Rohin Shah
Feb 5, 2020, 6:20 PM
14
points
2
comments
7
min read
LW
link
(mailchi.mp)
The Adventure: a new Utopia story
Stuart_Armstrong
Feb 5, 2020, 4:50 PM
101
points
37
comments
51
min read
LW
link
“But that’s your job”: why organisations can work
Stuart_Armstrong
Feb 5, 2020, 12:25 PM
77
points
12
comments
4
min read
LW
link
Training a tiny SupAmp model on easy tasks. The influence of failure rate on learning curves
rmoehn
Feb 5, 2020, 7:22 AM
5
points
0
comments
1
min read
LW
link
Physical alignment—do you have it? Take a minute & check.
leggi
Feb 5, 2020, 4:02 AM
4
points
4
comments
1
min read
LW
link
Open & Welcome Thread—February 2020
ryan_b
Feb 4, 2020, 8:49 PM
17
points
114
comments
1
min read
LW
link
Meta-Preference Utilitarianism
B Jacobs
Feb 4, 2020, 8:24 PM
10
points
30
comments
1
min read
LW
link
Philosophical self-ratification
jessicata
Feb 3, 2020, 10:48 PM
23
points
13
comments
5
min read
LW
link
(unstableontology.com)
Twenty-three AI alignment research project definitions
rmoehn
Feb 3, 2020, 10:21 PM
23
points
0
comments
6
min read
LW
link
Absent coordination, future technology will cause human extinction
Jeffrey Ladish
Feb 3, 2020, 9:52 PM
21
points
12
comments
5
min read
LW
link
Long Now, and Culture vs Artifacts
Raemon
Feb 3, 2020, 9:49 PM
26
points
3
comments
6
min read
LW
link
[Question]
Looking for books about software engineering as a field
mingyuan
Feb 3, 2020, 9:49 PM
14
points
15
comments
1
min read
LW
link
Category Theory Without The Baggage
johnswentworth
Feb 3, 2020, 8:03 PM
139
points
51
comments
13
min read
LW
link
Protecting Large Projects Against Mazedom
Zvi
Feb 3, 2020, 5:10 PM
78
points
11
comments
4
min read
LW
link
1
review
(thezvi.wordpress.com)
Pessimism About Unknown Unknowns Inspires Conservatism
michaelcohen
Feb 3, 2020, 2:48 PM
41
points
2
comments
5
min read
LW
link
Map Of Effective Altruism
Scott Alexander
Feb 3, 2020, 6:20 AM
17
points
1
comment
1
min read
LW
link
(slatestarcodex.com)
UML IX: Kernels and Boosting
Rafael Harth
Feb 2, 2020, 9:51 PM
13
points
1
comment
10
min read
LW
link
A point of clarification on infohazard terminology
eukaryote
Feb 2, 2020, 5:43 PM
52
points
21
comments
2
min read
LW
link
(eukaryotewritesblog.com)
[Question]
Money isn’t real. When you donate money to a charity, how does it actually help?
Dagon
Feb 2, 2020, 5:03 PM
15
points
28
comments
1
min read
LW
link
[Link] Beyond the hill: thoughts on ontologies for thinking, essay-completeness and forecasting
Bird Concept
Feb 2, 2020, 12:39 PM
33
points
6
comments
1
min read
LW
link
The Case for Artificial Expert Intelligence (AXI): What lies between narrow and general AI?
Yuli_Ban
Feb 2, 2020, 5:55 AM
8
points
2
comments
6
min read
LW
link
“Memento Mori”, Said The Confessor
namespace
Feb 2, 2020, 3:37 AM
34
points
4
comments
1
min read
LW
link
(www.thelastrationalist.com)
Bay Winter Solstice seating-scarcity
Raemon
Feb 1, 2020, 11:09 PM
2
points
3
comments
2
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel