Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
3
A Comment on Expected Utility Theory
DragonGod
Jun 5, 2017, 3:26 AM
0
points
5
comments
4
min read
LW
link
Mathematical System For Calibration
DragonGod
Jun 13, 2017, 12:01 PM
0
points
8
comments
1
min read
LW
link
Humans are not agents: short vs long term
Stuart_Armstrong
Jun 27, 2017, 1:04 PM
0
points
5
comments
1
min read
LW
link
New circumstances, new values?
Stuart_Armstrong
Jun 6, 2017, 8:18 AM
0
points
0
comments
1
min read
LW
link
Optimisation in manipulating humans: engineered fanatics vs yes-men
Stuart_Armstrong
Jun 2, 2017, 3:51 PM
0
points
0
comments
2
min read
LW
link
Acausal trade: trade barriers
Stuart_Armstrong
Jun 2, 2017, 3:32 PM
0
points
1
comment
2
min read
LW
link
Agents that don’t become maximisers
Stuart_Armstrong
Jun 2, 2017, 4:13 PM
0
points
0
comments
3
min read
LW
link
Understanding the important facts
Stuart_Armstrong
Jun 2, 2017, 4:15 PM
0
points
0
comments
1
min read
LW
link
Translation “counterfactual”
Stuart_Armstrong
Jun 2, 2017, 4:16 PM
0
points
0
comments
2
min read
LW
link
Indifference and compensatory rewards
Stuart_Armstrong
Jun 2, 2017, 4:19 PM
0
points
0
comments
1
min read
LW
link
The radioactive burrito and learning from positive examples
Stuart_Armstrong
Jun 2, 2017, 4:25 PM
0
points
2
comments
1
min read
LW
link
Corrigibility thoughts III: manipulating versus deceiving
Stuart_Armstrong
Jun 2, 2017, 4:27 PM
0
points
0
comments
1
min read
LW
link
Corrigibility thoughts II: the robot operator
Stuart_Armstrong
Jun 2, 2017, 4:27 PM
0
points
12
comments
2
min read
LW
link
Ontology, lost purposes, and instrumental goals
Stuart_Armstrong
Jun 2, 2017, 4:28 PM
0
points
1
comment
1
min read
LW
link
How to judge moral learning failure
Stuart_Armstrong
Jun 2, 2017, 4:32 PM
0
points
2
comments
2
min read
LW
link
The best value indifference method (so far)
Stuart_Armstrong
Jun 2, 2017, 4:33 PM
0
points
9
comments
5
min read
LW
link
Reward/value learning for reinforcement learning
Stuart_Armstrong
Jun 2, 2017, 4:34 PM
0
points
2
comments
2
min read
LW
link
Announcing AASAA—Accelerating AI Safety Adoption in Academia (and elsewhere)
null
Jun 15, 2017, 6:50 PM
0
points
0
comments
2
min read
LW
link
Am I/Was I a Cultist?
DragonGod
Jun 12, 2017, 11:21 PM
−1
points
12
comments
2
min read
LW
link
Back to first
Previous
Back to top