Archive
Sequences
About
Search
Log In
Home
Featured
All
Tags
Recent
Comments
Questions
Events
Shortform
Alignment Forum
AF Comments
Hide coronavirus posts
RSS
Page
1
Infra-Bayesianism Unwrapped
adamShimi
20 Jan 2021 13:35 UTC
12
points
0
comments
24
min read
LW
link
Against the Backward Approach to Goal-Directedness
adamShimi
19 Jan 2021 18:46 UTC
13
points
5
comments
4
min read
LW
link
Some thoughts on risks from narrow, non-agentic AI
Richard_Ngo
19 Jan 2021 0:04 UTC
25
points
10
comments
16
min read
LW
link
Short summary of mAIry’s room
Stuart_Armstrong
18 Jan 2021 18:11 UTC
24
points
2
comments
4
min read
LW
link
Birds, Brains, Planes, and AI: Against Appeals to the Complexity/Mysteriousness/Efficiency of the Brain
Daniel Kokotajlo
18 Jan 2021 12:08 UTC
100
points
14
comments
14
min read
LW
link
Literature Review on Goal-Directedness
adamShimi
,
Michele Campolo
and
Joe_Collman
18 Jan 2021 11:15 UTC
50
points
15
comments
31
min read
LW
link
Excerpt from Arbital Solomonoff induction dialogue
Richard_Ngo
17 Jan 2021 3:49 UTC
27
points
3
comments
5
min read
LW
link
(arbital.com)
Why I’m excited about Debate
Richard_Ngo
15 Jan 2021 23:37 UTC
62
points
12
comments
7
min read
LW
link
Thoughts on Iason Gabriel’s Artificial Intelligence, Values, and Alignment
alexflint
14 Jan 2021 12:58 UTC
36
points
14
comments
4
min read
LW
link
Some recent survey papers on (mostly near-term) AI safety, security, and assurance
alenglander
13 Jan 2021 21:50 UTC
11
points
0
comments
3
min read
LW
link
[AN #133]: Building machines that can cooperate (with humans, institutions, or other machines)
rohinmshah
13 Jan 2021 18:10 UTC
14
points
0
comments
9
min read
LW
link
(mailchi.mp)
Review of ‘Debate on Instrumental Convergence between LeCun, Russell, Bengio, Zador, and More’
TurnTrout
12 Jan 2021 3:57 UTC
37
points
1
comment
2
min read
LW
link
Transparency and AGI safety
jylin04
11 Jan 2021 18:51 UTC
49
points
12
comments
30
min read
LW
link
Prediction can be Outer Aligned at Optimum
Lanrian
10 Jan 2021 18:48 UTC
13
points
11
comments
11
min read
LW
link
Review of Soft Takeoff Can Still Lead to DSA
Daniel Kokotajlo
10 Jan 2021 18:10 UTC
62
points
10
comments
6
min read
LW
link
Imitative Generalisation (AKA ‘Learning the Prior’)
Beth Barnes
10 Jan 2021 0:30 UTC
59
points
7
comments
12
min read
LW
link
The Case for a Journal of AI Alignment
adamShimi
9 Jan 2021 18:13 UTC
38
points
27
comments
4
min read
LW
link
Eight claims about multi-agent AGI safety
Richard_Ngo
7 Jan 2021 13:34 UTC
68
points
17
comments
4
min read
LW
link
Review of ‘But exactly how complex and fragile?’
TurnTrout
6 Jan 2021 18:39 UTC
48
points
1
comment
8
min read
LW
link
[AN #132]: Complex and subtly incorrect arguments as an obstacle to debate
rohinmshah
6 Jan 2021 18:20 UTC
18
points
1
comment
19
min read
LW
link
(mailchi.mp)
Back to top
Next