Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
DanielFilan
Karma:
4,862
All
Posts
Comments
New
Top
Old
Page
1
Bottle Caps Aren’t Optimisers
DanielFilan
31 Aug 2018 18:30 UTC
76
points
21
comments
3
min read
LW
link
1
review
(danielfilan.com)
Announcing the Vitalik Buterin Fellowships in AI Existential Safety!
DanielFilan
21 Sep 2021 0:33 UTC
64
points
2
comments
1
min read
LW
link
(grants.futureoflife.org)
Test Cases for Impact Regularisation Methods
DanielFilan
6 Feb 2019 21:50 UTC
58
points
5
comments
12
min read
LW
link
(danielfilan.com)
AXRP Episode 9 - Finite Factored Sets with Scott Garrabrant
DanielFilan
24 Jun 2021 22:10 UTC
56
points
2
comments
58
min read
LW
link
Mechanistic Transparency for Machine Learning
DanielFilan
11 Jul 2018 0:34 UTC
54
points
9
comments
4
min read
LW
link
Security Mindset and Takeoff Speeds
DanielFilan
27 Oct 2020 3:20 UTC
54
points
23
comments
8
min read
LW
link
(danielfilan.com)
Announcing AXRP, the AI X-risk Research Podcast
DanielFilan
23 Dec 2020 20:00 UTC
54
points
6
comments
1
min read
LW
link
(danielfilan.com)
A Personal Rationality Wishlist
DanielFilan
27 Aug 2019 3:40 UTC
53
points
54
comments
4
min read
LW
link
(danielfilan.com)
An Analytic Perspective on AI Alignment
DanielFilan
1 Mar 2020 4:10 UTC
53
points
45
comments
8
min read
LW
link
(danielfilan.com)
Challenge: know everything that the best go bot knows about go
DanielFilan
11 May 2021 5:10 UTC
48
points
93
comments
2
min read
LW
link
(danielfilan.com)
A second example of conditional orthogonality in finite factored sets
DanielFilan
7 Jul 2021 1:40 UTC
46
points
0
comments
2
min read
LW
link
(danielfilan.com)
Cognitive mistakes I’ve made about COVID-19
DanielFilan
27 Dec 2020 0:50 UTC
45
points
3
comments
2
min read
LW
link
(danielfilan.com)
A simple example of conditional orthogonality in finite factored sets
DanielFilan
6 Jul 2021 0:36 UTC
43
points
3
comments
5
min read
LW
link
(danielfilan.com)
AXRP Episode 4 - Risks from Learned Optimization with Evan Hubinger
DanielFilan
18 Feb 2021 0:03 UTC
41
points
10
comments
86
min read
LW
link
What’s the chance a smart London resident dies of a Russian nuke in the next month?
DanielFilan
10 Mar 2022 19:20 UTC
40
points
8
comments
4
min read
LW
link
(danielfilan.com)
[LINK] Scott Aaronson on Integrated Information Theory
DanielFilan
22 May 2014 8:40 UTC
38
points
11
comments
1
min read
LW
link
Insights from ‘The Strategy of Conflict’
DanielFilan
4 Jan 2018 5:05 UTC
37
points
13
comments
7
min read
LW
link
AXRP Episode 12 - AI Existential Risk with Paul Christiano
DanielFilan
2 Dec 2021 2:20 UTC
36
points
0
comments
125
min read
LW
link
Verification and Transparency
DanielFilan
8 Aug 2019 1:50 UTC
34
points
6
comments
2
min read
LW
link
(danielfilan.com)
AXRP Episode 7 - Side Effects with Victoria Krakovna
DanielFilan
14 May 2021 3:50 UTC
34
points
6
comments
43
min read
LW
link
Back to top
Next