Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
Max Harms
Karma:
1,580
Also known as Raelifin:
https://www.lesswrong.com/users/raelifin
All
Posts
Comments
New
Top
Old
Serious Flaws in CAST
Max Harms
19 Nov 2025 17:27 UTC
105
points
10
comments
8
min read
LW
link
AI Corrigibility Debate: Max Harms vs. Jeremy Gillen
Liron
,
Max Harms
and
Jeremy Gillen
14 Nov 2025 4:09 UTC
46
points
1
comment
75
min read
LW
link
(doomdebates.com)
Worlds Where Iterative Design Succeeds?
Max Harms
23 Oct 2025 22:14 UTC
23
points
5
comments
8
min read
LW
link
Any corrigibility naysayers outside of MIRI?
Max Harms
22 Oct 2025 21:26 UTC
28
points
24
comments
1
min read
LW
link
Contra Collier on IABIED
Max Harms
20 Sep 2025 15:55 UTC
235
points
51
comments
20
min read
LW
link
Thoughts on AI 2027
Max Harms
9 Apr 2025 21:26 UTC
223
points
61
comments
21
min read
LW
link
(intelligence.org)
Instrumental vs Terminal Desiderata
Max Harms
26 Jun 2024 20:57 UTC
21
points
0
comments
3
min read
LW
link
Max Harms’s Shortform
Max Harms
13 Jun 2024 18:19 UTC
3
points
31
comments
1
min read
LW
link
5. Open Corrigibility Questions
Max Harms
10 Jun 2024 14:09 UTC
31
points
0
comments
7
min read
LW
link
4. Existing Writing on Corrigibility
Max Harms
10 Jun 2024 14:08 UTC
64
points
19
comments
106
min read
LW
link
3b. Formal (Faux) Corrigibility
Max Harms
9 Jun 2024 17:18 UTC
26
points
19
comments
17
min read
LW
link
3a. Towards Formal Corrigibility
Max Harms
9 Jun 2024 16:53 UTC
28
points
2
comments
19
min read
LW
link
2. Corrigibility Intuition
Max Harms
8 Jun 2024 15:52 UTC
74
points
10
comments
33
min read
LW
link
1. The CAST Strategy
Max Harms
7 Jun 2024 22:29 UTC
57
points
24
comments
38
min read
LW
link
0. CAST: Corrigibility as Singular Target
Max Harms
7 Jun 2024 22:29 UTC
156
points
19
comments
9
min read
LW
link
Back to top