Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
Max Harms
Karma:
819
Also known as Raelifin:
https://www.lesswrong.com/users/raelifin
All
Posts
Comments
New
Top
Old
Thoughts on AI 2027
Max Harms
Apr 9, 2025, 9:26 PM
219
points
61
comments
21
min read
LW
link
(intelligence.org)
Instrumental vs Terminal Desiderata
Max Harms
Jun 26, 2024, 8:57 PM
21
points
0
comments
3
min read
LW
link
Max Harms’s Shortform
Max Harms
Jun 13, 2024, 6:19 PM
3
points
1
comment
LW
link
5. Open Corrigibility Questions
Max Harms
Jun 10, 2024, 2:09 PM
30
points
0
comments
7
min read
LW
link
4. Existing Writing on Corrigibility
Max Harms
Jun 10, 2024, 2:08 PM
55
points
15
comments
106
min read
LW
link
3b. Formal (Faux) Corrigibility
Max Harms
Jun 9, 2024, 5:18 PM
26
points
13
comments
17
min read
LW
link
3a. Towards Formal Corrigibility
Max Harms
Jun 9, 2024, 4:53 PM
24
points
2
comments
19
min read
LW
link
2. Corrigibility Intuition
Max Harms
Jun 8, 2024, 3:52 PM
67
points
10
comments
33
min read
LW
link
1. The CAST Strategy
Max Harms
Jun 7, 2024, 10:29 PM
48
points
22
comments
38
min read
LW
link
0. CAST: Corrigibility as Singular Target
Max Harms
Jun 7, 2024, 10:29 PM
147
points
17
comments
8
min read
LW
link
Back to top