Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
So8res
(Nate Soares)
Karma:
13,792
All
Posts
Comments
New
Top
Old
Page
1
A mind needn’t be curious to reap the benefits of curiosity
So8res
2 Jun 2023 18:00 UTC
78
points
14
comments
1
min read
LW
link
Cosmopolitan values don’t come free
So8res
31 May 2023 15:58 UTC
127
points
81
comments
1
min read
LW
link
Sentience matters
So8res
29 May 2023 21:25 UTC
129
points
93
comments
2
min read
LW
link
Request: stop advancing AI capabilities
So8res
26 May 2023 17:42 UTC
150
points
23
comments
1
min read
LW
link
Would we even want AI to solve all our problems?
So8res
21 Apr 2023 18:04 UTC
96
points
15
comments
2
min read
LW
link
How could you possibly choose what an AI wants?
So8res
19 Apr 2023 17:08 UTC
104
points
19
comments
1
min read
LW
link
But why would the AI kill us?
So8res
17 Apr 2023 18:42 UTC
112
points
86
comments
2
min read
LW
link
Misgeneralization as a misnomer
So8res
6 Apr 2023 20:43 UTC
126
points
21
comments
4
min read
LW
link
If interpretability research goes well, it may get dangerous
So8res
3 Apr 2023 21:48 UTC
197
points
10
comments
2
min read
LW
link
Hooray for stepping out of the limelight
So8res
1 Apr 2023 2:45 UTC
278
points
23
comments
1
min read
LW
link
A rough and incomplete review of some of John Wentworth’s research
So8res
28 Mar 2023 18:52 UTC
168
points
16
comments
18
min read
LW
link
A stylized dialogue on John Wentworth’s claims about markets and optimization
So8res
25 Mar 2023 22:32 UTC
156
points
21
comments
8
min read
LW
link
Truth and Advantage: Response to a draft of “AI safety seems hard to measure”
So8res
22 Mar 2023 3:36 UTC
98
points
9
comments
5
min read
LW
link
Deep Deceptiveness
So8res
21 Mar 2023 2:51 UTC
215
points
56
comments
14
min read
LW
link
Comments on OpenAI’s “Planning for AGI and beyond”
So8res
3 Mar 2023 23:01 UTC
148
points
2
comments
14
min read
LW
link
Enemies vs Malefactors
So8res
28 Feb 2023 23:38 UTC
200
points
61
comments
1
min read
LW
link
AI alignment researchers don’t (seem to) stack
So8res
21 Feb 2023 0:48 UTC
182
points
38
comments
3
min read
LW
link
Hashing out long-standing disagreements seems low-value to me
So8res
16 Feb 2023 6:20 UTC
126
points
33
comments
4
min read
LW
link
Focus on the places where you feel shocked everyone’s dropping the ball
So8res
2 Feb 2023 0:27 UTC
385
points
59
comments
4
min read
LW
link
What I mean by “alignment is in large part about making cognition aimable at all”
So8res
30 Jan 2023 15:22 UTC
151
points
24
comments
2
min read
LW
link
Back to top
Next