So8res(Nate Soares)

Karma: 13,792

A mind needn’t be cu­ri­ous to reap the benefits of curiosity

So8res2 Jun 2023 18:00 UTC
78 points
14 comments1 min readLW link

Cos­mopoli­tan val­ues don’t come free

So8res31 May 2023 15:58 UTC
127 points
81 comments1 min readLW link

Sen­tience matters

So8res29 May 2023 21:25 UTC
129 points
93 comments2 min readLW link

Re­quest: stop ad­vanc­ing AI capabilities

So8res26 May 2023 17:42 UTC
150 points
23 comments1 min readLW link

Would we even want AI to solve all our prob­lems?

So8res21 Apr 2023 18:04 UTC
96 points
15 comments2 min readLW link

How could you pos­si­bly choose what an AI wants?

So8res19 Apr 2023 17:08 UTC
104 points
19 comments1 min readLW link

But why would the AI kill us?

So8res17 Apr 2023 18:42 UTC
112 points
86 comments2 min readLW link

Mis­gen­er­al­iza­tion as a misnomer

So8res6 Apr 2023 20:43 UTC
126 points
21 comments4 min readLW link

If in­ter­pretabil­ity re­search goes well, it may get dangerous

So8res3 Apr 2023 21:48 UTC
197 points
10 comments2 min readLW link

Hooray for step­ping out of the limelight

So8res1 Apr 2023 2:45 UTC
278 points
23 comments1 min readLW link

A rough and in­com­plete re­view of some of John Went­worth’s research

So8res28 Mar 2023 18:52 UTC
168 points
16 comments18 min readLW link

A stylized di­alogue on John Went­worth’s claims about mar­kets and optimization

So8res25 Mar 2023 22:32 UTC
156 points
21 comments8 min readLW link

Truth and Ad­van­tage: Re­sponse to a draft of “AI safety seems hard to mea­sure”

So8res22 Mar 2023 3:36 UTC
98 points
9 comments5 min readLW link

Deep Deceptiveness

So8res21 Mar 2023 2:51 UTC
215 points
56 comments14 min readLW link

Com­ments on OpenAI’s “Plan­ning for AGI and be­yond”

So8res3 Mar 2023 23:01 UTC
148 points
2 comments14 min readLW link

Ene­mies vs Malefactors

So8res28 Feb 2023 23:38 UTC
200 points
61 comments1 min readLW link

AI al­ign­ment re­searchers don’t (seem to) stack

So8res21 Feb 2023 0:48 UTC
182 points
38 comments3 min readLW link

Hash­ing out long-stand­ing dis­agree­ments seems low-value to me

So8res16 Feb 2023 6:20 UTC
126 points
33 comments4 min readLW link

Fo­cus on the places where you feel shocked ev­ery­one’s drop­ping the ball

So8res2 Feb 2023 0:27 UTC
385 points
59 comments4 min readLW link

What I mean by “al­ign­ment is in large part about mak­ing cog­ni­tion aimable at all”

So8res30 Jan 2023 15:22 UTC
151 points
24 comments2 min readLW link