RSS

rosehadshar

Karma: 709

New re­port: A re­view of the em­piri­cal ev­i­dence for ex­is­ten­tial risk from AI via mis­al­igned power-seeking

4 Apr 2024 23:41 UTC
30 points
5 comments1 min readLW link
(blog.aiimpacts.org)

Re­sults from an Ad­ver­sar­ial Col­lab­o­ra­tion on AI Risk (FRI)

11 Mar 2024 20:00 UTC
60 points
3 comments9 min readLW link
(forecastingresearch.org)

[Question] Strongest real-world ex­am­ples sup­port­ing AI risk claims?

rosehadshar5 Sep 2023 15:12 UTC
41 points
7 comments1 min readLW link

Short timelines and slow, con­tin­u­ous take­off as the safest path to AGI

21 Jun 2023 8:56 UTC
62 points
15 comments7 min readLW link

The self-un­al­ign­ment problem

14 Apr 2023 12:10 UTC
144 points
22 comments10 min readLW link

Why Si­mu­la­tor AIs want to be Ac­tive In­fer­ence AIs

10 Apr 2023 18:23 UTC
86 points
8 comments8 min readLW link

Cur­rent UK gov­ern­ment lev­ers on AI development

rosehadshar10 Apr 2023 13:16 UTC
16 points
0 comments1 min readLW link

Les­sons from Con­ver­gent Evolu­tion for AI Alignment

27 Mar 2023 16:25 UTC
53 points
9 comments8 min readLW link

The space of sys­tems and the space of maps

22 Mar 2023 14:59 UTC
39 points
0 comments5 min readLW link

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

22 Feb 2023 16:09 UTC
103 points
9 comments6 min readLW link

What’s go­ing on with ‘crunch time’?

rosehadshar20 Jan 2023 9:42 UTC
53 points
6 comments4 min readLW link

In­ter­nal com­mu­ni­ca­tion framework

15 Nov 2022 12:41 UTC
38 points
14 comments12 min readLW link

The econ­omy as an anal­ogy for ad­vanced AI systems

15 Nov 2022 11:16 UTC
28 points
0 comments5 min readLW link