RSS

Vika

Karma: 3,041

Victoria Krakovna. Research scientist at DeepMind working on AI safety, and cofounder of the Future of Life Institute. Website and blog: vkrakovna.wordpress.com

Deep­Mind al­ign­ment team opinions on AGI ruin arguments

Vika12 Aug 2022 21:06 UTC
376 points
37 comments14 min readLW link1 review

Pos­si­ble take­aways from the coro­n­avirus pan­demic for slow AI takeoff

Vika31 May 2020 17:51 UTC
135 points
36 comments3 min readLW link1 review

[Linkpost] Some high-level thoughts on the Deep­Mind al­ign­ment team’s strategy

7 Mar 2023 11:55 UTC
128 points
13 comments5 min readLW link
(drive.google.com)

Strate­gic choice of identity

Vika8 Mar 2014 16:27 UTC
119 points
58 comments2 min readLW link

When dis­cussing AI risks, talk about ca­pa­bil­ities, not intelligence

Vika11 Aug 2023 13:38 UTC
116 points
7 comments3 min readLW link
(vkrakovna.wordpress.com)

Refin­ing the Sharp Left Turn threat model, part 1: claims and mechanisms

12 Aug 2022 15:17 UTC
85 points
4 comments3 min readLW link1 review
(vkrakovna.wordpress.com)

Op­ti­miza­tion Con­cepts in the Game of Life

16 Oct 2021 20:51 UTC
74 points
16 comments11 min readLW link

Clas­sify­ing speci­fi­ca­tion prob­lems as var­i­ants of Good­hart’s Law

Vika19 Aug 2019 20:40 UTC
72 points
5 comments5 min readLW link1 review

New or­ga­ni­za­tion—Fu­ture of Life In­sti­tute (FLI)

Vika14 Jun 2014 23:00 UTC
70 points
35 comments1 min readLW link

Speci­fi­ca­tion gam­ing: the flip side of AI ingenuity

6 May 2020 23:51 UTC
65 points
9 comments6 min readLW link

Power-seek­ing can be prob­a­ble and pre­dic­tive for trained agents

28 Feb 2023 21:10 UTC
56 points
22 comments9 min readLW link
(arxiv.org)

Paradigms of AI al­ign­ment: com­po­nents and enablers

Vika2 Jun 2022 6:19 UTC
53 points
4 comments8 min readLW link

Mov­ing on from com­mu­nity living

Vika17 Apr 2024 17:02 UTC
49 points
7 comments3 min readLW link
(vkrakovna.wordpress.com)

Speci­fi­ca­tion gam­ing ex­am­ples in AI

Vika3 Apr 2018 12:30 UTC
45 points
9 comments1 min readLW link2 reviews

New Deep­Mind AI Safety Re­search Blog

Vika27 Sep 2018 16:28 UTC
43 points
0 comments1 min readLW link
(medium.com)

Refin­ing the Sharp Left Turn threat model, part 2: ap­ply­ing al­ign­ment techniques

25 Nov 2022 14:36 UTC
39 points
9 comments6 min readLW link
(vkrakovna.wordpress.com)

To con­tribute to AI safety, con­sider do­ing AI research

Vika16 Jan 2016 20:42 UTC
39 points
39 comments2 min readLW link

Trade­off be­tween de­sir­able prop­er­ties for baseline choices in im­pact measures

Vika4 Jul 2020 11:56 UTC
37 points
24 comments5 min readLW link

Fu­ture of Life In­sti­tute ex­is­ten­tial risk news site

Vika19 Mar 2015 14:33 UTC
35 points
2 comments1 min readLW link