RSS

Zach Stein-Perlman(Zachary Stein-Perlman)

Karma: 4,576

AI strategy & governance. ailabwatch.org. Looking for new projects.

As of late May 2024, I’m focusing on blogging. In June I expect to focus on exploring a version of ailabwatch.org that could get more attention. I’m most excited to receive offers to help with projects like ailabwatch.org. I’m also excited to be pitched blogposts/​projects.

Ilya Sutskever and Jan Leike re­sign from OpenAI [up­dated]

Zach Stein-Perlman15 May 2024 0:45 UTC
246 points
95 comments3 min readLW link

AI com­pa­nies aren’t re­ally us­ing ex­ter­nal evaluators

Zach Stein-Perlman24 May 2024 16:01 UTC
231 points
12 comments4 min readLW link

In­tro­duc­ing AI Lab Watch

Zach Stein-Perlman30 Apr 2024 17:00 UTC
219 points
31 comments1 min readLW link
(ailabwatch.org)

Deep­Mind’s “​​Fron­tier Safety Frame­work” is weak and unambitious

Zach Stein-Perlman18 May 2024 3:00 UTC
156 points
13 comments4 min readLW link

Maybe An­thropic’s Long-Term Benefit Trust is powerless

Zach Stein-Perlman27 May 2024 13:00 UTC
143 points
11 comments2 min readLW link

FLI open let­ter: Pause gi­ant AI experiments

Zach Stein-Perlman29 Mar 2023 4:04 UTC
126 points
123 comments2 min readLW link
(futureoflife.org)

The pub­lic sup­ports reg­u­lat­ing AI for safety

Zach Stein-Perlman17 Feb 2023 4:10 UTC
114 points
9 comments1 min readLW link
(aiimpacts.org)

Deep­Mind: Model eval­u­a­tion for ex­treme risks

Zach Stein-Perlman25 May 2023 3:00 UTC
94 points
11 comments1 min readLW link
(arxiv.org)

Ques­tions for labs

Zach Stein-Perlman30 Apr 2024 22:15 UTC
76 points
11 comments8 min readLW link

New vol­un­tary com­mit­ments (AI Seoul Sum­mit)

Zach Stein-Perlman21 May 2024 11:00 UTC
75 points
16 comments7 min readLW link
(www.gov.uk)

OpenAI: Pre­pared­ness framework

Zach Stein-Perlman18 Dec 2023 18:30 UTC
70 points
23 comments4 min readLW link
(openai.com)

Deep­Mind: Fron­tier Safety Framework

Zach Stein-Perlman17 May 2024 17:30 UTC
64 points
0 comments3 min readLW link
(deepmind.google)

Deep­Mind: Eval­u­at­ing Fron­tier Models for Danger­ous Capabilities

Zach Stein-Perlman21 Mar 2024 3:00 UTC
61 points
1 comment1 min readLW link
(arxiv.org)

OpenAI-Microsoft partnership

Zach Stein-Perlman3 Oct 2023 20:01 UTC
51 points
18 comments1 min readLW link

Slow­ing AI: Read­ing list

Zach Stein-Perlman17 Apr 2023 14:30 UTC
45 points
3 comments4 min readLW link

Slow­ing AI: Foundations

Zach Stein-Perlman17 Apr 2023 14:30 UTC
45 points
11 comments17 min readLW link

[Question] Which pos­si­ble AI sys­tems are rel­a­tively safe?

Zach Stein-Perlman21 Aug 2023 17:00 UTC
42 points
20 comments1 min readLW link

ARC Evals: Re­spon­si­ble Scal­ing Policies

Zach Stein-Perlman28 Sep 2023 4:30 UTC
40 points
9 comments2 min readLW link
(evals.alignment.org)

Fram­ing AI strategy

Zach Stein-Perlman7 Feb 2023 19:20 UTC
33 points
1 comment18 min readLW link
(aiimpacts.org)

An­thropic, Google, Microsoft & OpenAI an­nounce Ex­ec­u­tive Direc­tor of the Fron­tier Model Fo­rum & over $10 mil­lion for a new AI Safety Fund

Zach Stein-Perlman25 Oct 2023 15:20 UTC
31 points
8 comments4 min readLW link
(www.frontiermodelforum.org)