RSS

ozhang

Karma: 368

ML Align­ment The­ory Pro­gram un­der Evan Hubinger

6 Dec 2021 0:03 UTC
82 points
3 comments2 min readLW link

AI Safety Newslet­ter #3: AI policy pro­pos­als and a new challenger approaches

ozhang25 Apr 2023 16:15 UTC
33 points
0 comments1 min readLW link

AI Safety Newslet­ter #4: AI and Cy­ber­se­cu­rity, Per­sua­sive AIs, Weaponiza­tion, and Ge­offrey Hin­ton talks AI risks

2 May 2023 18:41 UTC
32 points
0 comments5 min readLW link
(newsletter.safe.ai)

AI Safety Newslet­ter #2: ChaosGPT, Nat­u­ral Selec­tion, and AI Safety in the Media

18 Apr 2023 18:44 UTC
30 points
0 comments4 min readLW link
(newsletter.safe.ai)

$250K in Prizes: SafeBench Com­pe­ti­tion An­nounce­ment

ozhang3 Apr 2024 22:07 UTC
24 points
0 comments1 min readLW link