RSS

peterbarnett

Karma: 158

EA and AI safety

AI safety meta research at Nonlinear

Theoretical Alignment research at SERI MATS

https://​​peterbarnett.org/​​

Fram­ings of De­cep­tive Alignment

peterbarnett26 Apr 2022 4:25 UTC
15 points
6 comments5 min readLW link

How to be­come an AI safety researcher

peterbarnett15 Apr 2022 11:41 UTC
10 points
0 comments14 min readLW link

Thoughts on Danger­ous Learned Optimization

peterbarnett19 Feb 2022 10:46 UTC
4 points
2 comments4 min readLW link

pe­ter­bar­nett’s Shortform

peterbarnett16 Feb 2022 17:24 UTC
3 points
1 comment1 min readLW link

Align­ment Prob­lems All the Way Down

peterbarnett22 Jan 2022 0:19 UTC
25 points
7 comments10 min readLW link

[Question] What ques­tions do you have about do­ing work on AI safety?

peterbarnett21 Dec 2021 16:36 UTC
13 points
8 comments1 min readLW link

Some mo­ti­va­tions to gra­di­ent hack

peterbarnett17 Dec 2021 3:06 UTC
6 points
0 comments6 min readLW link

Un­der­stand­ing Gra­di­ent Hacking

peterbarnett10 Dec 2021 15:58 UTC
26 points
5 comments30 min readLW link