RSS

peterbarnett

Karma: 1,260

Researcher at MIRI

EA and AI safety

https://​​peterbarnett.org/​​

Labs should be ex­plicit about why they are build­ing AGI

peterbarnett17 Oct 2023 21:09 UTC
187 points
16 comments1 min readLW link

Scott Aaron­son is join­ing OpenAI to work on AI safety

peterbarnett18 Jun 2022 4:06 UTC
117 points
31 comments1 min readLW link
(scottaaronson.blog)

Un­der­stand­ing Gra­di­ent Hacking

peterbarnett10 Dec 2021 15:58 UTC
41 points
5 comments30 min readLW link

When Should the Fire Alarm Go Off: A model for op­ti­mal thresholds

peterbarnett28 Apr 2021 12:27 UTC
40 points
4 comments5 min readLW link
(peterbarnett.org)

Fram­ings of De­cep­tive Alignment

peterbarnett26 Apr 2022 4:25 UTC
32 points
7 comments5 min readLW link

Align­ment Prob­lems All the Way Down

peterbarnett22 Jan 2022 0:19 UTC
26 points
7 comments11 min readLW link

A Story of AI Risk: In­struc­tGPT-N

peterbarnett26 May 2022 23:22 UTC
24 points
0 comments8 min readLW link

Try­ing to al­ign hu­mans with in­clu­sive ge­netic fitness

peterbarnett11 Jan 2024 0:13 UTC
23 points
5 comments10 min readLW link

Con­fu­sions in My Model of AI Risk

peterbarnett7 Jul 2022 1:05 UTC
22 points
9 comments5 min readLW link

How to be­come an AI safety researcher

peterbarnett15 Apr 2022 11:41 UTC
22 points
0 comments14 min readLW link

Why I’m Wor­ried About AI

peterbarnett23 May 2022 21:13 UTC
22 points
2 comments12 min readLW link