RSS

Cen­ter for Hu­man-Com­pat­i­ble AI (CHAI)

TagLast edit: 5 Feb 2023 21:04 UTC by _will_

The Center for Human-Compatible AI is a research institute at UC Berkeley, founded and led by Stuart Russell. Its stated objective is to prevent building unfriendly AI by focusing research on provably beneficial behaviour.

External links:

An­nounc­ing the Align­ment Newsletter

Rohin Shah9 Apr 2018 21:16 UTC
29 points
3 comments1 min readLW link

Our take on CHAI’s re­search agenda in un­der 1500 words

Alex Flint17 Jun 2020 12:24 UTC
112 points
18 comments5 min readLW link

CHAI In­tern­ship ap­pli­ca­tions are due by 12/​15

martinfukui13 Dec 2019 9:19 UTC
24 points
0 comments1 min readLW link

[AN #69] Stu­art Rus­sell’s new book on why we need to re­place the stan­dard model of AI

Rohin Shah19 Oct 2019 0:30 UTC
60 points
12 comments15 min readLW link
(mailchi.mp)

Stu­art Rus­sell: AI value al­ign­ment prob­lem must be an “in­trin­sic part” of the field’s main­stream agenda

Rob Bensinger26 Nov 2014 11:02 UTC
55 points
38 comments3 min readLW link

Re­but­tal piece by Stu­art Rus­sell and FHI Re­search As­so­ci­ate Allan Dafoe: “Yes, the ex­perts are wor­ried about the ex­is­ten­tial risk of ar­tifi­cial in­tel­li­gence.”

crmflynn3 Nov 2016 17:54 UTC
14 points
0 comments1 min readLW link
(www.technologyreview.com)

Stu­art Rus­sell’s Cen­ter for Hu­man Com­pat­i­ble AI is look­ing for an As­sis­tant Director

crmflynn25 Apr 2017 10:21 UTC
4 points
0 comments1 min readLW link
(humancompatible.ai)

De­bate on In­stru­men­tal Con­ver­gence be­tween LeCun, Rus­sell, Ben­gio, Zador, and More

Ben Pace4 Oct 2019 4:08 UTC
221 points
61 comments15 min readLW link2 reviews

Hu­man beats SOTA Go AI by learn­ing an ad­ver­sar­ial policy

Vanessa Kosoy19 Feb 2023 9:38 UTC
57 points
32 comments1 min readLW link
(goattack.far.ai)

Re­search in­tern­ship po­si­tion at CHAI

DanielFilan16 Jan 2018 6:25 UTC
10 points
3 comments1 min readLW link
(humancompatible.ai)

2019 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks19 Dec 2019 3:00 UTC
130 points
18 comments62 min readLW link

2018 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

Larks18 Dec 2018 4:46 UTC
190 points
26 comments62 min readLW link1 review

AXRP Epi­sode 8 - As­sis­tance Games with Dy­lan Had­field-Menell

DanielFilan8 Jun 2021 23:20 UTC
22 points
1 comment72 min readLW link

CHAI in­tern­ship ap­pli­ca­tions are open (due Nov 13)

Erik Jenner26 Oct 2023 0:53 UTC
34 points
0 comments3 min readLW link

Learn­ing prefer­ences by look­ing at the world

Rohin Shah12 Feb 2019 22:25 UTC
43 points
10 comments7 min readLW link
(bair.berkeley.edu)

The Align­ment Prob­lem Needs More Pos­i­tive Fiction

Netcentrica21 Aug 2022 22:01 UTC
5 points
2 comments5 min readLW link

UC Berkeley launches Cen­ter for Hu­man-Com­pat­i­ble Ar­tifi­cial Intelligence

ignoranceprior29 Aug 2016 22:43 UTC
15 points
1 comment2 min readLW link

Ten­sor Trust: An on­line game to un­cover prompt in­jec­tion vulnerabilities

1 Sep 2023 19:31 UTC
30 points
0 comments5 min readLW link
(tensortrust.ai)

Cur­rent AI Safety Roles for Soft­ware Engineers

ozziegooen9 Nov 2018 20:57 UTC
70 points
9 comments4 min readLW link

2017 AI Safety Liter­a­ture Re­view and Char­ity Com­par­i­son

Larks24 Dec 2017 18:52 UTC
41 points
5 comments23 min readLW link

CHAI In­tern­ship Application

martinfukui11 Nov 2020 21:10 UTC
31 points
0 comments1 min readLW link

Swim­ming Up­stream: A Case Study in In­stru­men­tal Rationality

TurnTrout3 Jun 2018 3:16 UTC
76 points
7 comments8 min readLW link

Ro­hin Shah on rea­sons for AI optimism

abergal31 Oct 2019 12:10 UTC
40 points
58 comments1 min readLW link
(aiimpacts.org)
No comments.