Ver­ifi­ca­tion and Transparency

DanielFilanAug 8, 2019, 1:50 AM
35 points
6 comments2 min readLW link
(danielfilan.com)

AI Align­ment Open Thread Au­gust 2019

habrykaAug 4, 2019, 10:09 PM
35 points
96 comments1 min readLW link

AI Fore­cast­ing Re­s­olu­tion Coun­cil (Fore­cast­ing in­fras­truc­ture, part 2)

Aug 29, 2019, 5:35 PM
35 points
2 comments3 min readLW link

[Question] What au­thors con­sis­tently give ac­cu­rate pic­tures of com­plex top­ics they dis­cuss?

seezAug 21, 2019, 12:09 AM
34 points
3 comments1 min readLW link

[Site Up­date] Weekly/​Monthly/​Yearly on All Posts

RaemonAug 2, 2019, 12:39 AM
33 points
7 comments1 min readLW link

“Can We Sur­vive Tech­nol­ogy” by von Neumann

Ben PaceAug 18, 2019, 6:58 PM
33 points
2 comments1 min readLW link
(geosci.uchicago.edu)

[Question] What ex­per­i­ments would demon­strate “up­per limits of aug­mented work­ing mem­ory?”

RaemonAug 15, 2019, 10:09 PM
33 points
6 comments2 min readLW link

AI Align­ment Writ­ing Day Roundup #1

Ben PaceAug 30, 2019, 1:26 AM
32 points
12 comments1 min readLW link

Cal­ibrat­ing With Cards

lifelonglearnerAug 8, 2019, 6:44 AM
32 points
3 comments3 min readLW link

Dis­tance Func­tions are Hard

Grue_SlinkyAug 13, 2019, 5:33 PM
31 points
19 comments6 min readLW link

Don’t Pull a Bro­ken Chain

johnswentworthAug 28, 2019, 1:21 AM
31 points
6 comments5 min readLW link

[Question] What ex­plana­tory power does Kah­ne­man’s Sys­tem 2 pos­sess?

Richard_NgoAug 12, 2019, 3:23 PM
31 points
2 comments1 min readLW link

When do util­ity func­tions con­strain?

HoagyAug 23, 2019, 5:19 PM
30 points
8 comments7 min readLW link

Self-Su­per­vised Learn­ing and AGI Safety

Steven ByrnesAug 7, 2019, 2:21 PM
30 points
9 comments12 min readLW link

Help fore­cast study repli­ca­tion in this so­cial sci­ence pre­dic­tion market

rosiecamAug 7, 2019, 6:18 PM
29 points
3 comments1 min readLW link

A Sur­vey of Early Im­pact Measures

Matthew BarnettAug 6, 2019, 1:22 AM
29 points
0 comments8 min readLW link

[Question] Could we solve this email mess if we all moved to paid emails?

Bird ConceptAug 11, 2019, 4:31 PM
29 points
50 comments4 min readLW link

In­spec­tion Para­dox as a Driver of Group Separation

ShmiAug 17, 2019, 9:47 PM
29 points
0 comments1 min readLW link

[Question] What are the rea­sons to *not* con­sider re­duc­ing AI-Xrisk the high­est pri­or­ity cause?

David Scott Krueger (formerly: capybaralet)Aug 20, 2019, 9:45 PM
29 points
27 comments1 min readLW link

Pre­dicted AI al­ign­ment event/​meet­ing calendar

rmoehnAug 14, 2019, 7:14 AM
29 points
14 comments1 min readLW link

An­nounce­ment: Writ­ing Day To­day (Thurs­day)

Ben PaceAug 22, 2019, 4:48 AM
29 points
5 comments1 min readLW link

GPT-2: 6-Month Fol­low-Up

lifelonglearnerAug 21, 2019, 5:06 AM
28 points
1 comment1 min readLW link

“De­sign­ing agent in­cen­tives to avoid re­ward tam­per­ing”, DeepMind

gwernAug 14, 2019, 4:57 PM
28 points
15 commentsLW link
(medium.com)

[AN #62] Are ad­ver­sar­ial ex­am­ples caused by real but im­per­cep­ti­ble fea­tures?

Rohin ShahAug 22, 2019, 5:10 PM
28 points
10 comments9 min readLW link
(mailchi.mp)

Al­gorith­mic Similarity

LukasMAug 23, 2019, 4:39 PM
28 points
10 comments11 min readLW link

[Question] What is the state of the ego de­ple­tion field?

Eli TyreAug 9, 2019, 8:30 PM
27 points
10 comments1 min readLW link

[Question] Why are the peo­ple who could be do­ing safety re­search, but aren’t, do­ing some­thing else?

Adam SchollAug 29, 2019, 8:51 AM
27 points
19 comments1 min readLW link

Raph Koster on Vir­tual Wor­lds vs Games (notes)

RaemonAug 18, 2019, 7:01 PM
26 points
8 comments2 min readLW link

Rev­ersible changes: con­sider a bucket of water

Stuart_ArmstrongAug 26, 2019, 10:55 PM
25 points
18 comments2 min readLW link

Pro­ject Pro­posal: Con­sid­er­a­tions for trad­ing off ca­pa­bil­ities and safety im­pacts of AI research

David Scott Krueger (formerly: capybaralet)Aug 6, 2019, 10:22 PM
25 points
11 comments2 min readLW link

In­ver­sion of the­o­rems into defi­ni­tions when generalizing

riceissaAug 4, 2019, 5:44 PM
25 points
3 comments5 min readLW link

Good­hart’s Curse and Limi­ta­tions on AI Alignment

Gordon Seidoh WorleyAug 19, 2019, 7:57 AM
25 points
18 comments10 min readLW link

Why Gra­di­ents Van­ish and Explode

Matthew BarnettAug 9, 2019, 2:54 AM
25 points
9 comments3 min readLW link

Which of these five AI al­ign­ment re­search pro­jects ideas are no good?

rmoehnAug 8, 2019, 7:17 AM
25 points
13 comments1 min readLW link

[Question] Why do hu­mans not have built-in neu­ral i/​o chan­nels?

Richard_NgoAug 8, 2019, 1:09 PM
25 points
23 comments1 min readLW link

Nega­tive “eeny meeny miny moe”

jefftkAug 20, 2019, 2:48 AM
25 points
6 comments1 min readLW link

A Primer on Ma­trix Calcu­lus, Part 1: Ba­sic review

Matthew BarnettAug 12, 2019, 11:44 PM
25 points
4 comments7 min readLW link

Emo­tions are not beliefs

Chris_LeongAug 7, 2019, 6:27 AM
25 points
2 comments2 min readLW link

Im­pli­ca­tions of Quan­tum Com­put­ing for Ar­tifi­cial In­tel­li­gence Align­ment Research

Aug 22, 2019, 10:33 AM
24 points
3 comments13 min readLW link

Un­der­stand­ing understanding

mthqAug 23, 2019, 6:10 PM
24 points
1 comment2 min readLW link

July 2019 gw­ern.net newsletter

gwernAug 1, 2019, 4:19 PM
23 points
0 commentsLW link
(www.gwern.net)

[Site Up­date] Be­hind the scenes data-layer and caching improvements

habrykaAug 7, 2019, 12:49 AM
23 points
3 comments1 min readLW link

Car­to­graphic Processes

johnswentworthAug 27, 2019, 8:02 PM
23 points
3 comments4 min readLW link

[Question] Do you do weekly or daily re­views? What are they like?

benwrAug 5, 2019, 1:23 AM
23 points
8 comments1 min readLW link

Prac­ti­cal con­se­quences of im­pos­si­bil­ity of value learning

Stuart_Armstrong2 Aug 2019 23:06 UTC
23 points
13 comments3 min readLW link

A Primer on Ma­trix Calcu­lus, Part 2: Ja­co­bi­ans and other fun

Matthew Barnett15 Aug 2019 1:13 UTC
22 points
7 comments7 min readLW link

In defense of Or­a­cle (“Tool”) AI research

Steven Byrnes7 Aug 2019 19:14 UTC
22 points
11 comments4 min readLW link

Four Ways An Im­pact Mea­sure Could Help Alignment

Matthew Barnett8 Aug 2019 0:10 UTC
21 points
1 comment9 min readLW link

[Question] Is LW mak­ing progress?

zulupineapple24 Aug 2019 0:32 UTC
21 points
11 comments1 min readLW link

Prob­lems with AI debate

Stuart_Armstrong26 Aug 2019 19:21 UTC
21 points
3 comments5 min readLW link